09.11.2012 Views

friday morning, 1 december 2006 lanai room, 7 - DTU Orbit

friday morning, 1 december 2006 lanai room, 7 - DTU Orbit

friday morning, 1 december 2006 lanai room, 7 - DTU Orbit

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

FRIDAY MORNING, 1 DECEMBER <strong>2006</strong> LANAI ROOM, 7:30 TO 11:55 A.M.<br />

Session 4aAA<br />

Architectural Acoustics: Measurement of Room Acoustics I<br />

Fumiaki Satoh, Cochair<br />

Chiba Inst. of Technology, Tsudanuma 2-17-1 Narashino-shi, Chiba 275-0016, Japan<br />

Boaz Rafaely, Cochair<br />

Ben Gurion Univ., Electrical and Computer Engineering Dept., 84105, Beer Sheva, Israel<br />

Chair’s Introduction—7:30<br />

Invited Papers<br />

7:35<br />

4aAA1. Warped-time-stretched pulse: An acoustic test signal robust against ambient noise. Masanori Morise, Toshio Irino,<br />

Hideki Banno, and Hideki Kawahara �Wakayama Univ., 930, Sakaedani, Wakayama, 640-8510, Japan,<br />

s055068@sys.wakayama-u.ac.jp�<br />

A new acoustic measurement signal that is a hybrid signal of time-stretched pulse �TSP�, or lin-TSP, and logarithmic TSP<br />

�log-TSP� is proposed. The signal, referred to as warped-TSP �Morise et al., IEICE Trans. Fundamentals, A, J89-A�1�, 7–14 �<strong>2006</strong>��,<br />

has a single parameter to adjust for better measurements in accordance with ambient noise conditions. It also provides a means to<br />

eliminate harmonic distortions produced mainly by loudspeaker systems. In this lecture, the definition and features of the warped-TSP<br />

in comparison with the lin-TSP and log-TSP are introduced. The following were shown: �1� the relationship between the parameters,<br />

the amplitude frequency characteristics, and the effect on the harmonic distortion components; �2� a method to select the optimal<br />

parameters of the warped-TSP for a specific measuring environment; and �3� the experimental results for a series of impulse response<br />

measurements under different ambient noise conditions. Those results show that the proposed method outperformed the lin-TSP and<br />

log-TSP under all conditions in terms of SNR of the measured impulse response. �This research was supported partly by grants-in-aid<br />

for scientific research �15300061 and 15650032� and a grant from the Faculty of Systems Engineering at Wakayama University.�<br />

7:55<br />

4aAA2. Simultaneous estimation of reverberation times and their uncertainties from <strong>room</strong> impulse responses using a<br />

single-measurement procedure. Ning Xiang and Tomislav Jasa �Grad. Program in Architecture Acoust., and Dept. of Elec.,<br />

Comput., and Systems Eng., Rensselaer Polyt. Inst, Troy, NY 12180�<br />

Accurate measurements of reverberation times are of fundamental importance in <strong>room</strong> acoustics. A number of test procedures for<br />

characterizing acoustics in performing arts venues, quantifying acoustic properties of materials in chamber measurements, rely on<br />

experimental determination of reverberation times. In addition, decay-time estimation in acoustically coupled spaces has been found<br />

to be very demanding. Our recent work has demonstrated that model-based Bayesian approaches �Xiang et al., J. Acoust. Soc. Am.<br />

110, 1415–1424 �2001�; 113, 2685–2697 �2003�; 117, 3705–3715 �2005�� can be very useful for such analysis in architectural<br />

acoustics measurements. This paper discusses the recent development of probabilistic tools for estimating both reverberation �decay�<br />

times and their uncertainties within Bayesian framework. This work shows that Bayesian probabilistic inference can be used as a<br />

useful tool for sound energy decay analysis in both single-space halls and coupled spaces. Bayesian decay analysis simultaneously<br />

provides architectural acousticians with reverberation times, diverse decay times, related derivations, and interdependencies to quantify<br />

uncertainties of the estimation from a single measurement of <strong>room</strong> impulse responses followed by Schroeder backward integrations.<br />

8:15<br />

4aAA3. Permissible number of synchronous averaging times to obtain reverberation time from impulse response under<br />

time-variance conditions. Fumiaki Satoh, Yukiteru Hayashi �Chiba Inst. of Technol., Tsudanuma 2-17-1, Narashino-shi, Chiba,<br />

275-0016, Japan�, Shinichi Sakamoto �Univ. of Tokyo, Meguro-ku, Tokyo, 153-8505, Japan�, and Hideki Tachibana �Chiba Inst. of<br />

Technol., Narashino-shi, Chiba, 275-0016, Japan�<br />

In the measurement of <strong>room</strong> impulse response, the synchronous averaging technique and such new methods as the MLS and the<br />

swept-sine methods are being widely used to improve the signal-to-noise ratio. In actual measurement conditions, however, the air in<br />

a <strong>room</strong> is continuously moving and the temperature is changing to some degree. The measured value of the reverberation time in such<br />

a <strong>room</strong> tends to be shorter at higher frequencies when applying the synchronous averaging. Therefore, the assumption of a time<br />

invariant has to be carefully considered, and, on this point, some research has been conducted to date. We also have reported various<br />

research results concerning the impulse response measurement under the time-variance conditions. In this paper, the permissible<br />

number of synchronous averaging times for reverberation measurement is studied through some field experiments. In each field, many<br />

3223 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3223<br />

4a FRI. AM


time impulse response measurements were taken between a fixed pair of sound source and receiving positions by the swept-sine<br />

method, without averaging. After the measurements, the characteristics and the extent of the time-variance under measuring were<br />

estimated by a short-term running cross-correlation function between each impulse response. The influence of the time variance on the<br />

synchronous averaging result was studied based on the estimated time variance.<br />

8:35<br />

4aAA4. Selection of receiving positions suitable for evaluating acoustical parameters. Taeko Akama, Hisaharu Suzuki, and<br />

Akira Omoto �Omoto Lab., Dept. of Acoust. Design, Faculty of Design, Kyushu Univ., Shiobaru 4-9-1, Minami, Fukuoka 811-8540,<br />

Japan�<br />

Many physical parameters show characteristics of large sound fields such as concert halls. Some of them are adopted in the Annex<br />

of ISO 3382. That definition is clearly provided in ISO. However, practical measurement methods for them remain obscure. Our<br />

research is intended to examine an effective selection method of receiving positions based on the distribution of acoustical parameters<br />

in a real field. For that purpose, impulse responses are measured at more than 1400 seat positions to elucidate the distribution of<br />

acoustical parameters in an existing concert hall. The acoustical parameters, which are reverberation time, early decay time, clarity,<br />

and center time at each seat, are then calculated for 500-Hz, 1-kHz, and 2-kHz octave bands. The distributions of reverberation time<br />

are quite even at all seats. However, the distributions of other parameters show symmetrical patterns at 500 Hz. At 1 and 2 kHz<br />

frequencies, the distributions show asymmetrical patterns in this hall. Based on the results obtained in this study, an effective method<br />

to select the receiving position can be proposed.<br />

8:55<br />

4aAA5. Hybrid measurement method in <strong>room</strong> acoustics using dodecahedron speakers and a subwoofer. Hideo Miyazaki �Ctr.<br />

for Adv. Sound Technologies, Yamaha Corp., 203 Matsunokijima, Iwata, Shizuoka 438-0192, miyazaki@beat.yamaha.co.jp�<br />

A dodecahedron speaker is usually utilized for measurement in <strong>room</strong> acoustics under the hypothesis of omni directional point<br />

source. But generally speakers used for a dodecahedron speaker cannot playback low-frequency sound such as under 100 Hz, which<br />

is important especially for auralization, while the one constructed of units with large diameter to support low-frequency sounds cannot<br />

be considered as an omni-directional speaker in high frequencies. To meet these requirements, a hybrid system combining a dodecahedron<br />

speaker and a subwoofer has been developed and actually used for measurements of impulse responses in acoustical design of<br />

concert halls. The summary of this method will be presented. The feasibility of this method will be also discussed while evaluating the<br />

measurement results in concert halls by changing measurement conditions such as speaker locations and comparing these results with<br />

those of conventional methods.<br />

9:15<br />

4aAA6. The perception of apparent source width and its dependence on frequency and loudness. Ingo B. Witew and Johannes<br />

A. Buechler �Inst. of Tech. Acoust., RWTH Aachen Univ., Templergraben 55, 52066 Aachen, Germany�<br />

While it is widely accepted that apparent source width �ASW� is an important factor in characterizing the acoustics of a concert<br />

hall, there is still a lively discussion on how to refine the physical measures for ASW. A lot of experience has been gathered with<br />

interaural-cross-correlation and lateral-sound-incidence measures during the last years. As a result it was learned that different<br />

frequencies contribute differently to the perception of ASW and that the level of a sound also influences the perception of the apparent<br />

width of a source. With many technical measures having an influence on the perceptual aspect of ASW, the design of psychometric<br />

experiments becomes challenging as it is desirable to avoid the interaction of different objective parameters. In the experiments for the<br />

study presented, the perception of ASW is investigated for frequencies ranging from 100 Hz to 12.5 kHz at different levels of<br />

loudness. It is shown how the frequency and the loudness of a sound influence the perception of ASW.<br />

9:35<br />

4aAA7. Sound source with adjustable directivity. Gottfried K. Behler �Inst. fuer Technische Akustik, RWTH Aachen Univ.,<br />

D-52056 Aachen, Germany�<br />

Omni-directional sound sources are used to measure <strong>room</strong>-acoustical parameters in accordance with ISO 3382. To record a<br />

detailed <strong>room</strong> impulse response �RIR� with the aim of auralization, an extended frequency range is required that is not covered by the<br />

often-used building acoustics sound sources. To obtain this target, a loudspeaker with dedicated sources for low, mid, and high<br />

frequencies was designed, providing a smooth omni-directionality up to 6 kHz and a usable frequency range from 40 Hz up to 20 kHz.<br />

However, a realistic auralization of sources like musical instruments is not possible with an omni-directional measured RIR. To<br />

include the directional characteristics of instruments in the measuring setup, the directivity of the sound source has to be frequency<br />

dependent and must be matched to the �measured� directivity of the real instrument. This can be obtained by using a dodecahedron<br />

loudspeaker with independently operating systems and an appropriate complex FIR filtering of the frequency response of each driver.<br />

The directivity is a result of parameters like magnitude and phase and the interference sum of all systems. To create the appropriate<br />

directivity, optimization algorithms are used to achieve minimum error between measured and adapted directivity.<br />

3224 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3224


9:55–10:15 Break<br />

10:15<br />

4aAA8. Objective measures for evaluating tonal balance of sound fields. Daiji Takahashi �Dept. of Urban and Environ. Eng.,<br />

Kyoto Univ., Kyoto Univ. Katsura, Nishikyo-ku, Kyoto 615-8540, Japan, tkhs@archi.kyoto-u.ac.jp�, Kanta Togawa �FUJITEC Co.,<br />

Ltd., Hikone Shiga 522-8588, Japan�, and Tetsuo Hotta �YAMAHA Corp., Hamamatsu Shizuoka 430-8650, Japan�<br />

The purpose of this study is to derive objective measures, which can well represent the characteristics of the sound field regarding<br />

the tonal balance corresponding to our hearing sense. Two kinds of listening test were conducted in the form of paired comparison, in<br />

which subjects were tested using sound fields produced by convoluting some anechoic music sources with some impulse responses.<br />

In the first listening test, impulse responses were calculated theoretically for a simple structure of sound field having a direct sound<br />

and reflections, and, in the second test, impulse responses were measured at the various seats of existing concert halls. In the latter<br />

case, impulse responses which give almost the same reverberance were used for the listening tests. From this investigation, it is found<br />

that one objective measure called the DL �deviation of level� has a possibility of an effective measure, which can be used as an<br />

appropriate measure for evaluating the tonal balance of sound fields. The index DL is calculated from the data based on the<br />

logarithmic scale in both the frequency and magnitude. This fact is not inconsistent with the past findings that human response<br />

corresponds to a logarithmic scale of stimulus.<br />

10:35<br />

4aAA9. Measuring impulse responses containing complete spatial information. Angelo Farina, Paolo Martignon, Andrea Capra,<br />

and Simone Fontana �Industrial Eng. Dept., Univ. of Parma, via delle Scienze 181/A, 43100 Parma, Italy�<br />

Traditional impulse response measurements did capture limited spatial information. Often just omnidirectional sources and microphones<br />

are employed. In some cases it was attempted to get more spatial information employing directive transdudcers: known<br />

examples are binaural microphones, figure-of-8 microphones, and directive loudspeakers. However, these approaches are not scientifically<br />

based and do not provide an easy way to process and visualize the spatial information. On the other side, psychoacoustics<br />

studies demonstrated that ‘‘spatial hearing’’ is one of the dominant factors for the acoustic quality of <strong>room</strong>s, particularly for theatres<br />

and concert halls. Of consequence, it is necessarily to reformulate the problem entirely, describing the transfer function between a<br />

source and a receiver as a time/space filter. This requires us to ‘‘sample’’ the impulse response not only in time, but also in space. This<br />

is possible employing spherical harmonics for describing, with a predefined accuracy, the directivity pattern of both source and<br />

receiver. It is possible to build arrays of microphones and of loudspeakers, which, by means of digital filters, can provide the required<br />

directive patterns. It can be shown how this makes it possible to extract useful information about the acoustical behavior of the <strong>room</strong><br />

and to make high-quality auralization.<br />

10:55<br />

4aAA10. Spherical and hemispherical microphone arrays for capture and analysis of sound fields. Ramani Duraiswami,<br />

Zhiyun Li, Dmitry N. Zotkin, and Elena Grassi �Perceptual Interfaces and Reality Lab., Inst. for Adv. Comput. Studies, Univ. of<br />

Maryland, College Park, MD 20742�<br />

The capture of the spatial structure of a sound field and analysis is important in many fields including creating virtual environments,<br />

source localization and detection, noise suppression, and beamforming. Spherical microphone arrays are a promising development<br />

to help achieve such capture and analysis, and have been studied by several groups. We develop a practical spherical<br />

microphone array and demonstrate its utility in applications for sound capture, <strong>room</strong> measurement and for beamforming and tracking.<br />

To accommodate equipment failure and manufacturing imprecision we extend their theory to handle arbitrary microphone placement.<br />

To handle speech capture and surveillance we describe the development of a new sensor, the hemispherical microphone array. For<br />

each array the practical performance follows that predicted by theory. Future applications and improvements are also discussed. �Work<br />

supported by NSF.�<br />

11:15<br />

4aAA11. High-order wave decomposition using a dual-radius spherical microphone array. Boaz Rafaely, Ilya Balmages, and<br />

Limor Eger �Dept. of Elec. and Comput. Eng., Ben-Gurion Univ. of the Negev, Beer-Sheva 84105, Israel�<br />

The acoustic performance of an auditorium is influenced by the manner in which sound propagates from the stage into the seating<br />

areas. In particular, the spatial and temporal distribution of early reflections is considered important for sound perception in the<br />

auditorium. Previous studies presented measurement and analysis methods based on spherical microphone arrays and plane-wave<br />

decomposition that could provide information on the direction and time of arrival of early reflections. This paper presents recent<br />

results of <strong>room</strong> acoustics analysis based on a spherical microphone array, which employs high spherical harmonics order for improved<br />

spatial resolution, and a dual-radius spherical measurement array to avoid ill-conditioning at the null frequencies of the spherical<br />

Bessel function. Spatial-temporal analysis is performed to produce directional impulse responses, while time-windowed spacefrequency<br />

analysis is employed to detect direction of arrival of individual reflections. Experimental results of sound-field analysis in<br />

a real auditorium will also be presented.<br />

3225 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3225<br />

4a FRI. AM


11:35<br />

4aAA12. Impulse response measurement system and its recent applications. Kazuhiro Takashima, Hiroshi Nakagawa, Natsu<br />

Tanaka, and Daiki Sekito �1-21-10, Midori, Sumida-Ku, Tokyo 130-0021, Japan�<br />

Our impulse response measurement system has been developed for ten years. During this decade, the environment related to this<br />

measurement has changed significantly. In this article, the features and notes on the measurement system using the sound card, and our<br />

brand new system, which is expanded for multichannel inputs, will be presented. Finally, a new technique, which combines multichannel<br />

impulse response measurement and signal processing with microphone array, will be presented. The microphone array was<br />

designed for noise analysis for automobile interiors. The array consists of 31 microphones on the surface of an acoustically hard<br />

sphere. Moreover, 12 cameras are arranged on the surface of the sphere to take photos. Some applications and future development will<br />

be presented.<br />

FRIDAY MORNING, 1 DECEMBER <strong>2006</strong> KOHALA/KONA ROOM, 8:00 TO 11:45 A.M.<br />

Session 4aAB<br />

Animal Bioacoustics: Marine Mammal Acoustics I<br />

Paul E. Nachtigall, Chair<br />

Hawaii Inst. of Marine Biology, P.O. Box 1106, Kailua, HI 96734<br />

8:00<br />

4aAB1. Development of evoked-potential audiometry in odontocetes.<br />

Alexander Supin �Inst. of Ecology and Evolution, 33 Leninsky prospect,<br />

119071 Moscow, Russia�<br />

Evoked-potential methods are widely used for investigation of hearing<br />

in whales, dolphins, and porpoises. For this purpose, mostly the auditory<br />

brainstem response �ABR� or rhythmic trains of ABRs, the envelopefollowing<br />

response �EFR�, are used. Although very productive, these<br />

methods require further elaboration. �i� Traditionally the EFR is provoked<br />

by sinusoidally amplitude-modulated tones �SAM�. SAM stimuli have narrow<br />

frequency band, which makes them little effective to produce the<br />

EFR, because response amplitude depends on the stimulus bandwidth. A<br />

solution of the problem is the use of trains of short tone pips instead of<br />

SAM tones. Such stimuli produce several times higher EFR than SAM<br />

tones. This makes the threshold determination much more confident and<br />

precise. The effect is achievable at stimulus bandwidths, which still do not<br />

influence negatively the precision of attribution of the threshold to a certain<br />

frequency. �ii� To extract low-amplitude evoked potentials from noise,<br />

the average technique is traditionally used. This operation returns a mean<br />

value of averaged traces. Effectively diminishing stationary noise, this<br />

method poorly eliminates big artifacts, which may spoil the record even it<br />

if appeared once or twice during acquisition. With this respect, computation<br />

of the median instead of mean is much more effective.<br />

8:15<br />

4aAB2. Towards a predictive model of noise-induced temporary<br />

threshold shift for an amphibious marine mammal, the California sea<br />

lion „Zalophus californianus…. David Kastak, Marla M. Holt, Jason<br />

Mulsow, Colleen J. Reichmuth Kastak, Ronald J. Schusterman �UCSC<br />

Long Marine Lab., 100 Shaffer Rd., Santa Cruz, CA 95060�, and Brandon<br />

L. Southall �Natl. Marine Fisheries Service, Silver Spring, MD 20910�<br />

A California sea lion that had previously been tested under water was<br />

assessed for noise-induced temporary threshold shift �TTS� in air. One<br />

hundred ninety-two controlled exposures of octave-band noise centered at<br />

2.5 kHz were conducted over a 3-year period. The noise was varied in<br />

level �to 133 dB SPL re: 20�Pa� and duration �to 50 min� to generate a<br />

variety of equal sound exposure levels �SELs�. Behavioral psychophysics<br />

was used to measure hearing sensitivity at 2.5 kHz before, immediately<br />

following, and 24 h following noise exposure. The levels of threshold<br />

Contributed Papers<br />

shifts obtained ranged up to 30 dB. In cases where TTS exceeded 20 dB,<br />

thresholds were obtained at regular intervals until recovery occurred. The<br />

average slope of the long-term recovery function was 10 dB per log-<br />

�minute�. Results show that the threshold shifts correlated with SEL; however,<br />

the equal-energy trading rule did not apply in all circumstances, with<br />

exposure duration contributing more than exposure level. Repeated testing<br />

showed no evidence of a permanent threshold shift at 2.5 kHz or octave<br />

higher. The amphibious sea lions appear to be equally susceptible to noise<br />

in air and under water, provided that the exposure levels are referenced to<br />

the subjects thresholds in both media.<br />

8:30<br />

4aAB3. Electrophysiological investigation of temporal resolution in<br />

three pinniped species: Adaptive implications. Jason Mulsow and<br />

Colleen Reichmuth Kastak �Univ. of California Santa Cruz, Long Marine<br />

Lab., 100 Shaffer Rd., Santa Cruz, CA 95060�<br />

Electrophysiological studies of auditory temporal processing in marine<br />

mammals have traditionally focused on the role of highly refined temporal<br />

resolution in dolphin echolocation. Studies in manatees, however, have<br />

found their temporal resolution to be better than expected, leading to<br />

speculation that such capabilities are an adaptation for underwater sound<br />

localization. This study measured the ability of auditory brainstem responses<br />

to follow rhythmic click stimuli in California sea lions �Zalophus<br />

californianus�, harbor seals �Phoca vitulina�, and northern elephant seals<br />

�Mirounga angustirostris�. Trains of 640-s clicks were presented in air at<br />

repetition rates of 125–1500 per second and averaged rate-following responses<br />

were recorded. Rate-following responses were detected in both<br />

the harbor seal and the sea lion at rates up to 1000 clicks per second,<br />

indicating that pinnipeds, like manatees, possess temporal resolution<br />

greater than humans but inferior to dolphins. While this finding might<br />

support an underwater sound localization hypothesis, comparable results<br />

were obtained in preliminary testing of a dog �Canis familiaris�, suggesting<br />

that increased temporal resolution in pinnipeds may not be the result of<br />

the evolutionary pressure of an aquatic environment, but rather a result of<br />

increased high-frequency hearing essential to mammalian sound localization.<br />

�Work supported by NOPP, ONR, and NMFS.�<br />

3226 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3226


8:45<br />

4aAB4. Click and tone-pip auditory evoked potentials in a large<br />

marine mammal, the northern elephant seal. Dorian S. Houser<br />

�BIOMIMETICA, 7951 Shantung Dr., Santee, CA 92071� and James J.<br />

Finneran �Space and Naval Warfare Systems Ctr., San Diego, CA 92152�<br />

The use of auditory-evoked potentials �AEPs� to study the hearing of<br />

mysticete whales is challenged by access to animals, their large size, and<br />

proportionately smaller brain relative to odontocetes. One means by which<br />

AEP techniques can be adapted to these larger animals is by application to<br />

more readily available proxy species. The northern elephant seal �Mirounga<br />

angustirostris� is a large pinniped, potentially in excess of 2000 kg,<br />

with a thick dermis, large skull, relatively small auditory nerve, and a<br />

low-frequency vocal communication system. AEP collection in elephant<br />

seals provides similar challenges to those of the mysticetes but at a scale<br />

that provides a greater opportunity for success. AEP tests were conducted<br />

on northern elephant seals at Año Nuevo State Reserve, the natural haulout<br />

site of the elephant seal. Subjects were chemically immobilized with<br />

tiletamine/zolazepam and chemical restraint was maintained with bolus<br />

injections of ketamine. Click-evoked potentials were collected from four<br />

weanling and two adult male elephant seals and tone-pip-evoked potentials<br />

were collected from a 2-year-old female. Results demonstrate that<br />

AEPs can be recorded from large pinniped species, providing a step towards<br />

the application of similar techniques to larger cetacean species.<br />

9:00<br />

4aAB5. Acoustic field measurements and bottlenose dolphin hearing<br />

thresholds using single-frequency and frequency-modulated tones.<br />

James J. Finneran �U.S. Navy Marine Mammal Program,<br />

SPAWARSYSCEN San Diego, Code 2351, 49620 Beluga Rd., San Diego,<br />

CA 92152, james.finneran@navy.mil� and Carolyn E. Schlundt �EDO<br />

Professional Services, San Diego, CA 92110�<br />

Studies of underwater hearing are often hampered by the behavior of<br />

sound waves in small experimental tanks. At lower frequencies, tank dimensions<br />

are often not sufficient for free-field conditions, resulting in<br />

large spatial variations of sound pressure. These effects may be mitigated<br />

somewhat by increasing the frequency bandwidth of the sound stimulus,<br />

so effects of multipath interference average out over many frequencies. In<br />

this study, acoustic fields and bottlenose dolphin �Tursiops truncatus�<br />

hearing thresholds were compared for pure-tone and frequency-modulated<br />

stimuli. Experiments were conducted in a vinyl-walled, seawater-filled<br />

pool approximately 4�5�1.5 m. Sound stimuli consisted of 500-ms<br />

tones at 13 carrier frequencies between 1 and 100 kHz. Frequencymodulated<br />

stimuli featured both linear and sinusoidal modulating waveforms<br />

with 5%, 10%, and 20% bandwidths. Acoustic fields were measured<br />

�without the dolphin present� at three depths over a 60�65-cm grid with a<br />

5-cm spacing. Hearing thresholds were measured using a behavioral response<br />

paradigm and up/down staircase technique. Frequency-modulated<br />

stimuli with a 10% bandwidth resulted in significant improvements to the<br />

sound field without substantially affecting the dolphins hearing thresholds.<br />

�Work supported by ONR.�<br />

9:15<br />

4aAB6. Hearing frequency selectivity in four species of toothed<br />

whales as revealed by the evoked-potential method. Vladimir Popov<br />

�Inst. of Ecology and Evolution, 33 Leninsky Prosp., 119071 Moscow,<br />

Russia popov_vl@sevin.ru�<br />

Frequency tuning curves were obtained using a tone-tone simultaneous<br />

masking paradigm in conjunction with the evoked potential recording. The<br />

masker was a continuous tone and the test was a sinusoidal amplitudemodulated<br />

�SAM� tonal signal, which evoked the envelope following response<br />

�EFR�. The EFR was recorded in unanaesthetized animals from a<br />

head surface with the use of suction-cup electrodes. The obtained tuning<br />

curves featured very sharp tuning with Q�ERB� �quality estimated by the<br />

equivalent rectangular bandwidth� from35inTursiops truncatus to nearly<br />

50 in Delphinapterus leucas. This acuteness is several times better than in<br />

humans and many animals. The Q�ERB� dependence on probe frequency<br />

could be approximated by regression lines with a slope from 0.18 in Tur-<br />

siops trucatus to 0.83–0.86 in Phocoena phocoena and Neophocoena phocoenoides.<br />

Thus, the frequency representation in the odontocete auditory<br />

system may be either near constant quality �in Tursiops� or near constant<br />

bandwidth �in porpoises�. �Work supported by The Russian Foundation for<br />

Basic Research and Russian President Grant.�<br />

9:30<br />

4aAB7. Growth and recovery of temporary threshold shifts in a<br />

dolphin exposed to midfrequency tones with durations up to 128 s.<br />

Carolyn E. Schlundt �EDO Professional Services, 3276 Rosecrans St.,<br />

San Diego, CA 92110, carolyn.melka@edocorp.com�, Randall L. Dear<br />

�Sci. Applications Intl. Corp., San Diego, CA 92110�, Donald A. Carder,<br />

and James J. Finneran �Space and Naval Warfare Systems Ctr., San<br />

Diego, San Diego, CA 92152�<br />

Auditory thresholds at 4.5 kHz were measured in a bottlenose dolphin<br />

�Tursiops truncatus� before and after exposure to midfrequency tones at 3<br />

kHz. Experiments were conducted in relatively quiet pools with low ambient<br />

noise levels at frequencies above 1 kHz. Behavioral hearing tests<br />

allowed for thresholds to be routinely measured within 4 min postexposure,<br />

and tracked recovery for at least 30 min postexposure. Exposure<br />

durations ranged from 4 to 128 s at sound pressure levels ranging from<br />

149 to 200 dB re: 1�Pa. Sound exposure levels ranged from 155 to 217<br />

dB re: 1�Pa 2 /s. Temporary threshold shifts at 4 min postexposure (TTS 4)<br />

of up to 23 dB were observed. All thresholds recovered to baseline and<br />

pre-exposure levels, most within 30 min of exposure. �Work supported by<br />

the U.S. ONR.�<br />

9:45<br />

4aAB8. Auditory brainstem response recovery rates during doublepulse<br />

presentation in the false killer whale „Pseudorca crassidens…: A<br />

mechanism of automatic gain control? Paul E. Nachtigall �Marine<br />

Mammal Res. Program, Hawaii Inst. of Marine Biol., P.O. Box 1106,<br />

Kailua, HI 96734�, Alexander Ya. Supin �Russian Acad. of Sci., Moscow,<br />

Russia�, and Marlee Breese �Hawaii Inst. of Marine Biol., Kailua, HI<br />

96734�<br />

The outgoing echolocation pulse and the return echo response can be<br />

approximately examined in the auditory system of an echolocating animal<br />

by presenting two pulses and determining the forward-masking effect of<br />

the first pulse on the response to the second pulse using auditory-evoked<br />

potential procedures. False killer whale, Pseudorca crassidens, auditory<br />

brainstem responses �ABR� were recorded using a double-click stimulation<br />

paradigm specifically measuring the recovery of the second �test�<br />

response �to the second click� as a function of the length of the interclick<br />

interval �ICI� following various levels of the first �conditioning� click. At<br />

all click intensities, the slopes of the recovery functions were almost constant:<br />

0.60.8 V per ICI decade. Therefore, even when the conditioning-toclick<br />

level ratio was kept constant, the duration of recovery was intensity<br />

dependent: the higher intensity the longer the recovery. The conditioningto-test-click<br />

level ratio strongly influenced the recovery time: the higher<br />

the ratio, the longer the recovery. This dependence was nearly linear, using<br />

a logarithmic ICI scale with a rate of 2530 dB per ICI decade. These data<br />

were used for modeling the interaction between the emitted click and the<br />

echo in the auditory system during echolocation.<br />

10:00–10:15 Break<br />

10:15<br />

4aAB9. Temporary threshold shifts in the bottlenose dolphin<br />

„Tursiops truncatus…, varying noise duration and intensity. T. Aran<br />

Mooney �Dept. of Zoology and Hawaii Inst. of Marlne Biol., Univ. of<br />

Hawaii, 46-007 Lilipuna Rd., Kaneohe, HI 96744�, Paul E. Nachtigall,<br />

Whitlow W. L. Au, Marlee Breese, and Stephanie Vlachos �Univ. of<br />

Hawaii, Kaneohe, HI 96744�<br />

There is much concern regarding increasing noise levels in the ocean<br />

and how it may affect marine mammals. However, there is a little information<br />

regarding how sound affects marine mammals and no published<br />

3227 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3227<br />

4a FRI. AM


data examining the relationship between broadband noise intensity and<br />

exposure duration. This study explored the effects of octave-band noise on<br />

the hearing of a bottlenose dolphin by inducing temporary hearing threshold<br />

shifts �TTS�. Sound pressure level �SPL� and exposure duration were<br />

varied to measure the effects of noise duration and intensity. Hearing<br />

thresholds were measured using auditory evoked potentials before and<br />

after sound exposure to track and map TTS and recovery. Shifts were<br />

frequency dependent and recovery time depended on shift and frequency,<br />

but full recovery was relatively rapid, usually within 20 min and always<br />

within 40 min. As exposure time was halved, TTS generally occurred with<br />

an increase in noise SPL. However, with shorter, louder noise, threshold<br />

shifts were not linear but rather shorter sounds required greater sound<br />

exposure levels to induce TTS, a contrast to some published literature.<br />

From the data a novel algorithm was written that predicts the physiological<br />

effects of anthropogenic noise if the intensity and duration of exposure<br />

are known.<br />

10:30<br />

4aAB10. Estimates of bio-sonar characteristics of a free-ranging<br />

Ganges river dolphin. Tamaki Ura, Harumi Sugimatsu, Tomoki Inoue<br />

�Underwater Technol. Res. Ctr., Inst. of Industrial Sci., Univ. of Tokyo,<br />

4-6-1 Komaba, Meguro, Tokyo 153-8505, Japan�, Rajendar Bahl �IIT<br />

Delhi, New Delhi 110016, India�, Junichi Kojima �KDDI R&D Labs.<br />

Inc., Saitama 356-8502, Japan�, Tomonari Akamatsu �Fisheries Res.<br />

Agency, Ibaraki 314-0408, Japan�, Sandeep Behera �Freshwater &<br />

Wetlands Prog., New Delhi 110003, India�, Ajit Pattnaik, Muntaz Kahn<br />

�Chilika Dev. Authority, Orissa, India�, Sudhakar Kar, Chandra Sekhar<br />

Kar �Off. of the Princip. CCF �Wildlife� & Chief Wildlife Warden,<br />

Blaubaneswar 751012, India�, Tetsuo Fukuchi, Hideyuki Takashi<br />

�System Giken Co. Ltd., Kanagawa 253-0085, Japan�, and Debabrata<br />

Swain �Simpilipal Biosphere and Tiger Reserve, Orissa, India�<br />

This paper reports the first known studies of the bio-sonar characteristics<br />

of an isolated free-ranging Ganges river dolphin, Platanista<br />

gangetica. The animal preferred to roam in a deeper tract of the otherwise<br />

shallow river. The click sounds of the dolphin were recorded over a period<br />

of 2 days on a 3.2-meter-long high-frequency hydrophone array composed<br />

of three hydrophones forming an equispaced linear array and another two<br />

hydrophones in conjunction with the central hydrophone forming an SSBL<br />

triangular array in a plane perpendicular to the array axis. The array was<br />

deployed both in horizontal and vertical configurations. The array structure<br />

provided 3-D measurements of the source location through measurement<br />

of the interelement time delay. Bio-sonar characteristics such as click<br />

duration, bandwidth, and interclick intervals in click trains have been reported.<br />

Measurements of dolphin track and the relative click levels on the<br />

array hydrophones have been used to obtain a preliminary characterization<br />

of the animal’s beam pattern.<br />

10:45<br />

4aAB11. Discriminating between the clicks produced by a bottlenose<br />

dolphin when searching for and identifying an object during a search<br />

task. Sandra Bohn, Stan Kuczaj �Univ. of Southern Mississippi, 118<br />

College Dr., #5025, Hattiesburg, MS 39406, sandra.bohn@usm.edu�, and<br />

Dorian Houser �BIOMIMETICA, Santee, CA 92071�<br />

Clicks collected from an echolocating bottlenose dolphin completing a<br />

search task were compared in order to determine if the clicks produced<br />

when the dolphin was acquiring the target differed from the clicks produced<br />

when the dolphin was searching for the target. The clicks produced<br />

by a free-swimming dolphin completing the search task were recorded<br />

using a biosonar measurement tool �BMT�, an instrumentation package<br />

carried by the dolphin that collected both the outgoing clicks and the<br />

returning echoes. A discriminant function analysis classified the clicks as<br />

search or acquisition using the variables of peak-to-peak amplitude, duration,<br />

peak frequency, center frequency, and bandwidth. The acquisition<br />

clicks were classified more accurately than the search clicks. Acquisition<br />

clicks and search clicks were significantly different across all five of the<br />

variables. These results suggest that the clicks produced by bottlenose<br />

dolphins acquiring a target are different than those produced by dolphins<br />

searching for a target.<br />

11:00<br />

4aAB12. Echo highlight amplitude and temporal difference<br />

resolutions of an echolocating Tursiops truncatus. Mark W. Muller,<br />

Whitlow W. L. Au, Paul E. Nachtigall, Marlee Breese �Marine Mammal<br />

Res. Program, Hawai’i Inst. of Marine Biol., 46-007 Lilipuna Rd.,<br />

Kaneohe, HI 96744�, and John S. Allen III �Univ. of Hawai’i at Manoa,<br />

Honolulu, HI 96822�<br />

A dolphin’s ability to discriminate targets may greatly depend on the<br />

relative amplitudes and the time separations of echo highlights within the<br />

received signal. Previous experiments with dolphins have varied the physical<br />

parameters of targets, but did not fully investigate how changes in<br />

these parameters corresponded with the composition of the scattered<br />

acoustic waveforms and the dolphin’s subsequent response. A novel experiment<br />

utilizes a phantom echo system to test a dolphin’s detection<br />

response of relative amplitude differences of secondary echo highlights<br />

and the time separation differences of all the echo highlights both within<br />

and outside the animal’s integration window. By electronically manipulating<br />

these echoes, the underlying acoustic classification cues can be more<br />

efficiently investigated. In the first study, the animal successfully discriminated<br />

between a standard echo signal and one with the middle highlight<br />

amplitude at �7 dB. When the middle highlight amplitude was raised to<br />

�6 dB, the animal’s discrimination performance radically dropped to<br />

65%. This study suggests the animal may not be as sensitive to the secondary<br />

echo highlights as previously proposed. The experiments were repeated<br />

for the trailing highlight amplitude and the time separations between<br />

the primary and middle highlights and the middle and trailing<br />

highlights.<br />

11:15<br />

4aAB13. A background noise reduction technique for improving false<br />

killer whale „Pseudorca crassidens… localization. Craig R. McPherson,<br />

Owen P. Kenny, Phil Turner �Dept. of Elec. and Comput. Eng., James<br />

Cook Univ., Douglas 4811, Queensland, Australia�, and Geoff R.<br />

McPherson �Queensland Dept. of Primary Industries and Fisheries,<br />

Cairns, 4870, Queensland Australia�<br />

The passive localization of false killer whales �Pseudorca crassidens�<br />

in acoustic environments comprised of discontinuous ambient, anthropogenic,<br />

and animal sounds is a challenging problem. A background noise<br />

reduction technique is required to improve the quality of sampled recordings,<br />

which will assist localization using auditory modeling and signal<br />

correlation at extended ranges. The algorithm developed meets this requirement<br />

using a combination of adaptive percentile estimation, a<br />

median-based tracker, and Gaussian windowing. The results indicate successful<br />

improvement of the signal-to-noise ratio, and consequently a significant<br />

increase in the detection range of false killer whales in acoustically<br />

complex environments.<br />

11:30<br />

4aAB14. Analysis of Australian humpback whale song using<br />

information theory. Jennifer L. Miksis-Olds, John R. Buck �School for<br />

Marine Sci. and Technol., Univ. of Massachusetts Dartmouth, New<br />

Bedford, MA 02744, jmiksis@umassd.edu�, Michael J. Noad �Univ. of<br />

Queensland, St. Lucia, QLD 4072 Australia�, Douglas H. Cato �Defence<br />

Sci. & Tech. Org., Pyrmont, NSW 2009 Australia�, and Dale Stokes<br />

�Scripps Inst. of Oceanogr., La Jolla, CA 92093�<br />

Songs produced by migrating whales were recorded off the coast of<br />

Queensland, Australia over 6 consecutive weeks in 2003. Approximately<br />

50 songs were analyzed using information theory techniques. The average<br />

length of the songs estimated by correlation analysis was approximately<br />

100 units, with song sessions lasting from 300 to over 3100 units. Song<br />

entropy, a measure of structural constraints and complexity, was estimated<br />

using three different methodologies: �1� the independently identically distributed<br />

model; �2� first-order Markov model; and �3� the nonparametric<br />

sliding window match length �SWML� method, as described in Suzuki<br />

et al. �J. Acoust. Soc. Am. 119, 1849 �<strong>2006</strong>��. The analysis finds the songs<br />

of migrating Australian whales are consistent with the hierarchical structure<br />

proposed by Payne and McVay �Science 173, 585�597 �1971��, and<br />

3228 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3228


ecently confirmed by Suzuki et al. for singers on the breeding grounds.<br />

Both the SWML entropy estimates and the song lengths for the Australian<br />

singers were lower than that reported by Suzuki et al. for Hawaiian<br />

whales in 1976�1978. These lower SWML entropy values indicate a<br />

higher level of predictability within songs. The average total information<br />

in the Australian sequence of song units was approximately 35 bits/song.<br />

Aberrant songs �10%� yielded entropies similar to the typical songs.<br />

�Sponsored by ONR and DSTO.�<br />

FRIDAY MORNING, 1 DECEMBER <strong>2006</strong> KAHUKU ROOM, 7:55 A.M. TO 12:00 NOON<br />

Session 4aBB<br />

Biomedical UltrasoundÕBioresponse to Vibration: Interaction of Cavitation Bubbles with Cells and Tissue<br />

John S. Allen, Cochair<br />

Univ. of Hawaii, Dept. of Mechanical Engineering, 2540 Dole St., Honolulu, HI 96822<br />

Yoshiki Yamakoshi, Cochair<br />

Gunma Univ., Faculty of Engineering, 1-5-1 Tenjin-cho, Kiryu-shi, Gunma 376-8515, Japan<br />

Chair’s Introduction—7:55<br />

Invited Papers<br />

8:00<br />

4aBB1. Ultra-high-speed imaging of bubbles interacting with cells and tissue. Michel Versluis, Philippe Marmottant, Sascha<br />

Hilgenfeldt, Claus-Dieter Ohl �Phys. of Fluids, Univ. of Twente, P.O. Box 217, 7500 AE Enschede, The Netherlands�, Chien T.<br />

Chin, Annemieke van Wamel, Nico de Jong �Erasmus MC, 3000 DR Rotterdam, The Netherlands�, and Detlef Lohse �Univ. of<br />

Twente, 7500 AE Enschede, The Netherlands�<br />

Ultrasound contrast microbubbles are exploited in molecular imaging, where bubbles are directed to target cells and where their<br />

high-scattering cross section to ultrasound allows for the detection of pathologies at a molecular level. In therapeutic applications<br />

vibrating bubbles close to cells may alter the permeability of cell membranes, and these systems are therefore highly interesting for<br />

drug and gene delivery applications using ultrasound. In a more extreme regime bubbles are driven through shock waves to sonoporate<br />

or kill cells through intense stresses or jets following inertial bubble collapse. Here, we elucidate some of the underlying mechanisms<br />

using the 25-Mfps camera Brandaris128, resolving the bubble dynamics and its interactions with cells. We quantify acoustic microstreaming<br />

around oscillating bubbles close to rigid walls and evaluate the shear stresses on nonadherent cells. In a study on the fluid<br />

dynamical interaction of cavitation bubbles with adherent cells, we find that the nonspherical collapse of bubbles is responsible for cell<br />

detachment. We also visualized the dynamics of vibrating microbubbles in contact with endothelial cells followed by fluorescent<br />

imaging of the transport of propidium iodide, used as a membrane integrity probe, into these cells showing a direct correlation<br />

between cell deformation and cell membrane permeability.<br />

8:20<br />

4aBB2. Sonoporation: Mechanisms of cell membrane perforation and rapid resealing. Nobuki Kudo and Katsuyuki Yamamoto<br />

�Grad. School of Information Sci. and Technol., Hokkaido Univ., Sapporo 060-0814 Japan, kudo@bme.ist.hokudai.ac.jp�<br />

Sonoporation is a technique for making membrane perforation by exposure of cells to ultrasound, and it is an attractive method for<br />

realizing nonvirus gene transfection. A continuous or quasicontinuous wave is frequently used for this technique because a higher duty<br />

ratio gives higher efficiency of sonoporation. Addition of microbubbles during insonification greatly improves the efficiency of<br />

sonoporation, and, especially when microbubbles exist in the vicinity of the cells, ultrasound pulses from diagnostic ultrasound<br />

equipment can cause sonoporation. In this study, we examined sonoporation induced by single-shot pulsed ultrasound and the role of<br />

microbubbles in induction of cell membrane perforation. Bubble behavior and cell membrane damage were observed using a highspeed<br />

camera and light and scanning electron microscopes. Results of observation showed that mechanical stress induced by bubble<br />

motion could cause cell membrane perforation. We also studied repair of the perforation using a fluorescence microscope and found<br />

that the membrane of mammalian cells has the ability to reseal the perforation within several seconds. �Research partially supported<br />

by a Grant-in-Aid for Scientific Research from the Ministry of Education, Science, Sports and Culture, Japan.�<br />

8:40<br />

4aBB3. Quantitative imaging of tumor blood flow with contrast ultrasound. Peter N. Burns, Raffi Karshafian, and John Hudson<br />

�Dept. Medical Biophys., 2075 Bayview Ave., Toronto ON, M4N 3M5, Canada�<br />

The point at which a solid cancer develops its own blood supply marks the onset of malignant progression. This process, known<br />

as angiogenesis, makes oxygen and nutrients available for growth and provides a path for metastatic spread. Angiogenesis is not only<br />

of interest as a diagnostic hallmark of malignancy, but also as a target for new therapeutic strategies. Assessing antiangiogenic<br />

therapies noninvasively poses problems—flow velocities (�1 mm/s� and vessel diameters (�50 �m� are below resolutions of direct<br />

imaging. Vessels are disorganized without the tree-like structure of normal vasculature. We have investigated the potential role of<br />

3229 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3229<br />

4a FRI. AM


microbubble disruption-replenishment flow measurement in monitoring antivascular treatment of an animal tumor. The currently used<br />

monexponential model incorrectly considers the vasculature a perfect mixing chamber. Simple fractal models of the circulation<br />

provide a distribution of vessel diameters which, combined with the geometry of the disruption and detection beams, produce better<br />

models of replenishment following acoustic bubble disruption. These not only measure flow, but also predicts differences between<br />

organized and disorganized circulations, even with equal flow and vascular volume. If detectable, such differences might be used to<br />

characterize vascular organization below the resolution limit of an ultrasound image.<br />

9:00<br />

4aBB4. Dynamics of laser-trapped microbubbles. Hiroyuki Takahira �Dept. of Mech. Eng., Osaka Prefecture Univ., 1-1<br />

Gakuen-cho, Naka-ku, Sakai, Osaka 599-8531, Japan�<br />

A laser-trapping method is utilized for microbubbles. Bubbles of the order of 10 microns in diameter are trapped and manipulated<br />

successfully using a dry objective lens with large working distance. The growth or shrinkage of a laser-trapped microbubble and the<br />

merger of microbubbles are observed with a high-speed camera to investigate the influence of gas diffusion on the stability of<br />

microbubbles. Two kinds of equilibrium radii are found for shrinking microbubbles. The first one is related to the equilibrium surface<br />

concentration of surfactant. The other is related to the decrease of the surface tension due to the compression of the surface area at the<br />

maximum surfactant concentration. The simulations in which the dynamic surface tension is considered are in good agreement with<br />

the experiments. The laser trapping technique is also applied to the motion of a microbubble in a shear flow. It is shown that the bubble<br />

escapes from the laser trap being repelled by the optical force in the shear flow. There is overall agreement between the experiments<br />

and the simulations in which the buoyancy force, the fluid dynamic forces, and the optical force are taken into account.<br />

9:20<br />

4aBB5. Novel methods of micro-object trapping by acoustic radiation force. Yoshiki Yamakoshi �1-5-1 Tenjin-cho, Kiryushi,<br />

Gunma 376-8515 Japan, yamakosi@el.gunma-u.ac.jp�<br />

It is expected that micro object trapping by acoustic radiation force is a useful method in future drug delivery systems in order to<br />

concentrate the payloads at desired position. In this paper, two novel methods of micro object trapping are presented. First method is<br />

micro object trapping by seed bubbles. This method uses seed bubbles, which have higher sensitivity to the ultrasonic wave, in order<br />

to trap micro objects, which are difficult to trap by conventional methods due to low volumetric oscillation under the ultrasonic wave.<br />

The Bjerkne’s force, which is produced by a secondary wave radiated from the seed bubbles, traps the target objects making bi-layer<br />

seed bubbletarget object mass. The Second method is micro bubble trapping by bubble nonlinear oscillation. Two ultrasonic waves<br />

with different frequencies �pumping and control waves� are introduced simultaneously. The frequency of the control wave is set to a<br />

harmonic frequency of the pumping wave. If the bubbles flow into the cross area of two waves, nonlinear oscillation by high intensity<br />

pumping wave generates the Bjerkne’s force, producing multiple traps with narrow separation along the control wave propagation<br />

direction. In order to demonstrate these methods, experiments using an ultrasonic wave contrast agent are shown.<br />

9:40<br />

4aBB6. Mechanical properties of HeLa cells at different stages of cell cycle by time-resolved acoustic microscope. Pavel V.<br />

Zinin �School of Ocean and Earth Sci. and Technol., Univ. of Hawaii, 2525 Correa Rd., Honolulu, HI 96822-2219�, Eike C. Weiss,<br />

Pavlos Anastasiadis, and Robert M. Lemor �Fraunhofer Inst. for Biomed. Technol., St. Ingbert, Germany�<br />

Scanning acoustic microscopy �SAM�, particularly time-resolved acoustic microscopy, is one of the few techniques for study of<br />

the mechanical properties of only the cell’s interior, cytosol and nucleus. Unfortunately, time-resolved acoustic microscopes typically<br />

do not provide sufficient resolution to study the elasticity of single cells. We demonstrate that the high-frequency, time-resolved<br />

acoustic microscope developed at the Fraunhofer Institute for Biomedical Technology �IBMT�, Germany, is capable of imaging and<br />

characterizing elastic properties of micron size structures in cell’s cytoskeleton with a theoretical resolution limit of 10 m/s for sound<br />

speed measurements. Measurements were performed on cells of the HeLa cell line derived from human cervics carcinoma. SAM<br />

measurements of the sound speed of adherent HeLa cells at different states of the cell cycle were conducted. They yielded an average<br />

value of 1540 m/s. B-Scan images of HeLa cells at different states of the cell cycle show distinct patterns inside the cell. A method<br />

for estimating sound attenuation inside HeLa cells is outlined as such a method is critical for the determination of a cell’s viscoelasticity.<br />

�Work supported by Alexander von Humboldt Foundation and the European Framework Program 6, Project ‘‘CellProm.’’�<br />

10:00–10:10 Break<br />

10:10<br />

4aBB7. Assessment of shock wave lithotripters via cavitation potential. Jonathan I. Iloreta, Andrew J. Szeri �UC Berkeley, 6119<br />

Etcheverry Hall, M.S. 1740, Berkeley, CA 94720-1740�, Yufeng Zhou, Georgii Sankin, and Pei Zhong �Duke Univ., Durham, NC<br />

27708-0300�<br />

An analysis of bubbles in elastic media has been made in order to characterize shock wave lithotripters by gauging the potential<br />

for cavitation associated with the lithotripter shock wave �LSW�. The method uses the maximum radius achieved by a bubble<br />

subjected to a LSW as the key parameter that defines the potential damage a lithotripter could cause at any point in the domain. The<br />

maximum radius is determined by an energy analysis. A new index—similar in spirit to the Mechanical Index of Holland and Apfel<br />

for diagnostic ultrasound—is proposed for use in gauging the likelihood of cavitation damage.<br />

3230 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3230


10:30<br />

4aBB8. Formation of water pore in a bilayer induced by shock wave:<br />

Molecular dynamics simulation. Kenichiro Koshiyama, Takeru Yano,<br />

Shigeo Fujikawa �Lab. of Adv. Fluid Mech., Hokkaido Univ., Sapporo<br />

060-8628, Japan, koshi@ring-me.eng.hokudai.ac.jp�, and Tetsuya<br />

Kodama �Tohoku Univ., Aobaku, Sendai 980-8575, Japan�<br />

The irradiation of a shock wave or ultrasound with micro-bubbles has<br />

the potential to make transient pores on cell membranes. Although such<br />

pores are believed to contribute to the molecular delivery thorough the<br />

membrane, the detailed mechanisms of the pore formation with shock<br />

waves and the subsequent molecular delivery through the pores into cells<br />

are still unclear. To investigate the mechanism at a molecular level, the<br />

molecular dynamics simulations of the interaction of the shock wave with<br />

a lipid bilayer are conducted. The water penetration into the hydrophobic<br />

region by the shock wave is observed in picoseconds. As a next step,<br />

structural changes of the bilayer containing water molecules in the hydrophobic<br />

region are investigated. The water pore is formed in 3 ns when the<br />

large number of water molecules is inserted. The lifetime of the water pore<br />

is more than 70 ns. The radius of the water pore is ca. 1.0 nm, which is<br />

three times larger than the Stoke’s radius of a typical anticancer drug<br />

�5FU�. Finally, the diffusion of the anticancer drug in the water pore is<br />

investigated.<br />

10:45<br />

4aBB9. Ultrasonic spore lysis and the release of intracellular content<br />

in a microfluidic channel. Oana C. Marina, Michael D. Ward, John M.<br />

Dunbar, and Gregory Kaduchak �MPA-11, Los Alamos Natl. Lab., P.O.<br />

Box 1663, MS D-429, Los Alamos, NM, 87545�<br />

Ultrasonic lysis of suspended spores in a microfluidic channel is a<br />

promising alternative to conventional spore disruption techniques that include<br />

bead beating as the spore lysis gold standard. Our overall research<br />

goal is to obtain an automated detection system with complete sample<br />

preparation and lysis steps in a microfluidic channel. Previously, much<br />

work in this area has focused on organism viability rather than the release<br />

of intracellular material. Our research focuses on quantifying the amount<br />

of intracellular content �e.g., DNA, proteins, etc.� that is released by<br />

acoustic lysis for detection by the sensor. Elucidating the efficacy of<br />

acoustics on the release of intracellular material requires reliable methods<br />

to quantify the released intracellular content �nucleic acids and proteins�.<br />

The device used for lysing spores consists of a microfluidic chamber with<br />

one acoustically active wall. The chamber depths are in the range of 100–<br />

200 m. Channels tested in the 70-kHz to 1-MHz frequency range show<br />

that the efficiency of intracellular release depends on the operating frequency<br />

of the device and the properties �concentration, composition� of<br />

the spore suspensions. Experimental results on viability and released intracellular<br />

content are discussed. �Work supported by LANL LDRD.�<br />

11:00<br />

4aBB10. The correlation between cavitation noise power and bubbleinduced<br />

heating in high-intensity focused ultrasound. Caleb H. Farny,<br />

Tianming Wu, R. Glynn Holt, and Ronald A. Roy �Dept. of Aerosp. and<br />

Mech. Eng., Boston Univ., 110 Cummington St., Boston, MA 02215,<br />

cfarny@bu.edu�<br />

It has been established that inertial cavitation is responsible for elevated<br />

heating during high-intensity focused ultrasound �HIFU� application<br />

for certain intensity regimes. The contribution of bubble-induced<br />

heating can be an important factor to consider, as it can be several times<br />

that expected from absorption of the primary ultrasound energy. Working<br />

in agar-graphite tissue phantoms with a 1.1-MHz HIFU transducer, an<br />

embedded type-E thermocouple, and a 15-MHz passive cavitation detector<br />

�PCD�, the temperature and cavitation signal near the focus were measured<br />

for 5-s continuous wave HIFU insonations. The measured temperature<br />

was corrected for heating predicted from the primary ultrasound absorption<br />

and the transient thermocouple viscous heating artifact to isolate<br />

the temperature rise from the bubble activity. We have found that the<br />

Contributed Papers<br />

temperature rise induced from the bubble activity correlates well with the<br />

instantaneous cavitation noise power as indicated by the mean square voltage<br />

output of the PCD. The results suggest that careful processing of the<br />

cavitation signals could serve as a proxy for measuring the heating contribution<br />

from inertial cavitation. �Work supported by the Dept. of the<br />

Army �Award No. DAMD17-02-2-0014� and the Center for Subsurface<br />

Sensing and Imaging Systems �NSF ERC Award No. EEC-9986821�.�<br />

11:15<br />

4aBB11. Membrane permeabilization of adherent cells with laserinduced<br />

cavitation bubbles. Rory Dijkink, Claus-Dieter Ohl �Phys. of<br />

Fluids, Univ. of Twente, Postbus 217, 7500 AE Enschede, The<br />

Netherlands�, Erwin Nijhuis, Sèverine Le Gac �Univ. of Twente, 7500 AE<br />

Enschede, The Netherlands�, and Istvàn Vermes �Medical Spectrum<br />

Twente Hospital Group, 7500 KA Enschede, The Netherlands�<br />

Strongly oscillating bubbles close to cells can cause the opening of the<br />

cell’s membrane, thus to stimulate the uptake of molecules from the exterior.<br />

However, the volume oscillations of bubbles induce complex fluid<br />

flows, especially when bubble-bubble interaction takes place. Here, we<br />

report on an experiment where a single cavitation bubble is created close<br />

to a layer of adherent HeLa cells. The interaction distance between the<br />

bubble and the cell layer is controlled by adjusting the focus of the pulsed<br />

laser light, which creates the cavitation bubble. The dynamics of the<br />

bubble and the cells is recorded with high-speed photography. The poration<br />

of the cells is probed with different fluorescent stains to distinguish<br />

viable and permanent poration and programmed cell death �apopotosis�.<br />

Quantitative data are presented as a function of the radial distance from<br />

the stagnation point. Our main finding is the importance of the asymmetrical<br />

collapse and the high-speed jet flow: After impact of the jet onto the<br />

substrate a strong boundary layer flow is responsible for shearing the cells.<br />

11:30<br />

4aBB12. Antitumor effectiveness of cisplatin with ultrasound and<br />

nanobubbles. Tetsuya Kodama, Yukiko Watanabe, Kiyoe Konno,<br />

Sachiko Horie �Res. Organization, Tohoku Univ., 2-1 Seiryo-machi,<br />

Aoba-ku, Sendai, Miyagi 980-8575, Japan�, Atsuko Aoi �Tohoku Univ.,<br />

Sendai 980-8575, Japan�, Geroges Vassaux �Bart’s and The London<br />

School of Medicine and Dentistry, UK�, and Shiro Mori �Tohoku Univ.<br />

Hospital, Sendai 980-8575, Japan�<br />

The potentiation of antitumor effect of cis-diamminedichloroplatinum<br />

�II�, cisplatin, with ultrasound �1 MHz, 0.6 MPa� and lipid-shelled<br />

nanobubbles in vitro �EMT6, C26, MCF7, A549� and in vivo on s.c. tumor<br />

in mice �HT29-expressing luciferase� were evaluated. In vitro and in vivo<br />

antitumor effects were measured by an MTT assay and a real-time in vivo<br />

imaging, respectively. The effective antitumor effect was seen both in vitro<br />

and in vivo when ultrasound and nanobubbles were used, while other<br />

treatment groups with cisplatin with ultrasound did not show the effectiveness.<br />

The antitumor effect was not attributed to necrosis but apoptosis,<br />

which was confirmed by increase in the activity of the pro-apoptosis signal<br />

caspase-3 and Bax. In conclusion, the combination of ultrasound and<br />

nanobubbles with cisplatin is an effective chemotherapy of solid tumors<br />

and may prove useful in clinical application.<br />

11:45<br />

4aBB13. Sonoporation by single-shot pulsed ultrasound with<br />

microbubbles—Little effect of sonochemical reaction of inertial<br />

cavitation. Kengo Okada, Nobuki Kudo, and Katsuyuki Yamamoto<br />

�Grad. School of Information Sci. and Technol., Hokkaido Univ., Kita 14<br />

Nishi 9, Kita-ku, Sapporo 060-0814, Japan�<br />

Sonoporation is a technique for introducing large molecules into a cell<br />

by exposure to ultrasound, and it has a potential application for gene<br />

transfection. Although continuous-wave ultrasound is generally used for<br />

this technique, we have been using single-shot pulsed ultrasound with<br />

microbubbles. To determine the contribution of the sonochemical effect of<br />

3231 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3231<br />

4a FRI. AM


inertial cavitation under the condition of single-shot exposure, we compared<br />

rates of cell membrane damage in the presence and absence of a free<br />

radical scavenger �cysteamine, 5 mM�. Cells with microbubbles in their<br />

vicinity were exposed to pulsed ultrasound of 1.1 MPa in negative peak<br />

pressure under microscopic observation, and the numbers of total and<br />

damaged cells in the view field were counted. The damage rates were<br />

8.1�4.0% and 10.3�6.3% in the presence (n�17) and absence (n<br />

�25) of the scavenger, respectively, and the average number of total cells<br />

was 772�285. Since there was no significant difference, we concluded<br />

that the cell membrane damage observed in our exposure condition was<br />

not caused by the sonochemical effect but by the mechanical effect of<br />

inertial cavitation. �Research was supported by a grant-in-aid for scientific<br />

research from the Ministry of Education, Science, Sports and Culture,<br />

Japan.�<br />

FRIDAY MORNING, 1 DECEMBER <strong>2006</strong> OAHU ROOM, 8:00 TO 11:50 A.M.<br />

Session 4aEA<br />

Engineering Acoustics and ASA Committee on Standards: Developments in Microphones: Calibrations,<br />

Standards, and Measures<br />

George S. K. Wong, Cochair<br />

National Research Council, Inst. for National Measurement Standards, 1500 Montreal Rd., Ottawa,<br />

Ontario K1A 0R6, Canada<br />

Masakasu Iwaki, Cochair<br />

NHK Science and Technology Research Labs., 1-10-11 Kinuta, Setagaya-ku, Tokyo 157-8510, Japan<br />

Chair’s Introduction—8:00<br />

Invited Papers<br />

8:05<br />

4aEA1. Current developments at the National Institute for Standards and Technology in pressure calibration of laboratory<br />

standard microphones. Victor Nedzelnitsky, Randall P. Wagner, and Steven E. Fick �National Inst. of Standards and Technol.<br />

�NIST�, 100 Bureau Dr., Stop 8220, Gaithersburg, MD 20899-8220, victor.nedzelnitsky@nist.gov�<br />

Current research effort aims at improving the apparatus and methods for determining the pressure sensitivities of IEC types LS1Pn<br />

and LS2aP laboratory standard microphones. Among the improvements that are being systematically incorporated in an evolving test<br />

bed is the capability to operate at adjustable power line frequencies other than the usual 60 Hz. Suitable choices of line frequency<br />

relative to frequencies of calibration and adjustable bandpass filter characteristics can be used to improve the signal-to-noise ratios of<br />

measurements performed near the usual line frequency and its first few harmonics. This can enable the use of relatively large volume<br />

couplers for which uncertainties in microphone front cavity volume and equivalent volume, capillary tube effects, and heat conduction<br />

corrections have a lesser influence than they have for small-volume couplers. Another improvement aims to control and to stabilize the<br />

ambient static pressure during microphone calibrations, to reduce or eliminate the effects of barometric pressure fluctuations on these<br />

calibrations.<br />

8:25<br />

4aEA2. Free-field reciprocity calibration of laboratory standard „LS… microphones using a time selective technique. Knud<br />

Rasmussen and Salvador Barrera-Figueroa �Danish Primary Lab. of Acoust. �DPLA�, Danish Fundamental Metrology, Danish Tech.<br />

Univ., Bldg. 307, 2800 Kgs., Lyngby, Denmark�<br />

Although the basic principle of reciprocity calibration of microphones in a free field is simple, the practical problems are<br />

complicated due to the low signal-to-noise ratio and the influence of cross talk and reflections from the surroundings. The influence<br />

of uncorrelated noise can be reduced by conventional narrow-band filtering and time averaging, while correlated signals like cross talk<br />

and reflections can be eliminated by using time-selective postprocessing techniques. The technique used at DPLA overcomes both<br />

these problems using a B&K Pulse analyzer in the SSR mode �steady state response� and an FFT-based time-selective technique. The<br />

complex electrical transfer impedance is measured in linear frequency steps from a few kHz to about three times the resonance<br />

frequency of the microphones. The missing values at low frequencies are estimated from a detailed knowledge of the pressure<br />

sensitivities. Next an inverse FFT is applied and a time window around the main signal is used to eliminate cross talk and reflections.<br />

Finally, the signal is transformed back to the frequency domain and the free field sensitivities calculated. The standard procedure at<br />

DPLA involves measurements at four distances and the repeatability of the calibrations over time is within �0.03 dB up to about 1.5<br />

times the resonance frequency of the microphones.<br />

3232 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3232


8:45<br />

4aEA3. Microphone calibration by comparison. George S. K. Wong �Acoust. Standards, Inst. for Natl. Measurement Standards,<br />

National Res. Council Canada, Ottawa, ON K1A 0R6, Canada�<br />

The absolute method of microphone calibration by the reciprocity method �IEC 61094-2 1992� provides the highest accuracy of<br />

approximately 0.04 to 0.05 dB, and the procedure requires three changes of microphone in the ‘‘driver-receiver combination’’ that<br />

needs approximately 1 to 2 days. The system capital cost is relatively high. The NRC interchange microphone method for microphone<br />

calibration by comparison has been adopted internationally by the International Electrotechnical Commission as IEC 61094-5 �2001-<br />

10�. With this method, the test microphone is compared with a reference microphone calibrated by the reciprocity method and the<br />

procedure requires approximately 3 h. The uncertainty of the comparison method is between 0.08 to 0.09 dB, which satisfies most<br />

industrial needs.<br />

9:05<br />

4aEA4. Development of a laser-pistonphone for an infrasonic measurement standard. Ryuzo Horiuchi, Takeshi Fujimori, and<br />

Sojun Sato �Natl. Metrology Inst. of Japan �NMIJ�, AIST, Tsukuba Central 3, 1-1-1 Umezono, Tsukuba, 305-8563, Japan�<br />

Acoustical standards for audio frequencies are based on pressure sensitivities of laboratory standard microphones calibrated using<br />

a coupler reciprocity technique. There is a growing need to extend the frequency range downward for reliable infrasonic measurement.<br />

The reciprocity technique, however, has limitations on low-frequency calibration �1–20 Hz� because signal-to-noise ratio deteriorates<br />

and a sound leak occurs from capillary tubes that equalize the static pressure inside and outside of the coupler. These factors rapidly<br />

increase the measurement uncertainty as the frequency is lowered. NMIJ has therefore recently developed a laser-pistonphone<br />

prototype, which enables precise calibration of microphones at low frequencies. Compared with the reciprocity technique, the<br />

laser-pistonphone produces a higher sound pressure within a cavity by the sinusoidal motion of a piston and has a significantly<br />

improved signal-to-noise ratio. Sound pressure is calculated from the piston displacement, which is determined via a Michelson<br />

interferometer. A test microphone is inserted into the cavity, exposed to the sound pressure, and its open-circuit voltage is measured.<br />

Static pressure equalization is realized through the gap between the piston and its guide. Careful design of the dimensions and relative<br />

position of the cavity and piston minimizes sound leakage and friction between them.<br />

9:25<br />

4aEA5. Anechoic measurements of particle-velocity probes compared to pressure gradient and pressure microphones.<br />

Wieslaw Woszczyk �CIRMMT, McGill Univ., 555 Sherbrooke St. West, Montreal, QC, Canada H3A 1E3,<br />

wieslaw@music.mcgill.ca�, Masakazu Iwaki, Takehiro Sugimoto, Kazuho Ono �NHK Sci. & Tech. Res. Labs., Setagaya-ku, Tokyo<br />

157-8510, Japan�, and Hans-Elias de Bree �R&D Microflown Technologies�<br />

Microflown probes are true figure-of-eight-pattern velocity microphones having extended response down to below the lowest<br />

audible frequencies, low noise, and high output. Unlike pressure-gradient microphones, velocity probes do not measure acoustic<br />

pressure at two points to derive a pressure gradient. When particle velocity is present, acoustical particle velocity sensors measure the<br />

temperature difference of the two closely spaced and heated platinum wire resistors, and quantify particle velocity from the temperature<br />

measurement. Microflown probes do not require a membrane and the associated mechanical vibration system. A number of<br />

anechoic measurements of velocity probes are compared to measurements of pressure-gradient and pressure microphones made under<br />

identical acoustical conditions at varying distances from a point source having a wide frequency response. Detailed measurements<br />

show specific response changes affected by the distance to the source, and focus on the importance of transducer calibration with<br />

respect to distance. Examples are given from field work using microflown probes to record acoustic response of <strong>room</strong>s to test signals.<br />

The probe’s cosine directional selectivity can be used to change the ratio between early reflections and the diffuse sound since only<br />

the 1<br />

3 of the power in the diffuse sound field is measured with the particle velocity probe.<br />

9:45<br />

4aEA6. Sensitivity change with practical use of electret condenser microphone. Yoshinobu Yasuno �Panasonic Semiconductor<br />

Device Solutions Co., Ltd. 600, Saedo-cho, Tsuzuki-ku, Yokohama, 224-8539, Japan� and Kenzo Miura �Panasonic Mobile Commun.<br />

Eng. Co., Ltd., Yokohama, Japan�<br />

Dr. Sessler and Dr. West invented the electret condenser microphone �ECM� in 1966. It has since been applied in various ways as<br />

a sound input device. The ECM has become an important component as a microphone for communications because of its stable<br />

sensitivity frequency characteristic. Materials and production methods have been improved continually up to the present. In particular,<br />

the ECM reliability is based on the electret’s stability. For that reason, the electret surface charge decay is the main factor in ECM<br />

sensitivity degradation. This study analyzed the changes of an ECM preserved for 28 years in the laboratory and actually used for an<br />

outdoor interphone unit for 29 years. The change of diaphragm stiffness and electret surface voltage were compared with the<br />

evaluation result of a heat-acceleration test and verified. A degradation estimate of sensitivity change of ECM was performed.<br />

Regarding the life of the electret predicted in the report of former study �K. Miura and Y. Yasuno, J. Acoust. Soc. Jpn. �E� 18�1�,<br />

29–35 �1997��, the validity was verified using actual data from this long-term observation.<br />

3233 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3233<br />

4a FRI. AM


10:05–10:20 Break<br />

10:20<br />

4aEA7. Development of a small size narrow directivity microphone. Masakazu Iwaki, Kazuho Ono, Takehiro Sugimoto �NHK<br />

Sci. & Technol. Res. Labs., 1-10-11 Kinuta, Setagaya-ku, Tokyo, 157-8510, Japan�, Takeshi Ishii, and Keishi Inamaga �Sanken<br />

Microphone Co., Ltd, 2-8-8 Ogikubo, Suginami-ku, Tokyo, 167-0051, Japan�<br />

We developed a new microphone that has very sharp directivity even in the low-frequency band. In an ordinary environment of<br />

sound pick-up, the energy of background noise is distributed mainly in frequencies lower than 1000 Hz. In such frequencies, a typical<br />

cardioid microphone has the directivity pattern close to that of omni-cardiod. Consequently, it is difficult to pick up the objective<br />

sounds clearly from background noises. To suppress the noises with very small level, the directivity pattern should be also sharpened<br />

in the low-frequency band. In this report, we describe a new method to sharpen directivity for the low-frequency band. The method<br />

requires three microphone capsules. One capsule is the main microphone with a very short acoustic pipe. The others compose a<br />

second-order gradient microphone to cancel the signal that comes from behind the main microphone. A special feature of this<br />

microphone is to control a dip frequency of behind sensitivity without changing the frequency response of the front sensitivity.<br />

10:40<br />

4aEA8. Two-wafer bulk-micromachined silicon microphones. Jianmin Miao and Chee Wee Tan �Micromachines Ctr., School of<br />

Mech. and Aerosp. Eng., Nanyang Technolog. Univ., 50 Nanyang Ave., Singapore 639798, mjmmiao@ntu.edu.sg�<br />

A two-wafer concept is proposed for silicon microphone manufacturing by using bulk-micromachining and wafer bonding technologies.<br />

Acoustical holes of the backplate in one wafer are micromachined by deep reactive ion etching and the diaphragm on another<br />

wafer is created by wet-chemical etching. The two wafers are then bonded together to form silicon condenser microphones. In order<br />

to minimize the mechanical-thermal noise and increase the sensitivity within the required bandwidth, an analytical model based on<br />

Zuckerwar’s equations has been developed to find the optimum location of the acoustical holes in the backplate of microphones.<br />

According to our study, this analytical modeling has shown excellent agreement between the simulated and measured results for the<br />

B&K MEMS microphone. Silicon condenser microphones have been further optimized in terms of the air gap, number and location<br />

of acoustical holes to achieve the best performance with a low polarization voltage, and easy fabrication for possible commercialization.<br />

Details of analytical modeling, fabrication, and measurement results will be presented.<br />

11:00<br />

4aEA9. Infrasound calibration of measurement microphones. Erling Frederiksen �Bruel & Kjaer, Skodsborgvej 307, 2850<br />

Naerum, Denmark, erlingfred@bksv.com�<br />

Increasing interest for traceable infrasound measurements has caused the Consultative Committee for Acoustics, Ultrasound and<br />

Vibration �CCAUV� of BIPM to initiate a key comparison calibration project �CCAUV.A-K2� on pressure reciprocity calibration<br />

down to 2 Hz. Ten national metrology institutes, including the Danish Primary Laboratory of Acoustics �DPLA�, take part in this<br />

project. In addition DPLA has started its own infrasound calibration project, which is described in this paper. The purposes of this<br />

project are verification of the CCAUV results and development of methods for calibration of general types of measurement microphone<br />

between 0.1 and 250 Hz. The project includes the design of an active comparison coupler, an experimental low-frequency<br />

reference microphone, and new methods for its frequency response calibration. One method applies an electrostatic actuator and<br />

requires a low-pressure measurement tank, while the other requires an additional microphone, whose design is closely related to that<br />

of the reference microphone that is to be calibrated. The overall calibration uncertainty (k�2) for ordinary measurement microphones<br />

is estimated to less than 0.05 dB down to 1 Hz and less than 0.1 dB down to 0.1 Hz, if the reference is calibrated in the latter<br />

mentioned way, i.e., by the related microphones method.<br />

11:20<br />

4aEA10. Free-field calibration of 1Õ4 inch microphones for ultrasound<br />

by reciprocity technique. Hironobu Takahashi, Takeshi Fujimori,<br />

Ryuzo Horiuchi, and Sojun Sato �Natl. Metrology Inst. of Japan, AIST,<br />

Tsukuba Central 3, 1-1-1 Umezono, Tsukuba, 305-8563 Japan�<br />

Recently, equipment that radiates ultrasound radiation at frequencies<br />

far beyond the audible range is increasing in our environment. Such electronic<br />

equipment has switching regulators or inverter circuits, and many<br />

devices are unintended sources of ultrasound radiation. However, the effects<br />

of airborne ultrasound on human hearing and the human body have<br />

not been well investigated. To estimate the potential damage of airborne<br />

ultrasound radiation quantitatively, it is necessary to establish an acoustic<br />

standard for airborne ultrasound because the standard is a basis of acoustic<br />

measurement. With the intention of establishing a standard on airborne<br />

ultrasound, a free-field calibration system with an anechoic chamber was<br />

produced. The principle of free-field calibration techniques is introduced<br />

in this presentation. Type WS3 microphones �B&K 4939� were calibrated<br />

in the system to examine the calibration ability to be achieved. Results<br />

Contributed Papers<br />

showed that it can calibrate a microphone from 10 to 100 kHz with dispersion<br />

of less than 1 dB. In addition, the effects that were dependent on<br />

the uncertainty of the calibration are discussed based on those results.<br />

11:35<br />

4aEA11. An environmentally robust silicon diaphragm microphone.<br />

Norihiro Arimura, Juro Ohga �Shibaura Inst. of Technol.,3-7-5 Toyosu,<br />

Koto-ku, Tokyo, 135-8548, Japan�, Norio Kimura, and Yoshinobu Yasuno<br />

�Panasonic Semiconductor Device Solutions Co., Ltd., Saedo-cho,<br />

Tsuzuki-ku, Yokohama, Japan�<br />

Recently, many small microphones installed in cellular phones are the<br />

electret condenser microphones �ECMs� that contain an organic film diaphragm.<br />

Although FEP of fluorocarbon polymer is generally used as the<br />

electret material, silicon dioxide is also used. Recently ECMs have been<br />

made small and thin while maintaining the basic sound performance according<br />

to the market demand. In addition, environment tests and the<br />

reflow soldering mounting process have been adjusted to meet market<br />

requirements. On the other hand, the examination satisfied the demand as<br />

3234 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3234


the high temperature resistance was insufficient. This paper depicts an<br />

examination and a comparison of conventional ECM with the experimental<br />

model, a silicon diaphragm condenser microphone produced using the<br />

MEMS method. The silicon diaphragm satisfies high-temperature resis-<br />

tance and stable temperature characteristics because of its very small coefficient<br />

of linear expansion and it is measured total harmonic distortion<br />

�THD� on high pressure sound. Finally, it will be able to be used in high<br />

temperature and high pressure sound conditions in the future.<br />

FRIDAY MORNING, 1 DECEMBER <strong>2006</strong> IAO NEEDLE/AKAKA FALLS ROOM, 8:15 TO 11:15 A.M.<br />

Session 4aMU<br />

Musical Acoustics: Music Information and Communication<br />

Bozena Kostek, Cochair<br />

Gdansk Univ. of Technology, Multimedia Systems Dept., Narutowicza 11- 12, 80-952 Gdansk, Poland<br />

Masuzo Yanagida, Cochair<br />

Doshisha Univ., Dept. of Information Science and Intelligent Systems, 1-3 Tatara-Miyakodani, Kyo-Tanabe,<br />

Kyoto 610-0321, Japan<br />

Invited Papers<br />

8:15<br />

4aMU1. Introduction of the Real World Computing music database. Masataka Goto �Natl. Inst. of Adv. Industrial Sci. and<br />

Technol. �AIST�, 1-1-1 Umezono, Tsukuba, Ibaraki 305-8568, Japan, m.goto@aist.go.jp�<br />

This paper introduces the RWC (Real World Computing) Music Database, a copyright-cleared music database that is available to<br />

researchers as a common foundation for research. Shared databases are common in other research fields and have contributed<br />

importantly to progress in those fields. The field of music information processing, however, has lacked a common database of musical<br />

pieces and a large-scale database of musical instrument sounds. The RWC Music Database was therefore built in fiscal 2000 and 2001<br />

as the world’s first large-scale music database compiled specifically for research purposes. It contains six original collections: the<br />

Popular Music Database �100 pieces�, Royalty-Free Music Database �15 pieces�, Classical Music Database �50 pieces�, Jazz Music<br />

Database �50 pieces�, Music Genre Database �100 pieces�, and Musical Instrument Sound Database �50 instruments�. To address<br />

copyright issues, all 315 musical pieces were originally composed, arranged, or performed, and all instrumental sounds were originally<br />

recorded. The database has already been distributed to more than 200 research groups and is widely used. In addition, a<br />

continuous effort has been undertaken to manually annotate a set of music-scene descriptions for the musical pieces, called AIST<br />

Annotation, which consists of the beat structure, melody line, and chorus sections.<br />

8:35<br />

4aMU2. Japanese traditional singing on the same lyrics. Ichiro Nakayama �Osaka Univ. of Arts, 469, Higashiyama, Kanan-cho,<br />

Minami-Kawachi-gun, Osaka, 585-8555 Japan� and Masuzo Yanagida �Doshisha Univ., Kyo-Tanabe, 610-0321 Japan�<br />

Described is a database of Japanese traditional singing together with supplementary recording of Bel Canto for comparative<br />

studies. Singing sounds and spoken speech by the same singers are recorded in pair to form the body of the database. This database<br />

covers most of genres of Japanese traditional singing, such as Shinto prayers, Buddist prayers, Nor, Kyogen, Heikyoku, Sokyoku,<br />

Gidayu-bushi, Kabuki, Nagauta, Tokiwazu, Kiyomoto, Itchu-bushi, Shinnai, Kouta, Zokkyoku, Rokyoku, Shigin, Ryukyu-clasico,<br />

Goze-uta, etc. All the sounds were recorded in anechoic chambers belonging to local institutions, mainly in Osaka and Tokyo, asking<br />

78 professional singers including 18 ‘‘Living National Treasures’’ to sing as informants. The most important point of this database is<br />

that an original lyric especially prepared for this recording is commonly used to make comparative studies easy. All the subjects are<br />

asked to sing the common lyrics in their own singing styles. Shown here are comparisons of formant shifts in vowels from ordinary<br />

speaking to singing for some singers, and comparison of temporal features of fundamental frequency between Japanese traditional<br />

singing and Western Bel Canto. �Work supported by the Academic Frontier Project, Doshisha University.�<br />

8:55<br />

4aMU3. Computational intelligence approach to archival musical recordings. Andrzej Czyzewski, Lukasz Litwic, and<br />

Przemyslaw Maziewski �Gdansk Univ. of Technol., Narutowicza 11/12, 80-952 Gdansk, Poland�<br />

An algorithmic approach to wow defect estimation in archival musical recordings is presented. The wow estimation is based on the<br />

simultaneous analysis of many sinusoidal components, which are assumed to depict the defect. The rough determination of sinusoidal<br />

components in analyzed musical recording is performed by standard sinusoidal modeling procedures employing magnitude and phase<br />

spectra analysis. Since archival recordings tend to contain distorted tonal structure, the basic sinusoidal modeling approach is often<br />

found insufficient, resulting in audible distortions in the restored signal. It is found that the standard sinusoidal modeling approach is<br />

prone to errors, especially when strong frequency or amplitude variations of sinusoidal components occur. It may result in gaps or<br />

inappropriately matched components, leading to incorrect estimation of the wow distortion. Hence, some refinements to sinusoidal<br />

component analysis including interpolation and extrapolation of tonal components are proposed. As it was demonstrated in experi-<br />

3235 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3235<br />

4a FRI. AM


ments, due to the nonlinear nature of wow distortion, the enhancement of sinusoidal analysis can be performed by means of a neural<br />

network. The paper demonstrates implemented algorithms for parasite frequency modulation in archival recordings together with<br />

obtained results. �Work supported by the Commission of the European Communities, within the Integrated Project No. FP6-507336:<br />

PRESTOSPACE.�<br />

9:15<br />

4aMU4. Music information retrieval seen from the communication technology perspective. Bozena Kostek �Gdansk Univ. of<br />

Technol., Narutowicza 11/12, PL-80-952 Gdansk, Poland�<br />

Music information retrieval �MIR� is a multidiscipline area. Within this domain one can see various approaches to musical<br />

instrument recognition, musical phrase classification, melody classification �e.g., query-by-humming systems�, rhythm retrieval,<br />

high-level-based music retrieval such as looking for emotions in music or differences in expressiveness, and music search based on<br />

listeners’ preferences. One may also find research that tries to correlate low-level descriptor analysis to high-level human perception.<br />

Researchers from musical acoustics, musicology, and music domains on one side, and communication technology on the other side,<br />

work together within this area. This may foster a framework for broader and deeper comprehension of contributions from all these<br />

disciplines and, in addition, translate the automated access to music information, gathered in various forms around the World Wide<br />

Web, as a fully understandable process to all participants regardless of their background. The semantic description is becoming a basis<br />

of the next Web generation. Several important concepts have been introduced recently by the researchers associated with the MIR<br />

community with regard to semantic data processing including techniques for computing with words. In this presentation some aspects<br />

related to MIR are briefly reviewed in the context of possible and actual applications of ontology-based approach to this domain.<br />

9:35<br />

4aMU5. Accompaniment included song waveform retrieval based on framewise phoneme recognition. Yuichi Yaguchi and<br />

Ryuichi Oka �Univ. of Aizu, Tsuruga, Ikkimachi, Aizuwakamatsu, Fukushima, 965-8580 Japan�<br />

A novel approach is presented for a retrieval method that is useful for waveforms of songs with accompaniment. Audio signals of<br />

songs have some different acoustical characteristics from speech signals. Furthermore, the length per mora of signals is longer than<br />

that of speech. Therefore, the authors suggest a sound retrieval system for application to musical compositions, including songs, that<br />

extracts framewise acoustical characteristics and uses a retrieval method for absorbing phoneme length. First, the system prepares two<br />

sets of phoneme identification functions that have corresponding order, but for which phoneme sets belong to different environments<br />

of accompaniment-included or accompaniment-reduced. Next, musical compositions are put into database and the query song wave<br />

converts a waveform to a label sequence using framewise phoneme recognition derived by Bayesian estimation that applies each<br />

phoneme identification function according to whether it is accompaniment-included or not. Finally, the system extracts an interval<br />

area, such as query data, from a database using spotting recognition that is derived using continuous dynamic programming �CDP�.<br />

Retrieval method results agree well with earlier results �Y. Yaguchi and R. Oka, AIRS2005, LNCS3689, 503–509 �2005�� that applied<br />

the same musical composition set without accompaniment.<br />

9:55–10:10 Break<br />

10:10<br />

4aMU6. Design of an impression-based music retrieval system. Kimiko Ohta, Tadahiko Kumamoto, and Hitoshi Isahara �NICT,<br />

Keihanna Kyoto 619-0289, Japan, kimiko@nict.go.jp�<br />

Impression-based music retrieval helps users to find musical pieces that suit their preferences, feelings, or mental states from<br />

among a huge volume of a music database. Users are asked to select one or more pairs of impression words from among multiple pairs<br />

that are presented by the system and to estimate each selected pair on a seven-step scale to input their impressions into the system. For<br />

instance, if they want to locate musical pieces that will create a happy impression, they should check the radio button ‘‘Happy’’ in the<br />

impression scale: Very happy–Happy–A little happy–Neutral–A little sad–Sad–Very sad. A pair of impression words with a sevenstep<br />

scale is called an impression scale in this paper. The system calculates the distance between the impressions of each musical piece<br />

in a user-specified music database and the impressions that are input by the user. Subsequently, it selects candidate musical pieces to<br />

be presented as retrieval results. The impressions of musical pieces are expressed numerically by vectors that are generated from a<br />

musical piece’s pitch, strength, and length of every tone using n-gram statistics.<br />

10:30<br />

4aMU7. Automatic discrimination between singing and speaking<br />

voices for a flexible music retrieval system. Yasunori Ohishi, Masataka<br />

Goto, Katunobu Itou, and Kazuya Takeda �Grad. School of Information<br />

Sci., Nagoya Univ., Furo-cho 1, Chikusa-ku, Nagoya, Aichi, 464-8603,<br />

Japan, ohishi@sp.m.is.nagoya-u.ac.jp�<br />

This paper describes a music retrieval system that enables a user to<br />

retrieve a song by two different methods: by singing its melody or by<br />

saying its title. To allow the user to use those methods seamlessly without<br />

changing a voice input mode, a method of automatically discriminating<br />

between singing and speaking voices is indispensable. We therefore first<br />

investigated measures that characterize differences between singing and<br />

Contributed Papers<br />

speaking voices. From subjective experiments, we found that human listeners<br />

discriminated between these two voices with 70% accuracy for<br />

200-ms signals. These results showed that even short-term characteristics<br />

such as the spectral envelope represented as MFCC can be used as a<br />

discrimination cue, while the temporal structure is the most important cue<br />

when longer signals are given. According to these results, we then developed<br />

the automatic method of discriminating between singing and speaking<br />

voices by combining two measures: MFCC and an F0 �voice pitch�<br />

contour. Experimental results with our method showed that 68.1% accuracy<br />

was obtained for 200-ms signals and 87.3% accuracy was obtained<br />

for 2-s signals. Based on this method, we finally built a music retrieval<br />

system that can accept both singing voices for the melody and speaking<br />

voices for the title.<br />

3236 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3236


10:45<br />

4aMU8. Various acoustical aspects of an Asian „South… Indian<br />

classical music concert. M. G. Prasad �Dept. of Mech. Eng., Stevens<br />

Inst. of Technol., Hoboken, NJ 07030�, V. K. Raman �Flautist,<br />

Germantown, MD 20874�, and Rama Jagadishan �Edison, NJ 08820�<br />

An Asian �South� Indian classical music concert is an integrated acoustical<br />

experience for both the audience and the player�s�. A typical concert<br />

team, either vocal or instrumental, consists of a main vocalist �or an instrumentalist�<br />

accompanied by a violinist, up to three percussion instrument<br />

players, and a reference drone. The concert is comprised of many<br />

songs. Each song has two main parts, namely Alapana and Kriti. The<br />

Alapana is an elaboration of a raga �tune� and the Kriti refers to the lyrics<br />

of the song. The violinist actively follows and supports the main musician<br />

during the concert. The percussion player�s� are provided an opportunity<br />

to present a solo of their rhythmic skills. The players and the audience<br />

communicate emotionally and intellectually with each other. Elements<br />

such as aesthetics, rhythm, skill, and emotional aspects of the players are<br />

evaluated and appreciated by the audience. This talk will present various<br />

aspects of a concert that brings about an integrated and holistic experience<br />

for both the audience and the player�s�. Some samples from live vocal and<br />

instrumental music concerts will be presented.<br />

11:00<br />

4aMU9. Musical scales, signals, quantum mechanics. Alpar Sevgen<br />

�Dept. of Phys., Bogazici Univ., Bebek 34342, Istanbul, Turkey�<br />

Scales, being finite length signals, allow themselves to be treated algebraically:<br />

key signatures are related to the ‘‘ring’’ property of the scale<br />

labels; cyclically permuted scales and their mirror images have the same<br />

number of sharps and flats; and complementary scales �like major and<br />

pentatonic� have their sharp and flat numbers exchanged. A search for<br />

minimum principles to select among all possible scales those employed in<br />

music yields two possibilities: �a� minimize total number of accidentals<br />

and �b� minimize frequency fluctuations in a scale. Either of these minimum<br />

principles helps filter those scales employed in music from the universe<br />

of all scales, setting up very different criteria than the harmonic<br />

ratios used by musicians. The notes of the scales employed in music seem<br />

to prefer to stay as far apart from each other as possible. Operators that<br />

step through the multiplet members of scales with N semitones form a<br />

complete set of operators together with those that step through their eigenvectors.<br />

The mathematics reveals the discrete Fourier transformations<br />

�DFT� and is identical to finite state quantum mechanics of N-level Stern-<br />

Gerlach filters worked out by J. Schwinger.<br />

FRIDAY MORNING, 1 DECEMBER <strong>2006</strong> MAUI ROOM, 7:30 A.M. TO 12:15 P.M.<br />

Session 4aNS<br />

Noise and Architectural Acoustics: Soundscapes and Cultural Perception I<br />

Brigitte Schulte-Fortkamp, Cochair<br />

Technical Univ. Berlin, Inst. of Technical Acoustics, Secr TA 7, Einsteinufer 25, 10587 Berlin, Germany<br />

Bennett M. Brooks, Cochair<br />

Brooks Acoustics Corp., 27 Hartford Turnpike, Vernon, CT 06066<br />

Invited Papers<br />

7:30<br />

4aNS1. Soundscape in the old town of Naples: Signs of cultural identity. Giovanni Brambilla �CNR Istituto di Acustica ‘‘O.M.<br />

Corbino’’ Via del Fosso del Cavaliere 100, 00133 Roma, Italy�, Luigi Maffei, Leda De Gregorio, and Massimiliano Masullo �Second<br />

Univ. of Naples, 81031 Aversa �Ce�, Italy�<br />

Like all cities in Magna Grecia, the ancient Neapolis was built along three main parallel, tight, and straight streets called<br />

decumani. Since then and during the following centuries, commercial and handicraft activities, as well as social life, have been<br />

developed along these streets. The narrow ground <strong>room</strong>s forced shopkeepers to occupy the main street to show their merchandise<br />

using vocal appeals to magnify their product, and handicrafts to work directly on the street �hammering, sawing, etc.�. Music artists<br />

had their performance on the streets too. The soundscape in the area was a strong symbol of the Neapolitan cultural identity.<br />

Nowadays decumani have kept the main features of the past but some of these are overrun by road traffic. To investigate in which way<br />

the traffic noise has modified the soundscape perception and cultural identity, sound walks were registered during day and night time.<br />

A number of residents were interviewed and laboratory listening tests were carried out. Despite the congested urban environment and<br />

high sound levels, preliminary results have shown that part of the residential population is still able to identify the soundscape more<br />

related to Neapolitan historical identity.<br />

7:50<br />

4aNS2. Soundscape design in public spaces: Concept, method, and practice. Hisao Nakamura-Funaba and Shin-ichiro Iwamiya<br />

�Kyushu Univ., 4-9-1.Shiobaru, Minami-ku, Fukuoka 815-8540, Japan�<br />

Soundscape design of public spaces necessitates consideration of whether a space has important meaning for the user. It is<br />

essential to imagine an ideal sound environment of that space. We designed an actual soundscape from the viewpoint of the<br />

environment, information, and decoration. In many cases, producing some silence in an environment becomes the first step of<br />

soundscape design. There is neither a special technology nor a technique when designing. It merely requires use of a general<br />

technology and techniques concerning sound according to location. A key point is knowledge of know how to coordinate these<br />

technologies and techniques. For instance, silence was made first at the renewal project of Tokyo Tower observatory through<br />

cancellation of its commercial and call broadcasting functions and installation of sound-absorbing panels to the ceiling. Next, suitable<br />

and small sounds were added at various points. Guests can take time to enjoy viewing as a result.<br />

3237 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3237<br />

4a FRI. AM


8:10<br />

4aNS3. The daily rhythm of soundscape. Brigitte Schulte-Fortkamp �TU-Berlin, Einsteinufer 25 TA 7, 10587 Germany,<br />

brigitte.schulte-fortkamp@tu-berlin.de� and Andr Fiebig �HEAD acoustics GmbH, 52134 Herzogenrath, Germany�<br />

With respect to people’s minds, soundscapes can be considered as dynamic systems characterized by the time-dependent occurrence<br />

of particular sound events embedded in specific environments. Therefore, an adequate evaluation of environmental noise will<br />

reflect the continually varying acoustical scenery and its specific perception. An acoustical diary shall provide information about the<br />

daily routine and subjectively perceived sound exposure of residents. It relies on cognitive and emotional aspects of perceiving and<br />

evaluating sounds. It gives insight into evaluating processes and their contextual parameters because of its spontaneous character. It<br />

includes and refers to long-term acoustic measurements. Simultaneous measurements for elaborate acoustical analyses will be taken<br />

outside the homes regarding the moments of essential sound events. The aim is to collect information about the daily rhythm regarding<br />

acoustic events, whereby the focus also should be placed on sound events that are well accepted to deepen the explication of data with<br />

respect to the analysis. Procedure and results will be discussed.<br />

8:30<br />

4aNS4. Ecological explorations of soundscapes: From verbal analysis to experimental settings. Daniele Dubois �CNRS, 11 rue<br />

de Lourmel, 75015 Paris, France� and Catherine Guastavino �McGill Univ., Montreal, QC H3A 1Y1, Canada�<br />

Scientific studies rely on rigorous methods that must be adapted to the object of study. Besides integrating acoustic features,<br />

soundscapes as complex cognitive representations also have the properties of being global, meaningful, multimodal, and categorical.<br />

Investigating these specificities, new paradigms were developed involving linguistics and ecological psychology to complement the<br />

psychophysical approach: cognitive linguistic analyses of discourse to address semantic properties of soundscapes, and categorization<br />

tasks and distances from prototypes to investigate their cognitive organization. As a methodological consequence, experimental<br />

settings must be designed to ensure the ecological validity of the stimuli processing, �the ‘‘realism’’ evaluated from a psychological<br />

point of view, stimuli being processed as in a real-life situation�. This point will be illustrated with perceptual evaluations of spatial<br />

auditory displays for soundscape reproduction. Data processing techniques should also take into consideration the intrinsic properties<br />

of the representations they account for. Examples of free-sorting tasks will be presented with measurements in terms of family<br />

resemblance of sets of properties defining categories rather than dimensional scales. New ways of coupling physical measurement and<br />

psychological evaluations will be presented in order to simulate or reproduce soundscapes in both a realistic and controlled manner for<br />

experimental purposes.<br />

8:50<br />

4aNS5. Artificial neural network models of sound signals in urban open spaces. Lei Yu and Jian Kang �School of Architecture,<br />

Sheffield Univ., Western Bank, Sheffield S10 2TN, UK�<br />

Sound signals, known as foreground sounds, are important components of soundscape in urban open spaces. Previous studies in<br />

this area have shown that sound preferences are different according to social and demographic factors of individual users �W. Yang<br />

and J. Kang, J. Urban Des., 10, 69–88�2005��. This study develops artificial neural network �ANN� models of sound signals for<br />

architects at design stage, simulating subjective evaluation of sound signals. A database for ANN modeling has been established based<br />

on large-scale social surveys in European and Chinese cities. The ANN models have consequently been built, where individual’s social<br />

and demographic factors, activities, and acoustic features of the space and sounds are used as input variables while the sound<br />

preference is defined as the output. Through the process of training and testing the ANN models, considerable convergences have been<br />

achieved, which means that the models can be applied as practical tools for architects to design sound signals in urban open spaces,<br />

taking the characteristics of potential users into account. Currently ANN models combining foreground and background sounds are<br />

being developed.<br />

9:10<br />

4aNS6. Describing soundscape and its effects on people where soundscape is understood as an expansion of the concept of<br />

noise engineering. Keiji Kawai �Grad. School of Sci. and Technol., Kumamoto Univ., 2-39-1 Kurokami, Kumamoto 860-8555,<br />

Japan, kawai@arch.kumamoto-u.ac.jp�<br />

This study discusses how to describe sound environment and people in terms of ‘‘soundscape’’ as an expansion of the ‘‘noise<br />

engineering.’’ In the framework of the conventional study field of noise evaluation, typically, sound environments are represented by<br />

loudness-based indices such as A-weighted sound pressure levels, and the impact of sound environments on people is represented by<br />

annoyance response or some physiological metrics. In the case of soundscape studies, however, the description should be expanded<br />

beyond what has been used in noise engineering. This matter has already been frequently discussed, but it doesn’t seem that much<br />

consensus has been achieved concerning it yet. With respect to the effects of sound environment on people, since the concept of<br />

soundscape focuses on personal and social meanings of environmental sounds including the historical or aesthetic contexts, the effects<br />

are considered to be represented not by a singular concept such as comfortableness or quietness, but by multiple dimensions of<br />

emotional and aesthetic concepts. Also, descriptions of sound environment should include some qualitative aspects, such as what types<br />

of sounds can be heard at what extent. In this paper, the methodology to describe human-soundscape relationships is discussed through<br />

a review of related studies.<br />

3238 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3238


9:30<br />

4aNS7. The sound environmental education aided by automated bioacoustic identification in view of soundscape recognition.<br />

Teruyo Oba �Natural History Museum & Inst., Chiba, 955-2 Aoba-cho, Chuo-ku, Chiba-shi, Chiba-ken 260-8682 Japan,<br />

oba@chiba-muse.or.jp�<br />

From the 2003–2004 JST projects, where the automated bioacoustic identification device Kikimimi-Zukin was introduced to the<br />

nature observation and environmental studies, it was learned that the activities encouraged children to take notice of sounds, become<br />

aware of the sound environment, and gain an insight into the soundscape. Sounds are often riddles to us, and hearing is the process<br />

to find out the facts and causes. It is more important to let children obtain appropriate clues to differentiate sounds by hearing and<br />

thinking for themselves than give them an immediate answer. Program on the Strength of Hearing was formulated to challenge<br />

children and have them enjoy hearing to identify, sharing what they hear with others, and observing environment through sounds.<br />

Kikimimi-Zukin reinforced the program by a step- by-step guide through the hearing process of scanning, focusing, characterizing,<br />

associating with relevant factors, and judging the identity. The experience not only brought them confidence in hearing but incentive<br />

to study nature and environment. With Kikimimi-Zukin children collected recordings and relevant information. Using the sound<br />

database, the local singing map and three-dimensional sound map were prepared. They facilitated communication on the local sound<br />

environment among children and with adults, leading to realization of their inner soundscape.<br />

9:50<br />

4aNS8. Acoustic environmental problems at temporary shelters for victims of the Mid-Niigata Earthquake. Koji Nagahata,<br />

Norio Suzuki, Megumi Sakamoto, Fuminori Tanba �Fukushima Univ., Kanayagawa 1, Fukushima City, Fukushima, 960-1296, Japan,<br />

nagahata@sss.fukushima-u.ac.jp�, Shin-ya Kaneko, and Tetsuhito Fukushima �Fukushima Medical Univ., Fukushima, 960-1295,<br />

Japan�<br />

An earthquake on 23 October 2004 inflicted heavy damage on the Mid-Niigata district. The earthquake isolated Yamakoshi village;<br />

consequently, all the village residents were forced to evacuate to temporary shelters in neighboring Nagaoka city for 2 months. Two<br />

types of temporary shelters were used: gymnasiums, and buildings with large separated <strong>room</strong>s similar to community centers. A<br />

questionnaire survey and interviews (N�95) were conducted to elucidate problems of the living environment at the temporary<br />

shelters. This study analyzed acoustic environmental problems there. Noise-related problems were noted by 40 respondents �46.5%�:<br />

they were the fifth most frequently cited environmental problems. Several serious complaints, e.g., general annoyance at the shelters<br />

and footsteps of refugees at night, were only indicated by respondents who had evacuated to the gymnasiums. However, some<br />

problems, e.g., the clamor of children, including crying babies and voices of other refugees, were indicated by respondents irrespective<br />

of the type of the shelters to which they had been evacuated. Therefore, buildings like community centers were more desirable for<br />

temporary shelters, at least from the perspective of noise problems.<br />

10:10–10:30 Break<br />

10:30<br />

4aNS9. The burden of cardiovascular diseases due to road traffic noise. Wolfgang Babisch �Dept. of Environ. Hygiene, Federal<br />

Environ. Agency, Corrensplatz 1, 14195 Berlin, Germany, wolfgang.babisch@uba.de� and Rokho Kim �WHO/EURO Ctr. for<br />

Environment and Health, 53113 Bonn, Germany�<br />

Epidemiological studies suggest a higher risk of cardiovascular diseases, including high blood pressure and myocardial infarction,<br />

in subjects chronically exposed to high levels of road or air traffic noise. A new meta-analysis was carried out to assess a doseresponse<br />

curve, which can be used for a quantitative risk assessment and to estimate the burden of cardiovascular disease attributable<br />

to environmental noise in European regions. Noise exposure was grouped according to 5 dB�A�-categories for the daytime outdoor<br />

average A-weighted sound pressure level, (L day ,16h:6–22h�, which was considered in most studies. Information on night-time<br />

exposure (L night ,8h:22–6hor23–7h� was seldom available. However, approximations can be made with respect to L den according<br />

to the European directive on the assessment and management of environmental noise. The strongest evidence of an association<br />

between community noise and cardiovascular endpoints was found for ischaemic heart diseases, including myocardial infarction and<br />

road traffic noise. The disability-adjusted life years lost for ischemic heart disease attributable to transport noise were estimated<br />

conservatively, assuming the same exposure patterns across the countries with an impact fraction 3% in the western European<br />

countries.<br />

10:50<br />

4aNS10. Soundscape, moderator effects, and economic implications. Cay Hehner �Henry George School of Social Sci., 121 E.<br />

30th St., New York, NY 10016, chehner.hengeoschool@att.net� and Brigitte Schulte-Fortkamp �TU-Berlin, Berlin, Germany�<br />

Soundscape is considered with respect to moderator effects and the contribution of economics. It will be questioned whether<br />

soundscapes can work as a moderator concerning noise annoyance. As shown by the different investigations concerning soundscapes,<br />

a definition of the meaning of soundscapes is necessary. Evidently, the moderating effect of a given environment and its soundscape<br />

has to be discussed on three levels: �1� extension of factors that describe annoyance,�2� peculiar feature of burdensome noise contexts,<br />

and �3� discrepancies of the social and economic status of people living in areas where the rebuilding will change the quality of the<br />

area. It has to be determined and analyzed to what extent the Georgist method of resources taxation, as recently exemplified, e.g., in<br />

Alaska and in Wyoming, can be instrumental in funding soundscapes to moderate noise annoyance as it has been the case in funding<br />

free education and allowing the distribution of a citizen’s dividend.<br />

3239 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3239<br />

4a FRI. AM


11:10<br />

4aNS11. Socio-cultural soundscape concepts to support policies for managing the acoustic environment. Michiko So Finegold,<br />

Lawrence S. Finegold �Finegold & So, Consultants, 1167 Bournemouth Court Centerville, OH 45459-2647, m-so@pb3.so-net.ne.jp�,<br />

and Kozo Hiramatsu �Kyoto Univ., Sakyou-ku, Kyoto 606-8501 Japan�<br />

In the past half-century, considerable effort has been invested in the academic, technological, and political arenas to achieve an<br />

adequate acoustic environment. Various national and international policy guidance documents have made reasonable progress in<br />

establishing a framework for a common approach to minimizing environmental noise, such as documents from various national<br />

Environmental Protection Agencies, the World Health Organization, and the European Union. Although these documents have provided<br />

useful information for global application, they only address minimizing the negative side of the acoustic environment �i.e.,<br />

noise�, they focus primarily on acoustics issues at the national or international level, and they still have not adequately considered<br />

implementation issues related to socio-cultural differences. To deal with the practical problems that exist in the acoustic environment<br />

in the context of different cultures, continuing research and new policy guidance are needed to address different local situations and<br />

in a variety of cultural contexts. The Soundscape approach has been developing tools for describing the acoustic environment at the<br />

local level to address both the positive and negative aspects of the acoustic environment. In this paper, the evolving interdisciplinary<br />

aspect of the Socio-Cultural Soundscape will be discussed and key topics for future work will be recommended.<br />

11:30<br />

4aNS12. Initial steps for the determination of environmental noise<br />

quality—The perception-related evaluation of traffic noise. Klaus<br />

Genuit, Sandro Guidati, Sebastian Rossberg, and Andr Fiebig �HEAD<br />

Acoust. GmbH, Ebertstrasse 30a, 52134 Herzogenrath, Germany,<br />

klaus.genuit@head-acoustics.de�<br />

Directives call for actions against noise pollution and noise annoyance.<br />

But, how do we eliminate harmful effects including annoyance due to the<br />

exposure of environmental noise without understanding the perception and<br />

evaluation of environmental noise? How do we preserve environmental<br />

noise quality where it is good �Directive 2002/49/EC� without identifying<br />

descriptors for noise quality? Various soundscape approaches based on<br />

different methodologies have been developed in the past. But, the measurement<br />

procedures must be realizable without much effort in order to<br />

achieve acceptance from legislation. Therefore, procedures have to be developed<br />

that capture the complexity of human hearing, on the one hand,<br />

and are feasible with respect to economy and time, on the other hand. The<br />

European project Quiet City �6FP PL516420� is dealing with, among other<br />

aspects, vehicle pass-by noise as a typical environmental noise source and<br />

its evaluation. Results of the analysis based on subjective assessments and<br />

psychoacoustic analyses carried out with respect to the development of an<br />

annoyance index will be presented and discussed. Such an index will<br />

Contributed Papers<br />

Contributed Poster Paper<br />

provide valuable information for effective improvement of noise quality.<br />

The final aim is to develop a descriptor valid for complete traffic noise<br />

scenarios predicting environmental noise quality adequately.<br />

11:45<br />

4aNS13. When objective permissible noise limits of a municipal<br />

planning process and a subjective noise ordinance conflict. Marlund<br />

Hale �Adv. Eng. Acoust., 663 Bristol Ave., Simi Valley, CA 93065,<br />

mehale@aol.com�<br />

In most communities, proposed new building projects are required to<br />

conform with planning, community development, zoning, and/or building<br />

and safety specifications and standards. In cases of allowable noise exposure<br />

and noise limits, where certain of these requirements are quite specific<br />

while others are purposefully vague, conflicts between residential and<br />

commercial neighbors can lead to extreme disagreement and needless litigation.<br />

This paper describes a recent situation occurring in an upscale<br />

beach community, the resulting conflict over existing noise sources that<br />

comply with the limits of the city planning and permitting process, and the<br />

interesting findings of the court following the civil and criminal litigation<br />

that followed. Some suggestions are given to avoid these conflicting<br />

policy situations.<br />

Poster paper 4aNS14 will be on display from 7:30 a.m. to 12:15 p.m. The author will be at the poster from 12:00 noon to 12:15 p.m.<br />

4aNS14. The complexity of environmental sound as a function of<br />

seasonal variation. Hideo Shibayama �3-7-5 Koutou-ku Tokyo,<br />

135-8548, Japan, sibayama@sic.shibaura-it.ac.jp�<br />

Residential land is performed for a surrounding area of a suburban<br />

local city. As a result of urbanization, an area of rich natural environments<br />

became narrow. For the animals and plants for whom a river and a forest<br />

are necessary, it becomes difficult to live. Environmental sound produced<br />

by the tiny insects in this area is changing from year to year. Catching the<br />

conditions for the developmental observations and environmental preservation<br />

in natural environments, we continue to measure the environmental<br />

sound as the time-series data. We estimate the complexity for these waveforms<br />

of the measured environmental sound in the season when insects<br />

chirp and do not chirp. For estimation of the complexity, we evaluate by<br />

the fractal dimension of the environmental sound. Environmental sound in<br />

early autumn is mainly generated by insects in the grass and on the trees.<br />

And, the fractal dimension for the sound waveforms of chirping of insects<br />

is up to 1.8.<br />

3240 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3240


FRIDAY MORNING, 1 DECEMBER <strong>2006</strong> WAIANAE ROOM, 7:30 A.M. TO 12:20 P.M.<br />

Session 4aPA<br />

Physical Acoustics and Biomedical UltrasoundÕBioresponse to Vibration: Sound Propagation in<br />

Inhomogeneous Media I<br />

James G. Miller, Cochair<br />

Washington Univ., Dept. of Physics, 1 Brookings Dr., St. Louis, MO 63130<br />

Mami Matsukawa, Cochair<br />

Doshisha Univ., Lab. of Ultrasonic Electronics, Kyoto 610-0321, Japan<br />

Chair’s Introduction—7:30<br />

Invited Papers<br />

7:35<br />

4aPA1. In vivo measurement of mass density and elasticity of cancellous bone using acoustic parameters for fast and slow<br />

waves. Takahiko Otani �Faculty of Eng., Doshisha Univ., Kyotanabe 610-0321 Japan�<br />

Cancellous bone �spongy bone� is comprised of a porous network of numerous trabecular elements with soft tissue in the pore<br />

space. The effect of decreasing bone density, namely a symptom of osteoporosis, is greater for cancellous bone than for dense cortical<br />

bone �compact bone�. Two longitudinal waves, the fast and slow waves, are clearly observed in cancellous bone, which correspond to<br />

‘‘waves of the first and second kinds’’ as predicted by Biot’s theory. According to experimental and theoretical studies, the propagation<br />

speed of the fast wave increases with bone density and that of the slow wave is almost constant. Experimental results show that the<br />

fast wave amplitude increases proportionally and the slow wave amplitude decreases inversely with bone density. However, the<br />

attenuation constant of the fast wave is almost independent of bone density and the attenuation constant of the slow wave increases<br />

with bone density. The<br />

in vivo ultrasonic wave propagation path is composed of soft tissue, cortical bone, and cancellous bone and is modeled to specify the<br />

causality between ultrasonic wave parameters and bone mass density of cancellous bone. Then, mass density and elasticity are<br />

quantitatively formulated and estimated.<br />

7:55<br />

4aPA2. Is ultrasound appropriate to measure bone quality factors? Pascal Laugier �Univ. Pierre et Marie Curie., UMR CNRS<br />

7623, 15 rue de l’Ecole de medecine, 7506 Paris, France, laugier@lip.bhdc.jussieu.fr�<br />

Theoretical considerations support the concept that quantitative ultrasound variables measured in transmission are mainly determined<br />

by bone microstructure and material properties. All these properties are attributes of bone other than its bone mineral density<br />

�BMD� that may contribute to its quality and thus to strength or fragility. However, the limitations of this approach for a BMDindependent<br />

characterization of bone quality, long questioned, have become indisputable. Such considerations have prompted a<br />

research aiming at the development of new methods capable of measuring bone quality factors. New ultrasonic approaches are being<br />

investigated that use ultrasonic backscatter, guided waves, or nonlinear acoustics for studying bone microstructure or microdamage.<br />

These approaches, combined with sophisticated theoretical models or powerful computational tools, are advancing ideas regarding<br />

ultrasonic assessment of bone quality, which is not satisfactorily measured by x-ray techniques.<br />

8:15<br />

4aPA3. Scanning acoustic microscopy studies of cortical and trabecular bone in the femur and mandible. J. Lawrence Katz,<br />

Paulette Spence, Yong Wang �Univ. of Missouri-Kansas City, 650 E. 25th St., Kansas City, MO 64108, katzjl@umkc.edu�, Anil<br />

Misra, Orestes Marangos �Univ. of Missouri-Kansas City, Kansas City, MO 64110�, Dov Hazony �Case Western Reserve Univ.,<br />

Cleveland, OH 44106�, and Tsutomu Nomura �Niigata Univ. Grad. School of Medical and Dental Sci., Niigata, Japan�<br />

Scanning acoustic microscopy �SAM� has been used to study the micromechanical properties of cortical and trabecular bone in<br />

both the human femur and mandible. SAM images vary in gray level reflecting the variations in reflectivity of the material under<br />

investigation. The reflection coefficient, r�(Z2�Z1)/(Z2�Z1), where the acoustic impedance �AI�, Z�dv, d is the materials local<br />

density and v is the speed of sound at the focal point; Z2 represents the AI of the material, Z1 that of the fluid coupling the acoustic<br />

wave from the lens to the material. Femoral cortical bone consists of haversian systems �secondary osteons� and interstitial lamellae,<br />

both of which show systematic variations of high and low AI from lamella to lamella. The lamellar components defining the edges of<br />

trabecular cortical bone exhibit the same lamellar variations as seen in cortical bone. Mandibular bone, while oriented perpendicular<br />

to the direction of gravitational attraction, exhibits the same cortical and trabecular structural organizations as found in the femur. It<br />

also exhibits the same systematic alternations in lamellar AI as found in femoral bone. Both femoral and mandibular cortical bone<br />

have transverse isotropic symmetry. Thus, modeling elastic properties requires only five independent measurements.<br />

3241 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3241<br />

4a FRI. AM


8:35<br />

4aPA4. The interaction between ultrasound and human cancellous bone. Keith Wear �US Food and Drug Administration, 12720<br />

Twinbrook Pkwy., Rockville, MD 20852, keith.wear@fda.hhs.gov�<br />

Attenuation is much greater in cancellous bone than in soft tissues, and varies approximately linearly with frequency between 300<br />

kHz and 1.7 MHz. At diagnostic frequencies �300 to 700 kHz�, sound speed is slightly faster in cancellous bone than in soft tissues.<br />

A linear-systems model can account for errors in through-transmission-based measurements of group velocity due to frequencydependent<br />

attenuation and dispersion. The dependence of phase velocity on porosity may be predicted from theory of propagation in<br />

fluid-filled porous solids. The dependence of phase velocity on frequency �negative dispersion� can be explained using a stratified<br />

two-component model. At diagnostic frequencies, scattering varies as frequency to the nth power where 3�n�3.5. This may be<br />

explained by a model that represents trabeculae as finite-length cylindrical scatterers.<br />

8:55<br />

4aPA5. Dependence of phase velocity on porosity in cancellous bone: Application of the modified Biot-Attenborough model.<br />

Suk Wang Yoon and Kang Il Lee �Dept. of Phys. and Inst. of Basic Sci., SungKyunKwan Univ., Suwon 440-746, Republic of Korea�<br />

This study aims to apply the modified Biot-Attenborough �MBA� model to predict the dependence of phase velocity on porosity<br />

in cancellous bone. The MBA model predicted that the phase velocity decreases nonlinearly with porosity. The optimum values for<br />

input parameters of the MBA model, such as compressional speed c m of solid bone and phase velocity parameter s 2, were determined<br />

by comparing the prediction with the previously published measurements in human calcaneus and bovine cancellous bones. The value<br />

of the phase velocity parameter s 2�1.23 was obtained by curve fitting to the experimental data only for 53 human calcaneus samples<br />

with a compressional speed c m�2500 m/s of solid bone. The root-mean-square error �rmse� of the curve fit was 15.3 m/s. The<br />

optimized value of s 2 for all 75 cancellous bone samples �53 human and 22 bovine samples� was 1.42 with the rmse of 55 m/s. The<br />

latter fit was obtained by using c m�3200 m/s. Although the MBA model relies on empirical parameters determined from the<br />

experimental data, it is expected that the model can be usefully employed as a practical tool in the field of clinical ultrasonic bone<br />

assessment.<br />

9:15<br />

4aPA6. Simulation of fast and slow wave propagations through cancellous bone using three-dimensional elastic and Biot’s<br />

trabecular models. Atsushi Hosokawa �Dept. of Elect. & Comp. Eng., Akashi Natl. Coll. of Tech., 679-3 Nishioka, Uozumi,<br />

Akashi, 674-8501 Hyogo, Japan, hosokawa@akashi.ac.jp�<br />

The propagation of ultrasonic pulse waves in cancellous �trabecular� bone was numerically simulated by using three-dimensional<br />

finite-difference time-domain �FDTD� methods. In previous research �A. Hosokawa, J. Acoust. Soc. Am. 118, 1782–1789 �2005��,<br />

two two-dimensional FDTD models, the commonly used elastic FDTD model and an FDTD model based on Biot’s theory for elastic<br />

wave propagation in an isotropic fluid-saturated porous medium, were used to simulate the fast and slow longitudinal waves propagating<br />

through cancellous bone in the direction parallel to the main trabecular orientation. In the present study, the extended<br />

three-dimensional viscoelastic and Biot’s anisotropic models were developed to investigate the effect of trabecular structure on the fast<br />

and slow wave propagations. Using the viscoelastic model of the trabecular frame comprised of numerous pore spaces in the solid<br />

bone, the effect of the trabecular irregularity, that is the scattering effect, on both the fast and slow waves could be investigated. The<br />

effect of the anisotropic viscous resistance of the fluid in the trabecular pore spaces on the slow wave could be considered using Biot’s<br />

anisotropic model.<br />

9:35<br />

4aPA7. Ultrasonic characteristics of in vitro human cancellous bone.<br />

Isao Mano �OYO Electric Co., Ltd., Joyo 610-0101 Japan�, Tadahito<br />

Yamamoto, Hiroshi Hagino, Ryota Teshima �Tottori Univ., Yonago<br />

683-8503 Japan�, Toshiyuki Tsujimoto �Horiba, Ltd., Kyoto 601-8510<br />

Japan�, and Takahiko Otani �Doshisha Univ., Kyotanabe 610-0321 Japan�<br />

Cancellous bone is comprised of a connected network of trabeculae<br />

and is considered as an inhomogeneous and anisotropic acoustic medium.<br />

The fast and slow longitudinal waves are clearly observed when the ultrasonic<br />

wave propagates parallel to the direction of the trabeculae. The<br />

propagation speed of the fast wave increases with bone density and that of<br />

the slow wave is almost constant. The fast wave amplitude increases proportionally<br />

and the slow wave amplitude decreases inversely with bone<br />

density. Human in vitro femoral head was sectioned to 10-mm-thick slices<br />

perpendicularly to the femoral cervical axis. These cancellous bone<br />

samples were subjected to the ultrasonic measurement system LD-100<br />

using a narrow focused beam. The propagation speed and the amplitude of<br />

the transmitted wave both for the fast and slow waves were measured at<br />

1-mm intervals. The local bone density corresponding to the measured<br />

points was obtained using a microfocus x-ray CT system. Experimental<br />

results show that the propagation speeds and amplitudes for the fast and<br />

slow waves are characterized not only by the local bone density, but also<br />

by the local trabecular structure.<br />

Contributed Papers<br />

9:50<br />

4aPA8. The effect of hydroxyapatite crystallite orientation on<br />

ultrasonic velocity in bovine cortical bone. Yu Yamato, Kaoru<br />

Yamazaki, Akira Nagano �Orthopaedic Surgery, Hamamatsu Univ.<br />

School of Medicine, 1-20-1 Hamamatsu Shizuoka 431-3192, Japan,<br />

yy14@hama-med.ac.jp�, Hirofumi Mizukawa, Takahiko Yanagitani, and<br />

Mami Matsukawa �Doshisha Univ., Kyotanabe, Kyoto-fu 610-0321,<br />

Japan�<br />

Cortical bone is recognized as a composite material of diverse elastic<br />

anisotropy, composed of hydroxyapatite �HAp� crystallite and type 1 collagen<br />

fiber. The aim of this study is to investigate the effects of HAp<br />

orientation on elastic anisotropy in bovine cortical bone. Eighty cubic<br />

samples were made from the cortical bone of two left bovine femurs.<br />

Longitudinal wave velocity in orthogonal three axes was measured using a<br />

conventional pulse echo system. For evaluating the orientation of HAp<br />

crystallite, x-ray diffraction profiles were obtained from three surfaces of<br />

each cubic sample. The preference of crystallites showed strong dependence<br />

on the direction of surface. The C-axis of crystallites was clearly<br />

preferred to the bone axial direction. The velocity in the axial direction<br />

was significantly correlated with the amounts of HAp crystallites aligning<br />

to the axial axis. The HAp orientation and velocity varied according to the<br />

microstructures of the samples. The samples with Haversian structure<br />

3242 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3242


showed larger preference of crystallites than plexiform samples in the<br />

axial direction. These results show clear effects of crystallites orientation<br />

on velocity.<br />

10:05–10:20 Break<br />

10:20<br />

4aPA9. Frequency variations of attenuation and velocity in cortical<br />

bone in vitro. Magali Sasso, Guillaume Haiat �Universite Paris 12,<br />

Laboratoire de Mecanique Physique, UMR CNRS 7052 B2OA, 61, avenue<br />

du General de Gaulle, 94010 Creteil, France�, Yu Yamato �Hamamatsu<br />

Univ. School of Medicine, Hamamastu, Shizuoka, 431-3192, Japan�,<br />

Salah Naili �Universite Paris 12, 94010 Creteil, France�, and Mami<br />

Matsukawa �Doshisha Univ., Kyotanabe, Kyoto-fu 610-0321, Japan�<br />

The development of ultrasonic characterization devices for cortical<br />

bone requires a better understanding of ultrasonic propagation in this heterogeneous<br />

medium. The aim of this work is to investigate the frequency<br />

dependence of the attenuation coefficient and of phase velocity and to<br />

relate them with bone microstructure and anatomical position. One hundred<br />

twenty parallelepipedic samples �4–11 mm side� have been cut from<br />

three bovine specimens and measured four times with repositioning in<br />

transmission with a pair of 8-MHz central frequency transducers. Phase<br />

velocity and BUA could be determined with acceptable precision: coefficients<br />

of variation of 0.8% and 13%, respectively. Velocity dispersion and<br />

BUA values are comprised between �13 and 40 m/s/MHz and 2 and 12<br />

dB/MHz/cm, respectively. Negative dispersion values were measured<br />

�similarly to trabecular bone� for 2% of the measured samples. BUA values<br />

were found to be smaller in plexiform than in Haversian structure and<br />

higher for porotic structure. BUA values were found to be the greatest in<br />

the postero-lateral distal part and the smallest in the anterior-medial center<br />

part of the bone. The same tendency was found for velocity dispersion.<br />

Our results show the sensitivity of the frequency dependence of ultrasound<br />

to anatomical position and micro-architectural properties of bone.<br />

10:35<br />

4aPA10. 3-D numerical simulations of wave propagation in trabecular<br />

bone predicts existence of the Biot fast compressional wave.<br />

Guillaume Haiat �Universite Paris 12, Laboratoire de Mecanique<br />

Physique, UMR CNRS 7052 B2OA, 61, avenue du General de Gaulle,<br />

94010 Creteil, France�, Frederic Padilla, and Pascal Laugier �Universite<br />

Pierre et Marie Curie, 75006 Paris, France�<br />

Trabecular bone is a poroelastic medium in which the propagation of<br />

two longitudinal waves �fast and slow� has been observed. The 3-D finitedifference<br />

time-domain simulations neglecting absorption coupled to 3-D<br />

microstructural models of 34 trabecular bone reconstructed from synchrotron<br />

radiation microtomography are shown to be suitable to predict both<br />

types of compressional wave in the three orthogonal directions. The influence<br />

of bone volume fraction �BV/TV� on the existence of the fast and<br />

slow waves was studied using a dedicated iterative image processing algorithm<br />

�dilation, erosion� in order to modify all 34 initial 3-D microstructures.<br />

An automatic criteria aiming at determining the existence of both<br />

wave modes was developed from the analysis of the transmitted signals in<br />

the frequency domain. For all samples, the fast wave disappears when<br />

bone volume fraction decreases. Both propagation modes were observed<br />

for BV/TV superior to a critical value for 2, 13, and 17 samples according<br />

to the direction of propagation. Above this critical value, the velocity of<br />

the fast �slow� wave increases �decreases� with BV/TV, consistent with<br />

Biot’s theoretical predictions. This critical value decreases when the degree<br />

of anisotropy increases, showing coupling between structural anisotropy<br />

and the existence of the fast wave.<br />

10:50<br />

4aPA11. 3-D numerical simulation of wave propagation in porous<br />

media: Influence of the microstructure and of the material properties<br />

of trabecular bone. Guillaume Haiat �Universite Paris 12, Laboratoire<br />

de Mecanique Physique, UMR CNRS 7052 B2OA, 61, avenue du General<br />

de Gaulle, 94010 Creteil, France�, Frederic Padilla, and Pascal Laugier<br />

�Universite Pierre et Marie Curie, 75006 Paris, France�<br />

Finite-difference time domain simulations coupled to 3-D microstructural<br />

models of 30 trabecular bones reconstructed from synchrotron radiation<br />

microtomography were employed herein to compare and quantify the<br />

effects of bone volume fraction, microstructure, and material properties of<br />

trabecular bone on QUS parameters. Scenarios of trabecular thinning and<br />

thickening using an iterative dedicated algorithm allowed the estimation of<br />

the sensitivity of QUS parameters to bone volume fraction. The sensitivity<br />

to bone material properties was assessed by considering independent<br />

variations of density and stiffness. The effect of microstructure was qualitatively<br />

assessed by producing virtual bone specimens of identical bone<br />

volume fraction �13%�. Both BUA and SOS show a strong correlation<br />

with BV/TV (r 2 �0.94 p10 �4 ) and vary quasi-linearly with BV/TV at an<br />

approximate rate of 2 dB/cm/MHz and 11 m/s per % increase of BV/TV,<br />

respectively. Bone alterations caused by variation in BV/TV �BUA: 40<br />

dB/cm.MHz; SOS: 200 m/s� is much more detrimental to QUS variables<br />

than that caused by alterations of material properties or diversity in microarchitecture<br />

�BUA: 7.8 dB/cm.MHz; SOS: 36 m/s�. QUS variables are<br />

changed more dramatically by BV/TV than by changes in material properties<br />

or microstructural diversity. However, material properties and structure<br />

also appear to play a role.<br />

11:05<br />

4aPA12. Singular value decomposition-based wave extraction<br />

algorithm for ultrasonic characterization of cortical bone in vivo.<br />

Magali Sasso, Guillaume Haiat �Universite Paris 12, Laboratoire de<br />

Mecanique Physique, UMR CNRS 7052 B2OA, 61, avenue du General de<br />

Gaulle, 94010 Creteil, France�, Maryline Talmant, Pascal Laugier<br />

�Universite Pierre et Marie Curie, 75006 Paris, France�, and Salah Naili<br />

�Universite Paris 12, 94010 Creteil, France�<br />

In the context of bone status assessment, the axial transmission technique<br />

allows ultrasonic evaluation of cortical bone using a multielement<br />

probe. The current processing uses the first arriving signal to evaluate the<br />

velocity while later contributions are potentially valuable and are not yet<br />

analyzed. However, all those contributions interfere, which disrupts the<br />

signal analysis. A novel ultrasonic wave extraction algorithm using a singular<br />

value decomposition method is proposed. This algorithm aims at<br />

characterizing a given energetic low-frequency �ELF� contribution observed<br />

in vivo at around 1 MHz. To evaluate the performances of the<br />

proposed algorithm, a simulated data set was constructed taking into account<br />

the influence of noise and of random interfering wavefront. The<br />

velocity of the ELF contribution is estimated on simulated datasets and<br />

compared to the input velocity. For a signal-to-noise ratio of 10 dB, the<br />

mean error associated with this method is 5.2%, to be compared with 34%<br />

with a classical signal analysis. The algorithm was also tested on real in<br />

vivo measurements. Results show the ability to accurately identify and<br />

possibly remove this wavefront contribution. Results are promising for<br />

multiple ultrasonic parameters evaluation from different wavefront contributions<br />

in our configuration.<br />

11:20<br />

4aPA13. Direct evaluation of cancellous bone porosity using<br />

ultrasound. Peiying Liu, Matthew Lewis, and Peter Antich �Grad.<br />

Program of Biomed. Eng., Univ. of Texas Southwestern Medical Ctr. at<br />

Dallas, 5323 Harry Hines Blvd, Dallas, TX 75390-9058�<br />

Quantitative measurements of trabecular bone porosity would be a<br />

great advantage in diagnosis and prognosis of osteoporosis. This study<br />

focused on evaluation of the relationship between ultrasonic reflection and<br />

the density and porosity of cancellous bone. Theoretical simulation using<br />

MATLAB and Field II ultrasound simulator predicted a linear model of a<br />

material’s porosity and parameters of the reflected ultrasonic signals. Ex-<br />

3243 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3243<br />

4a FRI. AM


perimentally, four plastic phantoms fabricated with different porosities<br />

were tested by a 5-MHz ultrasound transducer and the results agreed with<br />

simulations. Twelve specimens of bovine cancellous bone were measured.<br />

The porosities of these specimens were estimated by calculating the ratio<br />

of the mass in air to the wetted mass when they were immersed in water<br />

and all the air was removed from the pores. Among all the parameters, the<br />

peak value of the reflected ultrasound signal demonstrated highly significant<br />

linear correlations with porosity (R 2 �0.911) and density (R 2<br />

�0.866). This encouraging result shows that this technique has the potential<br />

to be used to monitor porosity changes noninvasively for clinical<br />

purpose such as noninvasive assessment of osteoporotic fracture risk.<br />

�Work supported by Pak Foundation.�<br />

11:35<br />

4aPA14. Transmission of ultrasound through bottlenose dolphin<br />

(tursiops truncatus) jaw and skull bone. Michael D. Gray, James S.<br />

Martin, and Peter H. Rogers �Woodruff School of Mech. Eng., Georgia<br />

Inst. of Technol., Atlanta, GA 30332-0405, michael.gray@gtri.gatech.edu�<br />

Measurements of ultrasound transmission through jaw �pan bone� and<br />

skull �temporal fossa� samples from an Atlantic bottlenose dolphin were<br />

performed as part of an investigation of the feasibility of performing in<br />

vivo elastography on cetacean head tissues. The pan bone and temporal<br />

fossa are both relatively thin and smooth, and are therefore of interest for<br />

acoustic imaging of the underlying tissues of the ear and brain, respectively.<br />

Field scan data will be presented, showing the influence of the bone<br />

on the quality of the focus and overall pressure levels generated by a<br />

spherically focused single-element transducer. �Work supported by ONR.�<br />

11:50<br />

4aPA15. Assessment of bone health by analyzing propagation<br />

parameters of various modes of acoustic waves. Armen Sarvazyan,<br />

Vladimir Egorov, and Alexey Tatarinov �Artann Labs., Inc., 1457 Lower<br />

Ferry Rd., West Trenton, NJ 08618�<br />

A method for assessment of bone based on comprehensive analysis of<br />

waveforms of ultrasound signals propagating in the bone is presented. A<br />

set of ultrasound propagation parameters, which are differentially sensitive<br />

to bone material properties, structure, and cortical thickness, are evaluated.<br />

The parameters include various features of different modes of acoustics<br />

waves, such as bulk, surface, and guided ultrasonic waves in a wide range<br />

of carrier frequencies. Data processing algorithms were developed for obtaining<br />

axial profiles of waveform parameters. Such profiles are capable of<br />

revealing axial heterogeneity of long bones and spatially nonuniform<br />

pathological processes, such as osteoporosis. The examination procedure<br />

may include either one long bone in the skeleton, like the tibia, radius of<br />

the forearm, etc., or several bones in sequence to provide a more comprehensive<br />

assessment of the skeletal system. Specifically, for tibia assessment,<br />

a multi-parametric linear classifier based on a DEXA evaluation of<br />

skeleton conditions has been developed. Preliminary results of the pilot<br />

clinical studies involving 149 patients have demonstrated early stages of<br />

osteoporosis detection sensitivity of 80% and specificity of 67% based on<br />

DEXA data as the gold standard. �Work was supported by NIH and NASA<br />

grants.�<br />

12:05<br />

4aPA16. Toward bone quality assessment: Interference of fast and<br />

slow wave modes with positive dispersion can account for the<br />

apparent negative dispersion. James G. Miller, Karen Marutyan, and<br />

Mark R. Holland �Dept. of Phys., Washington Univ., St. Louis, MO<br />

63130�<br />

The goal of this work was to show that apparently negative dispersion<br />

in bone can arise from interference between fast wave and slow wave<br />

longitudinal modes, each of positive dispersion. Simulations were carried<br />

out using two approaches, one based on the Biot-Johnson model and the<br />

second independent of that model. The resulting propagating fast wave<br />

and slow wave modes accumulated phase and suffered loss with distance<br />

traveled. Results of both types of simulation served as the input into a<br />

phase and magnitude spectroscopy algorithm �previously validated with<br />

experimental data� in order to determine the phase velocity as a function<br />

of frequency. Results based on both methods of simulation were mutually<br />

consistent. Depending upon the relative magnitudes and time relationships<br />

between the fast and slow wave modes, the apparent phase velocities as<br />

functions of frequency demonstrated either negative or positive dispersions.<br />

These results appear to account for measurements from many laboratories<br />

that report that the phase velocity of ultrasonic waves propagating<br />

in cancellous bone decreases with increasing frequency �negative dispersion�<br />

in about 90% of specimens but increases with frequency in about<br />

10%. �Work supported in part by Grant NIH R37HL40302.�<br />

3244 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3244


FRIDAY MORNING, 1 DECEMBER <strong>2006</strong> WAIALUA ROOM, 8:30 TO 11:25 A.M.<br />

Session 4aPP<br />

Psychological and Physiological Acoustics and ASA Committee on Standards: New Insights on Loudness<br />

and Hearing Thresholds<br />

Rhona P. Hellman, Cochair<br />

Northeastern Univ., Dept. of Speech-Language Pathology and Audiology, Boston, MA 02115<br />

Yôiti Suzuki, Cochair<br />

Tohoku Univ., Research Inst. of Electrical Communication, Katarahira 2-1-1, Aoba-ku, Sendai 980-8577, Japan<br />

Chair’s Introduction—8:30<br />

Invited Papers<br />

8:35<br />

4aPP1. Threshold of hearing for pure tones between 16 and 30 kHz. Kaoru Ashihara �AIST, Tsukuba Central, 6 Tsukuba, Ibaraki<br />

305-8566 Japan, ashihara-k@aist.go.jp�<br />

Hearing thresholds for pure tones were obtained at 2-kHz intervals between 16 and 30 kHz in an anechoic chamber. Measured 50<br />

cm from the sound source, the maximum presentation sound pressure level ranged from 105 to 112 dB depending on the frequency.<br />

To prevent the listener from detecting the quantization noise or subharmonic distortions at low frequencies, a pink noise was added as<br />

a masker. Using a 3-down 1-up transformed up-down method, thresholds were obtained at 26 kHz for 10 out of 32 ears. Even at 28<br />

kHz threshold values were obtained for 3 ears, but none were observed for a tone at 30 kHz. Above 24 kHz, the thresholds always<br />

exceeded 90 dB SPL. Between 16 and 20 kHz thresholds increased abruptly, whereas above 20 kHz the threshold increase was more<br />

gradual.<br />

8:55<br />

4aPP2. Do we have better hearing sensitivity than people in the past? Kenji Kurakata and Tazu Mizunami �Natl. Inst. of Adv.<br />

Industrial Sci. and Technol. �AIST�, 1-1-1 Higashi, Tsukuba, Ibaraki, 305-8566 Japan, kurakata-k@aist.go.jp�<br />

Our hearing sensitivity to tones of various frequencies declines progressively as we become older. ISO 7029 describes a method<br />

for calculating expected hearing threshold values as a function of age. However, more than 30 years have passed since the ISO<br />

standard source data were published. An earlier paper of the present authors �K. Kurakata and T. Mizunami, Acoust. Sci. Technol. 26,<br />

381–383 �2005�� compared hearing threshold data of Japanese people in recent years to the ISO standard values. The results of that<br />

comparison showed that the age-related sensitivity decrease of Japanese people was smaller, on average, than those described in the<br />

standard. A large discrepancy was apparent at 4000 and 8000 Hz: more than 10 dB for older males. In response to that inference, the<br />

ISO/TC43/WG1 ‘‘threshold of hearing’’ initiated a project in 2005 to explore the possibility of updating ISO 7029. This paper presents<br />

a summary of those comparison results of audiometric data and the work in WG1 for revising the standard.<br />

9:15<br />

4aPP3. Reliability and frequency specificity of auditory steady-state response. Masaru Aoyagi, Tomoo Watanabe, Tsukasa Ito,<br />

and Yasuhiro Abe �Dept. of Otolaryngol., Yamagata Univ. School of Medicine, 2-2-2 Iida-Nishi, Yamagata, 990-9585, Japan�<br />

The reliability and frequency specificity of 80-Hz auditory steady-state response �80-Hz ASSR� elicited by sinusoidally amplitudemodulated<br />

�SAM tones� tones and detected by phase coherence were evaluated as a measure of the hearing level in young children.<br />

The 80-Hz ASSR at a carrier frequency of 1000 Hz was monitored in 169 ears of 125 hearing-impaired children and auditory<br />

brainstem response �ABR� elicited by tone pips was examined in 93 ears. Both responses were examined during sleep, and the<br />

thresholds were compared with the behavioral hearing threshold, which was determined by standard pure-tone audiometry or play<br />

audiometry. In 24 ears with various patterns of audiogram, 80-Hz ASSRs were examined at different carrier frequencies, and the<br />

threshold patterns were compared with the audiograms to investigate the frequency specificity of ASSR. The correlation coefficient<br />

between the threshold of 80-Hz ASSR and pure-tone threshold (r�0.863) was higher than that for ABR (r�0.828). The threshold<br />

patterns of 80-Hz ASSR clearly followed the corresponding audiogram patterns in all types of hearing impairment. These findings<br />

suggest that 80-Hz ASSR elicited by SAM tones and detected by phase coherence is a useful audiometric device for the determination<br />

of hearing level in a frequency-specific manner in children.<br />

3245 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3245<br />

4a FRI. AM


9:35<br />

4aPP4. Use of perceptual weights to test a model of loudness summation. Lori J. Leibold and Walt Jesteadt �Boys Town Natl.<br />

Res. Hospital, 555 N 30th St., Omaha, NE 68131�<br />

We recently examined the contribution of individual components of a multitone complex to judgments of overall loudness by<br />

computing the perceptual weight listeners assign to each component in a loudness-matching task �Leibold et al., J. Acoust. Soc. Am.<br />

117, 2597 �2005��. Stimuli were five-tone complexes centered on 1000 Hz, with six different logarithmic frequency spacings, ranging<br />

from 1.012 to 1.586. When all components fell within the same critical band, weights varied little across components. In contrast, the<br />

range of weights increased with increasing frequency separation, with greater weight given to the lowest and highest frequency<br />

components. Perceptual weights were largely in agreement with the Moore et al. loudness model �J. Audio Eng. Soc. 45, 224–237<br />

�1997��, except at the widest bandwidth. In the current study we further examined predictions of the loudness model, focusing on the<br />

widest frequency-spacing condition. Masked thresholds and jnds for intensity discrimination were measured for each component and<br />

were compared to weights. The model predicts more interaction in the widely spaced conditions than simple critical band models, but<br />

underestimates the true interactions in conditions where components are widely spaced. Central factors appear to influence loudness,<br />

masked thresholds, and intensity discrimination in these conditions. �Work supported by NIH/NIDCD.�<br />

9:55<br />

4aPP5. Increased loudness effect at the absolute threshold of hearing. Junji Yoshida, Masao Kasuga �Grad. School of Eng.,<br />

Utsunomiya Univ., 7-1-2 Yoto, Utsunomiya-shi, Tochigi-ken, 321-8585, Japan�, and Hiroshi Hasegawa �Utsunomiya Univ.,<br />

Tochigi-ken, 321-8585, Japan�<br />

This study investigated the effects of a previous sound on loudness at the absolute threshold of hearing. Change of the absolute<br />

threshold of hearing was measured when a pure tone preceded the test tone in the measurement of the threshold. The previous sound<br />

at 60 dB SPL was presented first in one ear, followed by the presentation of the test sound in either the contralateral or ipsilateral ear<br />

at an interval of 0.5 s. Both the previous and test sounds had the same frequency of 500 Hz, and the same duration of 3 s. The change<br />

of the threshold was obtained from the difference between the thresholds with and without the previous sound. The threshold was<br />

found to be decreased significantly by approximately 2 dB when the previous sound was presented in the contralateral ear. On the<br />

other hand, the threshold was only slightly changed when the previous sound was presented in the ipsilateral ear. These experiment<br />

results suggest that loudness was increased by perceiving the previous sound in the contralateral ear.<br />

10:15–10:25 Break<br />

10:25<br />

4aPP6. Induced loudness reduction. Michael Epstein �Auditory Modeling and Processing Lab., Inst. for Hearing, Speech & Lang.,<br />

Dept. of Speech-Lang. Path. and Aud., Northeastern Univ., Boston, MA 02115� and Mary Florentine �Northeastern Univ., Boston,<br />

MA�<br />

Induced loudness reduction �ILR� is a phenomenon by which a preceding higher-level tone �an inducer tone� reduces the loudness<br />

of a lower-level tone �a test tone�. Under certain conditions, ILR can result in loudness reductions of 10 to 15 phons for pure tones.<br />

The strength of this effect depends on a number of parameters including: �1� the levels of both the inducer and test tones; �2� the<br />

frequency separation between the inducer and test tones; �3� the durations of the inducer and test tones; �4� the time separation<br />

between the inducer and test tones; �5� individual differences; and, possibly �6� the number of exposures to inducers. Because of the<br />

sensitivity of ILR to procedural conditions, it is quite important to consider the potential effects of ILR when considering any<br />

experimental design in which level varies. The understanding of ILR has also given insight into a number of unexplained discrepancies<br />

between data sets that were collected using different procedures. In fact, some of the variability known to affect loudness<br />

judgments may be due to ILR. �Work supported by NIH/NIDCD Grant R01DC02241.�<br />

10:45<br />

4aPP7. Loudness growth in individual listeners with hearing loss. Jeremy Marozeau and Mary Florentine �Commun. Res. Lab.,<br />

Inst. for Hearing, Speech & Lang., Northeastern Univ., Boston, MA 02115�<br />

Recent research indicates that there are large individual differences in how loudness grows with level for listeners with sensorineural<br />

hearing losses of primarily cochlear origin. Studies of loudness discomfort level suggest that loudness for most of these<br />

impaired listeners approaches the loudness experienced by the normal listeners at high levels. Loudness growth at low levels is still<br />

controversial. Although it is now well established that loudness at threshold is greater than zero, its exact value is unknown. If this<br />

value is the same for normal and impaired listeners, then the loudness growth for the impaired listeners must be steeper in order to<br />

catch up to normal at high levels. This phenomenon is called recruitment. If the loudness at threshold for impaired listeners is higher<br />

than that for normal listeners, then the impaired listeners will no longer be able to perceive sounds as soft. This phenomenon is called<br />

softness imperception. Results from two experiments suggest that: �1� individual differences are more important for impaired listeners<br />

than for normal listeners; �2� some impaired listeners seem to show recruitment, others softness imperception; and �3� averaging the<br />

results across the impaired listeners will mask these differences. �Work supported by NIH/NIDCD grant R01DC02241.�<br />

3246 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3246


11:05<br />

4aPP8. Growth of loudness in cochlear implant listening. Robert L. Smith and Nicole Sanpetrino �Inst. for Sensory Res. and<br />

Dept. of Biomed. and Chemical Eng., Syracuse Univ., 621 Skytop Rd., Syracuse NY, 13244�<br />

Cochlear implants �CIs� roughly mimic the transformation from sound frequency to cochlear place that occurs in acoustic hearing.<br />

However, CIs are relatively less capable of creating the intensive transformations that normal peripheral auditory processing provides.<br />

This is partly because CIs have small operating ranges on the order of 10:1 in electric current, compared to the 1 000 000:1 operating<br />

range for sound-pressure level �SPL� in acoustic hearing. Furthermore, loudness in acoustic hearing grows as a compressive power<br />

function of SPL. In contrast, loudness reportedly grows as a more expansive function of current for CI users, i.e., a power law with<br />

a large exponent or an exponential function. Our results, obtained using the minimally biased method of magnitude balance without<br />

an arbitrary standard, reveal a previously unreported range of shapes of CI loudness functions, going from linear to power laws with<br />

exponents of 5 or more. The shapes seem related in part to the etiology of deafness preceding cochlear implantation, although the<br />

shapes can vary with stimulating conditions within a subject. Furthermore, differential sensitivity to changes in current appears to be<br />

related to the shape of the corresponding loudness function. Implications for sound processing in electric and acoustic hearing will be<br />

discussed.<br />

FRIDAY MORNING, 1 DECEMBER <strong>2006</strong> MOLOKAI ROOM, 8:00 A.M. TO 12:00 NOON<br />

Session 4aSC<br />

Speech Communication: Perception „Poster Session…<br />

Katsura Aoyama, Cochair<br />

Texas Tech. Univ., Health Science Ctr., Dept. of Speech-Language and Hearing Science, Lubbock, TX 79430-6073<br />

Masato Akagi, Cochair<br />

JAIST, School of Information Science, 1-1 Asahidai, Nomi, Ishikawa 923-1292, Japan<br />

Contributed Papers<br />

All posters will be on display from 8:00 a.m. to 12:00 noon. To allow contributors an opportunity to see other posters, contributors of<br />

odd-numbered papers will be at their posters from 8:00 a.m. to 10:00 a.m. and contributors of even-numbered papers will be at their<br />

posters from 10:00 a.m. to 12:00 noon.<br />

4aSC1. Functional load of segments and features. Mafuyu Kitahara<br />

�School of Law, Waseda Univ., 1-6-1 Nishiwaseda, Shinjuku-Ku, Tokyo,<br />

Japan�<br />

The present paper proposes a new measure of functional load for segments<br />

and features. In a nut shell, it is based on word frequencies and the<br />

number of minimal pairs in which the relevant segment/feature is crucial<br />

in distinction. For example, minimal pairs distinguished only by /t/ are<br />

most frequent in English while those distinguished by /k/ are most frequent<br />

in Japanese. As for functional load of features, single-feature contrasts<br />

and multiple-feature contrasts are incorporated in the calculation. In<br />

Japanese, �high� alone distinguishes the largest number of minimal pairs<br />

while �voice� distinguishes words most frequently in cooperation with<br />

other features. Word frequency and familiarity database for English and<br />

Japanese are used to observe the commonalities and differences in both<br />

languages with respect to the proposed measure of functional load. This<br />

line of analysis suggests a better account for a certain phonological process<br />

being more typical in one language but not in the other. Functional<br />

load can be thought of as a part of the top-down information from the<br />

lexicon, which interacts with the bottom-up perceptual information in the<br />

process of spoken word recognition. Not only the ease of articulation and<br />

perceptual salience but also the functional load drives phonological processes.<br />

4aSC2. Cortical representation of processing Japanese phonemic and<br />

phonetic contrasts. Seiya Funatsu �The Prefectural Univ. of Hiroshima,<br />

1-1-71 Ujinahigashi, Minami-ku, Hiroshima, 734-8558 Japan�, Satoshi<br />

Imaizumi �The Prefectural Univ. of Hiroshima, Gakuen-machi Mihara,<br />

723-0053 Japan�, Akira Hashizume, and Kaoru Kurisu �Hiroshima Univ.,<br />

Minami-ku Hiroshima, 734-8551 Japan�<br />

This study investigated how Japanese speakers process phonemic and<br />

phonetic contrasts using voiced and devoiced vowel /u/ and /�u/. During<br />

six oddball experiments, brain responses were measured using magnetoencephalography.<br />

Under the phonemic condition, a frequent stimulus<br />

/ch�ita/ was contrasted with a deviant /ts�uta/, and a frequent /ts�uta/<br />

with a deviant /ch�ita/. Under the phonetic condition, a frequent /ts�uta/<br />

was contrasted with a deviant /tsuta/, and a frequent /tsuta/ with a deviant<br />

/ts�uta/. Under the segment condition, vowel segments, /�u/ and /u/,<br />

extracted from spoken words, were contrasted. The subjects were 13 native<br />

Japanese speakers. The equivalent current dipole moment �ECDM�<br />

was estimated from the mismatch field. Under the phonetic condition, the<br />

ECDM elicited by the voiced deviant was significantly larger than that<br />

elicited by the devoiced deviant in both hemispheres (p�0.01), while<br />

there were no significant deviant-related differences in ECDM under the<br />

phonemic condition in both hemispheres. Under the segment condition,<br />

the ECDM elicited by the voiced deviant and devoiced deviant did not<br />

differ significantly in either hemispheres. These results suggested that the<br />

3247 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3247<br />

4a FRI. AM


ECDM asymmetries between the voiced and the devoiced deviant observed<br />

under the phonetic condition did not originate from the acoustical<br />

difference itself, but from the phonetic environment.<br />

4aSC3. Evaluating a model to estimate breathiness in vowels. Rahul<br />

Shrivastav, Arturo Camacho, and Sona Patel �Dept. of Commun. Sci. and<br />

Disord., Univ. of Florida, Gainesville, FL 32611�<br />

The perception of breathiness in vowels is cued by changes in aspiration<br />

noise �AH� and the open quotient �OQ� �Klatt and Klatt, J. Acoust.<br />

Soc. Am. 87�2�, 820–857 �1990��. A loudness model can be used to determine<br />

the extent to which AH masks the harmonic components in voice.<br />

The resulting partial loudness �PL� and loudness of AH �noise loudness;<br />

NL� have been shown to be good predictors of perceived breathiness<br />

�Shrivastav and Sapienza, J. Acoust. Soc. Am. 114�1�, 2218–2224 �2005��.<br />

The levels of AH and OQ were systematically manipulated for ten synthetic<br />

vowels. Perceptual judgments of breathiness were obtained and regression<br />

functions to predict breathiness from NL/PL were derived. Results<br />

show breathiness to be a power function of NL/PL when NL/PL is<br />

above a certain threshold. This threshold appears to be affected by the<br />

stimulus pitch. A second experiment was conducted to determine if the<br />

resulting power function could be used to estimate breathiness in natural<br />

voices. The breathiness of novel stimuli, both natural and synthetic, was<br />

determined in a listening test. For comparison, breathiness for the same<br />

stimuli was also estimated using the power function obtained previously.<br />

Results show the extent to which findings can be generalized. �Research<br />

supported by NIH/R21DC006690.�<br />

4aSC4. Predicting vowel inventories: The dispersion-focalization<br />

theory revisited. Roy Becker �Dept. of Linguist., Univ. of California<br />

Los Angeles, 3125 Campbell Hall, Los Angeles, CA 90095-1543�<br />

A revision of the dispersion-focalization theory �DFT� �Schwartz<br />

et al., J. Phonetics 25, 233–253 �1997�� is presented. Like DFT, the current<br />

computational model incorporates the center of gravity effect �COG�<br />

of 3.5-Bark spectral integration, but it deviates from DFT in that the COG<br />

contributes to the actual values and reliability weights of the perceived<br />

formants of vowels. The COG is reinterpreted as a domain of acceleration<br />

towards formant merger: the percepts of formants less than 3.5 Barks apart<br />

are perturbed towards one another in a nonlinear yet continuous fashion<br />

and their weights are increased, but perceptual merger and weight maximization<br />

occur only when the acoustic distance is about 2 Bark. Like other<br />

dispersion-based models, inventories are evaluated predominantly according<br />

to the least dispersed vowel pair, where dispersion is measured as the<br />

weighted Euclidean distance between the vowels coordinates �the first two<br />

perceived formants�. Yet in the current model the weights are determined<br />

dynamically, in a well-principled manner. This model improves existing<br />

models in predicting certain universal traits, such as series of front<br />

rounded vowels in large vowel inventories, as emergent properties of certain<br />

local maxima of the inventory dispersion evaluation function, without<br />

sacrificing predictive adequacy for smaller inventories.<br />

4aSC5. Matching fundamental and formant frequencies in vowels.<br />

Peter F. Assmann �School of Behavioral and Brain Sci., Univ. of Texas at<br />

Dallas, Box 830688, Richardson, TX 75083�, Terrance M. Nearey �Univ.<br />

of AB, Edmonton, AB, Canada T6E 2G2�, and Derrick Chen �Univ. of<br />

Texas at Dallas, Richardson, TX 75083�<br />

In natural speech, there is a moderate correlation between fundamental<br />

frequency �F0� and formant frequencies �FF� associated with differences<br />

in larynx and vocal tract size across talkers. This study asks whether<br />

listeners prefer combinations of mean F0 and mean FF that mirror the<br />

covariation of these properties. The stimuli were vowel triplets �/i/-/a/-/u/�<br />

spoken by two men and two women and subsequently processed by Kawahara’s<br />

STRAIGHT vocoder. Experiment 1 included two continua, each<br />

containing 25 vowel triplets: one with the spectrum envelope �FF� scale<br />

factor fixed at 1.0 �i.e., unmodified� and F0 varied over �2 oct, the other<br />

with F0 scale factor fixed at 1.0 and FF scale factors between 0.63 and<br />

1.58. Listeners used a method of adjustment procedure to find the ‘‘best<br />

voice’’ in each set. For each continuum, best matches followed a unimodal<br />

distribution centered on the mean F0 or mean FF �F1, F2, F3� observed in<br />

measurements of vowels spoken by adult males and females. Experiment<br />

2 showed comparable results when male vowels were scaled to the female<br />

range and vice versa. Overall the results suggest that listeners have an<br />

implicit awareness of the natural covariation of F0 and FF in human<br />

voices.<br />

4aSC6. Acoustic cues for distinguishing consonant sequences in<br />

Russian. Lisa Davidson and Kevin Roon �Linguist. Dept., New York<br />

Univ., 719 Broadway, 4th Fl, New York, NY 10003,<br />

lisa.davidson@nyu.edu�<br />

In Russian, the same consonant sequences are permitted in various<br />

environments. Consequently, the presence of a word boundary or reduced<br />

vowel can be phonologically contrastive �e.g. �z.d0vat j ] ‘‘to assign,’’<br />

�zd0vat j ] ‘‘to turn in’’�, and both learners and experienced listeners likely<br />

rely on fine acoustic cues to discriminate the phonotactic structures they<br />

hear. In this study, the acoustic characteristics of consonant sequences are<br />

examined to establish which cues may distinguish �a� word-initial clusters<br />

�C1C2�; �b� consonant-schwa-consonant sequences �C1VC2�; and�c� sequences<br />

divided by a word boundary �C1#C2�. For all three sequence<br />

types, native Russian speakers produced phrases containing three categories<br />

of target sequences: stop�consonant, fricative�consonant, and nasal<br />

�consonant. Results show no significant differences in C1 burst duration<br />

for initial stops, though a longer interconsonantal duration is a reliable cue<br />

to schwa presence in C1VC2. C2 is significantly longer in C#C than in<br />

other sequences. For clusters, when C1 is a stop, there are no significant<br />

differences in duration with other sequence types, but other C1’s are significantly<br />

shorter. This suggests that articulatory overlap, which may lead<br />

to C1 shortening for fricative or nasal-initial clusters, is reduced in stopinitial<br />

clusters to ensure that the stop is released and recoverable. �Research<br />

supported by NSF.�<br />

4aSC7. Lipread me now, hear me better later: Crossmodal transfer of<br />

talker familiarity effects. Kauyumari Sanchez, Lawrence D.<br />

Rosenblum, and Rachel M. Miller �Dept. of Psych., Univ. of California,<br />

Riverside, Riverside, CA�<br />

There is evidence that for both auditory and visual speech perception<br />

�lipreading� familiarity with a talker facilitates speech recognition<br />

�Nygaard et al., Psychol. Sci. 5, 42�1994�; Yakel et al., Percept. Psychophys.<br />

62, 1405 �2000��. Explanations of these effects have concentrated on<br />

the retention of talker information specific to each of these modalities. It<br />

could be, however, that some amodal, talker-specific articulatory style<br />

information is retained to facilitate speech perception in both modalities. If<br />

this is true, then experience with a talker in one modality should facilitate<br />

perception of speech from that talker in the other modality. To test this<br />

prediction, subjects were given one hour of experience lipreading from a<br />

talker and were then asked to recover speech-in-noise from either this<br />

same talker or from a different talker. Results revealed that subjects who<br />

lipread and heard speech from the same talker performed better on the<br />

speech-in-noise task than did subjects who lipread from one talker and<br />

then heard speech from a different talker.<br />

4aSC8. Acoustic patterns of Japanese voiced velar stops. James<br />

Dembowski and Katsura Aoyama �Dept. Speech-Lang. & Hearing Sci.,<br />

Texas Tech Univ. Health Sci. Ctr., 3601 4th St., Lubbock, TX 79430-6073,<br />

james.dembowski@ttuhsc.edu�<br />

This presentation focuses on Japanese voiced velar /g/. The phoneme<br />

/g/ in VCV contexts is said to be characterized by a distinctive wedgeshaped<br />

formant pattern in which F2 and F3 converge toward one fre-<br />

3248 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3248


quency in the transition from vowel to stop closure, and then diverge as<br />

the vocal tract opens from the stop release to the following vowel. This<br />

pattern was examined using acoustic and kinematic data from an x-ray<br />

microbeam database of Japanese speakers, which is comparable to the<br />

English language x-ray microbeam speech production database �Hashi<br />

et al., J. Acoust. Soc. Am. 104, 2426–2437 �1998��. Japanese speakers<br />

produced the expected wedge-shaped formant pattern in isolated VCV<br />

nonsense syllables, but rarely, if ever, produced this pattern in connected<br />

speech. In contrast, English speakers more frequently produced the expected<br />

formant pattern in connected speech, though the pattern was less<br />

reliably present than in isolated VCVs and varied considerably within and<br />

across speakers. These observations highlight substantial differences between<br />

controlled laboratory speech and meaningful connected speech, as<br />

well as differences in the ways that phonemes are manifested by different<br />

linguistic communities. These data also potentially illuminate the relationship<br />

among phonetic, acoustic, and kinematic levels of speech production.<br />

4aSC9. Finding perceptual categories in multidimensional acoustic<br />

spaces. Eric Oglesbee and Kenneth de Jong �Dept. of Linguist., Indiana<br />

Univ., Bloomington, IN 47405, eoglesbe@indiana.edu�<br />

Examining phonetic categorization in multidimensional acoustic<br />

spaces poses a number of practical problems. The traditional method of<br />

forced identification of an entire stimulus space becomes prohibitive when<br />

the number and size of acoustic dimensions becomes increasingly large. In<br />

response to this, Iverson and Evans �ICPhS �2003�� proposed an adaptive<br />

tracking algorithm for finding best exemplars of vowels in a multidimensional<br />

acoustic space. Their algorithm converged on best exemplars in a<br />

relatively small number of trials; however, the search method took advantage<br />

of special properties of the vowel space in order to achieve rapid<br />

convergence. In this paper, a more general multidimensional search algorithm<br />

is proposed and analyzed for inherent biases. Then, using the algorithm,<br />

the phonetic categorization of /p/ and /b/ in a five-dimensional<br />

acoustic space by native speakers of English is tested. Results showed that<br />

�a� there were no substantial long-term biases in the search method and �b�<br />

the algorithm appeared to identify important acoustic dimensions in the<br />

identification of /p/ and /b/ using relatively few trials. �Work supported by<br />

NSF BCS-04406540.�<br />

4aSC10. On the perception of epenthetic stops in American English.<br />

Amalia Arvaniti, Ryan Shosted, and Cynthia Kilpatrick �Dept. of<br />

Linguist., UCSD, 9500 Gilman Dr., La Jolla, CA 92093,<br />

amalia@ling.ucsd.edu�<br />

This study examines the perception of epenthetic stops in American<br />

English. Stimuli were created from prince, prints, mince, mints, quince,<br />

and quints by removing all traces of /t/ and splicing in 0–72 ms of silence,<br />

in 12-ms steps, with or without a following burst. Subjects saw the minimal<br />

pairs on screen and selected the word they heard. It was hypothesized<br />

that stimuli with bursts and longer closure would result in more t responses<br />

�prince identified as prints� and that frequent words �prince/<br />

prints� would be more difficult to distinguish than infrequent words<br />

�quince/quints�, as our production results suggest that frequent pairs are<br />

more likely to be produced similarly. Results from 19 subjects show<br />

shorter response times with longer silence intervals, but no effect of burst<br />

or stimulus identity. Similarly, stimuli with bursts were not identified as<br />

nts words more frequently than those without. Generally, stimuli originating<br />

from nts words were more likely to be identified as such if they<br />

contained a burst, while the opposite was true for stimuli from nce words.<br />

Finally, frequent words were less likely to be correctly identified than<br />

infrequent words, suggesting that /t/ epenthesis is not as widespread<br />

throughout the lexicon as generally believed.<br />

4aSC11. Phonetic alignment to visual speech. Rachel M. Miller,<br />

Lawrence D. Rosenblum, and Kauyumari Sanchez �Dept. of Psych.,<br />

Univ. of California, Riverside, Riverside, CA 92521�<br />

Talkers are known to produce allophonic variation based, in part, on<br />

the speech of the person with whom they are talking. This subtle imitation,<br />

or phonetic alignment, occurs during live conversation and when a talker<br />

is asked to shadow recorded words �e.g., Shockley, et al., Percept. Psychophys.<br />

66, 422 �2004��. What is yet to be determined is the nature of the<br />

information to which talkers align. To examine whether this information is<br />

restricted to the acoustic modality, experiments were conducted to test if<br />

talkers align to visual speech �lipread� information. Normal-hearing subjects<br />

were asked to watch an actor silently utter words, and to identify<br />

these words by saying them out loud as quickly as possible. These shadowed<br />

responses were audio recorded and naive raters compared these<br />

responses to the actors auditory words �which had been recorded along<br />

with the actors visual tokens�. Raters judged the shadowed words as<br />

sounding more like the actors words than did baseline words, which had<br />

been spoken by subjects before the shadowing task. These results show<br />

that phonetic alignment can be based on visual speech, suggesting that its<br />

informational basis is not restricted to the acoustic signal.<br />

4aSC12. New anthropomorphic talking robot—investigation of the<br />

three-dimensional articulation mechanism and improvement of the<br />

pitch range. Kotaro Fukui, Yuma Ishikawa, Takashi Sawa, Eiji Shintaku<br />

�Dept. of Mech. Eng., Waseda Univ., 3-4-1 Ookubo, Shinjuku-ku, Tokyo,<br />

Japan�, Masaaki Honda �Waseda Univ., Saitama, Japan�, and Atsuo<br />

Takanishi �Waseda Univ., Shinjuku-ku, Tokyo, Japan�<br />

We developed a new three-dimensional talking robot WT-6 �Waseda<br />

Talker-No. 6�, which generates speech sounds by mechanically simulating<br />

articulatory motions as well as aeroacoustic phenomenon in the vocal<br />

tract. WT-6 consists of a 17-DOF mechanism �1-DOF lungs, 5-DOF vocal<br />

cords, 1-DOF jaws, 5-DOF tongue, and 4-DOF lips�. It has 3-D lips,<br />

tongue, jaw, and velum, which form the 3-D vocal tract structure. It also<br />

has an independent jaw opening/closing mechanism, which controls the<br />

relative tongue position in the vocal tract as well as the oral opening. The<br />

previous robot, which had a 2-D tongue �J. Acoust. Soc. Am. 117, 2541<br />

�2005��, was not enough to realize precise closure to produce humanlike<br />

consonants such as /s/ or /r/. The new tongue, which could be controlled to<br />

form the 3-D shape, makes it possible to produce more realistic vocal tract<br />

shape. The vocal cord model was also improved by adding a new pitch<br />

control mechanism pushing from the side of the vocal cords. The pitch<br />

range becomes broader than the previous robot, which is enough to reproduce<br />

normal human speech. Preliminary experimental results showed that<br />

synthesized speech quality improves for vowels /a/, /u/ and /o/. Some<br />

experimental results and video demonstration of the talking robot will be<br />

presented.<br />

4aSC13. The role of familiarity, semantic context, and amplitudemodulation<br />

on sentence intelligibility. Tom Carrell �Univ. of<br />

Nebraska—Lincoln, Lincoln, NE 68583, tcarrell@unl.edu� and Dawna<br />

Lewis �Boys Town Natl. Res. Hospital, Omaha, NE 68131�<br />

Amplitude modulation has been demonstrated to greatly improve the<br />

intelligibility of time-varying sinusoidal �TVS� sentences �Carrell and<br />

Opie, Percept. Psychophys. 52 �1992�; Barker and Cooke, Speech Commun.<br />

27 �1999�; Hu and Wang, Proceedings of ICASSP-02 �2002��. Ithas<br />

been argued that the improvement is due to a bottom-up process that<br />

causes the acoustically independent components of the sentences to be<br />

perceptually grouped for further analysis by the auditory system. A recent<br />

study �Shapley and Carrell, J. Acoust. Soc. Am. 118 �2005�� indicated that<br />

3249 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3249<br />

4a FRI. AM


semantic information did not influence intelligibility levels of TVS or<br />

modulated TVS sentences. In virtually all other studies in which speech<br />

was distorted or degraded its intelligibility was improved by appropriate<br />

semantic context �Miller, et al., JEP41 �1951��. It is possible that listeners’<br />

unfamiliarity with TVS speech might account for the difference. With<br />

one exception every study that has employed this type of stimulus provided<br />

the listener with very few practice sentences �Lewis, AAA �2005��.<br />

The present experiment manipulated listeners’ familiarity with TVS sentences<br />

to test this notion. Participants were presented with high- and lowpredictability<br />

TVS and modulated-TVS sentences. Familiarity had a large<br />

effect on perception and intelligibility. Interpretation of previous findings<br />

is reconsidered in this light.<br />

4aSC14. On the relation of apparent naturalness to phonetic<br />

perceptual resolution of consonant manner. Robert E. Remez, Claire<br />

A. Landau, Daria F. Ferro, Judith Meer, and Kathryn Dubowski �Dept. of<br />

Psych., Barnard College, 3009 Broadway, New York, NY 10027�<br />

How does the qualitative experience of speech influence phonetic perception?<br />

Our perceptual studies of consonant place and voicing have revealed<br />

a dichotomous relation between phonetic sensitivity and naturalness.<br />

Auditory quality and phonetic sensitivity sometimes co-vary, while<br />

in other conditions phonetic sensitivity is indifferent to huge variation in<br />

naturalness. New tests are reported extending the research to the dimension<br />

of manner, a contrast correlated with qualitatively distinct acoustic<br />

constituents in normal production. Speech synthesis techniques were used<br />

to create naturalness variants through �1� variation in the excitation of a<br />

synthetic voicing source and �2� variation in the bandwidth of the formant<br />

centers. Listeners calibrated the relative naturalness of items drawn from<br />

the test series, and the acuity of perceivers to the contrast between fricative<br />

and stop manner was estimated with cumulative d� across the series in<br />

identification tests. Combined with our prior findings, these new tests<br />

show how intelligibility and naturalness can be either perceptually orthogonal<br />

or contingent aspects of consonant dimensions, offering a tool to<br />

understand normative functions in speech perception. �Research supported<br />

by NIH �DC00308�.�<br />

4aSC15. Effects of signal levels on vowel formant discrimination for<br />

normal-hearing listeners. Chang Liu �Dept. of Commun. Sci. and<br />

Disord., Wichita State Univ., Wichita, KS 67260�<br />

The goal of this study was to examine effects of signal levels on vowel<br />

formant discrimination. Thresholds of formant discrimination were measured<br />

for F1 and F2 of four vowels in isolated vowels and sentences at<br />

three intensity levels: 70, 85, and 100 dB SPL for normal-hearing listeners<br />

using a three-interval, two-alternative forced-choice procedure with a twodown,<br />

one-up tracking algorithm. Results showed that formant thresholds<br />

were significantly affected by formant frequency, linguistic context, and<br />

signal levels. Thresholds of formant discrimination were increased as formant<br />

frequency increased and linguistic context changed from isolated<br />

vowels to sentences. The signal level indicated a rollover effect in which<br />

formant thresholds at 85 dB SPL are lower than at either 70 or 100 dB SPL<br />

in both isolated vowels and sentences. This rollover level effect could be<br />

due to reduced frequency selectivity and reduction in active cochlear nonlinearity<br />

at high signal levels for normal-hearing listeners. Excitation and<br />

loudness models will be explored to account for the level effect on formant<br />

discrimination of isolated vowels.<br />

4aSC16. Study on nonaudible murmur speaker verification using<br />

multiple session data. Mariko Kojima, Hiromichi Kawanami, Hiroshi<br />

Saruwatari, Kiyohiro Shikano �Nara Inst. of Sci. and Technol. 8916-5<br />

Takayama-cho Ikoma-shi, Nara, 630-0192 Japan�, and Tomoko Matsui<br />

�The Inst. of Statistical Mathematics, Minato-ku, Tokyo, 106-8569 Japan�<br />

A study on speaker verification with nonaudible murmur �NAM� segments<br />

using multiple session data was conducted. NAM is different from<br />

normal speech and is difficult for other people to catch. Therefore, a textdependent<br />

verification strategy can be used in which each user utters her/<br />

his own keyword phrase so that not only speaker-specific but also<br />

keyword-specific acoustic information is utilized. A special device called a<br />

NAM microphone worn on the surface of the skin below the mastoid bone<br />

is used to catch NAM because it is too low to be recorded using ordinary<br />

microphones. However, it is tolerant to exterior noise. This strategy is<br />

expected to yield relatively high performance. NAM segments, which consist<br />

of multiple short-term feature vectors, are used as input vectors to<br />

capture keyword-specific acoustic information well. To handle segments<br />

with a large number of dimensions, a support vector machine �SVM� is<br />

used. In experiments using NAM data uttered by 19 male and 10 female<br />

speakers in several different sessions, robustness against session-tosession<br />

data variation is examined. The effect of segment length is also<br />

investigated. The proposed approach achieves equal error rates of 0.04%<br />

�male� and 1.1% �female� when using 145-ms-long NAM segments.<br />

4aSC17. Sequential contrast or compensation for coarticulation?<br />

John Kingston, Daniel Mash, Della Chambless, and Shigeto Kawahara<br />

�Linguist. Dept., Univ. of Massachusetts, Amherst, MA 01003-9274�<br />

English listeners identify a stop ambiguous between /t/ and /k/ more<br />

often as ‘‘k’’ after /s/ than /sh/ �Mann and Repp, 1981�. Judgments shift<br />

similarly after a fricative ambiguous between /s/ and /sh/ when its identity<br />

is predictable from a transitional probability bias but perhaps not from a<br />

lexical bias �Pitt and McQueen, 1998; cf. Samuel and Pitt, 2003�. In replicating<br />

these experiments, we add a discrimination task to distinguish<br />

between the predictions of competing explanations for these findings: listeners<br />

respond ‘‘k’’ more often after /s/ because they compensate for the<br />

fronting of the stop expected from coarticulation with /s/ or because a stop<br />

with an F3 onset frequency midway between /t/’s high value and /k/’s low<br />

value sounds lower after the /s/’s high-frequency energy concentration.<br />

The second explanation predicts listeners will discriminate /s-k/ and /sh-t/<br />

sequences better than /s-t/ and /sh-k/ sequences because sequential contrast<br />

exaggerates the spectral differences between /s-k/’s high-low intervals<br />

and /sh-t/’s low-high intervals and distinguishes them more perceptually<br />

than /s-t/’s high-high intervals and /sh-k/’s low-low intervals.<br />

Compensation for coarticulation predicts no difference in discriminability<br />

between the two pairs because it does not exaggerate differences between<br />

the two intervals. �Work supported by NIH.�<br />

4aSC18. Acoustics and perception of coarticulation at a distance.<br />

Karen Jesney, Kathryn Pruitt, and John Kingston �Ling. Dept., Univ. of<br />

Massachusetts, Amherst, MA 01003-9274�<br />

CVC syllables were recorded from two speakers of American English,<br />

in which the initial and final stops ranged over /b,d,g/ and the vowel<br />

ranged over /i,I,e,E,ae,u,U,o,O,a/. F2 locus equations differed systematically<br />

as a function of the place of articulation of the other stop. These<br />

equation’s slope and y intercepts were used to synthesize initial /g-b/ and<br />

/g-d/ continua in CVC syllables in which the final stop ranged over /b,d,g/<br />

and the vowel over /e,o,a/, and the resulting stimuli were presented to<br />

listeners for identification. Listeners responded g significantly more often<br />

to both continua when the final stop was /d/ rather than /b/; the number of<br />

g responses fell between the /d/ and /b/ extremes for final /g/. This difference<br />

between final /d/ vs. /b/ is only observed when the intervening vowel<br />

is back /o,a/ and is actually reversed weakly when it is front /e/. Listeners<br />

also respond g significantly more often when the final stop is /g/ rather<br />

than /b/ and the vowel is /o,a/ but not �e�. Segments do coarticulate at a<br />

distance, listeners take this coarticulation into account, and perceptual adjustments<br />

depend on the segments through which the coarticulation is<br />

expressed. �Supported by NIH.�<br />

3250 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3250


4aSC19. Diphones, lexical access, and the verbal transformation<br />

effect. James A. Bashford, Jr., Richard M. Warren, and Peter W. Lenz<br />

�Dept. of Psych., Univ. of Wisconsin—Milwaukee, P.O. Box 413,<br />

Milwaukee, WI 53201-0413, bashford@uwm.edu�<br />

When listeners are presented with a repeating verbal stimulus, adaptation<br />

occurs and perception of the stimulus is replaced by perception of a<br />

competitor. The present study examines the first of these verbal transformations<br />

reported by 180 listeners who were presented with lexical and<br />

nonlexical consonantvowel �CV� syllables that varied in frequencyweighted<br />

neighborhood density �FWND�. These stimuli were produced by<br />

pairing the six English stop consonants with a set of three vowels. As<br />

expected, the majority of initial illusory forms �78%� were neighbors,<br />

differing from the stimulus by a single phoneme, and the proportion of<br />

lexical neighbors increased with stimulus FWND. Interestingly, FWND<br />

had opposite effects upon the lability of consonants and vowels: There was<br />

a strong positive correlation �r�0.79, F(17)�26.2, p�0.0001] between<br />

FWND and the number of consonant transformations, and in contrast,<br />

there was a strong negative correlation �r��0.78, F(17)�24.9, p<br />

�0.0001] between FWND and the number of vowel transformations. The<br />

implications of these and other findings with these simple diphones will be<br />

discussed in relation to current activation-competition theories of spoken<br />

word recognition. �Work supported by NIH.�<br />

4aSC20. Acoustic analysis and perceptual evaluation of nasalized ÕgÕ<br />

consonant in continuous Japanese. Hisao Kuwabara �Teikyo Univ. of<br />

Sci. & Technol., Uenohara, Yamanshi 409-0193, Japan�<br />

It is well known that the /g/ consonant, a velar voiced plosive, in<br />

Japanese continuous speech is often nasalized unless it appears at the<br />

word-initial position. The nasalized /g/ consonant takes place in dialects<br />

mainly spoken in northern districts including the Tokyo area where the<br />

standard Japanese is spoken. However, the number of nasalized /g/ consonants<br />

is said to be decreasing year by year according to a survey. This<br />

paper deals with acoustic and perceptual analysis of this phenomenon. Test<br />

materials used in this experiment are read version of Japanese short sentences<br />

by NHK’s �Japan Broadcasting Corporation� professional announcers.<br />

Each sentence includes at least one /g/ consonant that would likely be<br />

pronounced as nasalized. An evaluation test reveals that less than 60% of<br />

nasalization has been found to occur for /g/ consonants for which 100%<br />

nasalization had been observed decades ago. Acoustic analysis for nasalized<br />

and non-nasalized /g/ sounds has been performed mainly through<br />

waveform parameters. It has been found that the power ratio between<br />

consonant and vowel is the most effective parameter for distinguishing<br />

nasals from non-nasals, but it is highly speaker dependent.<br />

4aSC21. Production and perception of place of articulation errors.<br />

Adrienne M. Stearns and Stefan A. Frisch �Univ. of South Florida, 4202<br />

E. Fowler Ave., PCD1017, Tampa, FL 33620�<br />

Using ultrasound to examine speech production is gaining popularity<br />

because of its portability and noninvasiveness. This study examines ultrasound<br />

recordings of speech errors. In experiment 1, ultrasound images of<br />

participants’ tongues were recorded while they read tongue twisters. Onset<br />

stop closures were measured using the angle of the tongue blade and<br />

elevation of the tongue dorsum. Measurements of tongue twisters were<br />

compared to baseline production measures to examine the ways in which<br />

erroneous productions differ from normal productions. It was found that<br />

an error could create normal productions of the other category �categorical<br />

errors� or abnormal productions that fell outside the normal categories<br />

�gradient errors�. Consonant productions extracted from experiment 1<br />

were presented auditory-only to naive listeners in experiment 2 for identification<br />

of the onset consonant. Overwhelmingly, the participants heard<br />

normal productions and gradient error productions as the intended sound.<br />

Categorical error productions were judged to be different from the intended<br />

sound. The only effect of erroneous production on perception appears<br />

to be a slight increase in reaction time, which may suggest that error<br />

tokens are abnormal in some way not measured in this study.<br />

4aSC22. Role of linguistic experience on audio-visual perception of<br />

English fricatives in quiet and noise backgrounds. Yue Wang,<br />

Haisheng Jiang, Chad Danyluck �Dept. of Linguist., Simon Fraser Univ.,<br />

Burnaby, BC, V5A 1S6, Canada�, and Dawn Behne �Norwegian Univ. of<br />

Sci. and Technol., Trondheim, Norway�<br />

Previous research shows that for native perceivers, visual information<br />

enhances speech perception, especially when auditory distinctiveness decreases.<br />

This study examines how linguistic experience affects audiovisual<br />

�AV� perception of non-native �L2� speech. Native Canadian English<br />

perceivers and Mandarin perceivers with two levels of English<br />

exposure �early and late arrival in Canada� were presented with English<br />

fricative-initial syllables in a quiet and a caf-noise background in four<br />

ways: audio-only �A�, visual-only �V�, congruent AV, and incongruent AV.<br />

Identification results show that for all groups, performance was better in<br />

the congruent AV than A or V condition, and better in quiet than in cafnoise<br />

background. However, whereas Mandarin early arrivals approximate<br />

the native English patterns, the late arrivals showed poorer identification,<br />

more reliance on visual information, and greater audio-visual integration<br />

with the incongruent AV materials. These findings indicate that although<br />

non-natives were more attentive to visual information, they failed to use<br />

the linguistically significant L2 visual cues, suggesting language-specific<br />

AV processing. Nonetheless, these cues were adopted by the early arrivals<br />

who had more L2 exposure. Moreover, similarities across groups indicate<br />

possible perceptual universals involved. Together they point to an integrated<br />

network in speech processing across modalities and linguistic backgrounds.<br />

�Work supported by SSHRC.�<br />

4aSC23. Voicing of ÕhÕ in the Texas Instrument MIT „TIMIT…<br />

database. Laura Koenig �Haskins Labs, 300 George St., New Haven,<br />

CT 06511 and Long Island Univ.-Brooklyn�<br />

Although English /h/ is traditionally described as voiceless, authors<br />

have long recognized that voiced allophones exist, especially in unstressed,<br />

intervocalic positions. In past work, we have suggested that fully<br />

voiced /h/ may be more common in men than women, but our subject<br />

population was limited in number and dialectal diversity. In this study, we<br />

use the TIMIT database to obtain measures of /h/ production in men and<br />

women speaking multiple dialects of American English. Our analysis focuses<br />

on the /h/ initiating the word ‘‘had’’ in a sentence produced by all<br />

speakers in the database: ‘‘She had your dark suit...’’Eachtokenof/h/<br />

is coded for whether �a� the /h/ is deleted �i.e., auditorily imperceptible�;<br />

and, if /h/ is present, whether �b� voicing continues unbroken and �c� there<br />

is visible aspiration noise in the speech signal. This analysis will provide<br />

baseline data on /h/ realization in a common sentence context. We will<br />

also carry out follow-up analyses on selected utterances to gain more<br />

insight into the effects of phonetic context, stress, and lexical type �e.g.,<br />

content versus function word� on the characteristics of English /h/. �Work<br />

supported by NIH.�<br />

4aSC24. On distinctions between stops and similar-place weak<br />

fricatives. James M. Hillenbrand �Speech Pathol. and Audiol., MS5355,<br />

Western Michigan Univ., Kalamazoo, MI 49008� and Robert A. Houde<br />

�Ctr. for Commun. Res., Rochester, NY 14623�<br />

There is an extensive body of literature on the acoustic properties of<br />

both stops and fricatives. However, little attention has been paid to the<br />

acoustic features that might distinguish these two sound categories. This is<br />

unsurprising in the case of stops versus strong fricatives since these<br />

sounds are rarely confused. Stops and weak fricatives, on the other hand,<br />

are frequently confused �G.A. Miller and P.E. Nicely, J. Acoust. Soc. Am.<br />

27, 338–352 �1955��. The present study was undertaken in a search for<br />

acoustic features that might distinguish the stop/weak fricative pairs<br />

/b/–/v/ and /d/–/dh/ �i.e., /d/ vs voiced th�. Speech material consisted of<br />

CV and VCV syllables spoken by five men and five women, using the<br />

vowels /a/, /i/, and /u/. A combination of two features reliably separated<br />

3251 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3251<br />

4a FRI. AM


the stops from weak fricatives: �1� intensity during the consonant occlusion<br />

interval �typically greater for the fricatives�, and �2� the rate of increase<br />

in mid- and high-frequency energy �above 1 kHz� associated with<br />

consonant release �typically greater for the stops�.<br />

4aSC25. Salience of virtual formants as a function of the frequency<br />

separation between spectral components. Robert A. Fox, Ewa<br />

Jacewicz, Chiung-Yun Chang, and Jason D. Fox �Speech Acoust. and<br />

Percpt. Labs, Dept. of Speech and Hearing Sci., Ohio State, 110 Pressey<br />

Hall, 1070 Carmack Rd., Columbus, OH 43210-1002�<br />

The center-of-gravity �COG� hypothesis proposed by Chistovich and<br />

others for the perception of static vowels suggests that auditory spectral<br />

integration may occur when two or more formants fall within a 3.5 bark<br />

bandwidth. While several studies have examined the bandwidth limits of<br />

such integration, this study examines the extent to which spectral integration<br />

is uniform within this putative 3.5-bark range. We examine the perceptual<br />

salience of virtual formants produced by modifying the spectral<br />

COG of two closely spaced narrow-bandwidth resonances. Three different<br />

vowel series were created: �i-(�, �#-� and �.-�. A second set of vowels<br />

was then created in which one of the formants �F1 in �i-(�, F2in�#-� and<br />

F3 in �.-�� was replaced by a virtual formant whose COG matched that<br />

of the formant that had been removed. The frequency separation between<br />

the two component resonances was then systematically varied between 1.5<br />

and 3.5 barks and a singleinterval 2AFC vowel identification task was<br />

used to obtain estimates of vowel quality for each series step. Results will<br />

be discussed in terms of whether the spectral integration effects within the<br />

3.5 bark decline as the frequency separation between the resonance components<br />

increases. �Work supported by NIDCD R01DC00679-01A1.�<br />

4aSC26. Frequency effects in phoneme processing. Danny R. Moates,<br />

Noah E. Watkins, Zinny S. Bond, and Verna Stockmal �Inst. for the<br />

Empirical Study of Lang., Ohio Univ., Athens, OH 45701,<br />

moates@ohio.edu�<br />

Are phonological segments activated during word recognition in proportion<br />

to their frequency of use? Previous evidence for this hypothesis<br />

�Moates et al., Laboratory Phonology 7, edited by Gussenhoven and<br />

Warner �Mouton de Gruyter, 2002�� used a word reconstruction task. The<br />

present study used an online task, the gating task, in which progressively<br />

longer fragments of a word are presented to listeners who must identify<br />

the word in as few gates as possible. High- and low-frequency segments<br />

were contrasted by presenting them in word pairs that differed only in<br />

those two segments, e.g., label versus cable, where /l/ is used more frequently<br />

than /k/ �M. K. Vitevich and P. A. Luce, Behav. Res. Methods,<br />

Instrum. Comput. 36, 481–487 �2004��. We constructed 30 pairs of twosyllable<br />

words for which the contrasting segments were at the first syllable<br />

onset, 30 more for the second syllable onset, and 30 more for the coda of<br />

a syllable. Identification judgments were gathered from 120 participants. t<br />

tests showed high-frequency segments to be identified at significantly earlier<br />

gates than their matched low-frequency segments for first onset and<br />

coda, but not second onset. These results offer further evidence for sublexical<br />

processes in spoken word recognition.<br />

4aSC27. The clustering coefficient of phonological neighborhoods<br />

influences spoken word recognition. Michael Vitevitch �Dept. of<br />

Psych., Univ. of Kansas, 1415 Jayhawk Blvd., Lawrence, KS 66045-7556�<br />

Neighborhood density refers to the number of words, or neighbors,<br />

that are phonologically related to a given word. For example, the words<br />

BAT, MAT, CUT, and CAN �among others� are considered phonological<br />

neighbors of the word CAT. In contrast, the clustering coefficient of the<br />

neighborhood refers to the proportion of phonological neighbors that are<br />

also neighbors of each other. Among the neighbors of CAT, the words<br />

BAT and MAT are neighbors of each other, but the words BAT and CAN<br />

are not neighbors of each other. Despite the stimulus words having the<br />

same number of neighbors overall, the results of an auditory lexical decision<br />

task showed that words with a high clustering coefficient �i.e., most<br />

neighbors were also neighbors of each other� were responded to more<br />

quickly than words with a low clustering coefficient �i.e., few neighbors<br />

were also neighbors of each other�. These results suggest that some aspects<br />

of phonological similarity �i.e., clustering coefficient� might facilitate<br />

lexical activation, whereas other aspects of phonological similarity<br />

�i.e., neighborhood density� influence a later, decision stage of processing<br />

characterized by competition among activated word-forms. �Work supported<br />

by NIH.�<br />

4aSC28. The influence of noise and reverberation on vowel<br />

recognition: Response time. Magdalena Blaszak and Leon Rutkowski<br />

�Div. of Rm. Acoust. and Psychoacoustics, Adam Mickiewicz Univ.,<br />

Umultowska 85, 61-614 Poznan, Poland�<br />

This study examines the perceptual effect of two types of noise and<br />

reverberation on vowel recognition. Multitalker babble and traffic noise<br />

�European Standard EN 1793–3� were generated simultaneously with Polish<br />

vowels /a, e, i, o, u, y/ in two different sound fields and an anechoic<br />

chamber. The experiment was performed under various conditions of<br />

signal-to-noise ratio (�9, �6, �3, 0, �3, no noise�. A new procedure<br />

for listeners’ selection based on the Bourdon’s psychometrics test was<br />

proposed. The effects of noise and reverberation were quantified in terms<br />

of �a� vowel recognition scores for young normal-hearing listeners �YNH�<br />

and �b� ease of listening based on the time of response and subjective<br />

estimation of difficulty. Results of the experiment have shown that �a� the<br />

response time can be a good measure of the effect of noise and reverberation<br />

on the speech intelligibility is <strong>room</strong>s, and �b� in this type of experiment,<br />

of great significance is the choice of the subjects based on the<br />

psychometric tests.<br />

4aSC29. Quantifying the benefits of sentence repetition on the<br />

intelligibility of speech in continuous and fluctuating noises. Isabelle<br />

Mercille, Roxanne Larose, Christian Giguère, and Chantal Laroche<br />

�Univ. of Ottawa, 451 Smyth Rd., Ottawa, ON, Canada K1H 8M5�<br />

Good verbal communication is essential to ensure safety in the workplace<br />

and social participation during daily activities. In many situations,<br />

speech comprehension is difficult due to hearing problems, the presence of<br />

noise, or other factors. As a result, listeners must often ask the speaker to<br />

repeat what was said in order to understand the complete message. However,<br />

there has been little research describing the exact benefits of this<br />

commonly used strategy. This study reports original data quantifying the<br />

effect of sentence repetition on speech intelligibility as a function of<br />

signal-to-noise ratio and noise type. Speech intelligibility data were collected<br />

using 18 normal-hearing individuals. The speech material consisted<br />

of the sentences from the Hearing In Noise Test �HINT� presented in<br />

modulated and unmodulated noises. Results show that repeating a sentence<br />

decreases the speech reception threshold �SRT�, as expected, but<br />

also increases the slope of the intelligibility function. Repetition was also<br />

found to be more beneficial in modulated noises �decrease in SRT by 3.2<br />

to 5.4 dB� than in the unmodulated noise �decrease in SRT by 2.0 dB�.The<br />

findings of this study could be useful in a wider context to develop predictive<br />

tools to assess speech comprehension under various conditions.<br />

4aSC30. The effect of the spectral shape changes on voice perception.<br />

Mika Ito, Bruce R. Gerratt, Norma Antonanzas-Barroso, and Jody<br />

Kreiman �Div. of Head/Neck Surgery, UCLA School of Medicine, 31-24<br />

Rehab Ctr., Los Angeles, CA 90095-1794, jkreiman@ucla.edu�<br />

Researchers have long known that the shape of the vocal source spectrum<br />

is an important determinant of vocal quality, but the details regarding<br />

the importance of individual spectral features remains unclear. Previous<br />

research indicates four spectral features, H1-H2, the spectral slope above 4<br />

kHz, the slope from 1.5–2 kHz, and the slope from 2–4 kHz, account for<br />

3252 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3252


virtually all the variability in spectral shapes. The present study provides<br />

preliminary evidence about the perceptual importance of these four features.<br />

Four series of stimuli were synthesized for each spectral parameter,<br />

in which that parameter varied in small steps. Because the perceptual<br />

salience of source parameters depends on F0 and on the spectrum of the<br />

inharmonic part of the source, series differed in the sex of the speaker<br />

�male/female� and in the NSR �noise-free/very noisy�. Listeners heard all<br />

possible pairs of voices within each series and were asked to determine<br />

whether stimuli were the same or different. We hypothesize that listeners<br />

sensitivity to H1-H2 and the slope of the spectrum from 1.5–2 kHz will be<br />

independent of noise, but that sensitivity to changes in the spectral shape<br />

above 2 kHz will depend on the amount of noise excitation present in the<br />

voice.<br />

4aSC31. The use of auditory and visual information in the perception<br />

of stress in speech. James Harnsberger, Daniel Kahan, and Harry<br />

Hollien �Inst. for Adv. Study of the Commun. Proc., Univ. of Florida,<br />

Gainesville, FL 32611�<br />

Prior work on the acoustic correlates of the perception of psychological<br />

stress in speech has suffered from the problem of quantifying and<br />

verifying the extent to which a speaker was under stress during articulation.<br />

Two experiments were conducted to address this issue. First, stressed<br />

and unstressed speech samples were elicited from 78 speakers of American<br />

English. Stressed samples were recorded by having subjects read a<br />

standard passage while under the threat of the administration of mild electric<br />

shock. Both visual and audio recordings were collected. Stress was<br />

quantified in terms of four measures: two physiological �pulse rate and<br />

galvanic skin response� and two self-report scales. Sentences from the 16<br />

speakers showing the largest differences between the stressed and unstressed<br />

conditions were then presented in a paired comparison task to 90<br />

naive listeners, 30 each in three conditions: �1� audio-only presentation of<br />

the stimuli, �2� visual-only presentation of the stimuli, and �3� audiovisual<br />

presentation of the stimuli. The results indicate that individual listeners are<br />

sensitive to stress cues in speech in all three conditions.<br />

4aSC32. Selective attention and perceptual learning of speech.<br />

Alexander L. Francis �Dept. of Speech, Lang. & Hearing Sci., Purdue<br />

Univ., West Lafayette, IN 47907�, Natalya Kaganovich, and Courtney<br />

Driscoll �Purdue Univ., West Lafayette, IN 47907�<br />

Phonetic experience can change the perceptual distance between<br />

speech sounds, increasing both within-category similarity and betweencategory<br />

distinctiveness. Such warping of perceptual space is frequently<br />

characterized in terms of changes in selective attention: Listeners are assumed<br />

to attend more strongly to category-differentiating features while<br />

ignoring less relevant ones. However, the link between the distribution of<br />

selective attention and categorization-related differences in perceptual distance<br />

has not been empirically demonstrated. To explore this relationship,<br />

listeners were given 6hoftraining to categorize sounds according to one<br />

of two acoustic features while ignoring the other. The features were voice<br />

onset time and onset f 0, which are normally correlated and can both serve<br />

as a cue to consonant voicing. Before and after training, listener’s performance<br />

on a Garner selective attention task was compared with assessment<br />

of the perceptual distance between tokens. Results suggest that training<br />

can induce both warping of perceptual space and changes in the distribution<br />

of selective attention, but the two phenomena are not necessarily<br />

related. Results are consistent with a two-stage model of perceptual learning,<br />

involving both preattentive adjustment of acoustic cue weighting and<br />

higher-level changes in the distribution of selective attention between<br />

acoustic cues. �Work supported by NIH-NIDCD 1R03DC006811.�<br />

4aSC33. Investigation of interaction between speech perception and<br />

production using auditory feedback. Masato Akagi, Jianwu Dang,<br />

Xugang Lu, and Taichi Uchiyamada �School of Information Sci., JAIST,<br />

1-1 Asahidai, Nomi, Ishikawa 923-1292, Japan�<br />

This study employed an auditory feedback paradigm with perturbed<br />

fed-back speech to investigate interaction between speech perception and<br />

production by measuring simultaneous fluctuations of speech production<br />

organs using the electromyographic �EMG� signals, articulatory movements,<br />

as well as spectral analyses, where the articulatory data were obtained<br />

by the electromagnetic articulographic �EMA� system. Chinese<br />

vowels pair �i�-�y� and Japanese vowels pairs �e�-�a�, �e�-�i� and �e�-�u�<br />

were chosen as the experimental objects. When the speaker is maintaining<br />

the first vowel, the feedback sound is randomly changed from the first<br />

vowel to the second one in each pair by manipulating the first three formants.<br />

Spectral analysis showed that a clear compensation was seen in the<br />

first and second formants of the vowels. Analyses of EMG and EMA<br />

signals also showed muscle reactivation and tongue movements to compensate<br />

for the perturbations. Latency of the compensating response is<br />

about 150 ms to start and about 290 ms for maximum compensation from<br />

the onset of the perturbation. According to the measurements, it seems that<br />

in most cases the speaker attempts to compensate for the ‘‘error’’ caused<br />

by the auditory perturbation by a real-time monitoring, and the auditory<br />

feedback takes place simultaneously often during speech production.<br />

4aSC34. Cross-ear suppression of the verbal transformation effect:<br />

Tweaking an acoustic-phonetic level. Peter W. Lenz, James A.<br />

Bashford, Jr., and Richard M. Warren �Dept. of Psych., Univ. of<br />

Wisconsin—Milwaukee, P.O. Box 413, Milwaukee, WI 53201-0413,<br />

plenz@uwm.edu�<br />

A recorded word repeating over and over undergoes a succession of<br />

illusory changes. When two images of the same repeating word are presented<br />

dichotically, with a half-cycle delay preventing fusion, the two<br />

images of the word each undergo independent illusory transformations at a<br />

rate equivalent to that of a single image �Lenz et al., J. Acoust. Soc. Am.<br />

107, 2857 �2000��. However, with one phoneme difference �e.g., ‘‘dark’’<br />

versus ‘‘dart’’�, transition rate is dramatically suppressed �Bashford et al.,<br />

J. Acoust. Soc. Am. 110, 2658 �2001��. Rates decrease with extent of<br />

feature mismatch at a single phoneme position �roughly 30% reduction<br />

with one feature mismatch and 45% with three�. Rates also decrease with<br />

the number of mismatched phonemes �about 80% rate reduction with three<br />

out of four�, suggesting a strong acoustic-phonetic basis for verbal transformation<br />

suppression. In contrast, semantic relations had no effect �e.g.,<br />

transformations for ‘‘light’’ were suppressed equally by contralateral night<br />

and ‘‘might’’�. Dichotic competition appears to allow us to access and<br />

selectively influence a prelexical stage of linguistic analysis. �Work supported<br />

by NIH.�<br />

4aSC35. Perceptually balanced filter response for binaural dichotic<br />

presentation to reduce the effect of spectral masking. Pandurangarao<br />

N. Kulkarni, Prem C. Pandey �Elec. Eng. Dept, Indian Inst. of Technol.<br />

Bombay, Powai Mumbai 400076, India, pcpandey@ee.iitb.ac.in�, and<br />

Dakshayani S. Jangamashetti �Basaveshwar Eng. College Bagalkot,<br />

Bagalkot Karnataka 587102, India�<br />

Earlier investigations show that the scheme of binaural dichotic presentation<br />

with spectral splitting of speech signal helps in reducing the<br />

effect of spectral masking for persons with moderate bilateral sensorineural<br />

hearing impairment. Speech perception improved by employing filters<br />

with interband crossover gain adjusted between 4 and 6 dB below the pass<br />

band gain. The relationship between scaling factors for a tone presented to<br />

the two ears, so that perceived loudness is that of a monaural presentation,<br />

is investigated for design of comb filters with improved perceptually balanced<br />

response. Results from the listening tests show that, for perceptual<br />

balance, the sum of the two scaling factors should be constant, indicating<br />

that the magnitude response of the comb filters should be complementary<br />

on a linear scale.<br />

3253 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3253<br />

4a FRI. AM


4aSC36. The organization of bilingual perceptual consonant space:<br />

EnglishÕSpanish bilingual perception of Malayalam nasal consonants.<br />

Jenna Silver and James Harnsberger �Inst. for Adv. Study of the Commun.<br />

Proc., Univ. of Florida, Gainesville, FL 32611�<br />

This study examines the capacity of English/Spanish bilinguals to discriminate<br />

between consonants that exist in only one of their respective<br />

phonetic inventories. Two non-native nasal consonant contrasts were<br />

tested: dental versus alveolar and the palatal versus velar, both found in<br />

Malayalam. The dental and palatal nasals appear in Spanish, while the<br />

alveolar and velar nasals occur in English. Poorer performance in discrimination<br />

was interpreted as indicative of a common nasal category<br />

subsuming the Spanish dental and English alveolar nasals; better performance<br />

was taken as evidence of the maintenance of separate categories<br />

from both languages. Two other tests were administered to aid in the<br />

interpretation of the discrimination test scores: forced-choice identification<br />

and perceptual similarity ratings. The findings of this research will be used<br />

to characterize the perceptual consonant space in terms of continuum between<br />

two possible bilingual systems: one that collapses similar categories<br />

across languages or one that maintains two distinct phonological systems<br />

that can be accessed simultaneously. It is believed that bilinguals will be<br />

able to discriminate between these contrasts more consistently than their<br />

monolingual peers; however, there is no prediction about performance<br />

relative to the monolingual group from Malayalam.<br />

4aSC37. Agreement and reliability using reference-matching<br />

paradigm in perceptual voice quality rating in Chinese and English.<br />

Mandy Ho and Edwin Yiu �Voice Res. Lab., Div. of Speech & Hearing<br />

Sci., Univ. of Hong Kong, 5/F Prince Philip Dental Hospital, Hong Kong�<br />

Perceptual voice judgment is commonly used in clinical voice quality<br />

evaluation. The use of a referencematching paradigm in perceptual ratings<br />

has been shown to improve both agreement and reliability �Yiu et al., in<br />

press�. This study set out to investigate the agreement and reliability in<br />

rating Chinese and English dysphonic stimuli using the referencematching<br />

paradigm. Experiment 1 aimed to synthesize Chinese and English<br />

dysphonic stimuli with different breathy and rough severity levels<br />

using the HLSyn Speech Synthesizer. Seven representative anchors �references�<br />

for each of the rough and breathy series in Chinese and English<br />

were chosen by three judges to be used in experiment 2. Acoustic analysis<br />

of the anchor series indicated they were of increasing severity. Experiment<br />

2 recruited ten native Chinese and ten native English subjects to rate the<br />

quality of Chinese and English dysphonic voice samples using the synthesized<br />

anchor as references. Results showed that listeners achieved nearly<br />

90% agreement in rating the Chinese stimuli and 80% agreement in rating<br />

the English stimuli regardless of their language background. The study<br />

showed that the reference-matching paradigm was a reliable method in<br />

rating dysphonic stimuli across listeners with different language backgrounds.<br />

4aSC38. Learning to perceive non-native speech sounds: The role of<br />

test stimulus variability. McNeel Jantzen and Betty Tuller �Ctr. for<br />

Complex Systems and Brain Sci., Florida Atlantic Univ., 777 Glades Rd.,<br />

Boca Raton, FL 33431�<br />

Natural speech stimuli used in studies of phonological learning usually<br />

include several in talkers and phonetic environments because variability<br />

aids learning �e.g., Lively, Logan, and Pisoni, J. Acoust. Soc. Am. �1993��.<br />

The present study investigated whether nonphonetic variability in the synthetic<br />

test set has a similar effect. First, a perceptual mapping procedure<br />

was performed using a synthetic continuum that ranged from the Malayalam<br />

voiced, unaspirated, dental stop consonant to the American English<br />

alveolar �d�, with three F0 contours �low, mid, and high�. Subjects identified<br />

the stimuli �2AFC� and judged their goodness as exemplars of each<br />

category. Subjects then received 15 sessions �one/day� of 2AFC training<br />

with feedback using natural stimuli produced by native Malayalam speakers,<br />

and performed difference ratings on a subset of pairs from the syn-<br />

thetic stimuli. The perceptual mapping procedure was repeated at 1 and 14<br />

days post-training and results compared with a parallel experiment that<br />

included only the midlevel F0 contour in the synthetic test set. �Work<br />

supported by NSF.�<br />

4aSC39. Influence of the prosody of spoken language on recognition<br />

and memory for vocal quality. Sumi Shigeno �Aoyama Gakuin Univ.,<br />

4-4-25 Shibuya, Shibuya-ku, Tokyo, 150-8366 Japan�<br />

This study examined whether recognition and memory for vocal quality<br />

of a speaker who speaks either a native language or non-native language<br />

should be influenced by the prosody of the language that the<br />

speaker utters. Voices of 12 speakers were recorded. They were six Japanese<br />

people and six Americans and Britons. All speakers uttered short<br />

sentences in their respective native languages �i.e., Japanese for Japanese<br />

speakers and English for Americans and Britons� and in a non-native<br />

language �i.e., English for Japanese speakers and Japanese for Americans<br />

and Britons�. Ten Japanese participants rated the vocal quality of speakers<br />

in the first session. After 1 week the same experiment was again conducted<br />

in the second session. Results showed that the performance of identification<br />

of speakers as Japanese or as non-Japanese was comparatively accurate<br />

even though the ratings on the speakers’ voices were varied as the<br />

language spoken by the speakers. Ratings of the voices were compared<br />

further between two sessions and little difference was found, irrespective<br />

of a 1-week blank. Results suggest that the memory for vocal quality is<br />

robust, but that the recognition of vocal quality is dependent on the<br />

prosody of the language spoken by speakers.<br />

4aSC40. Brain activity during auditory processing affected by<br />

expectation of speech versus nonspeech. Yukiko Nota �ATR CIS<br />

BAIC, 2-2-2 Hikaridai, Keihanna Sci. City, Kyoto 619-0288, Japan,<br />

ynota@atr.jp�<br />

fMRI was used to clarify whether there is any differential brain activity<br />

invoked by expectation for speech versus nonspeech sounds. Auditory<br />

stimuli were created by acoustically morphing between either sustained<br />

vowels or tones, respectively, and a buzz sound. The two sets of interpolation<br />

were performed in nine nonlinear steps; the stimuli retained for<br />

perceptual experiments were only the three most vowel-like, the three<br />

most tone-like, and the three most buzz-like tokens morphed from the<br />

vowels. In the ‘‘speech expectation’’ session, subjects were instructed to<br />

discriminate the vowel-like and buzz-like stimuli; in the ‘‘nonspeech expectation’’<br />

session, subjects were instructed to discriminate the tone-like<br />

and buzz-like stimuli without knowing that the buzz stimuli had been<br />

morphed from the vowels. Thus the buzz-like stimuli in both experiments<br />

were the same, but the subjects’ expectations were different because they<br />

were told to expect either speech �vowel-like� or nonspeech �tone-like�<br />

stimuli. Comparison of brain activation during processing of the buzz-like<br />

stimuli under these two conditions revealed that BA40 and thalamus were<br />

more activated in speech expectation, while right BA20 was more activated<br />

in nonspeech expectation. These results suggest that subjects’<br />

speech/nonspeech expectation for sound stimuli influences brain activity<br />

for actual auditory processing. �Work supported by MEXT.�<br />

4aSC41. Representations involved in short-term versus long-term<br />

word learning by preschool children with and without phonological<br />

disorders. Holly Storkel, Jill Hoover, and Junko Maekawa �Dept. of<br />

Speech-Lang.-Hearing, Univ. of Kansas, 1000 Sunnyside Ave., 3001 Dole<br />

Ctr., Lawrence, KS 66045-7555, hstorkel@ku.edu�<br />

This study explores whether sublexical �i.e., individual sound� and/or<br />

lexical �i.e., whole-word� representations contribute to word learning and<br />

whether these contributions change across short-term versus long-term<br />

learning. Sublexical representations were indexed by phonotactic probability,<br />

the likelihood of occurrence of a sound sequence, whereas lexical<br />

representations were indexed by neighborhood density, the number of<br />

3254 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3254


similar sounding words. Thirty-four preschool children participated in a<br />

short-term word learning task that exposed them to nonwords varying in<br />

phonotactic probability and neighborhood density and tested learning of<br />

these nonwords. Long-term learning was assessed through comprehension<br />

and production of real words varying in phonotactic probability and neighborhood<br />

density. Results showed that phonotactic probability and neighborhood<br />

density equally influenced short-term word learning. In contrast,<br />

long-term word learning was affected primarily by neighborhood density.<br />

Thus, both sublexical and lexical representations appeared to play a role in<br />

short-term learning, but only lexical representations played a primary role<br />

in long-term retention. This pattern held for both children with normal<br />

phonological development and children with phonological delays. However,<br />

the direction of the effect of neighborhood density for short-term<br />

word learning varied by group status, suggesting differences in the use of<br />

lexical representations during short-term learning. �Work supported by<br />

NIH.�<br />

4aSC42. Changes in formant frequencies associated with postural<br />

change in adult male speakers over 50 years old. Michiko Hashi,<br />

Tomoki Nanto, and Natsuki Ohta �Prefectural Univ. of Hiroshima, 1-1<br />

Gakuen-cho, Mihara, Hiroshima, Japan�<br />

It is possible that changes of direction of gravity relative to the vocal<br />

tract associated with changes in posture influence acoustic characteristics<br />

of speech including vowel formant frequencies. Studies examining such<br />

effects had produced mixed results and demonstrated the possibility of<br />

substantive interspeaker variability in the effect of postural changes on<br />

vowel formant frequencies. Recent work by Takakura et al. �‘‘Changes in<br />

formant frequencies associated with postural change,’’ paper presented at<br />

the Fall meeting of Acoustical Society of Japan �<strong>2006</strong>��, using young adult<br />

male speakers, revealed a small number of speakers demonstrating<br />

changes in vowel formant frequencies and suggested effect of age. The<br />

present study attempts to examine changes of vowel formant frequencies<br />

in upright and supine position among older male speakers. Attempts will<br />

be made to eliminate the effect of differences in neck position between the<br />

postures through the use of a power-bead-based neck stabilizer. The results<br />

will be compared with data from young normal speakers in the previous<br />

study and inferences will be made relative to speech production models.<br />

4aSC43. The effect of viewing angle on the visual contribution to<br />

speech intelligibility in noise. Eugene Brandewie, Douglas Brungart,<br />

Nandini Iyer, and Brian Simpson �Air Force Res. Lab., Wright–Patterson<br />

AFB, Ohio 45433-7901�<br />

Visual cues are known to assist speech comprehension in noisy environments,<br />

but relatively little is known about the impact that viewing<br />

angle has on the visual contribution to speech intelligibility. In this experiment,<br />

four digital cameras were used to make simultaneous recordings of<br />

test phrases from the Modified Rhyme Test at four different viewing<br />

angles: 0, 45, 90, and 135 deg. These test phrases were used to measure<br />

the effect of viewing angle on the intelligibility of noise-masked speech<br />

stimuli that were presented with and without visual cues at seven different<br />

signal-to-noise ratios �SNRs�. When the face was viewed from the front,<br />

the visual cues provided an intelligibility improvement that was equivalent<br />

to a 6–10-dB increase in SNR. This visual benefit remained roughly constant<br />

for viewing angles up to 90 deg, but it dropped off rapidly �to less<br />

than 2 dB� when the viewing angle increased to 135 deg. The results<br />

suggest that the visual contribution to speech perception is severely impaired<br />

when the observer does not have access to an unobstructed view of<br />

the talker’s mouth.<br />

4aSC44. Towards estimation of Japanese intelligibility scores using<br />

objective voice quality assessment measures. Rui Kaga, Kazuhiro<br />

Kondo, Kiyoshi Nakagawa, and Masaya Fuzimori �Yamagata Univ.,<br />

Jounan 4-3-16, Yonezawa, 992-8510, Yamagata, Japan�<br />

We investigated the use of objective quality measures to estimate the<br />

intelligibility of Japanese speech. We initially focused on PESQ �perceptual<br />

evaluation of speech quality�, which is the state-of-the-art objective<br />

assessment method, and can estimate the mean opinion scores �MOS� at<br />

an extremely high accuracy. Since we can assume that speech quality is<br />

correlated with intelligibility, it should be possible to estimate the intelligibility<br />

from the estimated opinion scores or some of its derivatives. We<br />

tried to estimate the intelligibility of the Japanese Rhyme Test �DRT�. The<br />

DRT uses minimal word pairs whose initial phone differs by one phonetic<br />

feature. We estimated the MOS of the DRT word samples mixed with<br />

noise and tried to correlate this with the measured intelligibility. However,<br />

the estimated MOS showed no difference between phonetic features.<br />

However, the difference in the estimated MOS between the word pairs<br />

seems to differ by a phonetic feature for SNR above 10 dB, which suggests<br />

that the estimation of intelligibility by a phonetic feature may be<br />

possible. We also plan to selectively use the internal data used to calculate<br />

the MOS estimate for better intelligibility estimation.<br />

3255 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3255<br />

4a FRI. AM


FRIDAY MORNING, 1 DECEMBER <strong>2006</strong> HONOLULU ROOM, 8:00 TO 11:55 A.M.<br />

Session 4aSP<br />

Signal Processing in Acoustics, Underwater Acoustics, and Acoustical Oceanography: Adaptive Signal<br />

Processing<br />

Juan I. Arvelo, Jr., Cochair<br />

Johns Hopkins Univ., Applied Physics Lab., National Security Technology Dept., 11100 Johns Hopkins Rd.,<br />

Laurel, MD 20723-6099<br />

Kensaku Fujii, Cochair<br />

Univ. of Hyogo, School of Engineering and Graduate School of Engineering, 2-67 Shosha, Himeji, Hyogo 671-2201, Japan<br />

Chair’s Introduction—8:00<br />

Invited Papers<br />

8:05<br />

4aSP1. Adaptive beamforming for multipath environments. Henry Cox �Lockheed Martin, IS&S, AC3DI, 4350 North Fairfax<br />

Dr., Ste. 470, Arlington, VA 22203�<br />

Coherent multipaths present a significant challenge to most adaptive beamformers because they violate the common assumption of<br />

a rank-one plane wave or geometrically focused signal. When the multipath arrivals that characterize shallow water propagation are<br />

resolvable by the array’s aperture, the mismatch between the assumed and the true signal spatial structure causes signal suppression.<br />

If the amplitude and phase relationships among the various multipaths were known, they could, in principle, be included in a matched<br />

field beamforming approach. This is usually impractical due to inadequate knowledge of propagation parameters, especially bottom<br />

characteristics, and source/receiver motion. A generalization of the standard MVDR approach, called multirank MVDR, assumes only<br />

that the signal lies in a subspace of multiple rank rather than the usual rank-one assumption. An example of a subspace is a small fan<br />

of beams that cover the potential multipath directions. The signal may be rank one corresponding to fully coherent multipath or higher<br />

rank corresponding to incoherent or partially coherent multipath. The multirank approach is applied to the shallow water multipath<br />

problem and compared with the related technique of employing multiple linear constraints. Realistic simulations of alternative<br />

beamforming approaches for a large horizontal array are presented. �Work supported by ONR.�<br />

8:25<br />

4aSP2. Robust adaptive algorithm based on nonlinear error cost function for acoustic echo cancellation. Suehiro Shimauchi,<br />

Yoichi Haneda, and Akitoshi Kataoka �NTT Cyber Space Labs., NTT Corp., 3-9-11, Midori-cho, Musashino-shi, Tokyo, 180-8585,<br />

Japan, shimauchi.suehiro@lab.ntt.co.jp�<br />

Motivated by recent progress in the blind source separation �BSS� technique, a robust echo cancellation algorithm is investigated,<br />

which would inherently identify the echo path even during double-talk by separating the acoustic echo from the local speech. An<br />

adaptive filter has been introduced into acoustic echo cancellers to identify the acoustic echo path impulse response and generate the<br />

echo replica. However, most adaptive algorithms suffer from instability during double-talk. Although step-size control cooperating<br />

with a double-talk detector �DTD� is a promising approach to stop the adaptation temporarily during double-talk, it cannot handle the<br />

echo path change during double-talk. To overcome this problem, novel robust algorithms are derived by applying nonlinearity to the<br />

cost function of a conventional echo cancellation algorithm such as the normalized least mean squares algorithm �NLMS� or the affine<br />

projection algorithm �APA�. Using the simulation results, there is a discussion about how the robustness of the derived algorithms<br />

depends on the choice of the nonlinear function and the original algorithm.<br />

8:45<br />

4aSP3. Realistic modeling of adaptive beamformer performance in nonstationary noise. Bruce K. Newhall �Appl. Phys. Lab.,<br />

Johns Hopkins Univ., 11100 Johns Hopkins Rd., Laurel, MD 20723�<br />

Most adaptive beamformers �ABFs� operate under the assumption that the noise field is quasistationary. They estimate the present<br />

noise field by averaging, assuming stationarity over the estimation time. The adaptive beamformer responds to slow changes in the<br />

noise field across multiple estimation intervals. Unfortunately, in many low-frequency underwater sound applications, the shipping<br />

noise may change rapidly, due to nearby ship motion. This motion can be significant during the estimation interval and degrade ABF<br />

performance. A realistic model has been developed, including two effects of source motion on horizontal towed arrays. Bearing rate<br />

produces a differential Doppler shift across the array. Range rate produces an amplitude modulation as the multipath interference<br />

pattern shifts along the array. The ABF model begins with a realization of ship locations and motion based on historical shipping<br />

density. Each ship generates realistic random noise composed of tonals in a broadband background. The noise is propagated from each<br />

ship to each hydrophone by a normal mode model. For each time sample the position of each ship is updated and propagation<br />

recalculated. The ability of a variety of ABF algorithms to reduce shipping noise clutter is simulated and compared.<br />

3256 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3256


9:05<br />

4aSP4. Multichannel active noise control system without secondary path models using the simultaneous perturbation<br />

algorithm. Yoshinobu Kajikawa and Yasuo Nomura �Faculty of Enginnering, Kansai Univ., 3-3-35, Yamate-cho, Suita-shi, Osaka<br />

564-8680, Japan, kaji@joho.densi.kansai-u.ac.jp�<br />

This paper presents a novel multichannel active noise control �ANC� system without secondary path models. This ANC system<br />

uses a simultaneous perturbation algorithm as the updating algorithm and has an advantage that secondary path models �estimation of<br />

secondary paths� are not required, unlike the MEFX �multiple error filtered-X�-based ANC. This system can consequently control<br />

noise stably because there are not modeling errors that cause system instability. The computational complexity is also very small.<br />

Experimental results demonstrate that the proposed multi-channel ANC system can operate stably under the environment where the<br />

error microphones always move.<br />

9:25<br />

4aSP5. Two-microphone system using linear prediction and noise<br />

reconstruction. Hirofumi Nakano, Kensaku Fujii �Dept. of Comput.<br />

Eng., Univ. of Hyogo, 2167 Shosya, Himeji 671-2201, Japan,<br />

er06j025@steng.u-hyogo.ac.jp�, Tomohiro Amitani, Satoshi Miyata<br />

�TOA Corp., Takarazuka Hyogo 665-0043, Japan�, and Yoshio Itoh<br />

�Tottori Univ., Japan�<br />

This study proposes a new adaptive microphone system that is characterized<br />

by a linear prediction circuit inserted previous to a noise reconstruction<br />

filter corresponding to an adaptive delay used in conventional<br />

systems. This insertion provides various advantages to the adaptive microphone<br />

system. One is that steering a null for a noise source becomes<br />

possible irrespective of the incidence of speech signal. Another is that the<br />

new microphone system works as an omnidirectional microphone for the<br />

speech signal. Consequently, setting the microphones at arbitrary intervals<br />

is possible. For example, building one microphone in a handset and another<br />

into a telephone baseset becomes possible, which provides higher<br />

noise reduction effect. In practical use, microphone systems must function<br />

in reflective surroundings. In this study, such performance of the proposed<br />

system is verified first using computer simulations and then using an experimental<br />

system put in an ordinary <strong>room</strong>. This study also presents experimental<br />

results verifying that the proposed system can successfully reduce<br />

noise incident from the same direction as a speech signal, as well as<br />

crowd noise recorded in an airport.<br />

9:40<br />

4aSP6. Adaptive beamformer trade-off study of an expendable array<br />

for biologic vocalizations. Juan Arvelo �Appl. Phys. Lab., Johns<br />

Hopkins Univ., 11100 Johns Hopkins Rd., Laurel, MD 20723-6099�<br />

Adaptive beamformers exploit the ambient noise anisotropy to increase<br />

the array gain against this background, enhance target detection,<br />

and increase the resolution of the beam response. A prototype expendable<br />

array was developed and tested for high-frequency passive detection and<br />

localization of marine mammals. This latest array consists of vertical poly-<br />

�vinylidene fluoride� �PVDF� wire elements arranged in a horizontal 6<br />

�6 grid with the corner elements removed for a total of 32 lines. The<br />

length of the wires forms a vertical beam response that allows exploitation<br />

of the ambient noise directionality in elevation while the horizontal aperture<br />

provides full coverage in azimuth. The performance and computational<br />

demand of selected adaptive and conventional beamformers are<br />

compared in a trade-off study to determine their possible use in this expendable<br />

system. This trade-off study accounts for the demand of computational<br />

resources in addition to the predicted system performance as adjuncts<br />

to ocean observatories. �This effort is partly supported by JHU/APL<br />

and the Office of Naval Research �ONR�.�<br />

Contributed Papers<br />

9:55–10:10 Break<br />

10:10<br />

4aSP7. Adaptive matched field processing enhancements to forward<br />

sector beamforming. Jeffrey A. Dunne �Appl. Phys. Lab., Johns<br />

Hopkins Univ., 11100 Johns Hopkins Rd., Laurel, MD 20723�<br />

A study was undertaken to examine the potential benefit of adaptive<br />

matched field processing �AMFP� to the forward sector capability of<br />

single-line, twin-line, and volumetric arrays. Comparisons are made with<br />

conventional MFP �CMFP� and adapative and conventional plane-wave<br />

beamforming �APWB and CPWB� in order to assess the degree of ownship<br />

noise reduction obtainable and any corresponding improvement to the<br />

signal-to-noise ratio �SNR�. A minimum variance distortionless response<br />

beamformer using dominant mode rejection was implemented, applied to<br />

both uniform and distorted array shapes. Significant improvement over<br />

CMFP and CPWB in tracking and SNR was seen for modeled data in both<br />

cases, with the distorted array showing, not surprisingly, better left-right<br />

rejection capability. �Work was undertaken with support from the Defense<br />

Advanced Research Projects Agency �DARPA� Advanced Technology Office<br />

�ATO�.�<br />

10:25<br />

4aSP8. Vector sensor array sensitivity and mismatch: Generalization<br />

of the Gilbert-Morgan formula. Andrew J. Poulsen and Arthur B.<br />

Baggeroer �MIT, Cambridge, MA 02139, poulsen@mit.edu�<br />

The practical implementation of any sensing platform is susceptible to<br />

imperfections in system components. This mismatch or difference between<br />

the assumed and actual sensor configuration can significantly impact system<br />

performance. This paper addresses the sensitivity of an acoustic vector<br />

sensor array to system mismatch by generalizing the approach used by<br />

Gilbert and Morgan for an array with scalar, omnidirectional elements<br />

�E.N. Gilbert and S.P. Morgan, Bell Syst. Tech. J. 34, �1955��. As such, the<br />

sensor orientation is not an issue because it does not affect performance<br />

for an array of omnidirectional sensors. Since vector sensors measure both<br />

the scalar acoustic pressure and acoustic particle velocity or acceleration,<br />

the sensor orientation must also be measured to place the vector measurement<br />

in a global reference frame. Here, theoretical expressions for the<br />

mean and variance of the vector sensor array spatial response are derived<br />

using a Gaussian perturbation model. Such analysis leads to insight into<br />

theoretical limits of both conventional and adaptive processing in the presence<br />

of system imperfections. Comparisons of theoretical results and<br />

simulations are excellent. One noteworthy result is the variance is now a<br />

function of the steering angle. �Work supported by the PLUSNeT program<br />

of the Office of Naval Research.�<br />

3257 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3257<br />

4a FRI. AM


10:40<br />

4aSP9. Adaptive filtering using harmonic structure of voiced speech<br />

for reducing nonstationary known noise. Kenko Ota, Masuzo<br />

Yanagida �Doshisha Univ., 1-3, Tarata-Miyakodani, Kyotanabe, Kyoto,<br />

610-0321, Japan, etf1704@mail4.doshisha.ac.jp�, and Tatsuya Yamazaki<br />

�NICT, 619-0289, Kyoto, Japan, yamazaki@nict.go.jp�<br />

Proposed is an effective method for reducing nonstationary known<br />

noise. The objective of this research is to develop a scheme of preprocessing<br />

for speech recognition that keeps the same speech recognition rate<br />

even in a worse acoustic environment and to realize a TV control system<br />

using speech recognition. The basic idea of the proposed method is to<br />

estimate the frequency spectrum of noise including sounds from the TV<br />

itself and to remove noise components from the frequency spectrum of the<br />

received signal. A transfer function from the TV set to the microphone is<br />

calculated in an iterative manner estimating the noise signal at the microphone.<br />

Traditional ANC techniques do not directly use spectral features<br />

such as harmonic structure or fundamental frequency, for example. As the<br />

proposed method uses harmonic structure of vocalic segments in command<br />

speech, it is expected to have an advantage compared with the<br />

traditional adaptive methods. Results of evaluation using speech recognition<br />

show that the proposed method significantly improves speech recognition<br />

rate compared with conventional spectral subtraction. �Work supported<br />

by Knowledge Cluster Project, MEXT, and by Academic Frontier<br />

Project Doshisha University.�<br />

10:55<br />

4aSP10. Robust matched-field processor in the presence of<br />

geoacoustic inversion uncertainty. Chen-Fen Huang, Peter Gerstoft,<br />

and William S. Hodgkiss �Marine Physical Lab., Scripps Inst. of<br />

Oceanogr., UCSD, La Jolla, CA 92037-0238, chenfen@ucsd.edu�<br />

This presentation examines the performance of a matched-field processor<br />

incorporating geoacoustic inversion uncertainty. Uncertainty of geoacoustic<br />

parameters is described via a joint posterior probability distribution<br />

�PPD� of the estimated environmental parameters, which is found by<br />

formulating and solving the geoacoustic inversion problem in a Bayesian<br />

framework. The geoacoustic inversion uncertainty is mapped into uncertainty<br />

in the acoustic pressure field. The resulting acoustic field uncertainty<br />

is incorporated in the matched-field processor using the minimum variance<br />

beamformer with environmental perturbation constraints �MV-EPC�. The<br />

constraints are estimated using the ensemble of acoustic pressure fields<br />

derived from the PPD of the estimated environmental parameters. Using a<br />

data set from the ASIAEX 2001 East China Sea experiment, tracking<br />

performance of the MV-EPC beamformer is compared with the Bartlett<br />

beamformer using the best-fit model.<br />

11:10<br />

4aSP11. A study on combining acoustic echo cancelers with impulse<br />

response shortening. Stefan Goetze, Karl-Dirk Kammeyer �Dept. of<br />

Commun. Univ. of Bremen, Eng., D-28334 Bremen, Germany�, Markus<br />

Kallinger, and Alfred Mertins �Carl von Ossietzky-Univ., Oldenburg,<br />

D-26111 Oldenburg, Germany�<br />

In hands-free video conferencing systems acoustic echo cancelers<br />

�AECs� have to face the problem of very high-order impulse responses<br />

�IRs�, which have to be compensated. Time-domain algorithms for adaptation<br />

often suffer from slow convergence �as the NLMS algorithm, e.g.�<br />

or high computational complexity �e.g., the RLS�. On the other hand<br />

frequency-domain algorithms introduce undesired delays �S. Haykin, Filter<br />

Theory, 2002�. For high-quality hands-free systems IR shortening concepts<br />

and IR shaping concepts developed for listening <strong>room</strong> compensation<br />

�LRC� �M.Kallinger and A. Mertins, in Proc. Asilomar, 2005� can be applied<br />

to increase speech intelligibility for the near-end speaker. The aim of<br />

this study is the synergetic combination of LRC concepts with acoustic<br />

echo cancellation. For this scenario two different forms of concatenating<br />

the subsystems are possible. Either the AEC filter follows the LRC or vice<br />

versa. In the first case the equalization filter reduces the length of the<br />

effective IR seen by the AEC filter. Thus, shorter AEC filters can be used<br />

which results in faster convergence. However, an estimation algorithm for<br />

the <strong>room</strong> IR is necessary for the LRC subsystem. In the second case the<br />

AEC delivers an estimate of the <strong>room</strong> IR which can be used as an input for<br />

the LRC filter. Experimental results confirm the superiority of the new<br />

combined approach.<br />

11:25<br />

4aSP12. Krylov and predictive sequential least-squares methods for<br />

dimensionality reduction in adaptive signal processing and system<br />

identification. James Preisig and Weichang Li �Woods Hole Oceanogr.<br />

Inst., Woods Hole, MA 02543�<br />

Rapid time variation of the environment, a large number of parameters<br />

which need to be adjusted, and the presence of a reduced subset of the<br />

parameters that are relevant at any point in time create significant challenges<br />

for adaptive signal-processing algorithms in underwater acoustic<br />

applications. In applications such as underwater acoustic communications,<br />

the environment is represented by the ‘‘taps’’ of the time-varying impulse<br />

response. The instability of estimation algorithms or inability to track<br />

rapid channel fluctuations are among the problems that are encountered.<br />

An approach to addressing these challenges is to dynamically select a<br />

subspace in which the adjustment of taps takes place. Here, two algorithms<br />

for doing this are presented. The first is based upon using subspace<br />

basis vectors, which form a Krylov subspace with respect to the channel<br />

input correlation matrix and the channel input/output correlation vector.<br />

This method does not use a prediction residual error to select the basis<br />

vectors. A second algorithm is a new variant of the matching pursuit<br />

algorithm. In this case, ‘‘important’’ taps of the channel impulse response<br />

are selected to minimize a forward prediction residual error. The properties<br />

and performance of these two algorithms are presented and compared<br />

using simulation and field data.<br />

11:40<br />

4aSP13. Expectation maximization joint channel impulse response<br />

and dynamic parameter estimation and its impact on adaptive<br />

equalization. Weichang Li and James C. Preisig �Dept. of Appl. Ocean<br />

Phys. and Eng., Woods Hole Oceanograph. Inst., Woods Hole, MA 02543�<br />

Joint estimation of channel impulse response and its dynamic parameters<br />

using the expectation maximization �EM� algorithm and its MAP<br />

variant is derived for broadband shallow-water acoustic communication<br />

channels. Based on state-space channel modeling, the EM algorithms estimate<br />

the channel dynamic parameters from the sequence of channel impulse<br />

response estimates. The estimated parameters are then used in the<br />

Kalman smoother, which estimates the channel impulse response. The<br />

stability of the algorithm is shown to be related to an extended persistent<br />

excitation �EPE� condition, which requires that both the symbol sequence<br />

and the channel estimates be persistently exciting. Modified algorithms are<br />

proposed for broadband multipath channels to avoid the issue of insufficient<br />

excitation. Efficient suboptimal algorithms are also derived from the<br />

EM algorithms that alternatively estimate the parameter and the channel<br />

impulse response while allowing slow parameter variations. The performance<br />

of these channel estimation algorithms as well as their impact on<br />

the subsequent equalizer are demonstrated through experimental data<br />

analysis. �Work supported by ONR Ocean Acoustics.�<br />

3258 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3258


FRIDAY MORNING, 1 DECEMBER <strong>2006</strong> KAUAI ROOM, 7:55 TO 10:00 A.M.<br />

Session 4aUWa<br />

Underwater Acoustics: Sonar Performance<br />

Lisa M. Zurk, Cochair<br />

Portland State Univ., Electrical and Computer Engineering Dept., 1900 S. W. Fourth Ave., Portland, OR 97207<br />

Hiroshi Ochi, Cochair<br />

Japan Agency for Marine-Earth Science and Technology (JAMSTEC), 2-1-5 Natsushima-cho, Yokosuka,<br />

Kanagawa 237-0061, Japan<br />

8:00<br />

4aUWa1. Automatic detection performance comparisons of three<br />

different fluctuation-based signal processors. Ronald A. Wagstaff<br />

�Natl. Ctr. for Physical Acoust., 1 Coliseum Dr., Univ., MS 38677�<br />

The three most common fluctuation-based signal processor �FBP� algorithms<br />

achieve gain by exploiting either the reciprocal of the spectral<br />

amplitude, the log-differential amplitude, or aligned-phase angles. Two<br />

important features of these processors, for the underwater acoustics community,<br />

is their ability to detect and automatically identify signals which<br />

originated from submerged sources, and to provide unusually large signalto-noise<br />

ratio gains. Similar benefits are of interest to the atmosphere<br />

acoustic community. An example is the automatic detection and<br />

identification/classification of hostile airborne and ground vehicles by unattended<br />

ground sensors �UGS�. The three different generic types of FBP<br />

algorithms will be defined. The manner in which each exploits fluctuations<br />

to achieve gain will be explained. Corresponding performances will be<br />

compared using both underwater and atmosphere acoustic data. The ocean<br />

acoustic results came from towed array beamformed spectral data and will<br />

include spectral plots and grams. Corresponding single sensor spectral<br />

results will also be presented for atmosphere acoustic vehicle data. �Work<br />

supported by ARDEC and SMDC.�<br />

8:15<br />

4aUWa2. Evolution of modern fluctuation-based processing. Ronald<br />

Wagstaff �Natl. Ctr. for Physical Acoust., 1 Coliseum Dr., Univ., MS<br />

38677�<br />

Modern fluctuation-based processors �FBPs� are relatively new on the<br />

signal processing scene. They started in the mid-1980s with the realization<br />

that averaging the reciprocal acoustic powers and inverting the result back,<br />

i.e., the harmonic mean, could yield 6- to 8-dB signal-to-noise ratio gain<br />

over the corresponding average power. Because of its significant noise<br />

attenuation, this processor was designated WISPR, i.e., reduces noise to a<br />

whisper. WISPR had a unique, potentially more valuable, capability.<br />

Based on the decibel difference, or ratio, between the average power and<br />

WISPR, it could be determined whether the received signals were from<br />

ships, or from sources of sound that were submerged. After much time and<br />

experience with WISPR at sea, acquiring and processing towed array<br />

ocean acoustic data, and continuing data processing in the laboratory, the<br />

phenomena that were responsible for WISPRs performance, acoustic fluctuations<br />

generated near the sea surface, became better understood and<br />

WISPRs credibility increased. This led to the development of many other<br />

FBPs with similar capabilities, but with significantly enhanced performances.<br />

A brief account of post-WISPR development will be presented,<br />

including a description of the exploitable parameters, how they are used,<br />

and the range of gains that they achieve.<br />

Chair’s Introduction—7:55<br />

Contributed Papers<br />

8:30<br />

4aUWa3. A comprehensive unbiased third party evaluation of a signal<br />

processor for detecting submerged sources among clutter signals and<br />

noise. Ronald Wagstaff �Natl. Ctr. for Physical Acoust., 1 Coliseum Dr.,<br />

Univ., MS 38677�<br />

The Wagstaff’s integration silencing processor, WISPR’, was developed<br />

to detect and identify signals in the ocean from sources that are<br />

submerged well below the sea surface. WISPR is the type of signal processor<br />

that exploits the reciprocal of the spectral power amplitude, rather<br />

than the amplitude as the average power processor does. Processing the<br />

reciprocal of the power represented a significant departure in the prevailing<br />

signal processing philosophy that governed most conventional signal<br />

processing algorithms that were in use when WISPR first appeared on the<br />

scene several years ago. Furthermore, WISPR’s claimed submergedsource<br />

detection capability made it an attractive candidate for some high<br />

interest signal processing applications. Accordingly, one influential national<br />

organization considered its potential use in their mission and decided<br />

to commission a credible third party laboratory to conduct an unbiased<br />

evaluation of the WISPR processor. The emphasis was to be on its<br />

performance for automatic unalerted detection of signals from submerged<br />

sources. The techniques and evaluation methods used to test the WISPR<br />

processor will be described. The results of the evaluation will be presented,<br />

and the influence of those results on the development of other,<br />

more advanced, fluctuation-based processors will be discussed.<br />

8:45<br />

4aUWa4. The estimated ocean detector: Predicted performance for<br />

continuous time signals in a randomÕuncertain ocean. Jeffrey A.<br />

Ballard, R. Lee Culver, Leon H. Sibul �Appl. Res. Lab. and Grad.<br />

Program in Acoust., Penn State Univ. P.O. Box 30, State College, PA<br />

16804�, Colin W. Jemmott, and H. John Camin �Penn State Univ., State<br />

College, PA 16804�<br />

This paper addresses implementation of the maximum likelihood �ML�<br />

detector for passive SONAR detection of continuous time stochastic signals<br />

that have propagated through a random or uncertain ocean. We have<br />

shown previously that Monte Carlo simulation and the maximum entropy<br />

method can make use of knowledge of environmental variability to construct<br />

signal and noise parameter probability density functions �pdf’s� belonging<br />

to the exponential class. For these cases, the ML detector has an<br />

estimator-correlator and noise-canceller implementation. The estimatorcorrelator<br />

detector computes the conditional mean estimate of the signal<br />

conditioned on the received data and correlates it with a function of the<br />

received data, hence the name estimated ocean detector �EOD�. Here we<br />

derive the detector structure for continuous time stochastic signals and<br />

Gaussian noise and present receiver operating characteristic �ROC� curves<br />

for the detector as a function of the signal-to-noise ratio. �Work supported<br />

by ONR Undersea Signal Processing Code 321US.�<br />

3259 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3259<br />

4a FRI. AM


9:00<br />

4aUWa5. Echo detection enhancement using multiple guide sources in<br />

shallow water. David C. Calvo, Charles F. Gaumond, David M. Fromm,<br />

and Richard Menis �Naval Res. Lab., Acoust. Div. Code 7145, 4555<br />

Overlook Ave. SW, Washington, DC 20375�<br />

The use of a guide source has been proposed as a way of compensating<br />

for multipath by forming a spatial-temporal cross correlation of the received<br />

target and guide source signals across a vertical array in the frequency<br />

domain �Siderius et al., J. Acoust. Soc. Am. 102, 3439–3449�.<br />

This processing has the effect of creating a virtual receiver at the guide<br />

source position. In general, the performance of a virtual receiver degrades<br />

if the spatial integration is not carried out over the span of the array with<br />

significant signal. In our study, we have pursued an alternative approach of<br />

using guide sources which does not require this integration in general. The<br />

guide source signal is simply used as a matched filter. Although this does<br />

not correspond to a virtual receiver, it is useful as a means of improving<br />

active or passive detection of signals in noise. In general, the signal gain<br />

using this alternative technique is dependent on the guide source position.<br />

To compensate for this, we construct a separable-kernel-receiver filter<br />

bank using multiple randomly positioned guide source signals. Improvement<br />

of ROC curves in both passive and active scenarios is obtained using<br />

experimental and simulated data. �Work sponsored by ONR.�<br />

9:15<br />

4aUWa6. Incorporating environmental variability into received signal<br />

statistics. H. John Camin, R. Lee Culver, Leon H. Sibul �Appl. Res.<br />

Lab. and Grad. Program in Acoust., Penn State Univ., P.O. Box 30, State<br />

College, PA 16804�, Jeffrey A. Ballard, and Colin W. Jemmott �Penn<br />

State Univ., State College, PA 16804�<br />

We have developed a Monte Carlo-based method for estimating the<br />

variability of acoustic signal parameters caused by uncertain ocean environments.<br />

The method begins with a physics-based model for the environmental<br />

properties and uses the maximum entropy �MaxEnt� method to<br />

construct probability density functions �pdf’s� describing the measured<br />

deviations from the model mean. Random realizations of environmental<br />

variability, with proper depth correlation, are constructed from the pdf’s<br />

and added to the mean model parameters. A parabolic equation code �RAM�<br />

is used to propagate acoustic energy through each realization of the environment.<br />

Fourier synthesis is used to recreate the arrival structure. The<br />

method is demonstrated using measurements from the Strait of Gibraltar,<br />

which is a particularly complicated region dominated by strong tidal fluctuations<br />

and internal waves. During 1996, an international group carried<br />

out the Strait of Gibraltar Acoustic Monitoring Experiment �SGAME�, in<br />

which detailed environmental and 250-Hz acoustic data were collected.<br />

Here, pdf’s of the received signal level are compared with results of the<br />

Monte Carlo method to demonstrate performance. �Gibraltar data and SVP<br />

model provided by Chris Tiemann �ARL:UT� and Peter Worcester �SIO�.<br />

Work supported by ONR Undersea Signal Processing.�<br />

9:30<br />

4aUWa7. Motion compensation of multiple sources. Joung-Soo Park<br />

�Agency for Defense Development, P.O. Box18, Jinhae, Kyung-Nam,<br />

645-600, Korea�, Jae-Soo Kim �Korea Maritime Univ., Young-Do, Busan,<br />

Korea�, and Young-Gyu Kim �Agency for Defense Development, Jinhae,<br />

Kyung-Nam, 645-600, Korea�<br />

Matched field processing has a advantage of detection of multiple<br />

targets. But, if a strong interferer is moving fast near a quiet target, detection<br />

of the target is difficult due to the motion effect of the interferer. The<br />

motion of the interferer introduces energy spreading and results in poorer<br />

detection. A waveguide-invariant-based motion compensation algorithm<br />

was proposed to mitigate the motion effect of a dominant signal component,<br />

which is estimated by eigenvalue method. The eigenvalue method is<br />

good for a strong interferer, but not good for multiple targets. In this<br />

presentation, we will propose a steered beam processing method to mitigate<br />

the motion effect of multiple targets. We will verify the proposed<br />

method with numerical simulations and SwellEx96 data processing.<br />

9:45<br />

4aUWa8. Predicting sonar performance using observations of<br />

mesoscale eddies. Harry DeFerrari �Div. of Appl. Marine Phys.,<br />

RSMAS, Univ. of Miami, 4600 Rickenbacker Cswy, Miami, FL 33149�<br />

A predictive relationship has been observed between the location of<br />

offshore mesoscale eddies and the performance of active and passive sonar<br />

on the shallow water shelf area inside of the eddy. The passage of an eddy<br />

produces a prograde front that modifies acoustic propagation by two<br />

mechanisms. First, the density gradient serves as a conduit for offshore<br />

internal waves to propagate onto the shelf. A long-lived front can result in<br />

order of magnitude increases in potential energy of the internal wave field<br />

and corresponding increases in sound speed variability. Second, the circulation<br />

of the eddy produces a unique sound speed profile that is strongly<br />

downward refracting but has a nearly iso-velocity layer near the bottom<br />

owing to turbulent mixing. The shape of the profile closely approximates a<br />

hyperbolic cosine. Such a profile has mode group velocities that are equal<br />

for all refracted modes, thus producing strong focusing and a caustic at the<br />

depth of the source at all ranges. The experimental observations are confirmed<br />

with oceanographic and acoustic propagation models and, in turn,<br />

the models predict FOM fluctuations of as much as 15 dB for passive<br />

sonar and 24 dB for active sonar, depending on location of the eddy.<br />

3260 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3260


FRIDAY MORNING, 1 DECEMBER <strong>2006</strong> KAUAI ROOM, 10:15 A.M. TO 12:00 NOON<br />

Session 4aUWb<br />

Underwater Acoustics: Session in Honor of Leonid Brekhovskikh I<br />

William A. Kuperman, Cochair<br />

Scripps Inst. of Oceanography, Univ. of California, San Diego, Marine Physical Lab., MC0238, San Diego,<br />

La Jolla, CA 92093-0238<br />

Oleg A. Godin, Cochair<br />

NOAA, Earth System Research Lab., 325 Broadway, Boulder, CO 80305-3328<br />

Chair’s Introduction—10:15<br />

Invited Papers<br />

10:20<br />

4aUWb1. Phenomenon of Leonid Maximovich Brekhovskikh as a man and a scientist. Nikolay Dubrovskiy �Andreyev Acoust.<br />

Inst., Shvernika Ul 4, Moscow 117036, Russia�<br />

Leonid Maximovich Brekhovskikh made phenomenal contributions in acoustics: discovery of the underwater sound channel,<br />

development of the fundamental theory of wave propagation in layered media, and working out a tangent plane approximation in the<br />

wave scattering theory. Brekhovskikh contributed greatly to the organization of research teams and the dissemination of information<br />

on acoustics and oceanography through his popular books and lecturing. He also made a major breakthrough as a public figure and a<br />

statesman. He became the first Director of the Acoustics Institute at the age of 36. He served as a secretary of the Russian Academy<br />

of Sciences Branch involved in oceanography, geography, and atmospheric physics research. Brekhovskikh’s achievements in science<br />

and science leadership were marked by multiple highest USSR awards and many international awards. He became an Honorary<br />

Fellow of the Acoustical Society of America and the Russian Acoustical Society. He received the Lord Raleigh Medal for the<br />

discovery that preserved its urgency for 30 years. Brekhovskikh’s phenomenon is regarded here from the viewpoint of his personality<br />

as well as specific circumstances of his family and social life.<br />

10:40<br />

4aUWb2. Some aspects of Leonid Brekhovskikh’s influence on oceanographic acoustics. W. A. Kuperman and W. Munk<br />

�Scripps Inst. of Oceanogr., Univ. of California, San Diego, La Jolla, CA 92093-0238�<br />

Waveguide physics describes the basic features of long-range sound propagation in the ocean. Over the last half century the theory<br />

has progressed from describing ideal waveguides to more complicated layered structures to range-dependent structures to timevarying,<br />

range-dependent structures. The theme of Brekhovskikh’s pioneering work was the development of robust formulations that<br />

permitted understanding basic ocean acoustics while also laying the foundation to progress to the next levels of realistic complexity.<br />

Early on, he realized that acoustic data were not consistent with known oceanography. His seminal oceanographic experiments<br />

established the pervasive presence of mesoscale phenomena, which to this day are still not fully incorporated into rigorous formulations<br />

of the forward and inverse acoustics problems. We discuss only a very small part of his work and its subsequent influence.<br />

11:00<br />

4aUWb3. Underwater sound propagation: 49 years with L. M. Brekhovskikh’s Waves in Layered Media. Oleg A. Godin<br />

�CIRES, Univ. of Colorado and NOAA, Earth System Res. Lab., 325 Broadway, Boulder, CO 80305, oleg.godin@noaa.gov�<br />

In his first 10 years of research on wave propagation in layered media, L. M. Brekhovskikh created a theory that remains a basis<br />

for physical understanding and mathematical modeling of underwater sound propagation. Summarized in his celebrated book Waves<br />

in Layered Media, first published in 1957, the theory includes spectral �quasi-plane wave� representations of wave fields, normal mode<br />

theory for open waveguides, extensions of the ray and WKBJ methods, and a clear treatment of diffraction phenomena attendant to<br />

caustics, lateral waves, and reflection of wave beams and pulses. The book also charted the ways forward that have been and are<br />

followed by numerous researchers around the globe. Some of the resulting progress was documented in subsequent editions of Waves<br />

in Layered Media and in later books L. M. Brekhovskikh coauthored with his students. This paper will discuss diverse, groundbreaking<br />

contributions L. M. Brekhovskikh made to the wave propagation theory from the prospective offered by recent developments in<br />

underwater acoustics.<br />

11:20<br />

4aUWb4. L. M. Brekhovskikh’s studies on nonlinear wave interaction and atmospheric sound. Konstantin Naugolnykh �Univ.<br />

of Colorado, NOAA, ESRL/Zeltaech LLD, Boulder, CO�<br />

Nonlinear interaction of waves in a compressible fluid is an underlying factor in many geophysical effects, and L. M.<br />

Brekhovskikh made essential contributions to investigation of these phenomena. In particular, he suggested the mechanism of the<br />

infrasound generation by stormy areas in the ocean based on the nonlinear interaction of the counter-propagating sea-surface gravity<br />

3261 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3261<br />

4a FRI. AM


waves. The estimates of the order of magnitude of sound intensities were made indicating that the main part of the infrasound<br />

generated by the surface waves is absorbed in the upper layers of the atmosphere, resulting in the heating of these layers. The other<br />

part of the sound energy can be trapped by the atmospheric acoustic waveguide and then returned to earth at distances of hundreds of<br />

kilometers, producing the voice of the sea.<br />

11:40<br />

4aUWb5. Tangent-plane approximation by L. M. Brekhovskikh and connected methods in the theory of wave scattering from<br />

rough surfaces. Alexander G. Voronovich �NOAA, Earth System Res. Lab., Physical Sci. Div., 325 Broadway, Boulder, CO 80305,<br />

alexander.voronovich@noaa.gov�<br />

Starting from pioneering work by Rayleigh in 1907, scattering of waves from rough surfaces was restricted by the case of small<br />

Rayleigh parameter. In this case perturbation analysis describing the process of Bragg scattering applies. Apparently, smallness of the<br />

roughness is too restrictive for many applications. In 1952 L. M. Brekhovskikh suggested a tangent-plane approximation �TPA�. For<br />

ideal boundary conditions it represents the first iteration of the appropriate boundary integral equation. However, for more complex<br />

situations �e.g., dielectric or solid-fluid interfaces� appropriate boundary integral equations are rather complicated and, even worse,<br />

they cannot be readily iterated. The TPA allows bypassing this step providing the answer in closed form for arbitrary boundary<br />

conditions and for scalar or vector waves in terms of the local reflection coefficient. Unfortunately, the TPA does not correctly describe<br />

the Bragg scattering. However, later it was realized that the TPA allows simple generalization, which treats both low- and highfrequency<br />

limits within single theoretical scheme. This is achieved by considering the local reflection coefficient as an operator rather<br />

than a factor. New methods going beyond the two classical ones with much wider regions of validity were developed based on this<br />

idea. Some of them will be reviewed in this talk.<br />

FRIDAY AFTERNOON, 1 DECEMBER <strong>2006</strong> LANAI ROOM, 1:00 TO 4:45 P.M.<br />

Session 4pAA<br />

Architectural Acoustics: Measurement of Room Acoustics II<br />

Boaz Rafaely, Cochair<br />

Ben Gurion Univ., Electrical and Computer Engineering Dept., 84105 Beer Sheva, Israel<br />

Hideo Miyazaki, Cochair<br />

Yamaha Corp., Ctr. for Advanced Sound Technologies, 203 Matsunokijima, Iwata, Shizuoka 438-0192, Japan<br />

1:00<br />

4pAA1. Impulse response measurements based on music and speech<br />

signals. Wolfgang Ahnert, Stefan Feistel, Alexandru Miron, and Enno<br />

Finder �Ahnert Feistel Media Group, Berlin, Germany�<br />

All known software based measurement systems, including TEF,<br />

MLSSA, SMAART, and EASERA, derive results using predetermined excitation<br />

signals like Sweep, MLS, or Noise. This work extends the range<br />

of excitations to natural signals like speech and music. In this context<br />

selected parameters like frequency range, dynamic range, and fluctuation<br />

of the signal and the signal duration are investigated in order to reach<br />

conclusions about the conditions required to obtain results comparable<br />

with standard excitation signals. Also the limitations of the standard<br />

stimuli and the proposed natural stimuli are discussed.<br />

1:15<br />

4pAA2. Assessment of reverberation time in halls through analysis of<br />

running music. David Conant �McKay Conant Brook Inc., 5655<br />

Lindero Canyon Rd., Ste. 325, Westlake Village, CA 91362,<br />

dconant@mcbinc.com�<br />

The source signal to excite a <strong>room</strong>’s reverberant field sufficient for<br />

detailed measurement of reverberation time �RT60� and other measures<br />

has been the subject of considerable investigation over several decades. It<br />

is generally acknowledged that the best sources are �depending on the<br />

researcher� swept tones, MLS, MLS variations, stopped noise, cannon<br />

shots, etc. All can be characterized as highly audience unfriendly. In the<br />

Contributed Papers<br />

interest of obtaining useful approximations of measured midfrequency<br />

RT60 in the presence of live audiences, this paper discusses several approaches<br />

that may be fruitful while being entirely unobtrusive to the concert<br />

experience.<br />

1:30<br />

4pAA3. Comparison of measurement techniques for speech<br />

intelligibility. Bruce C. Olson �Olson Sound Design, 8717 Humboldt<br />

Ave N, Brooklyn Park, MN 55444, bco@olsonsound.com�<br />

A comparison of measurement techniques for speech intelligibility between<br />

two recently released measurement systems is made. EASERA<br />

�Electronic and Acoustic System Evaluation and Response Analysis� uses<br />

a standard PC and an EASERA Gateway interface attached via Firewire.<br />

The software postprocesses a variety of stimuli in order to derive the<br />

impulse response for the <strong>room</strong> under test. This impulse response is then<br />

further processed and the results are presented to the user in both graphical<br />

and textual presentations. The Ivie Technologies IE-35 is based on a<br />

Pocket PC system and uses an external modulated noise source as stimulus<br />

to produce an intelligibility score as a single number or average of a series<br />

of measurements. This paper will explore a variety of measurements made<br />

in the same locations in a <strong>room</strong> by both systems. Results will also be<br />

shown for a variety of other acoustic measures that quantify the acoustical<br />

parameters of the <strong>room</strong>.<br />

3262 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3262


1:45<br />

4pAA4. Under-balcony acoustics in concert halls: Single source versus<br />

an array of multiple sources. Youngmin Kwon and Gary W. Siebein<br />

�Architecture Technol. Res. Ctr., Univ. of Florida, 134 ARCH, P.O. Box<br />

115702, Gainesville, FL 32611, ymkwon@ufl.edu�<br />

The conventional measurement protocol using a single omnidirectional<br />

sound source may have limits or uncertainty in objective acoustical analysis<br />

of a performance hall. This study conducted monaural and binaural<br />

impulse response measurements with an array of 16 directional loudspeakers<br />

for quantitative acoustical assessment of specifically under-balcony<br />

area in a concert hall. The measurements were executed in a real performance<br />

hall. The measured time- and frequency-domain responses as well<br />

as the results of <strong>room</strong> acoustical parameters including binaural parameters<br />

were compared to the ones measured with a single omnidirectional source.<br />

The results were also compared to the ones taken at the main orchestra<br />

seating area. The time-domain responses showed a clear distinction particularly<br />

in early responses between single source and multiple sources.<br />

On the other hand, the magnitude of frequency response showed significantly<br />

lower at frequencies above 1 kHz than the one measured at the<br />

main area. The results of a binaural parameter, IACC, were found to be<br />

marginal between single source and multiple sources but critically different<br />

between under-balcony area and main area. Variations were also observed<br />

in the results of other <strong>room</strong> acoustical parameters when compared<br />

either between single source and multiple sources or between underbalcony<br />

area and main area.<br />

2:00<br />

4pAA5. Alternative metrics for the directivity of acoustic sources.<br />

Timothy W. Leishman �Acoust. Res. Group, Dept. of Phys. and Astron.,<br />

Brigham Young Univ., Provo, UT 84602�<br />

While the directivity of an acoustic source at a given frequency is<br />

thoroughly characterized by a directivity function over the angular coordinates,<br />

it may also be characterized to a lesser degree by a single-number<br />

directivity factor. The directivity index �i.e., the logarithmic version of the<br />

directivity factor� is a related figure of merit. Recent efforts to quantify the<br />

directivity of sources for architectural acoustics measurements have led to<br />

several alternatives to these values. One example is the area-weighted<br />

spatial standard deviation of radiated levels over a free-field measurement<br />

sphere. This paper presents and compares this and other directivity metrics<br />

for several types of sources, and discusses their benefits.<br />

2:15<br />

4pAA6. Room volume estimation from diffuse field theory. Martin<br />

Kuster and Maarten van Walstijn �Sonic Arts Res. Ctr., Queen’s Univ.<br />

Belfast, BT7 1NN Belfast, Northern Ireland, m.kuster@qub.ac.uk�<br />

Among the parameters relevant in <strong>room</strong> acoustics, the <strong>room</strong> volume is<br />

one of the most important. The general course in <strong>room</strong> acoustics research<br />

is to use the <strong>room</strong> volume in the prediction of <strong>room</strong> acoustic parameters<br />

such as reverberation time or total relative sound pressure level. Contrary<br />

to this, it has been investigated to what extent the <strong>room</strong> volume can be<br />

retrieved from a measured <strong>room</strong> impulse response. The approach followed<br />

is based on <strong>room</strong> acoustic diffuse field theory and requires correctly measured<br />

<strong>room</strong> impulse responses with the initial time delay corresponding to<br />

the source to receiver distance. A total of ten <strong>room</strong>s of varying size and<br />

acoustic characteristics have been included. The results in three <strong>room</strong>s<br />

were unreliable, which was explained by the particular acoustic characteristics.<br />

In the remaining <strong>room</strong>s the results were numerically useful and<br />

consistent between different positions within the same <strong>room</strong> �relative standard<br />

deviation around 20%�. The influence of source and receiver directivity<br />

is also considered.<br />

2:30<br />

4pAA7. In situ measurements for evaluating the scattering surfaces in<br />

a concert hall. Jin Yong Jeon and Shin-ichi Sato �School of<br />

Architectural Eng., Hanyang Univ., Seoul 133-791, Korea,<br />

jyjeon@hanyang.ac.kr�<br />

Sound diffusion by a wall structure is one of the main concerns with<br />

respect to the sound quality of concert halls. There is a need to develop<br />

measurement and evaluation methods for determining the performance of<br />

scattering wall surfaces not only in a laboratory but also in actual halls. In<br />

this study, the acoustical measurements were conducted in a concert hall<br />

which has diffusers with ceramic cubic tiles on the side walls of the stage<br />

and the audience area. Binaural impulse responses were measured at all of<br />

the seats under two conditions, that is, with and without diffusers. The area<br />

which was affected by the diffusive wall was determined and quantified.<br />

The condition without diffusers was produced by covering them with the<br />

movable reflectors. From the binaural impulse responses, the temporal<br />

diffusion �H. Kuttruff, Room Acoustics, �Elsevier Science, London,<br />

1991��, which is calculated from the autocorrelation of the impulse response,<br />

and other acoustical parameters were analyzed. From the relationship<br />

between the scattering coefficient and the acoustical parameters,<br />

sound scattering index for real halls, which represents the degree of the<br />

diffusion of a hall, was proposed.<br />

2:45<br />

4pAA8. Further investigations on acoustically coupled spaces using<br />

scale-model technique. Zuhre Su, Ning Xiang �Grad. Program in<br />

Architectural Acoust., School of Architecture, Rensselaer Polytechnic<br />

Inst., Troy, NY 12180�, and Jason E. Summers �Naval Res. Lab.,<br />

Washington, DC 20024�<br />

Recently, architectural acousticians have been increasingly interested<br />

in halls that incorporate coupled-volume systems because of their potential<br />

for creating nonexponential sound energy decay. Effects of couplingaperture<br />

configuration and source and receiver locations on energy decay<br />

are essential aspects of acoustically coupled spaces that have not yet been<br />

extensively investigated. In order to further understand these effects on<br />

sound fields in coupled <strong>room</strong>s, a systematic experimental study is carried<br />

out. An acoustic scale model technique is used in collecting <strong>room</strong> impulse<br />

responses of a two-<strong>room</strong> coupled system for varying aperture configurations<br />

and surface-scattering conditions. Baseline behavior is established by<br />

varying aperture area for a fixed aperture shape and analyzing relevant<br />

energy-decay parameters at different locations. Effects of aperture shape<br />

and number are systematically investigated by varying these parameters<br />

while holding coupling area fixed. Similarly, effects of receiver location<br />

are systematically investigated by varying the distance of the receiver<br />

from the coupling aperture for a fixed aperture configuration. Schroeder<br />

decay-function decompositions by Bayesian analysis reveal sensitivities to<br />

receiver location and aperture configuration across different frequency<br />

bands.<br />

3:00–3:15 Break<br />

3:15<br />

4pAA9. Virtual microphone control: A comparison of measured to<br />

created impulse responses of various microphone techniques. Daniel<br />

Valente and Jonas Braasch �Rensselaer Polytechnic Inst., 110 8th St.,<br />

Troy, NY 12180, danvprod@yahoo.com�<br />

A method of rendering sound sources in 3-D space has been developed<br />

using virtual microphone control �ViMiC� �J. Acoust. Soc. Am. 117,<br />

2391�. This method has been used to create a flexible architecture for the<br />

creation and rendering of a virtual auditory environment based on microphone<br />

techniques. One of the advantages of ViMiC is the ability to simulate<br />

coincident, near-coincident, and spaced microphone recording techniques.<br />

This allows the user active spatial control over the recorded<br />

environment and the ability to shape the final rendering based on his or her<br />

specific auditory needs. In order to determine the accuracy of simulating<br />

the virtual microphone techniques, measurements of several acoustic<br />

3263 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3263<br />

4p FRI. PM


spaces in Troy, NY will be compared to the measurements of generated<br />

impulse responses of the same modeled spaces within the ViMiC environment.<br />

The data from the measured impulse responses will be used to adapt<br />

the ViMiC system in order to create a more realistic auditory rendering.<br />

Moreover, the ViMiC system can be improved for use as an educational<br />

tool for teaching recording engineers to hear the subtle differences between<br />

various microphone techniques.<br />

3:30<br />

4pAA10. The estimation of the <strong>room</strong> acoustic characteristic using the<br />

acoustic intensity method. Yong Tang, Hideo Shibayama, and Takumi<br />

Yosida �Dept. of Commun. Eng., Shibaura Inst. of Technol., 3-7-5 Toyosu<br />

Koutou-ku, Tokyo, 135-8548 Japan, m603101@sic.shibaura-it.ac.jp�<br />

When a sound radiates in <strong>room</strong>s, a lot of reflection sounds are generated.<br />

From estimation of the direction where the <strong>room</strong> reflection sound<br />

comes from, we can understand the diffusion situation in the <strong>room</strong> acoustic<br />

field. By using the acoustic intensity method, we can measure the<br />

strength and the direction of the sound. In this paper, we estimate the<br />

direction of the reflection sound in the time-space by the acoustic intensity<br />

method and show the acoustic characteristic of the <strong>room</strong>.<br />

3:45<br />

4pAA11. Binaural simulation in an enclosure using the phased beam<br />

tracing. Cheol-Ho Jeong and Jeong-Guon Ih �NOVIC, Dept. of Mech.<br />

Eng., KAIST, Sci. Town, Daejeon 305-701, Korea, chjeong@kaist.ac.kr�<br />

Binaural simulation in an enclosed space is important in the subjective<br />

evaluation of the enclosure acoustics in the design or refinement stage. A<br />

time domain scheme using the geometrical acoustics technique has been<br />

usually used in the binaural processing. However, one can calculate a<br />

pressure impulse response by using the phased beam-tracing method,<br />

which incorporates the phase information in the beam tracing process.<br />

Such phased method employs reflection coefficient and wave number,<br />

whereas the conventional method uses absorption coefficient and air attenuation<br />

factor. Impulse response can be obtained by the inverse Fourier<br />

transformation of the frequency domain result. This feature facilitates the<br />

binaural simulation because the convolution with the HRTF can be accomplished<br />

by a simple multiplication in frequency domain. Convolutions<br />

were conducted for all reflections one by one, and the convolved transfer<br />

functions were summed into one transfer function. Consequently binaural<br />

<strong>room</strong> impulse responses at receivers’ ear positions can be simulated. The<br />

measured binaural <strong>room</strong> impulse responses in the conference <strong>room</strong> were<br />

compared with the predicted results for octave bands of 125 Hz to 4 kHz.<br />

A good agreement with measurement was found, especially in the early<br />

part of impulse responses. �Work supported by BK21.�<br />

4:00<br />

4pAA12. Visualization methods of direct and early reflection sounds<br />

in small enclosures. Chiaki Koga, Akira Omoto �Omoto Lab., Dept. of<br />

Acoust. Design, Faculty of Design, Kyushu Univ., Shiobaru 4-9-1,<br />

Minami, Fukuoka 815-8540, Japan�, Atsuro Ikeda, Masataka Nakahara<br />

�SONA Corp., Nakno-ku, Tokyo, 164-0013, Japan�, Natsu Tanaka, and<br />

Hiroshi Nakagawa �Nittobo Acoust. Eng. Co., Ltd. Sumida-ku, Tokyo<br />

130-0021, Japan�<br />

Many parameters exist for evaluating large sound fields such as concert<br />

halls. However, it is difficult to apply those parameters for evaluation<br />

of a small <strong>room</strong> such as a recording studio because of their different sound<br />

fields. Widely useful common parameters have not been established.<br />

Moreover, early reflections are important in small <strong>room</strong>s for determining<br />

spatial acoustic impressions. Therefore, various methods that visualize<br />

spatial acoustic information obtained by early reflection in <strong>room</strong>s are proposed.<br />

For this study, sound fields �a music studio and a filmmaking studio�<br />

were measured using three kinds of different techniques: instantaneous<br />

intensity, mean intensity, and a sphere-baffled microphone array.<br />

This report compares the information of sound source directions obtained<br />

using these methods. Results show that every method can estimate the<br />

position of sound sources and important reflections with high accuracy. In<br />

the future, we shall propose a method that visualizes spatial acoustic information<br />

more precisely by combining the methods and establishing<br />

acoustic parameters that are available for evaluating and designing small<br />

<strong>room</strong>s.<br />

4:15<br />

4pAA13. Acoustic evaluation of worship spaces in the city of Curitiba,<br />

Brazil. Cristiane Pulsides, David Q. de Sant’Ana, Samuel Ansay<br />

�LAAICA/UFPR, Bloco 4 sala PG-05 81531-990 Curitiba, PR, Brasil,<br />

pulsides@gmail.com�, Paulo Henrique T. Zannin, and Suzana Damico<br />

�LAAICA/UFPR, 81531-990 Curitiba, PR, Brasil�<br />

This article searches acoustic parameters in religious buildings located<br />

in the city of Curitiba intending to study its behavior in this kind of<br />

facilities. The temples were analyzed according to type of ceremony, architectonic<br />

style, and construction date. The research was made through<br />

the impulsive response integration method for three energetic parameters:<br />

�1� reverberation time �RT�; �2� clarity �C80�; and �3� definition �D50�<br />

according recommendations of the ISO/3382:1997 Standard. Performed in<br />

between were six and eight impulsive responses in each <strong>room</strong> using sweep<br />

signals and omnidirectional microphones. The results were than compared<br />

with referential values already existing �W. Fasold and E. Veres, Schallschutz<br />

� Raumakustik in der Praxis, 136 �1998�� for acoustic characterizations.<br />

It is possible to observe in the measurements the direct connection<br />

between reverberation time and the parameters clarity or definition. Moreover,<br />

it is possible also to observe the influence of the geometric ratios and<br />

architectural elements of the <strong>room</strong>s, getting itself for equivalent volumes<br />

and rays of removal of the source, different levels of definition.<br />

4:30<br />

4pAA14. A consideration of the measurement time interval for<br />

obtaining a reliable equivalent level of noise from expressway.<br />

Mitsunobu Maruyama �Salesian Polytechnic, Oyamagaoka 4-6-8,<br />

Machida, Tokyo 194-0215, Japan� and Toshio Sone �Akita Prepectural<br />

Univ., Honjo, Akita 015-0055, Japan�<br />

The level of road traffic noise LA eq,T greatly depends on the maximum<br />

level during the measurement time interval tau, and the maximum level<br />

often appears at the moment when two consecutive heavy vehicles pass<br />

through the point adjacent to the observation point. A mathematical model<br />

is proposed for simulating the variation in traffic noise, especially from the<br />

point of heavy vehicles with passing. The mean time interval between a<br />

pair of two consecutive heavy vehicles with the minimum allowable distance<br />

is obtained from time-series data and the mean recurrence time h ij<br />

which can be calculated from the transition matrix P�p ij�. The comparative<br />

study is made among the numbers of heavy vehicles from 25 to 300<br />

�vehicles/hour� in traffic flow and the observation distances of 40 to 200 m<br />

from the road. The result shows that the measurement time interval required<br />

for the acquisition of reliable data is three to four times as long as<br />

tau or h ij .<br />

3264 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3264


FRIDAY AFTERNOON, 1 DECEMBER <strong>2006</strong> KOHALA/KONA ROOM, 1:15 TO 4:30 P.M.<br />

Session 4pABa<br />

Animal Bioacoustics: Marine Mammal Acoustics II<br />

David K. Mellinger, Chair<br />

Oregon State Univ., Hatfield Marine Science Ctr., Newport, OR 97365<br />

1:15<br />

4pABa1. Great ears: Functional comparisons of land and marine<br />

leviathan ears. D. R. Ketten �Harvard Med. School, Boston, MA;<br />

Woods Hole Oceanograph. Inst., Woods Hole, MA�, J. Arruda, S. Cramer,<br />

M. Yamato �Woods Hole Oceanograph. Inst., Woods Hole, MA�, J.<br />

O’Malley �Massachusetts Eye and Ear Infirmary, Boston, MA�, D.<br />

Manoussaki �Vanderbilt Univ., Nashville, TN�, E. K. Dimitriadis �NIH/<br />

NIDCD, Bethesda, MD�, J. Shoshani �Univ. of Asmara, Asmara, Eritrea�,<br />

and J. Meng �American Museum of Natural History, New York, NY�<br />

Elephants and baleen whales are massive creatures that respond to<br />

exceptionally low frequency signals. Although we have many elephant and<br />

whale vocalization recordings, little is known about their hearing. Playback<br />

experiments suggest hearing in both proboscideans and mysticetes is<br />

tuned similarly to low or even infrasonic signals. This raises several interesting<br />

issues. First, they emit and perceive signals in two media, air and<br />

water, with radically different physical acoustic properties: 4.5-fold differences<br />

in sound speed, three-fold magnitude difference in acoustic impedance,<br />

and, for common percepts, whales must accommodate 60-fold<br />

acoustic pressures. Also, a commonly held tenet is that upper hearing limit<br />

is inversely correlated with body mass, implying there should be virtually<br />

no whale-elephant hearing overlap given body mass differences. This<br />

study analyzed how inner ears in these groups are structured and specialized<br />

for low-frequency hearing. Computerized tomography and celloidin<br />

histology sections were analyzed in six baleen whale (n�15) and two<br />

elephant species (n�7). The data show mysticetes have a substantially<br />

greater hearing range than elephants but that coiling and apical cochlear<br />

structures are similar, suggesting common mechanical underpinnings for<br />

LF hearing, including cochlear radii consistent with the Whispering Gallery<br />

propagation effect. �Work supported by ONR, NIH, WHOI OLI,<br />

Seaver Foundation.�<br />

1:30<br />

4pABa2. Social context of the behavior and vocalizations of the gray<br />

whale Eschrichtius robustus. Sarah M. Rohrkasse �School for Field<br />

Studies, Ctr. for Coastal Studies, Apartado Postal 15, Puerto San Carlos,<br />

BCS, CP 23740 Mexico, sarro101@hotmail.com� and Margaret M.<br />

Meserve �Guilford College, Greensboro, NC 27410�<br />

Sound production and surface behavior of the gray whale were investigated<br />

at Bahia Magdalena, Mexico to determine if vocalizations have<br />

behavioral correlations or are used in specific social contexts. Fifteenminute<br />

sessions of behavioral observations and acoustic recordings of gray<br />

whales in various social contexts were collected from February to April<br />

<strong>2006</strong> (n�30). Analysis of sound production included proportional use of<br />

different call types and acoustic variables of each sound type. Preliminary<br />

acoustic analysis found no correlation with social contexts or behaviors,<br />

but proportional use of different vocalizations is similar to past studies in<br />

Baja �Dahlheim et al, The Gray Whale, pp. 511–541 �1984�, F. J. Ollervides,<br />

dissertation, Texas A&M University �2001��. Initial results indicate<br />

significant differences in frequencies of high surface behaviors (p<br />

�0.0477) of groups that include mother-calf pairs. As analysis continues,<br />

possible correlations between social context and use of sounds could allow<br />

for acoustics to be an indicator of group composition, seasonal movements,<br />

and social patterns and to help determine the functions of sounds.<br />

�Work supported by SFS and NFWF.�<br />

Contributed Papers<br />

1:45<br />

4pABa3. Ambient noise and gray whale Eschrichtius robustus<br />

behavior. Francisco Ollervides, Kristin Kuester, Hannah Plekon, Sarah<br />

Rohrkasse �School for Field Studies—Ctr. for Coastal Studies, Apartado<br />

Postal 15, Puerto San Carlos, BCS, CP 23740 Mexico,<br />

follervides@hotmail.com�,Kristin Kuester �Univ. of<br />

Wisconsin�Madison, Madison, WI 53706�, HannahPlekon �Davidson<br />

College, Davidson, NC�, andSarahRohrkasse �Texas A and M Univ.,<br />

College Station, TX 77843�<br />

Between 14 February and 13, April <strong>2006</strong>, we conducted 31 recording<br />

sessions of ambient noise and behavioral sampling of gray whales within<br />

Magdalena Bay, Mexico. This breeding lagoon does not have the same<br />

Marine Protected Area status compared to the other breeding lagoons of<br />

San Ignacio and Guerrero Negro in the Mexican Pacific coast. Poorly<br />

monitored guidelines and increasing boat traffic from whale�watching<br />

tourism in this area have the potential to affect the surface behavior of<br />

these animals and increase average ambient noise levels. Relative ambient<br />

noise<br />

levelswererecordedandcomparedtoapreviousstudy�Ollervides,2001�todetermine<br />

similarities or differences in the 5�year interval between both data sets.<br />

Although results are not comparable in decibel levels, probably due to<br />

equipment calibration problems, there was a significant difference between<br />

the different regions of the bay Kruskal�Wallis �p�0.0067�. Activity<br />

levels ranged from 0.005–0.196 behaviors/whale/minute. Ambient noise<br />

levels ranged from 35.70–64.32 dB Re: 1 Pa. No correlation was found<br />

between the ambient noise levels in the bay and the activity level of gray<br />

whales �correlation value�0.0126; log correlation value�0.172�. Further<br />

acoustic processing is currently underway.<br />

2:00<br />

4pABa4. Look who’s talking; social communication in migrating<br />

humpback whales. Rebecca A. Dunlop, Michael J. Noad �School of<br />

Veterinary Sci., Univ. of Queensland, St. Lucia, Qld 4072, Australia.<br />

r.dunlop@uq.edu.au�, Douglas H. Cato �Defence Sci. and Tech Org.,<br />

Pyrmont, NSW 2009, Australia�, and Dale Stokes �Scripps Inst. of<br />

Oceanogr., La Jolla, CA 92037�<br />

A neglected area of humpback acoustics concerns nonsong vocalizations<br />

and surface behaviors known collectively as social sounds. This<br />

study describes a portion of the nonsong vocal repertoire and explores the<br />

social relevance of individual sound types. A total of 622 different sounds<br />

were catalogued and measured from whales migrating along the east coast<br />

of Australia. Aural and spectral categorization found 35 different sound<br />

types, and discriminate functions supported 33 of these. Vocalizations<br />

were analyzed from 60 pods that were tracked visually from land and<br />

acoustically using a static hydrophone array. Nonsong vocalizations occurred<br />

in all pod compositions: lone whales, adult pairs, mother/calf pairs,<br />

mother/calf/escorts, and multiple-adult pods. Thwops and wops were<br />

likely to be sex-differentiated calls with wops from females and thwops<br />

from males. Sounds similar to song-units were almost all from joining<br />

pods and yaps were only heard in splitting pods. Other low-frequency calls<br />

�less than 60 Hz� were thought to be within-pod contact calls. Higherfrequency<br />

cries �fundamental 450–700 Hz� and other calls �above 700 Hz�<br />

and presumed underwater blows were heard more frequently in joining<br />

3265 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3265<br />

4p FRI. PM


pods displaying agonistic behaviors. This work demonstrates that humpbacks<br />

produce a great range of contextually different communication signals.<br />

�Work supported by ONR and DSTO.�<br />

2:15<br />

4pABa5. Seasonal ambient noise levels and impacts on<br />

communication in the North Atlantic right whale. Susan E. Parks,<br />

Christopher W. Clark, Kathryn A. Cortopassi, and Dimitri Ponirakis<br />

�Bioacoustics Res. Program, Cornell Univ., 159 Sapsucker Woods Rd.,<br />

Ithaca, NY 14850, sep6@cornell.edu�<br />

The North Atlantic right whale is a highly endangered species of baleen<br />

whale. Acoustic communication plays an important role in the social<br />

behavior of these whales. Right whales are found in coastal waters along<br />

the east coast of the United States, an area characterized by high levels of<br />

human activity. Most of these activities generate noise that is propagated<br />

into the coastal marine environment. The goals of this project are to characterize<br />

the noise, both natural and anthropogenic, in right whale habitat<br />

areas to determine what levels of noise the whales are regularly exposed<br />

to, and whether the acoustic behavior of right whales changes in response<br />

to increased noise. Continuous recordings were made from autonomous<br />

bottom-mounted recorders in three major habitat areas in 2004 and 2005;<br />

Cape Cod Bay �December–May�, Great South Channel �May�, and the<br />

Bay of Fundy, Canada �August� to passively detect right whales by recording<br />

their vocalizations. Here, we describe the ambient noise levels in these<br />

recordings to describe the daily acoustic environment of right whales, how<br />

noise varied over diel, weekly, and seasonal time scales, and whether noise<br />

levels correlated with any observed changes in acoustic behavior of the<br />

whales.<br />

2:30<br />

4pABa6. Blue whale calling in Australian waters. Robert D.<br />

McCauley, Chandra P. Salgado Kent �Curtin Univ. of Technol., G.P.O.<br />

Box U 1987, Perth 6845, Australia�, Christopher L.K. Burton �Western<br />

Whale Res. Hillarys 6923, WA Australia�, and Curt Jenner �Ctr. for Whale<br />

Res. �WA Inc.�, Fremantle WA, 6959 Australia�<br />

Calling from the Antarctic true blue whale �Balaenoptera musculus<br />

intermedia� and the tropical subspecies �brevicauda, or pygmy blue� have<br />

been recorded across southern Australia with the pygmy blue calls also<br />

recorded along the Western Australian �WA� coast. The subspecies have a<br />

believed common downsweep and markedly different longer, tonal calls.<br />

The frequency of most energy in the tonal calls is offset between the<br />

subspecies suggesting sound-space partitioning. The pygmy blue threepart<br />

tonal call is typically 120 s long repeated every 200 s, has several<br />

variants, and includes a complex two-source component. The nature of the<br />

pygmy blue call allows counts of instantaneous calling individuals, giving<br />

relative abundance. These estimates in the Perth Canyon, a localized seasonal<br />

feeding area, show patterns in usage of space and through time<br />

within and between seasons, such as the sudden departure of animals at a<br />

season end, which varies by approximately 2 weeks between years. Sea<br />

noise records along the WA coast indicate south-traveling animals arrive<br />

midway along the coast in October to November, animals fan out across<br />

southern Australian over December through May, then move north in the<br />

Austral winter. We have begun converting abundance estimates from relative<br />

to absolute for pygmy blue calling rates.<br />

2:45<br />

4pABa7. Acoustical monitoring of finback whale movements on the<br />

New Jersey Shelf. Altan Turgut �Naval Res. Lab., Acoust. Div.,<br />

Washington, DC 20375� and Christopher Lefler �Univ. of California Santa<br />

Barbara, Santa Barbara, CA 93106�<br />

Acoustical monitoring of finback whales is performed by using a data<br />

set collected over a 3-week period in December of 2003 on the New<br />

Jersey Shelf. One-second-duration 20-Hz signals of finback whales were<br />

recorded on three vertical line arrays �VLAs� and a bottomed horizontal<br />

line array �HLA.�. One-second-duration pulses are separated by about 10 s<br />

and there is an approximately 2-min-long silent period between 10- to<br />

18-min-long pulse trains. A 30- to 60-min silent period after 5 to 10 pulse<br />

trains is also common. Modal analysis of individual pulses indicated that<br />

most signals contained two acoustic modes. Arrival-time and group-speed<br />

differences of these modes are used for remote acoustic ranging. These<br />

modal characteristics are also exploited in a broadband matched-field algorithm<br />

for depth discrimination. Bearing estimation of individual whales<br />

is obtained by performing horizontal beamforming on the HLA data.<br />

Range estimation results are verified by time-of-flight triangulation using<br />

single hydrophone data from each VLA location. Acoustic monitoring results<br />

indicated that most finback whales traveled near the shelf break front<br />

where food might be abundant. Relations between silent periods and<br />

acoustic range/depth monitoring results are also investigated. �This work<br />

was supported by the ONR.�<br />

3:00–3:15 Break<br />

3:15<br />

4pABa8. Analysis of melon-headed whale aggregation in Hanalei Bay,<br />

July 2004. David M. Fromm �Naval Res. Lab., 4555 Overlook Ave. SW,<br />

Washington, DC 20375-5350�, Joseph R. Mobley, Jr. �Univ. of Hawaii at<br />

M_noa, Honolulu, HI 96822�, Stephen W. Martin �Space and Naval<br />

Warfare Systems Ctr. San Diego, San Diego, CA 92152-5001�, and Paul<br />

E. Nachtigall �Univ. of Hawaii at M_noa, Kailua, HI 96734�<br />

On 3 July 2004, an aggregation of ca. 150–200 melon-headed whales<br />

�Peponocephala electra� appeared in the shallow waters of Hanalei Bay,<br />

Kauai and congregated there for over 27 h. Preceding the whales’ appearance<br />

and partially coincident with their time in the Bay, midrange �3.5–5<br />

kHz� tactical sonars were intermittently deployed during the Rim of the<br />

Pacific 2004 �RIMPAC� joint military exercises being conducted in waters<br />

near Kauai by the U.S., Japan, and Australia Navies. An NOAA report<br />

�Southall et al., <strong>2006</strong>� attributed the active sonar usage as a plausible, if<br />

not likely, contributing factor. A detailed timeline and reconstruction of the<br />

RIMPAC activities is presented showing the worst-case estimates of the<br />

sonar sound levels in the waters surrounding Kauai. A re-examination of<br />

available evidence combined with a new report of a simultaneous and<br />

similar aggregation in Sasanhaya Bay, Rota, Commonwealth of the Northern<br />

Mariana Islands, brings the plausibility conclusion into question. �This<br />

work was sponsored by multiple sources. D. Fromm and S. Martin conducted<br />

acoustic analyses with funds provided by the U.S. Pacific Fleet. J.<br />

Mobley received funding from the U.S. Geological Survey. P. Nachtigall is<br />

sponsored by the Office of Naval Research for marine mammal audiometric<br />

studies.�<br />

3:30<br />

4pABa9. Midfrequency sound propagation in beaked whale<br />

environments. Eryn M. Wezensky, Thomas R. Stottlemyer, Glenn H.<br />

Mitchell �Naval Undersea Warfare Ctr., Newport Div., Newport, RI<br />

02841�, and Colin D. MacLeod �Univ. of Aberdeen, Aberdeen, U.K.�<br />

Recent mass strandings of beaked whales �Ziphiidae, Cetacea� coinciding<br />

with the use of midfrequency range �1–10 kHz� active sonar have<br />

caused speculation about the potentially adverse effects of these sound<br />

sources. Particular questions of the research and regulatory communities<br />

concern whether beaked whale sensitivity to midfrequency sound exposure<br />

is influenced by oceanographic characteristics present at the time of<br />

the mass stranding events. This study investigated the interaction between<br />

beaked whale habitat characteristics and the nature of a midfrequency<br />

signal by analyzing the oceanographic factors affecting underwater acoustic<br />

propagation. Three types of model sites were selected from five specific<br />

geographical locations where beaked whales have been regularly recorded<br />

or where a mass stranding event has been reported. A ray-trace acoustic<br />

propagation model was used to generate transmission loss for a 3-kHz<br />

signal over a representative 60-km transect at each locality. Model outputs<br />

visually demonstrated how the combination of site/event-specific oceanographic<br />

characteristics affects the sound propagation of a moving source.<br />

A parametric sensitivity comparison and statistical analysis were conducted<br />

to identify influential factors between environmental parameters,<br />

source depth, and the resulting transmission loss. Major findings of this<br />

study as well as future research direction are discussed. �Research supported<br />

by NAVSEA.�<br />

3266 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3266


3:45<br />

4pABa10. Examination and evaluation of the effects of fast rise-time<br />

signals on aquatic animals. Michael Stocker �Seaflow, Inc., 1062 Ft.<br />

Cronkhite, Sausalito, CA 94965�<br />

Increasingly human enterprise is subjecting the ocean environment to<br />

acoustic signals to which marine animals are not biologically adapted.<br />

This is evidenced by a marked rise in marine mammal strandings, as well<br />

as hearing and other physiological damage to fish and other marine organisms<br />

as a result of, or coincident to, human-generated noise events. Determining<br />

phonotoxic thresholds of marine organisms is complicated by the<br />

fact that various marine animals are adapted to sense either pressure gradient<br />

or particle motion acoustic energy, or some combination or gradient<br />

between the two. This has been addressed to some degree by exposure<br />

metrics that consider either net or accumulated acoustical flux densities<br />

from various noise sources. This paper examines the role and effects of<br />

signal rise time both in terms of physiological impulse response of the<br />

exposed organisms, as well as broadband saturation flux densities of fast<br />

rise-time signals on animal sense organs. Case studies from the literature<br />

will be presented to demonstrate the effects of fast rise time signals on<br />

fish. Acoustical signals with high crest factors and fast rise-time components<br />

will be compared to signals with dominantly sinusoidal components<br />

to illustrate the perceptual effects of these signals on human hearing.<br />

4:00<br />

4pABa11. Noise separation of underwater acoustic vocalization using<br />

auditory filter bank and Poisson rate estimation. Owen P. Kenny and<br />

Craig R. McPherson �Dept. of Elec. and Comput. Eng., James Cook<br />

Univ., Douglas 4811, Queensland, Australia�<br />

Formant vocalization tracking has been achieved using a mammalian<br />

periphery model and a Poisson rate estimator. This approach used a set of<br />

linear bandpass filters to simulate the mechanical displacement of the basilar<br />

membrane. The auditory model simulated neural firing by producing a<br />

spike at the positive going zero crossing for each filter output. Poisson<br />

intensity of the neural firing rate is controlled by the dominant frequency<br />

components of the signal present in the filter. This approach is extended by<br />

incorporating neural synchronization information to separate the formant<br />

structure from that of noise. The filter structure is designed to overlap the<br />

frequency range of adjacent filters. The presence of a formant structure in<br />

adjacent filters controls the interspike intervals of neural firing for both<br />

filters, which results in the neural firing from both filters being synchronized.<br />

If a noise-only component is present in either filter, then the spiking<br />

outputs from the adjacent filters are unsynchronized. Experimental results<br />

have shown that incorporating neural synchronization information between<br />

adjacent filters has enabled separation of signal components from<br />

noise. This technique enables easier signal and noise separation than allowed<br />

by traditional methods.<br />

4:15<br />

4pABa12. Using vocalizations of Antarctic seals to determine pupping<br />

habitats. T. L. Rogers, C. J. Hogg, M. B. Ciaglia �Australian Marine<br />

Mammal Res. Ctr., Zoological Parks Board of NSW/Faculty of Veterinary<br />

Sci., Univ. of Sydney, Mosman Australia�, and D. H. Cato �Defence Sci.<br />

& Technol. Organisation, Pyrmont, Australia�<br />

The Ross and Leopard seal use the floes of the Antarctic pack ice to<br />

whelp and raise their pups, But both species are rarely seen in summer<br />

throughout the pack ice. We now realize that this is because they are under<br />

the water ‘‘calling’’ during the austral summer as part of their breeding<br />

display, and so their presence is underestimated in traditional visual surveys.<br />

The period of ‘‘calling’’ overlaps with the time that females give<br />

birth, so their vocalizations can be used to determine seal distributions<br />

during this time. Acoustic recordings were made using sonobuoys deployed<br />

during ship based surveys in the pack ice and analyzed to determine<br />

the seal distributions. This was used to predict habitat preference of<br />

seals by relating their distributions to remotely sensed indices: ice cover,<br />

ice floe type, ice thickness, distance to ice edge, distance to shelf break,<br />

distance to land, sea surface temperature, and chlorophyll a.<br />

FRIDAY AFTERNOON, 1 DECEMBER <strong>2006</strong> KOHALA/KONA ROOM, 4:30 TO 5:15 P.M.<br />

Session 4pABb<br />

Animal Bioacoustics: Avian Acoustics<br />

Ann E. Bowles, Chair<br />

Hubbs Sea World Research Inst., 2595 Ingraham St., San Diego, CA 92109<br />

4:30<br />

4pABb1. Effective area of acoustic lure surveys for Mexican spotted<br />

owls „Strix occidentalis lucida…. Samuel L. Denes, Ann E. Bowles<br />

�Hubbs-SeaWorld Res. Inst., 2595 Ingraham St., San Diego, CA 92109,<br />

sdenes@hswri.org�, Kenneth Plotkin, Chris Hobbs �Wyle Labs.,<br />

Arlington, VA 22202�, John Kern �Kern Statistical Services, Sauk Rapids,<br />

MN 56379�, and Elizabeth Pruitt �GeoMarine, Inc., Hampton, VA 23666�<br />

During acoustic lure surveys for birds, topography and ambient noise<br />

are likely to be important determinants of detectability. Examinations of<br />

propagation were conducted for acoustic lures �human-made calls� and<br />

owl responses recorded during acoustic surveys for Mexican spotted owls<br />

in the Gila National Forest �2005�. Lure surveys were designed based on<br />

Contributed Papers<br />

formal agency protocols, which assumed a 0.43-km detection range under<br />

typical conditions. A total of 558 points was called over a heavily forested,<br />

topographically complex 20�24-km area. Real-time measurements of owl<br />

calls and lures were made with a calibrated recording system. Ambient<br />

noise was collected using an array of 39 Larson-Davis 820 and 824 soundlevel<br />

meters. The NMSIM �Wyle Laboratories� single-event propagation<br />

simulator was used to model propagation of both owl and human calls.<br />

The resulting model of survey effort was compared with a simple twodimensional<br />

statistical model. Probability of detecting owls did not fit the<br />

expectations of the agency protocol, suggesting that acoustic propagation<br />

should be considered during owl surveys. �Work supported by U.S. Air<br />

Force ACC/CEVP; USFWS Permit No. TE024429�<br />

3267 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3267<br />

4p FRI. PM


4:45<br />

4pABb2. Automated localization of antbirds and their interactions in<br />

a Mexican rainforest. Alexander N. G. Kirschel, Travis C. Collier,<br />

Kung Yao, and Charles E. Taylor �Univ. of California, Los Angeles, 621<br />

Charles E. Young Dr. South, Los Angeles, CA 90095�<br />

Tropical rainforests contain diverse avian communities incorporating<br />

species that compete vocally to propagate their signals to intended receivers.<br />

In order to effectively communicate with birds of the same species,<br />

birds need to organize their song performance temporally and spatially. An<br />

automated identification and localization system can provide information<br />

on the spatial and temporal arrangement of songs. Acoustic sensor arrays<br />

were tested for the ability to localize the source of songs of antbirds<br />

recorded in a Mexican rainforest. Pilot studies with a five-node array<br />

arranged in a rough circle with a 20-m diameter located the song of Dusky<br />

Antbird �Cercomacra tyrannina� with an error of 73 cm and Mexican<br />

Antthrush �Formicarius moniliger� with an error of 65 cm from the location<br />

of a source loudspeaker within the array. An additional source 21 m<br />

outside was also localized. Results will be presented for experiments and<br />

recordings of individuals at the Mexican rainforest site in October <strong>2006</strong>.<br />

Locations of birds of the same and different species during vocal performance<br />

will provide a greater understanding of how individuals interact<br />

spatially with each other based on their vocal performance, from which the<br />

role of song in ecological interactions can be inferred.<br />

5:00<br />

4pABb3. Nonintrusive acoustic identification of hermit thrush<br />

„Catharus guttatus… individuals. Dennis F. Jones �Defence R&D<br />

Canada—Atlantic, P.O. Box 1012, Dartmouth, NS, Canada B2Y 3Z7,<br />

dennis.jones@drdc-rddc.gc.ca�<br />

From mid-April well into the summer, the secretive hermit thrush �Catharus<br />

guttatus� can be heard singing throughout the woodlands of Nova<br />

Scotia. Its song is distinctive, beginning with a clear introductory note<br />

followed by a flurry of flutelike body notes, often cascading and reverberant<br />

in character. Despite this fine display of avian virtuosity, few studies<br />

have been reported that probe the differences between the calls, songs, and<br />

repertoires of individuals. From April 2003 to May <strong>2006</strong>, over 3000 songs<br />

from several birds were recorded using digital video cameras at study sites<br />

in and around the city of Halifax, Nova Scotia. The only birds recorded<br />

were those in close proximity to roads and trails. None of the birds were<br />

marked, banded, or deliberately disturbed in any way. Although the study<br />

birds remained hidden from view most of the time, in the few instances<br />

where the birds perched in the open, their behaviors while singing were<br />

captured on videotape. All of the birds were readily distinguishable from<br />

each other as no two individuals had a single song in common. The most<br />

significant finding was that individuals could be reidentified acoustically<br />

after 1 week, 3 months, and 1 year had elapsed.<br />

FRIDAY AFTERNOON, 1 DECEMBER <strong>2006</strong> KAHUKU ROOM, 1:00 TO 4:50 P.M.<br />

Session 4pBB<br />

Biomedical UltrasoundÕBioresponse to Vibration and Signal Processing in Acoustics: Elastic Imaging<br />

Peter J. Kaczkowski, Cochair<br />

Univ. of Washington, Applied Physics Lab., 1013 NE 40th Street, Seattle, WA 98105-6698<br />

Tsuyoshi Shiina, Cochair<br />

Univ. of Tsukuba, Graduate School of Systems and Information Engineering, 1-1-1 Tennodai, Tsukuba 305-8573, Japan<br />

Invited Papers<br />

1:00<br />

4pBB1. Present and future of elasticity imaging technology. Tsuyoshi Shiina �Grad. School of Systems and Information Eng.,<br />

Univ. of Tsukuba, 1-1-1 Tennodai Tsukuba, Japan� and Ei Ueno �Univ. of Tsukuba, Tsukuba, Japan�<br />

Elastic properties of tissues are expected to provide us novel diagnostic information since they are based on tissue characteristics<br />

and sensitively reflect its pathological state. So far, various techniques for tissue elasticity imaging have been proposed. However, it<br />

was not so easy to satisfy real-time operation and freehand manipulation of probe, which was required for practical equipment. To<br />

satisfy these conditions, we developed the combined autocorrelation method �CAM� and recently manufactured a commercial ultrasound<br />

scanner, for real-time tissue elasticity imaging by implementing the CAM algorithm, By slightly compressing or relaxing the<br />

body through freehand operation, the strain images are obtained with real-time and superimposed on B-mode images with a translucent<br />

color scale. In addition, we proposed elasticity scores of malignancy by categorizing patterns of elasticity images of breast tumors<br />

into five classes from malignant to benign. As a result of diagnosis based on the elasticity score, it was revealed that even nonexperts<br />

could attain precise diagnosis of breast cancer based on elasticity score as well as experts since the criterion on elasticity score is much<br />

simpler than conventional B-mode images. Finally, some prospects for the next stages of elasticity imaging technology well be<br />

surveyed.<br />

1:20<br />

4pBB2. Real-time tissue elasticity system—Development and clinical application. Takeshi Matsumura, Tsuyoshi Mitake<br />

�Hitachi Medical Corp. 2-1, Toyofuta, Hashiwa-Shi, Chiba-Ken, Japan�, Tsuyishi Tsuyishi, Makoto Yamakawa, Ei Ueno �Tsukuba<br />

Univ.�, Nobuhiro Fukunari �Shouwa Univ.�, and Kumi Tanaka �Nippon Medical Univ.�<br />

The progress of recent semiconductor technology has a remarkable thing. Thanks to progress of this semiconductor technology, the<br />

ultrasound scanner in medicine could come to hold enormousness computing power and has come to realize various complicated<br />

processing. At the same time, hardness of human tissue which, as you know, is used by palpation is already the information that is<br />

important in a diagnosis. But, we think that it does not have enough objectivity. To increase objectivity by visualizing hardness of<br />

3268 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3268


tissue, we adopted ECAM �extended combined autocorrelation method�, which was developed by Professor Shiina at Tsukuba<br />

University in Japan, and succeeded in developing the commercial ultrasound scanner, which could display a strain image in real time.<br />

From a clinical point of view, in the breast region, mammography examination is effective in a diagnosis, but a judgment of<br />

permeation degree is not superior in ultrasound image. And in a thyroid gland region, we begin to get experience with availability<br />

from a diagnosis of papillary cancer and follicular cancer. So, we would like to have the presentation about the development of a strain<br />

imaging function and some of our clinical experiences by using the developed system.<br />

1:40<br />

4pBB3. Elasticity of perfused tissue. Kirk W. Beach �Dept. of Surgery, Univ. of Washington, Box 356410, Seattle, WA<br />

98195-6410�, Barbrina Dunmire, and John C. Kucewicz �Univ. of Washington, Seattle, WA 98105-6698�<br />

Elastic imaging intends to measure Young’s modulus �tissue stiffness� or bulk modulus �tissue compressibility� of tissue subjected<br />

to an applied strain of several percent. Underlying elastic imaging is the assumption of a linear stress/strain relationship without<br />

hysteresis or other time-dependent behavior. Perfused tissue is a composite material comprised of a solid matrix of cells, fibers,<br />

interstitial fluid �occupying up to 50% of the tissue volume and varying slowly with time�, arterioles �pulsating high-pressure spaces<br />

that occupy 0.1% of the tissue volume�, capillaries, and venules �low-pressure spaces that occupy up to 3% of the tissue volume<br />

varying with respiration�. This talk will speculate on the nonlinear, nonstationary stress/strain relationships expected from dependent<br />

tissues �legs�, pressurized tissues �breast tumors�, and other living, perfused tissues. The pressure versus strain curve from each tissue<br />

voxel allows the measurement of arteriolar and venular volumes and pressures, and interstitial pressure within the tissues. These<br />

volumes and pressures may be key to classifying pathologies.<br />

2:00<br />

4pBB4. New developments in transient elastography. Mathias Fink, Mickael Tanter, Ralph Sinkus, and Gabriel Montaldo �LOA,<br />

ESPCI, 10 rue Vauquelin, 75005, Paris, France�<br />

An ultra-high-rate ultrasonic scanner has been developed that can give 5000 ultrasonic images per second of the body. With such<br />

a high frame rate, the propagation of transient shear waves can be followed, and from the spatio-temporal evolution of the displacement<br />

fields, various inversion algorithms allow us to recover the shear modulus map. A discussion on the various inversion algorithms<br />

will be presented. In order to obtain unbiased shear elasticity map, different configurations of shear sources induced by radiation<br />

pressure of focused transducer arrays are used. Both 2-D and 3-D imaging can obtained with this technique. In vitro and in vivo results<br />

on breast will be presented that demonstrate the interest of elasticity imaging with transient elastography.<br />

2:20<br />

4pBB5. Spectral characteristics of breast vibro-acoustography<br />

images. Azra Alizad, Dana H. Whaley, Mathew Urban, Randall R.<br />

Kinnick, James F. Greenleaf, and Mostafa Fatemi �Mayo Clinic College<br />

of Medicine, Rochester, MN 55905 aza@mayo.edu�<br />

Vibro-acoustography image is a function of the dynamic characteristics<br />

of the object at the vibration �difference� frequency �df�. The dynamic<br />

characteristic of tissue is closely related to pathology. Therefore, it is<br />

important to evaluate image features versus df. Here, the influence of df on<br />

breast vibro-acoustography images is studied by scanning human breast at<br />

various df values ranging from 20 to 90 kHz. The subjects were chosen<br />

from a group of volunteers with different breast abnormalities. Images<br />

were compared subjectively to study image features and the appearances<br />

of breast lesions versus df. It is demonstrated that having a collection of<br />

images of the same tissue at different df values generally provides a better<br />

perception of the tissue structure and improves lesion identification. In<br />

most cases, higher df resulted in a higher signal-to-noise ratio and thus a<br />

higher image quality. Finally, a frequency-compounded images was obtained<br />

by calculating the weighted sum of images at different df values. It<br />

is demonstrated that image compounding normally improves visualization<br />

of breast tissue and abnormalities. �Work supported by NIH Grant EB-<br />

00535 and Grant BCTR0504550 from the Susan G. Komen Breast Cancer<br />

Foundation. Disclosure: Parts of the techniques used here are patented by<br />

MF and JFG.�<br />

Contributed Papers<br />

2:35<br />

4pBB6. Tissue pulsatility imaging: Ultrasonic measurement of strain<br />

due to perfusion. John C. Kucewicz, Barbrina Dunmire, Lingyun<br />

Huang, Marla Paun �Univ. of Washington Appl. Phys. Lab., 1013 NE<br />

40th St., Seattle, WA 98105-6698�, and Kirk W. Beach �Univ. of<br />

Washington, Seattle, WA 98195-6410�<br />

Over each cardiac cycle perfused tissues expand and relax by a fraction<br />

of a percent as blood rapidly accumulates in the arterial vasculature<br />

during systole and then slowly drains through the venous vasculature during<br />

diastole. Tissue pulsatility imaging �TPI� is a variation on ultrasonic<br />

tissue strain imaging that estimates tissue perfusion from this natural, cyclic<br />

tissue expansion and relaxation. TPI is derived in principle from plethysmography,<br />

a century-old technology for measuring gross tissue volume<br />

change from a whole limb or other isolatable body part. With TPI, the<br />

plethysmographic signal is measured from hundreds or thousands of<br />

sample volumes within an ultrasound image plane to characterize the local<br />

perfusion throughout a body part. TPI measures tissue strain over the<br />

cardiac cycle and parametrizes the signal in terms of its amplitude and<br />

shape. The amplitude of the strain waveform is correlated with perfusion,<br />

and the shape of the waveform is correlated with vascular resistance.<br />

Results will be presented from the leg showing the change in the TPI<br />

signals as the muscles recover from exercise, from breast tumors, and<br />

from the brain as blood flow changes in response to visual stimulation.<br />

�Work supported in part by NIH 1-R01EB002198-01 and NIH N01-CO-<br />

07118.�<br />

3269 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3269<br />

4p FRI. PM


2:50<br />

4pBB7. Using human body shear wave noise for passive elastography.<br />

Karim G. Sabra, Stephane Conti, Philippe Roux, and William A.<br />

Kuperman �Scripps Inst. of Ocean., Univ. of California—San Diego,<br />

9500 Gilman Dr., San Diego, CA 92093-0238�<br />

An elastography imaging technique based on passive measurement of<br />

shear wave ambient noise generated in the human body �e.g., due to the<br />

heart, muscles twitches, and blood flow system� has been developed. This<br />

technique merges two recent research developments in medical imaging<br />

and physics: �1� recent work on the efficacy of elastographic imaging<br />

demonstrating that shear waves are excellent candidates to image tissue<br />

elasticity in the human body and �2� theory and experimental verification<br />

in ultrasonics, underwater acoustics, and seismology of the concept of<br />

extracting coherent Green’s function from random noise cross correlations.<br />

These results provide a means for coherent passive imaging using<br />

only the human body noise field, without the use of external active<br />

sources. Coherent arrivals of the cross correlations of recordings of human<br />

body noise in the frequency band 2–50 Hz using skin-mounted accelerometers<br />

allows us to estimate the local shear velocity of the tissues. The<br />

coherent arrivals emerge from a correlation process that accumulates contributions<br />

over time from noise sources whose propagation paths pass<br />

through both sensors. The application of this passive elastography technique<br />

for constructing biomechanical models of in vivo muscles’ properties<br />

will be discussed.<br />

3:05–3:20 Break<br />

3:20<br />

4pBB8. Dynamic radiation force of acoustic waves on solid elastic<br />

spheres. Glauber T. Silva �Instituto de Computação, Universidade<br />

Federal de Algoas, Maceió, AL, 57072-970, Brazil�<br />

The present study concerns the dynamic radiation force on solid elastic<br />

spheres exerted by a plane wave with two frequencies �bichromatic wave�<br />

considering the nonlinearity of the fluid. Our approach is based on solving<br />

the wave scattering for the sphere in the quasilinear approximation within<br />

the preshock wave range. The dynamic radiation force is then obtained by<br />

integrating the component of the momentum flux tensor at the difference<br />

of the primary frequencies over the boundary of the sphere. Effects of the<br />

fluid nonlinearity play a major role in dynamic radiation force, leading it<br />

to a regime of parametric amplification. The developed theory is used to<br />

calculate the dynamic radiation force on three different solid spheres �aluminum,<br />

silver, and tungsten�. The obtained spectrum of dynamic radiation<br />

force presents resonances with larger amplitude and better shape than<br />

those exhibited in static radiation force. Applications of the results to some<br />

elasticity imaging techniques based on dynamic radiation force will be<br />

presented.<br />

3:35<br />

4pBB9. Ultrasonic measurement of displacement distribution inside<br />

an object caused by dual acoustic radiation force for evaluation of<br />

muscular relax property due to acupuncture therapy. Yoshitaka<br />

Odagiri, Hideyuki Hasegawa, and Hiroshi Kanai �Grad. School of Eng.,<br />

Tohoku Univ., Sendai 980-8579, Japan, odagiri@us.ecei.tohoku.ac.jp�<br />

Many studies have been carried out on the measurement of mechanical<br />

properties of tissues by applying an ultrasound-induced acoustic radiation<br />

force. To assess mechanical properties, strain of an object must be generated.<br />

However, one radiation force is not sufficient because it also causes<br />

translational motion when the object is much harder than surrounding<br />

medium. In this study, two cyclic radiation forces are applied to a muscle<br />

phantom from two opposite horizontal directions so that the object is<br />

cyclically compressed in the horizontal direction. As a result, the object is<br />

vertically expanded due to the incompressibility. The resultant vertical<br />

displacement is measured using ultrasound. Two concave ultrasonic transducers<br />

for actuation were both driven by sums of two continuous sinusoidal<br />

signals at two slightly different frequencies of 1 MHz and (1M<br />

�5) Hz. Displacement, which fluctuates at 5 Hz, was measured by the<br />

ultrasonic phased tracking method proposed by our group. Results indicated<br />

that the surface of the phantom was cyclically actuated with an<br />

amplitude of a tenth of a few micrometers, which well coincided with that<br />

measured with laser vibrometer. In addition, upward and downward displacements<br />

at the surface and deeper region were found during the increase<br />

phase of radiation forces. Such displacements correspond to the<br />

horizontal compression.<br />

3:50<br />

4pBB10. A phantom study on ultrasonic measurement of arterial wall<br />

strain combined with tracking of translational motion. Hideyuki<br />

Hasegawa and Hiroshi Kanai �Grad. School of Eng., Tohoku Univ.,<br />

Aramaki-aza-Aoba 6-6-05, Sendai 980-8579, Japan,<br />

hasegawa@us.ecei.tohoku.ac.jp�<br />

Correlation-based techniques are often applied to ultrasonic rf echoes<br />

to obtain the arterial wall deformation �strain�. In such methods, the displacement<br />

estimates are biased due to changes in center frequency of<br />

echoes. One of the reasons for the change in center frequency is the<br />

interference of echoes from scatterers within the wall. In the phased tracking<br />

method previously proposed for strain estimation by our group, the<br />

estimated displacement contains both the components due to the translational<br />

motion and strain. The translational motion is larger than strain by a<br />

factor of 10 and, thus, the error in the estimated displacement due to the<br />

change in center frequency mainly depends on translational motion and is<br />

often larger than the minute displacement due to strain. To reduce this<br />

error, in this study, a method is proposed in which the translational motion<br />

is compensated using the displacement of the luminal boundary estimated<br />

by the phased tracking method before correlating echoes between the<br />

frame before deformation and that at the maximum deformation to estimate<br />

the strain distribution within the wall. In basic experiments using<br />

phantoms made of silicone rubber, the estimation error was much reduced<br />

to 15.6% in comparison with 36.4% obtained by the previous method.<br />

4:05<br />

4pBB11. Wave biomechanics of skeletal muscle. Oleg Rudenko �Dept.<br />

of. Blekinge Inst. of Technol., 371 79 Karlskrona, Sweden� and Armen<br />

Sarvazyan �Artann Labs., Inc., West Trenton, NJ 08618�<br />

Physiological functions of skeletal muscle, such as voluntary contraction<br />

and force development, are accompanied by dramatic changes of its<br />

mechanical and acoustical properties. Experimental data show that during<br />

contraction, the muscle’s Young’s modulus, shear viscosity, and anisotropy<br />

parameter are changed by over an order of magnitude. None of the existing<br />

models of muscle contraction and muscle biomechanics can adequately<br />

explain the phenomena observed. A new mathematical model �O.<br />

Rudenko and A. Sarvazyan, Acoust. Phys. �6�, �<strong>2006</strong>��, has been developed<br />

relating the shear wave propagation parameters to molecular structure<br />

of the muscle and to the kinetics of the mechanochemical crossbridges<br />

between the actin and myosin filaments. New analytical solutions<br />

describing waves in muscle including nonlinear phenomena are found. A<br />

molecular mechanism for the dependence of acoustical characteristics of<br />

muscle on its fiber orientation and the contractile state is proposed. It is<br />

shown that although the anisotropy connected with the preferential direction<br />

along the muscle fibers is characterized by five elastic moduli, only<br />

two of these moduli have independent values in the muscle. The potential<br />

implications of the proposed model in terms of the acoustical assessment<br />

of muscle function are explored.<br />

3270 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3270


4:20<br />

4pBB12. Phase aberration correction for a linear array transducer<br />

using ultrasound radiation force and vibrometry optimization:<br />

Simulation study. Matthew W. Urban and James F. Greenleaf �Dept. of<br />

Physiol. and Biomed. Eng., Mayo Clinic College of Medicine, 200 First St.<br />

SW, Rochester, MN 55905, urban.matthew@mayo.edu�<br />

Diagnostic ultrasound images suffer from degradation due to tissues<br />

with sound speed inhomogeneities causing phase shifts of propagating<br />

waves. These phase shifts defocus the ultrasound beam, reducing spatial<br />

resolution and image contrast in the resulting image. We describe a phase<br />

aberration correction method that uses dynamic ultrasound radiation force<br />

to harmonically excite a medium using amplitude-modulated continuous<br />

wave ultrasound created by summing two ultrasound frequencies at f 0<br />

�3.0 MHz and f 0�� f �3.0005 MHz. The phase of each element of a<br />

linear array transducer is sequentially adjusted to maximize the radiation<br />

force and obtain optimal focus of the ultrasound beam. The optimization is<br />

performed by monitoring the harmonic amplitude of the scatterer velocity<br />

in the desired focal region using Doppler techniques. Simulation results<br />

show the ability to regain a 3.0-MHz focused field after applying a phase<br />

screen with an rms time delay of 95.4 ns. The radiation force magnitude<br />

increased by 22 dB and the resolution of the field was regained. Simulation<br />

results show that the focus of the beam can be qualitatively and<br />

quantitatively improved with this method. �This study was supported in<br />

part by Grants EB002640 and EB002167 from the NIH.�<br />

4:35<br />

4pBB13. Application of the optoacoustic technique to visualization of<br />

lesions induced by high-intensity focused ultrasound. Tatiana<br />

Khokhlova, Ivan Pelivanov, Vladimir Solomatin, Alexander Karabutov<br />

�Intl. Laser Ctr., Moscow State Univ., 119992, Moscow, Russia<br />

t_khokhlova@ilc.edu.ru�, and Oleg Sapozhnikov �Moscow State Univ.,<br />

119992, Moscow, Russia�<br />

Today several techniques are being applied to monitoring of highintensity<br />

focused ultrasound �HIFU� therapy, including MRI, conventional<br />

ultrasound, and elastography. In this work a new method for noninvasive<br />

monitoring of HIFU therapy is proposed: the optoacoustic method. The<br />

optoacoustic technique is based on the excitation of wideband ultrasonic<br />

pulses through the absorption of pulsed laser radiation in tissue and subsequent<br />

expansion of the heated volume. The excited optoacoustic �OA�<br />

pulse contains information on the distribution of optical properties within<br />

the tissue—light scattering and absorption coefficients. Therefore, if thermal<br />

lesions have different optical properties than the untreated tissue, they<br />

will be detectable on the OA waveform. The considerable change in light<br />

scattering and absorption coefficients after tissue coagulation was measured<br />

using techniques previously developed by our group. Heating induced<br />

by HIFU also influences the OA signal waveform due to the rise of<br />

thermal expansion coefficient of tissue with temperature. This dependence<br />

was measured in order to evaluate the feasibility of the OA technique in<br />

temperature monitoring. An OA image of HIFU lesion induced by a 1.1<br />

MHz focused transducer in a liver sample was reconstructed using a 64element<br />

wideband array transducer for OA signal detection.<br />

FRIDAY AFTERNOON, 1 DECEMBER <strong>2006</strong> OAHU ROOM, 1:00 TO 3:00 P.M.<br />

Session 4pEAa<br />

Engineering Acoustics: New Electroacoustic Transducers Utilizing Advanced Technolgies and Materials<br />

Juro Ohga, Cochair<br />

Shibaura Inst. of Technology, 3-9-14 Shibaura, Minato-ku, Tokyo 108-8548, Japan<br />

James E. West, Cochair<br />

Johns Hopkins Univ., Dept. of Electrical and Computer Engineering, Barton 105, 3400 N. Charles St.,<br />

Baltimore, MD 21218-2686<br />

Invited Papers<br />

1:00<br />

4pEAa1. Solid-state photo-microphones or pressure sensors by total reflection. Yasushi Suzuki �Gunma Natl. College 08 Tech.,<br />

580, Toriba-cho, Maebashi-shi, Gunma, 371-8530 Japan., suzuki@elc.gunma-ct.ac.jp� and Ken’iti Kido �Tohoku Univ.,<br />

Yokohama-shi, Kanagawa, 226-0017 Japan�<br />

Solid-state photo-microphones or pressure sensors are proposed. These sensors use a new principle, involving the optical total<br />

reflection at the boundary surface between glass and air. The critical angle for total reflection changes by the refractive index of air,<br />

which depends on the air density. Sound pressure changes the air density. Therefore, the sound pressure is measurable by detecting the<br />

intensity of the reflected light from the total reflection area. The sensitivity of the sensor is investigated theoretically. It is expected that<br />

the sensor has sufficient sensitivity for practical use, employing laser light and a curved boundary surface with a large radius of<br />

curvature. Some experiments are carried out to verify the theoretical investigations. A He-Ne laser or a laser diode is employed as a<br />

light source in the experiments. Experimental results show that the sensor has equivalent sensitivity to that which was theoretically<br />

estimated, but that sensitivity is very low. The sensor is useful as a pressure sensor, but it is difficult to realize a microphone for<br />

general use at the present. The microphones have no diaphragm and the upper limit in the frequency range is extremely high in<br />

principle.<br />

3271 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3271<br />

4p FRI. PM


1:20<br />

4pEAa2. Micromachined microphones with diffraction-based optical interferometric readout. F. Levent Degertekin �G.W.<br />

Woodruff School of Mech. Eng., Georgia Inst. of Technol., Atlanta, GA 30332, levent@gatech.edu�, Neal A. Hall �Sandia Natl. Labs,<br />

Albuquerque, NM 87185-5800�, and Baris Bicen �Georgia Inst. of Technol., Atlanta, GA 30332�<br />

A diffraction-based optical method for integrated interferometric detection of micromachined microphone diaphragm displacement<br />

is described. With multichip optoelectronics integration, this approach yields highly sensitive optical microphones in mm-cube<br />

volumes. Since the microphone sensitivity does not depend on capacitance, this method changes the paradigm for the backplate and<br />

gap structure design. As a result, one can use millimeter size diaphragms to achieve wide frequency response and low thermal<br />

mechanical noise levels characteristic of precision measurement microphones. Furthermore, the electrical port of the device, which is<br />

freed by optical detection, is used for electrostatic actuation of the microphone diaphragm to tune microphone sensitivity and to<br />

generate self-characterization signals. Prototype optical microphone structures have been fabricated using Sandia National Laboratories’<br />

silicon based SwIFT-Lite TM process. Measurements on these diaphragms show an A-weighted diaphragm displacement noise of<br />

2.4 pm and flat electrostatic response up to 20 kHz. These results indicate the feasibility of realizing measurement microphones with<br />

1.5-mm-diam diaphragms, 15-dBA internal noise, and 40-kHz bandwidth. Application of the detection method in a bio-inspired<br />

directional microphone for hearing aids is also discussed. �Work partially supported by NIH Grant 5R01DC005762-03, Sensing and<br />

Processing for Hearing Aids.�<br />

1:40<br />

4pEAa3. Hardware and software technologies for improvement of hearing characteristics of headphone reproduction.<br />

Kiyofumi Inanaga and Yuji Yamada �Audio Codec Development Dept., Technol. Development Group, SONY Corp., Shinagawa Tec.,<br />

12-15-3, Tokyo, 108-6201 Japan�<br />

This report specifically describes commercialization technology of a headphone system with out-of-head localization applying<br />

dynamic head-related transfer functions �HRTFs� that can localize sound easily over a full 360 deg. A source image by output of<br />

conventional headphones is localized inside the listener’s head. However, the image can be localized outside the listener’s head by<br />

wearing headphones over a full 360 deg through accurate simulation of the listener’s HRTFs. Developments of headphone systems<br />

using signal processing technology for data correction have given rise to the static binaural reproduction system �SBRS�. The first part<br />

of this speech describes its psychoacoustic characteristics and challenges. A rotating dummy-head that is synchronized with the<br />

listener’s head movement was produced experimentally to create the dynamic binaural reproduction system �DBRS�. Using the<br />

DBRS, HRTFs synchronize with the listener’s head movement. Psychoacoustic characteristics and advantages of the system are also<br />

discussed in this report. Further developments were made to realize the commercialization of the DBRS in areas including piezoelectric<br />

gyroscope head-tracking technology, headphone technologies that can reproduce real sound characteristics, and simplification of<br />

HRTF signal processing employing a simulator with electronic circuits. Finally, future visions for these technologies will be touched<br />

upon.<br />

2:00<br />

4pEAa4. Piezoelectret microphones: A new and promising group of transducers. Gerhard M. Sessler and Joachim Hillenbrand<br />

�Darmstadt Univ. of Technol., Merckstrasse 25, 64283 Darmstadt, Germany, g.sessler@nt.tu-darmstadt.de�<br />

Piezoelectret microphones, first described a few years ago, are transducers based on the strong longitudinal piezoelectric effect of<br />

charged cellular polymers. Such microphones have recently been improved in two respects: Firstly, an expansion process was used to<br />

increase the piezoelectric d 33 coefficients of cellular polypropylene �PP� films in the audio frequency range up to 600 pC/N and,<br />

secondly, stacking of several films was applied to increase the microphone sensitivity. Transducers with six films now show opencircuit<br />

sensitivities of up to 15 mV/Pa, comparable to that of electret microphones. Other characteristics of piezoelectret microphones<br />

are their low equivalent noise level of about 26 dB�A� and the very small total harmonic distortion of less than 0.1% at 140 dB SPL.<br />

The piezoelectric activity of the PP films and the microphone sensitivities are stable at <strong>room</strong> temperature but start to decay above<br />

50 °C. Recently, directional piezoelectret microphones with various directional characteristics have been designed. Major advantages<br />

of piezoelectret microphones are their simple design, their low harmonic distortion, and their wide frequency range extending into the<br />

ultrasonic region.<br />

2:20<br />

4pEAa5. Expansion of frequency range for piezoelectric loudspeakers by new transducer construction. Juro Ohga �Shibaura<br />

Inst. of Technol., 3-7-5, Toyosu, Koto-ku, Tokyo 135-8548, Japan�<br />

Although simple construction of piezoelectric loudspeakers engenders various merits, expansion of its working frequency range to<br />

the very low region is difficult because the mechanically stiff characteristics of conventional piezoelectric ceramic diaphragms prevent<br />

their large amplitude operation. This paper proposes two sorts of new piezoelectric loudspeaker construction that are suitable for<br />

low-frequency signal radiation. One idea is the use of a tuck-shape diaphragm by a PVDF polymer film bimorph. It has large surface<br />

area with a very low resonant frequency. Resonant frequencies and sensitivity frequency characteristics are examined, and control<br />

methods of local diaphragm bending are discussed. The other idea is the use of continuous revolution of a piezoelectric ultrasonic<br />

motor. It produces a completely controlled large output force because its output mechanical impedance is much greater than that of<br />

any conventional transducer or motor. An ultrasonic motor, whose stator is connected to a direct-radiator loudspeaker cone by a rod<br />

and whose rotor is burdened by a heavy metal ring, rotates with a constant velocity. Modulation of the velocity by using an audio<br />

signal imparts a driving force to the diaphragm because the heavy ring tends to keep a constant velocity. Experimental models suggest<br />

that this construction is useful.<br />

3272 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3272


2:40<br />

4pEAa6. Modal array signal processing using circular microphone arrays applied to acoustic source detection and localization<br />

problems. Heinz Teutsch �Avaya Labs, 233 Mt. Airy Rd., Basking Ridge, NJ 07920, teutsch@avaya.com� and Walter Kellermann<br />

�Univ. of Erlangen-Nuremberg, Erlangen, Germany�<br />

Many applications of acoustic signal processing rely on estimates of several parameters present in the observed acoustic scene<br />

such as the number and location of acoustic sources. These parameters have been traditionally estimated by means of classical array<br />

signal processing �CASP� algorithms using microphone arrays. Algorithms for parameter estimation solely based on the paradigm of<br />

CASP often suffer from the narrowband assumption underlying the signal model. This restriction limits their usability when wideband<br />

signals, such as speech, are present in the wave field under observation. We investigate the parameter estimation problem by applying<br />

the notion of wave field decomposition using baffled circular microphone arrays. The obtained wave field representation is used as the<br />

basis for ‘‘modal array signal processing algorithms.’’ It is shown that by applying the notion of modal array signal processing, novel<br />

algorithms can be derived that have the potential to unambiguously detect and localize multiple simultaneously active wideband<br />

sources in the array’s full field-of-view. Performance evaluations by means of simulations, measurements, and real-time case studies<br />

are presented.<br />

FRIDAY AFTERNOON, 1 DECEMBER <strong>2006</strong> OAHU ROOM, 3:15 TO 6:00 P.M.<br />

Session 4pEAb<br />

Engineering Acoustics: Special Topics in Engineering Acoustics<br />

Timothy W. Leishman, Cochair<br />

Brigham Young Univ., Dept. of Physics and Astronomy, N247 ESC, Provo, UT 84602<br />

3:15<br />

4pEAb1. Enhanced voided piezoelectric polymer for underwater<br />

acoustic sensors. Juan Arvelo �Appl. Phys. Lab., Johns Hopkins Univ.,<br />

11100 Johns Hopkins Rd., Laurel, MD 20723-6099�, Ilene<br />

Busch-Vishniac, and James West �Johns Hopkins Univ., Baltimore, MD<br />

21218�<br />

A charged voided polymer has been shown to exhibit large piezoelectricity.<br />

This material consists of injected air bubbles into polypropylene.<br />

This sheet of voided material is then biaxially stretched to elongate the<br />

voids. After stretching this material, a strong electric field is applied to<br />

cause dielectric breakdown of the gas in the voids, creating electric<br />

charges that are trapped in the polymer frame. Since the sides of the voids<br />

have opposite charges, they form macroscopic dipoles. When an external<br />

force is applied to this material, the voids become narrower, causing stronger<br />

dipole strength. A simple model of this voided material was implemented<br />

to derive formulas to estimate its piezoelectric constant, electromechanical<br />

coupling factor, resonance frequency, and sensor sensitivity<br />

based on electrical and mechanical properties of the polymer and gas in<br />

the voids. These formulas and a survey of available polymers and gases<br />

yielded promising combinations that result in more sensitive voided materials<br />

that satisfy selected criteria. These criteria include high sensitivity<br />

and maximum service temperature, low dissipation factor, and high dynamic<br />

compressibility, but low hydrostatic compressibility. This talk will<br />

describe the model, derive the formulas, uncover measured properties of<br />

candidate polymers and gases, and show calculated sensitivity of selected<br />

polymer/gas combinations.<br />

Kiyofumi Inanaga, Cochair<br />

Sony Corp., Shinagawa Tec. 12-15-3, Tokyo 108-6201, Japan<br />

Contributed Papers<br />

3:30<br />

4pEAb2. Basic study on one-dimensional transducer array using<br />

hydrothermally synthesized lead zirconium titanete poly-crystalline<br />

film. Akito Endo, Tomohito Hasegawa, Norimichi Kawashima, Shinichi<br />

Takeuchi �1614, Kurogane-cho, Aoba-ku, Yokohama, Kanagawa,<br />

225-8502, Japan�, Mutsuo Ishikawa, and Minoru Kurosawa �Midori-ku,<br />

Yokohama, Kanagawa 226-8502, Japan�<br />

Recently, high-frequency miniature medical ultrasound probes with<br />

high resolution were actively developed. However, it is difficult to fabricate<br />

such tiny ultrasound probes using piezoelectric ceramic vibrator with<br />

thickness less than 100 �m. We deposited a PZT poly-crystalline film on<br />

a titanium substrate using the hydrothermal method and developed transducers<br />

using the PZT poly-crystalline film for ultrasound probes. In this<br />

study, we applied it to a miniature medical one-dimensional �1-D�-arraytype<br />

ultrasound probe with resonance frequency of 10 MHz. After sputtering<br />

of pure titanium on the surface of a hydroxyapatite substrate, the<br />

titanium film was etched using the photolithography method to form a 1-D<br />

titanium film electrode array with 75 �m element pitch, 40 �m element<br />

width, and 4 mm element length to scan an ultrasound beam electronically<br />

by sector scan mode using phased-array technique. Thereby we fabricated<br />

a miniature 1-D-array-type ultrasound probe. A transmitted ultrasound<br />

pulse from 10 MHz commercial ultrasound probe was received by this<br />

fabricated 1-D-array type ultrasound probe with hydrothermally synthesized<br />

PZT poly-crystalline film vibrators.<br />

3273 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3273<br />

4p FRI. PM


3:45<br />

4pEAb3. Analysis of a barrel-stave flextensional transducer using<br />

MAVART „model to analyze the vibrations and acoustic radiation of<br />

transducers… and ATILA „analysis of transducers by integration of<br />

LAplace equations… finite-element codes. Richard A. G. Fleming, Mark<br />

Kwiecinski, and Dennis F. Jones �Defence R&D Canada—Atlantic, P.O.<br />

Box 1012, Dartmouth, NS, Canada B2Y 3Z7,<br />

dennis.jones@drdc-rddc.gc.da�<br />

A small barrel-stave flextensional transducer, designed and tested at<br />

Defence Research and Development Canada—Atlantic, is a candidate<br />

sound source for underwater coastal surveillance and acoustic communications<br />

applications. This high-power transducer has an outside diameter,<br />

length, and mass of 5.7 cm, 12.7 cm, and 1.1 kg, respectively. The measured<br />

fundamental flexural resonance frequency was 1.8 kHz with a transmitting<br />

voltage response of 118 dB/1�Pa-m/V and an omnidirectional<br />

radiation pattern. Two finite-element models were developed for this transducer<br />

using the finite-element codes MAVART �Model to Analyze the<br />

Vibrations and Acoustic Radiation of Transducers� and ATILA �Analysis<br />

of Transducers by Integration of LAplace equations�. Comparisons are<br />

made between the calibration measurements and the model predictions.<br />

�Work supported in part by Sensor Technology Limited.�<br />

4:00<br />

4pEAb4. Thermal behavior of high-power active devices with the<br />

ATILA „analysis of transducers by integration of LAplace equations…<br />

finite-element code. Jean-Claude Debus �Institut Superieur de<br />

l’Electronique et du Numerique, 41 Bv Vauban, 59046 Lille, Cedex<br />

France�, John BlottmanIII, and Stephen Butler �Naval Undersea Warfare<br />

Ctr. Div. Newport, RI 02841�<br />

Many active devices using piezoelectric ceramics are driven with very<br />

high power densities and long pulse lengths. Due to mechanical and dielectric<br />

losses in the materials, this produces heat, causing a temperature<br />

rise in the devices, which may lead to their mechanical failure. The thermal<br />

issues have been shown to be the limiting device design criteria over<br />

electric field and mechanical stress limits, yet the effect of the temperature<br />

on performance is generally not considered in the numerical models used<br />

during the design stage. A coupled electro-mechanical thermal analysis is<br />

implemented in the ATILA code. For a steady-state or transient solution, a<br />

thermal behavior is weakly coupled to the electromechanical response.<br />

The method may take advantage of the order-of-magnitude-greater time<br />

constant for thermal effects compared to mechanical behavior. A two-step<br />

analysis is performed whereby the electromechanical behavior is first<br />

computed, and the resulting dissipated power is then applied as a heat<br />

generator to determine the resulting temperature of the device. A highdrive,<br />

31-mode, free flooded ring transducer and a sonar projector serve as<br />

validation of the numerical model. The approach addresses both the transient<br />

thermal response and the steady temperature profile that results from<br />

the high-power, high-duty-cycle drive.<br />

4:15<br />

4pEAb5. Development of multichannel optical sensor and<br />

visualization of vibration distribution. Jun Hasegawa and Kenji<br />

Kobayashi �Faculty of Eng., Takushoku Univ., 815-1 Tatemachi,<br />

Hachioji-shi, Tokyo 193-0985 Japan, jhase@es.takushoku-u.ac.jp�<br />

A multi-channel optical sensor system was developed to measure vibrations<br />

simultaneously with high spatial resolution. As sensor elements,<br />

optical displacement sensor units were developed not to disturb the natural<br />

vibration. Each sensor unit, which consists of the optical fiber bundle and<br />

focusing lens, can detect the displacement of the object as the variation of<br />

the reflected light power. The sensor unit has a displacement resolution of<br />

10 nm, a dynamic range of more than 90 dB, and a frequency band width<br />

of up to 80 kHz. Up to 64 sensor units can be arrayed as one sensor head,<br />

which realizes the simultaneous measurement of vibration distribution<br />

with the high spatial resolution of 4 mm. A calibrating function under the<br />

measurement circumstances was developed. Under calibration mode, the<br />

sensor array head is moved by a linear actuator, while the vibration of the<br />

object is stopped. Thus the calibrated data of each sensor unit can be<br />

obtained for the displacement magnitude. Measured vibration distributions<br />

can be monitored as the three-dimensional animations. With the system<br />

developed, several actuators for vibratory micro-injection were measured,<br />

and the system could reveal their detailed vibration distributions and could<br />

detect the existence of a failure portion of some actuator.<br />

4:30<br />

4pEAb6. Prediction of howling for a sound system in an acoustical<br />

environment with both reverberant and direct sounds. Hideki<br />

Akiyama and Juro Ohga �Shibaura Inst. of Technol., 3-7-5 Toyosu,<br />

Koto-ku, Tokyo 135-8548, Japan, m106003@shibaura-it.ac.jp�<br />

Prediction of howling is a key technology for a howling suppression<br />

design for a sound system with a loudspeaker and microphone. A howling<br />

occurrence prediction method for a sound system in a reverberant <strong>room</strong><br />

has already been presented �J. Ohga, J. Sakaguchi, ‘‘Prediction of howling<br />

of a sound system in a reverberant <strong>room</strong>,’’ W. C. Sabine Centennial Symposium<br />

�ASA, New York, 1994�, 2aAAd4�. It is apparently useful for<br />

ordinary public address systems whose distances of loudspeakers from<br />

microphones are large. However, this result was not perfect because the<br />

direct sound component is not negligible in hands-free telephones or teleconference<br />

systems whose loudspeakers and microphones are set close to<br />

each other. This report gives a quantitative howling occurrence prediction<br />

method for a sound system in an acoustical environment with both reverberant<br />

and direct sounds. The following design parameters are obtained:<br />

�1� the increase of howling occurrence level from the power average<br />

value, �2� the level occurrence probability, and �3� the critical level chart<br />

given by an equation as a function of direct and reverberant sounds ratio.<br />

Prediction results for particular examples are compared with calculations<br />

of sound-field transfer functions. Results confirmed that it is practical.<br />

4:45<br />

4pEAb7. Effect of background noise on dialogue in telephony. Koichi<br />

Amamoto and Juro Ohga �Sibaura Inst. of Technol., 3-7-5 Toyosu,<br />

Koto-ku, Tokyo, 135-8548, Japan, m106006@shibaura-it.ac.jp�<br />

Recent developments of mobile telephones include new sorts of impairments<br />

against speech. Conventional evaluation method for impairments<br />

by a talker and a few listeners cannot apply to these new ones,<br />

because they are brought by long signal delay. The effect of it cannot<br />

discriminate by ‘‘one-sided’’ test. This research relates to a speech quality<br />

evaluation by conversation between two persons. Variation of conversation<br />

stream is observed by addition of pink noise of various levels to a dialogue<br />

by microphones and earphones. Length of sentences and frequency of<br />

repeats are quantified and their meanings are discussed<br />

5:00<br />

4pEAb8. Best practices for auditory alarm design in space<br />

applications. Durand Begault and Martine Godfroy �Human Systems<br />

Integration Div., NASA Ames Res. Ctr., Moffett Field, CA 94035�<br />

This presentation reviews current knowledge in the design of auditory<br />

caution and warning signals, and sets criteria for development of ‘‘best<br />

practices’’ for designing new signals for NASA’s Crew Exploration Vehicle<br />

�CEV� and other future spacecraft, as well as for extra-vehicular<br />

operations. A design approach is presented that is based upon crossdisciplinary<br />

examination of psychoacoustic research, human factors experience,<br />

aerospace practices, and acoustical engineering requirements. Existing<br />

alarms currently in use with the NASA Space Shuttle flight deck are<br />

analyzed and then alternative designs are proposed that are compliant with<br />

ISO 7731 �‘‘Danger signals for work places Auditory Danger Signals’’�<br />

and that correspond to suggested methods in the literature to insure discrimination<br />

and audibility. Parallel analyses are shown for a sampling of<br />

medical equipment used in surgical, periop, and ICU contexts. Future<br />

development of auditory sonification techniques into the design of alarms<br />

will allow auditory signals to be extremely subtle, yet extremely useful in<br />

subtly indicating trends or root causes of failures. �Work funded by<br />

NASA’s Space Human Factors Engineering Project.�<br />

3274 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3274


5:15<br />

4pEAb9. Acoustic signal analysis for forensic applications. Durand<br />

Begault and Christopher Peltier �Audio Forensic Ctr., Charles M. Salter<br />

Assoc., 130 Sutter St., Ste. 500, San Francisco, CA 94104,<br />

durand.begault@cmsalter.com�<br />

Acoustical analysis of audio signals is important in many legal contexts<br />

for determining the authenticity, originality, and continuity of recorded<br />

media; determining the circumstances of events in question that<br />

may have been recorded; for determining the audibility of signals; and for<br />

identification or elimination of talkers as a match to an unknown exemplar.<br />

Recorded media are analyzed in forensic applications using both familiar<br />

techniques �waveform and spectral analyses� and more novel methods<br />

�e.g., ferro fluid development of media; specialized tape heads with nonstandard<br />

reproduction characteristics; crystal microscopy; detection and<br />

matching to power grid frequencies�. Audibility analyses frequently require<br />

careful reconstructive field measurements and criteria in excess of<br />

normally accepted standards. Voice identification-elimination protocols<br />

must account for examiner bias and exemplar quality and can be described<br />

using a receiver operator curve �ROC� model. This presentation gives an<br />

overview of these techniques and their comparative advantages for specific<br />

forensic applications.<br />

5:30<br />

4pEAb10. Without low-pass filter for the 1-bit digital amplifier.<br />

Kiyoshi Masuda �Coroprate Res. and Development Group, SHARP<br />

Corp., 2613-1 Ichinomoto-cho, Tenri-shi, Nara, Japan� and Yoshio<br />

Yamasaki �Waseda Univerity, Okubo, Shinzyuku-ku, Tokyo, Japan�<br />

SHARP collaborated with Waseda University from 1990 for 1-bit digital<br />

technology. SHARP had started to receive an order for the 1-bit digital<br />

amplifier ‘‘SM-SX100’’ on 20 August 1999. Until today, we have introduced<br />

the 1-bit digital amplifier for audio, flat panel TV �LCD TV�, and<br />

PC. These 1-bit amplifiers provided low-pass filter for the final stage,<br />

which is provided after 1-bit digital switching. We have to achieve more<br />

good sound and reduce deterioration of this low-pass filter. We have introduced<br />

new 1-bit digital amplifier without this low-pass filter beginning<br />

this April. This means we controlled the 1-bit digital signal to directly<br />

operate the speaker. We have proved a better effect for sound to compare<br />

the new 1-bit digital amplifier with the PWM switching amplifier, the<br />

A-class amplifier and the 1-bit digital amplifier with low-pass filter. If we<br />

do not measure any improvement for this new 1-bit digital amplifier, it has<br />

large radiation noise. We had achieved a reduction to the limit level of<br />

FCC, Denanhou, etc.<br />

5:45<br />

4pEAb11. Force-frequency effect of thickness mode langasite<br />

resonators. Haifeng Zhang �W317.4 Nebraska Hall, Univ. of Nebraska,<br />

Lincoln, NE 68588-0526, hfzhang@bigred.unl.edu�, Joseph A. Turner,<br />

Jiashi Yang �Univ. of Nebraska, Lincoln, NE 68588-0526�, and John A.<br />

Kosinski �U.S. Army CECOM, Fort Monmouth, NJ 07703-5211�<br />

Langasite resonators are of recent interest for a variety of applications<br />

because of their good temperature behavior, good piezoelectric coupling,<br />

low acoustic loss, and high Q factor. The force-frequency effect describes<br />

the shift in resonant frequency a resonator experiences due to the application<br />

of a mechanical load. A clear understanding of this effect is essential<br />

for many design applications such as pressure sensors. In this presentation,<br />

the frequency shift is analyzed theoretically and numerically for thin, circular<br />

langasite plates subjected to a diametrical force. The results are<br />

compared with experimental measurements of the same system for a variety<br />

of langasite resonators with various material orientations. In addition,<br />

the sensitivity of force-frequency effect is analyzed with respect to the<br />

nonlinear material constants. A comparison between the force-frequency<br />

effect of langasite and quartz resonators is also made. Finally, the application<br />

of such measurements for determining third-order elastic constants<br />

is discussed. �Work supported by ARO.�<br />

FRIDAY AFTERNOON, 1 DECEMBER <strong>2006</strong> IAO NEEDLE/AKAKA FALLS ROOM, 1:20 TO 4:25 P.M.<br />

Session 4pMU<br />

Musical Acoustics and Psychological and Physiological Acoustics: Acoustic Correlates of Timbre in Musical<br />

Instruments<br />

James W. Beauchamp, Cochair<br />

Univ. of Illinois Urbana-Champaign, School of Music, Dept. of Electrical and Computer Engineering, 1002 Eliot Dr.,<br />

Urbana, IL 61801<br />

Mashashi Yamada, Cochair<br />

Kanazawa Inst. of Technology, Dept. of Media Informatics, 3-1 Yatsukaho, Hakusan, Ishikawa 924-0838, Japan<br />

Invited Papers<br />

1:20<br />

4pMU1. A meta-analysis of acoustic correlates of timbre dimensions. Stephen McAdams, Bruno Giordano �CIRMMT, Schulich<br />

School of Music, McGill Univ., 555 Sherbrooke St. West, Montreal, QC, Canada H3A 1E3�, Patrick Susini, Geoffroy Peeters<br />

�STMS-IRCAM-CNRS, F-75004 Paris, France�, and Vincent Rioux �Maison des Arts Urbains, F-75020 Paris, France�<br />

A meta-analysis of ten published timbre spaces was conducted using multidimensional scaling analyses �CLASCAL� of dissimilarity<br />

ratings on recorded, resynthesized, or synthesized musical instrument tones. A set of signal descriptors derived from the tones<br />

was drawn from a large set developed at IRCAM, including parameters derived from the long-term amplitude spectrum �slope,<br />

centroid, spread, deviation, skewness, kurtosis�, from the waveform and amplitude envelope �attack time, fluctuation, roughness�, and<br />

from variations in the short-term amplitude spectrum �flux�. Relations among all descriptors across the 128 sounds were used to<br />

determine families of related descriptors and to reduce the number of descriptors tested as predictors. Subsequently multiple correlations<br />

between descriptors and the positions of timbres along perceptual dimensions determined by the CLASCAL analyses were<br />

3275 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3275<br />

4p FRI. PM


computed. The aim was �1� to select the subset of acoustic descriptors �or their linear combinations� that provided the most generalizable<br />

prediction of timbral relations and �2� to provide a signal-based model of timbral description for musical instrument tones.<br />

Four primary classes of descriptors emerge: spectral centroid, spectral spread, spectral deviation, and temporal envelope �effective<br />

duration/attack time�. �Work supported by CRC, CFI, NSERC, CUIDADO European Project.�<br />

1:40<br />

4pMU2. Perceptual acoustics of consonance and disssonance in multitimbral triads. Roger Kendall �Music Cognition and<br />

Acoust. Lab., Program in Systematic Musicology, UCLA, 405 Hilgard Ave., Los Angeles, CA 90095, kendall@ucla.edu� and Pantelis<br />

Vassilakis �DePaul Univ., Chicago, IL 60614�<br />

Most studies of consonance and dissonance assume a singular spectrum for the constituent intervals of a dyad. Recently, the<br />

principal author conducted experiments evaluating triads consisting of digitally mixed combinations drawn from the MUMS singlenote<br />

natural-instrument recordings. Results indicated that the main effect of ratings for consonance and dissonance correlated well<br />

with studies using artificial signals. However, interaction effects suggested perceptual differences related to the timbral differences<br />

across combinations. The present experiment evaluates perceptual and acoustical variables of the ten possible triadic combinations<br />

created with C4 as the lower and the ten with C5 as the upper notes. UCLA wind ensemble performers on oboe, flute, and clarinet,<br />

combinations designed to span timbral space, were digitally recorded. Analyses include perceptual ratings of consonance and dissonance,<br />

similarity, as well as acoustical analysis of roughness using a recently developed model. Since natural performances of any type<br />

vary in fundamental frequency, additional experiments will employ emulated oboe, flute, and clarinet �using the Kontakt Silver<br />

synthesizer in Sibelius 4� as well as purely synthetic stimuli, in order to ascertain the relationship of time-variant spectral properties<br />

to consonance, dissonance, and perceived similarity.<br />

2:00<br />

4pMU3. Multidimensional scaling analysis of centroid- and attackÕdecay-normalized musical instrument sounds. James W.<br />

Beauchamp �School of Music and Dept. of Elect. & Comput. Eng., Univ. of Illinois at Urbana–Champaign, Urbana, IL 61801,<br />

jwbeauch@uiuc.edu�, Andrew B. Horner �Hong Kong Univ. of Sci. & Technol., Kowloon, Hong Kong�, Hans-Friedrich Koehn, and<br />

Mert Bay �Univ. of Illinois at Urbana-Champaign, Urbana, IL 61801�<br />

Ten sustained musical instrument tones �bassoon, cello, clarinet, flute, horn, oboe, recorder, alto saxophone, trumpet, and violin�<br />

were spectrally analyzed and then equalized for duration, attack and decay time, fundamental frequency, number of harmonics,<br />

average spectral centroid, and presentation loudness. The tones were resynthesized both with time-varying harmonic amplitudes and<br />

frequencies �dynamic case� and fixed amplitudes and frequencies �static case�. Tone triads were presented to ten musically experienced<br />

listeners whose tasks were to specify the most dissimilar and most similar pairs in each triad. Based on the resulting dissimilarity<br />

matrix, multidimensional scaling �MDS� was used to position the instruments in two- and three-dimensional metric spaces. Two<br />

measures of instrument amplitude spectra were found to correlate strongly with MDS dimensions. For both the static- and dynamiccase<br />

2-D solutions, the ratio of even-to-odd rms amplitudes correlated strongly with one of the dimensions. For the dynamic case,<br />

spectral centroid variation correlated strongly with the second dimension. Also, 2-D solution instrument groupings agreed well with<br />

groupings based on coefficients of the first two components of a principle components analysis representing 90% of the instruments’<br />

spectral variance. �This work was supported by the Hong Kong Research Grants Council’s CERG Project 613505.�<br />

2:20<br />

4pMU4. Sound synthesis based on a new micro timbre notation. Naotoshi Osaka �School of Eng., Tokyo Denki Univ., 2-2,<br />

Kanda-Nishikicho, Chiyoda-ku, Tokyo, 101-8457, Japan, osaka@im.dendai.ac.jp�, Takayuki Baba, Nobuhiko Kitawaki, and Takeshi<br />

Yamada �Univ. of Tsukuba, Japan�<br />

Timbre has become a major musical factor in contemporary and computer music. However, sufficient timbre theory has not yet<br />

been established. The author is challenging to create new timbre theory for music composition. The first step of its construction is to<br />

make the timbre descriptive. A micro timbre is defined, which is a perceptual impression of a sound with approximately 50 to 100-ms<br />

duration, and describe sound as a micro timbre sequence. This can be used as a new notation system in place of common music<br />

notation. In dictation, micro timbre sequence and correspondent duration sequence are perceptually recorded. When synthesizing from<br />

this notation, sounds corresponding to the notation systems are either physically synthesized or searched for in a large sound database<br />

to generate sound data for a given duration. Two sequential sound data instances are first represented in sinusoidal representations and<br />

then are concatenated using a morphing technique. Sounds generated by a stream of water and similar sounds are described using the<br />

method as examples. Then scripts describing electronic sounds are introduced and explained. The ability to record, transmit to others,<br />

and resynthesize timbre is one of the useful functions of the theory.<br />

2:40<br />

4pMU5. Timbre representation for automatic classification of musical instruments. Bozena Kostek �Gdansk Univ. of Technol.,<br />

Narutowicza 11/12, PL-80-952 Gdansk, Poland�<br />

Human communication includes the capability of recognition. This is particularly true of auditory communication. Music information<br />

retrieval �MIR� turns out to be particularly challenging, since many problems remain still unsolved. Topics that should be<br />

included within the scope of MIR are automatic classification of musical instruments/phrases/styles, music representation and indexing,<br />

estimating musical similarity using both perceptual and musicological criteria, recognizing music using audio and/or semantic<br />

description, language modeling for music, auditory scene analysis, and others. Many features of music content description are based<br />

on perceptual phenomena and cognition. However, it can easily be observed that most of the low-level descriptors used, for example,<br />

in musical instrument classification are more data- than human-oriented. This is because the idea behind these features is to have data<br />

3276 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3276


defined and linked in such a way as to be able to use it for more effective automatic discovery, integration, and reuse in various<br />

applications. The ambitious task is, however, to provide seamless meaning to low- and high-level descriptors such as timbre descriptors<br />

and linking them together. In such a way data can be processed and shared by both systems and people. This paper presents a<br />

study related to timbre representation of musical instrument sounds.<br />

3:00–3:15 Break<br />

3:15<br />

4pMU6. An attempt to construct a quantitative scale of musical brightness for short melodies implementing timbral<br />

brightness. Masashi Yamada �Kanazawa Inst. of Technol., 3-1 Yatsukaho, Hakusan, Ishikawa 924-0838, Japan,<br />

m-yamada@neptune.kanazawa-it.ac.jp�<br />

It is known that a major tune is brighter than a minor one, and that music played in a faster tempo and a higher register is brighter<br />

than a slower and lower one. However, it has not been clarified how these factors quantitatively determine the musical brightness. On<br />

the other hand, it has been clarified that the timbral brightness of a tone corresponds well to the spectral centroid. In the present study,<br />

major and minor scales and two short melodies were played with pure tones, and listeners evaluated their musical brightness. For each<br />

performance, the spectral centroid was calculated for the overall-term spectrum during the performance on the transformed frequency<br />

scale of the ERB rate. The results showed that the musical brightness of the ascending scale increases proportionally as the spectral<br />

centroid shown in the ERB rate increases. Using this, a quantitative scale of musical brightness, BM, was constructed. The results also<br />

showed that the difference in the musical brightness between major and minor scales corresponded to the transposition of approximately<br />

5 ERB rate, and doubling the speed corresponded to the upper shift of the centroid in approximately 2.5 ERB rate.<br />

3:35<br />

4pMU7. Subjective congruency between a sound effect and a switching pattern of a visual image. Shinichiro Iwamiya,<br />

Motonori Arita, and Sun Su �Dept. of Acoust. Design, Kyushu Univ., 4-9-1, Shiobaru, Minami-ku, Fukuoka 185-8540, Japan�<br />

The relationship between the transformation of a visual image and the pitch pattern of sound can create formal congruency<br />

between sounds and moving pictures. This effect by switching patterns of enlarging and reducing images that were combined with<br />

ascending and descending pitch scales was examined. Rating experiments showed two congruent patterns of the combination of<br />

switching and scale patterns; one was a combination of an ascending pitch scale and an enlarging image pattern, and the other a<br />

combination of a descending pitch scale and a reducing image pattern. These forms of matching might be based on a Doppler illusion.<br />

An additional pair of congruent patterns for combinations of switching and scale patterns was also found: one was a combination of<br />

an ascending pitch scale and a sliding movement from left to right, and the other a combination of a descending pitch scale and a<br />

sliding movement from right to left. These forms of matching might be based on the correspondence of a progressive sensation.<br />

Further, the formal congruency between a pitch pattern and the formal transportation can contribute to integrating auditory and visual<br />

information and to making audio-visual products more impressive.<br />

3:55<br />

4pMU8. SRA: An online tool for spectral and roughness analysis of<br />

sound signals. Pantelis Vassilakis �School of Music, ITD, Libraries,<br />

DePaul Univ., 2350 N. Kenmore Ave., JTR 207, Chicago, IL 60614�<br />

SRA performs spectral and roughness analysis on user-submitted 250to<br />

1000-ms-long portions of sound files �.wav/.aif formats�. Spectral<br />

analysis incorporates an improved STFT algorithm �K. Fitz and L. Haken,<br />

J. Aud. Eng. Soc. 50�11�, 879–893 �2002�� and automates spectral peakpicking<br />

using the Loris open source C�� class library �Fitz and Haken<br />

�CERL Sound Group��. Users can manipulate three spectral analysis/peakpicking<br />

parameters: analysis bandwidth, spectral-amplitude normalization,<br />

and spectral-amplitude threshold. Instructions describe the parameters in<br />

detail and suggest settings appropriate to the submitted files and questions<br />

of interest. The spectral values obtained from the analysis enter a roughness<br />

estimation model �P. N. Vassilakis, Sele. Rep. in Ethnomusicol. 12,<br />

119–144 �2005��, outputting roughness values for each individual sinepair<br />

in the file’s spectrum and for the entire file. The roughness model<br />

quantifies the dependence of roughness on a sine-pair’s �a� intensity �combined<br />

amplitude of the sines�, �b� amplitude fluctuation degree �amplitude<br />

difference of the sines�, �c� amplitude fluctuation rate �frequency difference<br />

of the sines�, and �d� register �lower sine frequency�. Presentation of<br />

the roughness estimation model and the online tool will be followed by a<br />

discussion of research studies employing it and an outline of future possible<br />

applications. �Work supported by DePaul University and Eastern<br />

Washington University. Programmed by K. Fitz.�<br />

Contributed Papers<br />

4:10<br />

4pMU9. Further spectral correlations of timbral adjectives used by<br />

musicians. Alastair C. Disley, David M. Howard, and Andrew D. Hunt<br />

�Dept. of Electron., Univ. of York, Heslington, York, YO10 5DD, UK�<br />

As part of a project to develop a synthesis interface which nontechnical<br />

musicians should find intuitive, the adjectives musicians use to describe<br />

timbre have been studied in a large-scale listening test covering the<br />

timbre space of Western orchestral instruments. These were refined in<br />

previous work by the authors �A. C. Disley et al. ‘‘Spectral correlations of<br />

timbral adjectives used by musicians,’’ J. Acoust. Soc. Am. 119, 3333,<br />

�<strong>2006</strong>�� to a set of ten words which had good common understanding and<br />

discrimination between the samples �bright, clear, dull, gentle, harsh, nasal,<br />

percussive, ringing, thin, and warm�. To help explore potential relationships<br />

between these adjectives and spectral features, 20 listeners participated<br />

in a further listening experiment, comparing samples in pairs to<br />

produce dissimilarity data. Multidimensional scaling produced dimensions<br />

which were compared with a large number of spectral and time-domain<br />

analyses of the stimuli, suggesting a number of significantly correlated<br />

spectral cues with some of the adjectives. These results are compared with<br />

previous studies by the authors and others, showing both similarities and<br />

differences, suggesting that collective consideration of timbral adjectives<br />

is more likely to result in simultaneously applicable theories of acoustic<br />

correlates than individual consideration of words.<br />

3277 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3277<br />

4p FRI. PM


FRIDAY AFTERNOON, 1 DECEMBER <strong>2006</strong> MAUI ROOM, 1:00 TO 2:00 P.M.<br />

Session 4pNSa<br />

Noise and Architectural Acoustics: Soundscapes and Cultural Perception II<br />

Brigitte Schulte-Fortkamp, Cochair<br />

Technical Univ. Berlin, Inst. of Technical Acoustics, Secr TA 7, Einsteinufer 25, 10587 Berlin, Germany<br />

Bennett M. Brooks, Cochair<br />

Brooks Acoustics Corp., 27 Hartford Turnpike, Vernon, CT 06066<br />

1:00<br />

4pNSa1. Mapping soundscapes in urban quiet areas. Gaetano Licitra<br />

�ARPAT, Tuscany Regional Agency for Environ. Protection, Via N.<br />

Porpora, 22-50144, Firenze, Italy�, Gianluca Memoli �Memolix, Environ.<br />

Consultants, 56127 Pisa, Italy�, Mauro Cerchiai, and Luca Nencini<br />

�ARPAT, 56127 Pisa, Italy�<br />

Innovative action plans in noise-polluted environments require the description<br />

of the existing soundscape in terms of suitable indicators. The<br />

role of these indicators, giving the ‘‘fingerprint’’ of a fixed soundscape,<br />

would be not only to measure the improvement in the sound quality after<br />

the action taken, but also to guide the designer in the process, providing a<br />

reference benchmark. One of the open questions on new indicators is the<br />

way they relate to existing ones and to people’s perception. The present<br />

work will describe a ‘‘Sonic Garden’’ in Florence, using both the ‘‘slope’’<br />

indicator �constructed from the LA eq time history and related in previous<br />

studies to people’s perception� and classical psychoacoustical parameters<br />

�level, spectral structure, and perceived characteristics such as loudness,<br />

sharpness, fluctuation, and roughness�. The latter parameters will be acquired<br />

using a binaural technique.<br />

Contributed Papers<br />

1:30–2:00<br />

Panel Discussion<br />

1:15<br />

4pNSa2. A questionnaire survey of the attitude of Japanese and<br />

foreign residents in Japan to sound masking devices for toilets.<br />

Miwako Ueda and Shin-ichiro Iwamiya �Grad. School of Design, Kyushu<br />

Univ., Iwamiya Lab. 4-9-1, Shiobaru, Minami-ku, Fukuoka 815-8540<br />

Japan, amaria@white.livedoor.com�<br />

Unique sound masking devices for toilets can be used in women’s<br />

rest<strong>room</strong>s in Japan. Such devices function to produce the sound of flushing<br />

water without actual flushing. To mask the sound of bodily functions,<br />

women tended to flush the toilet continuously while using it, thereby wasting<br />

a large amount of water. In the circumstances, sound masking devices<br />

have been introduced to public toilets. We have recently conducted a questionnaire<br />

survey to clarify the attitude of people toward such sound masking<br />

devices for toilets. The results of the survey showed that many Japanese<br />

women know such devices and often use them, that foreign women<br />

currently living in Japan also know that such devices exist, and that some<br />

Japanese men have heard of such devices but never used them. Many<br />

Japanese women are quite embarrassed at the thought that someone else<br />

can hear them while they are on the toilet. Many noted the necessity of<br />

such devices and required a wide range of setting for toilets in public<br />

spaces. However, they are not satisfied with the sound quality of the playback<br />

toilet flush sounds of currently available devices. The above results<br />

suggest that the sound quality of such devices should be improved.<br />

FRIDAY AFTERNOON, 1 DECEMBER <strong>2006</strong> MAUI ROOM, 2:10 TO 4:35 P.M.<br />

Session 4pNSb<br />

Noise, Physical Acoustics, and Structural Acoustics and Vibration: Acoustics of Sports<br />

Joseph Pope, Cochair<br />

Pope Engineering Company, P.O. Box 590236, Newton, MA 02459-0002<br />

Kenji Kurakata, Cochair<br />

AIST, 1-1-1 Higashi, Tsukuba, Ibaraki 305-8566, Japan<br />

Chair’s Introduction—2:10<br />

Invited Papers<br />

2:15<br />

4pNSb1. A review of the vibration and sounds of the crack of the bat and player auditory clues. Robert Collier �Thayer School<br />

of Eng., 8000 Cummings Hall, Hanover, NH 03755�<br />

The purpose of this paper is to review the state-of-the-art in the acoustics of baseball. As is well known, the crack of the bat is an<br />

important phenomenon of solid wood bats and metal bats. Each has a very different sound signature. At the 148th meeting of the ASA<br />

in 2004, the author and coauthors Ken Kaliski and James Sherwood presented the results of laboratory and field tests, which showed<br />

3278 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3278


that the spectral characteristics of radiated sound are dependent on the ball-bat impact location and resultant bat vibration of both solid<br />

wood and tubular metal bats. These results will be reviewed together with those of other investigators in the context of player auditory<br />

clues and the player’s response in game situations.<br />

2:35<br />

4pNSb2. Measurements of the impact sound of golf clubs and risk of hearing impairment. Kenji Kurakata �Natl. Inst. of Adv.<br />

Industrial Sci. and Technol. �AIST�, 1-1-1 Higashi, Tsukuba, Ibaraki, 305-8566 Japan, kurakata-k@aist.go.jp�<br />

The ball-impact sounds of golf clubs with metal heads and of a club with a wood head were measured to investigate their different<br />

acoustic properties. Hitting was executed using either a swing machine or a human player. Results of these analyses showed that the<br />

metal-head clubs generated sounds around 100 dB (L pA,Fmax). This level was 5–15 dB higher than that of the wood-head club. The<br />

sounds of the metal-head clubs had greater power in the high-frequency region of 4 kHz and above compared to the wood-head club,<br />

which particularly increased the overall sound levels. These results suggest that it would be desirable to develop a metal head with<br />

pleasant sound qualities, keeping the sound level lower to minimize hearing damage. Some of these measurement data were published<br />

in Japanese in a previous paper �K. Kurakata, J. INCE/J 26, 60–63�2002��.<br />

2:55<br />

4pNSb3. New underwater sound system for synchronized swimming: The 9th International Swimming Federation<br />

Championships. Takayuki Watanabe, Shinji Kishinaga �YAMAHA Ctr. for Adv. Sound Technologies, 203 Matsunokijima, Iwata,<br />

Shizuoka, 438-0192 Japan�, Tokuzo Fukamachi �YAMAHA Motor Marine Operations, Arai, Hamana, 438-8501 Japan�, and Osamu<br />

Maeda �YAMAHA Motor Adv. Technol. Res. Div., Iwata, 438-8501 Japan�<br />

There have been concerns about the differences between underwater sound fields in a temporary fiberglass-reinforced plastic<br />

�FRP� pool and in a conventional reinforced concrete �RC� pool. A temporary FRP pool was to be used for competitions at the World<br />

Swimming Championships in Fukuoka. We considered three items as key factors for a swimming pool used for synchronized<br />

swimming: �1� the sound source itself �output level, fluctuations in frequency characteristics�; �2� the effect of materials used in pool<br />

construction upon sound source installation conditions; and �3� the effect of the mth mode low-frequency cutoff in ‘‘shallow water.’’<br />

To improve basic problems related to the first factor, we developed a new actuator-driven underwater sound system �YALAS�, which<br />

can eliminate the effect of installation conditions for underwater speakers in the FRP pool. This new underwater system has now seen<br />

practical use in competitions. The report summarizes this new underwater sound system and compares the system with conventional<br />

systems in terms of its acoustic characteristics. The system can offer music with sufficient audibility in water. We gained a good<br />

reputation with competitors because the system showed superior performance to conventional systems in sound volume and quality,<br />

and in uniformity of sound distribution.<br />

3:15<br />

4pNSb4. Acoustics of the Great Ball Court at Chichen Itza, Mexico. David Lubman �14301 Middletown Ln., Westminster, CA<br />

92683�<br />

The ball game has played a central role in Mayan religion and culture for 5000 years. Thousands of ball courts have been<br />

discovered. The Great Ball Court �GBC� at Chichen Itza is a late development and is architecturally unique. Two remarkable<br />

acoustical features were noticed during excavation in the 1920s, but never explained or interpreted. A whispering gallery permits voice<br />

communication between temples located about 460 feet �140 m� apart. A profound flutter echo is heard between the two massive<br />

parallel walls of the playing field, about 270 ft �82 m� long, 28 ft �8.5 m� high, and 119 ft �36 m� apart. Until recently, most<br />

archaeologists dismissed acoustical features at Mayan sites as unintended artifacts. That is now changing. Stimulated by archaeological<br />

acoustic studies and reports since 1999, eminent Mayanists Stephen Houston and Karl Taube have reinterpreted certain Mayan<br />

glyphs as vibrant sounds and ballcourt echoes, and have famously called for a new archaeology of the senses, especially hearing, sight,<br />

and smell �Cambridge Archaeol. J. 10 �2� 261–294 �2000��. By interpreting architectural, psychoacoustic, and cognitive features of the<br />

GBC in the context of ancient Mayan culture, this paper speculates that acoustical effects at the GBC may be original design features.<br />

3:35<br />

4pNSb5. Sleep disturbance caused by shooting sounds. Joos Vos<br />

�TNO Human Factors, P.O. Box 23, 3769 ZG Soesterberg, The<br />

Netherlands, joos.vos@tno.nl�<br />

In the present study relations between the sound level of shooting<br />

sounds and the probability of behaviorally confirmed noise-induced awakening<br />

reactions were determined. The sounds were presented by means of<br />

loudspeakers in the bed<strong>room</strong>s of 30 volunteers. The shooting sounds had<br />

been produced by a small and a medium-large firearm, and the stimuli<br />

consisted of individual bangs or volleys of ten isolated or partly overlapping<br />

impulses. Aircraft sound was included as a reference source. The<br />

sounds were presented during a 6-h period that started 75 min after the<br />

beginning of the sleeping period. The time period between the various<br />

stimuli varied between 12 and 18 min, with a mean of 15 min. To cope<br />

with at least a relevant portion of habituation effects, each subject participated<br />

in 18 nights to be completed within 4 weeks. Preliminary results are<br />

presented both for the awakening reactions described above, and for vari-<br />

Contributed Papers<br />

ous other dependent variables collected with the help of an actimeter or<br />

determined by means of subjective rating scales. �Work supported by the<br />

Dutch Ministry of Defense.�<br />

3:50<br />

4pNSb6. Sound inside a gymnasium. Sergio Beristain �ESIME, IPN,<br />

IMA., P.O. Box 12-1022, Narvarte, 03001, Mexico City, Mexico�<br />

A new gymnasium for a sports club was designed taking acoustic<br />

confort into consideration, in order to accomodate sports practice, sports<br />

events with the public, or musical and drama presentations, taking advatage<br />

of its large capacity for the public and performers. The floor plan<br />

included <strong>room</strong> enough for a basketball court with public space on one<br />

side, where grades for about 200 people will be permanently installed.<br />

Walls were treated in a way that is useful for the sports practice �hard<br />

surfaces�, with hidden absorption material to reduce the usual reverberant<br />

3279 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3279<br />

4p FRI. PM


field inside the court, and to allow for sound events with only the addition<br />

ofamat�to protect the floor woodwork� and extra grades and the sound<br />

reinforcement system.<br />

4:05<br />

4pNSb7. Occupational and recreational noise exposures at stock car<br />

racing circuits. Chucri A. Kardous, Thais Morata, and Luann E. Van<br />

Campen �Natl. Inst. for Occupational Safety Health, 4676 Columbia<br />

Pkwy., Cincinnati, OH 45226, ckardous@cdc.gov�<br />

Noise in stock car racing is accepted as a normal occurrence but the<br />

exposure levels associated with the sport have not been adequately characterized.<br />

Researchers from the National Institute for Occupational Safety<br />

and Health �NIOSH� conducted an exploratory assessment of noise exposures<br />

to drivers, pit crew, team staff, and spectators at three stock car<br />

racing events. Area measurements were made during race preparation,<br />

practice, qualification, and competition. Personal dosimetry measurements<br />

were conducted on drivers, crew members, infield staff, and spectators.<br />

Findings showed time-weighted averages �TWA� that ranged from 94<br />

decibels A-weighted �dBA� for spectators to 114 dBA for car drivers. Peak<br />

sound-pressure levels exceeded the maximum allowable limit of 140 decibels<br />

�dB� during race competitions. Personal exposure measurements exceeded<br />

the NIOSH recommended exposure limit �REL� of 85 dBA as an<br />

8-h TWA in less than a minute for one driver during practice, within 2 min<br />

for pit crew and infield staff, and 7 to 10 min for spectators during the<br />

race. Hearing protection use was variable and intermittent among crew,<br />

staff, and spectators. Among drivers and crew, there was greater concern<br />

for communication performance than for hearing protection.<br />

4:20<br />

4pNSb8. Sports acoustics: Using sound from resonant shells,<br />

vibrating cylinders, strumming shafts, and water impact to evaluate<br />

athletic performance. David G. Browning �Dept. of Phys., Univ. of<br />

Rhode Island, Kingston, RI 02881, decibeldb@aol.com� and Peter M.<br />

Scheifele �Univ. of Connecticut, Storrs, CT 06269�<br />

The sound from equipment used and/or specific acts during athletic<br />

competition, such as hitting a baseball with an aluminum bat, carries beyond<br />

the playing field and can provide a nonobtrusive method to evaluate<br />

athletic performance—such as where on the bat the ball was hit. Standardized<br />

equipment guarantees repeatability, for example, every volleyball<br />

resonates at the same frequency. Each major sport can have unique noise<br />

interference which in some circumstances can be overwhelming, and the<br />

distance from the sound source can vary significantly during a game. Still,<br />

it will be shown that useful performance information can be obtained<br />

under realistic conditions for at least the following sports: volleyball, softball,<br />

baseball, golf, swimming and diving, soccer, and football.<br />

FRIDAY AFTERNOON, 1 DECEMBER <strong>2006</strong> WAIANAE ROOM, 1:30 TO 6:20 P.M.<br />

Session 4pPA<br />

Physical Acoustics and Biomedical UltrasoundÕBioresponse to Vibration: Sound Propagation in<br />

Inhomogeneous Media II<br />

Takahiko Otani, Cochair<br />

Doshisha Univ., Lab. of Ultrasonic Electronics, Kyotonabe-shi, Kyoto 610-0321, Japan<br />

James G. Miller, Cochair<br />

Washington Univ., Dept. of Physics, 1 Brookings Dr., St. Louis, MO 63130<br />

Invited Paper<br />

1:30<br />

4pPA1. Observables and prediction modeling in the presence of ultra-wideband heterogeneity. John J. McCoy �The Catholic<br />

Univ. of America, Washington, DC <strong>2006</strong>4�<br />

Underlying virtually all propagation and scattering models is an intuitive understanding; the acoustic field is observable using a<br />

device of a sufficiently small size to obtain a sufficiently dense set of discrete measurements. This assumes the field variation cuts off<br />

at an inner length scale larger than the device size, assuring that no information is lost to the inherent spatial averaging in any<br />

measurement. This understanding is faulty in the presence of environment heterogeneity observed on an extreme range of length<br />

scales. The reason is that all physical devices have finite accuracy, which limits their ability to capture variation on scales significantly<br />

larger than their size, in the presence of variation on intermediate scales. A more refined understanding of the ability to observe a field<br />

requires multiple devices, an unbounded hierarchy in the limit, to obtain multiple dense sets of discrete ‘‘observables.’’ This, then,<br />

suggests a different class of prediction models for environments with ultra-wideband heterogeneity, expressed in multiple sets of<br />

discrete variables, each set describing field variation in a limited subband. A framework for formulating these prediction models and<br />

their application to a scenario for which environment heterogeneity has no inner scale cutoff is presented.<br />

3280 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3280


1:50<br />

4pPA2. Parameters arising from the Burridge-Keller formulation for<br />

poroelastic media, especially for granular media and marine<br />

sediments. Allan D. Pierce, William M. Carey, and Paul E. Barbone<br />

�Boston Univ., Boston, MA 02215�<br />

It was previously shown �Pierce et al. �<strong>2006</strong>�� that the Burridge-Keller<br />

formulation �J. Acoust. Soc. Am. �1981�� rigorously justifies the lowfrequency<br />

version ��1956a�� of Biot’s equations. Implementation involves<br />

two microscale problems: �1� incompressible viscous flow driven in a<br />

highly irregular space with rigid walls by a uniformly and externally applied<br />

apparent pressure distribution and �2� elastostatic deformation of an<br />

intricate elastic web caused by the joint influence of a distributed constant<br />

body force and by uniform tractions �including an external pressure� on<br />

the web’s exposed surface. Microscale averages produce the Biot ‘‘constants.’’<br />

Theoretical devices of applied mechanics and mathematics yield<br />

estimates of these and related parameters. In particular, it is shown that<br />

Wood’s equation is a reasonable first approximation for the sound speed in<br />

sediments in the low-frequency limit. The formulation also yields an estimation<br />

for the sound speed in the high-frequency limit, when the viscous<br />

boundary layers become thin. The well-known result that the attenuation<br />

varies as f 1/2 in the high-frequency limit also results without the necessity<br />

of Biot’s heuristic patching theory. Various heuristic approximations due<br />

to Gassmann, to Geertsma and Smit, and to Stoll and Bryant are analytically<br />

and numerically assessed.<br />

2:05<br />

4pPA3. Sound propagation in the mixtures of liquid and solid<br />

aggregate; similarities at micro- and nanoscales.. Hasson M. Tavossi<br />

�Dept. of Physical and Environ. Sci., Mesa State College, 1100 North Ave.,<br />

Grand Junction, CO 81504�<br />

Sound propagation phenomena in certain liquid and solid aggregate<br />

mixtures, at micrometer scales, in some cases resemble the wave propagation<br />

behaviors of materials observed at nanometer and atomic scales.<br />

For example, it can be shown that the sound wave dispersion, attenuation,<br />

and cutoff-frequency effects depend on the same structural parameters as<br />

those observed at nano or atomic levels and are similar at both scales.<br />

Therefore, to investigate theoretical models of wave and matter interactions<br />

it is more convenient to use, as experimental tools, the readily analyzable<br />

models of wave propagation, in mixtures of solid and liquid, constructed<br />

at micrometer scales. Theoretical findings on sound propagation<br />

in the mixtures of liquid and solid particles at micrometer scales will be<br />

discussed. These results show the resemblance to the behavior of acoustic<br />

phonons, the lattice thermal vibrations of crystalline structures, at radically<br />

different scales. Experimental data on wave dispersion, attenuation, bandpass,<br />

and cutoff frequency effects, measured for sound propagation, in<br />

inhomogeneous materials consisting of mixtures of solid and liquid will be<br />

presented, showing the similarities of wave propagation behaviors at<br />

micro- and nanoscales.<br />

2:20<br />

4pPA4. Nonlinear surface waves in soil. Evgenia A. Zabolotskaya,<br />

Yurii A. Ilinskii, and Mark F. Hamilton �Appl. Res. Labs., Univ. of Texas,<br />

P.O. Box 8029, Austin, TX 78713-8029�<br />

Nonlinear effects in surface waves propagating in soil are investigated<br />

theoretically. Analytic solutions are derived for the second harmonics and<br />

difference frequency waves generated by a bifrequency primary wave<br />

propagating at moderate amplitude. The soil is modeled as an isotropic<br />

solid. As such, its elastic properties are described by five elastic constants,<br />

two at second order in the strain energy density �the shear and bulk<br />

moduli� and three at third order. Nonlinear propagation of the surface<br />

waves is based on a theory developed previously �Zabolotskaya, J. Acoust.<br />

Soc. Am. 91, 2569–2575 �1992��. Elements of the nonlinearity matrix<br />

associated with the interacting spectral components are expressed in terms<br />

of the five elastic constants. It was found convenient to express the nonlinearity<br />

matrix for soil as a function of a nonlinearity parameter corre-<br />

Contributed Papers<br />

sponding to B/A for liquids, particularly for saturated soils exhibiting<br />

liquidlike properties. This nonlinearity parameter can vary by several orders<br />

of magnitude. For soils with shear wave speeds less than 20% of the<br />

compressional wave speeds, the nonlinearity of surfaces waves is found to<br />

be independent of the third-order elastic constants and dependent only on<br />

the shear modulus. �Work supported by ONR.�<br />

2:35<br />

4pPA5. The measurement of the hysteretic nonlinearity parameter of<br />

a field soil by the phase shift method: A long-term survey. Zhiqu Lu<br />

�Natl. Ctr. for Physical Acoust., The Univ. of Mississippi, Univ., MS<br />

38677�<br />

Soil properties significantly affect the performance of the acoustic<br />

landmine detection. The climate and seasonal changes cause the variations<br />

of soil properties and smear landmine signature over time. On the other<br />

hand, soil is a complicated granular material that exhibits strong nonlinear<br />

acoustic behaviors. To understand the weather and seasonal effects on<br />

nonlinear acoustic behaviors of soils, a phase shift method is used to<br />

measure the hysteretic nonlinearity parameter of a field soil. The technique<br />

is based on measuring the variation of phase difference between two transducers,<br />

i.e., the phase shift, induced by changing sound level. The hysteretic<br />

nonlinear parameter can be extracted from the measured phase shift as<br />

a function of sound level or dynamic strain. In a long-term survey, the<br />

nonlinearity parameter, sound speed, and environmental conditions such<br />

as temperature, moisture, soil water potential, and rainfall precipitation are<br />

measured. It is found that the nonlinearity parameter is much more sensitive<br />

than sound speed to the climate change. Soil water potential is the<br />

predominant factor that affects the nonlinearity parameter and sound speed<br />

of the shallow field soil.<br />

2:50<br />

4pPA6. Nonlinear acoustic landmine detection: Comparison of soil<br />

nonlinearity with soil-interface nonlinearity. Murray S. Korman,<br />

Kathleen E. Pauls, Sean A. Genis �Dept. of Phys., U. S. Naval Acad.,<br />

Annapolis, MD 21402�, and James M. Sabatier �Univ. of Mississippi,<br />

Univ., MS 38677�<br />

To model the soil-top plate interface in nonlinear acoustic landmine<br />

detection, the soil-plate oscillator was developed �J. Acoust. Soc. Am. 116,<br />

3354–3369 �2004��. A Lexan plate �2.39 mm thick, 18.5 cm diameter� is<br />

clamped at an inside diameter of 11.8 cm between two metal flanges. Dry<br />

sifted masonry sand �2-cm layer� is placed over the plate. Turning curves<br />

experiments are performed by driving a loudspeaker �located over the<br />

sand� by a swept sinusoid. The acceleration versus frequency is measured<br />

near resonance on a swept spectrum analyzer using an accelerometer centered<br />

on the surface. The corresponding backbone curve exhibits a linear<br />

decrease in resonant frequency f versus increasing acceleration, where a<br />

��a o( f � f o)/f o . Define a nonlinear parameter ��1/a o . When the elastic<br />

plate is replaced by a ‘‘rigid’’ plate, � decreased from 0.128 to 0.070<br />

(s 2 /m�, while f o increased from 191 to 466 Hz. When a cylindrical drumlike<br />

mine simulant �rigid walls, thin acrylic top-plate� wasburied2cm<br />

deep in a concrete sand box, ‘‘on the mine’’ results yielded ��0.30 (s 2 /m�<br />

with f o�147 Hz, while ‘‘off the mine,’’ ��0.03 (s 2 /m� at f o�147 Hz.<br />

�Work supported by ONR.�<br />

3:05<br />

4pPA7. Causality conditions and signal propagation in bubbly water.<br />

Gregory J. Orris, Dalcio K. Dacol, and Michael Nicholas �Naval Res.<br />

Lab., 4555 Overlook Ave. SW, Washington, DC 20375�<br />

Acoustic propagation through subsurface bubble clouds in the ocean<br />

can exhibit signal travel times with enormous variations depending on the<br />

acoustic signal frequency, bubble size distribution, and void fraction. Recent<br />

theories have predicted large variations in phase speeds and attenuation<br />

that have been largely validated for frequencies well below and well<br />

above bubble resonance. However, great care must be exercised when<br />

3281 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3281<br />

4p FRI. PM


theoretically treating signal propagation at frequencies near resonance,<br />

termed the ‘‘Anomalous Absorption Regime’’ nearly 100 years ago in the<br />

pioneering work of Sommerfeld �A. Sommerfeld, Physik. Z. 8, 841<br />

�1975�� while investigating aspects of electromagnetic causality. We will<br />

discuss similarities between acoustic propagation in bubbly media and<br />

electromagnetic propagation in the presence of a conducting medium. We<br />

show that the signal travel time is dependent on the behavior of the dispersion<br />

formula in the complex frequency plane and place limits on the<br />

range of validity of these formulas, leading naturally to the necessary<br />

modifications to the current dispersion formulas to bring them into compliance<br />

with causality. Finally, we present theoretical results for the velocity<br />

of signals for a representative environment of experimental work carried<br />

out at the Naval Research Laboratory. �Work supported by the ONR.�<br />

3:20<br />

4pPA8. Measurements of the attenuation and sound speed in bubbly<br />

salt water. Gregory J. Orris, Dalcio K. Dacol, and Michael Nicholas<br />

�Naval Res. Lab., 4555 Overlook Ave. SW, Washington, DC 20375�<br />

Bubble clouds were injected from below the surface of a 144-cubicmeter<br />

water tank, wherein hydrophones were placed at varying distances<br />

from an acoustic source. Measurements were made over a wide range of<br />

frequencies to verify and validate the theoretical predictions of the relevant<br />

dispersion formula. This work was undertaken under a variety of<br />

conditions by varying the relevant environmental parameters: void fraction,<br />

temperature, and salinity. Void fractions were varied from roughly<br />

0.02% to 0.1%. Temperatures ranged from 9 °C to 18 °C, and the salinity<br />

was varied from zero to approximately 10% of typical oceanic values.<br />

Particular attention was paid to tracking the phase of the transmitted signal<br />

as the frequency progressed toward resonance starting from 100 kHz. This<br />

yielded phase-speed measurements in an essentially free-field environment<br />

using a modified version of phase spectral analysis. Time-of-flight measurements<br />

gave signal velocities, while the received energy yielded the<br />

attenuation. Results are compared to theoretical calculations, leading to<br />

the conclusion that current theoretical dispersion formula requires modification.<br />

�This work supported by the ONR.�<br />

3:35<br />

4pPA9. Efficient computation of 3-D acoustical scattering from<br />

multiple arbitrarily shaped objects using the boundary element<br />

methodÕfast multipole method „BEMÕFMM…. Nail A. Gumerov and<br />

Ramani Duraiswami �Perceptual Interfaces and Reality Lab., Inst. for<br />

Adv. Comput. Studies, Univ. of Maryland, College Park, MD 20742�<br />

Many applications require computation of acoustic fields in systems<br />

consisting of a large number of scatterers, which may have complex shape.<br />

Despite the boundary element method being a well-known technique for<br />

solution of the boundary value problems for the Helmholtz equation, its<br />

capabilities are usually limited by the memory and speed of computers,<br />

and conventional methods can be applicable to relatively small �up to<br />

order of 10 000 boundary elements� problems. We developed and implemented<br />

an efficient computational technique, based on an iterative solver<br />

employing generalized minimal residual method in combination with<br />

matrix-vector multiplication speeded up with the fast multipole method.<br />

We demonstrate that this technique has O(N) memory and computational<br />

complexity and enables solution of problems with thousands of scatterers<br />

�millions of boundary elements� on a desktop PC. The test problems<br />

solved are of moderate frequency �up to kD�150, where k is the wavenumber<br />

and D is the size of the computational domain�. Solution of large<br />

scale scattering problems was tested by comparison with the FMM-based<br />

T-matrix method applicable for simple shape objects reported earlier<br />

�Gumerov and Duraiswami, J. Acoust. Soc. Am., 117�4�, 1744–1761<br />

�2005��, visualization, and physical interpretation of the results.<br />

3:50<br />

4pPA10. Fast acoustic integral-equation solver for complex<br />

inhomogeneous media. Elizabeth Bleszynski, Marek Bleszynski, and<br />

Thomas Jaroszewicz �Monopole Res., 739 Calle Sequoia, Thousand<br />

Oaks, CA 91360, elizabeth@monopoleresearch.com�<br />

We describe elements and representative applications of an integralequation<br />

solver for large-scale computations in acoustic wave propagation<br />

problems. In the solver construction we used elements of our previously<br />

developed fast integral-equation solver for Maxwell’s equations. In comparison<br />

with the conventional integral equation approach �method of moments�,<br />

our solver achieves significant reduction of execution time and<br />

memory through the FFT-based matrix compression. One particular aspect<br />

of the solver we discuss, pertinent to its high efficiency and accuracy, is an<br />

efficient treatment of problems associated with subwavelength discretization.<br />

We illustrate the approach and its application on the example of a<br />

numerical simulation of acoustic wave propagation through the human<br />

head. �Work was supported by a grant from AFOSR.�<br />

4:05<br />

4pPA11. Models for acoustic scattering in high contrast media. Max<br />

Denis, Charles Thompson, and Kavitha Chandra �Univ. Massachusetts<br />

Lowell, One University Ave., Lowell, MA 01854�<br />

In this work a numerical method for evaluating backscatter from a<br />

three-dimensional medium having high acoustic contrast is presented. The<br />

solution is sought in terms of a perturbation expansion in the contrast<br />

amplitude. It is shown that limitations of the regular perturbation expansion<br />

can be overcome by recasting the perturbation sequence as a rational<br />

fraction using Padé approximants. The resulting solution allows for an<br />

accurate representation of the pressure and allows for the poles in the<br />

frequency response to be modeled. The determination of the pulse-echo<br />

response for a high-contrast medium is discussed and presented.<br />

4:20<br />

4pPA12. Multiple scattering and visco-thermal effects. Aroune<br />

Duclos, Denis Lafarge, Vincent Pagneux �Laboratoire d’Acoustique de<br />

l’Universite du Maine, Ave. Olivier Messiaen, 72085 Le Mans, France�,<br />

and Andrea Cortis �Lawrence Berkeley Natl. Lab., Berkeley, CA 94720�<br />

For modeling sound propagation in a rigid-framed fluid-saturated porous<br />

material it is customary to use frequency-dependent density and compressibility<br />

functions. These functions, which describe ‘‘temporal’’ dispersion<br />

effects due to inertial/viscous and thermal effects, can be computed<br />

by FEM in simple geometries and give complete information about the<br />

long-wavelength properties of the medium. When the wavelength is reduced,<br />

new effects due to scattering must be considered. To study this, we<br />

consider solving the sound propagation problem in a 2-D ‘‘phononic crystal’’<br />

made of an infinite square lattice of solid cylinders embedded in a<br />

fluid. An exact multiple-scattering solution is first developed for an ideal<br />

saturating fluid and then generalized to the case of visco-thermal fluid, by<br />

using the concept of visco-thermal admittances. The condition to use this<br />

concept is that the viscous and thermal penetration depths are small compared<br />

to the cylinder radius. We validate our results in the longwavelength<br />

regime by direct comparisons with FEM data �A. Cortis, ‘‘Dynamic<br />

parameters of porous media,’’ Ph.D. dissertation �Delft U.P., Delft,<br />

�2002��. When frequency increases, differences appear between the longwavelength<br />

solution and the exact multiple-scattering solution, which<br />

could be interpreted in terms of ‘‘spatial’’ dispersion effects.<br />

4:35<br />

4pPA13. Effective parameters of periodic and random distributions of<br />

rigid cylinders in air. Daniel Torrent and José Sánchez-Dehesa �Wave<br />

Phenomena Group, Nanophotonics Technol. Ctr., Polytechnic Univ. of<br />

Valencia, C/Camino de vera s/n., E-46022 Valencia, Spain�<br />

The scattering of sound by finite-size clusters consisting of twodimensional<br />

distributions �periodic and random� of rigid cylinders in air is<br />

theoretically studied in the low-frequency limit �homogenization�. Analyti-<br />

3282 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3282


cal expressions for the effective density and sound speed obtained in the<br />

framework of multiple scattering will be reported. For the case of circularshaped<br />

cluster, we have theoretically analyzed the homogenization as a<br />

function of the filling fraction, the type of arrangement of the cylinders in<br />

the cluster �hexagonal and square lattice�, and the number of cylinders in<br />

the cluster. When the number of cylinders in the cluster is small we found<br />

that for certain ‘‘magic numbers’’ their effective parameters �sound speed<br />

and density� are the same as those of the corresponding infinite array.<br />

�Work supported by MEC of Spain.�<br />

4:50<br />

4pPA14. The application of k-space acoustic propagation models to<br />

biomedical photoacoustics. Benjamin T. Cox �Dept. of Med. Phys. and<br />

Bioengineering, Univ. College London, Gower St., London, WC1E 6BT,<br />

UK, bencox@mpb.ucl.ac.uk�, Simon. R. Arridge, and Paul C. Beard<br />

�Univ. College London, Gower St., London, WC1E 6BT, UK�<br />

k�space models for broadband acoustic pulse propagation differ from<br />

pseudo�spectral time domain �PSTD� models in their treatment of the<br />

time step. By replacing a finite�difference scheme with a propagator,<br />

exact for homogeneous media, larger time steps can be taken without loss<br />

of accuracy or stability and without introducing dispersion. Three<br />

k�space models for modeling photoacoustically generated �PA� pulses are<br />

described here. A very simple, exact, model of PA propagation in a homogeneous<br />

fluid is used to introduce the k�space propagator, and two models<br />

of propagation in heterogeneous media, originally designed for modeling<br />

scattering in soft tissue, are adapted for use in photoacoustics �Mast<br />

et al., IEEE Trans. UFFC 48, 341�354 �2001�; Tabei et al., J. Acoust.<br />

Soc. Am. 111,53�63 �2002��. Our motivation for describing these models<br />

comes from biomedical PA imaging, in which one of the current limitations<br />

is the assumption that soft tissue has a uniform sound speed. Efficient,<br />

accurate, and simple�to�encode forward models such as these are<br />

very useful for studying the effects of the heterogeneities encountered in<br />

practice. They may also be useful in designing PA imaging schemes that<br />

can account for acoustic heterogeneities. �This work was funded by the<br />

EPSRC, UK�<br />

5:05<br />

4pPA15. Emergence of the acoustic Green’s function from thermal<br />

noise. Oleg A. Godin �CIRES, Univ. of Colorado and NOAA, Earth<br />

System Res. Lab., 325 Broadway, Boulder, CO 80305,<br />

oleg.godin@noaa.gov�<br />

Recently proposed applications of noise cross-correlation measurements<br />

to passive remote sensing range from ultrasonics and acoustic<br />

oceanography to helioseismology and geophysics, at wave frequencies<br />

that differ by more than ten orders of magnitude. At the heart of these<br />

applications is the possibility to retrieve an estimate of a deterministic<br />

Green’s function from long-range correlations of diffuse noise fields. Apparently,<br />

S. M. Rytov �A Theory of Electrical Fluctuations and Thermal<br />

Radiation �USSR Academy of Sciences, Moscow, 1953�� was the first to<br />

establish theoretically a simple relation between the Green’s function and<br />

the two-point correlation function of fluctuations of wave fields generated<br />

by random sources. He used reciprocity considerations to analyze fluctuations<br />

of electromagnetic fields. In this paper, an acoustic counterpart of<br />

Rytov’s approach is applied to derive exact and asymptotic relations between<br />

respective acoustic Green’s functions and cross-correlation of thermal<br />

noise in inhomogeneous fluid, solid, and fluid-solid media. Parameters<br />

of the media are assumed to be time independent, but can be arbitrary<br />

functions of spatial coordinates. Theoretical results obtained are compared<br />

to those previously reported in the literature.<br />

5:20<br />

4pPA16. Simulation of elastic wave scattering in living tissue at the<br />

cellular level. Timothy E. Doyle and Keith H. Warnick �Dept. of Phys.,<br />

Utah State Univ., 4415 Old Main Hill, Logan, UT 84322-4415,<br />

timdoyle@cc.usu.edu�<br />

Elastic wave scattering in biological tissue has been simulated at the<br />

cellular level by incorporating a first-order approximation of the cell structure<br />

and multiple scattering between cells. The cells were modeled with a<br />

concentric spherical shell-core structure embedded in a medium, with the<br />

core, shell, and medium representing the cell nucleus, the cell cytoplasm,<br />

and the extracellular matrix, respectively. Using vector multipole expansions<br />

and boundary conditions, scattering solutions were derived for a<br />

single cell with either solid or fluid properties for each of the cell components.<br />

Multiple scattering between cells was simulated using addition<br />

theorems to translate the multipole fields from cell to cell and using an<br />

iterative process to refine the scattering solutions. Backscattering simulations<br />

of single cells demonstrated that changes in the nuclear diameter had<br />

the greatest effect on the frequency spectra as compared to changes in cell<br />

size, density, and shear modulus. Wave field images and spectra from<br />

clusters of up to several hundred cells were also simulated, and they exhibited<br />

phenomena such as wave field enhancement at the cell membrane<br />

and nuclear envelope due to the scattering processes. Relevant applications<br />

for these models include ultrasonic tissue characterization and<br />

ultrasound-mediated gene transfection and drug delivery.<br />

5:35<br />

4pPA17. Acoustic analog of electronic Bloch oscillations and Zener<br />

tunneling. José Sánchez-Dehesa, Helios Sanchis-Alepuz, Yu. A.<br />

Kosevich, and Daniel Torrent �Wave Phenomena Group, Polytechnic<br />

Univ. of Valencia, C/Camino de Vera s.n., E-46022 Valencia, Spain�<br />

The observation of Bloch oscillations in sound propagation through a<br />

multilayer of two different fluidlike components is predicted. In order to<br />

obtain the equivalent to the acoustic analog of a Wannier-Stark ladder �E.<br />

E. Mendez et al., Phys. Rev. Lett. 60, 2426–2429 �1988��, a set of cavities<br />

with increasing thickness is employed. Bloch oscillations were theoretically<br />

predicted as time-resolved oscillations in transmission in direct analogy<br />

to electronic Bloch oscillations in semiconductor superlattices �J.<br />

Feldmann et al., Phys. Rev. B 46, R7252–R7255 �1992��. Finally, an experimental<br />

setup is proposed to observe the phenomenon by using arrays<br />

of cylindrical rods in air, which acoustically behaves as a fluidlike system<br />

with effective sound velocity and density �D. Torrent et al., Phys. Rev.<br />

Lett. 96, 204302 �<strong>2006</strong>��. For the proposed system, Bloch oscillations and<br />

Zener tunneling are confirmed by using multiple scattering simulations.<br />

�Work supported by MEC of Spain.�<br />

5:50<br />

4pPA18. Comparison of time reversal acoustic and prefiltering<br />

methods of focusing of tone burst signals. Bok Kyoung Choi �Korea<br />

Ocean Res. and Development Inst., Sangrok-gu, 426-744, Korea�,<br />

Alexander Sutin, and Armen Sarvazyan �Artann Labs., Inc., West<br />

Trenton, NJ 08618�<br />

The concept of time reversal acoustics �TRA� provides an elegant<br />

possibility of both temporal and spatial concentration of acoustic energy in<br />

highly inhomogeneous media. TRA-based focusing is typically used for<br />

generation of short acoustic pulses, however, in some medical and industrial<br />

applications, longer pulses are required. TRA focusing of longer signals<br />

leads to an increase of side lobes in temporal and spatial domains.<br />

Another method for focusing, known as prefiltering, is based on measurements<br />

of the impulse response, which relates the signal at the TRA transmitter<br />

to that at the focusing point. After evaluating the impulse response,<br />

the excitation signal may be calculated to generate the desired waveform<br />

in the focus point. This method allows signal generation with any desired<br />

form including long tone-burst signals. Experiments on comparison TRA<br />

and prefiltering methods of ultrasound focusing were conducted in the<br />

frequency band of 200–1000 kHz. In the experiments, focused acoustic<br />

pulses with various forms and duration were generated: triangular, rectan-<br />

3283 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3283<br />

4p FRI. PM


gular, and amplitude-modulated tone burst signals. The prefiltering modes<br />

provide better temporal compression of the focused signal, and the signal<br />

energy outside the main pulse in the prefiltering mode was shown to be<br />

much lower than that in standard TRA focusing.<br />

6:05<br />

4pPA19. Modeling quasi-one-dimensional sound propagation in ducts<br />

having two propagation media using a cross-sectional averaging<br />

theory. Donald Bliss and Lisa Burton �Dept. of Mech. Eng. and Mater.<br />

Sci., Duke Univ., Durham, NC 27708�<br />

Sound propagation of quasi-one-dimensional waves through a uniform<br />

duct partially filled with porous material has been studied theoretically and<br />

experimentally. The porous material makes the effective propagation wave<br />

number in the duct complex. A fairly simple theory based on cross-<br />

sectional averaging is derived and tested and found to work extremely<br />

well up to fairly high frequency. Interestingly, the basic theory depends<br />

only on the ratio of cross-sectional areas and the properties of the individual<br />

propagation media, but not on the specific configuration of material<br />

in a cross section. A higher order correction is developed to achieve excellent<br />

accuracy to very high frequency. This correction includes a coefficient<br />

that does depend on the specific cross-sectional configuration. Results<br />

are compared to exact solutions for layered and annular<br />

configurations, and also to experimental measurements with open cell<br />

foam as the porous material. An interesting application is to use measured<br />

wave numbers to predict the complex effective density and sound speed of<br />

porous media samples partially filling the duct. Other applications include<br />

fairly simple improved predictions of the behavior of sound in ducts lined<br />

with, or partially filled with, bulk reacting absorbing material.<br />

FRIDAY AFTERNOON, 1 DECEMBER <strong>2006</strong> WAIALUA ROOM, 1:30 TO 4:20 P.M.<br />

Session 4pPP<br />

Psychological and Physiological Acoustics: Auditory Physiology<br />

G. Christopher Stecker, Cochair<br />

Univ. of Washington, Dept. of Speech and Hearing Science, 1417 NE 42nd St., Seattle, WA 98105<br />

Shigeto Furukawa, Cochair<br />

NTT Communication Science Labs., Human and Information Science Lab., 3-1 Morinosato-wakamiya, Atsugi-shi,<br />

Kanagawa-ken 243-0198, Japan<br />

1:35<br />

4pPP1. Transmission of bone-conducted sound measured acoustically<br />

and psycho-acoustically. Sabine Reinfeldt �Signals & Systems,<br />

Chalmers Univ. of Technol., SE-412 96 Goteborg, Sweden�, Stefan<br />

Stenfelt �Linkoping Univ., SE-581 83 Linkoping, Sweden�, and Bo<br />

Hakansson �Chalmers Univ. of Technol., SE-412 96 Goteborg, Sweden�<br />

The transcranial transmission is important in the bone-conducted �BC�<br />

audiometry where the BC hearing thresholds depend on the stimulation<br />

position. It is also important for fitting of BC hearing aids; the transcranial<br />

transmission determines the amount of the sound that reaches the contralateral<br />

cochlea. Previous reported transcranial transmission results seem<br />

to depend on the method used. Here, a comparison between the transcranial<br />

transmission measured with BC hearing thresholds and ECSP is performed<br />

for both open and occluded ear canal. A BC transducer provided<br />

stimulation at both mastoids and the forehead. The ECSP was measured<br />

with a probe microphone and the BC hearing thresholds were obtained<br />

while masking the nontest ear. The transcranial transmission was determined<br />

as the BC hearing threshold or the ECSP for contralateral stimulation<br />

relative ipsilateral stimulation. The transmission from the forehead<br />

was calculated in a similar way. The transcranial transmission was similar<br />

for BC hearing thresholds and ECSP above 800 Hz; this indicates that the<br />

ECSP can be used as an estimator of the relative hearing perception by<br />

BC. The transcranial transmission results are also similar to vibration measurements<br />

of the cochleae made in earlier studies. Hence, vibration measurements<br />

of the cochleae can also estimate relative BC hearing.<br />

Chair’s Introduction—1:30<br />

Contributed Papers<br />

1:50<br />

4pPP2. Customization of head-related transfer functions using<br />

principal components analysis in the time domain. Ki H. Shin and<br />

Youngjin Park �Dept. of Mech. Eng., Korea Adv. Inst. of Sci. and<br />

Technol. �KAIST�, Sci.e Town, Daejeon, 305-701, Republic of Korea�<br />

Pinna responses were separated from HRIRs �head-related impulse<br />

responses� of 45 subjects in the CIPIC HRTF �head-related transfer function�<br />

database and modeled as linear combinations of five basic temporal<br />

shapes �basis functions� by PCA �principal components analysis� accounting<br />

for more than 90% of the variance in the original pinna responses per<br />

each selected elevation angle in the median plane. By adjusting the weight<br />

of each basis function computed for a specific height to replace the pinna<br />

response in the KEMAR HRIR at the same height with the resulting pinna<br />

response and listening to the filtered stimuli over headphones, four subjects<br />

were able to create a set of median HRIRs that outperformed the<br />

KEMAR HRIRs in producing elevation effects in the median plane. Since<br />

the monoaural spectral features due to the pinna are strongly dependent on<br />

elevation instead of azimuth, similar elevation effects could also be generated<br />

at different azimuthal positions simply by inserting the customized<br />

median pinna responses into the KEMAR HRIRs at other azimuths and<br />

varying the ITD �interaural time difference� according to the direction as<br />

well as the size of the subject’s own head.<br />

3284 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3284


2:05<br />

4pPP3. An electrophysiological measure of basilar membrane<br />

nonlinearity in humans. Christopher J. Plack �Dept. of Psych.,<br />

Lancaster Univ., Lancaster, LA1 4YF, England� and Ananthanarayan<br />

Krishnan �Purdue Univ., West Lafayette, IN 47906�<br />

A behavioral measure of the basilar membrane response can be obtained<br />

by comparing the growth in forward masking for maskers at, and<br />

well below, the signal frequency. Since the off-frequency masker is assumed<br />

to be processed linearly at the signal place, the difference in masking<br />

growth with level is thought to reflect the compressive response to the<br />

on-frequency masker. The present experiment used an electrophysiological<br />

analog of this technique, based on measurements of the latency of wave V<br />

of the auditory brainstem response elicited by a 4-kHz, 4-ms pure tone,<br />

presented at 65 dB SPL. Responses were obtained in quiet and in the<br />

presence of either an on-frequency or an off-frequency �1.8 kHz� puretone<br />

forward masker. Wave V latency increased with masker level, although<br />

the increase was greater for the off-frequency masker than for the<br />

on-frequency masker, consistent with a more compressive response to the<br />

latter. Response functions generated from the data showed the characteristic<br />

shape, with a nearly linear response at lower levels, and 5:1 compression<br />

at higher levels. However, the breakpoint between the linear region<br />

and the compressive region was at about 60 dB SPL, higher than expected<br />

on the basis of previous physiological and psychophysical measures.<br />

2:20<br />

4pPP4. Possible involvement of the spiral limbus in chinchilla<br />

cochlear mechanics. William S. Rhode �Dept. of Physiol., 1300<br />

University Ave., Madison, WI 53706, rhode@physiology.wisc.edu�<br />

Differences between cochlear mechanical tuning curves and those of<br />

auditory nerve fibers �ANFs� exist. In particular, mechanical transfer functions<br />

exhibit a high-frequency plateau; ANFs frequency threshold curves<br />

�FTCs� do not. ANF-FTCs may have a low-frequency slope due to a<br />

velocity forcing function operating on inner hair cells at low frequencies.<br />

Neither basilar membrane velocity nor displacement adequately explain<br />

the entire ANF tuning curve. A displacement sensitive interferometer was<br />

used to study basilar membrane and spiral limbus mechanics in the 6-kHz<br />

region of the chinchilla cochlea. The spiral limbus vibrates at the same<br />

phase as the basilar membrane nearly up to the location’s characteristic<br />

frequency. In the plateau region, the limbus appears to vibrate 0 to 20 dB<br />

less than the basilar membrane. The basilar membrane/limbus amplitude<br />

transfer function has a low-frequency slope of �3 dB/oct at low frequencies<br />

and is �10 dB lower than the basilar membrane amplitude at 1 kHz.<br />

It appears that spiral limbus vibration may contribute to the excitation of<br />

the cilia of the inner hair cells. �Work supported by NIDCD grant R01<br />

DC001910.�<br />

2:35<br />

4pPP5. The effects of antioxidants on cochlear mechanics. Barbara<br />

Acker-Mills, Melinda Hill, and Angeline Ebuen �U.S. Army Aeromedical<br />

Res. Lab., P.O. Box 620577, Fort Rucker, AL 36362-0577,<br />

barbara.acker@us.army.mil�<br />

Several studies have evaluated the effect of N-acetylcysteine �NAC�<br />

on temporary threshold shifts �TTSs� in humans. Work at USAARL found<br />

that NAC did not reduce TTSs compared to a placebo, but suppressed<br />

otoacoustic emissions �OAEs� more than a placebo, indicating that NAC<br />

may reduce outer hair cell activity. Kramer et al. �JAAA, 17�4�, �<strong>2006</strong>��<br />

found similar results, where NAC did not affect thresholds in people who<br />

had TTSs from exposure to loud music. However, OAEs also did not<br />

differ between NAC and placebo. Toppilla et al. �XXII Barany Society<br />

Meeting �2002�� measured thresholds and balance in people exposed to<br />

loud music and found that while NAC did not affect TTS, it reduced<br />

noise-induced balance impairment. The current study administered NAC<br />

and measured cochlear microphonics, compound action potentials, and<br />

OAEs to evaluate cochlear function. The vestibular myogenic potential<br />

was measured to assess the effect of NAC on the saccule. The results<br />

provide a comprehensive analysis of the effect of NAC on the auditory<br />

system and one component of the vestibular system. �Work supported by<br />

the U.S. Army ILIR program. The work is that of the authors and is not<br />

necessarily endorsed by the U.S. Army or the Department of Defense.�<br />

2:50–3:05 Break<br />

3:05<br />

4pPP6. Time characteristics of distortion product otoacoustic<br />

emissions recovery function after moderate sound exposure. Miguel<br />

Angel Aranda de Toro, Rodrigo Ordoñez, and Dorte Hammersho”i �Dept.<br />

of Acoust., Aalborg Univ., Fredrik Bajers Vej 7 B5, DK-9220, Aalborg,<br />

Denmark, maat@acoustics.aau.dk�<br />

Exposure to sound of moderate level temporarily attenuates the amplitude<br />

of distortion product otoacoustic emissions �DPOAEs�. These<br />

changes are similar to the changes observed in absolute hearing thresholds<br />

after similar sound exposures. To be able to assess changes over time<br />

across a broad frequency range, a detailed model of the recovery time<br />

characteristics is necessary. In the present study, the methodological aspects<br />

needed in order to monitor changes in DPOAEs from human subjects<br />

measured with high time resolution are presented. The issues treated<br />

are �1� time resolution of the measurements, �2� number of frequency<br />

points required, and �3� effects in fine structures, are they affected with the<br />

exposure? �Work supported by the Danish Research Council for Technology<br />

and Production.�<br />

3:20<br />

4pPP7. Probability characteristics of neural coincidence detectors in<br />

the brainstem. Ram Krips and Miriam Furst �Dept. of Elec. Eng.<br />

Systems, Faculty of Eng., Tel Aviv Univ., Tel Aviv 69978, Israel,<br />

mira@eng.tau.ac.il�<br />

Auditory neural activity in the periphery is usually described as nonhomogeneous<br />

Poisson process �NHPP�, characterized as either EE or EI,<br />

which operates as a coincidence detector. In order to apply a general<br />

probabilistic analysis of those brainstem nuclei activity, the stochastic<br />

properties of the axons that exit the EE and EI nuclei is essential. An<br />

analytical analysis of the probability characteristics of the output of an EE<br />

nucleus �EEout� will be presented. Assuming that an EE nucleus receives<br />

inputs from two neurons, each behaves as an NHPP with instantaneous<br />

rates �1 and �2, and an output spike is generated if both spike at a coincidence<br />

window (�) which is relatively small �this matches biological<br />

findings�. Then EEout is also a NHPP with instantaneous rate r(t)<br />

t<br />

t<br />

�� 1(t)� t���<br />

2(�) d��� 2(t)� t���<br />

1(�) d�. A similar derivation was applied<br />

for an EI nucleus. We found also that the output activity is NHPP for<br />

a relatively small coincidence window (�). The obtained IR is r(t)<br />

t<br />

�� e(t)�1�� t���<br />

i(�) d��, where �E and �I are the excitatory and inhibitory<br />

input IRs. On the other hand, for larger �, the output activity is not a<br />

Poisson process. Those derivations enable theoretical analysis and performance<br />

evaluation of large neural networks, which perform binaural tasks<br />

as ITD and ILD.<br />

3:35<br />

4pPP8. Musical expertise and concurrent sound segregation.<br />

Benjamin Rich Zendel and Claude Alain �Rotman Res. Inst., Baycrest<br />

Ctr. & Dept. of Psych., Univ. of Toronto, Toronto, ON M6A 2E1, Canada�<br />

There is growing evidence suggesting that musical training can improve<br />

performance in various auditory perceptual tasks. These improvements<br />

can be paralleled by changes in scalp recorded auditory evoked<br />

potentials �AEPs�. The present study examined whether musical training<br />

modulates the ability to segregate concurrent auditory objects using behavioral<br />

measures and AEPs. Expert musicians and nonmusicians were<br />

presented with complex sounds comprised of six harmonics �220, 440, 660<br />

Hz, etc.�. The third harmonic was either tuned or mistuned by 1%–16% of<br />

its original value. Mistuning a component of a harmonic complex results<br />

in the precept of a second auditory object. Stimuli were presented passively<br />

�no response� and actively �participants responded by indicating if<br />

they heard one sound or two sounds�. Behaviorally both musicians and<br />

3285 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3285<br />

4p FRI. PM


nonmusicians perceived a second auditory object at similar levels of mistuning.<br />

In both groups, complex sounds generated N1 and P2 waves at<br />

fronto-central scalp regions. The perception of concurrent auditory objects<br />

was paralleled by an increased negativity around 150 ms post-stimulus<br />

onset. This increased negativity is referred to as object-related negativity<br />

�ORN�. Small differences between musicians and nonmusicians were<br />

noted in the ORN. The implication of these results will be discussed in<br />

terms of current auditory scene analysis theory.<br />

3:50<br />

4pPP9. Study of auditory temporal processing of language and music<br />

in stroke patients. Li Hsieh and Tamara Baubie �Dept. of Commun.<br />

Sci. and Disord., Wayne State Univ., 207 Rackham Hall, 60 Farnsworth,<br />

Detroit, MI 48202�<br />

This study focused on functional lateralization of temporal processing<br />

of language and music for nine stroke patients and nine normal controls.<br />

Subjects were asked to discriminate short versus long sounds in ABX<br />

tasks, and then to reproduce sequences of short and long sounds they<br />

heard. The reproduction tasks consisted of sequences of 3, 4, and 5<br />

sounds. The linguistic stimuli were nonsense CVC syllables, and the musical<br />

stimuli were computer-generated musical notes with the timbres of<br />

French horn and trombone. Both linguistic and music stimuli were controlled<br />

for frequency and intensity, and varied only for duration �i.e., 500<br />

and 750 ms�. Our findings are consistent with previous studies; the left<br />

hemisphere specializes in linguistics, while the right hemisphere in music.<br />

Moreover, both hemispheres appeared to work closely together in processing<br />

temporal information. Both left- and right-hemisphere-damaged patients<br />

performed worse than controls. Most subjects performed better with<br />

music than language. Patients with left posterior lesions performed worse<br />

compared to patients with left or right anterior lesions, particularly with<br />

linguistic stimuli. Patients with right anterior lesions not only involved<br />

with temporal processing in music, but also in linguistic stimuli. Our study<br />

provided additional information regarding transient temporal processing in<br />

language and music.<br />

4:05<br />

4pPP10. Human bioacoustic biology: Acoustically anomalous vocal<br />

patterns used to detect biometric expressions relating to structural<br />

integrity and states of health. Sharry Edwards �Sound Health Res.<br />

Inst., P.O. Box 267, Sauber Res. Ctr., Albany, OH 45710�<br />

Computerized analyses of acoustically anomalous vocal patterns are<br />

being used as biomarkers for predictive, prediagnostic, and efficient management<br />

of individual biological form and function. To date, biometrically<br />

distinct vocal data have resulted in outcomes that would be considered<br />

improbable by contemporary medical standards. For instance, independent<br />

EMG conclusions confirmed the regeneration of nerve tissue for a multiple<br />

sclerosis patient who used acoustic bioinformation to guide his primary<br />

therapy. Another study monitored the amounts of synthetic labor hormones<br />

present during induced labor. False labor costs amount to millions of dollars<br />

each year in insurance and hospital resources. The use of noninvasive,<br />

possibly remote, vocal profiling could ameliorate such costs. Anomalous<br />

vocal acoustics are being investigated by many health-related organizations<br />

including Pfizer Pharmaceuticals and the Institute of Automatic Control<br />

Engineering in Taiwan. Complementary research studying molecular<br />

frequencies of cellular chemistry is being done by James Gimjewski,<br />

Ph.D., UCLA, Department of Chemistry and Biochemistry. Known as<br />

BioAcoustic Biology, this research modality has the potential to advance<br />

current health care standards for biological function, disease processes,<br />

and metabolism. Organizations such as the Acoustical Society of America<br />

are considering standards for technically defining human bioacoustics.<br />

This paper would expand that language.<br />

FRIDAY AFTERNOON, 1 DECEMBER <strong>2006</strong> HONOLULU ROOM, 1:00 TO 3:20 P.M.<br />

Session 4pSAa<br />

Structural Acoustics and Vibration: Vehicle Interior Noise and Vibration<br />

Courtney B. Burroughs, Cochair<br />

Noise Control Engineering, Inc. 1241 Smithfield St., State College, PA 16801<br />

Hiroaki Morimura, Cochair<br />

Japan Science and Technology Agency, Hosei Univ., Dept. of Mechanical Engineering, 3-7-2 Kajino-cho, Kogamei-city,<br />

Tokyo, Japan<br />

Invited Papers<br />

1:00<br />

4pSAa1. Analytical model for flow-excited interior cavity resonance and its application to the Stratospheric Observatory for<br />

Infrared Astronomy „SOFIA…. Jerry H. Ginsberg �G. W. Woodruff School of Mech. Eng., Georgia Inst. of Technol., Atlanta, GA<br />

30332-0405�<br />

The Stratospheric Observatory for Infrared Astronomy �SOFIA� is a joint effort between NASA and the German Space Agency<br />

that has installed a 20 000-kg telescope in a modified 747-SP. The modifications entailed constructing bulkheads, one of which is used<br />

to provide the active mount for the telescope, and a door that rotates to open as much as one-quarter of the fuselage circumference to<br />

the atmosphere. This configuration represents a large compartment that can exhibit acoustic resonances at low frequencies. Concern<br />

arose that a Rossiter mode, which is an aerodynamic resonance in which vortices shed from the leading edge of a gap form a coherent<br />

standing pattern at certain speeds, would create a strong acoustic source for acoustic and structural modes, whose frequencies might<br />

coincide. A model consisting of a two-dimensional hard-walled waveguide having a Rossiter mode source and an elastic plate at one<br />

end was analyzed in order to understand these issues. Unlike conventional analyses of interior cabin noise, in which vibrating walls<br />

are the acoustic source, the elastic plate represents a compliant boundary that couples with the acoustic modes. The results lead to<br />

some general insights to the the possibility of a ‘‘triple resonance,’’ as well as the role of structural compliance for cavities that are<br />

excited by turbulent external flows.<br />

3286 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3286


1:20<br />

4pSAa2. Energy finite energy analysis for shipboard noise. Raymond Fischer, Leo Boroditsky �Noise Control Eng. Inc., 799<br />

Middlesex Tnpk, Billerica, MA 01821�, Layton Gilfroy �DRDC-Atlantic�, and David Brennan �Martec Ltd.�<br />

Machinery-induced habitability noise is difficult to model efficiently and accurately. The potential of energy finite-element analysis<br />

�EFEA� is compared to other prediction tools such as statistical energy analysis �SEA�. This paper will explore the benefits and costs<br />

of EFEA with respect to SEA for acoustic modeling. The focus will be on issues relating to structural modeling for EFEA purposes.<br />

EFEA techniques will be evaluated to see if they possess the capabilities of verified SEA approaches for predicting habitability and<br />

radiated noise, where it is necessary to account for the impact of diverse marine constructions and sources such as the lack of<br />

machinery source information with respect to force or moment inputs or the finite impedance of machinery foundations. The effort<br />

proposed herein will provide the necessary engineering to research and identify salient features of EFEA that are potentially applicable<br />

for the detailed analysis of the acoustic environment and response of surface ships to various excitation sources. The paper will also<br />

address the pros and cons of SEA versus energy-finite element analysis �EFEA� methods used to predict the habitability noise of<br />

surface ship platforms. �This work is supported by an Office of Navy Research contract.�<br />

1:40<br />

4pSAa3. Spectral-based multicoordinate substructuring model for vehicle noise, vibration, and harshness refinement. Teik C.<br />

Lim �Mech., Industrial and Nuclear Eng., 598 Rhodes Hall, Box 210072, Univ. of Cincinnati, Cincinnati, OH 45221,<br />

teik.lim@uc.edu�<br />

The success of vehicle NVH �noise, vibration, and harshness� refinement often depends on the ability to identify and understand<br />

noise and vibration transmission paths within the mid-frequency range, i.e., 200–1000 Hz, throughout the assembled structure. Due to<br />

the complexity of the dynamics in this frequency range, most modal or finite element-based methods do not possess the fidelity<br />

needed. To address this gap, a multicoordinate substructuring theory applying measured structural-acoustic and vibration spectra is<br />

applied. Three forms of substructuring formulation, namely the nondiagonal, block-diagonal, and purely diagonal coupling cases, are<br />

developed. The performances of these approaches are studied numerically, and the net effects of these coupling formulations on the<br />

predicted joint and free substructure dynamic characteristics, and system response, are determined. Conditions for applying the<br />

simpler coupler that can simplify the testing process and overcome computational deficiencies are also derived. When the measured<br />

data is noise contaminated, the singular value decomposition �SVD� algorithm is found to be quite helpful. Using an actual vehicle,<br />

a comprehensive analysis of the measured and predicted vehicle system responses is performed. The results are employed to develop<br />

an understanding of the primary controlling factors and transfer paths and to cascade system requirements to the substructure level.<br />

2:00<br />

4pSAa4. Practical application of digital simulation with physical test in vehicle virtual. Toshiro Abe �ESTECH Corp., 89-1<br />

Yamashita-cho, Naka-ku, Yokohama-shi, Kanagawa-ken, Japan 231-0023, toshiro.abe@estech.co.jp�<br />

In the current vehicle design and development program, the Virtual Product Development process �hereafter, VPD process� is the<br />

innovation for automotive industry, which improves product quality and shortens time to market. In general, valid CAE applications<br />

are the key component of the VPD process as well as physical tests being indispensable to create valid simulation technologies. This<br />

presentation explains how physical-test-based CAE leverages the VPD process. In particular, the presentation is based on how<br />

physical testing supports the VPD process and why the ESTECH philosophy is that ‘‘The essence of CAE lies in its synergy with<br />

Testing;’’ a philosophy that differentiates the company from the competition. To demonstrate these activities, case studies based on<br />

automotive dynamic and real time simulation will be presented. In the case studies, vehicle body NVH and brake noise analysis will<br />

be used to show the interaction between physical test and computer simulation. Finally, practical application of the VPD process in<br />

one of the leading Japanese automotive companies will be shown, where the effectiveness of the front loading in the actual vehicle<br />

development program and the actual deployment of VPD process to the Functional Digital Vehicle project in the powertrain design are<br />

to be presented.<br />

2:20<br />

4pSAa5. Active noise control simulations using minimization of<br />

energy density in a mock helicopter cabin. Jared Thomas, Stephan P.<br />

Lovstedt, Jonathan Blotter, Scott D. Sommerfeldt �Brigham Young Univ.,<br />

Provo, UT 84602�, Norman S. Serrano, and Allan Egelston �Silver State<br />

Helicopters, Provo, UT 84601�<br />

Helicopter cabin noise is dominated by low-frequency tonal noise,<br />

making it an ideal candidate for active noise control. Previous work in<br />

active control of cabin noise suggests an energy density approach to be a<br />

good solution �B. Faber and S.D. Sommerfeldt, Global Control in a Mock<br />

Tractor Cabin Using Energy Density, Proc. ACTIVE 04, Sept. 2004.�<br />

Simulations for active noise control using energy density minimization<br />

have been made using recorded data from a Robinson R44 helicopter.<br />

Initial computer models show substantial noise reductions in excess of 6<br />

dB at the error sensor are possible. Performance results for computer<br />

models and simulations in a mock cab for different control arrangements<br />

will be compared.<br />

Contributed Papers<br />

2:35<br />

4pSAa6. Modeling airborne interior noise in full vehicles using<br />

statistical energy analysis. Arnaud Charpentier and Phil Shorter �ESI,<br />

12555 High Bluff Dr., Ste. 250, San Diego, CA 92130,<br />

arnaud.charpentier@esi-group-na.com�<br />

SEA is particularly well suited for predicting airborne noise in vehicles.<br />

The acoustic sources found in such environment are typically spatially<br />

distributed around the vehicle and can be well represented with SEA<br />

diffuse acoustic loads. Multiple transmission paths contribute to interior<br />

noise levels including �1� mass law transmission through trimmed panels,<br />

�2� resonant radiation from vibrating structures, and �3� flanking paths<br />

through gaskets, seals, and holes. All these transmission mechanisms may<br />

be modeled using SEA techniques. Finally, interior trim �including carpet,<br />

headliner, seats� is a key contributor to the acoustic performance of modern<br />

vehicles. The vehicle sound package has a significant impact on both<br />

the strength of the transmissions paths into the vehicle as well as the<br />

acoustic absorption in the cabin. Both these effects can be accounted for<br />

with SEA through detailed models of the trim. SEA models of full vehicles<br />

3287 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3287<br />

4p FRI. PM


are usually validated against experimental results at both component and<br />

system levels. The models can then be confidently used to �a� rank key<br />

design parameters governing interior levels and �b� quickly evaluate the<br />

impact of potential design changes. Example vehicle models and correlation<br />

results are presented here.<br />

2:50<br />

4pSAa7. Radiation from vibrating panels at high frequency including<br />

an inquiry into the role of edges and drive points. Donald Bliss<br />

�Mech. Eng. and Mater. Sci., Duke Univ., Durham, NC 27708�<br />

In the high frequency limit, a vibrating finite panel is shown to have<br />

broadband power and directivity characteristics that can expressed analytically<br />

by a limited set of parameters. Two-dimensional problems with subsonic<br />

structural wave speed are considered. Three basic directivity patterns<br />

are identified, associated with right and left traveling waves and the correlation<br />

between them. The role of boundary conditions at the panel edge<br />

is illustrated, as are the effects of types of forcing. Overall, relatively<br />

simple broadband behaviors are revealed. The analytical characterization<br />

of the radiation is shown to be particularly straightforward in the high<br />

frequency broadband limit. Interestingly, the radiated mean-square pressures<br />

are independent of the panel length, indicating the radiation is associated<br />

with the edges and the drive point. However, the radiation patterns<br />

cannot be explained in terms of simple volumetric sources placed just at<br />

the edges and the drive point, showing that the often-stated idea of uncan-<br />

celed volumetric sources at these locations is not correct except under very<br />

restricted circumstances. A correct physical interpretation of the radiation<br />

is provided both in physical space and in terms of spatial Fourier transforms.<br />

3:05<br />

4pSAa8. Some effect of trim and body panels on the low-frequency<br />

interior noise in vehicles. Andrzej Pietrzyk �Volvo Car Corp., Noise<br />

and Vib. Ctr., 405 31 Gothenburg, Sweden�<br />

Structure borne noise dominates the interior noise in vehicles at low<br />

frequencies. One of the basic vinroacoustic characteristics of the trimmed<br />

body is the noise transfer function, i.e., the acoustic response at a selected<br />

position in the passenger compartment, e.g., driver’s ear due to a mechanical<br />

excitation at a selected body mount. Detailed CAE models based on<br />

the FE method are today available for calculating this characteristic at low<br />

frequencies, corresponding to the engine idling and road excitation. However,<br />

the accuracy of CAE predictions of interior noise is still considered<br />

insufficient for the so-called analytical sign off, i.e., zero-prototypes vision.<br />

The current paper describes some investigations into the contribution<br />

of individual body panels to the overall interior noise. Also, the effect of<br />

selected interior trim items on the area investigated. Relative errors of<br />

prediction at different trim levels and different frequency ranges are discussed.<br />

Both experimental and CAE results are provided. The aim is to<br />

better understand the way the interior noise in vehicles is created, and how<br />

it can be controlled.<br />

FRIDAY AFTERNOON, 1 DECEMBER <strong>2006</strong> HONOLULU ROOM, 3:45 TO 5:45 P.M.<br />

Session 4pSAb<br />

Structural Acoustics and Vibration: General Vibration and Measurement Technology<br />

Peter C. Herdic, Cochair<br />

Naval Research Lab., Physical Acoustics Branch, Code 7136, 4555 Overlook Ave., SW, Washington, DC 20375<br />

Naoya Kojima, Cochair<br />

Yamaguchi Univ., Dept. of Mechanical Engineering, 2-16-1 Tokiwadai, Ube, Yamaguchi 755-8611, Japan<br />

3:45<br />

4pSAb1. Direct computation of degenerate elastodynamic solution of<br />

elastic wave propagation in a thick plate. Jamal Ghorieshi �Eng.<br />

Dept., Wilkes Univ., Wilkes-Barre, PA 18766�<br />

The limiting form of elastodynamic solutions as frequency tends to<br />

zero leads to the elastostatic eigenvalue equations. However, this limiting<br />

procedure is not convenient. It is cumbersome when applied to the solutions<br />

obtained using Stokes potentials and, in the case of utilizing Lames<br />

potentials, it does not produce static solutions that are a function of position<br />

alone. In this paper it is shown that the exact solutions of elastostatic<br />

problems can, in general, be obtained in a straightforward manner by the<br />

use of harmonic potentials without recourse to any special limiting form of<br />

analysis. This method is applied to an infinite, elastic thick plate with<br />

traction-free parallel surfaces and the elastostatic eigenvalue equation. It is<br />

shown that the problem can be solved exactly in terms of harmonic functions,<br />

one of which is a scalar and the other one is a vector. It is noted that<br />

results are in agreement with the published solutions.<br />

Contributed Papers<br />

4:00<br />

4pSAb2. Wave propagation characteristics of an infinite fluid-loaded<br />

periodic plate. Abhijit Sarkar and Venkata R. Sonti �Indian Inst. of Sci.<br />

Mech. Eng., IISc., Bangalore-560012, India�<br />

A 1-D infinite periodic plate with simple supports placed along equidistant<br />

parallel lines is considered using the finite-element method. The<br />

plate is loaded with a finite-height fluid column covered on the top with a<br />

rigid plate. Results show a relation between the propagation constant of<br />

the fluid-loaded structure with its in vacuo counterpart. Since the acoustic<br />

medium is an additional wave carrier, the attenuation bands corresponding<br />

to the in vacuo structure turn out to be propagating. However, the presence<br />

of the fluid can also bring about attenuation regions within the in vacuo<br />

propagation bands. Primary propagation constants bring additional waves<br />

called space harmonics with them. Hence, a localized coincidence effect is<br />

seen where a particular harmonic falls below or above the acoustic wave<br />

number, leading to propagation or a mass loading effect. Occasionally, a<br />

complete attenuation band is created. This is verified by decomposing the<br />

3288 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3288


single span displacement profile into the space harmonics and also by<br />

computing the frequency response function �FRF� for a finite fluid-loaded<br />

periodic plate and observing the huge antiresonance dip in frequency in<br />

the exact same frequency band where an attenuation band was predicted<br />

for the infinite structure.<br />

4:15<br />

4pSAb3. The dynamic response of a plate subjected to various edge<br />

excitations. Baruch Karp �Faculty of Aerosp. Eng., Technion Israel Inst.<br />

of Technol., Haifa 32000, Israel�<br />

Plane strain response of a semi-infinite, elastic plate to harmonic edge<br />

excitation is investigated analytically. The exact solution to this problem is<br />

obtained as a series expansion of the Rayleigh-Lamb modes of the plate.<br />

The variation of energy partition among the propagating modes with frequency<br />

of the edge excitation was found for load and displacement �symmetrical�<br />

perturbations of uniform and cosine form. The biorthogonality<br />

relation was employed in deriving the relative amplitudes of each mode to<br />

the given perturbation. The emphasis here is on the sensitivity of the<br />

far-field response, represented by propagating waves, to the details of the<br />

excitation. Within the frequency range investigated it was found that the<br />

plate’s response is remarkably insensitive to whether the excitation is load<br />

or displacement type. The two types of edge excitation distributions, on<br />

the other hand, result in different patterns of energy partition above the<br />

first cutoff frequency with similar energy partition only within limited<br />

range of frequencies. The effect of the nature of the excitation on the<br />

dynamic response of the plate and a possible implication to dynamic<br />

equivalence is discussed.<br />

4:30<br />

4pSAb4. Measurement of structural intensity using patch near-field<br />

acoustical holography. Kenji Saijyou �Fifth Res. Ctr., TRDI, Japan<br />

Defense Agency, 3-13-1 Nagase, Yokosuka City, Kanagawa Prefecture,<br />

Japan, 239-0826�<br />

Measurement of power flow in a structure, called the structural intensity<br />

�SI�, is essential for vibration control and noise reduction. The nearfield<br />

acoustical holography �NAH�-based measurement method is suitable<br />

to analyze the interrelationship between SI and acoustic intensity �AI�<br />

because NAH-based methods provide measurements of SI and AI simultaneously.<br />

Use of NAH requires the measurement of a pressure field over<br />

a complete surface located exterior to the structure. Therefore, if the measurement<br />

aperture is smaller than the structure, reconstructed results from<br />

the pressure on the finite aperture are seriously contaminated. This finite<br />

aperture effect prevents implementation of this NAH-based method on an<br />

actual large-scale structure such as a ship. Patch NAH and regularization<br />

method for SI measurement are applied to overcome this difficulty. This<br />

method enables implementation of the NAH-based SI measurement<br />

method in a large-scale structure. The effectiveness of this method is demonstrated<br />

by experiment.<br />

4:45<br />

4pSAb5. Flexural component and extensional component of vibration<br />

energy in shell structure. Taito Ogushi, Manabu Yahara, Masato<br />

Mikami, and Naoya Kojima �Dept. of Mech. Eng., Yamaguchi Univ.,<br />

2-16-1 Tokiwadai, Ube, Yamaguchi 755-8611, Japan,<br />

j008ve@yamaguchi-u.ac.jp�<br />

In this research, the behavior of the flexural component and the extensional<br />

component of vibration intensity and their transmission in curved<br />

shells are presented. L-shaped shell model was employed as an analysis<br />

model of FEM. As FEM analysis methods, both the frequency response<br />

analysis and the transitional response analysis were employed. The flexural<br />

component and the extensional component of vibration intensity �VI�<br />

were calculated by the results of FEM analysis. In the flexural component<br />

of the VI, the vibration energy supplied in the flat part decreased at the<br />

boundary from the flat part to the curved part and VI vectors flew in<br />

circumferential direction in the curved part. In the extensional component<br />

of the VI, the vibration energy appeared at the boundary from the flat part<br />

to the curved part and most VI vectors flew parallel to the shell axis in the<br />

curved part. The total vibration energy of the flexural component and the<br />

extensional component was conserved. So, the vibration energy transformed<br />

to each other between the flexural component and the extensional<br />

component in L-shaped shell.<br />

5:00<br />

4pSAb6. SeismicÕacoustic detection of ground and air traffic for<br />

unattended ground sensor technology. Peter C. Herdic �Naval Res.<br />

Lab., Physical Acoust. Branch, Washington, DC 20375 and SFA Inc.,<br />

Crofton, MD�, Brian H. Houston, Phillip A. Frank, and Robert M. Baden<br />

�Naval Res. Lab., Washington, DC 20375�<br />

Human footfall and vehicle traffic create surface waves in soil media<br />

that can easily be detected by seismic sensors. Field measurement data<br />

have been acquired with a triaxial geophone at several experimental sites.<br />

The in-plane-surface wave components dominate the response and decay<br />

at a rate of approximately 1/R, where R is distance. This decay rate is due<br />

to the combined effect of spreading �1/sqrt(R)) and damping losses in the<br />

soil. Further, the detection range is dependent upon the level of environmental<br />

noise, soil compliance, moisture content, and topography. Human<br />

detection was achieved in rural environments at distances up to �30–40<br />

m, and vehicle detection was possible at much greater distances. Seismic<br />

signals due to aircraft are small when compared to the acoustic signature.<br />

Ground-based microphone measurements clearly show the blade passage<br />

frequency tones of propeller airplanes and the broader band signature of<br />

turbojet aircraft. Time- and frequency-domain signal-processing methods<br />

for the detection and identification will also be introduced. These experimental<br />

results will be discussed with particular emphasis placed on wave<br />

phenomenon, detection and identification algorithms, and the related physics.<br />

5:15<br />

4pSAb7. Modeling of acoustic and elastic wave phenomena using<br />

plane wave basis functions. Tomi Huttunen, Jari P. Kaipio �Dept. of<br />

Phys., Univ. of Kuopio, P.O. Box 1627, FI-70211 Kuopio, Finland�, and<br />

Peter Monk �Univ. of Delaware, Newark, DE 19716�<br />

When simulating acoustic, elastodynamic, or coupled fluid-solid vibration<br />

problems using standard finite element �FE� techniques, several elements<br />

per wavelength are needed to obtain a tolerable accuracy for engineering<br />

purposes. At high wave numbers, the requirement of dense meshes<br />

may lead to an overwhelmingly large computational burden, which significantly<br />

limits the feasibility of FE methods for the modeling of wave<br />

phenomena. A promising technique for reducing the computational complexity<br />

is to use plane wave basis functions opposed to the low-order<br />

polynomials that are used in conventional FE methods. A possible method<br />

for utilizing the plane wave basis is the ultra-weak variational formulation<br />

�UWVF�. The UWVF method can be used for acoustic Helmholtz problems,<br />

elastodynamic Navier problems, or fluid-solid systems characterized<br />

by a coupled Helmholtz-Navier model. A comparison of the UWVF technique<br />

with a low-order FE method shows reduced computational complexity<br />

and improved accuracy.<br />

3289 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3289<br />

4p FRI. PM


Contributed Poster Paper<br />

Poster paper 4pSAb8 will be on display from 1:00 p.m. to 5:45 p.m. The author will be at the poster from 5:30 p.m. to 5:45 p.m.<br />

4pSAb8. Application of vibration energy flow to evaluation of<br />

thickness. Akihiko Higashi �Dept. of Maritime Sci. and Technol., Japan<br />

Coast Guard Acad., 5-1 Wakaba-cho, Kure, Hiroshima, 737-8512, Japan�<br />

In this study, the possibility of the useful application of the vibration<br />

energy flow is investigated. The vibration energy flow means the propagation<br />

of the vibration energy of the flexural waves. The vibration energy<br />

flow is expressed by the structural intensity. Here, it is easy to input the<br />

flexural waves in the thin plates and beam elements. Then, large structures<br />

such as ships use many of these thin plates and beam elements. But the<br />

usual methods of the evaluation and the inspection of these large structures<br />

are inefficient. Then, we investigated the possibility of the evaluation of<br />

the changed thickness of the structure by using the vibration energy flow<br />

analysis. As the result of analysis, the structural intensity suddenly<br />

changes at the position of the changed thickness. The changed quantity of<br />

the structural intensity corresponds to the change quantity of the thickness.<br />

Then, the evaluation method of the thickness of the structure is proposed.<br />

As a result, it was found that the change of the structural intensity indicates<br />

the change of the thickness. And the evaluation of the change of<br />

thickness of beams could be possible by using the proposed method.<br />

FRIDAY AFTERNOON, 1 DECEMBER <strong>2006</strong> MOLOKAI ROOM, 1:00 TO 4:00 P.M.<br />

Session 4pSC<br />

Speech Communication: Variation in Production and Perception of Speech „Poster Session…<br />

Heriberto Avelino, Cochair<br />

Univ. of California—Berkeley, Dept. of Linguistics, 1203 Dwinelle Hall, Berkeley, CA 94720-2650<br />

Haruo Kubozono, Cochair<br />

Kobe Univ., Dept. of Linguistics, Faculty of Letters, Nada-ku, Kobe 657-8501, Japan<br />

Contributed Papers<br />

All posters will be on display from 1:00 p.m. to 4:00 p.m. To allow contributors an opportunity to see other posters, contributors of<br />

odd-numbered papers will be at their posters from 1:00 p.m. to 2:30 p.m. and contributors of even-numbered papers will be at their<br />

posters from 2:30 p.m. to 4:00 p.m.<br />

4pSC1. Cross-language perception of voice and affect. Christer Gobl,<br />

Irena Yanushevskaya, and Ailbhe N. Chasaide �Phonet. and Speech Lab.,<br />

School of Linguistic, Speech and Commun. Sci., Trinity College Dublin,<br />

Dublin 2, Ireland, yanushei@tcd.ie�<br />

The paper reports on a cross-language study of how voice quality and<br />

f 0 combine in the signaling of affect. Speakers of Irish-English and Japanese<br />

participated in perception tests. The stimuli consisted of a short utterance<br />

where f 0 and voice source parameters were varied using the LFmodel<br />

implementation of the KLSyn88a formant synthesizer, and were of<br />

three types: �1� VQ only involving voice quality variations and a neutral<br />

f 0 contour; �2� f 0 only, with different affect-related f 0 contours and<br />

modal voice; �3� VQ� f 0 stimuli, where the voice qualities of �1� combine<br />

with specific f 0 contours from �2�. Overall, stimuli involving voice quality<br />

variation were consistently associated with affect. In �2� only stimuli with<br />

high f 0 yielded high affective ratings. Striking differences emerge between<br />

the ratings obtained from the two language groups. The results<br />

show that not only were some affects consistently perceived by one language<br />

group and not the other, but also that specific voice qualities and<br />

pitch contours were associated with very different affects across the two<br />

groups. The results have important implications for expressive speech synthesis,<br />

indicating that language/culture-specific differences need to be considered.<br />

�This work is supported by the EU-funded Network of Excellence<br />

on Emotion, HUMAINE.�<br />

4pSC2. An articulatory study of coronal consonants in Arrernte.<br />

Marija Tabain �LaTrobe Univ., Melbourne, Australia�, Richard Beare<br />

�Monash Univ., Melbourne, Australia�, Catherine Best �Univ. of Western<br />

Sydney, Sydney, Australia�, and Louis Goldstein �Haskins Labs., CT�<br />

This paper presents electro-magnetic articulography �EMA� data on<br />

the four coronal stops of Arrernte, an Australian language. The stops are:<br />

the lamino-dental ‘‘th,’’ the apico-alveolar ‘‘t,’’ the apico-postalveolar �or<br />

‘‘retroflex’’� ‘‘rt,’’ and the lamino-palatal ‘‘ty.’’ Jaw, tongue tip �TT�, and<br />

tongue body �TB� data were collected for two female speakers of the<br />

language. Results for the first speaker show a fronted tongue position for<br />

the laminal consonants, with the TT reflecting a similar location for both<br />

the dental and the palatal. However, for the palatal, the TB position is<br />

much higher, whereas for the dental, the TB is very low. For the apical<br />

consonants, the TT is not as far forward, and the TB is not quite as high as<br />

for the lamino-palatal. For both TT and TB, apico-postalveolar is further<br />

back than apico-alveolar. For the second speaker, the TT sensor failed, but<br />

in line with the first speaker, the TB sensor showed a higher position for<br />

the palatal. The other stops were lower and more forward, with the postalveolar<br />

TB position higher than the laminal or alveolar stop position. For<br />

both speakers, the jaw position is lowest for the postalveolar. �Work supported<br />

by Australian Research Council and NIH: NIDCD.�<br />

3290 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3290


4pSC3. Symbolic phonetic features for pronunciation modeling.<br />

Rebecca A. Bates, a� , Mari Ostendorf �Dept. of Elec. Eng., Univ. of<br />

Washington, Box 352500, Seattle, WA 98195�, and Richard A. Wright<br />

�Univ. of Washington, Seattle, WA 98195�<br />

A significant source of variation in spontaneous speech is due to intraspeaker<br />

pronunciation changes, often realized as small feature changes,<br />

e.g., nasalized vowels or affricated stops, rather than full phone transformations.<br />

Previous computational modeling of pronunciation variation has<br />

typically involved transformations from one phone to another, partly because<br />

most speech processing systems use phone-based units. Here, a<br />

phonetic-feature-based prediction model is presented where phones are<br />

represented by a vector of symbolic features that can be on, off, unspecified,<br />

or unused. Feature interaction is examined using different groupings<br />

of possibly dependent features, and a hierarchical grouping with conditional<br />

dependencies led to the best results. Feature-based models are<br />

shown to be more efficient than phone-based models, in the sense of<br />

requiring fewer parameters to predict variation while giving smaller distance<br />

and perplexity values when comparing predictions to the handlabeled<br />

reference. A parsimonious model is better suited to incorporating<br />

new conditioning factors, and this work investigates high-level information<br />

sources, including both text �syntax, discourse� and prosody cues.<br />

Detailed results are under review with Speech Communication. �This research<br />

was supported in part by the NSF, Award No. IIS-9618926, an Intel<br />

Ph.D. Fellowship, and by a faculty improvement grant from Minnesota<br />

State University Mankato.� a� Currently at Minnesota State University,<br />

Mankato.<br />

4pSC4. Acoustic phonetic variability and auditory word recognition<br />

by dyslexic and nondyslexic children. Patricia Keating, Kuniko Nielsen<br />

�Phonet. Lab., Linguist., UCLA, Los Angeles, CA 90095-1543,<br />

keating@humnet.ucla.edu�, Frank Manis, and Jennifer Bruno �USC, Los<br />

Angeles, CA 90089�<br />

The hypothesis that dyslexia involves degraded phonological representations<br />

predicts impairments in behaviors that rely on these representations,<br />

such as auditory word recognition. Normal adult listeners recognize<br />

different pronunciations of a word as instances of the same lexical item,<br />

but more slowly and less accurately; dylexics should be even more impaired<br />

by acoustic phonetic variability. Children with and without dyslexia<br />

performed a word recognition task: on each trial, a child hears a target<br />

word, then eight probes �matching the target or not�, responding yes/no to<br />

each probe. On some trials, probes are spoken by multiple talkers who<br />

differ in age, sex, speech style, etc.; on some trials the match probes also<br />

differ from the target in final stop consonant allophone. Responses are<br />

scored for accuracy and speed. Research questions include: Do all children<br />

demonstrate less accurate/slower recognition of words spoken by multiple<br />

talkers versus by one talker? Do all children demonstrate less accurate/<br />

slower recognition of words spoken with different allophones? Do dyslexic<br />

children demonstrate less accurate/slower recognition than nondyslexic<br />

children and, if so, for all trials, only for multiple talker trials, and/or<br />

only for different allophone trials; for all dyslexic children, or only those<br />

with particular phonological impairments? �Work supported by NIH.�<br />

4pSC5. Intertalker differences in intelligibility of cochlear-implant<br />

simulated speech. Tessa Bent, Adam B. Buchwald, and David B. Pisoni<br />

�Indiana Univ., Dept. of Psychol. and Brain Sci., 1101 E. 10th St.,<br />

Bloomington, IN 47405, tbent@indiana.edu�<br />

Are the acoustic-phonetic factors that promote highly intelligible<br />

speech invariant across different listener populations? Two approaches<br />

have been taken to investigate intelligibility variation for a variety of<br />

listener populations including hearing-impaired listeners, second language<br />

learners, and listeners with cochlear implants: studies on how speaking<br />

style affects intelligibility and other research on how inherent differences<br />

among talkers influence intelligibility. Taking the second approach, we<br />

compared intertalker differences in intelligibility for normal-hearing listeners<br />

under cochlear implant �CI� simulation (n�120) and in quiet (n<br />

�200). Stimuli consisted of 20 native English talkers’ productions of 100<br />

sentences. These recordings were processed to simulate listening with an<br />

eight-channel CI. Both clear and CI-processed tokens were presented to<br />

listeners in a sentence transcription task. Results showed that the most<br />

intelligible talkers in quiet were not the most intelligible talkers under CI<br />

simulation. Furthermore, listeners demonstrated perceptual learning with<br />

the CI-simulated speech but showed little learning in the quiet. Some of<br />

the acoustic-phonetic properties that were correlated with intelligibility<br />

differed between the CI-simulated speech and the speech in the quiet.<br />

These results suggest that the intertalker variations that result in highly<br />

intelligible speech observed in earlier studies are dependent on listener<br />

characteristics. �Work supported by NIDCD.�<br />

4pSC6. The effect of phonological neighborhood density and word<br />

frequency on vowel production and perception in clear speech. Rajka<br />

Smiljanic, Josh Viau, and Ann Bradlow �Dept. of Linguist., Northwestern<br />

Univ., 2016 Sheridan Rd., Evanston, IL 60208�<br />

Previous research showed that phonological neighborhood density and<br />

word frequency influence word recognition �Luce and Pisoni, 1998� and<br />

vowel production �Wright, 2002; Munson and Solomon, 2004; Munson, to<br />

appear�, suggesting an interaction of lexical and phonetic factors in speech<br />

production and perception Here, we explore whether hyperarticulated,<br />

intelligibility-enhancing clear speech shows similar sensitivity to lexicallevel<br />

structure. Nine American English talkers �five females, four males�<br />

produced 40 monosyllabic easy �frequent words with few lexical neighbors�<br />

and hard �infrequent words with many lexical neighbors� words in<br />

conversational and clear speech. Twenty-four subjects participated in a<br />

word-in-noise listening test. Results revealed a large effect of style on<br />

intelligibility and vowel production: words were more intelligible and<br />

vowels were longer and more dispersed in clear compared to conversational<br />

speech. Moreover, the female talkers produced larger vowel spaces<br />

than male talkers in both speaking styles. Vowels in hard words were<br />

marginally more dispersed than vowels in easy words in both speaking<br />

styles. However, within both speaking styles, easy and hard words were<br />

equally intelligible and of approximately equal duration. These results<br />

showed that phonetic properties of vowels were enhanced equally in clear<br />

speech regardless of their lexical properties.<br />

4pSC7. Phoneme dependency of accuracy rates in familiar and<br />

unknown speaker identification. Kanae Amino, Takayuki Arai �Dept.<br />

of Elec. and Electron. Eng., Sophia Univ., 7-1 Kioi-cho, Chiyoda-ku,<br />

Tokyo, 102-8554 Japan, amino-k@sophia.ac.jp�, and Tsutomu Sugawara<br />

�Sophia Univ., Chiyoda-ku, Tokyo, 102-8554 Japan�<br />

For perceptual speaker identification, the identification accuracy depends<br />

on the speech contents presented to subjects. Our previous studies<br />

have shown that stimuli containing nasals are effective for identifying<br />

familiar speakers �Amino et al., Acoust. Sci. Tech. 27�4� �<strong>2006</strong>��. We have<br />

also presented the possibility that the interspeaker spectral distances reflect<br />

perceptual speaker similarities. In the present study, we conducted an experiment<br />

in which four unknown speakers were identified by 15 subjects.<br />

The stimuli were identical to those used in the previous study, in which ten<br />

speakers were identified by familiar listeners, although the speakers were<br />

fewer this time. Nine consonants in the CV structure were used as stimuli.<br />

The consonants were /d/, /t/, /z/, /s/, /r/, /j/, /m/, /n/, and /nj/; the vowel was<br />

restricted to /a/ for all CV syllables to simplify the experiment. The results<br />

showed that the nasals /n/ and /nj/ obtained higher scores. Tendencies in<br />

the differences among consonants were on the same order as those of the<br />

3291 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3291<br />

4p FRI. PM


previous experiment, but the average scores were lower than those for<br />

familiar listeners. �Work supported by Grant-in-Aid for JSPS Fellows 17-<br />

6901.�<br />

4pSC8. Speech style and stereotypical character in Japanese. Akiko<br />

Nakagawa �Grad. School of Cultural Studies and Human Sci., Kobe<br />

Univ., 1-2-1 Tsurukabuto, Nada-ku, Kobe 657-8501, Japan,<br />

akiko.nakagawa@atr.jp� and Hiroko Sawada �Kyoto Univ., Kyoto<br />

606-8501, Japan�<br />

This study shows that ‘‘stereotypical character’’ is necessary to understand<br />

Japanese speech communication in addition to existing conceptions<br />

such as emotion, communicative strategy, register, and so on. Stereotypical<br />

character is here defined as a complex entity, consisting of information<br />

about gender, age, social status, physical features, characteristics, and<br />

speech style. The necessity of stereotypical character was shown through<br />

an auditory experiment involving a total of 70 speech sounds comprised of<br />

15–19 short phrases �mean duration 1.4� selected from recordings of spontaneous<br />

speech of four adult female speakers of Japanese. Ten participants<br />

were asked to listen to these speech sounds randomly, and to classify them<br />

into four speakers. Each of the resulting auditory-perceptual categories<br />

was found to contain speech sounds from more than one speaker. Further<br />

analyses of these results suggested that the participants classified the<br />

speech sounds not according to invariant speaker characteristics but according<br />

to virtual stereotypical characters that are common in Japanese<br />

society. Therefore, such changeable speaker characteristics as ‘‘busybody’’<br />

‘‘thoughtful,’’ ‘‘high-handed,’’ and so on, can be elicited through speech<br />

sounds by Japanese speakers. �This work was partially supported by the<br />

Ministry of Education, Science, Sport, and Culture, Grant-in-Aid for Scientific<br />

Research �A�, 1620<strong>2006</strong>.�<br />

4pSC9. Perceived vocal age and its acoustic correlates. Hiroshi Kido<br />

�Dept. of Commun. Eng., Tohoku Inst. of Technol., Taihaku-ku, Sendai,<br />

Japan 989-8577, kidoh@tohtech.ac.jp� and Hideki Kasuya �Intl. Univ. of<br />

Health and Welfare, Otawara, Japan 324-8501�<br />

This study investigates relationships between perceived and chronological<br />

age of talkers and acoustic correlates of the perceived age. Most of<br />

the past studies were primarily concerned with the instability of the vocalfold<br />

vibration extracted from sustained vowels. This study focuses on the<br />

dynamic nature of sentence utterances. Talkers included 115 healthy men,<br />

aged 20–60 years, who read a short sentence in Japanese. Listeners consisted<br />

of 70 men and women, aged 20–40 years, who made direct estimations<br />

of age. The results showed a strong correlation (r�0.66) between<br />

the perceived and chronological age as well as the tendency toward overestimating<br />

the ages of younger talkers and underestimating those of older<br />

talkers, supporting past investigations �e.g., R. Huntley et al., J. Voice 1,<br />

49–52 �1987��. Acoustic parameters considered were median of the fundamental<br />

frequency (F0� contour, F0 range, declination of F0 contour,<br />

spectral tilt, median of the boundary frequencies above which irregularities<br />

dominate, and speaking rate. From both statistical graphical modeling<br />

and regression tree analysis, the speaking rate, F0 declination, and spectral<br />

tilt were found to be dominant acoustic correlates to the perceived age.<br />

�Work supported partly by a Grant-in-Aid for Scientific Research, JSPS<br />

�16300061�.�<br />

4pSC10. A cross-linguistic study of informational masking: English<br />

versus Chinese. Bruce A. Schneider, Liang Li, Meredyth Daneman<br />

�Dept. of Psych., Univ. of Toronto at Mississauga, Mississauga, ON, L5L<br />

1C6 Canada, bschneid@utm.utoronto.ca�, Xihong Wu, Zhigang Yang,<br />

Jing Chen, and Ying Huang �Peking Univ., Beijing, China 10087�<br />

The amount of release from informational masking in monolingual<br />

English �Toronto, Canada�, and Chinese �Beijing, China� listeners was<br />

measured using the paradigm developed by Freyman et al. �J. Acoust.<br />

Soc. Am. 106, 3578–3588�. Specifically, psychometric functions relating<br />

percent-correct word recognition to signal-to-noise ratio were determined<br />

under two conditions: �1� masker and target perceived as originating from<br />

the same position in space; �2� masker and target perceived as originating<br />

from different locations. The amount of release from masking due to spatial<br />

separation was the same for English and Chinese listeners when the<br />

masker was speech-spectrum noise or cross linguistic �Chinese speech<br />

masking English target sentences for English listeners or English speech<br />

masking Chinese target sentences for Chinese listeners�. However, there<br />

was a greater release from masking for same-language masking of English<br />

�English speech masking English target sentences� than for same-language<br />

masking of Chinese �Chinese speech masking Chinese target sentences�. It<br />

will be argued that the differences in same-language masking between<br />

English and Chinese listeners reflect structural differences between English<br />

and Mandarin Chinese. �Work supported by China NSF and CIHR.�<br />

4pSC11. Cross-linguistic differences in speech perception. Keith<br />

Johnson and Molly Babel �UC Berkeley, 1203 Dwinelle Hall, Berkeley,<br />

CA 94720-2650�<br />

This research explores language-specific perception of speech sounds.<br />

This paper discusses two experiments: experiment 1 is a speeded forcedchoice<br />

AX discrimination task and experiment 2 is a similarity rating task.<br />

Experiment 1 was intended to investigate the basic auditory perception of<br />

the listeners. It was predicted that listeners’ native languages would not<br />

influence responses in experiment 1. Experiment 2 asked subjects to rate<br />

the similarity between two tokens on a five-point equal interval scale; the<br />

purpose of this experiment was to explore listeners’ subjective impression<br />

of speech sounds. In experiment 2 it was predicted that listeners’ language<br />

would affect their responses. The same stimuli were used in both experiments.<br />

The stimuli consisted of vowel-fricative-vowel sequences produced<br />

by a trained phonetician. Six fricatives were used: /f, th, s, sh, x, h/. These<br />

fricatives were embedded in three vowel environments: /a_a/, /i_i/, and<br />

/u_u/. Tokens were presented to listeners over headphones with a 100-ms<br />

interval. Independent groups of 15 native Dutch and English listeners<br />

participated in each of the two experiments. Results suggest that listeners’<br />

language influenced responses in both experiments, albeit the result was<br />

larger in experiment 2. �Work supported by NIH.�<br />

4pSC12. Neural coding of perceptual interference at the preattentive<br />

level. Yang Zhang �Dept. of Speech-Lang.-Hearing Sci., Univ. of<br />

Minnesota, Minneapolis, MN 55455�, Patricia Kuhl, Toshiaki Imada<br />

�Univ. of Washington, Seattle, WA 98195�, Toshiaki Imada, and Masaki<br />

Kawakatsu �Tokyo Denki Univ., Inzai-shi, Chiba 270-1382, Japan�<br />

Language acquisition involves neural commitment to languagespecific<br />

auditory patterns, which may interfere with second language<br />

learning. This magnetoencephalography study tested whether perceptual<br />

interference could occur at the preattentive level. Auditory mismatch field<br />

�MMF� responses were recorded from ten American and ten Japanese<br />

adult subjects in the passive oddball paradigm. The subjects read selfchosen<br />

books and ignored the sounds. Three pairs of synthetic /ra-la/ syllables<br />

were used: one cross-category pair varied only in the third formant<br />

�F3�, and the other two within-category pairs varied only in the second<br />

formant �F2�. ANOVA results showed a main effect of acoustic dimension<br />

with significant interaction with subject groups �p�0.01). As reported<br />

earlier, American listeners showed larger but later MMF responses for the<br />

F3 change. By contrast, Japanese listeners showed larger and earlier<br />

MMFs than Americans for changes in F2. Moreover, Japanese listeners<br />

had larger and earlier MMF responses for the changes in F2 as against<br />

changes in F3, which was more prominent in the right hemisphere than in<br />

the left. These results provided further support for the hypothesis that<br />

language experience produces neural networks dedicated to the statistical<br />

properties of incoming speech experienced in infancy, which later interfere<br />

with second language acquisition.<br />

3292 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3292


4pSC13. Russian and Spanish listener’s perception of the English<br />

tenseÕlax vowel contrast: Contributions of native language allophony<br />

and individual experience. Maria V. Kondaurova �Program in<br />

Linguist., Purdue Univ., West Lafayette, IN 47907� and Alexander L.<br />

Francis �Purdue Univ., West Lafayette, IN 47906�<br />

We examined the influence of listeners native phonology on the perception<br />

of American English tense and lax front unrounded vowels ��i� and<br />

�(��. These vowels are distinguishable according to both spectral quality<br />

and duration. Nineteen Russian, 18 Spanish, and 16 American English<br />

listeners identified stimuli from a beat-bit continuum varying in nine spectral<br />

and nine duration steps. English listeners relied predominantly on<br />

spectral quality when identifying these vowels, but also showed some<br />

reliance on duration. Russian and Spanish speakers relied entirely on duration.<br />

Three additional tests examined listeners allophonic use of vowel<br />

duration in their native languages. Duration was found to be equally important<br />

for the perception of lexical stress for all three language groups.<br />

However, the use of duration as a cue to postvocalic consonant voicing<br />

differed due to phonotactic differences across the three languages. Group<br />

results suggest that non-native perception of the English tense/lax vowel<br />

contrast is governed by language-independent psychoacoustic factors<br />

and/or individual experience. Individual results show large variability<br />

within all three language groups, supporting the hypothesis that individual<br />

differences in perceptual sensitivity as well as the more frequently cited<br />

factors of second language education and experience play an important<br />

role in cross-language perception.<br />

4pSC14. An analysis of acoustic deviation manner in spontaneous<br />

speech. Norimichi Hosogai, Kanae Okita, Takuya Aida, and Shigeki<br />

Okawa �Chiba Inst. of Technol., 2-17-1 Tsudanuma, Narashino, Chiba<br />

275-0016, Japan�<br />

Natural speech typically contains various phenomena deviated from<br />

the formal mode such as read speech. It is well known that those paralinguistic<br />

phenomena have an important role to give the human emotions and<br />

the state of the speakers in speech communication. This study attempts to<br />

extract the deviation as an acoustic ‘‘vagueness,’’ defined by temporal and<br />

dynamical acoustic features of speech. Especially the change of the vagueness<br />

during a certain period of speech, such as a 10-minute presentation, is<br />

focused. As the acoustic features, it used �i� modulation spectrum and �ii�<br />

syllable speed, which may have relations to the speech clarity and the<br />

tempo. For the experiments, 70 academic presentation speech data in the<br />

Corpus of Spontaneous Japanese �CSJ� are used. As the experimental results,<br />

significant properties in the patterns of the modulation spectrum and<br />

the syllable speed are obtained as a difference of the beginning and the<br />

ending periods of the presentation. This result will contribute to a humanlike<br />

speech dialog system.<br />

4pSC15. Nondurational cues for durational contrast in Japanese.<br />

Kaori Idemaru �Dept. of East Asian Lang. and Lit., Univ. of Oregon, 1248<br />

Univ. of Oregon, Eugene, OR 97403� and Susan G Guion �Univ. of<br />

Oregon, 1290 Eugene, OR 97403�<br />

This study explores potential secondary cues to a durational contrast<br />

by examining short and long stop consonants in Japanese. Durational contrasts<br />

are susceptible to considerable variability in temporal dimensions<br />

caused by changes in speaking rate. In this study, the proposal is examined<br />

that multiple acoustic features covary with the stop length distinction and<br />

that these features may aid in accessing the percept intended by the<br />

speaker, even when the primary cue, closure duration, is unreliable. The<br />

results support the proposal, revealing the presence of multiple acoustic<br />

features covarying with the short versus long contrast. Not only are there<br />

durational correlates to this contrast—the preceding vowel is longer and<br />

the following vowel is shorter for geminates than singletons—but there<br />

also are nondurational features covarying with this contrast. Greater fundamental<br />

frequency and intensity drops are found from the preceding to<br />

the following vowel for the geminate than the singleton stops. These results<br />

suggest the possibility that systematic variation of these acoustic<br />

features is used in the perceptual categorization of the contrast in addition<br />

to the primary cue of closure duration. Moreover, the nondurational correlates<br />

are promising candidates for speech-rate resistant features.<br />

4pSC16. Different motor strategies for increasing speaking rate: Data<br />

and modeling. Majid Zandipour, Joseph Perkell, Mark Tiede, Frank<br />

Guenther �M.I.T., Res. Lab Electron., Speech Commun. Group, 50 Vassar<br />

St, Cambridge, MA 02139, majidz@speech.mit.edu�, Kiyoshi Honda<br />

�ATR Human Information Processing Res. Lab., Kyoto 619-0288, Japan�,<br />

and Emi Murano �Univ. Maryland Dental School, Baltimore, MD, 21209�<br />

Different motor strategies for increasing speaking rate: data and modeling<br />

EMG, kinematic and acoustic signals were recorded from two male<br />

subjects as they pronounced multiple repetitions of simple nonsense utterances.<br />

The resulting data indicate that the two subjects employed different<br />

motor strategies to increase speaking rate. When speaking faster, S1 significantly<br />

increased the size of the articulatory target region for his tongue<br />

movements, increased the speed of the tongue movements and the rate of<br />

EMG rise somewhat, while decreasing the movement duration significantly<br />

and movement distance slightly. In contrast, at the fast rate, S2 had<br />

the same size articulatory target region and rate of EMG rise as at the<br />

normal rate, but decreased the speed, distance, and duration of tongue<br />

movement slightly. Each subject had similar dispersions of acoustic targets<br />

in F1–F2 space at fast versus normal rates, but both shifted target centroids<br />

toward the center of the vowel space at the fast rate. Simulations<br />

with a biomechanical model of the vocal tract show how modulations of<br />

motor commands may account for such effects of speaking rate on EMG,<br />

kinematics, and acoustic outputs. �Work supported by NIDCD, NIH.�<br />

4pSC17. Effect of speaking rate on individual talker differences in<br />

voice-onset-time. Rachel M. Theodore, Joanne L. Miller, and David<br />

DeSteno �Dept. of Psych., 125 NI, Northeastern Univ., 360 Huntington<br />

Ave., Boston, MA, 02115-5000, r.theodore@neu.edu�<br />

Recent findings indicate that individual talkers systematically differ in<br />

phonetically relevant properties of speech. One such property is voiceonset-time<br />

�VOT� in word-initial voiceless stop consonants: at a given rate<br />

of speech, some talkers have longer VOTs than others. It is also known<br />

that for any given talker, VOT increases as speaking rate slows. We examined<br />

whether the pattern of individual differences in VOT holds across<br />

variation in rate. For example, if a given talker has relatively short VOTs<br />

at one rate, does that talker also have relatively short VOTs at a different<br />

rate? Numerous tokens of /ti/ were elicited from ten talkers across a range<br />

of rates using a magnitude-production procedure. VOT and syllable duration<br />

�a metric of speaking rate� were measured for each token. As expected,<br />

VOT increased as syllable duration increased �i.e., rate slowed� for<br />

each talker. However, the slopes as well as the intercepts of the functions<br />

relating VOT to syllable duration differed significantly across talkers. As a<br />

consequence, a talker with relatively short VOTs at one rate could have<br />

relatively long VOTs at another rate. Thus the pattern of individual talker<br />

differences in VOT is rate dependent. �Work supported by NIH/NIDCD.�<br />

4pSC18. Variation in vowel production. Joseph Perkell, Majid<br />

Zandipour, Satrajit Ghosh, Lucie Menard �Speech Commun. Group, Res.<br />

Lab. of Electron., Rm. 36-511, M.I.T., Cambridge, MA 02139�, Harlan<br />

Lane, Mark Tiede, and Frank Guenther �M.I.T., Cambridge, MA 02139�<br />

Acoustic and articulatory recordings were made of vowel productions<br />

by young adult speakers of American English—ten females and ten<br />

males—to investigate effects of speaker and speaking condition on measures<br />

of contrast and dispersion. The vowels in the words teat, tit, tet, tat,<br />

tot, and toot were embedded in two-syllable ‘‘compound words’’ consisting<br />

of two CVC syllables, in which each of the two syllables comprised a<br />

real word, the consonants were /p/, /t/ or /k/, the two adjoining consonants<br />

were always the same, the first syllable was unstressed and the second<br />

stressed. Variations of phonetic context and stress were used to induce<br />

3293 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3293<br />

4p FRI. PM


dispersion around each vowel centroid. The compound words were embedded<br />

in a carrier phrase and were spoken in normal, clear, and fast<br />

conditions. Initial analyses of F1 andF2 on 15 speakers have shown<br />

significant effects of speaker, speaking condition �and also vowel, stress,<br />

and context� on vowel contrast, and dispersion around means. Generally,<br />

dispersions increased and contrasts diminished going from clear to normal<br />

to fast conditions. Results of additional analyses will be reported. �Work<br />

supported by NIDCD, NIH.�<br />

4pSC19. Region, gender, and vowel quality: A word to the wise<br />

hearing scientist. Richard Wright �Dept. of Linguist., Univ. of<br />

Washington, Box 354340, Seattle, WA 98195-4340,<br />

rawright@u.washington.edu�, Stephanie Bor, and Pamela Souza �Univ. of<br />

Washington, Seattle, WA 98105�<br />

Sociophonetic research has established effects of regional accent and<br />

gender on spoken vowels. Many gender differences are due to sociolinguistic<br />

factors and thus vary by region. The implications for researchers<br />

and clinicians are important: gender variation must be controlled for according<br />

to the region of the listener and talker population. Moreover,<br />

speech perception stimuli used in research and in clinical applications<br />

have limited regional application. This poster illustrates these factors using<br />

the Pacific Northwest regional accent. The data, collected for a previous<br />

study on hearing aid processing, consist of three repetitions of eight vowels<br />

produced in real-word /h_d/ �or /_d/� contexts by six males and six<br />

females ranging in age from 19 to 60. Formants were measured using an<br />

LPC with an accompanying FFT and spectrogram for verification. The<br />

results revealed vowel-specific differences in the male and female speech<br />

over and above those typically associated with physiologic predictions,<br />

and different again from those observed in past studies from different<br />

regions. Taken as a whole, these data suggest that speech and hearing<br />

researchers should take care in selecting stimuli for general-use speech<br />

perception tests. �Work supported by NIDCD training grant �#DC00033�<br />

and NIH RO1 �1 RO1 DC006014�.�<br />

4pSC20. Acoustic characteristics of vowels in three regional dialects<br />

of American English. Ewa Jacewicz, Robert Allen Fox, Yolanda Holt<br />

�Speech Acoust. and Percept. Labs., Dept. of Speech and Hearing Sci., The<br />

Ohio State Univ., Columbus, OH 43210�, and Joseph Salmons �Univ. of<br />

Wisconsin—Madison, Madison, WI�<br />

Most of the comparative sociophonetic studies of regional dialect<br />

variation have focused on individual vowel differences across dialects as<br />

well as speaker variables. The present work seeks to define basic acoustic<br />

characteristics of entire vowel systems for three different regional variants<br />

of American English spoken in southeastern Wisconsin �affected by the<br />

Northern Cities Shift�, western North Carolina �affected by the Southern<br />

Vowel Shift�, and central Ohio �not considered to be affected currently by<br />

any vowel shift�. Three groups of speakers �men and women� aged 20–29<br />

years were recorded from each geographic area defined by two to three<br />

counties �creating a highly homogeneous set of speakers�. Acoustic measures<br />

for the set of 14 monophthongs and diphthongs in /h_d/ context<br />

included vowel space area for each speaker, global spectral rate of change<br />

for diphthongized vowels �defined over the first three formant slopes�, the<br />

amount of frequency change for F1 and F2 at two temporal points located<br />

close to vowel onset and offset �vector length�, and vowel duration. These<br />

measures will establish both systemic and vowel inherent characteristics<br />

across the three dialects, serving as a basis for future examination of<br />

conditioning factors on vowels in chain shifts. Dialectal differences will be<br />

discussed. �Work supported by NIH NIDCD R01 DC006871.�<br />

4pSC21. The rhythmic characterization of two varieties of<br />

Portuguese. Verna Stockmal, Emilia Alonso Marks, Audra Woods, and<br />

Z. S. Bond �Ohio Univ., Athens, OH 45701�<br />

As spoken in Europe, Portuguese is said to be stress-timed, while<br />

Brazilian Portuguese appears to display characteristics of both stress and<br />

syllable timing �P. A. Barbosa, D.E.L.T.A. 16, 369–402 �2000��. Weemployed<br />

the Ramus et al. metric, based on acoustic-phonetic measurements<br />

�Ramus et al., Cognition 73, 265–292 �1999��, to investigate the possibility<br />

of distinguishing between the two varieties of the language. Five native<br />

speakers of European Portuguese and five native speakers of Brazilian<br />

Portuguese recorded the same short prose passage taken from a magazine.<br />

The talkers were asked to read at a normal, comfortable rate. The reading<br />

time of the passage averaged 60 s, with considerable differences among<br />

the talkers. From the vocalic and consonantal intervals, the Ramus metrics,<br />

percent vocalic interval and standard deviation of consonantal and<br />

vocalic interval, were calculated. The five talkers of the two language<br />

varieties differed on the values of these parameters. The values of %V and<br />

SD-V showed overlap between the two talker groups, while the BP talkers<br />

tended to show lower values for SD-C. Apparently, the rhythmic characterization<br />

of the two varieties of the language is not clearly categorical, but<br />

rather ranges along a continuum.<br />

4pSC22. Indexical cues to talker sex and individual talker identity<br />

extracted from vowels produced in sentence-length utterances.<br />

Michael J. Berkowitz �Dept. of Psych., 301 Wilson Hall, Vanderbilt<br />

Univ., 111 21st Ave. South, Nashville, TN 37203,<br />

michael.j.berkowitz@vanderbilt.edu�, Jo-Anne Bachorowski �Vanderbilt<br />

Univ., Nashville, TN 37203�, and Michael J. Owren �Georgia State Univ.,<br />

Atlanta, GA 30302�<br />

The purpose of this study was to replicate and extend a previous study<br />

of indexical cuing �J.-A. Bachorowski and M. J. Owren, J. Acoust. Soc.<br />

Am. 106, 1054–1063 �1999�� by including more vowel sounds spoken in<br />

more diverse contexts. Specific combinations of acoustic parameters that<br />

should represent talker sex and individual talker identity were identified<br />

using predictions based on known sources of variation in vocal<br />

production-related anatomy. This study utilized 100 recordings of<br />

sentence-length utterances, produced by each of 43 male and 44 female<br />

undergraduates, as well as 22 stock-phrase recordings produced by these<br />

same participants. One of five vowel sounds �/æ, }, i,., u/� was isolated<br />

from each sentence and analyzed for F 0, F1, F2, F3, F4, vowel duration,<br />

jitter, shimmer, and harmonicity. Classification by talker sex was nearly<br />

perfect using a combination of cues related to both vocal-fold and vocaltract<br />

anatomy. The accuracy of classification by individual identity depended<br />

strongly on cues relating to vocal tract-variation within sex.<br />

4pSC23. Utterance-final position and projection of femininity in<br />

Japanese. Mie Hiramoto and Victoria Anderson �Dept. of Linguist.,<br />

Univ. of Hawaii, 1890 East-West Rd., Honolulu, HI 96822�<br />

Japanese female speakers frequent use of suprasegmental features such<br />

as higher pitch, longer duration, wider pitch range, and more instances of<br />

rising intonation vis-a-vis male speakers, is recognized as Japanese womens<br />

language �JWL� prosody. However, these features normally co-occur<br />

with gender-specific sentence-final particles �SFPs� like the strongly feminine<br />

‘‘kashira.’’ In this study, we examined the use of pitch and duration in<br />

utterances without SFPs, to determine whether JWL prosody is a function<br />

of SFPs or of utterance-final position. Eight male and eight female native<br />

Japanese speakers were instructed to read prepared sentences as though<br />

auditioning for a masculine theater role and then as though auditioning for<br />

a feminine role. Results indicate that utterance-final position is the projection<br />

point of JWL prosody even in the absence of SFPs. The data used for<br />

this study show high pitch, wide pitch range, long duration, and rising<br />

intonation at utterance-final positions when produced �by both men and<br />

women� in the feminine gender role. Conversely, in the masculine gender<br />

3294 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3294


ole, both men and women avoid the use of such prosodic features, and<br />

may even avoid using rising intonation in interrogative sentences, where<br />

the tonal grammar calls for it.<br />

4pSC24. Attitudinal correlate of final rise-fall intonation in Japanese.<br />

Toshiyuki Sadanobu �Kobe Univ., Tsurukabuto 1-2-1, Nada, Kobe,<br />

657-8501, Japan�<br />

Abrupt rise and subsequent fall intonation is common at the end of<br />

intonation units in Japanese, but its attitudinal correlate has not been fully<br />

elucidated yet. This intonation appears in the literature of the 1960’s as<br />

politicians’ way of speech, and nowadays not only politicians but many<br />

speakers including older generations often use it. However, this intonation<br />

is stigmatized as childish, and many people devalue it as an unintelligent<br />

way of speaking by young people. Where does this great gap between<br />

reality and image of this intonation come from? This presentation addresses<br />

this problem by focusing on natural conversation of Japanese daily<br />

life. The conclusions are as follows: �i� Rise-fall intonation often appears<br />

when the speaker talks about high-level knowledge, whereas it disappears<br />

when the speaker talks about their personal experience. �ii� Rise-fall intonation<br />

at the end of an intonation conveys the speaker’s being so occupied<br />

with speaking that intonation unit. The childish image comes from the<br />

speaker’s unawareness of their overall speech because of being occupied<br />

with local process. �Work supported by the Ministry of Education, Science,<br />

Sport, and Culture, Grant-in-Aid for Scientific Research �A�,<br />

1620<strong>2006</strong>, and by the Ministry of Internal Affairs and Communications,<br />

SCOPE 041307003.�<br />

4pSC25. Vowel devoicing in Japanese infant- and adult-directed<br />

speech. Laurel Fais, Janet Werker �Dept. of Psych., Univ. of BC, 2136<br />

West Mall, Vancouver, BC V6T 1Z4 Canada, jwlab@psych.ubc.ca�,<br />

Sachiyo Kajikawa, and Shigeaki Amano �NTT Commun. Sci. Labs.,<br />

Seika-cho, Soraku-gun, Kyoto 619-0237 Japan�<br />

It is well known that parents make systematic changes in the way they<br />

speak to infants; they use higher pitch overall, more pronounced pitch<br />

contours, more extreme point vowels, and simplified morphology and syntax<br />

�Andruski and Kuhl, 1996; Fernald et al., 1989�. Yet, they also preserve<br />

information crucial to the infants ability to acquire the phonology of<br />

the native language �e.g., phonemic length information, Werker et al.,<br />

<strong>2006</strong>�. The question examined in this paper is whether information other<br />

than phonemic segmental information is also preserved, namely, information<br />

concerning the phonological process of vowel devoicing. Devoicing<br />

of high vowels between voiceless consonants and word-finally after a<br />

voiceless consonant is a regular and well-attested phonological process in<br />

Japanese �Shibatani, 1990�. A corpus of speech by Japanese mothers addressed<br />

to their infants and addressed to another adult was examined, and<br />

the degree and frequency with which they apply vowel devoicing in each<br />

type of speech was analyzed. Rates of vowel devoicing in speech to adults<br />

and infants are compared, accommodations made to infants and to<br />

hearing-impaired children are discussed �Imaizumi et al., 1995�, and the<br />

implications of these accommodations for the acquisition of vowel devoicing<br />

by Japanese infants are explored.<br />

4pSC26. Language and gender differences in speech overlaps in<br />

conversation. Jiahong Yuan, Mark Liberman, and Christopher Cieri<br />

�Univ. of Pennsylvania, Philadelphia, PA 19104�<br />

Language and gender differences in speech overlaps in conversation<br />

were investigated, using the LDC CallHome telephone speech corpora of<br />

six languages: Arabic, English, German, Japanese, Mandarin, and Spanish.<br />

To automatically obtain the speech overlaps between two sides in a conversation,<br />

each side was segmented into pause and speaking segments, and<br />

the overlap segments during which both sides were speaking were time<br />

stamped. Two types of speech overlaps are distinguished: �1� One side<br />

takes over the turn before the other side finishes �turn-taking type�. �2�<br />

One side speaks in the middle of the other side’s turn �backchannel type�.<br />

It was found that Japanese conversations have more short �less than 500<br />

ms� turn-taking type of overlap segments than the other languages. The<br />

average number of such segments per 10 min of conversation for Japanese<br />

was 62.6 whereas the average numbers for the other languages ranged<br />

from 37.9 to 43.3. There were, however, no significant differences among<br />

the languages on the backchannel type of overlaps. Cross-linguistically,<br />

conversations between two females contained more speech overlaps �both<br />

types� than those between a male and a female or between two males, and<br />

there was no significant difference between the latter two.<br />

4pSC27. An acoustic study of laringeal contrast in three American<br />

Indian Languages. Heriberto Avelino �Dept. of Linguist., UC Berkeley,<br />

Berkeley, CA 94720-2650�<br />

A contrast between modal and nonmodal phonation is commonly<br />

found in American Indian languages. The use of laryngealized voice has<br />

been reported in a number of languages from different linguistic families.<br />

This paper investigates the acoustics of laryngealized phonation in three<br />

indigenous languages spoken in Mexico, Yalalag Zapotec, Yucatec Maya,<br />

and Mixe. These languages differ in terms of the use of other features<br />

controlled by action of the larynx, i.e., tone. In Zapotec there is a contrast<br />

between high, low, and falling tones; Maya has phonemic high and low<br />

tones, whereas Mixe does not present phonemic pitch. The results show<br />

that the production of phonemic laryngeal vowels differs from language to<br />

language. The data suggest that the specific implementation of laryngealization<br />

depends in part on the relationship with contrastive tone. The patterns<br />

of the languages investigated provide new evidence of the possible<br />

synchronization of phonation throughout the vowel. With this evidence, a<br />

typology of modal/nonmodal phonation in phonation-synchronizing languages<br />

is proposed.<br />

4pSC28. The comparison between Thai and Japanese temporal<br />

control characteristics using segmental duration models.<br />

Chatchawarn Hansakunbuntheung and Yoshinori Sagisaka �GITI, Waseda<br />

Univ., 29-7 Bldg. 1-3-10, Nishi-Waseda, Shinjuku-ku, Tokyo 169-0051,<br />

Japan, chatchawarnh@fuji.waseda.jp�<br />

This paper compares the temporal control characteristics between Thai<br />

and Japanese read speech data using segmental duration models. The same<br />

and the different control characteristics have been observed from phone<br />

level to sentence level. The language-dependent and language-independent<br />

control factors have also been observed. In phone and neighboring phone<br />

level, different characteristics are found. Japanese vowel durations are<br />

mainly compensated by only adjacent preceding and following phones,<br />

which results from mora timing. Unlike Japanese, Thai vowel durations<br />

are affected by two succeding phones. It can be guessed that the differences<br />

come from syllabic structures. In word level, most content words<br />

tend to have longer phone durations while function words have shorter<br />

ones. In phrase level, both languages express duration lengtening of<br />

syllable/mora at the phrase initial and final. For language-specific factors,<br />

Thai tones express small alteration on phone duration. The comparisions<br />

explore the duration characteristics of the languages and give more understanding<br />

to be used in speech synthesis and second-language learning<br />

research. �Work supported in part by Waseda Univ. RISE research project<br />

of ‘‘Analysis and modeling of human mechanism in speech and language<br />

processing’’ and Grant-in-Aid for Scientific Research A-2, No. 16200016<br />

of JSPS.�<br />

4pSC29. Articulatory settings of French and English monolinguals<br />

and bilinguals. Ian L. Wilson �Univ. of Aizu, Tsuruga, Ikki-machi,<br />

Aizu-Wakamatsu City, Fukushima, 965-8580, Japan, wilson@u-aizu.ac.jp�<br />

and Bryan Gick �Univ. of BC, Vancouver, BC V6T1Z1 Canada�<br />

This study investigated articulatory setting �AS�, a language’s underlying<br />

posture of the articulators. Interspeech posture �ISP� of the articulators<br />

�their position when motionless during interutterance pauses� was<br />

used as a measure of AS in Canadian English and Quebecois French.<br />

3295 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3295<br />

4p FRI. PM


Optotrak and ultrasound imaging were used to test whether ISP is language<br />

specific in monolingual and bilingual speakers. Results show significant<br />

differences in ISP across the monolingual groups, with English<br />

exhibiting a higher tongue tip, more protruded upper and lower lips, and<br />

narrower horizontal lip aperture. Results also show that upper and lower<br />

lip protrusion are greater for the English ISP than for the French ISP, in all<br />

bilinguals who were perceived as native speakers of both languages, but in<br />

no other bilinguals. Tongue tip height results mirror those of the monolingual<br />

groups, for half of the bilinguals perceived as native speakers of both<br />

languages. Finally, results show that there is no unique bilingual-mode<br />

ISP, but instead one that is equivalent to the monolingual-mode ISP of a<br />

bilingual’s currently most-used language. This research empirically confirms<br />

centuries of noninstrumental evidence for AS, and for bilinguals it<br />

suggests that differences between monolingual and bilingual modes do not<br />

hold at the phonetic level.<br />

4pSC30. Temporal and spectral variability of vowels within and<br />

across languages with small vowel inventories: Russian, Japanese,<br />

and Spanish. Franzo F. LawII �Speech Acoust. and Percept. Lab., City<br />

Univ. of New York—Grad. Ctr., 365 Fifth Ave., New York, NY<br />

10016-4309, flaw@gc.cuny.edu�, Yana D. Gilichinskaya, Kikuyo Ito,<br />

Miwako Hisagi, Shari Berkowitz, Mieko N. Sperbeck, Marisa<br />

Monteleone, and Winifred Strange �City Univ. of New York—Grad. Ctr.,<br />

New York, NY 10016-4309�<br />

Variability of vowels in three languages with small vowel inventories<br />

�Russian, Japanese, and Spanish� was explored. Three male speakers of<br />

each language produced vowels in two-syllable nonsense words �VCa� in<br />

isolation and three-syllable nonsense words �gaC1VC2a� embedded within<br />

carrier sentences in three contexts: bilabial stops in normal rate sentences<br />

and alveolar stops in both normal and rapid rate sentences. Dependent<br />

variables were syllable duration and formant frequency at syllable midpoint.<br />

Results showed very little variation across consonant and rate conditions<br />

in formants for /i/ in Russian and Japanese. Japanese short /u, o, a/<br />

showed fronting �F2 increases� in alveolar context, which was more pronounced<br />

in rapid sentences. Fronting of Japanese long vowels was less<br />

pronounced. Japanese long/short vowel ratios varied with speaking style<br />

�isolation versus sentences� and speaking rate. All Russian vowels except<br />

/i/ were fronted in alveolar context, but showed little change in either<br />

spectrum or duration with speaking rate. Spanish showed a strong effect of<br />

consonantal context: front vowels were backed in bilabial context and<br />

back vowels were fronted in alveolar context, also more pronounced in<br />

rapid sentences. Results will be compared to female productions of the<br />

same languages, as well as American English production patterns.<br />

4pSC31. Does infant-directed speech in Tagalog resemble infantdirected<br />

speech in American English? Emmylou Garza-Prisby, Shiri<br />

Katz-Gershon, and Jean Andruski �Aud. & Speech-Lang. Pathol. Dept.,<br />

Wayne State Univ., 207 Rackham, 60 Farnsworth St., Detroit, MI 48202�<br />

This study compared the speech of a Filipino mother in infant- and<br />

adult-directed speech in order to investigate whether the mother used the<br />

acoustic features typically found in the infant-directed speech of American<br />

English-speaking mothers. Little acoustic documentation is available on<br />

the acoustic features of Tagalog, and no acoustic comparison of speech<br />

registers has so far been conducted. Impressionistically, Tagalog-speaking<br />

mothers’ do not appear to use the features typically found in American<br />

mothers speech to their young infants. The mother was recorded talking to<br />

her infant and to another adult native speaker of Tagalog. Recordings were<br />

made in the mother’s home and visits occurred during the first 6 months of<br />

the infant’s life. Specific acoustic features analyzed include �a� vowel<br />

duration, �b� vowel format frequencies, �c� vowel triangle size, �d� rate of<br />

speech, �e� fundamental frequency, and �f� F0 range. Morphological and<br />

syntactic features were also analyzed, including �g� mean length of utterance<br />

and �h� sentence length. Results support a need for further study of<br />

speech registers in Filipino mothers.<br />

4pSC32. Restricting relativized faithfulness and local conjunction in<br />

optimality theory. Haruka Fukazawa �Keio Univ., 4-1-1 Hiyoshi,<br />

Kohoku-ku, Yokohama, Japan�<br />

Within the framework of optimality theory �OT�, the two mechanisms,<br />

relativized faithfulness and local conjunction, play inevitable roles, especially<br />

when a simple constraint ranking fails to account for the data. However,<br />

their domain of application are too unrestricted and even overlapping<br />

each other. For instance, there are some cases which could be explained<br />

both by the ranking with relativized faithfulness and by the one with local<br />

conjunction. Syllable-final neutralization in German and geminate devoicing<br />

in Japanese loanwords are of interest in this context. The present paper<br />

proposes formal restrictions mostly on the local conjunction mechanism:<br />

the soundness of constraint combination, the number of constraints involved<br />

in a conjunction, and the local domain of conjunction. They not<br />

only can simplify the analysis but also give a more principled account for<br />

the overlapping cases. In fact, relativized faithfulness approach wins over<br />

local conjunction approach both in German neutralization and in Japanese<br />

loanwords. It is desirable for the universal grammar to be more restricted.<br />

Removing an overlap of theoretical devices is an important step toward<br />

the goal.<br />

3296 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3296


FRIDAY AFTERNOON, 1 DECEMBER <strong>2006</strong> KAUAI ROOM, 1:00 TO 2:55 P.M.<br />

Session 4pUWa<br />

Underwater Acoustics: Session in Honor of Leonid Brekhovskikh II<br />

William A. Kuperman, Cochair<br />

Scripps Inst. of Oceanography, Univ. of California, San Diego, Marine Physical Lab., La Jolla, CA 92093-0238<br />

Oleg A. Godin, Cochair<br />

NOAA, Earth System Research Lab., 325 Broadway, Boulder, CO 80305-3328<br />

Invited Papers<br />

1:00<br />

4pUWa1. Underwater noise as source of information on conditions and dynamics of ocean environments. Alexander V.<br />

Furduev �N. N. Andreyev Acoust. Inst., 4 Shvernika St., Moscow 117036, Russia�<br />

Leonid Brekhovskikh wrote in his book The Ocean and the Human �1987�: ‘‘Ocean noise is as important oceanographic parameter<br />

as temperature, current, and wind.’’ Brekhovskikh created and headed the Laboratory of Acoustic Methods of Ocean Research in 1956.<br />

One of the scientific directions of the Laboratory was investigation of underwater noise both as interference for sound reception and<br />

a source of environmental information. Long-term studies on the unique acoustic research vessels created under the initiative of<br />

Brekhovskikh resulted in numerous important findings, including ambient noise spectra and envelopes of acoustic fluctuations, depth<br />

dependence of noise directivity, and mechanisms of ambient noise generation. Brekhovskikh was always eager to find practical<br />

applications of scientific results. Different methods of ensuring noise immunity of hydroacoustic arrays were developed under his<br />

supervision. Passive methods of acoustic navigation based on use of natural noise were suggested. Techniques for underwater acoustic<br />

monitoring of the ocean based either on ambient noise analysis or reemission of noise from a point away from the receiving system<br />

have been developed. The success of the team of scientists headed by Brekhovskikh was determined by the creative atmosphere<br />

around him: there was neither competition nor commercial interests. The common goal was knowledge of the ocean.<br />

1:20<br />

4pUWa2. Distributed acoustic sensing in shallow water. Henrik Schmidt �Ctr. for Ocean Eng., MIT, Cambridge, MA 02139�<br />

The significance of Leonid Brekhovskikh to the ocean acoustics is undisputed. He was pioneering not only in terms of fundamental<br />

understanding of the ocean acoustic waveguide, but also the development of efficient and numerically stable approaches to<br />

propagation of sound in a stratified ocean. As such he has been an inspiration to a whole generation of model developers, leading to<br />

today’s suite of extremely powerful wave theory models, capable of accurately representing the complexity of the shallow-water ocean<br />

waveguide physics. The availability of these computational tools have in turn led to major advances in adaptive, model-based<br />

signal-processing techniques. However, such computationally intensive approaches are not necessarily optimal for the next generation<br />

of acoustic sensing systems. Thus, ocean observation in general is currently experiencing a paradigm shift away from platform-centric<br />

sensing concepts toward distributed sensing systems, made possible by recent advances in underwater robotics. In addition to a fully<br />

autonomous capability, the latency and limited bandwidth of underwater communication make on-board processing essential for such<br />

systems to be operationally feasible. In addition, the reduced sensing capability of the smaller physical apertures may be compensated<br />

by using mobility and artificial intelligence to dynamically adapt the sonar configuration to the environment and the tactical situation,<br />

and by exploiting multiplatform collaborative sensing. The development of such integrated sensing and control concepts for detection,<br />

classification, and localization requires extensive use of artificial intelligence incorporating a fundamental understanding of the ocean<br />

acoustic waveguide. No other sources in literature provide this with the clarity and depth that is the trademark of Academician<br />

Brekovskikh’s articles and classical textbooks. �Work supported by ONR.�<br />

1:40<br />

4pUWa3. When the shear modulus approaches zero: Fluids don’t<br />

bend and Scholte leaves the <strong>room</strong>. Robert I. Odom �Appl. Phys. Lab.,<br />

Univ. of Washington, 1013 NE 40th St., Seattle, WA 98105-6698�, Donna<br />

L. G. Sylvester, and Caitlin P. McHugh �Seattle Univ., Seattle, WA<br />

98122-1090�<br />

The 4�4 linear system of differential equations describing the propagation<br />

of the displacements and tractions in an elastic layered medium<br />

becomes singular as the shear modulus of the elastic medium approaches<br />

zero. There are a number of approximate ways to handle this singularity in<br />

Contributed Papers<br />

order to impart numerical stability to the computation of the elastic waves<br />

in a layered medium. For example, one can impose an irrotational constraint<br />

on the displacements or introduce a massive elastic interface �MEI�.<br />

Both of these ways of handling the weak shear strength are approximate,<br />

but avoid the need for singular perturbation theory �Gilbert, 1998�. Scholte<br />

waves are interface waves that propagate along the interface between an<br />

elastic solid and a fluid. They have nodes near or on the interface and<br />

decay exponentially into the bounding media. Scholte waves do not occur<br />

at the boundary between fluids. As the shear speed in the bounding elastic<br />

medium approaches zero, the Scholte waves disappear from the spectrum.<br />

We investigate this disappearance by applying singular perturbation theory<br />

3297 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3297<br />

4p FRI. PM


to the coupled fluid-elastic system. Among other things, we will address<br />

the rate in wave-number space at which the Scholte waves disappear from<br />

the spectrum.<br />

1:55<br />

4pUWa4. Measurement of the plane-wave reflection coefficient of the<br />

ocean bottom and the legacy of Leonid Brekhovskikh. George V.<br />

Frisk �Florida Atlantic Univ., 101 N. Beach Rd., Dania Beach, FL 33004�<br />

and Luiz L. Souza �Bose Corp., Framingham, MA 01701�<br />

Leonid Brekhovskikh’s classic text �Waves in Layered Media �Academic,<br />

New York, 1980�� inspired the development of several techniques<br />

for measuring the plane-wave reflection coefficient of the ocean bottom.<br />

Specifically, his application of the geometrical acoustics approximation to<br />

the problem of reflection of a spherical wave from a horizontally stratified<br />

medium provided the theoretical foundation for evaluating the strengths<br />

and weaknesses of various measurement methods. The most popular<br />

method assumes that the reflected field also consists of a spherical wave<br />

multiplied by the reflection coefficient evaluated at the specular angle.<br />

However, Brekhovskikh’s work showed that this interpretation is confined<br />

to a limited range of angles and bottom structures and, if applied improperly,<br />

can lead to unphysical results such as negative bottom loss. This<br />

paper describes a technique which circumvents these deficiencies. It consists<br />

of measuring the pressure field magnitude and phase versus range due<br />

to a cw point source and Hankel transforming these data to obtain the<br />

depth-dependent Green’s function versus horizontal wavenumber. The reflection<br />

coefficient is then obtained from the Green’s function using the<br />

analytical relationship between the two quantities. The method is demonstrated<br />

using 220-Hz data obtained in a near-bottom geometry in the Icelandic<br />

Basin. �Work supported by ONR.�<br />

2:10<br />

4pUWa5. Field from a point source above a layered half-space; theory<br />

and observations on reflection from the seabed. Charles W. Holland<br />

�Appl. Res. Lab., The Penn State Univ., State College, PA�<br />

L. M. Brekovskikh’s book Waves in Layered Media has provided decades<br />

of graduate students and researchers alike with a comprehensive and<br />

enormously useful reference. One topic from that work, reflection from a<br />

point source above a plane-layered medium �the seabed�, is discussed<br />

here. Both theoretical underpinnings and observations of reflection from a<br />

homogeneous halfspace, a transition layer with smoothly varying density<br />

and velocity profiles, and discrete layered media are considered for various<br />

shallow water sediment fabrics. �Work supported by the Office of<br />

Naval Research and NATO Undersea Research Centre.�<br />

2:25<br />

4pUWa6. Plane-wave model and experimental measurements of the<br />

directivity of a Fabry-Perot, polymer film, ultrasound sensor.<br />

Benjamin T. Cox and Paul C. Beard �Dept. of Med. Phys. and<br />

Bioengineering, Univ. College London, Gower St., London, WC1E 6BT,<br />

UK, bencox@mpb.ucl.ac.uk�<br />

Optical detection of ultrasound is popular due to the small element<br />

sizes that can be achieved. One method exploits the thickness change of a<br />

Fabry-Perot �FP� interferometer caused by the passage of an acoustic wave<br />

to modulate a laser beam. This detection method can have greater sensitivity<br />

than piezoelectric detectors for sub-millimeter element sizes. The<br />

directivity of FP sensors and the smallest achievable effective element size<br />

are examined here. A plane-wave model of the frequency-dependent directional<br />

response of the sensor, based on Brekhovskikh’s work on elastic<br />

waves in layered media, is described and validated against experimental<br />

directivity measurements made over a frequency range of 15 MHz and<br />

from normal incidence to 80 deg. In terms of applications, the model may<br />

be used to provide a noise-free response function that can be deconvolved<br />

from sound field measurements in order to improve accuracy in highfrequency<br />

metrology and imaging applications, or, for example, as a predictive<br />

tool to improve sensor design. Here, the smallest achievable effective<br />

element radius was investigated by comparing the directivity with that<br />

of a rigid circular pressure transducer, and found to be �0.9d, where d is<br />

the thickness of the FP interferometer. �Funding was provided by the<br />

EPSRC, UK�<br />

2:40<br />

4pUWa7. The interference head wave and its parametric dependence.<br />

Jee Woong Choi and Peter H. Dahl �Appl. Phys. Lab., Univ. of<br />

Washington�<br />

The interference head wave that can exist in the presence of a soundspeed<br />

gradient in the sediment, is a precursor arrival representing a transition<br />

between the first-order head wave and the zeroth-order refracted<br />

wave. Using a parabolic equation �PE� simulation, Choi and Dahl �J.<br />

Acoust. Soc. Am. 119, 2660–2668 �<strong>2006</strong>�� showed how the small shift in<br />

the dominant frequency of the interference head wave behaves as a function<br />

of the nondimensional parameter zeta, which itself is a function of<br />

center frequency, gradient, and range. For example, it was shown that the<br />

maximum frequency shift occurring in the vicinity of zeta equals 2. In this<br />

work, we investigate the amplitude and additional spectral properties of<br />

the interference head wave and analyze the cause of the frequency shift<br />

phenomenon using the ray theory. The limitation on the application of ray<br />

method also will be discussed. Finally, the conclusion will be verified by<br />

the time-dependent simulation using the RAM PE algorithm. �Work supported<br />

by the ONR.�<br />

3298 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3298


FRIDAY AFTERNOON, 1 DECEMBER <strong>2006</strong> KAUAI ROOM, 3:10 TO 4:25 P.M.<br />

Session 4pUWb<br />

Underwater Acoustics: Session in Honor of Fredrick Fisher<br />

William A. Kuperman, Chair<br />

Scripps Inst. of Oceanography, Univ. of California, San Diego, Marine Physical Lab., La Jolla, CA 92093-0238<br />

Chair’s Introduction—3:10<br />

Invited Papers<br />

3:15<br />

4pUWb1. FLIP „Floating Instrument Platform…: A major Fred Fisher contribution to ocean science. Fred N. Spiess, Robert<br />

Pinkel, William S. Hodgkiss, John A. Hildebrand �Marine Physical Lab., Scripps Inst. of Oceanogr., UCSD 0205, 9500 Gilman Drive;<br />

La Jolla, CA 92093-0205, fspiess@ucsd.edu�, and Gerald L. D’Spain �UCSD, La Jolla, CA 92093-0205�<br />

Frederick H. Fisher, a loyal and zealous member of the Acoustical Society of America, was an imaginative and effective developer<br />

of new techniques for research in both laboratory and seagoing acoustics. Most notable among his contributions was his work in<br />

bringing into being and enhancing the usefulness of the spar buoy laboratory, FLIP, from its inception in 1960. Not only did Fred use<br />

FLIP in his own research, its existence and many of its ancillary capabilities constituted a base for the seagoing research of others. The<br />

authors of this paper have benefited from FLIP’s unique capabilities, starting with long-range sound propagation studies in the 1960’s<br />

and 1970’s. FLIP’s stability and deep draft structure provided the platform for development of acoustic Doppler techniques for the<br />

measurement of ocean currents. Most recently, FLIP has been involved in studies of marine mammal vocalizations and use of<br />

multielement arrays to investigate details of shallow-water propagation. Fred’s initial studies of sonar bearing accuracy, for which<br />

FLIP’s construction was funded, and his dedication to advancing FLIP’s ability to contribute to ocean science, constitute a legacy that<br />

is being utilized today, more than 40 years after FLIP’s launching.<br />

3:35<br />

4pUWb2. Absorption of sound in seawater and ocean ambient noise, the scientific passions of Fred Fisher. John A. Hildebrand<br />

�Scripps Inst. of Oceanogr., UCSD, La Jolla, CA 92093-0205�<br />

Fred Fisher made seminal contributions to ocean acoustics in the understanding of the absorption of sound in seawater and ocean<br />

ambient noise. Laboratory data and long-range sound propagation data revealed excess acoustic absorption in seawater. Fred Fisher<br />

spent much of his scientific career, beginning with his Ph.D dissertation, teasing out the contributions of various seawater components<br />

to sound absorption, and his work on this topic set the standard for understanding and modeling these phenomena. Ambient noise is<br />

an important aspect of underwater signal detection and is the focus of recent concerns about disturbance of marine organisms. Fred<br />

Fisher made important contributions to ambient noise studies by conducting measurements of vertical directionality, thereby testing<br />

models for ambient noise production. The value of archival ambient noise data and recent increases in ambient noise will be discussed.<br />

3:55<br />

4pUWb3. Fred Fisher’s high-pressure work with eyewash and epsom<br />

salts. Christian de Moustier �Ctr. for Coastal & Ocean Mapping, Chase<br />

Ocean Eng. Lab, Univ. of New Hampshire, 24 Colovos Rd., Durham, NH<br />

03824�<br />

Starting in 1957 Fred Fisher led research programs devoted to highpressure<br />

measurements related to the physical chemistry of sound absorption<br />

in seawater due to magnesium sulfate and other salts. As he put it, he<br />

spent his professional lifetime squeezing epsom salt. His interest in the<br />

low-frequency anomalous sound absorption in the ocean below 1 kHz led<br />

to the discovery of boric acid as the cause of the low-frequency relaxation.<br />

This paper is a short review of Fred Fisher’s contributions to our knowledge<br />

of sound absorption in seawater, based in part on his carefully handwritten<br />

lecture notes and numerous low-pressure discussions.<br />

Contributed Papers<br />

4:10<br />

4pUWb4. Fred Fisher and research with acoustic vector sensors;<br />

Marine Physical Laboratory’s vertical array of directional acoustic<br />

receivers and ocean noise. Gerald L. D’Spain and William S. Hodgkiss<br />

�Marine Physical Lab, Scripps Inst. of Oceanogr., La Jollla, CA<br />

93940-0701�<br />

Fred Fisher had boundless enthusiasm for all topics acoustic. A chance<br />

encounter with him in the hallway usually led to a half-hour discussion of<br />

the latest research efforts at the lab and recent results he found exciting. In<br />

the 1980s, Fred became interested in the problem of identifying the physical<br />

phenomena forming the pedestal about the horizontal in vertical directionality<br />

measurements of the deep ocean’s low-frequency noise field. Two<br />

competing mechanisms had been proposed: downslope conversion of<br />

coastal shipping and noise from high latitude winds coupling into the deep<br />

3299 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3299<br />

4p FRI. PM


sound channel due to the shoaling of the sound channel axis. The relative<br />

contributions of these two mechanisms possibly could be separated if the<br />

azimuthal ambiguity of a vertical line array of hydrophones somehow<br />

could be broken. Therefore, Fred proposed to build a vertical array of<br />

‘‘DIFAR’’ sensors, which led to the design and construction of the Marine<br />

Physical Lab’s Vertical ‘‘DIFAR’’ Array. This talk will reminisce a bit<br />

about Fred as well as present some results from an ambient noise experiment<br />

conducted in 1992 on the continental shelf using the Vertical DIFAR<br />

Array co-deployed with MPL’s freely drifting vector sensors, the Swallow<br />

floats. �Work supported by ONR and ONT.�<br />

FRIDAY EVENING, 1 DECEMBER <strong>2006</strong> HAWAII BALLROOM, 7:00 TO 10:00 P.M.<br />

Awards Ceremony<br />

Anthony A. Atchley, President<br />

Acoustical Society of America<br />

Yôiti Suzuki, President<br />

Acoustical Society of Japan<br />

Acknowledgment of Honolulu Local Meeting Organizing Committees<br />

Presentation of Fellowship Certificates<br />

Anders Askenfeldt James A. McAteer<br />

Sergio Beristain David R. Palmer<br />

Philippe Blanc-Benon Marehalli G. Prasad<br />

David A. Conant Hiroshi Riquimaroux<br />

Andes C. Gade Peter A. Rona<br />

Anthony W. Gummer Mark V. Trevorrow<br />

Charles W. Holland Michael Vörlander<br />

Jody E. Kreiman Joos Vos<br />

Kevin D. LePage Ben T. Zinn<br />

Science Writing Award in Acoustics for Journalists to Radek Boschetty<br />

Science Writing Award for Professionals in Acoustics to Edwin Thomas<br />

Announcement of 2005 A. B. Wood Medal and Prize to Aaron Thode<br />

Distinguished Service Citation to Thomas D. Rossing<br />

Silver Medal in Noise to Alan H. Marsh<br />

Silver Medal in Physical Acoustics to Henry E. Bass<br />

Silver Medal in Psychological and Physiological Acoustics to William A. Yost<br />

Wallace Clement Sabine Award to William J. Cavanaugh<br />

Recognition of Acoustical Society of Japan meeting organizers<br />

Recognition of Acoustical Society of America meeting organizers<br />

3300 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />

3300

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!