friday morning, 1 december 2006 lanai room, 7 - DTU Orbit
friday morning, 1 december 2006 lanai room, 7 - DTU Orbit
friday morning, 1 december 2006 lanai room, 7 - DTU Orbit
Create successful ePaper yourself
Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.
FRIDAY MORNING, 1 DECEMBER <strong>2006</strong> LANAI ROOM, 7:30 TO 11:55 A.M.<br />
Session 4aAA<br />
Architectural Acoustics: Measurement of Room Acoustics I<br />
Fumiaki Satoh, Cochair<br />
Chiba Inst. of Technology, Tsudanuma 2-17-1 Narashino-shi, Chiba 275-0016, Japan<br />
Boaz Rafaely, Cochair<br />
Ben Gurion Univ., Electrical and Computer Engineering Dept., 84105, Beer Sheva, Israel<br />
Chair’s Introduction—7:30<br />
Invited Papers<br />
7:35<br />
4aAA1. Warped-time-stretched pulse: An acoustic test signal robust against ambient noise. Masanori Morise, Toshio Irino,<br />
Hideki Banno, and Hideki Kawahara �Wakayama Univ., 930, Sakaedani, Wakayama, 640-8510, Japan,<br />
s055068@sys.wakayama-u.ac.jp�<br />
A new acoustic measurement signal that is a hybrid signal of time-stretched pulse �TSP�, or lin-TSP, and logarithmic TSP<br />
�log-TSP� is proposed. The signal, referred to as warped-TSP �Morise et al., IEICE Trans. Fundamentals, A, J89-A�1�, 7–14 �<strong>2006</strong>��,<br />
has a single parameter to adjust for better measurements in accordance with ambient noise conditions. It also provides a means to<br />
eliminate harmonic distortions produced mainly by loudspeaker systems. In this lecture, the definition and features of the warped-TSP<br />
in comparison with the lin-TSP and log-TSP are introduced. The following were shown: �1� the relationship between the parameters,<br />
the amplitude frequency characteristics, and the effect on the harmonic distortion components; �2� a method to select the optimal<br />
parameters of the warped-TSP for a specific measuring environment; and �3� the experimental results for a series of impulse response<br />
measurements under different ambient noise conditions. Those results show that the proposed method outperformed the lin-TSP and<br />
log-TSP under all conditions in terms of SNR of the measured impulse response. �This research was supported partly by grants-in-aid<br />
for scientific research �15300061 and 15650032� and a grant from the Faculty of Systems Engineering at Wakayama University.�<br />
7:55<br />
4aAA2. Simultaneous estimation of reverberation times and their uncertainties from <strong>room</strong> impulse responses using a<br />
single-measurement procedure. Ning Xiang and Tomislav Jasa �Grad. Program in Architecture Acoust., and Dept. of Elec.,<br />
Comput., and Systems Eng., Rensselaer Polyt. Inst, Troy, NY 12180�<br />
Accurate measurements of reverberation times are of fundamental importance in <strong>room</strong> acoustics. A number of test procedures for<br />
characterizing acoustics in performing arts venues, quantifying acoustic properties of materials in chamber measurements, rely on<br />
experimental determination of reverberation times. In addition, decay-time estimation in acoustically coupled spaces has been found<br />
to be very demanding. Our recent work has demonstrated that model-based Bayesian approaches �Xiang et al., J. Acoust. Soc. Am.<br />
110, 1415–1424 �2001�; 113, 2685–2697 �2003�; 117, 3705–3715 �2005�� can be very useful for such analysis in architectural<br />
acoustics measurements. This paper discusses the recent development of probabilistic tools for estimating both reverberation �decay�<br />
times and their uncertainties within Bayesian framework. This work shows that Bayesian probabilistic inference can be used as a<br />
useful tool for sound energy decay analysis in both single-space halls and coupled spaces. Bayesian decay analysis simultaneously<br />
provides architectural acousticians with reverberation times, diverse decay times, related derivations, and interdependencies to quantify<br />
uncertainties of the estimation from a single measurement of <strong>room</strong> impulse responses followed by Schroeder backward integrations.<br />
8:15<br />
4aAA3. Permissible number of synchronous averaging times to obtain reverberation time from impulse response under<br />
time-variance conditions. Fumiaki Satoh, Yukiteru Hayashi �Chiba Inst. of Technol., Tsudanuma 2-17-1, Narashino-shi, Chiba,<br />
275-0016, Japan�, Shinichi Sakamoto �Univ. of Tokyo, Meguro-ku, Tokyo, 153-8505, Japan�, and Hideki Tachibana �Chiba Inst. of<br />
Technol., Narashino-shi, Chiba, 275-0016, Japan�<br />
In the measurement of <strong>room</strong> impulse response, the synchronous averaging technique and such new methods as the MLS and the<br />
swept-sine methods are being widely used to improve the signal-to-noise ratio. In actual measurement conditions, however, the air in<br />
a <strong>room</strong> is continuously moving and the temperature is changing to some degree. The measured value of the reverberation time in such<br />
a <strong>room</strong> tends to be shorter at higher frequencies when applying the synchronous averaging. Therefore, the assumption of a time<br />
invariant has to be carefully considered, and, on this point, some research has been conducted to date. We also have reported various<br />
research results concerning the impulse response measurement under the time-variance conditions. In this paper, the permissible<br />
number of synchronous averaging times for reverberation measurement is studied through some field experiments. In each field, many<br />
3223 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3223<br />
4a FRI. AM
time impulse response measurements were taken between a fixed pair of sound source and receiving positions by the swept-sine<br />
method, without averaging. After the measurements, the characteristics and the extent of the time-variance under measuring were<br />
estimated by a short-term running cross-correlation function between each impulse response. The influence of the time variance on the<br />
synchronous averaging result was studied based on the estimated time variance.<br />
8:35<br />
4aAA4. Selection of receiving positions suitable for evaluating acoustical parameters. Taeko Akama, Hisaharu Suzuki, and<br />
Akira Omoto �Omoto Lab., Dept. of Acoust. Design, Faculty of Design, Kyushu Univ., Shiobaru 4-9-1, Minami, Fukuoka 811-8540,<br />
Japan�<br />
Many physical parameters show characteristics of large sound fields such as concert halls. Some of them are adopted in the Annex<br />
of ISO 3382. That definition is clearly provided in ISO. However, practical measurement methods for them remain obscure. Our<br />
research is intended to examine an effective selection method of receiving positions based on the distribution of acoustical parameters<br />
in a real field. For that purpose, impulse responses are measured at more than 1400 seat positions to elucidate the distribution of<br />
acoustical parameters in an existing concert hall. The acoustical parameters, which are reverberation time, early decay time, clarity,<br />
and center time at each seat, are then calculated for 500-Hz, 1-kHz, and 2-kHz octave bands. The distributions of reverberation time<br />
are quite even at all seats. However, the distributions of other parameters show symmetrical patterns at 500 Hz. At 1 and 2 kHz<br />
frequencies, the distributions show asymmetrical patterns in this hall. Based on the results obtained in this study, an effective method<br />
to select the receiving position can be proposed.<br />
8:55<br />
4aAA5. Hybrid measurement method in <strong>room</strong> acoustics using dodecahedron speakers and a subwoofer. Hideo Miyazaki �Ctr.<br />
for Adv. Sound Technologies, Yamaha Corp., 203 Matsunokijima, Iwata, Shizuoka 438-0192, miyazaki@beat.yamaha.co.jp�<br />
A dodecahedron speaker is usually utilized for measurement in <strong>room</strong> acoustics under the hypothesis of omni directional point<br />
source. But generally speakers used for a dodecahedron speaker cannot playback low-frequency sound such as under 100 Hz, which<br />
is important especially for auralization, while the one constructed of units with large diameter to support low-frequency sounds cannot<br />
be considered as an omni-directional speaker in high frequencies. To meet these requirements, a hybrid system combining a dodecahedron<br />
speaker and a subwoofer has been developed and actually used for measurements of impulse responses in acoustical design of<br />
concert halls. The summary of this method will be presented. The feasibility of this method will be also discussed while evaluating the<br />
measurement results in concert halls by changing measurement conditions such as speaker locations and comparing these results with<br />
those of conventional methods.<br />
9:15<br />
4aAA6. The perception of apparent source width and its dependence on frequency and loudness. Ingo B. Witew and Johannes<br />
A. Buechler �Inst. of Tech. Acoust., RWTH Aachen Univ., Templergraben 55, 52066 Aachen, Germany�<br />
While it is widely accepted that apparent source width �ASW� is an important factor in characterizing the acoustics of a concert<br />
hall, there is still a lively discussion on how to refine the physical measures for ASW. A lot of experience has been gathered with<br />
interaural-cross-correlation and lateral-sound-incidence measures during the last years. As a result it was learned that different<br />
frequencies contribute differently to the perception of ASW and that the level of a sound also influences the perception of the apparent<br />
width of a source. With many technical measures having an influence on the perceptual aspect of ASW, the design of psychometric<br />
experiments becomes challenging as it is desirable to avoid the interaction of different objective parameters. In the experiments for the<br />
study presented, the perception of ASW is investigated for frequencies ranging from 100 Hz to 12.5 kHz at different levels of<br />
loudness. It is shown how the frequency and the loudness of a sound influence the perception of ASW.<br />
9:35<br />
4aAA7. Sound source with adjustable directivity. Gottfried K. Behler �Inst. fuer Technische Akustik, RWTH Aachen Univ.,<br />
D-52056 Aachen, Germany�<br />
Omni-directional sound sources are used to measure <strong>room</strong>-acoustical parameters in accordance with ISO 3382. To record a<br />
detailed <strong>room</strong> impulse response �RIR� with the aim of auralization, an extended frequency range is required that is not covered by the<br />
often-used building acoustics sound sources. To obtain this target, a loudspeaker with dedicated sources for low, mid, and high<br />
frequencies was designed, providing a smooth omni-directionality up to 6 kHz and a usable frequency range from 40 Hz up to 20 kHz.<br />
However, a realistic auralization of sources like musical instruments is not possible with an omni-directional measured RIR. To<br />
include the directional characteristics of instruments in the measuring setup, the directivity of the sound source has to be frequency<br />
dependent and must be matched to the �measured� directivity of the real instrument. This can be obtained by using a dodecahedron<br />
loudspeaker with independently operating systems and an appropriate complex FIR filtering of the frequency response of each driver.<br />
The directivity is a result of parameters like magnitude and phase and the interference sum of all systems. To create the appropriate<br />
directivity, optimization algorithms are used to achieve minimum error between measured and adapted directivity.<br />
3224 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3224
9:55–10:15 Break<br />
10:15<br />
4aAA8. Objective measures for evaluating tonal balance of sound fields. Daiji Takahashi �Dept. of Urban and Environ. Eng.,<br />
Kyoto Univ., Kyoto Univ. Katsura, Nishikyo-ku, Kyoto 615-8540, Japan, tkhs@archi.kyoto-u.ac.jp�, Kanta Togawa �FUJITEC Co.,<br />
Ltd., Hikone Shiga 522-8588, Japan�, and Tetsuo Hotta �YAMAHA Corp., Hamamatsu Shizuoka 430-8650, Japan�<br />
The purpose of this study is to derive objective measures, which can well represent the characteristics of the sound field regarding<br />
the tonal balance corresponding to our hearing sense. Two kinds of listening test were conducted in the form of paired comparison, in<br />
which subjects were tested using sound fields produced by convoluting some anechoic music sources with some impulse responses.<br />
In the first listening test, impulse responses were calculated theoretically for a simple structure of sound field having a direct sound<br />
and reflections, and, in the second test, impulse responses were measured at the various seats of existing concert halls. In the latter<br />
case, impulse responses which give almost the same reverberance were used for the listening tests. From this investigation, it is found<br />
that one objective measure called the DL �deviation of level� has a possibility of an effective measure, which can be used as an<br />
appropriate measure for evaluating the tonal balance of sound fields. The index DL is calculated from the data based on the<br />
logarithmic scale in both the frequency and magnitude. This fact is not inconsistent with the past findings that human response<br />
corresponds to a logarithmic scale of stimulus.<br />
10:35<br />
4aAA9. Measuring impulse responses containing complete spatial information. Angelo Farina, Paolo Martignon, Andrea Capra,<br />
and Simone Fontana �Industrial Eng. Dept., Univ. of Parma, via delle Scienze 181/A, 43100 Parma, Italy�<br />
Traditional impulse response measurements did capture limited spatial information. Often just omnidirectional sources and microphones<br />
are employed. In some cases it was attempted to get more spatial information employing directive transdudcers: known<br />
examples are binaural microphones, figure-of-8 microphones, and directive loudspeakers. However, these approaches are not scientifically<br />
based and do not provide an easy way to process and visualize the spatial information. On the other side, psychoacoustics<br />
studies demonstrated that ‘‘spatial hearing’’ is one of the dominant factors for the acoustic quality of <strong>room</strong>s, particularly for theatres<br />
and concert halls. Of consequence, it is necessarily to reformulate the problem entirely, describing the transfer function between a<br />
source and a receiver as a time/space filter. This requires us to ‘‘sample’’ the impulse response not only in time, but also in space. This<br />
is possible employing spherical harmonics for describing, with a predefined accuracy, the directivity pattern of both source and<br />
receiver. It is possible to build arrays of microphones and of loudspeakers, which, by means of digital filters, can provide the required<br />
directive patterns. It can be shown how this makes it possible to extract useful information about the acoustical behavior of the <strong>room</strong><br />
and to make high-quality auralization.<br />
10:55<br />
4aAA10. Spherical and hemispherical microphone arrays for capture and analysis of sound fields. Ramani Duraiswami,<br />
Zhiyun Li, Dmitry N. Zotkin, and Elena Grassi �Perceptual Interfaces and Reality Lab., Inst. for Adv. Comput. Studies, Univ. of<br />
Maryland, College Park, MD 20742�<br />
The capture of the spatial structure of a sound field and analysis is important in many fields including creating virtual environments,<br />
source localization and detection, noise suppression, and beamforming. Spherical microphone arrays are a promising development<br />
to help achieve such capture and analysis, and have been studied by several groups. We develop a practical spherical<br />
microphone array and demonstrate its utility in applications for sound capture, <strong>room</strong> measurement and for beamforming and tracking.<br />
To accommodate equipment failure and manufacturing imprecision we extend their theory to handle arbitrary microphone placement.<br />
To handle speech capture and surveillance we describe the development of a new sensor, the hemispherical microphone array. For<br />
each array the practical performance follows that predicted by theory. Future applications and improvements are also discussed. �Work<br />
supported by NSF.�<br />
11:15<br />
4aAA11. High-order wave decomposition using a dual-radius spherical microphone array. Boaz Rafaely, Ilya Balmages, and<br />
Limor Eger �Dept. of Elec. and Comput. Eng., Ben-Gurion Univ. of the Negev, Beer-Sheva 84105, Israel�<br />
The acoustic performance of an auditorium is influenced by the manner in which sound propagates from the stage into the seating<br />
areas. In particular, the spatial and temporal distribution of early reflections is considered important for sound perception in the<br />
auditorium. Previous studies presented measurement and analysis methods based on spherical microphone arrays and plane-wave<br />
decomposition that could provide information on the direction and time of arrival of early reflections. This paper presents recent<br />
results of <strong>room</strong> acoustics analysis based on a spherical microphone array, which employs high spherical harmonics order for improved<br />
spatial resolution, and a dual-radius spherical measurement array to avoid ill-conditioning at the null frequencies of the spherical<br />
Bessel function. Spatial-temporal analysis is performed to produce directional impulse responses, while time-windowed spacefrequency<br />
analysis is employed to detect direction of arrival of individual reflections. Experimental results of sound-field analysis in<br />
a real auditorium will also be presented.<br />
3225 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3225<br />
4a FRI. AM
11:35<br />
4aAA12. Impulse response measurement system and its recent applications. Kazuhiro Takashima, Hiroshi Nakagawa, Natsu<br />
Tanaka, and Daiki Sekito �1-21-10, Midori, Sumida-Ku, Tokyo 130-0021, Japan�<br />
Our impulse response measurement system has been developed for ten years. During this decade, the environment related to this<br />
measurement has changed significantly. In this article, the features and notes on the measurement system using the sound card, and our<br />
brand new system, which is expanded for multichannel inputs, will be presented. Finally, a new technique, which combines multichannel<br />
impulse response measurement and signal processing with microphone array, will be presented. The microphone array was<br />
designed for noise analysis for automobile interiors. The array consists of 31 microphones on the surface of an acoustically hard<br />
sphere. Moreover, 12 cameras are arranged on the surface of the sphere to take photos. Some applications and future development will<br />
be presented.<br />
FRIDAY MORNING, 1 DECEMBER <strong>2006</strong> KOHALA/KONA ROOM, 8:00 TO 11:45 A.M.<br />
Session 4aAB<br />
Animal Bioacoustics: Marine Mammal Acoustics I<br />
Paul E. Nachtigall, Chair<br />
Hawaii Inst. of Marine Biology, P.O. Box 1106, Kailua, HI 96734<br />
8:00<br />
4aAB1. Development of evoked-potential audiometry in odontocetes.<br />
Alexander Supin �Inst. of Ecology and Evolution, 33 Leninsky prospect,<br />
119071 Moscow, Russia�<br />
Evoked-potential methods are widely used for investigation of hearing<br />
in whales, dolphins, and porpoises. For this purpose, mostly the auditory<br />
brainstem response �ABR� or rhythmic trains of ABRs, the envelopefollowing<br />
response �EFR�, are used. Although very productive, these<br />
methods require further elaboration. �i� Traditionally the EFR is provoked<br />
by sinusoidally amplitude-modulated tones �SAM�. SAM stimuli have narrow<br />
frequency band, which makes them little effective to produce the<br />
EFR, because response amplitude depends on the stimulus bandwidth. A<br />
solution of the problem is the use of trains of short tone pips instead of<br />
SAM tones. Such stimuli produce several times higher EFR than SAM<br />
tones. This makes the threshold determination much more confident and<br />
precise. The effect is achievable at stimulus bandwidths, which still do not<br />
influence negatively the precision of attribution of the threshold to a certain<br />
frequency. �ii� To extract low-amplitude evoked potentials from noise,<br />
the average technique is traditionally used. This operation returns a mean<br />
value of averaged traces. Effectively diminishing stationary noise, this<br />
method poorly eliminates big artifacts, which may spoil the record even it<br />
if appeared once or twice during acquisition. With this respect, computation<br />
of the median instead of mean is much more effective.<br />
8:15<br />
4aAB2. Towards a predictive model of noise-induced temporary<br />
threshold shift for an amphibious marine mammal, the California sea<br />
lion „Zalophus californianus…. David Kastak, Marla M. Holt, Jason<br />
Mulsow, Colleen J. Reichmuth Kastak, Ronald J. Schusterman �UCSC<br />
Long Marine Lab., 100 Shaffer Rd., Santa Cruz, CA 95060�, and Brandon<br />
L. Southall �Natl. Marine Fisheries Service, Silver Spring, MD 20910�<br />
A California sea lion that had previously been tested under water was<br />
assessed for noise-induced temporary threshold shift �TTS� in air. One<br />
hundred ninety-two controlled exposures of octave-band noise centered at<br />
2.5 kHz were conducted over a 3-year period. The noise was varied in<br />
level �to 133 dB SPL re: 20�Pa� and duration �to 50 min� to generate a<br />
variety of equal sound exposure levels �SELs�. Behavioral psychophysics<br />
was used to measure hearing sensitivity at 2.5 kHz before, immediately<br />
following, and 24 h following noise exposure. The levels of threshold<br />
Contributed Papers<br />
shifts obtained ranged up to 30 dB. In cases where TTS exceeded 20 dB,<br />
thresholds were obtained at regular intervals until recovery occurred. The<br />
average slope of the long-term recovery function was 10 dB per log-<br />
�minute�. Results show that the threshold shifts correlated with SEL; however,<br />
the equal-energy trading rule did not apply in all circumstances, with<br />
exposure duration contributing more than exposure level. Repeated testing<br />
showed no evidence of a permanent threshold shift at 2.5 kHz or octave<br />
higher. The amphibious sea lions appear to be equally susceptible to noise<br />
in air and under water, provided that the exposure levels are referenced to<br />
the subjects thresholds in both media.<br />
8:30<br />
4aAB3. Electrophysiological investigation of temporal resolution in<br />
three pinniped species: Adaptive implications. Jason Mulsow and<br />
Colleen Reichmuth Kastak �Univ. of California Santa Cruz, Long Marine<br />
Lab., 100 Shaffer Rd., Santa Cruz, CA 95060�<br />
Electrophysiological studies of auditory temporal processing in marine<br />
mammals have traditionally focused on the role of highly refined temporal<br />
resolution in dolphin echolocation. Studies in manatees, however, have<br />
found their temporal resolution to be better than expected, leading to<br />
speculation that such capabilities are an adaptation for underwater sound<br />
localization. This study measured the ability of auditory brainstem responses<br />
to follow rhythmic click stimuli in California sea lions �Zalophus<br />
californianus�, harbor seals �Phoca vitulina�, and northern elephant seals<br />
�Mirounga angustirostris�. Trains of 640-s clicks were presented in air at<br />
repetition rates of 125–1500 per second and averaged rate-following responses<br />
were recorded. Rate-following responses were detected in both<br />
the harbor seal and the sea lion at rates up to 1000 clicks per second,<br />
indicating that pinnipeds, like manatees, possess temporal resolution<br />
greater than humans but inferior to dolphins. While this finding might<br />
support an underwater sound localization hypothesis, comparable results<br />
were obtained in preliminary testing of a dog �Canis familiaris�, suggesting<br />
that increased temporal resolution in pinnipeds may not be the result of<br />
the evolutionary pressure of an aquatic environment, but rather a result of<br />
increased high-frequency hearing essential to mammalian sound localization.<br />
�Work supported by NOPP, ONR, and NMFS.�<br />
3226 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3226
8:45<br />
4aAB4. Click and tone-pip auditory evoked potentials in a large<br />
marine mammal, the northern elephant seal. Dorian S. Houser<br />
�BIOMIMETICA, 7951 Shantung Dr., Santee, CA 92071� and James J.<br />
Finneran �Space and Naval Warfare Systems Ctr., San Diego, CA 92152�<br />
The use of auditory-evoked potentials �AEPs� to study the hearing of<br />
mysticete whales is challenged by access to animals, their large size, and<br />
proportionately smaller brain relative to odontocetes. One means by which<br />
AEP techniques can be adapted to these larger animals is by application to<br />
more readily available proxy species. The northern elephant seal �Mirounga<br />
angustirostris� is a large pinniped, potentially in excess of 2000 kg,<br />
with a thick dermis, large skull, relatively small auditory nerve, and a<br />
low-frequency vocal communication system. AEP collection in elephant<br />
seals provides similar challenges to those of the mysticetes but at a scale<br />
that provides a greater opportunity for success. AEP tests were conducted<br />
on northern elephant seals at Año Nuevo State Reserve, the natural haulout<br />
site of the elephant seal. Subjects were chemically immobilized with<br />
tiletamine/zolazepam and chemical restraint was maintained with bolus<br />
injections of ketamine. Click-evoked potentials were collected from four<br />
weanling and two adult male elephant seals and tone-pip-evoked potentials<br />
were collected from a 2-year-old female. Results demonstrate that<br />
AEPs can be recorded from large pinniped species, providing a step towards<br />
the application of similar techniques to larger cetacean species.<br />
9:00<br />
4aAB5. Acoustic field measurements and bottlenose dolphin hearing<br />
thresholds using single-frequency and frequency-modulated tones.<br />
James J. Finneran �U.S. Navy Marine Mammal Program,<br />
SPAWARSYSCEN San Diego, Code 2351, 49620 Beluga Rd., San Diego,<br />
CA 92152, james.finneran@navy.mil� and Carolyn E. Schlundt �EDO<br />
Professional Services, San Diego, CA 92110�<br />
Studies of underwater hearing are often hampered by the behavior of<br />
sound waves in small experimental tanks. At lower frequencies, tank dimensions<br />
are often not sufficient for free-field conditions, resulting in<br />
large spatial variations of sound pressure. These effects may be mitigated<br />
somewhat by increasing the frequency bandwidth of the sound stimulus,<br />
so effects of multipath interference average out over many frequencies. In<br />
this study, acoustic fields and bottlenose dolphin �Tursiops truncatus�<br />
hearing thresholds were compared for pure-tone and frequency-modulated<br />
stimuli. Experiments were conducted in a vinyl-walled, seawater-filled<br />
pool approximately 4�5�1.5 m. Sound stimuli consisted of 500-ms<br />
tones at 13 carrier frequencies between 1 and 100 kHz. Frequencymodulated<br />
stimuli featured both linear and sinusoidal modulating waveforms<br />
with 5%, 10%, and 20% bandwidths. Acoustic fields were measured<br />
�without the dolphin present� at three depths over a 60�65-cm grid with a<br />
5-cm spacing. Hearing thresholds were measured using a behavioral response<br />
paradigm and up/down staircase technique. Frequency-modulated<br />
stimuli with a 10% bandwidth resulted in significant improvements to the<br />
sound field without substantially affecting the dolphins hearing thresholds.<br />
�Work supported by ONR.�<br />
9:15<br />
4aAB6. Hearing frequency selectivity in four species of toothed<br />
whales as revealed by the evoked-potential method. Vladimir Popov<br />
�Inst. of Ecology and Evolution, 33 Leninsky Prosp., 119071 Moscow,<br />
Russia popov_vl@sevin.ru�<br />
Frequency tuning curves were obtained using a tone-tone simultaneous<br />
masking paradigm in conjunction with the evoked potential recording. The<br />
masker was a continuous tone and the test was a sinusoidal amplitudemodulated<br />
�SAM� tonal signal, which evoked the envelope following response<br />
�EFR�. The EFR was recorded in unanaesthetized animals from a<br />
head surface with the use of suction-cup electrodes. The obtained tuning<br />
curves featured very sharp tuning with Q�ERB� �quality estimated by the<br />
equivalent rectangular bandwidth� from35inTursiops truncatus to nearly<br />
50 in Delphinapterus leucas. This acuteness is several times better than in<br />
humans and many animals. The Q�ERB� dependence on probe frequency<br />
could be approximated by regression lines with a slope from 0.18 in Tur-<br />
siops trucatus to 0.83–0.86 in Phocoena phocoena and Neophocoena phocoenoides.<br />
Thus, the frequency representation in the odontocete auditory<br />
system may be either near constant quality �in Tursiops� or near constant<br />
bandwidth �in porpoises�. �Work supported by The Russian Foundation for<br />
Basic Research and Russian President Grant.�<br />
9:30<br />
4aAB7. Growth and recovery of temporary threshold shifts in a<br />
dolphin exposed to midfrequency tones with durations up to 128 s.<br />
Carolyn E. Schlundt �EDO Professional Services, 3276 Rosecrans St.,<br />
San Diego, CA 92110, carolyn.melka@edocorp.com�, Randall L. Dear<br />
�Sci. Applications Intl. Corp., San Diego, CA 92110�, Donald A. Carder,<br />
and James J. Finneran �Space and Naval Warfare Systems Ctr., San<br />
Diego, San Diego, CA 92152�<br />
Auditory thresholds at 4.5 kHz were measured in a bottlenose dolphin<br />
�Tursiops truncatus� before and after exposure to midfrequency tones at 3<br />
kHz. Experiments were conducted in relatively quiet pools with low ambient<br />
noise levels at frequencies above 1 kHz. Behavioral hearing tests<br />
allowed for thresholds to be routinely measured within 4 min postexposure,<br />
and tracked recovery for at least 30 min postexposure. Exposure<br />
durations ranged from 4 to 128 s at sound pressure levels ranging from<br />
149 to 200 dB re: 1�Pa. Sound exposure levels ranged from 155 to 217<br />
dB re: 1�Pa 2 /s. Temporary threshold shifts at 4 min postexposure (TTS 4)<br />
of up to 23 dB were observed. All thresholds recovered to baseline and<br />
pre-exposure levels, most within 30 min of exposure. �Work supported by<br />
the U.S. ONR.�<br />
9:45<br />
4aAB8. Auditory brainstem response recovery rates during doublepulse<br />
presentation in the false killer whale „Pseudorca crassidens…: A<br />
mechanism of automatic gain control? Paul E. Nachtigall �Marine<br />
Mammal Res. Program, Hawaii Inst. of Marine Biol., P.O. Box 1106,<br />
Kailua, HI 96734�, Alexander Ya. Supin �Russian Acad. of Sci., Moscow,<br />
Russia�, and Marlee Breese �Hawaii Inst. of Marine Biol., Kailua, HI<br />
96734�<br />
The outgoing echolocation pulse and the return echo response can be<br />
approximately examined in the auditory system of an echolocating animal<br />
by presenting two pulses and determining the forward-masking effect of<br />
the first pulse on the response to the second pulse using auditory-evoked<br />
potential procedures. False killer whale, Pseudorca crassidens, auditory<br />
brainstem responses �ABR� were recorded using a double-click stimulation<br />
paradigm specifically measuring the recovery of the second �test�<br />
response �to the second click� as a function of the length of the interclick<br />
interval �ICI� following various levels of the first �conditioning� click. At<br />
all click intensities, the slopes of the recovery functions were almost constant:<br />
0.60.8 V per ICI decade. Therefore, even when the conditioning-toclick<br />
level ratio was kept constant, the duration of recovery was intensity<br />
dependent: the higher intensity the longer the recovery. The conditioningto-test-click<br />
level ratio strongly influenced the recovery time: the higher<br />
the ratio, the longer the recovery. This dependence was nearly linear, using<br />
a logarithmic ICI scale with a rate of 2530 dB per ICI decade. These data<br />
were used for modeling the interaction between the emitted click and the<br />
echo in the auditory system during echolocation.<br />
10:00–10:15 Break<br />
10:15<br />
4aAB9. Temporary threshold shifts in the bottlenose dolphin<br />
„Tursiops truncatus…, varying noise duration and intensity. T. Aran<br />
Mooney �Dept. of Zoology and Hawaii Inst. of Marlne Biol., Univ. of<br />
Hawaii, 46-007 Lilipuna Rd., Kaneohe, HI 96744�, Paul E. Nachtigall,<br />
Whitlow W. L. Au, Marlee Breese, and Stephanie Vlachos �Univ. of<br />
Hawaii, Kaneohe, HI 96744�<br />
There is much concern regarding increasing noise levels in the ocean<br />
and how it may affect marine mammals. However, there is a little information<br />
regarding how sound affects marine mammals and no published<br />
3227 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3227<br />
4a FRI. AM
data examining the relationship between broadband noise intensity and<br />
exposure duration. This study explored the effects of octave-band noise on<br />
the hearing of a bottlenose dolphin by inducing temporary hearing threshold<br />
shifts �TTS�. Sound pressure level �SPL� and exposure duration were<br />
varied to measure the effects of noise duration and intensity. Hearing<br />
thresholds were measured using auditory evoked potentials before and<br />
after sound exposure to track and map TTS and recovery. Shifts were<br />
frequency dependent and recovery time depended on shift and frequency,<br />
but full recovery was relatively rapid, usually within 20 min and always<br />
within 40 min. As exposure time was halved, TTS generally occurred with<br />
an increase in noise SPL. However, with shorter, louder noise, threshold<br />
shifts were not linear but rather shorter sounds required greater sound<br />
exposure levels to induce TTS, a contrast to some published literature.<br />
From the data a novel algorithm was written that predicts the physiological<br />
effects of anthropogenic noise if the intensity and duration of exposure<br />
are known.<br />
10:30<br />
4aAB10. Estimates of bio-sonar characteristics of a free-ranging<br />
Ganges river dolphin. Tamaki Ura, Harumi Sugimatsu, Tomoki Inoue<br />
�Underwater Technol. Res. Ctr., Inst. of Industrial Sci., Univ. of Tokyo,<br />
4-6-1 Komaba, Meguro, Tokyo 153-8505, Japan�, Rajendar Bahl �IIT<br />
Delhi, New Delhi 110016, India�, Junichi Kojima �KDDI R&D Labs.<br />
Inc., Saitama 356-8502, Japan�, Tomonari Akamatsu �Fisheries Res.<br />
Agency, Ibaraki 314-0408, Japan�, Sandeep Behera �Freshwater &<br />
Wetlands Prog., New Delhi 110003, India�, Ajit Pattnaik, Muntaz Kahn<br />
�Chilika Dev. Authority, Orissa, India�, Sudhakar Kar, Chandra Sekhar<br />
Kar �Off. of the Princip. CCF �Wildlife� & Chief Wildlife Warden,<br />
Blaubaneswar 751012, India�, Tetsuo Fukuchi, Hideyuki Takashi<br />
�System Giken Co. Ltd., Kanagawa 253-0085, Japan�, and Debabrata<br />
Swain �Simpilipal Biosphere and Tiger Reserve, Orissa, India�<br />
This paper reports the first known studies of the bio-sonar characteristics<br />
of an isolated free-ranging Ganges river dolphin, Platanista<br />
gangetica. The animal preferred to roam in a deeper tract of the otherwise<br />
shallow river. The click sounds of the dolphin were recorded over a period<br />
of 2 days on a 3.2-meter-long high-frequency hydrophone array composed<br />
of three hydrophones forming an equispaced linear array and another two<br />
hydrophones in conjunction with the central hydrophone forming an SSBL<br />
triangular array in a plane perpendicular to the array axis. The array was<br />
deployed both in horizontal and vertical configurations. The array structure<br />
provided 3-D measurements of the source location through measurement<br />
of the interelement time delay. Bio-sonar characteristics such as click<br />
duration, bandwidth, and interclick intervals in click trains have been reported.<br />
Measurements of dolphin track and the relative click levels on the<br />
array hydrophones have been used to obtain a preliminary characterization<br />
of the animal’s beam pattern.<br />
10:45<br />
4aAB11. Discriminating between the clicks produced by a bottlenose<br />
dolphin when searching for and identifying an object during a search<br />
task. Sandra Bohn, Stan Kuczaj �Univ. of Southern Mississippi, 118<br />
College Dr., #5025, Hattiesburg, MS 39406, sandra.bohn@usm.edu�, and<br />
Dorian Houser �BIOMIMETICA, Santee, CA 92071�<br />
Clicks collected from an echolocating bottlenose dolphin completing a<br />
search task were compared in order to determine if the clicks produced<br />
when the dolphin was acquiring the target differed from the clicks produced<br />
when the dolphin was searching for the target. The clicks produced<br />
by a free-swimming dolphin completing the search task were recorded<br />
using a biosonar measurement tool �BMT�, an instrumentation package<br />
carried by the dolphin that collected both the outgoing clicks and the<br />
returning echoes. A discriminant function analysis classified the clicks as<br />
search or acquisition using the variables of peak-to-peak amplitude, duration,<br />
peak frequency, center frequency, and bandwidth. The acquisition<br />
clicks were classified more accurately than the search clicks. Acquisition<br />
clicks and search clicks were significantly different across all five of the<br />
variables. These results suggest that the clicks produced by bottlenose<br />
dolphins acquiring a target are different than those produced by dolphins<br />
searching for a target.<br />
11:00<br />
4aAB12. Echo highlight amplitude and temporal difference<br />
resolutions of an echolocating Tursiops truncatus. Mark W. Muller,<br />
Whitlow W. L. Au, Paul E. Nachtigall, Marlee Breese �Marine Mammal<br />
Res. Program, Hawai’i Inst. of Marine Biol., 46-007 Lilipuna Rd.,<br />
Kaneohe, HI 96744�, and John S. Allen III �Univ. of Hawai’i at Manoa,<br />
Honolulu, HI 96822�<br />
A dolphin’s ability to discriminate targets may greatly depend on the<br />
relative amplitudes and the time separations of echo highlights within the<br />
received signal. Previous experiments with dolphins have varied the physical<br />
parameters of targets, but did not fully investigate how changes in<br />
these parameters corresponded with the composition of the scattered<br />
acoustic waveforms and the dolphin’s subsequent response. A novel experiment<br />
utilizes a phantom echo system to test a dolphin’s detection<br />
response of relative amplitude differences of secondary echo highlights<br />
and the time separation differences of all the echo highlights both within<br />
and outside the animal’s integration window. By electronically manipulating<br />
these echoes, the underlying acoustic classification cues can be more<br />
efficiently investigated. In the first study, the animal successfully discriminated<br />
between a standard echo signal and one with the middle highlight<br />
amplitude at �7 dB. When the middle highlight amplitude was raised to<br />
�6 dB, the animal’s discrimination performance radically dropped to<br />
65%. This study suggests the animal may not be as sensitive to the secondary<br />
echo highlights as previously proposed. The experiments were repeated<br />
for the trailing highlight amplitude and the time separations between<br />
the primary and middle highlights and the middle and trailing<br />
highlights.<br />
11:15<br />
4aAB13. A background noise reduction technique for improving false<br />
killer whale „Pseudorca crassidens… localization. Craig R. McPherson,<br />
Owen P. Kenny, Phil Turner �Dept. of Elec. and Comput. Eng., James<br />
Cook Univ., Douglas 4811, Queensland, Australia�, and Geoff R.<br />
McPherson �Queensland Dept. of Primary Industries and Fisheries,<br />
Cairns, 4870, Queensland Australia�<br />
The passive localization of false killer whales �Pseudorca crassidens�<br />
in acoustic environments comprised of discontinuous ambient, anthropogenic,<br />
and animal sounds is a challenging problem. A background noise<br />
reduction technique is required to improve the quality of sampled recordings,<br />
which will assist localization using auditory modeling and signal<br />
correlation at extended ranges. The algorithm developed meets this requirement<br />
using a combination of adaptive percentile estimation, a<br />
median-based tracker, and Gaussian windowing. The results indicate successful<br />
improvement of the signal-to-noise ratio, and consequently a significant<br />
increase in the detection range of false killer whales in acoustically<br />
complex environments.<br />
11:30<br />
4aAB14. Analysis of Australian humpback whale song using<br />
information theory. Jennifer L. Miksis-Olds, John R. Buck �School for<br />
Marine Sci. and Technol., Univ. of Massachusetts Dartmouth, New<br />
Bedford, MA 02744, jmiksis@umassd.edu�, Michael J. Noad �Univ. of<br />
Queensland, St. Lucia, QLD 4072 Australia�, Douglas H. Cato �Defence<br />
Sci. & Tech. Org., Pyrmont, NSW 2009 Australia�, and Dale Stokes<br />
�Scripps Inst. of Oceanogr., La Jolla, CA 92093�<br />
Songs produced by migrating whales were recorded off the coast of<br />
Queensland, Australia over 6 consecutive weeks in 2003. Approximately<br />
50 songs were analyzed using information theory techniques. The average<br />
length of the songs estimated by correlation analysis was approximately<br />
100 units, with song sessions lasting from 300 to over 3100 units. Song<br />
entropy, a measure of structural constraints and complexity, was estimated<br />
using three different methodologies: �1� the independently identically distributed<br />
model; �2� first-order Markov model; and �3� the nonparametric<br />
sliding window match length �SWML� method, as described in Suzuki<br />
et al. �J. Acoust. Soc. Am. 119, 1849 �<strong>2006</strong>��. The analysis finds the songs<br />
of migrating Australian whales are consistent with the hierarchical structure<br />
proposed by Payne and McVay �Science 173, 585�597 �1971��, and<br />
3228 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3228
ecently confirmed by Suzuki et al. for singers on the breeding grounds.<br />
Both the SWML entropy estimates and the song lengths for the Australian<br />
singers were lower than that reported by Suzuki et al. for Hawaiian<br />
whales in 1976�1978. These lower SWML entropy values indicate a<br />
higher level of predictability within songs. The average total information<br />
in the Australian sequence of song units was approximately 35 bits/song.<br />
Aberrant songs �10%� yielded entropies similar to the typical songs.<br />
�Sponsored by ONR and DSTO.�<br />
FRIDAY MORNING, 1 DECEMBER <strong>2006</strong> KAHUKU ROOM, 7:55 A.M. TO 12:00 NOON<br />
Session 4aBB<br />
Biomedical UltrasoundÕBioresponse to Vibration: Interaction of Cavitation Bubbles with Cells and Tissue<br />
John S. Allen, Cochair<br />
Univ. of Hawaii, Dept. of Mechanical Engineering, 2540 Dole St., Honolulu, HI 96822<br />
Yoshiki Yamakoshi, Cochair<br />
Gunma Univ., Faculty of Engineering, 1-5-1 Tenjin-cho, Kiryu-shi, Gunma 376-8515, Japan<br />
Chair’s Introduction—7:55<br />
Invited Papers<br />
8:00<br />
4aBB1. Ultra-high-speed imaging of bubbles interacting with cells and tissue. Michel Versluis, Philippe Marmottant, Sascha<br />
Hilgenfeldt, Claus-Dieter Ohl �Phys. of Fluids, Univ. of Twente, P.O. Box 217, 7500 AE Enschede, The Netherlands�, Chien T.<br />
Chin, Annemieke van Wamel, Nico de Jong �Erasmus MC, 3000 DR Rotterdam, The Netherlands�, and Detlef Lohse �Univ. of<br />
Twente, 7500 AE Enschede, The Netherlands�<br />
Ultrasound contrast microbubbles are exploited in molecular imaging, where bubbles are directed to target cells and where their<br />
high-scattering cross section to ultrasound allows for the detection of pathologies at a molecular level. In therapeutic applications<br />
vibrating bubbles close to cells may alter the permeability of cell membranes, and these systems are therefore highly interesting for<br />
drug and gene delivery applications using ultrasound. In a more extreme regime bubbles are driven through shock waves to sonoporate<br />
or kill cells through intense stresses or jets following inertial bubble collapse. Here, we elucidate some of the underlying mechanisms<br />
using the 25-Mfps camera Brandaris128, resolving the bubble dynamics and its interactions with cells. We quantify acoustic microstreaming<br />
around oscillating bubbles close to rigid walls and evaluate the shear stresses on nonadherent cells. In a study on the fluid<br />
dynamical interaction of cavitation bubbles with adherent cells, we find that the nonspherical collapse of bubbles is responsible for cell<br />
detachment. We also visualized the dynamics of vibrating microbubbles in contact with endothelial cells followed by fluorescent<br />
imaging of the transport of propidium iodide, used as a membrane integrity probe, into these cells showing a direct correlation<br />
between cell deformation and cell membrane permeability.<br />
8:20<br />
4aBB2. Sonoporation: Mechanisms of cell membrane perforation and rapid resealing. Nobuki Kudo and Katsuyuki Yamamoto<br />
�Grad. School of Information Sci. and Technol., Hokkaido Univ., Sapporo 060-0814 Japan, kudo@bme.ist.hokudai.ac.jp�<br />
Sonoporation is a technique for making membrane perforation by exposure of cells to ultrasound, and it is an attractive method for<br />
realizing nonvirus gene transfection. A continuous or quasicontinuous wave is frequently used for this technique because a higher duty<br />
ratio gives higher efficiency of sonoporation. Addition of microbubbles during insonification greatly improves the efficiency of<br />
sonoporation, and, especially when microbubbles exist in the vicinity of the cells, ultrasound pulses from diagnostic ultrasound<br />
equipment can cause sonoporation. In this study, we examined sonoporation induced by single-shot pulsed ultrasound and the role of<br />
microbubbles in induction of cell membrane perforation. Bubble behavior and cell membrane damage were observed using a highspeed<br />
camera and light and scanning electron microscopes. Results of observation showed that mechanical stress induced by bubble<br />
motion could cause cell membrane perforation. We also studied repair of the perforation using a fluorescence microscope and found<br />
that the membrane of mammalian cells has the ability to reseal the perforation within several seconds. �Research partially supported<br />
by a Grant-in-Aid for Scientific Research from the Ministry of Education, Science, Sports and Culture, Japan.�<br />
8:40<br />
4aBB3. Quantitative imaging of tumor blood flow with contrast ultrasound. Peter N. Burns, Raffi Karshafian, and John Hudson<br />
�Dept. Medical Biophys., 2075 Bayview Ave., Toronto ON, M4N 3M5, Canada�<br />
The point at which a solid cancer develops its own blood supply marks the onset of malignant progression. This process, known<br />
as angiogenesis, makes oxygen and nutrients available for growth and provides a path for metastatic spread. Angiogenesis is not only<br />
of interest as a diagnostic hallmark of malignancy, but also as a target for new therapeutic strategies. Assessing antiangiogenic<br />
therapies noninvasively poses problems—flow velocities (�1 mm/s� and vessel diameters (�50 �m� are below resolutions of direct<br />
imaging. Vessels are disorganized without the tree-like structure of normal vasculature. We have investigated the potential role of<br />
3229 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3229<br />
4a FRI. AM
microbubble disruption-replenishment flow measurement in monitoring antivascular treatment of an animal tumor. The currently used<br />
monexponential model incorrectly considers the vasculature a perfect mixing chamber. Simple fractal models of the circulation<br />
provide a distribution of vessel diameters which, combined with the geometry of the disruption and detection beams, produce better<br />
models of replenishment following acoustic bubble disruption. These not only measure flow, but also predicts differences between<br />
organized and disorganized circulations, even with equal flow and vascular volume. If detectable, such differences might be used to<br />
characterize vascular organization below the resolution limit of an ultrasound image.<br />
9:00<br />
4aBB4. Dynamics of laser-trapped microbubbles. Hiroyuki Takahira �Dept. of Mech. Eng., Osaka Prefecture Univ., 1-1<br />
Gakuen-cho, Naka-ku, Sakai, Osaka 599-8531, Japan�<br />
A laser-trapping method is utilized for microbubbles. Bubbles of the order of 10 microns in diameter are trapped and manipulated<br />
successfully using a dry objective lens with large working distance. The growth or shrinkage of a laser-trapped microbubble and the<br />
merger of microbubbles are observed with a high-speed camera to investigate the influence of gas diffusion on the stability of<br />
microbubbles. Two kinds of equilibrium radii are found for shrinking microbubbles. The first one is related to the equilibrium surface<br />
concentration of surfactant. The other is related to the decrease of the surface tension due to the compression of the surface area at the<br />
maximum surfactant concentration. The simulations in which the dynamic surface tension is considered are in good agreement with<br />
the experiments. The laser trapping technique is also applied to the motion of a microbubble in a shear flow. It is shown that the bubble<br />
escapes from the laser trap being repelled by the optical force in the shear flow. There is overall agreement between the experiments<br />
and the simulations in which the buoyancy force, the fluid dynamic forces, and the optical force are taken into account.<br />
9:20<br />
4aBB5. Novel methods of micro-object trapping by acoustic radiation force. Yoshiki Yamakoshi �1-5-1 Tenjin-cho, Kiryushi,<br />
Gunma 376-8515 Japan, yamakosi@el.gunma-u.ac.jp�<br />
It is expected that micro object trapping by acoustic radiation force is a useful method in future drug delivery systems in order to<br />
concentrate the payloads at desired position. In this paper, two novel methods of micro object trapping are presented. First method is<br />
micro object trapping by seed bubbles. This method uses seed bubbles, which have higher sensitivity to the ultrasonic wave, in order<br />
to trap micro objects, which are difficult to trap by conventional methods due to low volumetric oscillation under the ultrasonic wave.<br />
The Bjerkne’s force, which is produced by a secondary wave radiated from the seed bubbles, traps the target objects making bi-layer<br />
seed bubbletarget object mass. The Second method is micro bubble trapping by bubble nonlinear oscillation. Two ultrasonic waves<br />
with different frequencies �pumping and control waves� are introduced simultaneously. The frequency of the control wave is set to a<br />
harmonic frequency of the pumping wave. If the bubbles flow into the cross area of two waves, nonlinear oscillation by high intensity<br />
pumping wave generates the Bjerkne’s force, producing multiple traps with narrow separation along the control wave propagation<br />
direction. In order to demonstrate these methods, experiments using an ultrasonic wave contrast agent are shown.<br />
9:40<br />
4aBB6. Mechanical properties of HeLa cells at different stages of cell cycle by time-resolved acoustic microscope. Pavel V.<br />
Zinin �School of Ocean and Earth Sci. and Technol., Univ. of Hawaii, 2525 Correa Rd., Honolulu, HI 96822-2219�, Eike C. Weiss,<br />
Pavlos Anastasiadis, and Robert M. Lemor �Fraunhofer Inst. for Biomed. Technol., St. Ingbert, Germany�<br />
Scanning acoustic microscopy �SAM�, particularly time-resolved acoustic microscopy, is one of the few techniques for study of<br />
the mechanical properties of only the cell’s interior, cytosol and nucleus. Unfortunately, time-resolved acoustic microscopes typically<br />
do not provide sufficient resolution to study the elasticity of single cells. We demonstrate that the high-frequency, time-resolved<br />
acoustic microscope developed at the Fraunhofer Institute for Biomedical Technology �IBMT�, Germany, is capable of imaging and<br />
characterizing elastic properties of micron size structures in cell’s cytoskeleton with a theoretical resolution limit of 10 m/s for sound<br />
speed measurements. Measurements were performed on cells of the HeLa cell line derived from human cervics carcinoma. SAM<br />
measurements of the sound speed of adherent HeLa cells at different states of the cell cycle were conducted. They yielded an average<br />
value of 1540 m/s. B-Scan images of HeLa cells at different states of the cell cycle show distinct patterns inside the cell. A method<br />
for estimating sound attenuation inside HeLa cells is outlined as such a method is critical for the determination of a cell’s viscoelasticity.<br />
�Work supported by Alexander von Humboldt Foundation and the European Framework Program 6, Project ‘‘CellProm.’’�<br />
10:00–10:10 Break<br />
10:10<br />
4aBB7. Assessment of shock wave lithotripters via cavitation potential. Jonathan I. Iloreta, Andrew J. Szeri �UC Berkeley, 6119<br />
Etcheverry Hall, M.S. 1740, Berkeley, CA 94720-1740�, Yufeng Zhou, Georgii Sankin, and Pei Zhong �Duke Univ., Durham, NC<br />
27708-0300�<br />
An analysis of bubbles in elastic media has been made in order to characterize shock wave lithotripters by gauging the potential<br />
for cavitation associated with the lithotripter shock wave �LSW�. The method uses the maximum radius achieved by a bubble<br />
subjected to a LSW as the key parameter that defines the potential damage a lithotripter could cause at any point in the domain. The<br />
maximum radius is determined by an energy analysis. A new index—similar in spirit to the Mechanical Index of Holland and Apfel<br />
for diagnostic ultrasound—is proposed for use in gauging the likelihood of cavitation damage.<br />
3230 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3230
10:30<br />
4aBB8. Formation of water pore in a bilayer induced by shock wave:<br />
Molecular dynamics simulation. Kenichiro Koshiyama, Takeru Yano,<br />
Shigeo Fujikawa �Lab. of Adv. Fluid Mech., Hokkaido Univ., Sapporo<br />
060-8628, Japan, koshi@ring-me.eng.hokudai.ac.jp�, and Tetsuya<br />
Kodama �Tohoku Univ., Aobaku, Sendai 980-8575, Japan�<br />
The irradiation of a shock wave or ultrasound with micro-bubbles has<br />
the potential to make transient pores on cell membranes. Although such<br />
pores are believed to contribute to the molecular delivery thorough the<br />
membrane, the detailed mechanisms of the pore formation with shock<br />
waves and the subsequent molecular delivery through the pores into cells<br />
are still unclear. To investigate the mechanism at a molecular level, the<br />
molecular dynamics simulations of the interaction of the shock wave with<br />
a lipid bilayer are conducted. The water penetration into the hydrophobic<br />
region by the shock wave is observed in picoseconds. As a next step,<br />
structural changes of the bilayer containing water molecules in the hydrophobic<br />
region are investigated. The water pore is formed in 3 ns when the<br />
large number of water molecules is inserted. The lifetime of the water pore<br />
is more than 70 ns. The radius of the water pore is ca. 1.0 nm, which is<br />
three times larger than the Stoke’s radius of a typical anticancer drug<br />
�5FU�. Finally, the diffusion of the anticancer drug in the water pore is<br />
investigated.<br />
10:45<br />
4aBB9. Ultrasonic spore lysis and the release of intracellular content<br />
in a microfluidic channel. Oana C. Marina, Michael D. Ward, John M.<br />
Dunbar, and Gregory Kaduchak �MPA-11, Los Alamos Natl. Lab., P.O.<br />
Box 1663, MS D-429, Los Alamos, NM, 87545�<br />
Ultrasonic lysis of suspended spores in a microfluidic channel is a<br />
promising alternative to conventional spore disruption techniques that include<br />
bead beating as the spore lysis gold standard. Our overall research<br />
goal is to obtain an automated detection system with complete sample<br />
preparation and lysis steps in a microfluidic channel. Previously, much<br />
work in this area has focused on organism viability rather than the release<br />
of intracellular material. Our research focuses on quantifying the amount<br />
of intracellular content �e.g., DNA, proteins, etc.� that is released by<br />
acoustic lysis for detection by the sensor. Elucidating the efficacy of<br />
acoustics on the release of intracellular material requires reliable methods<br />
to quantify the released intracellular content �nucleic acids and proteins�.<br />
The device used for lysing spores consists of a microfluidic chamber with<br />
one acoustically active wall. The chamber depths are in the range of 100–<br />
200 m. Channels tested in the 70-kHz to 1-MHz frequency range show<br />
that the efficiency of intracellular release depends on the operating frequency<br />
of the device and the properties �concentration, composition� of<br />
the spore suspensions. Experimental results on viability and released intracellular<br />
content are discussed. �Work supported by LANL LDRD.�<br />
11:00<br />
4aBB10. The correlation between cavitation noise power and bubbleinduced<br />
heating in high-intensity focused ultrasound. Caleb H. Farny,<br />
Tianming Wu, R. Glynn Holt, and Ronald A. Roy �Dept. of Aerosp. and<br />
Mech. Eng., Boston Univ., 110 Cummington St., Boston, MA 02215,<br />
cfarny@bu.edu�<br />
It has been established that inertial cavitation is responsible for elevated<br />
heating during high-intensity focused ultrasound �HIFU� application<br />
for certain intensity regimes. The contribution of bubble-induced<br />
heating can be an important factor to consider, as it can be several times<br />
that expected from absorption of the primary ultrasound energy. Working<br />
in agar-graphite tissue phantoms with a 1.1-MHz HIFU transducer, an<br />
embedded type-E thermocouple, and a 15-MHz passive cavitation detector<br />
�PCD�, the temperature and cavitation signal near the focus were measured<br />
for 5-s continuous wave HIFU insonations. The measured temperature<br />
was corrected for heating predicted from the primary ultrasound absorption<br />
and the transient thermocouple viscous heating artifact to isolate<br />
the temperature rise from the bubble activity. We have found that the<br />
Contributed Papers<br />
temperature rise induced from the bubble activity correlates well with the<br />
instantaneous cavitation noise power as indicated by the mean square voltage<br />
output of the PCD. The results suggest that careful processing of the<br />
cavitation signals could serve as a proxy for measuring the heating contribution<br />
from inertial cavitation. �Work supported by the Dept. of the<br />
Army �Award No. DAMD17-02-2-0014� and the Center for Subsurface<br />
Sensing and Imaging Systems �NSF ERC Award No. EEC-9986821�.�<br />
11:15<br />
4aBB11. Membrane permeabilization of adherent cells with laserinduced<br />
cavitation bubbles. Rory Dijkink, Claus-Dieter Ohl �Phys. of<br />
Fluids, Univ. of Twente, Postbus 217, 7500 AE Enschede, The<br />
Netherlands�, Erwin Nijhuis, Sèverine Le Gac �Univ. of Twente, 7500 AE<br />
Enschede, The Netherlands�, and Istvàn Vermes �Medical Spectrum<br />
Twente Hospital Group, 7500 KA Enschede, The Netherlands�<br />
Strongly oscillating bubbles close to cells can cause the opening of the<br />
cell’s membrane, thus to stimulate the uptake of molecules from the exterior.<br />
However, the volume oscillations of bubbles induce complex fluid<br />
flows, especially when bubble-bubble interaction takes place. Here, we<br />
report on an experiment where a single cavitation bubble is created close<br />
to a layer of adherent HeLa cells. The interaction distance between the<br />
bubble and the cell layer is controlled by adjusting the focus of the pulsed<br />
laser light, which creates the cavitation bubble. The dynamics of the<br />
bubble and the cells is recorded with high-speed photography. The poration<br />
of the cells is probed with different fluorescent stains to distinguish<br />
viable and permanent poration and programmed cell death �apopotosis�.<br />
Quantitative data are presented as a function of the radial distance from<br />
the stagnation point. Our main finding is the importance of the asymmetrical<br />
collapse and the high-speed jet flow: After impact of the jet onto the<br />
substrate a strong boundary layer flow is responsible for shearing the cells.<br />
11:30<br />
4aBB12. Antitumor effectiveness of cisplatin with ultrasound and<br />
nanobubbles. Tetsuya Kodama, Yukiko Watanabe, Kiyoe Konno,<br />
Sachiko Horie �Res. Organization, Tohoku Univ., 2-1 Seiryo-machi,<br />
Aoba-ku, Sendai, Miyagi 980-8575, Japan�, Atsuko Aoi �Tohoku Univ.,<br />
Sendai 980-8575, Japan�, Geroges Vassaux �Bart’s and The London<br />
School of Medicine and Dentistry, UK�, and Shiro Mori �Tohoku Univ.<br />
Hospital, Sendai 980-8575, Japan�<br />
The potentiation of antitumor effect of cis-diamminedichloroplatinum<br />
�II�, cisplatin, with ultrasound �1 MHz, 0.6 MPa� and lipid-shelled<br />
nanobubbles in vitro �EMT6, C26, MCF7, A549� and in vivo on s.c. tumor<br />
in mice �HT29-expressing luciferase� were evaluated. In vitro and in vivo<br />
antitumor effects were measured by an MTT assay and a real-time in vivo<br />
imaging, respectively. The effective antitumor effect was seen both in vitro<br />
and in vivo when ultrasound and nanobubbles were used, while other<br />
treatment groups with cisplatin with ultrasound did not show the effectiveness.<br />
The antitumor effect was not attributed to necrosis but apoptosis,<br />
which was confirmed by increase in the activity of the pro-apoptosis signal<br />
caspase-3 and Bax. In conclusion, the combination of ultrasound and<br />
nanobubbles with cisplatin is an effective chemotherapy of solid tumors<br />
and may prove useful in clinical application.<br />
11:45<br />
4aBB13. Sonoporation by single-shot pulsed ultrasound with<br />
microbubbles—Little effect of sonochemical reaction of inertial<br />
cavitation. Kengo Okada, Nobuki Kudo, and Katsuyuki Yamamoto<br />
�Grad. School of Information Sci. and Technol., Hokkaido Univ., Kita 14<br />
Nishi 9, Kita-ku, Sapporo 060-0814, Japan�<br />
Sonoporation is a technique for introducing large molecules into a cell<br />
by exposure to ultrasound, and it has a potential application for gene<br />
transfection. Although continuous-wave ultrasound is generally used for<br />
this technique, we have been using single-shot pulsed ultrasound with<br />
microbubbles. To determine the contribution of the sonochemical effect of<br />
3231 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3231<br />
4a FRI. AM
inertial cavitation under the condition of single-shot exposure, we compared<br />
rates of cell membrane damage in the presence and absence of a free<br />
radical scavenger �cysteamine, 5 mM�. Cells with microbubbles in their<br />
vicinity were exposed to pulsed ultrasound of 1.1 MPa in negative peak<br />
pressure under microscopic observation, and the numbers of total and<br />
damaged cells in the view field were counted. The damage rates were<br />
8.1�4.0% and 10.3�6.3% in the presence (n�17) and absence (n<br />
�25) of the scavenger, respectively, and the average number of total cells<br />
was 772�285. Since there was no significant difference, we concluded<br />
that the cell membrane damage observed in our exposure condition was<br />
not caused by the sonochemical effect but by the mechanical effect of<br />
inertial cavitation. �Research was supported by a grant-in-aid for scientific<br />
research from the Ministry of Education, Science, Sports and Culture,<br />
Japan.�<br />
FRIDAY MORNING, 1 DECEMBER <strong>2006</strong> OAHU ROOM, 8:00 TO 11:50 A.M.<br />
Session 4aEA<br />
Engineering Acoustics and ASA Committee on Standards: Developments in Microphones: Calibrations,<br />
Standards, and Measures<br />
George S. K. Wong, Cochair<br />
National Research Council, Inst. for National Measurement Standards, 1500 Montreal Rd., Ottawa,<br />
Ontario K1A 0R6, Canada<br />
Masakasu Iwaki, Cochair<br />
NHK Science and Technology Research Labs., 1-10-11 Kinuta, Setagaya-ku, Tokyo 157-8510, Japan<br />
Chair’s Introduction—8:00<br />
Invited Papers<br />
8:05<br />
4aEA1. Current developments at the National Institute for Standards and Technology in pressure calibration of laboratory<br />
standard microphones. Victor Nedzelnitsky, Randall P. Wagner, and Steven E. Fick �National Inst. of Standards and Technol.<br />
�NIST�, 100 Bureau Dr., Stop 8220, Gaithersburg, MD 20899-8220, victor.nedzelnitsky@nist.gov�<br />
Current research effort aims at improving the apparatus and methods for determining the pressure sensitivities of IEC types LS1Pn<br />
and LS2aP laboratory standard microphones. Among the improvements that are being systematically incorporated in an evolving test<br />
bed is the capability to operate at adjustable power line frequencies other than the usual 60 Hz. Suitable choices of line frequency<br />
relative to frequencies of calibration and adjustable bandpass filter characteristics can be used to improve the signal-to-noise ratios of<br />
measurements performed near the usual line frequency and its first few harmonics. This can enable the use of relatively large volume<br />
couplers for which uncertainties in microphone front cavity volume and equivalent volume, capillary tube effects, and heat conduction<br />
corrections have a lesser influence than they have for small-volume couplers. Another improvement aims to control and to stabilize the<br />
ambient static pressure during microphone calibrations, to reduce or eliminate the effects of barometric pressure fluctuations on these<br />
calibrations.<br />
8:25<br />
4aEA2. Free-field reciprocity calibration of laboratory standard „LS… microphones using a time selective technique. Knud<br />
Rasmussen and Salvador Barrera-Figueroa �Danish Primary Lab. of Acoust. �DPLA�, Danish Fundamental Metrology, Danish Tech.<br />
Univ., Bldg. 307, 2800 Kgs., Lyngby, Denmark�<br />
Although the basic principle of reciprocity calibration of microphones in a free field is simple, the practical problems are<br />
complicated due to the low signal-to-noise ratio and the influence of cross talk and reflections from the surroundings. The influence<br />
of uncorrelated noise can be reduced by conventional narrow-band filtering and time averaging, while correlated signals like cross talk<br />
and reflections can be eliminated by using time-selective postprocessing techniques. The technique used at DPLA overcomes both<br />
these problems using a B&K Pulse analyzer in the SSR mode �steady state response� and an FFT-based time-selective technique. The<br />
complex electrical transfer impedance is measured in linear frequency steps from a few kHz to about three times the resonance<br />
frequency of the microphones. The missing values at low frequencies are estimated from a detailed knowledge of the pressure<br />
sensitivities. Next an inverse FFT is applied and a time window around the main signal is used to eliminate cross talk and reflections.<br />
Finally, the signal is transformed back to the frequency domain and the free field sensitivities calculated. The standard procedure at<br />
DPLA involves measurements at four distances and the repeatability of the calibrations over time is within �0.03 dB up to about 1.5<br />
times the resonance frequency of the microphones.<br />
3232 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3232
8:45<br />
4aEA3. Microphone calibration by comparison. George S. K. Wong �Acoust. Standards, Inst. for Natl. Measurement Standards,<br />
National Res. Council Canada, Ottawa, ON K1A 0R6, Canada�<br />
The absolute method of microphone calibration by the reciprocity method �IEC 61094-2 1992� provides the highest accuracy of<br />
approximately 0.04 to 0.05 dB, and the procedure requires three changes of microphone in the ‘‘driver-receiver combination’’ that<br />
needs approximately 1 to 2 days. The system capital cost is relatively high. The NRC interchange microphone method for microphone<br />
calibration by comparison has been adopted internationally by the International Electrotechnical Commission as IEC 61094-5 �2001-<br />
10�. With this method, the test microphone is compared with a reference microphone calibrated by the reciprocity method and the<br />
procedure requires approximately 3 h. The uncertainty of the comparison method is between 0.08 to 0.09 dB, which satisfies most<br />
industrial needs.<br />
9:05<br />
4aEA4. Development of a laser-pistonphone for an infrasonic measurement standard. Ryuzo Horiuchi, Takeshi Fujimori, and<br />
Sojun Sato �Natl. Metrology Inst. of Japan �NMIJ�, AIST, Tsukuba Central 3, 1-1-1 Umezono, Tsukuba, 305-8563, Japan�<br />
Acoustical standards for audio frequencies are based on pressure sensitivities of laboratory standard microphones calibrated using<br />
a coupler reciprocity technique. There is a growing need to extend the frequency range downward for reliable infrasonic measurement.<br />
The reciprocity technique, however, has limitations on low-frequency calibration �1–20 Hz� because signal-to-noise ratio deteriorates<br />
and a sound leak occurs from capillary tubes that equalize the static pressure inside and outside of the coupler. These factors rapidly<br />
increase the measurement uncertainty as the frequency is lowered. NMIJ has therefore recently developed a laser-pistonphone<br />
prototype, which enables precise calibration of microphones at low frequencies. Compared with the reciprocity technique, the<br />
laser-pistonphone produces a higher sound pressure within a cavity by the sinusoidal motion of a piston and has a significantly<br />
improved signal-to-noise ratio. Sound pressure is calculated from the piston displacement, which is determined via a Michelson<br />
interferometer. A test microphone is inserted into the cavity, exposed to the sound pressure, and its open-circuit voltage is measured.<br />
Static pressure equalization is realized through the gap between the piston and its guide. Careful design of the dimensions and relative<br />
position of the cavity and piston minimizes sound leakage and friction between them.<br />
9:25<br />
4aEA5. Anechoic measurements of particle-velocity probes compared to pressure gradient and pressure microphones.<br />
Wieslaw Woszczyk �CIRMMT, McGill Univ., 555 Sherbrooke St. West, Montreal, QC, Canada H3A 1E3,<br />
wieslaw@music.mcgill.ca�, Masakazu Iwaki, Takehiro Sugimoto, Kazuho Ono �NHK Sci. & Tech. Res. Labs., Setagaya-ku, Tokyo<br />
157-8510, Japan�, and Hans-Elias de Bree �R&D Microflown Technologies�<br />
Microflown probes are true figure-of-eight-pattern velocity microphones having extended response down to below the lowest<br />
audible frequencies, low noise, and high output. Unlike pressure-gradient microphones, velocity probes do not measure acoustic<br />
pressure at two points to derive a pressure gradient. When particle velocity is present, acoustical particle velocity sensors measure the<br />
temperature difference of the two closely spaced and heated platinum wire resistors, and quantify particle velocity from the temperature<br />
measurement. Microflown probes do not require a membrane and the associated mechanical vibration system. A number of<br />
anechoic measurements of velocity probes are compared to measurements of pressure-gradient and pressure microphones made under<br />
identical acoustical conditions at varying distances from a point source having a wide frequency response. Detailed measurements<br />
show specific response changes affected by the distance to the source, and focus on the importance of transducer calibration with<br />
respect to distance. Examples are given from field work using microflown probes to record acoustic response of <strong>room</strong>s to test signals.<br />
The probe’s cosine directional selectivity can be used to change the ratio between early reflections and the diffuse sound since only<br />
the 1<br />
3 of the power in the diffuse sound field is measured with the particle velocity probe.<br />
9:45<br />
4aEA6. Sensitivity change with practical use of electret condenser microphone. Yoshinobu Yasuno �Panasonic Semiconductor<br />
Device Solutions Co., Ltd. 600, Saedo-cho, Tsuzuki-ku, Yokohama, 224-8539, Japan� and Kenzo Miura �Panasonic Mobile Commun.<br />
Eng. Co., Ltd., Yokohama, Japan�<br />
Dr. Sessler and Dr. West invented the electret condenser microphone �ECM� in 1966. It has since been applied in various ways as<br />
a sound input device. The ECM has become an important component as a microphone for communications because of its stable<br />
sensitivity frequency characteristic. Materials and production methods have been improved continually up to the present. In particular,<br />
the ECM reliability is based on the electret’s stability. For that reason, the electret surface charge decay is the main factor in ECM<br />
sensitivity degradation. This study analyzed the changes of an ECM preserved for 28 years in the laboratory and actually used for an<br />
outdoor interphone unit for 29 years. The change of diaphragm stiffness and electret surface voltage were compared with the<br />
evaluation result of a heat-acceleration test and verified. A degradation estimate of sensitivity change of ECM was performed.<br />
Regarding the life of the electret predicted in the report of former study �K. Miura and Y. Yasuno, J. Acoust. Soc. Jpn. �E� 18�1�,<br />
29–35 �1997��, the validity was verified using actual data from this long-term observation.<br />
3233 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3233<br />
4a FRI. AM
10:05–10:20 Break<br />
10:20<br />
4aEA7. Development of a small size narrow directivity microphone. Masakazu Iwaki, Kazuho Ono, Takehiro Sugimoto �NHK<br />
Sci. & Technol. Res. Labs., 1-10-11 Kinuta, Setagaya-ku, Tokyo, 157-8510, Japan�, Takeshi Ishii, and Keishi Inamaga �Sanken<br />
Microphone Co., Ltd, 2-8-8 Ogikubo, Suginami-ku, Tokyo, 167-0051, Japan�<br />
We developed a new microphone that has very sharp directivity even in the low-frequency band. In an ordinary environment of<br />
sound pick-up, the energy of background noise is distributed mainly in frequencies lower than 1000 Hz. In such frequencies, a typical<br />
cardioid microphone has the directivity pattern close to that of omni-cardiod. Consequently, it is difficult to pick up the objective<br />
sounds clearly from background noises. To suppress the noises with very small level, the directivity pattern should be also sharpened<br />
in the low-frequency band. In this report, we describe a new method to sharpen directivity for the low-frequency band. The method<br />
requires three microphone capsules. One capsule is the main microphone with a very short acoustic pipe. The others compose a<br />
second-order gradient microphone to cancel the signal that comes from behind the main microphone. A special feature of this<br />
microphone is to control a dip frequency of behind sensitivity without changing the frequency response of the front sensitivity.<br />
10:40<br />
4aEA8. Two-wafer bulk-micromachined silicon microphones. Jianmin Miao and Chee Wee Tan �Micromachines Ctr., School of<br />
Mech. and Aerosp. Eng., Nanyang Technolog. Univ., 50 Nanyang Ave., Singapore 639798, mjmmiao@ntu.edu.sg�<br />
A two-wafer concept is proposed for silicon microphone manufacturing by using bulk-micromachining and wafer bonding technologies.<br />
Acoustical holes of the backplate in one wafer are micromachined by deep reactive ion etching and the diaphragm on another<br />
wafer is created by wet-chemical etching. The two wafers are then bonded together to form silicon condenser microphones. In order<br />
to minimize the mechanical-thermal noise and increase the sensitivity within the required bandwidth, an analytical model based on<br />
Zuckerwar’s equations has been developed to find the optimum location of the acoustical holes in the backplate of microphones.<br />
According to our study, this analytical modeling has shown excellent agreement between the simulated and measured results for the<br />
B&K MEMS microphone. Silicon condenser microphones have been further optimized in terms of the air gap, number and location<br />
of acoustical holes to achieve the best performance with a low polarization voltage, and easy fabrication for possible commercialization.<br />
Details of analytical modeling, fabrication, and measurement results will be presented.<br />
11:00<br />
4aEA9. Infrasound calibration of measurement microphones. Erling Frederiksen �Bruel & Kjaer, Skodsborgvej 307, 2850<br />
Naerum, Denmark, erlingfred@bksv.com�<br />
Increasing interest for traceable infrasound measurements has caused the Consultative Committee for Acoustics, Ultrasound and<br />
Vibration �CCAUV� of BIPM to initiate a key comparison calibration project �CCAUV.A-K2� on pressure reciprocity calibration<br />
down to 2 Hz. Ten national metrology institutes, including the Danish Primary Laboratory of Acoustics �DPLA�, take part in this<br />
project. In addition DPLA has started its own infrasound calibration project, which is described in this paper. The purposes of this<br />
project are verification of the CCAUV results and development of methods for calibration of general types of measurement microphone<br />
between 0.1 and 250 Hz. The project includes the design of an active comparison coupler, an experimental low-frequency<br />
reference microphone, and new methods for its frequency response calibration. One method applies an electrostatic actuator and<br />
requires a low-pressure measurement tank, while the other requires an additional microphone, whose design is closely related to that<br />
of the reference microphone that is to be calibrated. The overall calibration uncertainty (k�2) for ordinary measurement microphones<br />
is estimated to less than 0.05 dB down to 1 Hz and less than 0.1 dB down to 0.1 Hz, if the reference is calibrated in the latter<br />
mentioned way, i.e., by the related microphones method.<br />
11:20<br />
4aEA10. Free-field calibration of 1Õ4 inch microphones for ultrasound<br />
by reciprocity technique. Hironobu Takahashi, Takeshi Fujimori,<br />
Ryuzo Horiuchi, and Sojun Sato �Natl. Metrology Inst. of Japan, AIST,<br />
Tsukuba Central 3, 1-1-1 Umezono, Tsukuba, 305-8563 Japan�<br />
Recently, equipment that radiates ultrasound radiation at frequencies<br />
far beyond the audible range is increasing in our environment. Such electronic<br />
equipment has switching regulators or inverter circuits, and many<br />
devices are unintended sources of ultrasound radiation. However, the effects<br />
of airborne ultrasound on human hearing and the human body have<br />
not been well investigated. To estimate the potential damage of airborne<br />
ultrasound radiation quantitatively, it is necessary to establish an acoustic<br />
standard for airborne ultrasound because the standard is a basis of acoustic<br />
measurement. With the intention of establishing a standard on airborne<br />
ultrasound, a free-field calibration system with an anechoic chamber was<br />
produced. The principle of free-field calibration techniques is introduced<br />
in this presentation. Type WS3 microphones �B&K 4939� were calibrated<br />
in the system to examine the calibration ability to be achieved. Results<br />
Contributed Papers<br />
showed that it can calibrate a microphone from 10 to 100 kHz with dispersion<br />
of less than 1 dB. In addition, the effects that were dependent on<br />
the uncertainty of the calibration are discussed based on those results.<br />
11:35<br />
4aEA11. An environmentally robust silicon diaphragm microphone.<br />
Norihiro Arimura, Juro Ohga �Shibaura Inst. of Technol.,3-7-5 Toyosu,<br />
Koto-ku, Tokyo, 135-8548, Japan�, Norio Kimura, and Yoshinobu Yasuno<br />
�Panasonic Semiconductor Device Solutions Co., Ltd., Saedo-cho,<br />
Tsuzuki-ku, Yokohama, Japan�<br />
Recently, many small microphones installed in cellular phones are the<br />
electret condenser microphones �ECMs� that contain an organic film diaphragm.<br />
Although FEP of fluorocarbon polymer is generally used as the<br />
electret material, silicon dioxide is also used. Recently ECMs have been<br />
made small and thin while maintaining the basic sound performance according<br />
to the market demand. In addition, environment tests and the<br />
reflow soldering mounting process have been adjusted to meet market<br />
requirements. On the other hand, the examination satisfied the demand as<br />
3234 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3234
the high temperature resistance was insufficient. This paper depicts an<br />
examination and a comparison of conventional ECM with the experimental<br />
model, a silicon diaphragm condenser microphone produced using the<br />
MEMS method. The silicon diaphragm satisfies high-temperature resis-<br />
tance and stable temperature characteristics because of its very small coefficient<br />
of linear expansion and it is measured total harmonic distortion<br />
�THD� on high pressure sound. Finally, it will be able to be used in high<br />
temperature and high pressure sound conditions in the future.<br />
FRIDAY MORNING, 1 DECEMBER <strong>2006</strong> IAO NEEDLE/AKAKA FALLS ROOM, 8:15 TO 11:15 A.M.<br />
Session 4aMU<br />
Musical Acoustics: Music Information and Communication<br />
Bozena Kostek, Cochair<br />
Gdansk Univ. of Technology, Multimedia Systems Dept., Narutowicza 11- 12, 80-952 Gdansk, Poland<br />
Masuzo Yanagida, Cochair<br />
Doshisha Univ., Dept. of Information Science and Intelligent Systems, 1-3 Tatara-Miyakodani, Kyo-Tanabe,<br />
Kyoto 610-0321, Japan<br />
Invited Papers<br />
8:15<br />
4aMU1. Introduction of the Real World Computing music database. Masataka Goto �Natl. Inst. of Adv. Industrial Sci. and<br />
Technol. �AIST�, 1-1-1 Umezono, Tsukuba, Ibaraki 305-8568, Japan, m.goto@aist.go.jp�<br />
This paper introduces the RWC (Real World Computing) Music Database, a copyright-cleared music database that is available to<br />
researchers as a common foundation for research. Shared databases are common in other research fields and have contributed<br />
importantly to progress in those fields. The field of music information processing, however, has lacked a common database of musical<br />
pieces and a large-scale database of musical instrument sounds. The RWC Music Database was therefore built in fiscal 2000 and 2001<br />
as the world’s first large-scale music database compiled specifically for research purposes. It contains six original collections: the<br />
Popular Music Database �100 pieces�, Royalty-Free Music Database �15 pieces�, Classical Music Database �50 pieces�, Jazz Music<br />
Database �50 pieces�, Music Genre Database �100 pieces�, and Musical Instrument Sound Database �50 instruments�. To address<br />
copyright issues, all 315 musical pieces were originally composed, arranged, or performed, and all instrumental sounds were originally<br />
recorded. The database has already been distributed to more than 200 research groups and is widely used. In addition, a<br />
continuous effort has been undertaken to manually annotate a set of music-scene descriptions for the musical pieces, called AIST<br />
Annotation, which consists of the beat structure, melody line, and chorus sections.<br />
8:35<br />
4aMU2. Japanese traditional singing on the same lyrics. Ichiro Nakayama �Osaka Univ. of Arts, 469, Higashiyama, Kanan-cho,<br />
Minami-Kawachi-gun, Osaka, 585-8555 Japan� and Masuzo Yanagida �Doshisha Univ., Kyo-Tanabe, 610-0321 Japan�<br />
Described is a database of Japanese traditional singing together with supplementary recording of Bel Canto for comparative<br />
studies. Singing sounds and spoken speech by the same singers are recorded in pair to form the body of the database. This database<br />
covers most of genres of Japanese traditional singing, such as Shinto prayers, Buddist prayers, Nor, Kyogen, Heikyoku, Sokyoku,<br />
Gidayu-bushi, Kabuki, Nagauta, Tokiwazu, Kiyomoto, Itchu-bushi, Shinnai, Kouta, Zokkyoku, Rokyoku, Shigin, Ryukyu-clasico,<br />
Goze-uta, etc. All the sounds were recorded in anechoic chambers belonging to local institutions, mainly in Osaka and Tokyo, asking<br />
78 professional singers including 18 ‘‘Living National Treasures’’ to sing as informants. The most important point of this database is<br />
that an original lyric especially prepared for this recording is commonly used to make comparative studies easy. All the subjects are<br />
asked to sing the common lyrics in their own singing styles. Shown here are comparisons of formant shifts in vowels from ordinary<br />
speaking to singing for some singers, and comparison of temporal features of fundamental frequency between Japanese traditional<br />
singing and Western Bel Canto. �Work supported by the Academic Frontier Project, Doshisha University.�<br />
8:55<br />
4aMU3. Computational intelligence approach to archival musical recordings. Andrzej Czyzewski, Lukasz Litwic, and<br />
Przemyslaw Maziewski �Gdansk Univ. of Technol., Narutowicza 11/12, 80-952 Gdansk, Poland�<br />
An algorithmic approach to wow defect estimation in archival musical recordings is presented. The wow estimation is based on the<br />
simultaneous analysis of many sinusoidal components, which are assumed to depict the defect. The rough determination of sinusoidal<br />
components in analyzed musical recording is performed by standard sinusoidal modeling procedures employing magnitude and phase<br />
spectra analysis. Since archival recordings tend to contain distorted tonal structure, the basic sinusoidal modeling approach is often<br />
found insufficient, resulting in audible distortions in the restored signal. It is found that the standard sinusoidal modeling approach is<br />
prone to errors, especially when strong frequency or amplitude variations of sinusoidal components occur. It may result in gaps or<br />
inappropriately matched components, leading to incorrect estimation of the wow distortion. Hence, some refinements to sinusoidal<br />
component analysis including interpolation and extrapolation of tonal components are proposed. As it was demonstrated in experi-<br />
3235 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3235<br />
4a FRI. AM
ments, due to the nonlinear nature of wow distortion, the enhancement of sinusoidal analysis can be performed by means of a neural<br />
network. The paper demonstrates implemented algorithms for parasite frequency modulation in archival recordings together with<br />
obtained results. �Work supported by the Commission of the European Communities, within the Integrated Project No. FP6-507336:<br />
PRESTOSPACE.�<br />
9:15<br />
4aMU4. Music information retrieval seen from the communication technology perspective. Bozena Kostek �Gdansk Univ. of<br />
Technol., Narutowicza 11/12, PL-80-952 Gdansk, Poland�<br />
Music information retrieval �MIR� is a multidiscipline area. Within this domain one can see various approaches to musical<br />
instrument recognition, musical phrase classification, melody classification �e.g., query-by-humming systems�, rhythm retrieval,<br />
high-level-based music retrieval such as looking for emotions in music or differences in expressiveness, and music search based on<br />
listeners’ preferences. One may also find research that tries to correlate low-level descriptor analysis to high-level human perception.<br />
Researchers from musical acoustics, musicology, and music domains on one side, and communication technology on the other side,<br />
work together within this area. This may foster a framework for broader and deeper comprehension of contributions from all these<br />
disciplines and, in addition, translate the automated access to music information, gathered in various forms around the World Wide<br />
Web, as a fully understandable process to all participants regardless of their background. The semantic description is becoming a basis<br />
of the next Web generation. Several important concepts have been introduced recently by the researchers associated with the MIR<br />
community with regard to semantic data processing including techniques for computing with words. In this presentation some aspects<br />
related to MIR are briefly reviewed in the context of possible and actual applications of ontology-based approach to this domain.<br />
9:35<br />
4aMU5. Accompaniment included song waveform retrieval based on framewise phoneme recognition. Yuichi Yaguchi and<br />
Ryuichi Oka �Univ. of Aizu, Tsuruga, Ikkimachi, Aizuwakamatsu, Fukushima, 965-8580 Japan�<br />
A novel approach is presented for a retrieval method that is useful for waveforms of songs with accompaniment. Audio signals of<br />
songs have some different acoustical characteristics from speech signals. Furthermore, the length per mora of signals is longer than<br />
that of speech. Therefore, the authors suggest a sound retrieval system for application to musical compositions, including songs, that<br />
extracts framewise acoustical characteristics and uses a retrieval method for absorbing phoneme length. First, the system prepares two<br />
sets of phoneme identification functions that have corresponding order, but for which phoneme sets belong to different environments<br />
of accompaniment-included or accompaniment-reduced. Next, musical compositions are put into database and the query song wave<br />
converts a waveform to a label sequence using framewise phoneme recognition derived by Bayesian estimation that applies each<br />
phoneme identification function according to whether it is accompaniment-included or not. Finally, the system extracts an interval<br />
area, such as query data, from a database using spotting recognition that is derived using continuous dynamic programming �CDP�.<br />
Retrieval method results agree well with earlier results �Y. Yaguchi and R. Oka, AIRS2005, LNCS3689, 503–509 �2005�� that applied<br />
the same musical composition set without accompaniment.<br />
9:55–10:10 Break<br />
10:10<br />
4aMU6. Design of an impression-based music retrieval system. Kimiko Ohta, Tadahiko Kumamoto, and Hitoshi Isahara �NICT,<br />
Keihanna Kyoto 619-0289, Japan, kimiko@nict.go.jp�<br />
Impression-based music retrieval helps users to find musical pieces that suit their preferences, feelings, or mental states from<br />
among a huge volume of a music database. Users are asked to select one or more pairs of impression words from among multiple pairs<br />
that are presented by the system and to estimate each selected pair on a seven-step scale to input their impressions into the system. For<br />
instance, if they want to locate musical pieces that will create a happy impression, they should check the radio button ‘‘Happy’’ in the<br />
impression scale: Very happy–Happy–A little happy–Neutral–A little sad–Sad–Very sad. A pair of impression words with a sevenstep<br />
scale is called an impression scale in this paper. The system calculates the distance between the impressions of each musical piece<br />
in a user-specified music database and the impressions that are input by the user. Subsequently, it selects candidate musical pieces to<br />
be presented as retrieval results. The impressions of musical pieces are expressed numerically by vectors that are generated from a<br />
musical piece’s pitch, strength, and length of every tone using n-gram statistics.<br />
10:30<br />
4aMU7. Automatic discrimination between singing and speaking<br />
voices for a flexible music retrieval system. Yasunori Ohishi, Masataka<br />
Goto, Katunobu Itou, and Kazuya Takeda �Grad. School of Information<br />
Sci., Nagoya Univ., Furo-cho 1, Chikusa-ku, Nagoya, Aichi, 464-8603,<br />
Japan, ohishi@sp.m.is.nagoya-u.ac.jp�<br />
This paper describes a music retrieval system that enables a user to<br />
retrieve a song by two different methods: by singing its melody or by<br />
saying its title. To allow the user to use those methods seamlessly without<br />
changing a voice input mode, a method of automatically discriminating<br />
between singing and speaking voices is indispensable. We therefore first<br />
investigated measures that characterize differences between singing and<br />
Contributed Papers<br />
speaking voices. From subjective experiments, we found that human listeners<br />
discriminated between these two voices with 70% accuracy for<br />
200-ms signals. These results showed that even short-term characteristics<br />
such as the spectral envelope represented as MFCC can be used as a<br />
discrimination cue, while the temporal structure is the most important cue<br />
when longer signals are given. According to these results, we then developed<br />
the automatic method of discriminating between singing and speaking<br />
voices by combining two measures: MFCC and an F0 �voice pitch�<br />
contour. Experimental results with our method showed that 68.1% accuracy<br />
was obtained for 200-ms signals and 87.3% accuracy was obtained<br />
for 2-s signals. Based on this method, we finally built a music retrieval<br />
system that can accept both singing voices for the melody and speaking<br />
voices for the title.<br />
3236 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3236
10:45<br />
4aMU8. Various acoustical aspects of an Asian „South… Indian<br />
classical music concert. M. G. Prasad �Dept. of Mech. Eng., Stevens<br />
Inst. of Technol., Hoboken, NJ 07030�, V. K. Raman �Flautist,<br />
Germantown, MD 20874�, and Rama Jagadishan �Edison, NJ 08820�<br />
An Asian �South� Indian classical music concert is an integrated acoustical<br />
experience for both the audience and the player�s�. A typical concert<br />
team, either vocal or instrumental, consists of a main vocalist �or an instrumentalist�<br />
accompanied by a violinist, up to three percussion instrument<br />
players, and a reference drone. The concert is comprised of many<br />
songs. Each song has two main parts, namely Alapana and Kriti. The<br />
Alapana is an elaboration of a raga �tune� and the Kriti refers to the lyrics<br />
of the song. The violinist actively follows and supports the main musician<br />
during the concert. The percussion player�s� are provided an opportunity<br />
to present a solo of their rhythmic skills. The players and the audience<br />
communicate emotionally and intellectually with each other. Elements<br />
such as aesthetics, rhythm, skill, and emotional aspects of the players are<br />
evaluated and appreciated by the audience. This talk will present various<br />
aspects of a concert that brings about an integrated and holistic experience<br />
for both the audience and the player�s�. Some samples from live vocal and<br />
instrumental music concerts will be presented.<br />
11:00<br />
4aMU9. Musical scales, signals, quantum mechanics. Alpar Sevgen<br />
�Dept. of Phys., Bogazici Univ., Bebek 34342, Istanbul, Turkey�<br />
Scales, being finite length signals, allow themselves to be treated algebraically:<br />
key signatures are related to the ‘‘ring’’ property of the scale<br />
labels; cyclically permuted scales and their mirror images have the same<br />
number of sharps and flats; and complementary scales �like major and<br />
pentatonic� have their sharp and flat numbers exchanged. A search for<br />
minimum principles to select among all possible scales those employed in<br />
music yields two possibilities: �a� minimize total number of accidentals<br />
and �b� minimize frequency fluctuations in a scale. Either of these minimum<br />
principles helps filter those scales employed in music from the universe<br />
of all scales, setting up very different criteria than the harmonic<br />
ratios used by musicians. The notes of the scales employed in music seem<br />
to prefer to stay as far apart from each other as possible. Operators that<br />
step through the multiplet members of scales with N semitones form a<br />
complete set of operators together with those that step through their eigenvectors.<br />
The mathematics reveals the discrete Fourier transformations<br />
�DFT� and is identical to finite state quantum mechanics of N-level Stern-<br />
Gerlach filters worked out by J. Schwinger.<br />
FRIDAY MORNING, 1 DECEMBER <strong>2006</strong> MAUI ROOM, 7:30 A.M. TO 12:15 P.M.<br />
Session 4aNS<br />
Noise and Architectural Acoustics: Soundscapes and Cultural Perception I<br />
Brigitte Schulte-Fortkamp, Cochair<br />
Technical Univ. Berlin, Inst. of Technical Acoustics, Secr TA 7, Einsteinufer 25, 10587 Berlin, Germany<br />
Bennett M. Brooks, Cochair<br />
Brooks Acoustics Corp., 27 Hartford Turnpike, Vernon, CT 06066<br />
Invited Papers<br />
7:30<br />
4aNS1. Soundscape in the old town of Naples: Signs of cultural identity. Giovanni Brambilla �CNR Istituto di Acustica ‘‘O.M.<br />
Corbino’’ Via del Fosso del Cavaliere 100, 00133 Roma, Italy�, Luigi Maffei, Leda De Gregorio, and Massimiliano Masullo �Second<br />
Univ. of Naples, 81031 Aversa �Ce�, Italy�<br />
Like all cities in Magna Grecia, the ancient Neapolis was built along three main parallel, tight, and straight streets called<br />
decumani. Since then and during the following centuries, commercial and handicraft activities, as well as social life, have been<br />
developed along these streets. The narrow ground <strong>room</strong>s forced shopkeepers to occupy the main street to show their merchandise<br />
using vocal appeals to magnify their product, and handicrafts to work directly on the street �hammering, sawing, etc.�. Music artists<br />
had their performance on the streets too. The soundscape in the area was a strong symbol of the Neapolitan cultural identity.<br />
Nowadays decumani have kept the main features of the past but some of these are overrun by road traffic. To investigate in which way<br />
the traffic noise has modified the soundscape perception and cultural identity, sound walks were registered during day and night time.<br />
A number of residents were interviewed and laboratory listening tests were carried out. Despite the congested urban environment and<br />
high sound levels, preliminary results have shown that part of the residential population is still able to identify the soundscape more<br />
related to Neapolitan historical identity.<br />
7:50<br />
4aNS2. Soundscape design in public spaces: Concept, method, and practice. Hisao Nakamura-Funaba and Shin-ichiro Iwamiya<br />
�Kyushu Univ., 4-9-1.Shiobaru, Minami-ku, Fukuoka 815-8540, Japan�<br />
Soundscape design of public spaces necessitates consideration of whether a space has important meaning for the user. It is<br />
essential to imagine an ideal sound environment of that space. We designed an actual soundscape from the viewpoint of the<br />
environment, information, and decoration. In many cases, producing some silence in an environment becomes the first step of<br />
soundscape design. There is neither a special technology nor a technique when designing. It merely requires use of a general<br />
technology and techniques concerning sound according to location. A key point is knowledge of know how to coordinate these<br />
technologies and techniques. For instance, silence was made first at the renewal project of Tokyo Tower observatory through<br />
cancellation of its commercial and call broadcasting functions and installation of sound-absorbing panels to the ceiling. Next, suitable<br />
and small sounds were added at various points. Guests can take time to enjoy viewing as a result.<br />
3237 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3237<br />
4a FRI. AM
8:10<br />
4aNS3. The daily rhythm of soundscape. Brigitte Schulte-Fortkamp �TU-Berlin, Einsteinufer 25 TA 7, 10587 Germany,<br />
brigitte.schulte-fortkamp@tu-berlin.de� and Andr Fiebig �HEAD acoustics GmbH, 52134 Herzogenrath, Germany�<br />
With respect to people’s minds, soundscapes can be considered as dynamic systems characterized by the time-dependent occurrence<br />
of particular sound events embedded in specific environments. Therefore, an adequate evaluation of environmental noise will<br />
reflect the continually varying acoustical scenery and its specific perception. An acoustical diary shall provide information about the<br />
daily routine and subjectively perceived sound exposure of residents. It relies on cognitive and emotional aspects of perceiving and<br />
evaluating sounds. It gives insight into evaluating processes and their contextual parameters because of its spontaneous character. It<br />
includes and refers to long-term acoustic measurements. Simultaneous measurements for elaborate acoustical analyses will be taken<br />
outside the homes regarding the moments of essential sound events. The aim is to collect information about the daily rhythm regarding<br />
acoustic events, whereby the focus also should be placed on sound events that are well accepted to deepen the explication of data with<br />
respect to the analysis. Procedure and results will be discussed.<br />
8:30<br />
4aNS4. Ecological explorations of soundscapes: From verbal analysis to experimental settings. Daniele Dubois �CNRS, 11 rue<br />
de Lourmel, 75015 Paris, France� and Catherine Guastavino �McGill Univ., Montreal, QC H3A 1Y1, Canada�<br />
Scientific studies rely on rigorous methods that must be adapted to the object of study. Besides integrating acoustic features,<br />
soundscapes as complex cognitive representations also have the properties of being global, meaningful, multimodal, and categorical.<br />
Investigating these specificities, new paradigms were developed involving linguistics and ecological psychology to complement the<br />
psychophysical approach: cognitive linguistic analyses of discourse to address semantic properties of soundscapes, and categorization<br />
tasks and distances from prototypes to investigate their cognitive organization. As a methodological consequence, experimental<br />
settings must be designed to ensure the ecological validity of the stimuli processing, �the ‘‘realism’’ evaluated from a psychological<br />
point of view, stimuli being processed as in a real-life situation�. This point will be illustrated with perceptual evaluations of spatial<br />
auditory displays for soundscape reproduction. Data processing techniques should also take into consideration the intrinsic properties<br />
of the representations they account for. Examples of free-sorting tasks will be presented with measurements in terms of family<br />
resemblance of sets of properties defining categories rather than dimensional scales. New ways of coupling physical measurement and<br />
psychological evaluations will be presented in order to simulate or reproduce soundscapes in both a realistic and controlled manner for<br />
experimental purposes.<br />
8:50<br />
4aNS5. Artificial neural network models of sound signals in urban open spaces. Lei Yu and Jian Kang �School of Architecture,<br />
Sheffield Univ., Western Bank, Sheffield S10 2TN, UK�<br />
Sound signals, known as foreground sounds, are important components of soundscape in urban open spaces. Previous studies in<br />
this area have shown that sound preferences are different according to social and demographic factors of individual users �W. Yang<br />
and J. Kang, J. Urban Des., 10, 69–88�2005��. This study develops artificial neural network �ANN� models of sound signals for<br />
architects at design stage, simulating subjective evaluation of sound signals. A database for ANN modeling has been established based<br />
on large-scale social surveys in European and Chinese cities. The ANN models have consequently been built, where individual’s social<br />
and demographic factors, activities, and acoustic features of the space and sounds are used as input variables while the sound<br />
preference is defined as the output. Through the process of training and testing the ANN models, considerable convergences have been<br />
achieved, which means that the models can be applied as practical tools for architects to design sound signals in urban open spaces,<br />
taking the characteristics of potential users into account. Currently ANN models combining foreground and background sounds are<br />
being developed.<br />
9:10<br />
4aNS6. Describing soundscape and its effects on people where soundscape is understood as an expansion of the concept of<br />
noise engineering. Keiji Kawai �Grad. School of Sci. and Technol., Kumamoto Univ., 2-39-1 Kurokami, Kumamoto 860-8555,<br />
Japan, kawai@arch.kumamoto-u.ac.jp�<br />
This study discusses how to describe sound environment and people in terms of ‘‘soundscape’’ as an expansion of the ‘‘noise<br />
engineering.’’ In the framework of the conventional study field of noise evaluation, typically, sound environments are represented by<br />
loudness-based indices such as A-weighted sound pressure levels, and the impact of sound environments on people is represented by<br />
annoyance response or some physiological metrics. In the case of soundscape studies, however, the description should be expanded<br />
beyond what has been used in noise engineering. This matter has already been frequently discussed, but it doesn’t seem that much<br />
consensus has been achieved concerning it yet. With respect to the effects of sound environment on people, since the concept of<br />
soundscape focuses on personal and social meanings of environmental sounds including the historical or aesthetic contexts, the effects<br />
are considered to be represented not by a singular concept such as comfortableness or quietness, but by multiple dimensions of<br />
emotional and aesthetic concepts. Also, descriptions of sound environment should include some qualitative aspects, such as what types<br />
of sounds can be heard at what extent. In this paper, the methodology to describe human-soundscape relationships is discussed through<br />
a review of related studies.<br />
3238 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3238
9:30<br />
4aNS7. The sound environmental education aided by automated bioacoustic identification in view of soundscape recognition.<br />
Teruyo Oba �Natural History Museum & Inst., Chiba, 955-2 Aoba-cho, Chuo-ku, Chiba-shi, Chiba-ken 260-8682 Japan,<br />
oba@chiba-muse.or.jp�<br />
From the 2003–2004 JST projects, where the automated bioacoustic identification device Kikimimi-Zukin was introduced to the<br />
nature observation and environmental studies, it was learned that the activities encouraged children to take notice of sounds, become<br />
aware of the sound environment, and gain an insight into the soundscape. Sounds are often riddles to us, and hearing is the process<br />
to find out the facts and causes. It is more important to let children obtain appropriate clues to differentiate sounds by hearing and<br />
thinking for themselves than give them an immediate answer. Program on the Strength of Hearing was formulated to challenge<br />
children and have them enjoy hearing to identify, sharing what they hear with others, and observing environment through sounds.<br />
Kikimimi-Zukin reinforced the program by a step- by-step guide through the hearing process of scanning, focusing, characterizing,<br />
associating with relevant factors, and judging the identity. The experience not only brought them confidence in hearing but incentive<br />
to study nature and environment. With Kikimimi-Zukin children collected recordings and relevant information. Using the sound<br />
database, the local singing map and three-dimensional sound map were prepared. They facilitated communication on the local sound<br />
environment among children and with adults, leading to realization of their inner soundscape.<br />
9:50<br />
4aNS8. Acoustic environmental problems at temporary shelters for victims of the Mid-Niigata Earthquake. Koji Nagahata,<br />
Norio Suzuki, Megumi Sakamoto, Fuminori Tanba �Fukushima Univ., Kanayagawa 1, Fukushima City, Fukushima, 960-1296, Japan,<br />
nagahata@sss.fukushima-u.ac.jp�, Shin-ya Kaneko, and Tetsuhito Fukushima �Fukushima Medical Univ., Fukushima, 960-1295,<br />
Japan�<br />
An earthquake on 23 October 2004 inflicted heavy damage on the Mid-Niigata district. The earthquake isolated Yamakoshi village;<br />
consequently, all the village residents were forced to evacuate to temporary shelters in neighboring Nagaoka city for 2 months. Two<br />
types of temporary shelters were used: gymnasiums, and buildings with large separated <strong>room</strong>s similar to community centers. A<br />
questionnaire survey and interviews (N�95) were conducted to elucidate problems of the living environment at the temporary<br />
shelters. This study analyzed acoustic environmental problems there. Noise-related problems were noted by 40 respondents �46.5%�:<br />
they were the fifth most frequently cited environmental problems. Several serious complaints, e.g., general annoyance at the shelters<br />
and footsteps of refugees at night, were only indicated by respondents who had evacuated to the gymnasiums. However, some<br />
problems, e.g., the clamor of children, including crying babies and voices of other refugees, were indicated by respondents irrespective<br />
of the type of the shelters to which they had been evacuated. Therefore, buildings like community centers were more desirable for<br />
temporary shelters, at least from the perspective of noise problems.<br />
10:10–10:30 Break<br />
10:30<br />
4aNS9. The burden of cardiovascular diseases due to road traffic noise. Wolfgang Babisch �Dept. of Environ. Hygiene, Federal<br />
Environ. Agency, Corrensplatz 1, 14195 Berlin, Germany, wolfgang.babisch@uba.de� and Rokho Kim �WHO/EURO Ctr. for<br />
Environment and Health, 53113 Bonn, Germany�<br />
Epidemiological studies suggest a higher risk of cardiovascular diseases, including high blood pressure and myocardial infarction,<br />
in subjects chronically exposed to high levels of road or air traffic noise. A new meta-analysis was carried out to assess a doseresponse<br />
curve, which can be used for a quantitative risk assessment and to estimate the burden of cardiovascular disease attributable<br />
to environmental noise in European regions. Noise exposure was grouped according to 5 dB�A�-categories for the daytime outdoor<br />
average A-weighted sound pressure level, (L day ,16h:6–22h�, which was considered in most studies. Information on night-time<br />
exposure (L night ,8h:22–6hor23–7h� was seldom available. However, approximations can be made with respect to L den according<br />
to the European directive on the assessment and management of environmental noise. The strongest evidence of an association<br />
between community noise and cardiovascular endpoints was found for ischaemic heart diseases, including myocardial infarction and<br />
road traffic noise. The disability-adjusted life years lost for ischemic heart disease attributable to transport noise were estimated<br />
conservatively, assuming the same exposure patterns across the countries with an impact fraction 3% in the western European<br />
countries.<br />
10:50<br />
4aNS10. Soundscape, moderator effects, and economic implications. Cay Hehner �Henry George School of Social Sci., 121 E.<br />
30th St., New York, NY 10016, chehner.hengeoschool@att.net� and Brigitte Schulte-Fortkamp �TU-Berlin, Berlin, Germany�<br />
Soundscape is considered with respect to moderator effects and the contribution of economics. It will be questioned whether<br />
soundscapes can work as a moderator concerning noise annoyance. As shown by the different investigations concerning soundscapes,<br />
a definition of the meaning of soundscapes is necessary. Evidently, the moderating effect of a given environment and its soundscape<br />
has to be discussed on three levels: �1� extension of factors that describe annoyance,�2� peculiar feature of burdensome noise contexts,<br />
and �3� discrepancies of the social and economic status of people living in areas where the rebuilding will change the quality of the<br />
area. It has to be determined and analyzed to what extent the Georgist method of resources taxation, as recently exemplified, e.g., in<br />
Alaska and in Wyoming, can be instrumental in funding soundscapes to moderate noise annoyance as it has been the case in funding<br />
free education and allowing the distribution of a citizen’s dividend.<br />
3239 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3239<br />
4a FRI. AM
11:10<br />
4aNS11. Socio-cultural soundscape concepts to support policies for managing the acoustic environment. Michiko So Finegold,<br />
Lawrence S. Finegold �Finegold & So, Consultants, 1167 Bournemouth Court Centerville, OH 45459-2647, m-so@pb3.so-net.ne.jp�,<br />
and Kozo Hiramatsu �Kyoto Univ., Sakyou-ku, Kyoto 606-8501 Japan�<br />
In the past half-century, considerable effort has been invested in the academic, technological, and political arenas to achieve an<br />
adequate acoustic environment. Various national and international policy guidance documents have made reasonable progress in<br />
establishing a framework for a common approach to minimizing environmental noise, such as documents from various national<br />
Environmental Protection Agencies, the World Health Organization, and the European Union. Although these documents have provided<br />
useful information for global application, they only address minimizing the negative side of the acoustic environment �i.e.,<br />
noise�, they focus primarily on acoustics issues at the national or international level, and they still have not adequately considered<br />
implementation issues related to socio-cultural differences. To deal with the practical problems that exist in the acoustic environment<br />
in the context of different cultures, continuing research and new policy guidance are needed to address different local situations and<br />
in a variety of cultural contexts. The Soundscape approach has been developing tools for describing the acoustic environment at the<br />
local level to address both the positive and negative aspects of the acoustic environment. In this paper, the evolving interdisciplinary<br />
aspect of the Socio-Cultural Soundscape will be discussed and key topics for future work will be recommended.<br />
11:30<br />
4aNS12. Initial steps for the determination of environmental noise<br />
quality—The perception-related evaluation of traffic noise. Klaus<br />
Genuit, Sandro Guidati, Sebastian Rossberg, and Andr Fiebig �HEAD<br />
Acoust. GmbH, Ebertstrasse 30a, 52134 Herzogenrath, Germany,<br />
klaus.genuit@head-acoustics.de�<br />
Directives call for actions against noise pollution and noise annoyance.<br />
But, how do we eliminate harmful effects including annoyance due to the<br />
exposure of environmental noise without understanding the perception and<br />
evaluation of environmental noise? How do we preserve environmental<br />
noise quality where it is good �Directive 2002/49/EC� without identifying<br />
descriptors for noise quality? Various soundscape approaches based on<br />
different methodologies have been developed in the past. But, the measurement<br />
procedures must be realizable without much effort in order to<br />
achieve acceptance from legislation. Therefore, procedures have to be developed<br />
that capture the complexity of human hearing, on the one hand,<br />
and are feasible with respect to economy and time, on the other hand. The<br />
European project Quiet City �6FP PL516420� is dealing with, among other<br />
aspects, vehicle pass-by noise as a typical environmental noise source and<br />
its evaluation. Results of the analysis based on subjective assessments and<br />
psychoacoustic analyses carried out with respect to the development of an<br />
annoyance index will be presented and discussed. Such an index will<br />
Contributed Papers<br />
Contributed Poster Paper<br />
provide valuable information for effective improvement of noise quality.<br />
The final aim is to develop a descriptor valid for complete traffic noise<br />
scenarios predicting environmental noise quality adequately.<br />
11:45<br />
4aNS13. When objective permissible noise limits of a municipal<br />
planning process and a subjective noise ordinance conflict. Marlund<br />
Hale �Adv. Eng. Acoust., 663 Bristol Ave., Simi Valley, CA 93065,<br />
mehale@aol.com�<br />
In most communities, proposed new building projects are required to<br />
conform with planning, community development, zoning, and/or building<br />
and safety specifications and standards. In cases of allowable noise exposure<br />
and noise limits, where certain of these requirements are quite specific<br />
while others are purposefully vague, conflicts between residential and<br />
commercial neighbors can lead to extreme disagreement and needless litigation.<br />
This paper describes a recent situation occurring in an upscale<br />
beach community, the resulting conflict over existing noise sources that<br />
comply with the limits of the city planning and permitting process, and the<br />
interesting findings of the court following the civil and criminal litigation<br />
that followed. Some suggestions are given to avoid these conflicting<br />
policy situations.<br />
Poster paper 4aNS14 will be on display from 7:30 a.m. to 12:15 p.m. The author will be at the poster from 12:00 noon to 12:15 p.m.<br />
4aNS14. The complexity of environmental sound as a function of<br />
seasonal variation. Hideo Shibayama �3-7-5 Koutou-ku Tokyo,<br />
135-8548, Japan, sibayama@sic.shibaura-it.ac.jp�<br />
Residential land is performed for a surrounding area of a suburban<br />
local city. As a result of urbanization, an area of rich natural environments<br />
became narrow. For the animals and plants for whom a river and a forest<br />
are necessary, it becomes difficult to live. Environmental sound produced<br />
by the tiny insects in this area is changing from year to year. Catching the<br />
conditions for the developmental observations and environmental preservation<br />
in natural environments, we continue to measure the environmental<br />
sound as the time-series data. We estimate the complexity for these waveforms<br />
of the measured environmental sound in the season when insects<br />
chirp and do not chirp. For estimation of the complexity, we evaluate by<br />
the fractal dimension of the environmental sound. Environmental sound in<br />
early autumn is mainly generated by insects in the grass and on the trees.<br />
And, the fractal dimension for the sound waveforms of chirping of insects<br />
is up to 1.8.<br />
3240 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3240
FRIDAY MORNING, 1 DECEMBER <strong>2006</strong> WAIANAE ROOM, 7:30 A.M. TO 12:20 P.M.<br />
Session 4aPA<br />
Physical Acoustics and Biomedical UltrasoundÕBioresponse to Vibration: Sound Propagation in<br />
Inhomogeneous Media I<br />
James G. Miller, Cochair<br />
Washington Univ., Dept. of Physics, 1 Brookings Dr., St. Louis, MO 63130<br />
Mami Matsukawa, Cochair<br />
Doshisha Univ., Lab. of Ultrasonic Electronics, Kyoto 610-0321, Japan<br />
Chair’s Introduction—7:30<br />
Invited Papers<br />
7:35<br />
4aPA1. In vivo measurement of mass density and elasticity of cancellous bone using acoustic parameters for fast and slow<br />
waves. Takahiko Otani �Faculty of Eng., Doshisha Univ., Kyotanabe 610-0321 Japan�<br />
Cancellous bone �spongy bone� is comprised of a porous network of numerous trabecular elements with soft tissue in the pore<br />
space. The effect of decreasing bone density, namely a symptom of osteoporosis, is greater for cancellous bone than for dense cortical<br />
bone �compact bone�. Two longitudinal waves, the fast and slow waves, are clearly observed in cancellous bone, which correspond to<br />
‘‘waves of the first and second kinds’’ as predicted by Biot’s theory. According to experimental and theoretical studies, the propagation<br />
speed of the fast wave increases with bone density and that of the slow wave is almost constant. Experimental results show that the<br />
fast wave amplitude increases proportionally and the slow wave amplitude decreases inversely with bone density. However, the<br />
attenuation constant of the fast wave is almost independent of bone density and the attenuation constant of the slow wave increases<br />
with bone density. The<br />
in vivo ultrasonic wave propagation path is composed of soft tissue, cortical bone, and cancellous bone and is modeled to specify the<br />
causality between ultrasonic wave parameters and bone mass density of cancellous bone. Then, mass density and elasticity are<br />
quantitatively formulated and estimated.<br />
7:55<br />
4aPA2. Is ultrasound appropriate to measure bone quality factors? Pascal Laugier �Univ. Pierre et Marie Curie., UMR CNRS<br />
7623, 15 rue de l’Ecole de medecine, 7506 Paris, France, laugier@lip.bhdc.jussieu.fr�<br />
Theoretical considerations support the concept that quantitative ultrasound variables measured in transmission are mainly determined<br />
by bone microstructure and material properties. All these properties are attributes of bone other than its bone mineral density<br />
�BMD� that may contribute to its quality and thus to strength or fragility. However, the limitations of this approach for a BMDindependent<br />
characterization of bone quality, long questioned, have become indisputable. Such considerations have prompted a<br />
research aiming at the development of new methods capable of measuring bone quality factors. New ultrasonic approaches are being<br />
investigated that use ultrasonic backscatter, guided waves, or nonlinear acoustics for studying bone microstructure or microdamage.<br />
These approaches, combined with sophisticated theoretical models or powerful computational tools, are advancing ideas regarding<br />
ultrasonic assessment of bone quality, which is not satisfactorily measured by x-ray techniques.<br />
8:15<br />
4aPA3. Scanning acoustic microscopy studies of cortical and trabecular bone in the femur and mandible. J. Lawrence Katz,<br />
Paulette Spence, Yong Wang �Univ. of Missouri-Kansas City, 650 E. 25th St., Kansas City, MO 64108, katzjl@umkc.edu�, Anil<br />
Misra, Orestes Marangos �Univ. of Missouri-Kansas City, Kansas City, MO 64110�, Dov Hazony �Case Western Reserve Univ.,<br />
Cleveland, OH 44106�, and Tsutomu Nomura �Niigata Univ. Grad. School of Medical and Dental Sci., Niigata, Japan�<br />
Scanning acoustic microscopy �SAM� has been used to study the micromechanical properties of cortical and trabecular bone in<br />
both the human femur and mandible. SAM images vary in gray level reflecting the variations in reflectivity of the material under<br />
investigation. The reflection coefficient, r�(Z2�Z1)/(Z2�Z1), where the acoustic impedance �AI�, Z�dv, d is the materials local<br />
density and v is the speed of sound at the focal point; Z2 represents the AI of the material, Z1 that of the fluid coupling the acoustic<br />
wave from the lens to the material. Femoral cortical bone consists of haversian systems �secondary osteons� and interstitial lamellae,<br />
both of which show systematic variations of high and low AI from lamella to lamella. The lamellar components defining the edges of<br />
trabecular cortical bone exhibit the same lamellar variations as seen in cortical bone. Mandibular bone, while oriented perpendicular<br />
to the direction of gravitational attraction, exhibits the same cortical and trabecular structural organizations as found in the femur. It<br />
also exhibits the same systematic alternations in lamellar AI as found in femoral bone. Both femoral and mandibular cortical bone<br />
have transverse isotropic symmetry. Thus, modeling elastic properties requires only five independent measurements.<br />
3241 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3241<br />
4a FRI. AM
8:35<br />
4aPA4. The interaction between ultrasound and human cancellous bone. Keith Wear �US Food and Drug Administration, 12720<br />
Twinbrook Pkwy., Rockville, MD 20852, keith.wear@fda.hhs.gov�<br />
Attenuation is much greater in cancellous bone than in soft tissues, and varies approximately linearly with frequency between 300<br />
kHz and 1.7 MHz. At diagnostic frequencies �300 to 700 kHz�, sound speed is slightly faster in cancellous bone than in soft tissues.<br />
A linear-systems model can account for errors in through-transmission-based measurements of group velocity due to frequencydependent<br />
attenuation and dispersion. The dependence of phase velocity on porosity may be predicted from theory of propagation in<br />
fluid-filled porous solids. The dependence of phase velocity on frequency �negative dispersion� can be explained using a stratified<br />
two-component model. At diagnostic frequencies, scattering varies as frequency to the nth power where 3�n�3.5. This may be<br />
explained by a model that represents trabeculae as finite-length cylindrical scatterers.<br />
8:55<br />
4aPA5. Dependence of phase velocity on porosity in cancellous bone: Application of the modified Biot-Attenborough model.<br />
Suk Wang Yoon and Kang Il Lee �Dept. of Phys. and Inst. of Basic Sci., SungKyunKwan Univ., Suwon 440-746, Republic of Korea�<br />
This study aims to apply the modified Biot-Attenborough �MBA� model to predict the dependence of phase velocity on porosity<br />
in cancellous bone. The MBA model predicted that the phase velocity decreases nonlinearly with porosity. The optimum values for<br />
input parameters of the MBA model, such as compressional speed c m of solid bone and phase velocity parameter s 2, were determined<br />
by comparing the prediction with the previously published measurements in human calcaneus and bovine cancellous bones. The value<br />
of the phase velocity parameter s 2�1.23 was obtained by curve fitting to the experimental data only for 53 human calcaneus samples<br />
with a compressional speed c m�2500 m/s of solid bone. The root-mean-square error �rmse� of the curve fit was 15.3 m/s. The<br />
optimized value of s 2 for all 75 cancellous bone samples �53 human and 22 bovine samples� was 1.42 with the rmse of 55 m/s. The<br />
latter fit was obtained by using c m�3200 m/s. Although the MBA model relies on empirical parameters determined from the<br />
experimental data, it is expected that the model can be usefully employed as a practical tool in the field of clinical ultrasonic bone<br />
assessment.<br />
9:15<br />
4aPA6. Simulation of fast and slow wave propagations through cancellous bone using three-dimensional elastic and Biot’s<br />
trabecular models. Atsushi Hosokawa �Dept. of Elect. & Comp. Eng., Akashi Natl. Coll. of Tech., 679-3 Nishioka, Uozumi,<br />
Akashi, 674-8501 Hyogo, Japan, hosokawa@akashi.ac.jp�<br />
The propagation of ultrasonic pulse waves in cancellous �trabecular� bone was numerically simulated by using three-dimensional<br />
finite-difference time-domain �FDTD� methods. In previous research �A. Hosokawa, J. Acoust. Soc. Am. 118, 1782–1789 �2005��,<br />
two two-dimensional FDTD models, the commonly used elastic FDTD model and an FDTD model based on Biot’s theory for elastic<br />
wave propagation in an isotropic fluid-saturated porous medium, were used to simulate the fast and slow longitudinal waves propagating<br />
through cancellous bone in the direction parallel to the main trabecular orientation. In the present study, the extended<br />
three-dimensional viscoelastic and Biot’s anisotropic models were developed to investigate the effect of trabecular structure on the fast<br />
and slow wave propagations. Using the viscoelastic model of the trabecular frame comprised of numerous pore spaces in the solid<br />
bone, the effect of the trabecular irregularity, that is the scattering effect, on both the fast and slow waves could be investigated. The<br />
effect of the anisotropic viscous resistance of the fluid in the trabecular pore spaces on the slow wave could be considered using Biot’s<br />
anisotropic model.<br />
9:35<br />
4aPA7. Ultrasonic characteristics of in vitro human cancellous bone.<br />
Isao Mano �OYO Electric Co., Ltd., Joyo 610-0101 Japan�, Tadahito<br />
Yamamoto, Hiroshi Hagino, Ryota Teshima �Tottori Univ., Yonago<br />
683-8503 Japan�, Toshiyuki Tsujimoto �Horiba, Ltd., Kyoto 601-8510<br />
Japan�, and Takahiko Otani �Doshisha Univ., Kyotanabe 610-0321 Japan�<br />
Cancellous bone is comprised of a connected network of trabeculae<br />
and is considered as an inhomogeneous and anisotropic acoustic medium.<br />
The fast and slow longitudinal waves are clearly observed when the ultrasonic<br />
wave propagates parallel to the direction of the trabeculae. The<br />
propagation speed of the fast wave increases with bone density and that of<br />
the slow wave is almost constant. The fast wave amplitude increases proportionally<br />
and the slow wave amplitude decreases inversely with bone<br />
density. Human in vitro femoral head was sectioned to 10-mm-thick slices<br />
perpendicularly to the femoral cervical axis. These cancellous bone<br />
samples were subjected to the ultrasonic measurement system LD-100<br />
using a narrow focused beam. The propagation speed and the amplitude of<br />
the transmitted wave both for the fast and slow waves were measured at<br />
1-mm intervals. The local bone density corresponding to the measured<br />
points was obtained using a microfocus x-ray CT system. Experimental<br />
results show that the propagation speeds and amplitudes for the fast and<br />
slow waves are characterized not only by the local bone density, but also<br />
by the local trabecular structure.<br />
Contributed Papers<br />
9:50<br />
4aPA8. The effect of hydroxyapatite crystallite orientation on<br />
ultrasonic velocity in bovine cortical bone. Yu Yamato, Kaoru<br />
Yamazaki, Akira Nagano �Orthopaedic Surgery, Hamamatsu Univ.<br />
School of Medicine, 1-20-1 Hamamatsu Shizuoka 431-3192, Japan,<br />
yy14@hama-med.ac.jp�, Hirofumi Mizukawa, Takahiko Yanagitani, and<br />
Mami Matsukawa �Doshisha Univ., Kyotanabe, Kyoto-fu 610-0321,<br />
Japan�<br />
Cortical bone is recognized as a composite material of diverse elastic<br />
anisotropy, composed of hydroxyapatite �HAp� crystallite and type 1 collagen<br />
fiber. The aim of this study is to investigate the effects of HAp<br />
orientation on elastic anisotropy in bovine cortical bone. Eighty cubic<br />
samples were made from the cortical bone of two left bovine femurs.<br />
Longitudinal wave velocity in orthogonal three axes was measured using a<br />
conventional pulse echo system. For evaluating the orientation of HAp<br />
crystallite, x-ray diffraction profiles were obtained from three surfaces of<br />
each cubic sample. The preference of crystallites showed strong dependence<br />
on the direction of surface. The C-axis of crystallites was clearly<br />
preferred to the bone axial direction. The velocity in the axial direction<br />
was significantly correlated with the amounts of HAp crystallites aligning<br />
to the axial axis. The HAp orientation and velocity varied according to the<br />
microstructures of the samples. The samples with Haversian structure<br />
3242 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3242
showed larger preference of crystallites than plexiform samples in the<br />
axial direction. These results show clear effects of crystallites orientation<br />
on velocity.<br />
10:05–10:20 Break<br />
10:20<br />
4aPA9. Frequency variations of attenuation and velocity in cortical<br />
bone in vitro. Magali Sasso, Guillaume Haiat �Universite Paris 12,<br />
Laboratoire de Mecanique Physique, UMR CNRS 7052 B2OA, 61, avenue<br />
du General de Gaulle, 94010 Creteil, France�, Yu Yamato �Hamamatsu<br />
Univ. School of Medicine, Hamamastu, Shizuoka, 431-3192, Japan�,<br />
Salah Naili �Universite Paris 12, 94010 Creteil, France�, and Mami<br />
Matsukawa �Doshisha Univ., Kyotanabe, Kyoto-fu 610-0321, Japan�<br />
The development of ultrasonic characterization devices for cortical<br />
bone requires a better understanding of ultrasonic propagation in this heterogeneous<br />
medium. The aim of this work is to investigate the frequency<br />
dependence of the attenuation coefficient and of phase velocity and to<br />
relate them with bone microstructure and anatomical position. One hundred<br />
twenty parallelepipedic samples �4–11 mm side� have been cut from<br />
three bovine specimens and measured four times with repositioning in<br />
transmission with a pair of 8-MHz central frequency transducers. Phase<br />
velocity and BUA could be determined with acceptable precision: coefficients<br />
of variation of 0.8% and 13%, respectively. Velocity dispersion and<br />
BUA values are comprised between �13 and 40 m/s/MHz and 2 and 12<br />
dB/MHz/cm, respectively. Negative dispersion values were measured<br />
�similarly to trabecular bone� for 2% of the measured samples. BUA values<br />
were found to be smaller in plexiform than in Haversian structure and<br />
higher for porotic structure. BUA values were found to be the greatest in<br />
the postero-lateral distal part and the smallest in the anterior-medial center<br />
part of the bone. The same tendency was found for velocity dispersion.<br />
Our results show the sensitivity of the frequency dependence of ultrasound<br />
to anatomical position and micro-architectural properties of bone.<br />
10:35<br />
4aPA10. 3-D numerical simulations of wave propagation in trabecular<br />
bone predicts existence of the Biot fast compressional wave.<br />
Guillaume Haiat �Universite Paris 12, Laboratoire de Mecanique<br />
Physique, UMR CNRS 7052 B2OA, 61, avenue du General de Gaulle,<br />
94010 Creteil, France�, Frederic Padilla, and Pascal Laugier �Universite<br />
Pierre et Marie Curie, 75006 Paris, France�<br />
Trabecular bone is a poroelastic medium in which the propagation of<br />
two longitudinal waves �fast and slow� has been observed. The 3-D finitedifference<br />
time-domain simulations neglecting absorption coupled to 3-D<br />
microstructural models of 34 trabecular bone reconstructed from synchrotron<br />
radiation microtomography are shown to be suitable to predict both<br />
types of compressional wave in the three orthogonal directions. The influence<br />
of bone volume fraction �BV/TV� on the existence of the fast and<br />
slow waves was studied using a dedicated iterative image processing algorithm<br />
�dilation, erosion� in order to modify all 34 initial 3-D microstructures.<br />
An automatic criteria aiming at determining the existence of both<br />
wave modes was developed from the analysis of the transmitted signals in<br />
the frequency domain. For all samples, the fast wave disappears when<br />
bone volume fraction decreases. Both propagation modes were observed<br />
for BV/TV superior to a critical value for 2, 13, and 17 samples according<br />
to the direction of propagation. Above this critical value, the velocity of<br />
the fast �slow� wave increases �decreases� with BV/TV, consistent with<br />
Biot’s theoretical predictions. This critical value decreases when the degree<br />
of anisotropy increases, showing coupling between structural anisotropy<br />
and the existence of the fast wave.<br />
10:50<br />
4aPA11. 3-D numerical simulation of wave propagation in porous<br />
media: Influence of the microstructure and of the material properties<br />
of trabecular bone. Guillaume Haiat �Universite Paris 12, Laboratoire<br />
de Mecanique Physique, UMR CNRS 7052 B2OA, 61, avenue du General<br />
de Gaulle, 94010 Creteil, France�, Frederic Padilla, and Pascal Laugier<br />
�Universite Pierre et Marie Curie, 75006 Paris, France�<br />
Finite-difference time domain simulations coupled to 3-D microstructural<br />
models of 30 trabecular bones reconstructed from synchrotron radiation<br />
microtomography were employed herein to compare and quantify the<br />
effects of bone volume fraction, microstructure, and material properties of<br />
trabecular bone on QUS parameters. Scenarios of trabecular thinning and<br />
thickening using an iterative dedicated algorithm allowed the estimation of<br />
the sensitivity of QUS parameters to bone volume fraction. The sensitivity<br />
to bone material properties was assessed by considering independent<br />
variations of density and stiffness. The effect of microstructure was qualitatively<br />
assessed by producing virtual bone specimens of identical bone<br />
volume fraction �13%�. Both BUA and SOS show a strong correlation<br />
with BV/TV (r 2 �0.94 p10 �4 ) and vary quasi-linearly with BV/TV at an<br />
approximate rate of 2 dB/cm/MHz and 11 m/s per % increase of BV/TV,<br />
respectively. Bone alterations caused by variation in BV/TV �BUA: 40<br />
dB/cm.MHz; SOS: 200 m/s� is much more detrimental to QUS variables<br />
than that caused by alterations of material properties or diversity in microarchitecture<br />
�BUA: 7.8 dB/cm.MHz; SOS: 36 m/s�. QUS variables are<br />
changed more dramatically by BV/TV than by changes in material properties<br />
or microstructural diversity. However, material properties and structure<br />
also appear to play a role.<br />
11:05<br />
4aPA12. Singular value decomposition-based wave extraction<br />
algorithm for ultrasonic characterization of cortical bone in vivo.<br />
Magali Sasso, Guillaume Haiat �Universite Paris 12, Laboratoire de<br />
Mecanique Physique, UMR CNRS 7052 B2OA, 61, avenue du General de<br />
Gaulle, 94010 Creteil, France�, Maryline Talmant, Pascal Laugier<br />
�Universite Pierre et Marie Curie, 75006 Paris, France�, and Salah Naili<br />
�Universite Paris 12, 94010 Creteil, France�<br />
In the context of bone status assessment, the axial transmission technique<br />
allows ultrasonic evaluation of cortical bone using a multielement<br />
probe. The current processing uses the first arriving signal to evaluate the<br />
velocity while later contributions are potentially valuable and are not yet<br />
analyzed. However, all those contributions interfere, which disrupts the<br />
signal analysis. A novel ultrasonic wave extraction algorithm using a singular<br />
value decomposition method is proposed. This algorithm aims at<br />
characterizing a given energetic low-frequency �ELF� contribution observed<br />
in vivo at around 1 MHz. To evaluate the performances of the<br />
proposed algorithm, a simulated data set was constructed taking into account<br />
the influence of noise and of random interfering wavefront. The<br />
velocity of the ELF contribution is estimated on simulated datasets and<br />
compared to the input velocity. For a signal-to-noise ratio of 10 dB, the<br />
mean error associated with this method is 5.2%, to be compared with 34%<br />
with a classical signal analysis. The algorithm was also tested on real in<br />
vivo measurements. Results show the ability to accurately identify and<br />
possibly remove this wavefront contribution. Results are promising for<br />
multiple ultrasonic parameters evaluation from different wavefront contributions<br />
in our configuration.<br />
11:20<br />
4aPA13. Direct evaluation of cancellous bone porosity using<br />
ultrasound. Peiying Liu, Matthew Lewis, and Peter Antich �Grad.<br />
Program of Biomed. Eng., Univ. of Texas Southwestern Medical Ctr. at<br />
Dallas, 5323 Harry Hines Blvd, Dallas, TX 75390-9058�<br />
Quantitative measurements of trabecular bone porosity would be a<br />
great advantage in diagnosis and prognosis of osteoporosis. This study<br />
focused on evaluation of the relationship between ultrasonic reflection and<br />
the density and porosity of cancellous bone. Theoretical simulation using<br />
MATLAB and Field II ultrasound simulator predicted a linear model of a<br />
material’s porosity and parameters of the reflected ultrasonic signals. Ex-<br />
3243 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3243<br />
4a FRI. AM
perimentally, four plastic phantoms fabricated with different porosities<br />
were tested by a 5-MHz ultrasound transducer and the results agreed with<br />
simulations. Twelve specimens of bovine cancellous bone were measured.<br />
The porosities of these specimens were estimated by calculating the ratio<br />
of the mass in air to the wetted mass when they were immersed in water<br />
and all the air was removed from the pores. Among all the parameters, the<br />
peak value of the reflected ultrasound signal demonstrated highly significant<br />
linear correlations with porosity (R 2 �0.911) and density (R 2<br />
�0.866). This encouraging result shows that this technique has the potential<br />
to be used to monitor porosity changes noninvasively for clinical<br />
purpose such as noninvasive assessment of osteoporotic fracture risk.<br />
�Work supported by Pak Foundation.�<br />
11:35<br />
4aPA14. Transmission of ultrasound through bottlenose dolphin<br />
(tursiops truncatus) jaw and skull bone. Michael D. Gray, James S.<br />
Martin, and Peter H. Rogers �Woodruff School of Mech. Eng., Georgia<br />
Inst. of Technol., Atlanta, GA 30332-0405, michael.gray@gtri.gatech.edu�<br />
Measurements of ultrasound transmission through jaw �pan bone� and<br />
skull �temporal fossa� samples from an Atlantic bottlenose dolphin were<br />
performed as part of an investigation of the feasibility of performing in<br />
vivo elastography on cetacean head tissues. The pan bone and temporal<br />
fossa are both relatively thin and smooth, and are therefore of interest for<br />
acoustic imaging of the underlying tissues of the ear and brain, respectively.<br />
Field scan data will be presented, showing the influence of the bone<br />
on the quality of the focus and overall pressure levels generated by a<br />
spherically focused single-element transducer. �Work supported by ONR.�<br />
11:50<br />
4aPA15. Assessment of bone health by analyzing propagation<br />
parameters of various modes of acoustic waves. Armen Sarvazyan,<br />
Vladimir Egorov, and Alexey Tatarinov �Artann Labs., Inc., 1457 Lower<br />
Ferry Rd., West Trenton, NJ 08618�<br />
A method for assessment of bone based on comprehensive analysis of<br />
waveforms of ultrasound signals propagating in the bone is presented. A<br />
set of ultrasound propagation parameters, which are differentially sensitive<br />
to bone material properties, structure, and cortical thickness, are evaluated.<br />
The parameters include various features of different modes of acoustics<br />
waves, such as bulk, surface, and guided ultrasonic waves in a wide range<br />
of carrier frequencies. Data processing algorithms were developed for obtaining<br />
axial profiles of waveform parameters. Such profiles are capable of<br />
revealing axial heterogeneity of long bones and spatially nonuniform<br />
pathological processes, such as osteoporosis. The examination procedure<br />
may include either one long bone in the skeleton, like the tibia, radius of<br />
the forearm, etc., or several bones in sequence to provide a more comprehensive<br />
assessment of the skeletal system. Specifically, for tibia assessment,<br />
a multi-parametric linear classifier based on a DEXA evaluation of<br />
skeleton conditions has been developed. Preliminary results of the pilot<br />
clinical studies involving 149 patients have demonstrated early stages of<br />
osteoporosis detection sensitivity of 80% and specificity of 67% based on<br />
DEXA data as the gold standard. �Work was supported by NIH and NASA<br />
grants.�<br />
12:05<br />
4aPA16. Toward bone quality assessment: Interference of fast and<br />
slow wave modes with positive dispersion can account for the<br />
apparent negative dispersion. James G. Miller, Karen Marutyan, and<br />
Mark R. Holland �Dept. of Phys., Washington Univ., St. Louis, MO<br />
63130�<br />
The goal of this work was to show that apparently negative dispersion<br />
in bone can arise from interference between fast wave and slow wave<br />
longitudinal modes, each of positive dispersion. Simulations were carried<br />
out using two approaches, one based on the Biot-Johnson model and the<br />
second independent of that model. The resulting propagating fast wave<br />
and slow wave modes accumulated phase and suffered loss with distance<br />
traveled. Results of both types of simulation served as the input into a<br />
phase and magnitude spectroscopy algorithm �previously validated with<br />
experimental data� in order to determine the phase velocity as a function<br />
of frequency. Results based on both methods of simulation were mutually<br />
consistent. Depending upon the relative magnitudes and time relationships<br />
between the fast and slow wave modes, the apparent phase velocities as<br />
functions of frequency demonstrated either negative or positive dispersions.<br />
These results appear to account for measurements from many laboratories<br />
that report that the phase velocity of ultrasonic waves propagating<br />
in cancellous bone decreases with increasing frequency �negative dispersion�<br />
in about 90% of specimens but increases with frequency in about<br />
10%. �Work supported in part by Grant NIH R37HL40302.�<br />
3244 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3244
FRIDAY MORNING, 1 DECEMBER <strong>2006</strong> WAIALUA ROOM, 8:30 TO 11:25 A.M.<br />
Session 4aPP<br />
Psychological and Physiological Acoustics and ASA Committee on Standards: New Insights on Loudness<br />
and Hearing Thresholds<br />
Rhona P. Hellman, Cochair<br />
Northeastern Univ., Dept. of Speech-Language Pathology and Audiology, Boston, MA 02115<br />
Yôiti Suzuki, Cochair<br />
Tohoku Univ., Research Inst. of Electrical Communication, Katarahira 2-1-1, Aoba-ku, Sendai 980-8577, Japan<br />
Chair’s Introduction—8:30<br />
Invited Papers<br />
8:35<br />
4aPP1. Threshold of hearing for pure tones between 16 and 30 kHz. Kaoru Ashihara �AIST, Tsukuba Central, 6 Tsukuba, Ibaraki<br />
305-8566 Japan, ashihara-k@aist.go.jp�<br />
Hearing thresholds for pure tones were obtained at 2-kHz intervals between 16 and 30 kHz in an anechoic chamber. Measured 50<br />
cm from the sound source, the maximum presentation sound pressure level ranged from 105 to 112 dB depending on the frequency.<br />
To prevent the listener from detecting the quantization noise or subharmonic distortions at low frequencies, a pink noise was added as<br />
a masker. Using a 3-down 1-up transformed up-down method, thresholds were obtained at 26 kHz for 10 out of 32 ears. Even at 28<br />
kHz threshold values were obtained for 3 ears, but none were observed for a tone at 30 kHz. Above 24 kHz, the thresholds always<br />
exceeded 90 dB SPL. Between 16 and 20 kHz thresholds increased abruptly, whereas above 20 kHz the threshold increase was more<br />
gradual.<br />
8:55<br />
4aPP2. Do we have better hearing sensitivity than people in the past? Kenji Kurakata and Tazu Mizunami �Natl. Inst. of Adv.<br />
Industrial Sci. and Technol. �AIST�, 1-1-1 Higashi, Tsukuba, Ibaraki, 305-8566 Japan, kurakata-k@aist.go.jp�<br />
Our hearing sensitivity to tones of various frequencies declines progressively as we become older. ISO 7029 describes a method<br />
for calculating expected hearing threshold values as a function of age. However, more than 30 years have passed since the ISO<br />
standard source data were published. An earlier paper of the present authors �K. Kurakata and T. Mizunami, Acoust. Sci. Technol. 26,<br />
381–383 �2005�� compared hearing threshold data of Japanese people in recent years to the ISO standard values. The results of that<br />
comparison showed that the age-related sensitivity decrease of Japanese people was smaller, on average, than those described in the<br />
standard. A large discrepancy was apparent at 4000 and 8000 Hz: more than 10 dB for older males. In response to that inference, the<br />
ISO/TC43/WG1 ‘‘threshold of hearing’’ initiated a project in 2005 to explore the possibility of updating ISO 7029. This paper presents<br />
a summary of those comparison results of audiometric data and the work in WG1 for revising the standard.<br />
9:15<br />
4aPP3. Reliability and frequency specificity of auditory steady-state response. Masaru Aoyagi, Tomoo Watanabe, Tsukasa Ito,<br />
and Yasuhiro Abe �Dept. of Otolaryngol., Yamagata Univ. School of Medicine, 2-2-2 Iida-Nishi, Yamagata, 990-9585, Japan�<br />
The reliability and frequency specificity of 80-Hz auditory steady-state response �80-Hz ASSR� elicited by sinusoidally amplitudemodulated<br />
�SAM tones� tones and detected by phase coherence were evaluated as a measure of the hearing level in young children.<br />
The 80-Hz ASSR at a carrier frequency of 1000 Hz was monitored in 169 ears of 125 hearing-impaired children and auditory<br />
brainstem response �ABR� elicited by tone pips was examined in 93 ears. Both responses were examined during sleep, and the<br />
thresholds were compared with the behavioral hearing threshold, which was determined by standard pure-tone audiometry or play<br />
audiometry. In 24 ears with various patterns of audiogram, 80-Hz ASSRs were examined at different carrier frequencies, and the<br />
threshold patterns were compared with the audiograms to investigate the frequency specificity of ASSR. The correlation coefficient<br />
between the threshold of 80-Hz ASSR and pure-tone threshold (r�0.863) was higher than that for ABR (r�0.828). The threshold<br />
patterns of 80-Hz ASSR clearly followed the corresponding audiogram patterns in all types of hearing impairment. These findings<br />
suggest that 80-Hz ASSR elicited by SAM tones and detected by phase coherence is a useful audiometric device for the determination<br />
of hearing level in a frequency-specific manner in children.<br />
3245 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3245<br />
4a FRI. AM
9:35<br />
4aPP4. Use of perceptual weights to test a model of loudness summation. Lori J. Leibold and Walt Jesteadt �Boys Town Natl.<br />
Res. Hospital, 555 N 30th St., Omaha, NE 68131�<br />
We recently examined the contribution of individual components of a multitone complex to judgments of overall loudness by<br />
computing the perceptual weight listeners assign to each component in a loudness-matching task �Leibold et al., J. Acoust. Soc. Am.<br />
117, 2597 �2005��. Stimuli were five-tone complexes centered on 1000 Hz, with six different logarithmic frequency spacings, ranging<br />
from 1.012 to 1.586. When all components fell within the same critical band, weights varied little across components. In contrast, the<br />
range of weights increased with increasing frequency separation, with greater weight given to the lowest and highest frequency<br />
components. Perceptual weights were largely in agreement with the Moore et al. loudness model �J. Audio Eng. Soc. 45, 224–237<br />
�1997��, except at the widest bandwidth. In the current study we further examined predictions of the loudness model, focusing on the<br />
widest frequency-spacing condition. Masked thresholds and jnds for intensity discrimination were measured for each component and<br />
were compared to weights. The model predicts more interaction in the widely spaced conditions than simple critical band models, but<br />
underestimates the true interactions in conditions where components are widely spaced. Central factors appear to influence loudness,<br />
masked thresholds, and intensity discrimination in these conditions. �Work supported by NIH/NIDCD.�<br />
9:55<br />
4aPP5. Increased loudness effect at the absolute threshold of hearing. Junji Yoshida, Masao Kasuga �Grad. School of Eng.,<br />
Utsunomiya Univ., 7-1-2 Yoto, Utsunomiya-shi, Tochigi-ken, 321-8585, Japan�, and Hiroshi Hasegawa �Utsunomiya Univ.,<br />
Tochigi-ken, 321-8585, Japan�<br />
This study investigated the effects of a previous sound on loudness at the absolute threshold of hearing. Change of the absolute<br />
threshold of hearing was measured when a pure tone preceded the test tone in the measurement of the threshold. The previous sound<br />
at 60 dB SPL was presented first in one ear, followed by the presentation of the test sound in either the contralateral or ipsilateral ear<br />
at an interval of 0.5 s. Both the previous and test sounds had the same frequency of 500 Hz, and the same duration of 3 s. The change<br />
of the threshold was obtained from the difference between the thresholds with and without the previous sound. The threshold was<br />
found to be decreased significantly by approximately 2 dB when the previous sound was presented in the contralateral ear. On the<br />
other hand, the threshold was only slightly changed when the previous sound was presented in the ipsilateral ear. These experiment<br />
results suggest that loudness was increased by perceiving the previous sound in the contralateral ear.<br />
10:15–10:25 Break<br />
10:25<br />
4aPP6. Induced loudness reduction. Michael Epstein �Auditory Modeling and Processing Lab., Inst. for Hearing, Speech & Lang.,<br />
Dept. of Speech-Lang. Path. and Aud., Northeastern Univ., Boston, MA 02115� and Mary Florentine �Northeastern Univ., Boston,<br />
MA�<br />
Induced loudness reduction �ILR� is a phenomenon by which a preceding higher-level tone �an inducer tone� reduces the loudness<br />
of a lower-level tone �a test tone�. Under certain conditions, ILR can result in loudness reductions of 10 to 15 phons for pure tones.<br />
The strength of this effect depends on a number of parameters including: �1� the levels of both the inducer and test tones; �2� the<br />
frequency separation between the inducer and test tones; �3� the durations of the inducer and test tones; �4� the time separation<br />
between the inducer and test tones; �5� individual differences; and, possibly �6� the number of exposures to inducers. Because of the<br />
sensitivity of ILR to procedural conditions, it is quite important to consider the potential effects of ILR when considering any<br />
experimental design in which level varies. The understanding of ILR has also given insight into a number of unexplained discrepancies<br />
between data sets that were collected using different procedures. In fact, some of the variability known to affect loudness<br />
judgments may be due to ILR. �Work supported by NIH/NIDCD Grant R01DC02241.�<br />
10:45<br />
4aPP7. Loudness growth in individual listeners with hearing loss. Jeremy Marozeau and Mary Florentine �Commun. Res. Lab.,<br />
Inst. for Hearing, Speech & Lang., Northeastern Univ., Boston, MA 02115�<br />
Recent research indicates that there are large individual differences in how loudness grows with level for listeners with sensorineural<br />
hearing losses of primarily cochlear origin. Studies of loudness discomfort level suggest that loudness for most of these<br />
impaired listeners approaches the loudness experienced by the normal listeners at high levels. Loudness growth at low levels is still<br />
controversial. Although it is now well established that loudness at threshold is greater than zero, its exact value is unknown. If this<br />
value is the same for normal and impaired listeners, then the loudness growth for the impaired listeners must be steeper in order to<br />
catch up to normal at high levels. This phenomenon is called recruitment. If the loudness at threshold for impaired listeners is higher<br />
than that for normal listeners, then the impaired listeners will no longer be able to perceive sounds as soft. This phenomenon is called<br />
softness imperception. Results from two experiments suggest that: �1� individual differences are more important for impaired listeners<br />
than for normal listeners; �2� some impaired listeners seem to show recruitment, others softness imperception; and �3� averaging the<br />
results across the impaired listeners will mask these differences. �Work supported by NIH/NIDCD grant R01DC02241.�<br />
3246 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3246
11:05<br />
4aPP8. Growth of loudness in cochlear implant listening. Robert L. Smith and Nicole Sanpetrino �Inst. for Sensory Res. and<br />
Dept. of Biomed. and Chemical Eng., Syracuse Univ., 621 Skytop Rd., Syracuse NY, 13244�<br />
Cochlear implants �CIs� roughly mimic the transformation from sound frequency to cochlear place that occurs in acoustic hearing.<br />
However, CIs are relatively less capable of creating the intensive transformations that normal peripheral auditory processing provides.<br />
This is partly because CIs have small operating ranges on the order of 10:1 in electric current, compared to the 1 000 000:1 operating<br />
range for sound-pressure level �SPL� in acoustic hearing. Furthermore, loudness in acoustic hearing grows as a compressive power<br />
function of SPL. In contrast, loudness reportedly grows as a more expansive function of current for CI users, i.e., a power law with<br />
a large exponent or an exponential function. Our results, obtained using the minimally biased method of magnitude balance without<br />
an arbitrary standard, reveal a previously unreported range of shapes of CI loudness functions, going from linear to power laws with<br />
exponents of 5 or more. The shapes seem related in part to the etiology of deafness preceding cochlear implantation, although the<br />
shapes can vary with stimulating conditions within a subject. Furthermore, differential sensitivity to changes in current appears to be<br />
related to the shape of the corresponding loudness function. Implications for sound processing in electric and acoustic hearing will be<br />
discussed.<br />
FRIDAY MORNING, 1 DECEMBER <strong>2006</strong> MOLOKAI ROOM, 8:00 A.M. TO 12:00 NOON<br />
Session 4aSC<br />
Speech Communication: Perception „Poster Session…<br />
Katsura Aoyama, Cochair<br />
Texas Tech. Univ., Health Science Ctr., Dept. of Speech-Language and Hearing Science, Lubbock, TX 79430-6073<br />
Masato Akagi, Cochair<br />
JAIST, School of Information Science, 1-1 Asahidai, Nomi, Ishikawa 923-1292, Japan<br />
Contributed Papers<br />
All posters will be on display from 8:00 a.m. to 12:00 noon. To allow contributors an opportunity to see other posters, contributors of<br />
odd-numbered papers will be at their posters from 8:00 a.m. to 10:00 a.m. and contributors of even-numbered papers will be at their<br />
posters from 10:00 a.m. to 12:00 noon.<br />
4aSC1. Functional load of segments and features. Mafuyu Kitahara<br />
�School of Law, Waseda Univ., 1-6-1 Nishiwaseda, Shinjuku-Ku, Tokyo,<br />
Japan�<br />
The present paper proposes a new measure of functional load for segments<br />
and features. In a nut shell, it is based on word frequencies and the<br />
number of minimal pairs in which the relevant segment/feature is crucial<br />
in distinction. For example, minimal pairs distinguished only by /t/ are<br />
most frequent in English while those distinguished by /k/ are most frequent<br />
in Japanese. As for functional load of features, single-feature contrasts<br />
and multiple-feature contrasts are incorporated in the calculation. In<br />
Japanese, �high� alone distinguishes the largest number of minimal pairs<br />
while �voice� distinguishes words most frequently in cooperation with<br />
other features. Word frequency and familiarity database for English and<br />
Japanese are used to observe the commonalities and differences in both<br />
languages with respect to the proposed measure of functional load. This<br />
line of analysis suggests a better account for a certain phonological process<br />
being more typical in one language but not in the other. Functional<br />
load can be thought of as a part of the top-down information from the<br />
lexicon, which interacts with the bottom-up perceptual information in the<br />
process of spoken word recognition. Not only the ease of articulation and<br />
perceptual salience but also the functional load drives phonological processes.<br />
4aSC2. Cortical representation of processing Japanese phonemic and<br />
phonetic contrasts. Seiya Funatsu �The Prefectural Univ. of Hiroshima,<br />
1-1-71 Ujinahigashi, Minami-ku, Hiroshima, 734-8558 Japan�, Satoshi<br />
Imaizumi �The Prefectural Univ. of Hiroshima, Gakuen-machi Mihara,<br />
723-0053 Japan�, Akira Hashizume, and Kaoru Kurisu �Hiroshima Univ.,<br />
Minami-ku Hiroshima, 734-8551 Japan�<br />
This study investigated how Japanese speakers process phonemic and<br />
phonetic contrasts using voiced and devoiced vowel /u/ and /�u/. During<br />
six oddball experiments, brain responses were measured using magnetoencephalography.<br />
Under the phonemic condition, a frequent stimulus<br />
/ch�ita/ was contrasted with a deviant /ts�uta/, and a frequent /ts�uta/<br />
with a deviant /ch�ita/. Under the phonetic condition, a frequent /ts�uta/<br />
was contrasted with a deviant /tsuta/, and a frequent /tsuta/ with a deviant<br />
/ts�uta/. Under the segment condition, vowel segments, /�u/ and /u/,<br />
extracted from spoken words, were contrasted. The subjects were 13 native<br />
Japanese speakers. The equivalent current dipole moment �ECDM�<br />
was estimated from the mismatch field. Under the phonetic condition, the<br />
ECDM elicited by the voiced deviant was significantly larger than that<br />
elicited by the devoiced deviant in both hemispheres (p�0.01), while<br />
there were no significant deviant-related differences in ECDM under the<br />
phonemic condition in both hemispheres. Under the segment condition,<br />
the ECDM elicited by the voiced deviant and devoiced deviant did not<br />
differ significantly in either hemispheres. These results suggested that the<br />
3247 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3247<br />
4a FRI. AM
ECDM asymmetries between the voiced and the devoiced deviant observed<br />
under the phonetic condition did not originate from the acoustical<br />
difference itself, but from the phonetic environment.<br />
4aSC3. Evaluating a model to estimate breathiness in vowels. Rahul<br />
Shrivastav, Arturo Camacho, and Sona Patel �Dept. of Commun. Sci. and<br />
Disord., Univ. of Florida, Gainesville, FL 32611�<br />
The perception of breathiness in vowels is cued by changes in aspiration<br />
noise �AH� and the open quotient �OQ� �Klatt and Klatt, J. Acoust.<br />
Soc. Am. 87�2�, 820–857 �1990��. A loudness model can be used to determine<br />
the extent to which AH masks the harmonic components in voice.<br />
The resulting partial loudness �PL� and loudness of AH �noise loudness;<br />
NL� have been shown to be good predictors of perceived breathiness<br />
�Shrivastav and Sapienza, J. Acoust. Soc. Am. 114�1�, 2218–2224 �2005��.<br />
The levels of AH and OQ were systematically manipulated for ten synthetic<br />
vowels. Perceptual judgments of breathiness were obtained and regression<br />
functions to predict breathiness from NL/PL were derived. Results<br />
show breathiness to be a power function of NL/PL when NL/PL is<br />
above a certain threshold. This threshold appears to be affected by the<br />
stimulus pitch. A second experiment was conducted to determine if the<br />
resulting power function could be used to estimate breathiness in natural<br />
voices. The breathiness of novel stimuli, both natural and synthetic, was<br />
determined in a listening test. For comparison, breathiness for the same<br />
stimuli was also estimated using the power function obtained previously.<br />
Results show the extent to which findings can be generalized. �Research<br />
supported by NIH/R21DC006690.�<br />
4aSC4. Predicting vowel inventories: The dispersion-focalization<br />
theory revisited. Roy Becker �Dept. of Linguist., Univ. of California<br />
Los Angeles, 3125 Campbell Hall, Los Angeles, CA 90095-1543�<br />
A revision of the dispersion-focalization theory �DFT� �Schwartz<br />
et al., J. Phonetics 25, 233–253 �1997�� is presented. Like DFT, the current<br />
computational model incorporates the center of gravity effect �COG�<br />
of 3.5-Bark spectral integration, but it deviates from DFT in that the COG<br />
contributes to the actual values and reliability weights of the perceived<br />
formants of vowels. The COG is reinterpreted as a domain of acceleration<br />
towards formant merger: the percepts of formants less than 3.5 Barks apart<br />
are perturbed towards one another in a nonlinear yet continuous fashion<br />
and their weights are increased, but perceptual merger and weight maximization<br />
occur only when the acoustic distance is about 2 Bark. Like other<br />
dispersion-based models, inventories are evaluated predominantly according<br />
to the least dispersed vowel pair, where dispersion is measured as the<br />
weighted Euclidean distance between the vowels coordinates �the first two<br />
perceived formants�. Yet in the current model the weights are determined<br />
dynamically, in a well-principled manner. This model improves existing<br />
models in predicting certain universal traits, such as series of front<br />
rounded vowels in large vowel inventories, as emergent properties of certain<br />
local maxima of the inventory dispersion evaluation function, without<br />
sacrificing predictive adequacy for smaller inventories.<br />
4aSC5. Matching fundamental and formant frequencies in vowels.<br />
Peter F. Assmann �School of Behavioral and Brain Sci., Univ. of Texas at<br />
Dallas, Box 830688, Richardson, TX 75083�, Terrance M. Nearey �Univ.<br />
of AB, Edmonton, AB, Canada T6E 2G2�, and Derrick Chen �Univ. of<br />
Texas at Dallas, Richardson, TX 75083�<br />
In natural speech, there is a moderate correlation between fundamental<br />
frequency �F0� and formant frequencies �FF� associated with differences<br />
in larynx and vocal tract size across talkers. This study asks whether<br />
listeners prefer combinations of mean F0 and mean FF that mirror the<br />
covariation of these properties. The stimuli were vowel triplets �/i/-/a/-/u/�<br />
spoken by two men and two women and subsequently processed by Kawahara’s<br />
STRAIGHT vocoder. Experiment 1 included two continua, each<br />
containing 25 vowel triplets: one with the spectrum envelope �FF� scale<br />
factor fixed at 1.0 �i.e., unmodified� and F0 varied over �2 oct, the other<br />
with F0 scale factor fixed at 1.0 and FF scale factors between 0.63 and<br />
1.58. Listeners used a method of adjustment procedure to find the ‘‘best<br />
voice’’ in each set. For each continuum, best matches followed a unimodal<br />
distribution centered on the mean F0 or mean FF �F1, F2, F3� observed in<br />
measurements of vowels spoken by adult males and females. Experiment<br />
2 showed comparable results when male vowels were scaled to the female<br />
range and vice versa. Overall the results suggest that listeners have an<br />
implicit awareness of the natural covariation of F0 and FF in human<br />
voices.<br />
4aSC6. Acoustic cues for distinguishing consonant sequences in<br />
Russian. Lisa Davidson and Kevin Roon �Linguist. Dept., New York<br />
Univ., 719 Broadway, 4th Fl, New York, NY 10003,<br />
lisa.davidson@nyu.edu�<br />
In Russian, the same consonant sequences are permitted in various<br />
environments. Consequently, the presence of a word boundary or reduced<br />
vowel can be phonologically contrastive �e.g. �z.d0vat j ] ‘‘to assign,’’<br />
�zd0vat j ] ‘‘to turn in’’�, and both learners and experienced listeners likely<br />
rely on fine acoustic cues to discriminate the phonotactic structures they<br />
hear. In this study, the acoustic characteristics of consonant sequences are<br />
examined to establish which cues may distinguish �a� word-initial clusters<br />
�C1C2�; �b� consonant-schwa-consonant sequences �C1VC2�; and�c� sequences<br />
divided by a word boundary �C1#C2�. For all three sequence<br />
types, native Russian speakers produced phrases containing three categories<br />
of target sequences: stop�consonant, fricative�consonant, and nasal<br />
�consonant. Results show no significant differences in C1 burst duration<br />
for initial stops, though a longer interconsonantal duration is a reliable cue<br />
to schwa presence in C1VC2. C2 is significantly longer in C#C than in<br />
other sequences. For clusters, when C1 is a stop, there are no significant<br />
differences in duration with other sequence types, but other C1’s are significantly<br />
shorter. This suggests that articulatory overlap, which may lead<br />
to C1 shortening for fricative or nasal-initial clusters, is reduced in stopinitial<br />
clusters to ensure that the stop is released and recoverable. �Research<br />
supported by NSF.�<br />
4aSC7. Lipread me now, hear me better later: Crossmodal transfer of<br />
talker familiarity effects. Kauyumari Sanchez, Lawrence D.<br />
Rosenblum, and Rachel M. Miller �Dept. of Psych., Univ. of California,<br />
Riverside, Riverside, CA�<br />
There is evidence that for both auditory and visual speech perception<br />
�lipreading� familiarity with a talker facilitates speech recognition<br />
�Nygaard et al., Psychol. Sci. 5, 42�1994�; Yakel et al., Percept. Psychophys.<br />
62, 1405 �2000��. Explanations of these effects have concentrated on<br />
the retention of talker information specific to each of these modalities. It<br />
could be, however, that some amodal, talker-specific articulatory style<br />
information is retained to facilitate speech perception in both modalities. If<br />
this is true, then experience with a talker in one modality should facilitate<br />
perception of speech from that talker in the other modality. To test this<br />
prediction, subjects were given one hour of experience lipreading from a<br />
talker and were then asked to recover speech-in-noise from either this<br />
same talker or from a different talker. Results revealed that subjects who<br />
lipread and heard speech from the same talker performed better on the<br />
speech-in-noise task than did subjects who lipread from one talker and<br />
then heard speech from a different talker.<br />
4aSC8. Acoustic patterns of Japanese voiced velar stops. James<br />
Dembowski and Katsura Aoyama �Dept. Speech-Lang. & Hearing Sci.,<br />
Texas Tech Univ. Health Sci. Ctr., 3601 4th St., Lubbock, TX 79430-6073,<br />
james.dembowski@ttuhsc.edu�<br />
This presentation focuses on Japanese voiced velar /g/. The phoneme<br />
/g/ in VCV contexts is said to be characterized by a distinctive wedgeshaped<br />
formant pattern in which F2 and F3 converge toward one fre-<br />
3248 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3248
quency in the transition from vowel to stop closure, and then diverge as<br />
the vocal tract opens from the stop release to the following vowel. This<br />
pattern was examined using acoustic and kinematic data from an x-ray<br />
microbeam database of Japanese speakers, which is comparable to the<br />
English language x-ray microbeam speech production database �Hashi<br />
et al., J. Acoust. Soc. Am. 104, 2426–2437 �1998��. Japanese speakers<br />
produced the expected wedge-shaped formant pattern in isolated VCV<br />
nonsense syllables, but rarely, if ever, produced this pattern in connected<br />
speech. In contrast, English speakers more frequently produced the expected<br />
formant pattern in connected speech, though the pattern was less<br />
reliably present than in isolated VCVs and varied considerably within and<br />
across speakers. These observations highlight substantial differences between<br />
controlled laboratory speech and meaningful connected speech, as<br />
well as differences in the ways that phonemes are manifested by different<br />
linguistic communities. These data also potentially illuminate the relationship<br />
among phonetic, acoustic, and kinematic levels of speech production.<br />
4aSC9. Finding perceptual categories in multidimensional acoustic<br />
spaces. Eric Oglesbee and Kenneth de Jong �Dept. of Linguist., Indiana<br />
Univ., Bloomington, IN 47405, eoglesbe@indiana.edu�<br />
Examining phonetic categorization in multidimensional acoustic<br />
spaces poses a number of practical problems. The traditional method of<br />
forced identification of an entire stimulus space becomes prohibitive when<br />
the number and size of acoustic dimensions becomes increasingly large. In<br />
response to this, Iverson and Evans �ICPhS �2003�� proposed an adaptive<br />
tracking algorithm for finding best exemplars of vowels in a multidimensional<br />
acoustic space. Their algorithm converged on best exemplars in a<br />
relatively small number of trials; however, the search method took advantage<br />
of special properties of the vowel space in order to achieve rapid<br />
convergence. In this paper, a more general multidimensional search algorithm<br />
is proposed and analyzed for inherent biases. Then, using the algorithm,<br />
the phonetic categorization of /p/ and /b/ in a five-dimensional<br />
acoustic space by native speakers of English is tested. Results showed that<br />
�a� there were no substantial long-term biases in the search method and �b�<br />
the algorithm appeared to identify important acoustic dimensions in the<br />
identification of /p/ and /b/ using relatively few trials. �Work supported by<br />
NSF BCS-04406540.�<br />
4aSC10. On the perception of epenthetic stops in American English.<br />
Amalia Arvaniti, Ryan Shosted, and Cynthia Kilpatrick �Dept. of<br />
Linguist., UCSD, 9500 Gilman Dr., La Jolla, CA 92093,<br />
amalia@ling.ucsd.edu�<br />
This study examines the perception of epenthetic stops in American<br />
English. Stimuli were created from prince, prints, mince, mints, quince,<br />
and quints by removing all traces of /t/ and splicing in 0–72 ms of silence,<br />
in 12-ms steps, with or without a following burst. Subjects saw the minimal<br />
pairs on screen and selected the word they heard. It was hypothesized<br />
that stimuli with bursts and longer closure would result in more t responses<br />
�prince identified as prints� and that frequent words �prince/<br />
prints� would be more difficult to distinguish than infrequent words<br />
�quince/quints�, as our production results suggest that frequent pairs are<br />
more likely to be produced similarly. Results from 19 subjects show<br />
shorter response times with longer silence intervals, but no effect of burst<br />
or stimulus identity. Similarly, stimuli with bursts were not identified as<br />
nts words more frequently than those without. Generally, stimuli originating<br />
from nts words were more likely to be identified as such if they<br />
contained a burst, while the opposite was true for stimuli from nce words.<br />
Finally, frequent words were less likely to be correctly identified than<br />
infrequent words, suggesting that /t/ epenthesis is not as widespread<br />
throughout the lexicon as generally believed.<br />
4aSC11. Phonetic alignment to visual speech. Rachel M. Miller,<br />
Lawrence D. Rosenblum, and Kauyumari Sanchez �Dept. of Psych.,<br />
Univ. of California, Riverside, Riverside, CA 92521�<br />
Talkers are known to produce allophonic variation based, in part, on<br />
the speech of the person with whom they are talking. This subtle imitation,<br />
or phonetic alignment, occurs during live conversation and when a talker<br />
is asked to shadow recorded words �e.g., Shockley, et al., Percept. Psychophys.<br />
66, 422 �2004��. What is yet to be determined is the nature of the<br />
information to which talkers align. To examine whether this information is<br />
restricted to the acoustic modality, experiments were conducted to test if<br />
talkers align to visual speech �lipread� information. Normal-hearing subjects<br />
were asked to watch an actor silently utter words, and to identify<br />
these words by saying them out loud as quickly as possible. These shadowed<br />
responses were audio recorded and naive raters compared these<br />
responses to the actors auditory words �which had been recorded along<br />
with the actors visual tokens�. Raters judged the shadowed words as<br />
sounding more like the actors words than did baseline words, which had<br />
been spoken by subjects before the shadowing task. These results show<br />
that phonetic alignment can be based on visual speech, suggesting that its<br />
informational basis is not restricted to the acoustic signal.<br />
4aSC12. New anthropomorphic talking robot—investigation of the<br />
three-dimensional articulation mechanism and improvement of the<br />
pitch range. Kotaro Fukui, Yuma Ishikawa, Takashi Sawa, Eiji Shintaku<br />
�Dept. of Mech. Eng., Waseda Univ., 3-4-1 Ookubo, Shinjuku-ku, Tokyo,<br />
Japan�, Masaaki Honda �Waseda Univ., Saitama, Japan�, and Atsuo<br />
Takanishi �Waseda Univ., Shinjuku-ku, Tokyo, Japan�<br />
We developed a new three-dimensional talking robot WT-6 �Waseda<br />
Talker-No. 6�, which generates speech sounds by mechanically simulating<br />
articulatory motions as well as aeroacoustic phenomenon in the vocal<br />
tract. WT-6 consists of a 17-DOF mechanism �1-DOF lungs, 5-DOF vocal<br />
cords, 1-DOF jaws, 5-DOF tongue, and 4-DOF lips�. It has 3-D lips,<br />
tongue, jaw, and velum, which form the 3-D vocal tract structure. It also<br />
has an independent jaw opening/closing mechanism, which controls the<br />
relative tongue position in the vocal tract as well as the oral opening. The<br />
previous robot, which had a 2-D tongue �J. Acoust. Soc. Am. 117, 2541<br />
�2005��, was not enough to realize precise closure to produce humanlike<br />
consonants such as /s/ or /r/. The new tongue, which could be controlled to<br />
form the 3-D shape, makes it possible to produce more realistic vocal tract<br />
shape. The vocal cord model was also improved by adding a new pitch<br />
control mechanism pushing from the side of the vocal cords. The pitch<br />
range becomes broader than the previous robot, which is enough to reproduce<br />
normal human speech. Preliminary experimental results showed that<br />
synthesized speech quality improves for vowels /a/, /u/ and /o/. Some<br />
experimental results and video demonstration of the talking robot will be<br />
presented.<br />
4aSC13. The role of familiarity, semantic context, and amplitudemodulation<br />
on sentence intelligibility. Tom Carrell �Univ. of<br />
Nebraska—Lincoln, Lincoln, NE 68583, tcarrell@unl.edu� and Dawna<br />
Lewis �Boys Town Natl. Res. Hospital, Omaha, NE 68131�<br />
Amplitude modulation has been demonstrated to greatly improve the<br />
intelligibility of time-varying sinusoidal �TVS� sentences �Carrell and<br />
Opie, Percept. Psychophys. 52 �1992�; Barker and Cooke, Speech Commun.<br />
27 �1999�; Hu and Wang, Proceedings of ICASSP-02 �2002��. Ithas<br />
been argued that the improvement is due to a bottom-up process that<br />
causes the acoustically independent components of the sentences to be<br />
perceptually grouped for further analysis by the auditory system. A recent<br />
study �Shapley and Carrell, J. Acoust. Soc. Am. 118 �2005�� indicated that<br />
3249 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3249<br />
4a FRI. AM
semantic information did not influence intelligibility levels of TVS or<br />
modulated TVS sentences. In virtually all other studies in which speech<br />
was distorted or degraded its intelligibility was improved by appropriate<br />
semantic context �Miller, et al., JEP41 �1951��. It is possible that listeners’<br />
unfamiliarity with TVS speech might account for the difference. With<br />
one exception every study that has employed this type of stimulus provided<br />
the listener with very few practice sentences �Lewis, AAA �2005��.<br />
The present experiment manipulated listeners’ familiarity with TVS sentences<br />
to test this notion. Participants were presented with high- and lowpredictability<br />
TVS and modulated-TVS sentences. Familiarity had a large<br />
effect on perception and intelligibility. Interpretation of previous findings<br />
is reconsidered in this light.<br />
4aSC14. On the relation of apparent naturalness to phonetic<br />
perceptual resolution of consonant manner. Robert E. Remez, Claire<br />
A. Landau, Daria F. Ferro, Judith Meer, and Kathryn Dubowski �Dept. of<br />
Psych., Barnard College, 3009 Broadway, New York, NY 10027�<br />
How does the qualitative experience of speech influence phonetic perception?<br />
Our perceptual studies of consonant place and voicing have revealed<br />
a dichotomous relation between phonetic sensitivity and naturalness.<br />
Auditory quality and phonetic sensitivity sometimes co-vary, while<br />
in other conditions phonetic sensitivity is indifferent to huge variation in<br />
naturalness. New tests are reported extending the research to the dimension<br />
of manner, a contrast correlated with qualitatively distinct acoustic<br />
constituents in normal production. Speech synthesis techniques were used<br />
to create naturalness variants through �1� variation in the excitation of a<br />
synthetic voicing source and �2� variation in the bandwidth of the formant<br />
centers. Listeners calibrated the relative naturalness of items drawn from<br />
the test series, and the acuity of perceivers to the contrast between fricative<br />
and stop manner was estimated with cumulative d� across the series in<br />
identification tests. Combined with our prior findings, these new tests<br />
show how intelligibility and naturalness can be either perceptually orthogonal<br />
or contingent aspects of consonant dimensions, offering a tool to<br />
understand normative functions in speech perception. �Research supported<br />
by NIH �DC00308�.�<br />
4aSC15. Effects of signal levels on vowel formant discrimination for<br />
normal-hearing listeners. Chang Liu �Dept. of Commun. Sci. and<br />
Disord., Wichita State Univ., Wichita, KS 67260�<br />
The goal of this study was to examine effects of signal levels on vowel<br />
formant discrimination. Thresholds of formant discrimination were measured<br />
for F1 and F2 of four vowels in isolated vowels and sentences at<br />
three intensity levels: 70, 85, and 100 dB SPL for normal-hearing listeners<br />
using a three-interval, two-alternative forced-choice procedure with a twodown,<br />
one-up tracking algorithm. Results showed that formant thresholds<br />
were significantly affected by formant frequency, linguistic context, and<br />
signal levels. Thresholds of formant discrimination were increased as formant<br />
frequency increased and linguistic context changed from isolated<br />
vowels to sentences. The signal level indicated a rollover effect in which<br />
formant thresholds at 85 dB SPL are lower than at either 70 or 100 dB SPL<br />
in both isolated vowels and sentences. This rollover level effect could be<br />
due to reduced frequency selectivity and reduction in active cochlear nonlinearity<br />
at high signal levels for normal-hearing listeners. Excitation and<br />
loudness models will be explored to account for the level effect on formant<br />
discrimination of isolated vowels.<br />
4aSC16. Study on nonaudible murmur speaker verification using<br />
multiple session data. Mariko Kojima, Hiromichi Kawanami, Hiroshi<br />
Saruwatari, Kiyohiro Shikano �Nara Inst. of Sci. and Technol. 8916-5<br />
Takayama-cho Ikoma-shi, Nara, 630-0192 Japan�, and Tomoko Matsui<br />
�The Inst. of Statistical Mathematics, Minato-ku, Tokyo, 106-8569 Japan�<br />
A study on speaker verification with nonaudible murmur �NAM� segments<br />
using multiple session data was conducted. NAM is different from<br />
normal speech and is difficult for other people to catch. Therefore, a textdependent<br />
verification strategy can be used in which each user utters her/<br />
his own keyword phrase so that not only speaker-specific but also<br />
keyword-specific acoustic information is utilized. A special device called a<br />
NAM microphone worn on the surface of the skin below the mastoid bone<br />
is used to catch NAM because it is too low to be recorded using ordinary<br />
microphones. However, it is tolerant to exterior noise. This strategy is<br />
expected to yield relatively high performance. NAM segments, which consist<br />
of multiple short-term feature vectors, are used as input vectors to<br />
capture keyword-specific acoustic information well. To handle segments<br />
with a large number of dimensions, a support vector machine �SVM� is<br />
used. In experiments using NAM data uttered by 19 male and 10 female<br />
speakers in several different sessions, robustness against session-tosession<br />
data variation is examined. The effect of segment length is also<br />
investigated. The proposed approach achieves equal error rates of 0.04%<br />
�male� and 1.1% �female� when using 145-ms-long NAM segments.<br />
4aSC17. Sequential contrast or compensation for coarticulation?<br />
John Kingston, Daniel Mash, Della Chambless, and Shigeto Kawahara<br />
�Linguist. Dept., Univ. of Massachusetts, Amherst, MA 01003-9274�<br />
English listeners identify a stop ambiguous between /t/ and /k/ more<br />
often as ‘‘k’’ after /s/ than /sh/ �Mann and Repp, 1981�. Judgments shift<br />
similarly after a fricative ambiguous between /s/ and /sh/ when its identity<br />
is predictable from a transitional probability bias but perhaps not from a<br />
lexical bias �Pitt and McQueen, 1998; cf. Samuel and Pitt, 2003�. In replicating<br />
these experiments, we add a discrimination task to distinguish<br />
between the predictions of competing explanations for these findings: listeners<br />
respond ‘‘k’’ more often after /s/ because they compensate for the<br />
fronting of the stop expected from coarticulation with /s/ or because a stop<br />
with an F3 onset frequency midway between /t/’s high value and /k/’s low<br />
value sounds lower after the /s/’s high-frequency energy concentration.<br />
The second explanation predicts listeners will discriminate /s-k/ and /sh-t/<br />
sequences better than /s-t/ and /sh-k/ sequences because sequential contrast<br />
exaggerates the spectral differences between /s-k/’s high-low intervals<br />
and /sh-t/’s low-high intervals and distinguishes them more perceptually<br />
than /s-t/’s high-high intervals and /sh-k/’s low-low intervals.<br />
Compensation for coarticulation predicts no difference in discriminability<br />
between the two pairs because it does not exaggerate differences between<br />
the two intervals. �Work supported by NIH.�<br />
4aSC18. Acoustics and perception of coarticulation at a distance.<br />
Karen Jesney, Kathryn Pruitt, and John Kingston �Ling. Dept., Univ. of<br />
Massachusetts, Amherst, MA 01003-9274�<br />
CVC syllables were recorded from two speakers of American English,<br />
in which the initial and final stops ranged over /b,d,g/ and the vowel<br />
ranged over /i,I,e,E,ae,u,U,o,O,a/. F2 locus equations differed systematically<br />
as a function of the place of articulation of the other stop. These<br />
equation’s slope and y intercepts were used to synthesize initial /g-b/ and<br />
/g-d/ continua in CVC syllables in which the final stop ranged over /b,d,g/<br />
and the vowel over /e,o,a/, and the resulting stimuli were presented to<br />
listeners for identification. Listeners responded g significantly more often<br />
to both continua when the final stop was /d/ rather than /b/; the number of<br />
g responses fell between the /d/ and /b/ extremes for final /g/. This difference<br />
between final /d/ vs. /b/ is only observed when the intervening vowel<br />
is back /o,a/ and is actually reversed weakly when it is front /e/. Listeners<br />
also respond g significantly more often when the final stop is /g/ rather<br />
than /b/ and the vowel is /o,a/ but not �e�. Segments do coarticulate at a<br />
distance, listeners take this coarticulation into account, and perceptual adjustments<br />
depend on the segments through which the coarticulation is<br />
expressed. �Supported by NIH.�<br />
3250 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3250
4aSC19. Diphones, lexical access, and the verbal transformation<br />
effect. James A. Bashford, Jr., Richard M. Warren, and Peter W. Lenz<br />
�Dept. of Psych., Univ. of Wisconsin—Milwaukee, P.O. Box 413,<br />
Milwaukee, WI 53201-0413, bashford@uwm.edu�<br />
When listeners are presented with a repeating verbal stimulus, adaptation<br />
occurs and perception of the stimulus is replaced by perception of a<br />
competitor. The present study examines the first of these verbal transformations<br />
reported by 180 listeners who were presented with lexical and<br />
nonlexical consonantvowel �CV� syllables that varied in frequencyweighted<br />
neighborhood density �FWND�. These stimuli were produced by<br />
pairing the six English stop consonants with a set of three vowels. As<br />
expected, the majority of initial illusory forms �78%� were neighbors,<br />
differing from the stimulus by a single phoneme, and the proportion of<br />
lexical neighbors increased with stimulus FWND. Interestingly, FWND<br />
had opposite effects upon the lability of consonants and vowels: There was<br />
a strong positive correlation �r�0.79, F(17)�26.2, p�0.0001] between<br />
FWND and the number of consonant transformations, and in contrast,<br />
there was a strong negative correlation �r��0.78, F(17)�24.9, p<br />
�0.0001] between FWND and the number of vowel transformations. The<br />
implications of these and other findings with these simple diphones will be<br />
discussed in relation to current activation-competition theories of spoken<br />
word recognition. �Work supported by NIH.�<br />
4aSC20. Acoustic analysis and perceptual evaluation of nasalized ÕgÕ<br />
consonant in continuous Japanese. Hisao Kuwabara �Teikyo Univ. of<br />
Sci. & Technol., Uenohara, Yamanshi 409-0193, Japan�<br />
It is well known that the /g/ consonant, a velar voiced plosive, in<br />
Japanese continuous speech is often nasalized unless it appears at the<br />
word-initial position. The nasalized /g/ consonant takes place in dialects<br />
mainly spoken in northern districts including the Tokyo area where the<br />
standard Japanese is spoken. However, the number of nasalized /g/ consonants<br />
is said to be decreasing year by year according to a survey. This<br />
paper deals with acoustic and perceptual analysis of this phenomenon. Test<br />
materials used in this experiment are read version of Japanese short sentences<br />
by NHK’s �Japan Broadcasting Corporation� professional announcers.<br />
Each sentence includes at least one /g/ consonant that would likely be<br />
pronounced as nasalized. An evaluation test reveals that less than 60% of<br />
nasalization has been found to occur for /g/ consonants for which 100%<br />
nasalization had been observed decades ago. Acoustic analysis for nasalized<br />
and non-nasalized /g/ sounds has been performed mainly through<br />
waveform parameters. It has been found that the power ratio between<br />
consonant and vowel is the most effective parameter for distinguishing<br />
nasals from non-nasals, but it is highly speaker dependent.<br />
4aSC21. Production and perception of place of articulation errors.<br />
Adrienne M. Stearns and Stefan A. Frisch �Univ. of South Florida, 4202<br />
E. Fowler Ave., PCD1017, Tampa, FL 33620�<br />
Using ultrasound to examine speech production is gaining popularity<br />
because of its portability and noninvasiveness. This study examines ultrasound<br />
recordings of speech errors. In experiment 1, ultrasound images of<br />
participants’ tongues were recorded while they read tongue twisters. Onset<br />
stop closures were measured using the angle of the tongue blade and<br />
elevation of the tongue dorsum. Measurements of tongue twisters were<br />
compared to baseline production measures to examine the ways in which<br />
erroneous productions differ from normal productions. It was found that<br />
an error could create normal productions of the other category �categorical<br />
errors� or abnormal productions that fell outside the normal categories<br />
�gradient errors�. Consonant productions extracted from experiment 1<br />
were presented auditory-only to naive listeners in experiment 2 for identification<br />
of the onset consonant. Overwhelmingly, the participants heard<br />
normal productions and gradient error productions as the intended sound.<br />
Categorical error productions were judged to be different from the intended<br />
sound. The only effect of erroneous production on perception appears<br />
to be a slight increase in reaction time, which may suggest that error<br />
tokens are abnormal in some way not measured in this study.<br />
4aSC22. Role of linguistic experience on audio-visual perception of<br />
English fricatives in quiet and noise backgrounds. Yue Wang,<br />
Haisheng Jiang, Chad Danyluck �Dept. of Linguist., Simon Fraser Univ.,<br />
Burnaby, BC, V5A 1S6, Canada�, and Dawn Behne �Norwegian Univ. of<br />
Sci. and Technol., Trondheim, Norway�<br />
Previous research shows that for native perceivers, visual information<br />
enhances speech perception, especially when auditory distinctiveness decreases.<br />
This study examines how linguistic experience affects audiovisual<br />
�AV� perception of non-native �L2� speech. Native Canadian English<br />
perceivers and Mandarin perceivers with two levels of English<br />
exposure �early and late arrival in Canada� were presented with English<br />
fricative-initial syllables in a quiet and a caf-noise background in four<br />
ways: audio-only �A�, visual-only �V�, congruent AV, and incongruent AV.<br />
Identification results show that for all groups, performance was better in<br />
the congruent AV than A or V condition, and better in quiet than in cafnoise<br />
background. However, whereas Mandarin early arrivals approximate<br />
the native English patterns, the late arrivals showed poorer identification,<br />
more reliance on visual information, and greater audio-visual integration<br />
with the incongruent AV materials. These findings indicate that although<br />
non-natives were more attentive to visual information, they failed to use<br />
the linguistically significant L2 visual cues, suggesting language-specific<br />
AV processing. Nonetheless, these cues were adopted by the early arrivals<br />
who had more L2 exposure. Moreover, similarities across groups indicate<br />
possible perceptual universals involved. Together they point to an integrated<br />
network in speech processing across modalities and linguistic backgrounds.<br />
�Work supported by SSHRC.�<br />
4aSC23. Voicing of ÕhÕ in the Texas Instrument MIT „TIMIT…<br />
database. Laura Koenig �Haskins Labs, 300 George St., New Haven,<br />
CT 06511 and Long Island Univ.-Brooklyn�<br />
Although English /h/ is traditionally described as voiceless, authors<br />
have long recognized that voiced allophones exist, especially in unstressed,<br />
intervocalic positions. In past work, we have suggested that fully<br />
voiced /h/ may be more common in men than women, but our subject<br />
population was limited in number and dialectal diversity. In this study, we<br />
use the TIMIT database to obtain measures of /h/ production in men and<br />
women speaking multiple dialects of American English. Our analysis focuses<br />
on the /h/ initiating the word ‘‘had’’ in a sentence produced by all<br />
speakers in the database: ‘‘She had your dark suit...’’Eachtokenof/h/<br />
is coded for whether �a� the /h/ is deleted �i.e., auditorily imperceptible�;<br />
and, if /h/ is present, whether �b� voicing continues unbroken and �c� there<br />
is visible aspiration noise in the speech signal. This analysis will provide<br />
baseline data on /h/ realization in a common sentence context. We will<br />
also carry out follow-up analyses on selected utterances to gain more<br />
insight into the effects of phonetic context, stress, and lexical type �e.g.,<br />
content versus function word� on the characteristics of English /h/. �Work<br />
supported by NIH.�<br />
4aSC24. On distinctions between stops and similar-place weak<br />
fricatives. James M. Hillenbrand �Speech Pathol. and Audiol., MS5355,<br />
Western Michigan Univ., Kalamazoo, MI 49008� and Robert A. Houde<br />
�Ctr. for Commun. Res., Rochester, NY 14623�<br />
There is an extensive body of literature on the acoustic properties of<br />
both stops and fricatives. However, little attention has been paid to the<br />
acoustic features that might distinguish these two sound categories. This is<br />
unsurprising in the case of stops versus strong fricatives since these<br />
sounds are rarely confused. Stops and weak fricatives, on the other hand,<br />
are frequently confused �G.A. Miller and P.E. Nicely, J. Acoust. Soc. Am.<br />
27, 338–352 �1955��. The present study was undertaken in a search for<br />
acoustic features that might distinguish the stop/weak fricative pairs<br />
/b/–/v/ and /d/–/dh/ �i.e., /d/ vs voiced th�. Speech material consisted of<br />
CV and VCV syllables spoken by five men and five women, using the<br />
vowels /a/, /i/, and /u/. A combination of two features reliably separated<br />
3251 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3251<br />
4a FRI. AM
the stops from weak fricatives: �1� intensity during the consonant occlusion<br />
interval �typically greater for the fricatives�, and �2� the rate of increase<br />
in mid- and high-frequency energy �above 1 kHz� associated with<br />
consonant release �typically greater for the stops�.<br />
4aSC25. Salience of virtual formants as a function of the frequency<br />
separation between spectral components. Robert A. Fox, Ewa<br />
Jacewicz, Chiung-Yun Chang, and Jason D. Fox �Speech Acoust. and<br />
Percpt. Labs, Dept. of Speech and Hearing Sci., Ohio State, 110 Pressey<br />
Hall, 1070 Carmack Rd., Columbus, OH 43210-1002�<br />
The center-of-gravity �COG� hypothesis proposed by Chistovich and<br />
others for the perception of static vowels suggests that auditory spectral<br />
integration may occur when two or more formants fall within a 3.5 bark<br />
bandwidth. While several studies have examined the bandwidth limits of<br />
such integration, this study examines the extent to which spectral integration<br />
is uniform within this putative 3.5-bark range. We examine the perceptual<br />
salience of virtual formants produced by modifying the spectral<br />
COG of two closely spaced narrow-bandwidth resonances. Three different<br />
vowel series were created: �i-(�, �#-� and �.-�. A second set of vowels<br />
was then created in which one of the formants �F1 in �i-(�, F2in�#-� and<br />
F3 in �.-�� was replaced by a virtual formant whose COG matched that<br />
of the formant that had been removed. The frequency separation between<br />
the two component resonances was then systematically varied between 1.5<br />
and 3.5 barks and a singleinterval 2AFC vowel identification task was<br />
used to obtain estimates of vowel quality for each series step. Results will<br />
be discussed in terms of whether the spectral integration effects within the<br />
3.5 bark decline as the frequency separation between the resonance components<br />
increases. �Work supported by NIDCD R01DC00679-01A1.�<br />
4aSC26. Frequency effects in phoneme processing. Danny R. Moates,<br />
Noah E. Watkins, Zinny S. Bond, and Verna Stockmal �Inst. for the<br />
Empirical Study of Lang., Ohio Univ., Athens, OH 45701,<br />
moates@ohio.edu�<br />
Are phonological segments activated during word recognition in proportion<br />
to their frequency of use? Previous evidence for this hypothesis<br />
�Moates et al., Laboratory Phonology 7, edited by Gussenhoven and<br />
Warner �Mouton de Gruyter, 2002�� used a word reconstruction task. The<br />
present study used an online task, the gating task, in which progressively<br />
longer fragments of a word are presented to listeners who must identify<br />
the word in as few gates as possible. High- and low-frequency segments<br />
were contrasted by presenting them in word pairs that differed only in<br />
those two segments, e.g., label versus cable, where /l/ is used more frequently<br />
than /k/ �M. K. Vitevich and P. A. Luce, Behav. Res. Methods,<br />
Instrum. Comput. 36, 481–487 �2004��. We constructed 30 pairs of twosyllable<br />
words for which the contrasting segments were at the first syllable<br />
onset, 30 more for the second syllable onset, and 30 more for the coda of<br />
a syllable. Identification judgments were gathered from 120 participants. t<br />
tests showed high-frequency segments to be identified at significantly earlier<br />
gates than their matched low-frequency segments for first onset and<br />
coda, but not second onset. These results offer further evidence for sublexical<br />
processes in spoken word recognition.<br />
4aSC27. The clustering coefficient of phonological neighborhoods<br />
influences spoken word recognition. Michael Vitevitch �Dept. of<br />
Psych., Univ. of Kansas, 1415 Jayhawk Blvd., Lawrence, KS 66045-7556�<br />
Neighborhood density refers to the number of words, or neighbors,<br />
that are phonologically related to a given word. For example, the words<br />
BAT, MAT, CUT, and CAN �among others� are considered phonological<br />
neighbors of the word CAT. In contrast, the clustering coefficient of the<br />
neighborhood refers to the proportion of phonological neighbors that are<br />
also neighbors of each other. Among the neighbors of CAT, the words<br />
BAT and MAT are neighbors of each other, but the words BAT and CAN<br />
are not neighbors of each other. Despite the stimulus words having the<br />
same number of neighbors overall, the results of an auditory lexical decision<br />
task showed that words with a high clustering coefficient �i.e., most<br />
neighbors were also neighbors of each other� were responded to more<br />
quickly than words with a low clustering coefficient �i.e., few neighbors<br />
were also neighbors of each other�. These results suggest that some aspects<br />
of phonological similarity �i.e., clustering coefficient� might facilitate<br />
lexical activation, whereas other aspects of phonological similarity<br />
�i.e., neighborhood density� influence a later, decision stage of processing<br />
characterized by competition among activated word-forms. �Work supported<br />
by NIH.�<br />
4aSC28. The influence of noise and reverberation on vowel<br />
recognition: Response time. Magdalena Blaszak and Leon Rutkowski<br />
�Div. of Rm. Acoust. and Psychoacoustics, Adam Mickiewicz Univ.,<br />
Umultowska 85, 61-614 Poznan, Poland�<br />
This study examines the perceptual effect of two types of noise and<br />
reverberation on vowel recognition. Multitalker babble and traffic noise<br />
�European Standard EN 1793–3� were generated simultaneously with Polish<br />
vowels /a, e, i, o, u, y/ in two different sound fields and an anechoic<br />
chamber. The experiment was performed under various conditions of<br />
signal-to-noise ratio (�9, �6, �3, 0, �3, no noise�. A new procedure<br />
for listeners’ selection based on the Bourdon’s psychometrics test was<br />
proposed. The effects of noise and reverberation were quantified in terms<br />
of �a� vowel recognition scores for young normal-hearing listeners �YNH�<br />
and �b� ease of listening based on the time of response and subjective<br />
estimation of difficulty. Results of the experiment have shown that �a� the<br />
response time can be a good measure of the effect of noise and reverberation<br />
on the speech intelligibility is <strong>room</strong>s, and �b� in this type of experiment,<br />
of great significance is the choice of the subjects based on the<br />
psychometric tests.<br />
4aSC29. Quantifying the benefits of sentence repetition on the<br />
intelligibility of speech in continuous and fluctuating noises. Isabelle<br />
Mercille, Roxanne Larose, Christian Giguère, and Chantal Laroche<br />
�Univ. of Ottawa, 451 Smyth Rd., Ottawa, ON, Canada K1H 8M5�<br />
Good verbal communication is essential to ensure safety in the workplace<br />
and social participation during daily activities. In many situations,<br />
speech comprehension is difficult due to hearing problems, the presence of<br />
noise, or other factors. As a result, listeners must often ask the speaker to<br />
repeat what was said in order to understand the complete message. However,<br />
there has been little research describing the exact benefits of this<br />
commonly used strategy. This study reports original data quantifying the<br />
effect of sentence repetition on speech intelligibility as a function of<br />
signal-to-noise ratio and noise type. Speech intelligibility data were collected<br />
using 18 normal-hearing individuals. The speech material consisted<br />
of the sentences from the Hearing In Noise Test �HINT� presented in<br />
modulated and unmodulated noises. Results show that repeating a sentence<br />
decreases the speech reception threshold �SRT�, as expected, but<br />
also increases the slope of the intelligibility function. Repetition was also<br />
found to be more beneficial in modulated noises �decrease in SRT by 3.2<br />
to 5.4 dB� than in the unmodulated noise �decrease in SRT by 2.0 dB�.The<br />
findings of this study could be useful in a wider context to develop predictive<br />
tools to assess speech comprehension under various conditions.<br />
4aSC30. The effect of the spectral shape changes on voice perception.<br />
Mika Ito, Bruce R. Gerratt, Norma Antonanzas-Barroso, and Jody<br />
Kreiman �Div. of Head/Neck Surgery, UCLA School of Medicine, 31-24<br />
Rehab Ctr., Los Angeles, CA 90095-1794, jkreiman@ucla.edu�<br />
Researchers have long known that the shape of the vocal source spectrum<br />
is an important determinant of vocal quality, but the details regarding<br />
the importance of individual spectral features remains unclear. Previous<br />
research indicates four spectral features, H1-H2, the spectral slope above 4<br />
kHz, the slope from 1.5–2 kHz, and the slope from 2–4 kHz, account for<br />
3252 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3252
virtually all the variability in spectral shapes. The present study provides<br />
preliminary evidence about the perceptual importance of these four features.<br />
Four series of stimuli were synthesized for each spectral parameter,<br />
in which that parameter varied in small steps. Because the perceptual<br />
salience of source parameters depends on F0 and on the spectrum of the<br />
inharmonic part of the source, series differed in the sex of the speaker<br />
�male/female� and in the NSR �noise-free/very noisy�. Listeners heard all<br />
possible pairs of voices within each series and were asked to determine<br />
whether stimuli were the same or different. We hypothesize that listeners<br />
sensitivity to H1-H2 and the slope of the spectrum from 1.5–2 kHz will be<br />
independent of noise, but that sensitivity to changes in the spectral shape<br />
above 2 kHz will depend on the amount of noise excitation present in the<br />
voice.<br />
4aSC31. The use of auditory and visual information in the perception<br />
of stress in speech. James Harnsberger, Daniel Kahan, and Harry<br />
Hollien �Inst. for Adv. Study of the Commun. Proc., Univ. of Florida,<br />
Gainesville, FL 32611�<br />
Prior work on the acoustic correlates of the perception of psychological<br />
stress in speech has suffered from the problem of quantifying and<br />
verifying the extent to which a speaker was under stress during articulation.<br />
Two experiments were conducted to address this issue. First, stressed<br />
and unstressed speech samples were elicited from 78 speakers of American<br />
English. Stressed samples were recorded by having subjects read a<br />
standard passage while under the threat of the administration of mild electric<br />
shock. Both visual and audio recordings were collected. Stress was<br />
quantified in terms of four measures: two physiological �pulse rate and<br />
galvanic skin response� and two self-report scales. Sentences from the 16<br />
speakers showing the largest differences between the stressed and unstressed<br />
conditions were then presented in a paired comparison task to 90<br />
naive listeners, 30 each in three conditions: �1� audio-only presentation of<br />
the stimuli, �2� visual-only presentation of the stimuli, and �3� audiovisual<br />
presentation of the stimuli. The results indicate that individual listeners are<br />
sensitive to stress cues in speech in all three conditions.<br />
4aSC32. Selective attention and perceptual learning of speech.<br />
Alexander L. Francis �Dept. of Speech, Lang. & Hearing Sci., Purdue<br />
Univ., West Lafayette, IN 47907�, Natalya Kaganovich, and Courtney<br />
Driscoll �Purdue Univ., West Lafayette, IN 47907�<br />
Phonetic experience can change the perceptual distance between<br />
speech sounds, increasing both within-category similarity and betweencategory<br />
distinctiveness. Such warping of perceptual space is frequently<br />
characterized in terms of changes in selective attention: Listeners are assumed<br />
to attend more strongly to category-differentiating features while<br />
ignoring less relevant ones. However, the link between the distribution of<br />
selective attention and categorization-related differences in perceptual distance<br />
has not been empirically demonstrated. To explore this relationship,<br />
listeners were given 6hoftraining to categorize sounds according to one<br />
of two acoustic features while ignoring the other. The features were voice<br />
onset time and onset f 0, which are normally correlated and can both serve<br />
as a cue to consonant voicing. Before and after training, listener’s performance<br />
on a Garner selective attention task was compared with assessment<br />
of the perceptual distance between tokens. Results suggest that training<br />
can induce both warping of perceptual space and changes in the distribution<br />
of selective attention, but the two phenomena are not necessarily<br />
related. Results are consistent with a two-stage model of perceptual learning,<br />
involving both preattentive adjustment of acoustic cue weighting and<br />
higher-level changes in the distribution of selective attention between<br />
acoustic cues. �Work supported by NIH-NIDCD 1R03DC006811.�<br />
4aSC33. Investigation of interaction between speech perception and<br />
production using auditory feedback. Masato Akagi, Jianwu Dang,<br />
Xugang Lu, and Taichi Uchiyamada �School of Information Sci., JAIST,<br />
1-1 Asahidai, Nomi, Ishikawa 923-1292, Japan�<br />
This study employed an auditory feedback paradigm with perturbed<br />
fed-back speech to investigate interaction between speech perception and<br />
production by measuring simultaneous fluctuations of speech production<br />
organs using the electromyographic �EMG� signals, articulatory movements,<br />
as well as spectral analyses, where the articulatory data were obtained<br />
by the electromagnetic articulographic �EMA� system. Chinese<br />
vowels pair �i�-�y� and Japanese vowels pairs �e�-�a�, �e�-�i� and �e�-�u�<br />
were chosen as the experimental objects. When the speaker is maintaining<br />
the first vowel, the feedback sound is randomly changed from the first<br />
vowel to the second one in each pair by manipulating the first three formants.<br />
Spectral analysis showed that a clear compensation was seen in the<br />
first and second formants of the vowels. Analyses of EMG and EMA<br />
signals also showed muscle reactivation and tongue movements to compensate<br />
for the perturbations. Latency of the compensating response is<br />
about 150 ms to start and about 290 ms for maximum compensation from<br />
the onset of the perturbation. According to the measurements, it seems that<br />
in most cases the speaker attempts to compensate for the ‘‘error’’ caused<br />
by the auditory perturbation by a real-time monitoring, and the auditory<br />
feedback takes place simultaneously often during speech production.<br />
4aSC34. Cross-ear suppression of the verbal transformation effect:<br />
Tweaking an acoustic-phonetic level. Peter W. Lenz, James A.<br />
Bashford, Jr., and Richard M. Warren �Dept. of Psych., Univ. of<br />
Wisconsin—Milwaukee, P.O. Box 413, Milwaukee, WI 53201-0413,<br />
plenz@uwm.edu�<br />
A recorded word repeating over and over undergoes a succession of<br />
illusory changes. When two images of the same repeating word are presented<br />
dichotically, with a half-cycle delay preventing fusion, the two<br />
images of the word each undergo independent illusory transformations at a<br />
rate equivalent to that of a single image �Lenz et al., J. Acoust. Soc. Am.<br />
107, 2857 �2000��. However, with one phoneme difference �e.g., ‘‘dark’’<br />
versus ‘‘dart’’�, transition rate is dramatically suppressed �Bashford et al.,<br />
J. Acoust. Soc. Am. 110, 2658 �2001��. Rates decrease with extent of<br />
feature mismatch at a single phoneme position �roughly 30% reduction<br />
with one feature mismatch and 45% with three�. Rates also decrease with<br />
the number of mismatched phonemes �about 80% rate reduction with three<br />
out of four�, suggesting a strong acoustic-phonetic basis for verbal transformation<br />
suppression. In contrast, semantic relations had no effect �e.g.,<br />
transformations for ‘‘light’’ were suppressed equally by contralateral night<br />
and ‘‘might’’�. Dichotic competition appears to allow us to access and<br />
selectively influence a prelexical stage of linguistic analysis. �Work supported<br />
by NIH.�<br />
4aSC35. Perceptually balanced filter response for binaural dichotic<br />
presentation to reduce the effect of spectral masking. Pandurangarao<br />
N. Kulkarni, Prem C. Pandey �Elec. Eng. Dept, Indian Inst. of Technol.<br />
Bombay, Powai Mumbai 400076, India, pcpandey@ee.iitb.ac.in�, and<br />
Dakshayani S. Jangamashetti �Basaveshwar Eng. College Bagalkot,<br />
Bagalkot Karnataka 587102, India�<br />
Earlier investigations show that the scheme of binaural dichotic presentation<br />
with spectral splitting of speech signal helps in reducing the<br />
effect of spectral masking for persons with moderate bilateral sensorineural<br />
hearing impairment. Speech perception improved by employing filters<br />
with interband crossover gain adjusted between 4 and 6 dB below the pass<br />
band gain. The relationship between scaling factors for a tone presented to<br />
the two ears, so that perceived loudness is that of a monaural presentation,<br />
is investigated for design of comb filters with improved perceptually balanced<br />
response. Results from the listening tests show that, for perceptual<br />
balance, the sum of the two scaling factors should be constant, indicating<br />
that the magnitude response of the comb filters should be complementary<br />
on a linear scale.<br />
3253 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3253<br />
4a FRI. AM
4aSC36. The organization of bilingual perceptual consonant space:<br />
EnglishÕSpanish bilingual perception of Malayalam nasal consonants.<br />
Jenna Silver and James Harnsberger �Inst. for Adv. Study of the Commun.<br />
Proc., Univ. of Florida, Gainesville, FL 32611�<br />
This study examines the capacity of English/Spanish bilinguals to discriminate<br />
between consonants that exist in only one of their respective<br />
phonetic inventories. Two non-native nasal consonant contrasts were<br />
tested: dental versus alveolar and the palatal versus velar, both found in<br />
Malayalam. The dental and palatal nasals appear in Spanish, while the<br />
alveolar and velar nasals occur in English. Poorer performance in discrimination<br />
was interpreted as indicative of a common nasal category<br />
subsuming the Spanish dental and English alveolar nasals; better performance<br />
was taken as evidence of the maintenance of separate categories<br />
from both languages. Two other tests were administered to aid in the<br />
interpretation of the discrimination test scores: forced-choice identification<br />
and perceptual similarity ratings. The findings of this research will be used<br />
to characterize the perceptual consonant space in terms of continuum between<br />
two possible bilingual systems: one that collapses similar categories<br />
across languages or one that maintains two distinct phonological systems<br />
that can be accessed simultaneously. It is believed that bilinguals will be<br />
able to discriminate between these contrasts more consistently than their<br />
monolingual peers; however, there is no prediction about performance<br />
relative to the monolingual group from Malayalam.<br />
4aSC37. Agreement and reliability using reference-matching<br />
paradigm in perceptual voice quality rating in Chinese and English.<br />
Mandy Ho and Edwin Yiu �Voice Res. Lab., Div. of Speech & Hearing<br />
Sci., Univ. of Hong Kong, 5/F Prince Philip Dental Hospital, Hong Kong�<br />
Perceptual voice judgment is commonly used in clinical voice quality<br />
evaluation. The use of a referencematching paradigm in perceptual ratings<br />
has been shown to improve both agreement and reliability �Yiu et al., in<br />
press�. This study set out to investigate the agreement and reliability in<br />
rating Chinese and English dysphonic stimuli using the referencematching<br />
paradigm. Experiment 1 aimed to synthesize Chinese and English<br />
dysphonic stimuli with different breathy and rough severity levels<br />
using the HLSyn Speech Synthesizer. Seven representative anchors �references�<br />
for each of the rough and breathy series in Chinese and English<br />
were chosen by three judges to be used in experiment 2. Acoustic analysis<br />
of the anchor series indicated they were of increasing severity. Experiment<br />
2 recruited ten native Chinese and ten native English subjects to rate the<br />
quality of Chinese and English dysphonic voice samples using the synthesized<br />
anchor as references. Results showed that listeners achieved nearly<br />
90% agreement in rating the Chinese stimuli and 80% agreement in rating<br />
the English stimuli regardless of their language background. The study<br />
showed that the reference-matching paradigm was a reliable method in<br />
rating dysphonic stimuli across listeners with different language backgrounds.<br />
4aSC38. Learning to perceive non-native speech sounds: The role of<br />
test stimulus variability. McNeel Jantzen and Betty Tuller �Ctr. for<br />
Complex Systems and Brain Sci., Florida Atlantic Univ., 777 Glades Rd.,<br />
Boca Raton, FL 33431�<br />
Natural speech stimuli used in studies of phonological learning usually<br />
include several in talkers and phonetic environments because variability<br />
aids learning �e.g., Lively, Logan, and Pisoni, J. Acoust. Soc. Am. �1993��.<br />
The present study investigated whether nonphonetic variability in the synthetic<br />
test set has a similar effect. First, a perceptual mapping procedure<br />
was performed using a synthetic continuum that ranged from the Malayalam<br />
voiced, unaspirated, dental stop consonant to the American English<br />
alveolar �d�, with three F0 contours �low, mid, and high�. Subjects identified<br />
the stimuli �2AFC� and judged their goodness as exemplars of each<br />
category. Subjects then received 15 sessions �one/day� of 2AFC training<br />
with feedback using natural stimuli produced by native Malayalam speakers,<br />
and performed difference ratings on a subset of pairs from the syn-<br />
thetic stimuli. The perceptual mapping procedure was repeated at 1 and 14<br />
days post-training and results compared with a parallel experiment that<br />
included only the midlevel F0 contour in the synthetic test set. �Work<br />
supported by NSF.�<br />
4aSC39. Influence of the prosody of spoken language on recognition<br />
and memory for vocal quality. Sumi Shigeno �Aoyama Gakuin Univ.,<br />
4-4-25 Shibuya, Shibuya-ku, Tokyo, 150-8366 Japan�<br />
This study examined whether recognition and memory for vocal quality<br />
of a speaker who speaks either a native language or non-native language<br />
should be influenced by the prosody of the language that the<br />
speaker utters. Voices of 12 speakers were recorded. They were six Japanese<br />
people and six Americans and Britons. All speakers uttered short<br />
sentences in their respective native languages �i.e., Japanese for Japanese<br />
speakers and English for Americans and Britons� and in a non-native<br />
language �i.e., English for Japanese speakers and Japanese for Americans<br />
and Britons�. Ten Japanese participants rated the vocal quality of speakers<br />
in the first session. After 1 week the same experiment was again conducted<br />
in the second session. Results showed that the performance of identification<br />
of speakers as Japanese or as non-Japanese was comparatively accurate<br />
even though the ratings on the speakers’ voices were varied as the<br />
language spoken by the speakers. Ratings of the voices were compared<br />
further between two sessions and little difference was found, irrespective<br />
of a 1-week blank. Results suggest that the memory for vocal quality is<br />
robust, but that the recognition of vocal quality is dependent on the<br />
prosody of the language spoken by speakers.<br />
4aSC40. Brain activity during auditory processing affected by<br />
expectation of speech versus nonspeech. Yukiko Nota �ATR CIS<br />
BAIC, 2-2-2 Hikaridai, Keihanna Sci. City, Kyoto 619-0288, Japan,<br />
ynota@atr.jp�<br />
fMRI was used to clarify whether there is any differential brain activity<br />
invoked by expectation for speech versus nonspeech sounds. Auditory<br />
stimuli were created by acoustically morphing between either sustained<br />
vowels or tones, respectively, and a buzz sound. The two sets of interpolation<br />
were performed in nine nonlinear steps; the stimuli retained for<br />
perceptual experiments were only the three most vowel-like, the three<br />
most tone-like, and the three most buzz-like tokens morphed from the<br />
vowels. In the ‘‘speech expectation’’ session, subjects were instructed to<br />
discriminate the vowel-like and buzz-like stimuli; in the ‘‘nonspeech expectation’’<br />
session, subjects were instructed to discriminate the tone-like<br />
and buzz-like stimuli without knowing that the buzz stimuli had been<br />
morphed from the vowels. Thus the buzz-like stimuli in both experiments<br />
were the same, but the subjects’ expectations were different because they<br />
were told to expect either speech �vowel-like� or nonspeech �tone-like�<br />
stimuli. Comparison of brain activation during processing of the buzz-like<br />
stimuli under these two conditions revealed that BA40 and thalamus were<br />
more activated in speech expectation, while right BA20 was more activated<br />
in nonspeech expectation. These results suggest that subjects’<br />
speech/nonspeech expectation for sound stimuli influences brain activity<br />
for actual auditory processing. �Work supported by MEXT.�<br />
4aSC41. Representations involved in short-term versus long-term<br />
word learning by preschool children with and without phonological<br />
disorders. Holly Storkel, Jill Hoover, and Junko Maekawa �Dept. of<br />
Speech-Lang.-Hearing, Univ. of Kansas, 1000 Sunnyside Ave., 3001 Dole<br />
Ctr., Lawrence, KS 66045-7555, hstorkel@ku.edu�<br />
This study explores whether sublexical �i.e., individual sound� and/or<br />
lexical �i.e., whole-word� representations contribute to word learning and<br />
whether these contributions change across short-term versus long-term<br />
learning. Sublexical representations were indexed by phonotactic probability,<br />
the likelihood of occurrence of a sound sequence, whereas lexical<br />
representations were indexed by neighborhood density, the number of<br />
3254 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3254
similar sounding words. Thirty-four preschool children participated in a<br />
short-term word learning task that exposed them to nonwords varying in<br />
phonotactic probability and neighborhood density and tested learning of<br />
these nonwords. Long-term learning was assessed through comprehension<br />
and production of real words varying in phonotactic probability and neighborhood<br />
density. Results showed that phonotactic probability and neighborhood<br />
density equally influenced short-term word learning. In contrast,<br />
long-term word learning was affected primarily by neighborhood density.<br />
Thus, both sublexical and lexical representations appeared to play a role in<br />
short-term learning, but only lexical representations played a primary role<br />
in long-term retention. This pattern held for both children with normal<br />
phonological development and children with phonological delays. However,<br />
the direction of the effect of neighborhood density for short-term<br />
word learning varied by group status, suggesting differences in the use of<br />
lexical representations during short-term learning. �Work supported by<br />
NIH.�<br />
4aSC42. Changes in formant frequencies associated with postural<br />
change in adult male speakers over 50 years old. Michiko Hashi,<br />
Tomoki Nanto, and Natsuki Ohta �Prefectural Univ. of Hiroshima, 1-1<br />
Gakuen-cho, Mihara, Hiroshima, Japan�<br />
It is possible that changes of direction of gravity relative to the vocal<br />
tract associated with changes in posture influence acoustic characteristics<br />
of speech including vowel formant frequencies. Studies examining such<br />
effects had produced mixed results and demonstrated the possibility of<br />
substantive interspeaker variability in the effect of postural changes on<br />
vowel formant frequencies. Recent work by Takakura et al. �‘‘Changes in<br />
formant frequencies associated with postural change,’’ paper presented at<br />
the Fall meeting of Acoustical Society of Japan �<strong>2006</strong>��, using young adult<br />
male speakers, revealed a small number of speakers demonstrating<br />
changes in vowel formant frequencies and suggested effect of age. The<br />
present study attempts to examine changes of vowel formant frequencies<br />
in upright and supine position among older male speakers. Attempts will<br />
be made to eliminate the effect of differences in neck position between the<br />
postures through the use of a power-bead-based neck stabilizer. The results<br />
will be compared with data from young normal speakers in the previous<br />
study and inferences will be made relative to speech production models.<br />
4aSC43. The effect of viewing angle on the visual contribution to<br />
speech intelligibility in noise. Eugene Brandewie, Douglas Brungart,<br />
Nandini Iyer, and Brian Simpson �Air Force Res. Lab., Wright–Patterson<br />
AFB, Ohio 45433-7901�<br />
Visual cues are known to assist speech comprehension in noisy environments,<br />
but relatively little is known about the impact that viewing<br />
angle has on the visual contribution to speech intelligibility. In this experiment,<br />
four digital cameras were used to make simultaneous recordings of<br />
test phrases from the Modified Rhyme Test at four different viewing<br />
angles: 0, 45, 90, and 135 deg. These test phrases were used to measure<br />
the effect of viewing angle on the intelligibility of noise-masked speech<br />
stimuli that were presented with and without visual cues at seven different<br />
signal-to-noise ratios �SNRs�. When the face was viewed from the front,<br />
the visual cues provided an intelligibility improvement that was equivalent<br />
to a 6–10-dB increase in SNR. This visual benefit remained roughly constant<br />
for viewing angles up to 90 deg, but it dropped off rapidly �to less<br />
than 2 dB� when the viewing angle increased to 135 deg. The results<br />
suggest that the visual contribution to speech perception is severely impaired<br />
when the observer does not have access to an unobstructed view of<br />
the talker’s mouth.<br />
4aSC44. Towards estimation of Japanese intelligibility scores using<br />
objective voice quality assessment measures. Rui Kaga, Kazuhiro<br />
Kondo, Kiyoshi Nakagawa, and Masaya Fuzimori �Yamagata Univ.,<br />
Jounan 4-3-16, Yonezawa, 992-8510, Yamagata, Japan�<br />
We investigated the use of objective quality measures to estimate the<br />
intelligibility of Japanese speech. We initially focused on PESQ �perceptual<br />
evaluation of speech quality�, which is the state-of-the-art objective<br />
assessment method, and can estimate the mean opinion scores �MOS� at<br />
an extremely high accuracy. Since we can assume that speech quality is<br />
correlated with intelligibility, it should be possible to estimate the intelligibility<br />
from the estimated opinion scores or some of its derivatives. We<br />
tried to estimate the intelligibility of the Japanese Rhyme Test �DRT�. The<br />
DRT uses minimal word pairs whose initial phone differs by one phonetic<br />
feature. We estimated the MOS of the DRT word samples mixed with<br />
noise and tried to correlate this with the measured intelligibility. However,<br />
the estimated MOS showed no difference between phonetic features.<br />
However, the difference in the estimated MOS between the word pairs<br />
seems to differ by a phonetic feature for SNR above 10 dB, which suggests<br />
that the estimation of intelligibility by a phonetic feature may be<br />
possible. We also plan to selectively use the internal data used to calculate<br />
the MOS estimate for better intelligibility estimation.<br />
3255 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3255<br />
4a FRI. AM
FRIDAY MORNING, 1 DECEMBER <strong>2006</strong> HONOLULU ROOM, 8:00 TO 11:55 A.M.<br />
Session 4aSP<br />
Signal Processing in Acoustics, Underwater Acoustics, and Acoustical Oceanography: Adaptive Signal<br />
Processing<br />
Juan I. Arvelo, Jr., Cochair<br />
Johns Hopkins Univ., Applied Physics Lab., National Security Technology Dept., 11100 Johns Hopkins Rd.,<br />
Laurel, MD 20723-6099<br />
Kensaku Fujii, Cochair<br />
Univ. of Hyogo, School of Engineering and Graduate School of Engineering, 2-67 Shosha, Himeji, Hyogo 671-2201, Japan<br />
Chair’s Introduction—8:00<br />
Invited Papers<br />
8:05<br />
4aSP1. Adaptive beamforming for multipath environments. Henry Cox �Lockheed Martin, IS&S, AC3DI, 4350 North Fairfax<br />
Dr., Ste. 470, Arlington, VA 22203�<br />
Coherent multipaths present a significant challenge to most adaptive beamformers because they violate the common assumption of<br />
a rank-one plane wave or geometrically focused signal. When the multipath arrivals that characterize shallow water propagation are<br />
resolvable by the array’s aperture, the mismatch between the assumed and the true signal spatial structure causes signal suppression.<br />
If the amplitude and phase relationships among the various multipaths were known, they could, in principle, be included in a matched<br />
field beamforming approach. This is usually impractical due to inadequate knowledge of propagation parameters, especially bottom<br />
characteristics, and source/receiver motion. A generalization of the standard MVDR approach, called multirank MVDR, assumes only<br />
that the signal lies in a subspace of multiple rank rather than the usual rank-one assumption. An example of a subspace is a small fan<br />
of beams that cover the potential multipath directions. The signal may be rank one corresponding to fully coherent multipath or higher<br />
rank corresponding to incoherent or partially coherent multipath. The multirank approach is applied to the shallow water multipath<br />
problem and compared with the related technique of employing multiple linear constraints. Realistic simulations of alternative<br />
beamforming approaches for a large horizontal array are presented. �Work supported by ONR.�<br />
8:25<br />
4aSP2. Robust adaptive algorithm based on nonlinear error cost function for acoustic echo cancellation. Suehiro Shimauchi,<br />
Yoichi Haneda, and Akitoshi Kataoka �NTT Cyber Space Labs., NTT Corp., 3-9-11, Midori-cho, Musashino-shi, Tokyo, 180-8585,<br />
Japan, shimauchi.suehiro@lab.ntt.co.jp�<br />
Motivated by recent progress in the blind source separation �BSS� technique, a robust echo cancellation algorithm is investigated,<br />
which would inherently identify the echo path even during double-talk by separating the acoustic echo from the local speech. An<br />
adaptive filter has been introduced into acoustic echo cancellers to identify the acoustic echo path impulse response and generate the<br />
echo replica. However, most adaptive algorithms suffer from instability during double-talk. Although step-size control cooperating<br />
with a double-talk detector �DTD� is a promising approach to stop the adaptation temporarily during double-talk, it cannot handle the<br />
echo path change during double-talk. To overcome this problem, novel robust algorithms are derived by applying nonlinearity to the<br />
cost function of a conventional echo cancellation algorithm such as the normalized least mean squares algorithm �NLMS� or the affine<br />
projection algorithm �APA�. Using the simulation results, there is a discussion about how the robustness of the derived algorithms<br />
depends on the choice of the nonlinear function and the original algorithm.<br />
8:45<br />
4aSP3. Realistic modeling of adaptive beamformer performance in nonstationary noise. Bruce K. Newhall �Appl. Phys. Lab.,<br />
Johns Hopkins Univ., 11100 Johns Hopkins Rd., Laurel, MD 20723�<br />
Most adaptive beamformers �ABFs� operate under the assumption that the noise field is quasistationary. They estimate the present<br />
noise field by averaging, assuming stationarity over the estimation time. The adaptive beamformer responds to slow changes in the<br />
noise field across multiple estimation intervals. Unfortunately, in many low-frequency underwater sound applications, the shipping<br />
noise may change rapidly, due to nearby ship motion. This motion can be significant during the estimation interval and degrade ABF<br />
performance. A realistic model has been developed, including two effects of source motion on horizontal towed arrays. Bearing rate<br />
produces a differential Doppler shift across the array. Range rate produces an amplitude modulation as the multipath interference<br />
pattern shifts along the array. The ABF model begins with a realization of ship locations and motion based on historical shipping<br />
density. Each ship generates realistic random noise composed of tonals in a broadband background. The noise is propagated from each<br />
ship to each hydrophone by a normal mode model. For each time sample the position of each ship is updated and propagation<br />
recalculated. The ability of a variety of ABF algorithms to reduce shipping noise clutter is simulated and compared.<br />
3256 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3256
9:05<br />
4aSP4. Multichannel active noise control system without secondary path models using the simultaneous perturbation<br />
algorithm. Yoshinobu Kajikawa and Yasuo Nomura �Faculty of Enginnering, Kansai Univ., 3-3-35, Yamate-cho, Suita-shi, Osaka<br />
564-8680, Japan, kaji@joho.densi.kansai-u.ac.jp�<br />
This paper presents a novel multichannel active noise control �ANC� system without secondary path models. This ANC system<br />
uses a simultaneous perturbation algorithm as the updating algorithm and has an advantage that secondary path models �estimation of<br />
secondary paths� are not required, unlike the MEFX �multiple error filtered-X�-based ANC. This system can consequently control<br />
noise stably because there are not modeling errors that cause system instability. The computational complexity is also very small.<br />
Experimental results demonstrate that the proposed multi-channel ANC system can operate stably under the environment where the<br />
error microphones always move.<br />
9:25<br />
4aSP5. Two-microphone system using linear prediction and noise<br />
reconstruction. Hirofumi Nakano, Kensaku Fujii �Dept. of Comput.<br />
Eng., Univ. of Hyogo, 2167 Shosya, Himeji 671-2201, Japan,<br />
er06j025@steng.u-hyogo.ac.jp�, Tomohiro Amitani, Satoshi Miyata<br />
�TOA Corp., Takarazuka Hyogo 665-0043, Japan�, and Yoshio Itoh<br />
�Tottori Univ., Japan�<br />
This study proposes a new adaptive microphone system that is characterized<br />
by a linear prediction circuit inserted previous to a noise reconstruction<br />
filter corresponding to an adaptive delay used in conventional<br />
systems. This insertion provides various advantages to the adaptive microphone<br />
system. One is that steering a null for a noise source becomes<br />
possible irrespective of the incidence of speech signal. Another is that the<br />
new microphone system works as an omnidirectional microphone for the<br />
speech signal. Consequently, setting the microphones at arbitrary intervals<br />
is possible. For example, building one microphone in a handset and another<br />
into a telephone baseset becomes possible, which provides higher<br />
noise reduction effect. In practical use, microphone systems must function<br />
in reflective surroundings. In this study, such performance of the proposed<br />
system is verified first using computer simulations and then using an experimental<br />
system put in an ordinary <strong>room</strong>. This study also presents experimental<br />
results verifying that the proposed system can successfully reduce<br />
noise incident from the same direction as a speech signal, as well as<br />
crowd noise recorded in an airport.<br />
9:40<br />
4aSP6. Adaptive beamformer trade-off study of an expendable array<br />
for biologic vocalizations. Juan Arvelo �Appl. Phys. Lab., Johns<br />
Hopkins Univ., 11100 Johns Hopkins Rd., Laurel, MD 20723-6099�<br />
Adaptive beamformers exploit the ambient noise anisotropy to increase<br />
the array gain against this background, enhance target detection,<br />
and increase the resolution of the beam response. A prototype expendable<br />
array was developed and tested for high-frequency passive detection and<br />
localization of marine mammals. This latest array consists of vertical poly-<br />
�vinylidene fluoride� �PVDF� wire elements arranged in a horizontal 6<br />
�6 grid with the corner elements removed for a total of 32 lines. The<br />
length of the wires forms a vertical beam response that allows exploitation<br />
of the ambient noise directionality in elevation while the horizontal aperture<br />
provides full coverage in azimuth. The performance and computational<br />
demand of selected adaptive and conventional beamformers are<br />
compared in a trade-off study to determine their possible use in this expendable<br />
system. This trade-off study accounts for the demand of computational<br />
resources in addition to the predicted system performance as adjuncts<br />
to ocean observatories. �This effort is partly supported by JHU/APL<br />
and the Office of Naval Research �ONR�.�<br />
Contributed Papers<br />
9:55–10:10 Break<br />
10:10<br />
4aSP7. Adaptive matched field processing enhancements to forward<br />
sector beamforming. Jeffrey A. Dunne �Appl. Phys. Lab., Johns<br />
Hopkins Univ., 11100 Johns Hopkins Rd., Laurel, MD 20723�<br />
A study was undertaken to examine the potential benefit of adaptive<br />
matched field processing �AMFP� to the forward sector capability of<br />
single-line, twin-line, and volumetric arrays. Comparisons are made with<br />
conventional MFP �CMFP� and adapative and conventional plane-wave<br />
beamforming �APWB and CPWB� in order to assess the degree of ownship<br />
noise reduction obtainable and any corresponding improvement to the<br />
signal-to-noise ratio �SNR�. A minimum variance distortionless response<br />
beamformer using dominant mode rejection was implemented, applied to<br />
both uniform and distorted array shapes. Significant improvement over<br />
CMFP and CPWB in tracking and SNR was seen for modeled data in both<br />
cases, with the distorted array showing, not surprisingly, better left-right<br />
rejection capability. �Work was undertaken with support from the Defense<br />
Advanced Research Projects Agency �DARPA� Advanced Technology Office<br />
�ATO�.�<br />
10:25<br />
4aSP8. Vector sensor array sensitivity and mismatch: Generalization<br />
of the Gilbert-Morgan formula. Andrew J. Poulsen and Arthur B.<br />
Baggeroer �MIT, Cambridge, MA 02139, poulsen@mit.edu�<br />
The practical implementation of any sensing platform is susceptible to<br />
imperfections in system components. This mismatch or difference between<br />
the assumed and actual sensor configuration can significantly impact system<br />
performance. This paper addresses the sensitivity of an acoustic vector<br />
sensor array to system mismatch by generalizing the approach used by<br />
Gilbert and Morgan for an array with scalar, omnidirectional elements<br />
�E.N. Gilbert and S.P. Morgan, Bell Syst. Tech. J. 34, �1955��. As such, the<br />
sensor orientation is not an issue because it does not affect performance<br />
for an array of omnidirectional sensors. Since vector sensors measure both<br />
the scalar acoustic pressure and acoustic particle velocity or acceleration,<br />
the sensor orientation must also be measured to place the vector measurement<br />
in a global reference frame. Here, theoretical expressions for the<br />
mean and variance of the vector sensor array spatial response are derived<br />
using a Gaussian perturbation model. Such analysis leads to insight into<br />
theoretical limits of both conventional and adaptive processing in the presence<br />
of system imperfections. Comparisons of theoretical results and<br />
simulations are excellent. One noteworthy result is the variance is now a<br />
function of the steering angle. �Work supported by the PLUSNeT program<br />
of the Office of Naval Research.�<br />
3257 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3257<br />
4a FRI. AM
10:40<br />
4aSP9. Adaptive filtering using harmonic structure of voiced speech<br />
for reducing nonstationary known noise. Kenko Ota, Masuzo<br />
Yanagida �Doshisha Univ., 1-3, Tarata-Miyakodani, Kyotanabe, Kyoto,<br />
610-0321, Japan, etf1704@mail4.doshisha.ac.jp�, and Tatsuya Yamazaki<br />
�NICT, 619-0289, Kyoto, Japan, yamazaki@nict.go.jp�<br />
Proposed is an effective method for reducing nonstationary known<br />
noise. The objective of this research is to develop a scheme of preprocessing<br />
for speech recognition that keeps the same speech recognition rate<br />
even in a worse acoustic environment and to realize a TV control system<br />
using speech recognition. The basic idea of the proposed method is to<br />
estimate the frequency spectrum of noise including sounds from the TV<br />
itself and to remove noise components from the frequency spectrum of the<br />
received signal. A transfer function from the TV set to the microphone is<br />
calculated in an iterative manner estimating the noise signal at the microphone.<br />
Traditional ANC techniques do not directly use spectral features<br />
such as harmonic structure or fundamental frequency, for example. As the<br />
proposed method uses harmonic structure of vocalic segments in command<br />
speech, it is expected to have an advantage compared with the<br />
traditional adaptive methods. Results of evaluation using speech recognition<br />
show that the proposed method significantly improves speech recognition<br />
rate compared with conventional spectral subtraction. �Work supported<br />
by Knowledge Cluster Project, MEXT, and by Academic Frontier<br />
Project Doshisha University.�<br />
10:55<br />
4aSP10. Robust matched-field processor in the presence of<br />
geoacoustic inversion uncertainty. Chen-Fen Huang, Peter Gerstoft,<br />
and William S. Hodgkiss �Marine Physical Lab., Scripps Inst. of<br />
Oceanogr., UCSD, La Jolla, CA 92037-0238, chenfen@ucsd.edu�<br />
This presentation examines the performance of a matched-field processor<br />
incorporating geoacoustic inversion uncertainty. Uncertainty of geoacoustic<br />
parameters is described via a joint posterior probability distribution<br />
�PPD� of the estimated environmental parameters, which is found by<br />
formulating and solving the geoacoustic inversion problem in a Bayesian<br />
framework. The geoacoustic inversion uncertainty is mapped into uncertainty<br />
in the acoustic pressure field. The resulting acoustic field uncertainty<br />
is incorporated in the matched-field processor using the minimum variance<br />
beamformer with environmental perturbation constraints �MV-EPC�. The<br />
constraints are estimated using the ensemble of acoustic pressure fields<br />
derived from the PPD of the estimated environmental parameters. Using a<br />
data set from the ASIAEX 2001 East China Sea experiment, tracking<br />
performance of the MV-EPC beamformer is compared with the Bartlett<br />
beamformer using the best-fit model.<br />
11:10<br />
4aSP11. A study on combining acoustic echo cancelers with impulse<br />
response shortening. Stefan Goetze, Karl-Dirk Kammeyer �Dept. of<br />
Commun. Univ. of Bremen, Eng., D-28334 Bremen, Germany�, Markus<br />
Kallinger, and Alfred Mertins �Carl von Ossietzky-Univ., Oldenburg,<br />
D-26111 Oldenburg, Germany�<br />
In hands-free video conferencing systems acoustic echo cancelers<br />
�AECs� have to face the problem of very high-order impulse responses<br />
�IRs�, which have to be compensated. Time-domain algorithms for adaptation<br />
often suffer from slow convergence �as the NLMS algorithm, e.g.�<br />
or high computational complexity �e.g., the RLS�. On the other hand<br />
frequency-domain algorithms introduce undesired delays �S. Haykin, Filter<br />
Theory, 2002�. For high-quality hands-free systems IR shortening concepts<br />
and IR shaping concepts developed for listening <strong>room</strong> compensation<br />
�LRC� �M.Kallinger and A. Mertins, in Proc. Asilomar, 2005� can be applied<br />
to increase speech intelligibility for the near-end speaker. The aim of<br />
this study is the synergetic combination of LRC concepts with acoustic<br />
echo cancellation. For this scenario two different forms of concatenating<br />
the subsystems are possible. Either the AEC filter follows the LRC or vice<br />
versa. In the first case the equalization filter reduces the length of the<br />
effective IR seen by the AEC filter. Thus, shorter AEC filters can be used<br />
which results in faster convergence. However, an estimation algorithm for<br />
the <strong>room</strong> IR is necessary for the LRC subsystem. In the second case the<br />
AEC delivers an estimate of the <strong>room</strong> IR which can be used as an input for<br />
the LRC filter. Experimental results confirm the superiority of the new<br />
combined approach.<br />
11:25<br />
4aSP12. Krylov and predictive sequential least-squares methods for<br />
dimensionality reduction in adaptive signal processing and system<br />
identification. James Preisig and Weichang Li �Woods Hole Oceanogr.<br />
Inst., Woods Hole, MA 02543�<br />
Rapid time variation of the environment, a large number of parameters<br />
which need to be adjusted, and the presence of a reduced subset of the<br />
parameters that are relevant at any point in time create significant challenges<br />
for adaptive signal-processing algorithms in underwater acoustic<br />
applications. In applications such as underwater acoustic communications,<br />
the environment is represented by the ‘‘taps’’ of the time-varying impulse<br />
response. The instability of estimation algorithms or inability to track<br />
rapid channel fluctuations are among the problems that are encountered.<br />
An approach to addressing these challenges is to dynamically select a<br />
subspace in which the adjustment of taps takes place. Here, two algorithms<br />
for doing this are presented. The first is based upon using subspace<br />
basis vectors, which form a Krylov subspace with respect to the channel<br />
input correlation matrix and the channel input/output correlation vector.<br />
This method does not use a prediction residual error to select the basis<br />
vectors. A second algorithm is a new variant of the matching pursuit<br />
algorithm. In this case, ‘‘important’’ taps of the channel impulse response<br />
are selected to minimize a forward prediction residual error. The properties<br />
and performance of these two algorithms are presented and compared<br />
using simulation and field data.<br />
11:40<br />
4aSP13. Expectation maximization joint channel impulse response<br />
and dynamic parameter estimation and its impact on adaptive<br />
equalization. Weichang Li and James C. Preisig �Dept. of Appl. Ocean<br />
Phys. and Eng., Woods Hole Oceanograph. Inst., Woods Hole, MA 02543�<br />
Joint estimation of channel impulse response and its dynamic parameters<br />
using the expectation maximization �EM� algorithm and its MAP<br />
variant is derived for broadband shallow-water acoustic communication<br />
channels. Based on state-space channel modeling, the EM algorithms estimate<br />
the channel dynamic parameters from the sequence of channel impulse<br />
response estimates. The estimated parameters are then used in the<br />
Kalman smoother, which estimates the channel impulse response. The<br />
stability of the algorithm is shown to be related to an extended persistent<br />
excitation �EPE� condition, which requires that both the symbol sequence<br />
and the channel estimates be persistently exciting. Modified algorithms are<br />
proposed for broadband multipath channels to avoid the issue of insufficient<br />
excitation. Efficient suboptimal algorithms are also derived from the<br />
EM algorithms that alternatively estimate the parameter and the channel<br />
impulse response while allowing slow parameter variations. The performance<br />
of these channel estimation algorithms as well as their impact on<br />
the subsequent equalizer are demonstrated through experimental data<br />
analysis. �Work supported by ONR Ocean Acoustics.�<br />
3258 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3258
FRIDAY MORNING, 1 DECEMBER <strong>2006</strong> KAUAI ROOM, 7:55 TO 10:00 A.M.<br />
Session 4aUWa<br />
Underwater Acoustics: Sonar Performance<br />
Lisa M. Zurk, Cochair<br />
Portland State Univ., Electrical and Computer Engineering Dept., 1900 S. W. Fourth Ave., Portland, OR 97207<br />
Hiroshi Ochi, Cochair<br />
Japan Agency for Marine-Earth Science and Technology (JAMSTEC), 2-1-5 Natsushima-cho, Yokosuka,<br />
Kanagawa 237-0061, Japan<br />
8:00<br />
4aUWa1. Automatic detection performance comparisons of three<br />
different fluctuation-based signal processors. Ronald A. Wagstaff<br />
�Natl. Ctr. for Physical Acoust., 1 Coliseum Dr., Univ., MS 38677�<br />
The three most common fluctuation-based signal processor �FBP� algorithms<br />
achieve gain by exploiting either the reciprocal of the spectral<br />
amplitude, the log-differential amplitude, or aligned-phase angles. Two<br />
important features of these processors, for the underwater acoustics community,<br />
is their ability to detect and automatically identify signals which<br />
originated from submerged sources, and to provide unusually large signalto-noise<br />
ratio gains. Similar benefits are of interest to the atmosphere<br />
acoustic community. An example is the automatic detection and<br />
identification/classification of hostile airborne and ground vehicles by unattended<br />
ground sensors �UGS�. The three different generic types of FBP<br />
algorithms will be defined. The manner in which each exploits fluctuations<br />
to achieve gain will be explained. Corresponding performances will be<br />
compared using both underwater and atmosphere acoustic data. The ocean<br />
acoustic results came from towed array beamformed spectral data and will<br />
include spectral plots and grams. Corresponding single sensor spectral<br />
results will also be presented for atmosphere acoustic vehicle data. �Work<br />
supported by ARDEC and SMDC.�<br />
8:15<br />
4aUWa2. Evolution of modern fluctuation-based processing. Ronald<br />
Wagstaff �Natl. Ctr. for Physical Acoust., 1 Coliseum Dr., Univ., MS<br />
38677�<br />
Modern fluctuation-based processors �FBPs� are relatively new on the<br />
signal processing scene. They started in the mid-1980s with the realization<br />
that averaging the reciprocal acoustic powers and inverting the result back,<br />
i.e., the harmonic mean, could yield 6- to 8-dB signal-to-noise ratio gain<br />
over the corresponding average power. Because of its significant noise<br />
attenuation, this processor was designated WISPR, i.e., reduces noise to a<br />
whisper. WISPR had a unique, potentially more valuable, capability.<br />
Based on the decibel difference, or ratio, between the average power and<br />
WISPR, it could be determined whether the received signals were from<br />
ships, or from sources of sound that were submerged. After much time and<br />
experience with WISPR at sea, acquiring and processing towed array<br />
ocean acoustic data, and continuing data processing in the laboratory, the<br />
phenomena that were responsible for WISPRs performance, acoustic fluctuations<br />
generated near the sea surface, became better understood and<br />
WISPRs credibility increased. This led to the development of many other<br />
FBPs with similar capabilities, but with significantly enhanced performances.<br />
A brief account of post-WISPR development will be presented,<br />
including a description of the exploitable parameters, how they are used,<br />
and the range of gains that they achieve.<br />
Chair’s Introduction—7:55<br />
Contributed Papers<br />
8:30<br />
4aUWa3. A comprehensive unbiased third party evaluation of a signal<br />
processor for detecting submerged sources among clutter signals and<br />
noise. Ronald Wagstaff �Natl. Ctr. for Physical Acoust., 1 Coliseum Dr.,<br />
Univ., MS 38677�<br />
The Wagstaff’s integration silencing processor, WISPR’, was developed<br />
to detect and identify signals in the ocean from sources that are<br />
submerged well below the sea surface. WISPR is the type of signal processor<br />
that exploits the reciprocal of the spectral power amplitude, rather<br />
than the amplitude as the average power processor does. Processing the<br />
reciprocal of the power represented a significant departure in the prevailing<br />
signal processing philosophy that governed most conventional signal<br />
processing algorithms that were in use when WISPR first appeared on the<br />
scene several years ago. Furthermore, WISPR’s claimed submergedsource<br />
detection capability made it an attractive candidate for some high<br />
interest signal processing applications. Accordingly, one influential national<br />
organization considered its potential use in their mission and decided<br />
to commission a credible third party laboratory to conduct an unbiased<br />
evaluation of the WISPR processor. The emphasis was to be on its<br />
performance for automatic unalerted detection of signals from submerged<br />
sources. The techniques and evaluation methods used to test the WISPR<br />
processor will be described. The results of the evaluation will be presented,<br />
and the influence of those results on the development of other,<br />
more advanced, fluctuation-based processors will be discussed.<br />
8:45<br />
4aUWa4. The estimated ocean detector: Predicted performance for<br />
continuous time signals in a randomÕuncertain ocean. Jeffrey A.<br />
Ballard, R. Lee Culver, Leon H. Sibul �Appl. Res. Lab. and Grad.<br />
Program in Acoust., Penn State Univ. P.O. Box 30, State College, PA<br />
16804�, Colin W. Jemmott, and H. John Camin �Penn State Univ., State<br />
College, PA 16804�<br />
This paper addresses implementation of the maximum likelihood �ML�<br />
detector for passive SONAR detection of continuous time stochastic signals<br />
that have propagated through a random or uncertain ocean. We have<br />
shown previously that Monte Carlo simulation and the maximum entropy<br />
method can make use of knowledge of environmental variability to construct<br />
signal and noise parameter probability density functions �pdf’s� belonging<br />
to the exponential class. For these cases, the ML detector has an<br />
estimator-correlator and noise-canceller implementation. The estimatorcorrelator<br />
detector computes the conditional mean estimate of the signal<br />
conditioned on the received data and correlates it with a function of the<br />
received data, hence the name estimated ocean detector �EOD�. Here we<br />
derive the detector structure for continuous time stochastic signals and<br />
Gaussian noise and present receiver operating characteristic �ROC� curves<br />
for the detector as a function of the signal-to-noise ratio. �Work supported<br />
by ONR Undersea Signal Processing Code 321US.�<br />
3259 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3259<br />
4a FRI. AM
9:00<br />
4aUWa5. Echo detection enhancement using multiple guide sources in<br />
shallow water. David C. Calvo, Charles F. Gaumond, David M. Fromm,<br />
and Richard Menis �Naval Res. Lab., Acoust. Div. Code 7145, 4555<br />
Overlook Ave. SW, Washington, DC 20375�<br />
The use of a guide source has been proposed as a way of compensating<br />
for multipath by forming a spatial-temporal cross correlation of the received<br />
target and guide source signals across a vertical array in the frequency<br />
domain �Siderius et al., J. Acoust. Soc. Am. 102, 3439–3449�.<br />
This processing has the effect of creating a virtual receiver at the guide<br />
source position. In general, the performance of a virtual receiver degrades<br />
if the spatial integration is not carried out over the span of the array with<br />
significant signal. In our study, we have pursued an alternative approach of<br />
using guide sources which does not require this integration in general. The<br />
guide source signal is simply used as a matched filter. Although this does<br />
not correspond to a virtual receiver, it is useful as a means of improving<br />
active or passive detection of signals in noise. In general, the signal gain<br />
using this alternative technique is dependent on the guide source position.<br />
To compensate for this, we construct a separable-kernel-receiver filter<br />
bank using multiple randomly positioned guide source signals. Improvement<br />
of ROC curves in both passive and active scenarios is obtained using<br />
experimental and simulated data. �Work sponsored by ONR.�<br />
9:15<br />
4aUWa6. Incorporating environmental variability into received signal<br />
statistics. H. John Camin, R. Lee Culver, Leon H. Sibul �Appl. Res.<br />
Lab. and Grad. Program in Acoust., Penn State Univ., P.O. Box 30, State<br />
College, PA 16804�, Jeffrey A. Ballard, and Colin W. Jemmott �Penn<br />
State Univ., State College, PA 16804�<br />
We have developed a Monte Carlo-based method for estimating the<br />
variability of acoustic signal parameters caused by uncertain ocean environments.<br />
The method begins with a physics-based model for the environmental<br />
properties and uses the maximum entropy �MaxEnt� method to<br />
construct probability density functions �pdf’s� describing the measured<br />
deviations from the model mean. Random realizations of environmental<br />
variability, with proper depth correlation, are constructed from the pdf’s<br />
and added to the mean model parameters. A parabolic equation code �RAM�<br />
is used to propagate acoustic energy through each realization of the environment.<br />
Fourier synthesis is used to recreate the arrival structure. The<br />
method is demonstrated using measurements from the Strait of Gibraltar,<br />
which is a particularly complicated region dominated by strong tidal fluctuations<br />
and internal waves. During 1996, an international group carried<br />
out the Strait of Gibraltar Acoustic Monitoring Experiment �SGAME�, in<br />
which detailed environmental and 250-Hz acoustic data were collected.<br />
Here, pdf’s of the received signal level are compared with results of the<br />
Monte Carlo method to demonstrate performance. �Gibraltar data and SVP<br />
model provided by Chris Tiemann �ARL:UT� and Peter Worcester �SIO�.<br />
Work supported by ONR Undersea Signal Processing.�<br />
9:30<br />
4aUWa7. Motion compensation of multiple sources. Joung-Soo Park<br />
�Agency for Defense Development, P.O. Box18, Jinhae, Kyung-Nam,<br />
645-600, Korea�, Jae-Soo Kim �Korea Maritime Univ., Young-Do, Busan,<br />
Korea�, and Young-Gyu Kim �Agency for Defense Development, Jinhae,<br />
Kyung-Nam, 645-600, Korea�<br />
Matched field processing has a advantage of detection of multiple<br />
targets. But, if a strong interferer is moving fast near a quiet target, detection<br />
of the target is difficult due to the motion effect of the interferer. The<br />
motion of the interferer introduces energy spreading and results in poorer<br />
detection. A waveguide-invariant-based motion compensation algorithm<br />
was proposed to mitigate the motion effect of a dominant signal component,<br />
which is estimated by eigenvalue method. The eigenvalue method is<br />
good for a strong interferer, but not good for multiple targets. In this<br />
presentation, we will propose a steered beam processing method to mitigate<br />
the motion effect of multiple targets. We will verify the proposed<br />
method with numerical simulations and SwellEx96 data processing.<br />
9:45<br />
4aUWa8. Predicting sonar performance using observations of<br />
mesoscale eddies. Harry DeFerrari �Div. of Appl. Marine Phys.,<br />
RSMAS, Univ. of Miami, 4600 Rickenbacker Cswy, Miami, FL 33149�<br />
A predictive relationship has been observed between the location of<br />
offshore mesoscale eddies and the performance of active and passive sonar<br />
on the shallow water shelf area inside of the eddy. The passage of an eddy<br />
produces a prograde front that modifies acoustic propagation by two<br />
mechanisms. First, the density gradient serves as a conduit for offshore<br />
internal waves to propagate onto the shelf. A long-lived front can result in<br />
order of magnitude increases in potential energy of the internal wave field<br />
and corresponding increases in sound speed variability. Second, the circulation<br />
of the eddy produces a unique sound speed profile that is strongly<br />
downward refracting but has a nearly iso-velocity layer near the bottom<br />
owing to turbulent mixing. The shape of the profile closely approximates a<br />
hyperbolic cosine. Such a profile has mode group velocities that are equal<br />
for all refracted modes, thus producing strong focusing and a caustic at the<br />
depth of the source at all ranges. The experimental observations are confirmed<br />
with oceanographic and acoustic propagation models and, in turn,<br />
the models predict FOM fluctuations of as much as 15 dB for passive<br />
sonar and 24 dB for active sonar, depending on location of the eddy.<br />
3260 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3260
FRIDAY MORNING, 1 DECEMBER <strong>2006</strong> KAUAI ROOM, 10:15 A.M. TO 12:00 NOON<br />
Session 4aUWb<br />
Underwater Acoustics: Session in Honor of Leonid Brekhovskikh I<br />
William A. Kuperman, Cochair<br />
Scripps Inst. of Oceanography, Univ. of California, San Diego, Marine Physical Lab., MC0238, San Diego,<br />
La Jolla, CA 92093-0238<br />
Oleg A. Godin, Cochair<br />
NOAA, Earth System Research Lab., 325 Broadway, Boulder, CO 80305-3328<br />
Chair’s Introduction—10:15<br />
Invited Papers<br />
10:20<br />
4aUWb1. Phenomenon of Leonid Maximovich Brekhovskikh as a man and a scientist. Nikolay Dubrovskiy �Andreyev Acoust.<br />
Inst., Shvernika Ul 4, Moscow 117036, Russia�<br />
Leonid Maximovich Brekhovskikh made phenomenal contributions in acoustics: discovery of the underwater sound channel,<br />
development of the fundamental theory of wave propagation in layered media, and working out a tangent plane approximation in the<br />
wave scattering theory. Brekhovskikh contributed greatly to the organization of research teams and the dissemination of information<br />
on acoustics and oceanography through his popular books and lecturing. He also made a major breakthrough as a public figure and a<br />
statesman. He became the first Director of the Acoustics Institute at the age of 36. He served as a secretary of the Russian Academy<br />
of Sciences Branch involved in oceanography, geography, and atmospheric physics research. Brekhovskikh’s achievements in science<br />
and science leadership were marked by multiple highest USSR awards and many international awards. He became an Honorary<br />
Fellow of the Acoustical Society of America and the Russian Acoustical Society. He received the Lord Raleigh Medal for the<br />
discovery that preserved its urgency for 30 years. Brekhovskikh’s phenomenon is regarded here from the viewpoint of his personality<br />
as well as specific circumstances of his family and social life.<br />
10:40<br />
4aUWb2. Some aspects of Leonid Brekhovskikh’s influence on oceanographic acoustics. W. A. Kuperman and W. Munk<br />
�Scripps Inst. of Oceanogr., Univ. of California, San Diego, La Jolla, CA 92093-0238�<br />
Waveguide physics describes the basic features of long-range sound propagation in the ocean. Over the last half century the theory<br />
has progressed from describing ideal waveguides to more complicated layered structures to range-dependent structures to timevarying,<br />
range-dependent structures. The theme of Brekhovskikh’s pioneering work was the development of robust formulations that<br />
permitted understanding basic ocean acoustics while also laying the foundation to progress to the next levels of realistic complexity.<br />
Early on, he realized that acoustic data were not consistent with known oceanography. His seminal oceanographic experiments<br />
established the pervasive presence of mesoscale phenomena, which to this day are still not fully incorporated into rigorous formulations<br />
of the forward and inverse acoustics problems. We discuss only a very small part of his work and its subsequent influence.<br />
11:00<br />
4aUWb3. Underwater sound propagation: 49 years with L. M. Brekhovskikh’s Waves in Layered Media. Oleg A. Godin<br />
�CIRES, Univ. of Colorado and NOAA, Earth System Res. Lab., 325 Broadway, Boulder, CO 80305, oleg.godin@noaa.gov�<br />
In his first 10 years of research on wave propagation in layered media, L. M. Brekhovskikh created a theory that remains a basis<br />
for physical understanding and mathematical modeling of underwater sound propagation. Summarized in his celebrated book Waves<br />
in Layered Media, first published in 1957, the theory includes spectral �quasi-plane wave� representations of wave fields, normal mode<br />
theory for open waveguides, extensions of the ray and WKBJ methods, and a clear treatment of diffraction phenomena attendant to<br />
caustics, lateral waves, and reflection of wave beams and pulses. The book also charted the ways forward that have been and are<br />
followed by numerous researchers around the globe. Some of the resulting progress was documented in subsequent editions of Waves<br />
in Layered Media and in later books L. M. Brekhovskikh coauthored with his students. This paper will discuss diverse, groundbreaking<br />
contributions L. M. Brekhovskikh made to the wave propagation theory from the prospective offered by recent developments in<br />
underwater acoustics.<br />
11:20<br />
4aUWb4. L. M. Brekhovskikh’s studies on nonlinear wave interaction and atmospheric sound. Konstantin Naugolnykh �Univ.<br />
of Colorado, NOAA, ESRL/Zeltaech LLD, Boulder, CO�<br />
Nonlinear interaction of waves in a compressible fluid is an underlying factor in many geophysical effects, and L. M.<br />
Brekhovskikh made essential contributions to investigation of these phenomena. In particular, he suggested the mechanism of the<br />
infrasound generation by stormy areas in the ocean based on the nonlinear interaction of the counter-propagating sea-surface gravity<br />
3261 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3261<br />
4a FRI. AM
waves. The estimates of the order of magnitude of sound intensities were made indicating that the main part of the infrasound<br />
generated by the surface waves is absorbed in the upper layers of the atmosphere, resulting in the heating of these layers. The other<br />
part of the sound energy can be trapped by the atmospheric acoustic waveguide and then returned to earth at distances of hundreds of<br />
kilometers, producing the voice of the sea.<br />
11:40<br />
4aUWb5. Tangent-plane approximation by L. M. Brekhovskikh and connected methods in the theory of wave scattering from<br />
rough surfaces. Alexander G. Voronovich �NOAA, Earth System Res. Lab., Physical Sci. Div., 325 Broadway, Boulder, CO 80305,<br />
alexander.voronovich@noaa.gov�<br />
Starting from pioneering work by Rayleigh in 1907, scattering of waves from rough surfaces was restricted by the case of small<br />
Rayleigh parameter. In this case perturbation analysis describing the process of Bragg scattering applies. Apparently, smallness of the<br />
roughness is too restrictive for many applications. In 1952 L. M. Brekhovskikh suggested a tangent-plane approximation �TPA�. For<br />
ideal boundary conditions it represents the first iteration of the appropriate boundary integral equation. However, for more complex<br />
situations �e.g., dielectric or solid-fluid interfaces� appropriate boundary integral equations are rather complicated and, even worse,<br />
they cannot be readily iterated. The TPA allows bypassing this step providing the answer in closed form for arbitrary boundary<br />
conditions and for scalar or vector waves in terms of the local reflection coefficient. Unfortunately, the TPA does not correctly describe<br />
the Bragg scattering. However, later it was realized that the TPA allows simple generalization, which treats both low- and highfrequency<br />
limits within single theoretical scheme. This is achieved by considering the local reflection coefficient as an operator rather<br />
than a factor. New methods going beyond the two classical ones with much wider regions of validity were developed based on this<br />
idea. Some of them will be reviewed in this talk.<br />
FRIDAY AFTERNOON, 1 DECEMBER <strong>2006</strong> LANAI ROOM, 1:00 TO 4:45 P.M.<br />
Session 4pAA<br />
Architectural Acoustics: Measurement of Room Acoustics II<br />
Boaz Rafaely, Cochair<br />
Ben Gurion Univ., Electrical and Computer Engineering Dept., 84105 Beer Sheva, Israel<br />
Hideo Miyazaki, Cochair<br />
Yamaha Corp., Ctr. for Advanced Sound Technologies, 203 Matsunokijima, Iwata, Shizuoka 438-0192, Japan<br />
1:00<br />
4pAA1. Impulse response measurements based on music and speech<br />
signals. Wolfgang Ahnert, Stefan Feistel, Alexandru Miron, and Enno<br />
Finder �Ahnert Feistel Media Group, Berlin, Germany�<br />
All known software based measurement systems, including TEF,<br />
MLSSA, SMAART, and EASERA, derive results using predetermined excitation<br />
signals like Sweep, MLS, or Noise. This work extends the range<br />
of excitations to natural signals like speech and music. In this context<br />
selected parameters like frequency range, dynamic range, and fluctuation<br />
of the signal and the signal duration are investigated in order to reach<br />
conclusions about the conditions required to obtain results comparable<br />
with standard excitation signals. Also the limitations of the standard<br />
stimuli and the proposed natural stimuli are discussed.<br />
1:15<br />
4pAA2. Assessment of reverberation time in halls through analysis of<br />
running music. David Conant �McKay Conant Brook Inc., 5655<br />
Lindero Canyon Rd., Ste. 325, Westlake Village, CA 91362,<br />
dconant@mcbinc.com�<br />
The source signal to excite a <strong>room</strong>’s reverberant field sufficient for<br />
detailed measurement of reverberation time �RT60� and other measures<br />
has been the subject of considerable investigation over several decades. It<br />
is generally acknowledged that the best sources are �depending on the<br />
researcher� swept tones, MLS, MLS variations, stopped noise, cannon<br />
shots, etc. All can be characterized as highly audience unfriendly. In the<br />
Contributed Papers<br />
interest of obtaining useful approximations of measured midfrequency<br />
RT60 in the presence of live audiences, this paper discusses several approaches<br />
that may be fruitful while being entirely unobtrusive to the concert<br />
experience.<br />
1:30<br />
4pAA3. Comparison of measurement techniques for speech<br />
intelligibility. Bruce C. Olson �Olson Sound Design, 8717 Humboldt<br />
Ave N, Brooklyn Park, MN 55444, bco@olsonsound.com�<br />
A comparison of measurement techniques for speech intelligibility between<br />
two recently released measurement systems is made. EASERA<br />
�Electronic and Acoustic System Evaluation and Response Analysis� uses<br />
a standard PC and an EASERA Gateway interface attached via Firewire.<br />
The software postprocesses a variety of stimuli in order to derive the<br />
impulse response for the <strong>room</strong> under test. This impulse response is then<br />
further processed and the results are presented to the user in both graphical<br />
and textual presentations. The Ivie Technologies IE-35 is based on a<br />
Pocket PC system and uses an external modulated noise source as stimulus<br />
to produce an intelligibility score as a single number or average of a series<br />
of measurements. This paper will explore a variety of measurements made<br />
in the same locations in a <strong>room</strong> by both systems. Results will also be<br />
shown for a variety of other acoustic measures that quantify the acoustical<br />
parameters of the <strong>room</strong>.<br />
3262 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3262
1:45<br />
4pAA4. Under-balcony acoustics in concert halls: Single source versus<br />
an array of multiple sources. Youngmin Kwon and Gary W. Siebein<br />
�Architecture Technol. Res. Ctr., Univ. of Florida, 134 ARCH, P.O. Box<br />
115702, Gainesville, FL 32611, ymkwon@ufl.edu�<br />
The conventional measurement protocol using a single omnidirectional<br />
sound source may have limits or uncertainty in objective acoustical analysis<br />
of a performance hall. This study conducted monaural and binaural<br />
impulse response measurements with an array of 16 directional loudspeakers<br />
for quantitative acoustical assessment of specifically under-balcony<br />
area in a concert hall. The measurements were executed in a real performance<br />
hall. The measured time- and frequency-domain responses as well<br />
as the results of <strong>room</strong> acoustical parameters including binaural parameters<br />
were compared to the ones measured with a single omnidirectional source.<br />
The results were also compared to the ones taken at the main orchestra<br />
seating area. The time-domain responses showed a clear distinction particularly<br />
in early responses between single source and multiple sources.<br />
On the other hand, the magnitude of frequency response showed significantly<br />
lower at frequencies above 1 kHz than the one measured at the<br />
main area. The results of a binaural parameter, IACC, were found to be<br />
marginal between single source and multiple sources but critically different<br />
between under-balcony area and main area. Variations were also observed<br />
in the results of other <strong>room</strong> acoustical parameters when compared<br />
either between single source and multiple sources or between underbalcony<br />
area and main area.<br />
2:00<br />
4pAA5. Alternative metrics for the directivity of acoustic sources.<br />
Timothy W. Leishman �Acoust. Res. Group, Dept. of Phys. and Astron.,<br />
Brigham Young Univ., Provo, UT 84602�<br />
While the directivity of an acoustic source at a given frequency is<br />
thoroughly characterized by a directivity function over the angular coordinates,<br />
it may also be characterized to a lesser degree by a single-number<br />
directivity factor. The directivity index �i.e., the logarithmic version of the<br />
directivity factor� is a related figure of merit. Recent efforts to quantify the<br />
directivity of sources for architectural acoustics measurements have led to<br />
several alternatives to these values. One example is the area-weighted<br />
spatial standard deviation of radiated levels over a free-field measurement<br />
sphere. This paper presents and compares this and other directivity metrics<br />
for several types of sources, and discusses their benefits.<br />
2:15<br />
4pAA6. Room volume estimation from diffuse field theory. Martin<br />
Kuster and Maarten van Walstijn �Sonic Arts Res. Ctr., Queen’s Univ.<br />
Belfast, BT7 1NN Belfast, Northern Ireland, m.kuster@qub.ac.uk�<br />
Among the parameters relevant in <strong>room</strong> acoustics, the <strong>room</strong> volume is<br />
one of the most important. The general course in <strong>room</strong> acoustics research<br />
is to use the <strong>room</strong> volume in the prediction of <strong>room</strong> acoustic parameters<br />
such as reverberation time or total relative sound pressure level. Contrary<br />
to this, it has been investigated to what extent the <strong>room</strong> volume can be<br />
retrieved from a measured <strong>room</strong> impulse response. The approach followed<br />
is based on <strong>room</strong> acoustic diffuse field theory and requires correctly measured<br />
<strong>room</strong> impulse responses with the initial time delay corresponding to<br />
the source to receiver distance. A total of ten <strong>room</strong>s of varying size and<br />
acoustic characteristics have been included. The results in three <strong>room</strong>s<br />
were unreliable, which was explained by the particular acoustic characteristics.<br />
In the remaining <strong>room</strong>s the results were numerically useful and<br />
consistent between different positions within the same <strong>room</strong> �relative standard<br />
deviation around 20%�. The influence of source and receiver directivity<br />
is also considered.<br />
2:30<br />
4pAA7. In situ measurements for evaluating the scattering surfaces in<br />
a concert hall. Jin Yong Jeon and Shin-ichi Sato �School of<br />
Architectural Eng., Hanyang Univ., Seoul 133-791, Korea,<br />
jyjeon@hanyang.ac.kr�<br />
Sound diffusion by a wall structure is one of the main concerns with<br />
respect to the sound quality of concert halls. There is a need to develop<br />
measurement and evaluation methods for determining the performance of<br />
scattering wall surfaces not only in a laboratory but also in actual halls. In<br />
this study, the acoustical measurements were conducted in a concert hall<br />
which has diffusers with ceramic cubic tiles on the side walls of the stage<br />
and the audience area. Binaural impulse responses were measured at all of<br />
the seats under two conditions, that is, with and without diffusers. The area<br />
which was affected by the diffusive wall was determined and quantified.<br />
The condition without diffusers was produced by covering them with the<br />
movable reflectors. From the binaural impulse responses, the temporal<br />
diffusion �H. Kuttruff, Room Acoustics, �Elsevier Science, London,<br />
1991��, which is calculated from the autocorrelation of the impulse response,<br />
and other acoustical parameters were analyzed. From the relationship<br />
between the scattering coefficient and the acoustical parameters,<br />
sound scattering index for real halls, which represents the degree of the<br />
diffusion of a hall, was proposed.<br />
2:45<br />
4pAA8. Further investigations on acoustically coupled spaces using<br />
scale-model technique. Zuhre Su, Ning Xiang �Grad. Program in<br />
Architectural Acoust., School of Architecture, Rensselaer Polytechnic<br />
Inst., Troy, NY 12180�, and Jason E. Summers �Naval Res. Lab.,<br />
Washington, DC 20024�<br />
Recently, architectural acousticians have been increasingly interested<br />
in halls that incorporate coupled-volume systems because of their potential<br />
for creating nonexponential sound energy decay. Effects of couplingaperture<br />
configuration and source and receiver locations on energy decay<br />
are essential aspects of acoustically coupled spaces that have not yet been<br />
extensively investigated. In order to further understand these effects on<br />
sound fields in coupled <strong>room</strong>s, a systematic experimental study is carried<br />
out. An acoustic scale model technique is used in collecting <strong>room</strong> impulse<br />
responses of a two-<strong>room</strong> coupled system for varying aperture configurations<br />
and surface-scattering conditions. Baseline behavior is established by<br />
varying aperture area for a fixed aperture shape and analyzing relevant<br />
energy-decay parameters at different locations. Effects of aperture shape<br />
and number are systematically investigated by varying these parameters<br />
while holding coupling area fixed. Similarly, effects of receiver location<br />
are systematically investigated by varying the distance of the receiver<br />
from the coupling aperture for a fixed aperture configuration. Schroeder<br />
decay-function decompositions by Bayesian analysis reveal sensitivities to<br />
receiver location and aperture configuration across different frequency<br />
bands.<br />
3:00–3:15 Break<br />
3:15<br />
4pAA9. Virtual microphone control: A comparison of measured to<br />
created impulse responses of various microphone techniques. Daniel<br />
Valente and Jonas Braasch �Rensselaer Polytechnic Inst., 110 8th St.,<br />
Troy, NY 12180, danvprod@yahoo.com�<br />
A method of rendering sound sources in 3-D space has been developed<br />
using virtual microphone control �ViMiC� �J. Acoust. Soc. Am. 117,<br />
2391�. This method has been used to create a flexible architecture for the<br />
creation and rendering of a virtual auditory environment based on microphone<br />
techniques. One of the advantages of ViMiC is the ability to simulate<br />
coincident, near-coincident, and spaced microphone recording techniques.<br />
This allows the user active spatial control over the recorded<br />
environment and the ability to shape the final rendering based on his or her<br />
specific auditory needs. In order to determine the accuracy of simulating<br />
the virtual microphone techniques, measurements of several acoustic<br />
3263 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3263<br />
4p FRI. PM
spaces in Troy, NY will be compared to the measurements of generated<br />
impulse responses of the same modeled spaces within the ViMiC environment.<br />
The data from the measured impulse responses will be used to adapt<br />
the ViMiC system in order to create a more realistic auditory rendering.<br />
Moreover, the ViMiC system can be improved for use as an educational<br />
tool for teaching recording engineers to hear the subtle differences between<br />
various microphone techniques.<br />
3:30<br />
4pAA10. The estimation of the <strong>room</strong> acoustic characteristic using the<br />
acoustic intensity method. Yong Tang, Hideo Shibayama, and Takumi<br />
Yosida �Dept. of Commun. Eng., Shibaura Inst. of Technol., 3-7-5 Toyosu<br />
Koutou-ku, Tokyo, 135-8548 Japan, m603101@sic.shibaura-it.ac.jp�<br />
When a sound radiates in <strong>room</strong>s, a lot of reflection sounds are generated.<br />
From estimation of the direction where the <strong>room</strong> reflection sound<br />
comes from, we can understand the diffusion situation in the <strong>room</strong> acoustic<br />
field. By using the acoustic intensity method, we can measure the<br />
strength and the direction of the sound. In this paper, we estimate the<br />
direction of the reflection sound in the time-space by the acoustic intensity<br />
method and show the acoustic characteristic of the <strong>room</strong>.<br />
3:45<br />
4pAA11. Binaural simulation in an enclosure using the phased beam<br />
tracing. Cheol-Ho Jeong and Jeong-Guon Ih �NOVIC, Dept. of Mech.<br />
Eng., KAIST, Sci. Town, Daejeon 305-701, Korea, chjeong@kaist.ac.kr�<br />
Binaural simulation in an enclosed space is important in the subjective<br />
evaluation of the enclosure acoustics in the design or refinement stage. A<br />
time domain scheme using the geometrical acoustics technique has been<br />
usually used in the binaural processing. However, one can calculate a<br />
pressure impulse response by using the phased beam-tracing method,<br />
which incorporates the phase information in the beam tracing process.<br />
Such phased method employs reflection coefficient and wave number,<br />
whereas the conventional method uses absorption coefficient and air attenuation<br />
factor. Impulse response can be obtained by the inverse Fourier<br />
transformation of the frequency domain result. This feature facilitates the<br />
binaural simulation because the convolution with the HRTF can be accomplished<br />
by a simple multiplication in frequency domain. Convolutions<br />
were conducted for all reflections one by one, and the convolved transfer<br />
functions were summed into one transfer function. Consequently binaural<br />
<strong>room</strong> impulse responses at receivers’ ear positions can be simulated. The<br />
measured binaural <strong>room</strong> impulse responses in the conference <strong>room</strong> were<br />
compared with the predicted results for octave bands of 125 Hz to 4 kHz.<br />
A good agreement with measurement was found, especially in the early<br />
part of impulse responses. �Work supported by BK21.�<br />
4:00<br />
4pAA12. Visualization methods of direct and early reflection sounds<br />
in small enclosures. Chiaki Koga, Akira Omoto �Omoto Lab., Dept. of<br />
Acoust. Design, Faculty of Design, Kyushu Univ., Shiobaru 4-9-1,<br />
Minami, Fukuoka 815-8540, Japan�, Atsuro Ikeda, Masataka Nakahara<br />
�SONA Corp., Nakno-ku, Tokyo, 164-0013, Japan�, Natsu Tanaka, and<br />
Hiroshi Nakagawa �Nittobo Acoust. Eng. Co., Ltd. Sumida-ku, Tokyo<br />
130-0021, Japan�<br />
Many parameters exist for evaluating large sound fields such as concert<br />
halls. However, it is difficult to apply those parameters for evaluation<br />
of a small <strong>room</strong> such as a recording studio because of their different sound<br />
fields. Widely useful common parameters have not been established.<br />
Moreover, early reflections are important in small <strong>room</strong>s for determining<br />
spatial acoustic impressions. Therefore, various methods that visualize<br />
spatial acoustic information obtained by early reflection in <strong>room</strong>s are proposed.<br />
For this study, sound fields �a music studio and a filmmaking studio�<br />
were measured using three kinds of different techniques: instantaneous<br />
intensity, mean intensity, and a sphere-baffled microphone array.<br />
This report compares the information of sound source directions obtained<br />
using these methods. Results show that every method can estimate the<br />
position of sound sources and important reflections with high accuracy. In<br />
the future, we shall propose a method that visualizes spatial acoustic information<br />
more precisely by combining the methods and establishing<br />
acoustic parameters that are available for evaluating and designing small<br />
<strong>room</strong>s.<br />
4:15<br />
4pAA13. Acoustic evaluation of worship spaces in the city of Curitiba,<br />
Brazil. Cristiane Pulsides, David Q. de Sant’Ana, Samuel Ansay<br />
�LAAICA/UFPR, Bloco 4 sala PG-05 81531-990 Curitiba, PR, Brasil,<br />
pulsides@gmail.com�, Paulo Henrique T. Zannin, and Suzana Damico<br />
�LAAICA/UFPR, 81531-990 Curitiba, PR, Brasil�<br />
This article searches acoustic parameters in religious buildings located<br />
in the city of Curitiba intending to study its behavior in this kind of<br />
facilities. The temples were analyzed according to type of ceremony, architectonic<br />
style, and construction date. The research was made through<br />
the impulsive response integration method for three energetic parameters:<br />
�1� reverberation time �RT�; �2� clarity �C80�; and �3� definition �D50�<br />
according recommendations of the ISO/3382:1997 Standard. Performed in<br />
between were six and eight impulsive responses in each <strong>room</strong> using sweep<br />
signals and omnidirectional microphones. The results were than compared<br />
with referential values already existing �W. Fasold and E. Veres, Schallschutz<br />
� Raumakustik in der Praxis, 136 �1998�� for acoustic characterizations.<br />
It is possible to observe in the measurements the direct connection<br />
between reverberation time and the parameters clarity or definition. Moreover,<br />
it is possible also to observe the influence of the geometric ratios and<br />
architectural elements of the <strong>room</strong>s, getting itself for equivalent volumes<br />
and rays of removal of the source, different levels of definition.<br />
4:30<br />
4pAA14. A consideration of the measurement time interval for<br />
obtaining a reliable equivalent level of noise from expressway.<br />
Mitsunobu Maruyama �Salesian Polytechnic, Oyamagaoka 4-6-8,<br />
Machida, Tokyo 194-0215, Japan� and Toshio Sone �Akita Prepectural<br />
Univ., Honjo, Akita 015-0055, Japan�<br />
The level of road traffic noise LA eq,T greatly depends on the maximum<br />
level during the measurement time interval tau, and the maximum level<br />
often appears at the moment when two consecutive heavy vehicles pass<br />
through the point adjacent to the observation point. A mathematical model<br />
is proposed for simulating the variation in traffic noise, especially from the<br />
point of heavy vehicles with passing. The mean time interval between a<br />
pair of two consecutive heavy vehicles with the minimum allowable distance<br />
is obtained from time-series data and the mean recurrence time h ij<br />
which can be calculated from the transition matrix P�p ij�. The comparative<br />
study is made among the numbers of heavy vehicles from 25 to 300<br />
�vehicles/hour� in traffic flow and the observation distances of 40 to 200 m<br />
from the road. The result shows that the measurement time interval required<br />
for the acquisition of reliable data is three to four times as long as<br />
tau or h ij .<br />
3264 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3264
FRIDAY AFTERNOON, 1 DECEMBER <strong>2006</strong> KOHALA/KONA ROOM, 1:15 TO 4:30 P.M.<br />
Session 4pABa<br />
Animal Bioacoustics: Marine Mammal Acoustics II<br />
David K. Mellinger, Chair<br />
Oregon State Univ., Hatfield Marine Science Ctr., Newport, OR 97365<br />
1:15<br />
4pABa1. Great ears: Functional comparisons of land and marine<br />
leviathan ears. D. R. Ketten �Harvard Med. School, Boston, MA;<br />
Woods Hole Oceanograph. Inst., Woods Hole, MA�, J. Arruda, S. Cramer,<br />
M. Yamato �Woods Hole Oceanograph. Inst., Woods Hole, MA�, J.<br />
O’Malley �Massachusetts Eye and Ear Infirmary, Boston, MA�, D.<br />
Manoussaki �Vanderbilt Univ., Nashville, TN�, E. K. Dimitriadis �NIH/<br />
NIDCD, Bethesda, MD�, J. Shoshani �Univ. of Asmara, Asmara, Eritrea�,<br />
and J. Meng �American Museum of Natural History, New York, NY�<br />
Elephants and baleen whales are massive creatures that respond to<br />
exceptionally low frequency signals. Although we have many elephant and<br />
whale vocalization recordings, little is known about their hearing. Playback<br />
experiments suggest hearing in both proboscideans and mysticetes is<br />
tuned similarly to low or even infrasonic signals. This raises several interesting<br />
issues. First, they emit and perceive signals in two media, air and<br />
water, with radically different physical acoustic properties: 4.5-fold differences<br />
in sound speed, three-fold magnitude difference in acoustic impedance,<br />
and, for common percepts, whales must accommodate 60-fold<br />
acoustic pressures. Also, a commonly held tenet is that upper hearing limit<br />
is inversely correlated with body mass, implying there should be virtually<br />
no whale-elephant hearing overlap given body mass differences. This<br />
study analyzed how inner ears in these groups are structured and specialized<br />
for low-frequency hearing. Computerized tomography and celloidin<br />
histology sections were analyzed in six baleen whale (n�15) and two<br />
elephant species (n�7). The data show mysticetes have a substantially<br />
greater hearing range than elephants but that coiling and apical cochlear<br />
structures are similar, suggesting common mechanical underpinnings for<br />
LF hearing, including cochlear radii consistent with the Whispering Gallery<br />
propagation effect. �Work supported by ONR, NIH, WHOI OLI,<br />
Seaver Foundation.�<br />
1:30<br />
4pABa2. Social context of the behavior and vocalizations of the gray<br />
whale Eschrichtius robustus. Sarah M. Rohrkasse �School for Field<br />
Studies, Ctr. for Coastal Studies, Apartado Postal 15, Puerto San Carlos,<br />
BCS, CP 23740 Mexico, sarro101@hotmail.com� and Margaret M.<br />
Meserve �Guilford College, Greensboro, NC 27410�<br />
Sound production and surface behavior of the gray whale were investigated<br />
at Bahia Magdalena, Mexico to determine if vocalizations have<br />
behavioral correlations or are used in specific social contexts. Fifteenminute<br />
sessions of behavioral observations and acoustic recordings of gray<br />
whales in various social contexts were collected from February to April<br />
<strong>2006</strong> (n�30). Analysis of sound production included proportional use of<br />
different call types and acoustic variables of each sound type. Preliminary<br />
acoustic analysis found no correlation with social contexts or behaviors,<br />
but proportional use of different vocalizations is similar to past studies in<br />
Baja �Dahlheim et al, The Gray Whale, pp. 511–541 �1984�, F. J. Ollervides,<br />
dissertation, Texas A&M University �2001��. Initial results indicate<br />
significant differences in frequencies of high surface behaviors (p<br />
�0.0477) of groups that include mother-calf pairs. As analysis continues,<br />
possible correlations between social context and use of sounds could allow<br />
for acoustics to be an indicator of group composition, seasonal movements,<br />
and social patterns and to help determine the functions of sounds.<br />
�Work supported by SFS and NFWF.�<br />
Contributed Papers<br />
1:45<br />
4pABa3. Ambient noise and gray whale Eschrichtius robustus<br />
behavior. Francisco Ollervides, Kristin Kuester, Hannah Plekon, Sarah<br />
Rohrkasse �School for Field Studies—Ctr. for Coastal Studies, Apartado<br />
Postal 15, Puerto San Carlos, BCS, CP 23740 Mexico,<br />
follervides@hotmail.com�,Kristin Kuester �Univ. of<br />
Wisconsin�Madison, Madison, WI 53706�, HannahPlekon �Davidson<br />
College, Davidson, NC�, andSarahRohrkasse �Texas A and M Univ.,<br />
College Station, TX 77843�<br />
Between 14 February and 13, April <strong>2006</strong>, we conducted 31 recording<br />
sessions of ambient noise and behavioral sampling of gray whales within<br />
Magdalena Bay, Mexico. This breeding lagoon does not have the same<br />
Marine Protected Area status compared to the other breeding lagoons of<br />
San Ignacio and Guerrero Negro in the Mexican Pacific coast. Poorly<br />
monitored guidelines and increasing boat traffic from whale�watching<br />
tourism in this area have the potential to affect the surface behavior of<br />
these animals and increase average ambient noise levels. Relative ambient<br />
noise<br />
levelswererecordedandcomparedtoapreviousstudy�Ollervides,2001�todetermine<br />
similarities or differences in the 5�year interval between both data sets.<br />
Although results are not comparable in decibel levels, probably due to<br />
equipment calibration problems, there was a significant difference between<br />
the different regions of the bay Kruskal�Wallis �p�0.0067�. Activity<br />
levels ranged from 0.005–0.196 behaviors/whale/minute. Ambient noise<br />
levels ranged from 35.70–64.32 dB Re: 1 Pa. No correlation was found<br />
between the ambient noise levels in the bay and the activity level of gray<br />
whales �correlation value�0.0126; log correlation value�0.172�. Further<br />
acoustic processing is currently underway.<br />
2:00<br />
4pABa4. Look who’s talking; social communication in migrating<br />
humpback whales. Rebecca A. Dunlop, Michael J. Noad �School of<br />
Veterinary Sci., Univ. of Queensland, St. Lucia, Qld 4072, Australia.<br />
r.dunlop@uq.edu.au�, Douglas H. Cato �Defence Sci. and Tech Org.,<br />
Pyrmont, NSW 2009, Australia�, and Dale Stokes �Scripps Inst. of<br />
Oceanogr., La Jolla, CA 92037�<br />
A neglected area of humpback acoustics concerns nonsong vocalizations<br />
and surface behaviors known collectively as social sounds. This<br />
study describes a portion of the nonsong vocal repertoire and explores the<br />
social relevance of individual sound types. A total of 622 different sounds<br />
were catalogued and measured from whales migrating along the east coast<br />
of Australia. Aural and spectral categorization found 35 different sound<br />
types, and discriminate functions supported 33 of these. Vocalizations<br />
were analyzed from 60 pods that were tracked visually from land and<br />
acoustically using a static hydrophone array. Nonsong vocalizations occurred<br />
in all pod compositions: lone whales, adult pairs, mother/calf pairs,<br />
mother/calf/escorts, and multiple-adult pods. Thwops and wops were<br />
likely to be sex-differentiated calls with wops from females and thwops<br />
from males. Sounds similar to song-units were almost all from joining<br />
pods and yaps were only heard in splitting pods. Other low-frequency calls<br />
�less than 60 Hz� were thought to be within-pod contact calls. Higherfrequency<br />
cries �fundamental 450–700 Hz� and other calls �above 700 Hz�<br />
and presumed underwater blows were heard more frequently in joining<br />
3265 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3265<br />
4p FRI. PM
pods displaying agonistic behaviors. This work demonstrates that humpbacks<br />
produce a great range of contextually different communication signals.<br />
�Work supported by ONR and DSTO.�<br />
2:15<br />
4pABa5. Seasonal ambient noise levels and impacts on<br />
communication in the North Atlantic right whale. Susan E. Parks,<br />
Christopher W. Clark, Kathryn A. Cortopassi, and Dimitri Ponirakis<br />
�Bioacoustics Res. Program, Cornell Univ., 159 Sapsucker Woods Rd.,<br />
Ithaca, NY 14850, sep6@cornell.edu�<br />
The North Atlantic right whale is a highly endangered species of baleen<br />
whale. Acoustic communication plays an important role in the social<br />
behavior of these whales. Right whales are found in coastal waters along<br />
the east coast of the United States, an area characterized by high levels of<br />
human activity. Most of these activities generate noise that is propagated<br />
into the coastal marine environment. The goals of this project are to characterize<br />
the noise, both natural and anthropogenic, in right whale habitat<br />
areas to determine what levels of noise the whales are regularly exposed<br />
to, and whether the acoustic behavior of right whales changes in response<br />
to increased noise. Continuous recordings were made from autonomous<br />
bottom-mounted recorders in three major habitat areas in 2004 and 2005;<br />
Cape Cod Bay �December–May�, Great South Channel �May�, and the<br />
Bay of Fundy, Canada �August� to passively detect right whales by recording<br />
their vocalizations. Here, we describe the ambient noise levels in these<br />
recordings to describe the daily acoustic environment of right whales, how<br />
noise varied over diel, weekly, and seasonal time scales, and whether noise<br />
levels correlated with any observed changes in acoustic behavior of the<br />
whales.<br />
2:30<br />
4pABa6. Blue whale calling in Australian waters. Robert D.<br />
McCauley, Chandra P. Salgado Kent �Curtin Univ. of Technol., G.P.O.<br />
Box U 1987, Perth 6845, Australia�, Christopher L.K. Burton �Western<br />
Whale Res. Hillarys 6923, WA Australia�, and Curt Jenner �Ctr. for Whale<br />
Res. �WA Inc.�, Fremantle WA, 6959 Australia�<br />
Calling from the Antarctic true blue whale �Balaenoptera musculus<br />
intermedia� and the tropical subspecies �brevicauda, or pygmy blue� have<br />
been recorded across southern Australia with the pygmy blue calls also<br />
recorded along the Western Australian �WA� coast. The subspecies have a<br />
believed common downsweep and markedly different longer, tonal calls.<br />
The frequency of most energy in the tonal calls is offset between the<br />
subspecies suggesting sound-space partitioning. The pygmy blue threepart<br />
tonal call is typically 120 s long repeated every 200 s, has several<br />
variants, and includes a complex two-source component. The nature of the<br />
pygmy blue call allows counts of instantaneous calling individuals, giving<br />
relative abundance. These estimates in the Perth Canyon, a localized seasonal<br />
feeding area, show patterns in usage of space and through time<br />
within and between seasons, such as the sudden departure of animals at a<br />
season end, which varies by approximately 2 weeks between years. Sea<br />
noise records along the WA coast indicate south-traveling animals arrive<br />
midway along the coast in October to November, animals fan out across<br />
southern Australian over December through May, then move north in the<br />
Austral winter. We have begun converting abundance estimates from relative<br />
to absolute for pygmy blue calling rates.<br />
2:45<br />
4pABa7. Acoustical monitoring of finback whale movements on the<br />
New Jersey Shelf. Altan Turgut �Naval Res. Lab., Acoust. Div.,<br />
Washington, DC 20375� and Christopher Lefler �Univ. of California Santa<br />
Barbara, Santa Barbara, CA 93106�<br />
Acoustical monitoring of finback whales is performed by using a data<br />
set collected over a 3-week period in December of 2003 on the New<br />
Jersey Shelf. One-second-duration 20-Hz signals of finback whales were<br />
recorded on three vertical line arrays �VLAs� and a bottomed horizontal<br />
line array �HLA.�. One-second-duration pulses are separated by about 10 s<br />
and there is an approximately 2-min-long silent period between 10- to<br />
18-min-long pulse trains. A 30- to 60-min silent period after 5 to 10 pulse<br />
trains is also common. Modal analysis of individual pulses indicated that<br />
most signals contained two acoustic modes. Arrival-time and group-speed<br />
differences of these modes are used for remote acoustic ranging. These<br />
modal characteristics are also exploited in a broadband matched-field algorithm<br />
for depth discrimination. Bearing estimation of individual whales<br />
is obtained by performing horizontal beamforming on the HLA data.<br />
Range estimation results are verified by time-of-flight triangulation using<br />
single hydrophone data from each VLA location. Acoustic monitoring results<br />
indicated that most finback whales traveled near the shelf break front<br />
where food might be abundant. Relations between silent periods and<br />
acoustic range/depth monitoring results are also investigated. �This work<br />
was supported by the ONR.�<br />
3:00–3:15 Break<br />
3:15<br />
4pABa8. Analysis of melon-headed whale aggregation in Hanalei Bay,<br />
July 2004. David M. Fromm �Naval Res. Lab., 4555 Overlook Ave. SW,<br />
Washington, DC 20375-5350�, Joseph R. Mobley, Jr. �Univ. of Hawaii at<br />
M_noa, Honolulu, HI 96822�, Stephen W. Martin �Space and Naval<br />
Warfare Systems Ctr. San Diego, San Diego, CA 92152-5001�, and Paul<br />
E. Nachtigall �Univ. of Hawaii at M_noa, Kailua, HI 96734�<br />
On 3 July 2004, an aggregation of ca. 150–200 melon-headed whales<br />
�Peponocephala electra� appeared in the shallow waters of Hanalei Bay,<br />
Kauai and congregated there for over 27 h. Preceding the whales’ appearance<br />
and partially coincident with their time in the Bay, midrange �3.5–5<br />
kHz� tactical sonars were intermittently deployed during the Rim of the<br />
Pacific 2004 �RIMPAC� joint military exercises being conducted in waters<br />
near Kauai by the U.S., Japan, and Australia Navies. An NOAA report<br />
�Southall et al., <strong>2006</strong>� attributed the active sonar usage as a plausible, if<br />
not likely, contributing factor. A detailed timeline and reconstruction of the<br />
RIMPAC activities is presented showing the worst-case estimates of the<br />
sonar sound levels in the waters surrounding Kauai. A re-examination of<br />
available evidence combined with a new report of a simultaneous and<br />
similar aggregation in Sasanhaya Bay, Rota, Commonwealth of the Northern<br />
Mariana Islands, brings the plausibility conclusion into question. �This<br />
work was sponsored by multiple sources. D. Fromm and S. Martin conducted<br />
acoustic analyses with funds provided by the U.S. Pacific Fleet. J.<br />
Mobley received funding from the U.S. Geological Survey. P. Nachtigall is<br />
sponsored by the Office of Naval Research for marine mammal audiometric<br />
studies.�<br />
3:30<br />
4pABa9. Midfrequency sound propagation in beaked whale<br />
environments. Eryn M. Wezensky, Thomas R. Stottlemyer, Glenn H.<br />
Mitchell �Naval Undersea Warfare Ctr., Newport Div., Newport, RI<br />
02841�, and Colin D. MacLeod �Univ. of Aberdeen, Aberdeen, U.K.�<br />
Recent mass strandings of beaked whales �Ziphiidae, Cetacea� coinciding<br />
with the use of midfrequency range �1–10 kHz� active sonar have<br />
caused speculation about the potentially adverse effects of these sound<br />
sources. Particular questions of the research and regulatory communities<br />
concern whether beaked whale sensitivity to midfrequency sound exposure<br />
is influenced by oceanographic characteristics present at the time of<br />
the mass stranding events. This study investigated the interaction between<br />
beaked whale habitat characteristics and the nature of a midfrequency<br />
signal by analyzing the oceanographic factors affecting underwater acoustic<br />
propagation. Three types of model sites were selected from five specific<br />
geographical locations where beaked whales have been regularly recorded<br />
or where a mass stranding event has been reported. A ray-trace acoustic<br />
propagation model was used to generate transmission loss for a 3-kHz<br />
signal over a representative 60-km transect at each locality. Model outputs<br />
visually demonstrated how the combination of site/event-specific oceanographic<br />
characteristics affects the sound propagation of a moving source.<br />
A parametric sensitivity comparison and statistical analysis were conducted<br />
to identify influential factors between environmental parameters,<br />
source depth, and the resulting transmission loss. Major findings of this<br />
study as well as future research direction are discussed. �Research supported<br />
by NAVSEA.�<br />
3266 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3266
3:45<br />
4pABa10. Examination and evaluation of the effects of fast rise-time<br />
signals on aquatic animals. Michael Stocker �Seaflow, Inc., 1062 Ft.<br />
Cronkhite, Sausalito, CA 94965�<br />
Increasingly human enterprise is subjecting the ocean environment to<br />
acoustic signals to which marine animals are not biologically adapted.<br />
This is evidenced by a marked rise in marine mammal strandings, as well<br />
as hearing and other physiological damage to fish and other marine organisms<br />
as a result of, or coincident to, human-generated noise events. Determining<br />
phonotoxic thresholds of marine organisms is complicated by the<br />
fact that various marine animals are adapted to sense either pressure gradient<br />
or particle motion acoustic energy, or some combination or gradient<br />
between the two. This has been addressed to some degree by exposure<br />
metrics that consider either net or accumulated acoustical flux densities<br />
from various noise sources. This paper examines the role and effects of<br />
signal rise time both in terms of physiological impulse response of the<br />
exposed organisms, as well as broadband saturation flux densities of fast<br />
rise-time signals on animal sense organs. Case studies from the literature<br />
will be presented to demonstrate the effects of fast rise time signals on<br />
fish. Acoustical signals with high crest factors and fast rise-time components<br />
will be compared to signals with dominantly sinusoidal components<br />
to illustrate the perceptual effects of these signals on human hearing.<br />
4:00<br />
4pABa11. Noise separation of underwater acoustic vocalization using<br />
auditory filter bank and Poisson rate estimation. Owen P. Kenny and<br />
Craig R. McPherson �Dept. of Elec. and Comput. Eng., James Cook<br />
Univ., Douglas 4811, Queensland, Australia�<br />
Formant vocalization tracking has been achieved using a mammalian<br />
periphery model and a Poisson rate estimator. This approach used a set of<br />
linear bandpass filters to simulate the mechanical displacement of the basilar<br />
membrane. The auditory model simulated neural firing by producing a<br />
spike at the positive going zero crossing for each filter output. Poisson<br />
intensity of the neural firing rate is controlled by the dominant frequency<br />
components of the signal present in the filter. This approach is extended by<br />
incorporating neural synchronization information to separate the formant<br />
structure from that of noise. The filter structure is designed to overlap the<br />
frequency range of adjacent filters. The presence of a formant structure in<br />
adjacent filters controls the interspike intervals of neural firing for both<br />
filters, which results in the neural firing from both filters being synchronized.<br />
If a noise-only component is present in either filter, then the spiking<br />
outputs from the adjacent filters are unsynchronized. Experimental results<br />
have shown that incorporating neural synchronization information between<br />
adjacent filters has enabled separation of signal components from<br />
noise. This technique enables easier signal and noise separation than allowed<br />
by traditional methods.<br />
4:15<br />
4pABa12. Using vocalizations of Antarctic seals to determine pupping<br />
habitats. T. L. Rogers, C. J. Hogg, M. B. Ciaglia �Australian Marine<br />
Mammal Res. Ctr., Zoological Parks Board of NSW/Faculty of Veterinary<br />
Sci., Univ. of Sydney, Mosman Australia�, and D. H. Cato �Defence Sci.<br />
& Technol. Organisation, Pyrmont, Australia�<br />
The Ross and Leopard seal use the floes of the Antarctic pack ice to<br />
whelp and raise their pups, But both species are rarely seen in summer<br />
throughout the pack ice. We now realize that this is because they are under<br />
the water ‘‘calling’’ during the austral summer as part of their breeding<br />
display, and so their presence is underestimated in traditional visual surveys.<br />
The period of ‘‘calling’’ overlaps with the time that females give<br />
birth, so their vocalizations can be used to determine seal distributions<br />
during this time. Acoustic recordings were made using sonobuoys deployed<br />
during ship based surveys in the pack ice and analyzed to determine<br />
the seal distributions. This was used to predict habitat preference of<br />
seals by relating their distributions to remotely sensed indices: ice cover,<br />
ice floe type, ice thickness, distance to ice edge, distance to shelf break,<br />
distance to land, sea surface temperature, and chlorophyll a.<br />
FRIDAY AFTERNOON, 1 DECEMBER <strong>2006</strong> KOHALA/KONA ROOM, 4:30 TO 5:15 P.M.<br />
Session 4pABb<br />
Animal Bioacoustics: Avian Acoustics<br />
Ann E. Bowles, Chair<br />
Hubbs Sea World Research Inst., 2595 Ingraham St., San Diego, CA 92109<br />
4:30<br />
4pABb1. Effective area of acoustic lure surveys for Mexican spotted<br />
owls „Strix occidentalis lucida…. Samuel L. Denes, Ann E. Bowles<br />
�Hubbs-SeaWorld Res. Inst., 2595 Ingraham St., San Diego, CA 92109,<br />
sdenes@hswri.org�, Kenneth Plotkin, Chris Hobbs �Wyle Labs.,<br />
Arlington, VA 22202�, John Kern �Kern Statistical Services, Sauk Rapids,<br />
MN 56379�, and Elizabeth Pruitt �GeoMarine, Inc., Hampton, VA 23666�<br />
During acoustic lure surveys for birds, topography and ambient noise<br />
are likely to be important determinants of detectability. Examinations of<br />
propagation were conducted for acoustic lures �human-made calls� and<br />
owl responses recorded during acoustic surveys for Mexican spotted owls<br />
in the Gila National Forest �2005�. Lure surveys were designed based on<br />
Contributed Papers<br />
formal agency protocols, which assumed a 0.43-km detection range under<br />
typical conditions. A total of 558 points was called over a heavily forested,<br />
topographically complex 20�24-km area. Real-time measurements of owl<br />
calls and lures were made with a calibrated recording system. Ambient<br />
noise was collected using an array of 39 Larson-Davis 820 and 824 soundlevel<br />
meters. The NMSIM �Wyle Laboratories� single-event propagation<br />
simulator was used to model propagation of both owl and human calls.<br />
The resulting model of survey effort was compared with a simple twodimensional<br />
statistical model. Probability of detecting owls did not fit the<br />
expectations of the agency protocol, suggesting that acoustic propagation<br />
should be considered during owl surveys. �Work supported by U.S. Air<br />
Force ACC/CEVP; USFWS Permit No. TE024429�<br />
3267 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3267<br />
4p FRI. PM
4:45<br />
4pABb2. Automated localization of antbirds and their interactions in<br />
a Mexican rainforest. Alexander N. G. Kirschel, Travis C. Collier,<br />
Kung Yao, and Charles E. Taylor �Univ. of California, Los Angeles, 621<br />
Charles E. Young Dr. South, Los Angeles, CA 90095�<br />
Tropical rainforests contain diverse avian communities incorporating<br />
species that compete vocally to propagate their signals to intended receivers.<br />
In order to effectively communicate with birds of the same species,<br />
birds need to organize their song performance temporally and spatially. An<br />
automated identification and localization system can provide information<br />
on the spatial and temporal arrangement of songs. Acoustic sensor arrays<br />
were tested for the ability to localize the source of songs of antbirds<br />
recorded in a Mexican rainforest. Pilot studies with a five-node array<br />
arranged in a rough circle with a 20-m diameter located the song of Dusky<br />
Antbird �Cercomacra tyrannina� with an error of 73 cm and Mexican<br />
Antthrush �Formicarius moniliger� with an error of 65 cm from the location<br />
of a source loudspeaker within the array. An additional source 21 m<br />
outside was also localized. Results will be presented for experiments and<br />
recordings of individuals at the Mexican rainforest site in October <strong>2006</strong>.<br />
Locations of birds of the same and different species during vocal performance<br />
will provide a greater understanding of how individuals interact<br />
spatially with each other based on their vocal performance, from which the<br />
role of song in ecological interactions can be inferred.<br />
5:00<br />
4pABb3. Nonintrusive acoustic identification of hermit thrush<br />
„Catharus guttatus… individuals. Dennis F. Jones �Defence R&D<br />
Canada—Atlantic, P.O. Box 1012, Dartmouth, NS, Canada B2Y 3Z7,<br />
dennis.jones@drdc-rddc.gc.ca�<br />
From mid-April well into the summer, the secretive hermit thrush �Catharus<br />
guttatus� can be heard singing throughout the woodlands of Nova<br />
Scotia. Its song is distinctive, beginning with a clear introductory note<br />
followed by a flurry of flutelike body notes, often cascading and reverberant<br />
in character. Despite this fine display of avian virtuosity, few studies<br />
have been reported that probe the differences between the calls, songs, and<br />
repertoires of individuals. From April 2003 to May <strong>2006</strong>, over 3000 songs<br />
from several birds were recorded using digital video cameras at study sites<br />
in and around the city of Halifax, Nova Scotia. The only birds recorded<br />
were those in close proximity to roads and trails. None of the birds were<br />
marked, banded, or deliberately disturbed in any way. Although the study<br />
birds remained hidden from view most of the time, in the few instances<br />
where the birds perched in the open, their behaviors while singing were<br />
captured on videotape. All of the birds were readily distinguishable from<br />
each other as no two individuals had a single song in common. The most<br />
significant finding was that individuals could be reidentified acoustically<br />
after 1 week, 3 months, and 1 year had elapsed.<br />
FRIDAY AFTERNOON, 1 DECEMBER <strong>2006</strong> KAHUKU ROOM, 1:00 TO 4:50 P.M.<br />
Session 4pBB<br />
Biomedical UltrasoundÕBioresponse to Vibration and Signal Processing in Acoustics: Elastic Imaging<br />
Peter J. Kaczkowski, Cochair<br />
Univ. of Washington, Applied Physics Lab., 1013 NE 40th Street, Seattle, WA 98105-6698<br />
Tsuyoshi Shiina, Cochair<br />
Univ. of Tsukuba, Graduate School of Systems and Information Engineering, 1-1-1 Tennodai, Tsukuba 305-8573, Japan<br />
Invited Papers<br />
1:00<br />
4pBB1. Present and future of elasticity imaging technology. Tsuyoshi Shiina �Grad. School of Systems and Information Eng.,<br />
Univ. of Tsukuba, 1-1-1 Tennodai Tsukuba, Japan� and Ei Ueno �Univ. of Tsukuba, Tsukuba, Japan�<br />
Elastic properties of tissues are expected to provide us novel diagnostic information since they are based on tissue characteristics<br />
and sensitively reflect its pathological state. So far, various techniques for tissue elasticity imaging have been proposed. However, it<br />
was not so easy to satisfy real-time operation and freehand manipulation of probe, which was required for practical equipment. To<br />
satisfy these conditions, we developed the combined autocorrelation method �CAM� and recently manufactured a commercial ultrasound<br />
scanner, for real-time tissue elasticity imaging by implementing the CAM algorithm, By slightly compressing or relaxing the<br />
body through freehand operation, the strain images are obtained with real-time and superimposed on B-mode images with a translucent<br />
color scale. In addition, we proposed elasticity scores of malignancy by categorizing patterns of elasticity images of breast tumors<br />
into five classes from malignant to benign. As a result of diagnosis based on the elasticity score, it was revealed that even nonexperts<br />
could attain precise diagnosis of breast cancer based on elasticity score as well as experts since the criterion on elasticity score is much<br />
simpler than conventional B-mode images. Finally, some prospects for the next stages of elasticity imaging technology well be<br />
surveyed.<br />
1:20<br />
4pBB2. Real-time tissue elasticity system—Development and clinical application. Takeshi Matsumura, Tsuyoshi Mitake<br />
�Hitachi Medical Corp. 2-1, Toyofuta, Hashiwa-Shi, Chiba-Ken, Japan�, Tsuyishi Tsuyishi, Makoto Yamakawa, Ei Ueno �Tsukuba<br />
Univ.�, Nobuhiro Fukunari �Shouwa Univ.�, and Kumi Tanaka �Nippon Medical Univ.�<br />
The progress of recent semiconductor technology has a remarkable thing. Thanks to progress of this semiconductor technology, the<br />
ultrasound scanner in medicine could come to hold enormousness computing power and has come to realize various complicated<br />
processing. At the same time, hardness of human tissue which, as you know, is used by palpation is already the information that is<br />
important in a diagnosis. But, we think that it does not have enough objectivity. To increase objectivity by visualizing hardness of<br />
3268 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3268
tissue, we adopted ECAM �extended combined autocorrelation method�, which was developed by Professor Shiina at Tsukuba<br />
University in Japan, and succeeded in developing the commercial ultrasound scanner, which could display a strain image in real time.<br />
From a clinical point of view, in the breast region, mammography examination is effective in a diagnosis, but a judgment of<br />
permeation degree is not superior in ultrasound image. And in a thyroid gland region, we begin to get experience with availability<br />
from a diagnosis of papillary cancer and follicular cancer. So, we would like to have the presentation about the development of a strain<br />
imaging function and some of our clinical experiences by using the developed system.<br />
1:40<br />
4pBB3. Elasticity of perfused tissue. Kirk W. Beach �Dept. of Surgery, Univ. of Washington, Box 356410, Seattle, WA<br />
98195-6410�, Barbrina Dunmire, and John C. Kucewicz �Univ. of Washington, Seattle, WA 98105-6698�<br />
Elastic imaging intends to measure Young’s modulus �tissue stiffness� or bulk modulus �tissue compressibility� of tissue subjected<br />
to an applied strain of several percent. Underlying elastic imaging is the assumption of a linear stress/strain relationship without<br />
hysteresis or other time-dependent behavior. Perfused tissue is a composite material comprised of a solid matrix of cells, fibers,<br />
interstitial fluid �occupying up to 50% of the tissue volume and varying slowly with time�, arterioles �pulsating high-pressure spaces<br />
that occupy 0.1% of the tissue volume�, capillaries, and venules �low-pressure spaces that occupy up to 3% of the tissue volume<br />
varying with respiration�. This talk will speculate on the nonlinear, nonstationary stress/strain relationships expected from dependent<br />
tissues �legs�, pressurized tissues �breast tumors�, and other living, perfused tissues. The pressure versus strain curve from each tissue<br />
voxel allows the measurement of arteriolar and venular volumes and pressures, and interstitial pressure within the tissues. These<br />
volumes and pressures may be key to classifying pathologies.<br />
2:00<br />
4pBB4. New developments in transient elastography. Mathias Fink, Mickael Tanter, Ralph Sinkus, and Gabriel Montaldo �LOA,<br />
ESPCI, 10 rue Vauquelin, 75005, Paris, France�<br />
An ultra-high-rate ultrasonic scanner has been developed that can give 5000 ultrasonic images per second of the body. With such<br />
a high frame rate, the propagation of transient shear waves can be followed, and from the spatio-temporal evolution of the displacement<br />
fields, various inversion algorithms allow us to recover the shear modulus map. A discussion on the various inversion algorithms<br />
will be presented. In order to obtain unbiased shear elasticity map, different configurations of shear sources induced by radiation<br />
pressure of focused transducer arrays are used. Both 2-D and 3-D imaging can obtained with this technique. In vitro and in vivo results<br />
on breast will be presented that demonstrate the interest of elasticity imaging with transient elastography.<br />
2:20<br />
4pBB5. Spectral characteristics of breast vibro-acoustography<br />
images. Azra Alizad, Dana H. Whaley, Mathew Urban, Randall R.<br />
Kinnick, James F. Greenleaf, and Mostafa Fatemi �Mayo Clinic College<br />
of Medicine, Rochester, MN 55905 aza@mayo.edu�<br />
Vibro-acoustography image is a function of the dynamic characteristics<br />
of the object at the vibration �difference� frequency �df�. The dynamic<br />
characteristic of tissue is closely related to pathology. Therefore, it is<br />
important to evaluate image features versus df. Here, the influence of df on<br />
breast vibro-acoustography images is studied by scanning human breast at<br />
various df values ranging from 20 to 90 kHz. The subjects were chosen<br />
from a group of volunteers with different breast abnormalities. Images<br />
were compared subjectively to study image features and the appearances<br />
of breast lesions versus df. It is demonstrated that having a collection of<br />
images of the same tissue at different df values generally provides a better<br />
perception of the tissue structure and improves lesion identification. In<br />
most cases, higher df resulted in a higher signal-to-noise ratio and thus a<br />
higher image quality. Finally, a frequency-compounded images was obtained<br />
by calculating the weighted sum of images at different df values. It<br />
is demonstrated that image compounding normally improves visualization<br />
of breast tissue and abnormalities. �Work supported by NIH Grant EB-<br />
00535 and Grant BCTR0504550 from the Susan G. Komen Breast Cancer<br />
Foundation. Disclosure: Parts of the techniques used here are patented by<br />
MF and JFG.�<br />
Contributed Papers<br />
2:35<br />
4pBB6. Tissue pulsatility imaging: Ultrasonic measurement of strain<br />
due to perfusion. John C. Kucewicz, Barbrina Dunmire, Lingyun<br />
Huang, Marla Paun �Univ. of Washington Appl. Phys. Lab., 1013 NE<br />
40th St., Seattle, WA 98105-6698�, and Kirk W. Beach �Univ. of<br />
Washington, Seattle, WA 98195-6410�<br />
Over each cardiac cycle perfused tissues expand and relax by a fraction<br />
of a percent as blood rapidly accumulates in the arterial vasculature<br />
during systole and then slowly drains through the venous vasculature during<br />
diastole. Tissue pulsatility imaging �TPI� is a variation on ultrasonic<br />
tissue strain imaging that estimates tissue perfusion from this natural, cyclic<br />
tissue expansion and relaxation. TPI is derived in principle from plethysmography,<br />
a century-old technology for measuring gross tissue volume<br />
change from a whole limb or other isolatable body part. With TPI, the<br />
plethysmographic signal is measured from hundreds or thousands of<br />
sample volumes within an ultrasound image plane to characterize the local<br />
perfusion throughout a body part. TPI measures tissue strain over the<br />
cardiac cycle and parametrizes the signal in terms of its amplitude and<br />
shape. The amplitude of the strain waveform is correlated with perfusion,<br />
and the shape of the waveform is correlated with vascular resistance.<br />
Results will be presented from the leg showing the change in the TPI<br />
signals as the muscles recover from exercise, from breast tumors, and<br />
from the brain as blood flow changes in response to visual stimulation.<br />
�Work supported in part by NIH 1-R01EB002198-01 and NIH N01-CO-<br />
07118.�<br />
3269 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3269<br />
4p FRI. PM
2:50<br />
4pBB7. Using human body shear wave noise for passive elastography.<br />
Karim G. Sabra, Stephane Conti, Philippe Roux, and William A.<br />
Kuperman �Scripps Inst. of Ocean., Univ. of California—San Diego,<br />
9500 Gilman Dr., San Diego, CA 92093-0238�<br />
An elastography imaging technique based on passive measurement of<br />
shear wave ambient noise generated in the human body �e.g., due to the<br />
heart, muscles twitches, and blood flow system� has been developed. This<br />
technique merges two recent research developments in medical imaging<br />
and physics: �1� recent work on the efficacy of elastographic imaging<br />
demonstrating that shear waves are excellent candidates to image tissue<br />
elasticity in the human body and �2� theory and experimental verification<br />
in ultrasonics, underwater acoustics, and seismology of the concept of<br />
extracting coherent Green’s function from random noise cross correlations.<br />
These results provide a means for coherent passive imaging using<br />
only the human body noise field, without the use of external active<br />
sources. Coherent arrivals of the cross correlations of recordings of human<br />
body noise in the frequency band 2–50 Hz using skin-mounted accelerometers<br />
allows us to estimate the local shear velocity of the tissues. The<br />
coherent arrivals emerge from a correlation process that accumulates contributions<br />
over time from noise sources whose propagation paths pass<br />
through both sensors. The application of this passive elastography technique<br />
for constructing biomechanical models of in vivo muscles’ properties<br />
will be discussed.<br />
3:05–3:20 Break<br />
3:20<br />
4pBB8. Dynamic radiation force of acoustic waves on solid elastic<br />
spheres. Glauber T. Silva �Instituto de Computação, Universidade<br />
Federal de Algoas, Maceió, AL, 57072-970, Brazil�<br />
The present study concerns the dynamic radiation force on solid elastic<br />
spheres exerted by a plane wave with two frequencies �bichromatic wave�<br />
considering the nonlinearity of the fluid. Our approach is based on solving<br />
the wave scattering for the sphere in the quasilinear approximation within<br />
the preshock wave range. The dynamic radiation force is then obtained by<br />
integrating the component of the momentum flux tensor at the difference<br />
of the primary frequencies over the boundary of the sphere. Effects of the<br />
fluid nonlinearity play a major role in dynamic radiation force, leading it<br />
to a regime of parametric amplification. The developed theory is used to<br />
calculate the dynamic radiation force on three different solid spheres �aluminum,<br />
silver, and tungsten�. The obtained spectrum of dynamic radiation<br />
force presents resonances with larger amplitude and better shape than<br />
those exhibited in static radiation force. Applications of the results to some<br />
elasticity imaging techniques based on dynamic radiation force will be<br />
presented.<br />
3:35<br />
4pBB9. Ultrasonic measurement of displacement distribution inside<br />
an object caused by dual acoustic radiation force for evaluation of<br />
muscular relax property due to acupuncture therapy. Yoshitaka<br />
Odagiri, Hideyuki Hasegawa, and Hiroshi Kanai �Grad. School of Eng.,<br />
Tohoku Univ., Sendai 980-8579, Japan, odagiri@us.ecei.tohoku.ac.jp�<br />
Many studies have been carried out on the measurement of mechanical<br />
properties of tissues by applying an ultrasound-induced acoustic radiation<br />
force. To assess mechanical properties, strain of an object must be generated.<br />
However, one radiation force is not sufficient because it also causes<br />
translational motion when the object is much harder than surrounding<br />
medium. In this study, two cyclic radiation forces are applied to a muscle<br />
phantom from two opposite horizontal directions so that the object is<br />
cyclically compressed in the horizontal direction. As a result, the object is<br />
vertically expanded due to the incompressibility. The resultant vertical<br />
displacement is measured using ultrasound. Two concave ultrasonic transducers<br />
for actuation were both driven by sums of two continuous sinusoidal<br />
signals at two slightly different frequencies of 1 MHz and (1M<br />
�5) Hz. Displacement, which fluctuates at 5 Hz, was measured by the<br />
ultrasonic phased tracking method proposed by our group. Results indicated<br />
that the surface of the phantom was cyclically actuated with an<br />
amplitude of a tenth of a few micrometers, which well coincided with that<br />
measured with laser vibrometer. In addition, upward and downward displacements<br />
at the surface and deeper region were found during the increase<br />
phase of radiation forces. Such displacements correspond to the<br />
horizontal compression.<br />
3:50<br />
4pBB10. A phantom study on ultrasonic measurement of arterial wall<br />
strain combined with tracking of translational motion. Hideyuki<br />
Hasegawa and Hiroshi Kanai �Grad. School of Eng., Tohoku Univ.,<br />
Aramaki-aza-Aoba 6-6-05, Sendai 980-8579, Japan,<br />
hasegawa@us.ecei.tohoku.ac.jp�<br />
Correlation-based techniques are often applied to ultrasonic rf echoes<br />
to obtain the arterial wall deformation �strain�. In such methods, the displacement<br />
estimates are biased due to changes in center frequency of<br />
echoes. One of the reasons for the change in center frequency is the<br />
interference of echoes from scatterers within the wall. In the phased tracking<br />
method previously proposed for strain estimation by our group, the<br />
estimated displacement contains both the components due to the translational<br />
motion and strain. The translational motion is larger than strain by a<br />
factor of 10 and, thus, the error in the estimated displacement due to the<br />
change in center frequency mainly depends on translational motion and is<br />
often larger than the minute displacement due to strain. To reduce this<br />
error, in this study, a method is proposed in which the translational motion<br />
is compensated using the displacement of the luminal boundary estimated<br />
by the phased tracking method before correlating echoes between the<br />
frame before deformation and that at the maximum deformation to estimate<br />
the strain distribution within the wall. In basic experiments using<br />
phantoms made of silicone rubber, the estimation error was much reduced<br />
to 15.6% in comparison with 36.4% obtained by the previous method.<br />
4:05<br />
4pBB11. Wave biomechanics of skeletal muscle. Oleg Rudenko �Dept.<br />
of. Blekinge Inst. of Technol., 371 79 Karlskrona, Sweden� and Armen<br />
Sarvazyan �Artann Labs., Inc., West Trenton, NJ 08618�<br />
Physiological functions of skeletal muscle, such as voluntary contraction<br />
and force development, are accompanied by dramatic changes of its<br />
mechanical and acoustical properties. Experimental data show that during<br />
contraction, the muscle’s Young’s modulus, shear viscosity, and anisotropy<br />
parameter are changed by over an order of magnitude. None of the existing<br />
models of muscle contraction and muscle biomechanics can adequately<br />
explain the phenomena observed. A new mathematical model �O.<br />
Rudenko and A. Sarvazyan, Acoust. Phys. �6�, �<strong>2006</strong>��, has been developed<br />
relating the shear wave propagation parameters to molecular structure<br />
of the muscle and to the kinetics of the mechanochemical crossbridges<br />
between the actin and myosin filaments. New analytical solutions<br />
describing waves in muscle including nonlinear phenomena are found. A<br />
molecular mechanism for the dependence of acoustical characteristics of<br />
muscle on its fiber orientation and the contractile state is proposed. It is<br />
shown that although the anisotropy connected with the preferential direction<br />
along the muscle fibers is characterized by five elastic moduli, only<br />
two of these moduli have independent values in the muscle. The potential<br />
implications of the proposed model in terms of the acoustical assessment<br />
of muscle function are explored.<br />
3270 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3270
4:20<br />
4pBB12. Phase aberration correction for a linear array transducer<br />
using ultrasound radiation force and vibrometry optimization:<br />
Simulation study. Matthew W. Urban and James F. Greenleaf �Dept. of<br />
Physiol. and Biomed. Eng., Mayo Clinic College of Medicine, 200 First St.<br />
SW, Rochester, MN 55905, urban.matthew@mayo.edu�<br />
Diagnostic ultrasound images suffer from degradation due to tissues<br />
with sound speed inhomogeneities causing phase shifts of propagating<br />
waves. These phase shifts defocus the ultrasound beam, reducing spatial<br />
resolution and image contrast in the resulting image. We describe a phase<br />
aberration correction method that uses dynamic ultrasound radiation force<br />
to harmonically excite a medium using amplitude-modulated continuous<br />
wave ultrasound created by summing two ultrasound frequencies at f 0<br />
�3.0 MHz and f 0�� f �3.0005 MHz. The phase of each element of a<br />
linear array transducer is sequentially adjusted to maximize the radiation<br />
force and obtain optimal focus of the ultrasound beam. The optimization is<br />
performed by monitoring the harmonic amplitude of the scatterer velocity<br />
in the desired focal region using Doppler techniques. Simulation results<br />
show the ability to regain a 3.0-MHz focused field after applying a phase<br />
screen with an rms time delay of 95.4 ns. The radiation force magnitude<br />
increased by 22 dB and the resolution of the field was regained. Simulation<br />
results show that the focus of the beam can be qualitatively and<br />
quantitatively improved with this method. �This study was supported in<br />
part by Grants EB002640 and EB002167 from the NIH.�<br />
4:35<br />
4pBB13. Application of the optoacoustic technique to visualization of<br />
lesions induced by high-intensity focused ultrasound. Tatiana<br />
Khokhlova, Ivan Pelivanov, Vladimir Solomatin, Alexander Karabutov<br />
�Intl. Laser Ctr., Moscow State Univ., 119992, Moscow, Russia<br />
t_khokhlova@ilc.edu.ru�, and Oleg Sapozhnikov �Moscow State Univ.,<br />
119992, Moscow, Russia�<br />
Today several techniques are being applied to monitoring of highintensity<br />
focused ultrasound �HIFU� therapy, including MRI, conventional<br />
ultrasound, and elastography. In this work a new method for noninvasive<br />
monitoring of HIFU therapy is proposed: the optoacoustic method. The<br />
optoacoustic technique is based on the excitation of wideband ultrasonic<br />
pulses through the absorption of pulsed laser radiation in tissue and subsequent<br />
expansion of the heated volume. The excited optoacoustic �OA�<br />
pulse contains information on the distribution of optical properties within<br />
the tissue—light scattering and absorption coefficients. Therefore, if thermal<br />
lesions have different optical properties than the untreated tissue, they<br />
will be detectable on the OA waveform. The considerable change in light<br />
scattering and absorption coefficients after tissue coagulation was measured<br />
using techniques previously developed by our group. Heating induced<br />
by HIFU also influences the OA signal waveform due to the rise of<br />
thermal expansion coefficient of tissue with temperature. This dependence<br />
was measured in order to evaluate the feasibility of the OA technique in<br />
temperature monitoring. An OA image of HIFU lesion induced by a 1.1<br />
MHz focused transducer in a liver sample was reconstructed using a 64element<br />
wideband array transducer for OA signal detection.<br />
FRIDAY AFTERNOON, 1 DECEMBER <strong>2006</strong> OAHU ROOM, 1:00 TO 3:00 P.M.<br />
Session 4pEAa<br />
Engineering Acoustics: New Electroacoustic Transducers Utilizing Advanced Technolgies and Materials<br />
Juro Ohga, Cochair<br />
Shibaura Inst. of Technology, 3-9-14 Shibaura, Minato-ku, Tokyo 108-8548, Japan<br />
James E. West, Cochair<br />
Johns Hopkins Univ., Dept. of Electrical and Computer Engineering, Barton 105, 3400 N. Charles St.,<br />
Baltimore, MD 21218-2686<br />
Invited Papers<br />
1:00<br />
4pEAa1. Solid-state photo-microphones or pressure sensors by total reflection. Yasushi Suzuki �Gunma Natl. College 08 Tech.,<br />
580, Toriba-cho, Maebashi-shi, Gunma, 371-8530 Japan., suzuki@elc.gunma-ct.ac.jp� and Ken’iti Kido �Tohoku Univ.,<br />
Yokohama-shi, Kanagawa, 226-0017 Japan�<br />
Solid-state photo-microphones or pressure sensors are proposed. These sensors use a new principle, involving the optical total<br />
reflection at the boundary surface between glass and air. The critical angle for total reflection changes by the refractive index of air,<br />
which depends on the air density. Sound pressure changes the air density. Therefore, the sound pressure is measurable by detecting the<br />
intensity of the reflected light from the total reflection area. The sensitivity of the sensor is investigated theoretically. It is expected that<br />
the sensor has sufficient sensitivity for practical use, employing laser light and a curved boundary surface with a large radius of<br />
curvature. Some experiments are carried out to verify the theoretical investigations. A He-Ne laser or a laser diode is employed as a<br />
light source in the experiments. Experimental results show that the sensor has equivalent sensitivity to that which was theoretically<br />
estimated, but that sensitivity is very low. The sensor is useful as a pressure sensor, but it is difficult to realize a microphone for<br />
general use at the present. The microphones have no diaphragm and the upper limit in the frequency range is extremely high in<br />
principle.<br />
3271 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3271<br />
4p FRI. PM
1:20<br />
4pEAa2. Micromachined microphones with diffraction-based optical interferometric readout. F. Levent Degertekin �G.W.<br />
Woodruff School of Mech. Eng., Georgia Inst. of Technol., Atlanta, GA 30332, levent@gatech.edu�, Neal A. Hall �Sandia Natl. Labs,<br />
Albuquerque, NM 87185-5800�, and Baris Bicen �Georgia Inst. of Technol., Atlanta, GA 30332�<br />
A diffraction-based optical method for integrated interferometric detection of micromachined microphone diaphragm displacement<br />
is described. With multichip optoelectronics integration, this approach yields highly sensitive optical microphones in mm-cube<br />
volumes. Since the microphone sensitivity does not depend on capacitance, this method changes the paradigm for the backplate and<br />
gap structure design. As a result, one can use millimeter size diaphragms to achieve wide frequency response and low thermal<br />
mechanical noise levels characteristic of precision measurement microphones. Furthermore, the electrical port of the device, which is<br />
freed by optical detection, is used for electrostatic actuation of the microphone diaphragm to tune microphone sensitivity and to<br />
generate self-characterization signals. Prototype optical microphone structures have been fabricated using Sandia National Laboratories’<br />
silicon based SwIFT-Lite TM process. Measurements on these diaphragms show an A-weighted diaphragm displacement noise of<br />
2.4 pm and flat electrostatic response up to 20 kHz. These results indicate the feasibility of realizing measurement microphones with<br />
1.5-mm-diam diaphragms, 15-dBA internal noise, and 40-kHz bandwidth. Application of the detection method in a bio-inspired<br />
directional microphone for hearing aids is also discussed. �Work partially supported by NIH Grant 5R01DC005762-03, Sensing and<br />
Processing for Hearing Aids.�<br />
1:40<br />
4pEAa3. Hardware and software technologies for improvement of hearing characteristics of headphone reproduction.<br />
Kiyofumi Inanaga and Yuji Yamada �Audio Codec Development Dept., Technol. Development Group, SONY Corp., Shinagawa Tec.,<br />
12-15-3, Tokyo, 108-6201 Japan�<br />
This report specifically describes commercialization technology of a headphone system with out-of-head localization applying<br />
dynamic head-related transfer functions �HRTFs� that can localize sound easily over a full 360 deg. A source image by output of<br />
conventional headphones is localized inside the listener’s head. However, the image can be localized outside the listener’s head by<br />
wearing headphones over a full 360 deg through accurate simulation of the listener’s HRTFs. Developments of headphone systems<br />
using signal processing technology for data correction have given rise to the static binaural reproduction system �SBRS�. The first part<br />
of this speech describes its psychoacoustic characteristics and challenges. A rotating dummy-head that is synchronized with the<br />
listener’s head movement was produced experimentally to create the dynamic binaural reproduction system �DBRS�. Using the<br />
DBRS, HRTFs synchronize with the listener’s head movement. Psychoacoustic characteristics and advantages of the system are also<br />
discussed in this report. Further developments were made to realize the commercialization of the DBRS in areas including piezoelectric<br />
gyroscope head-tracking technology, headphone technologies that can reproduce real sound characteristics, and simplification of<br />
HRTF signal processing employing a simulator with electronic circuits. Finally, future visions for these technologies will be touched<br />
upon.<br />
2:00<br />
4pEAa4. Piezoelectret microphones: A new and promising group of transducers. Gerhard M. Sessler and Joachim Hillenbrand<br />
�Darmstadt Univ. of Technol., Merckstrasse 25, 64283 Darmstadt, Germany, g.sessler@nt.tu-darmstadt.de�<br />
Piezoelectret microphones, first described a few years ago, are transducers based on the strong longitudinal piezoelectric effect of<br />
charged cellular polymers. Such microphones have recently been improved in two respects: Firstly, an expansion process was used to<br />
increase the piezoelectric d 33 coefficients of cellular polypropylene �PP� films in the audio frequency range up to 600 pC/N and,<br />
secondly, stacking of several films was applied to increase the microphone sensitivity. Transducers with six films now show opencircuit<br />
sensitivities of up to 15 mV/Pa, comparable to that of electret microphones. Other characteristics of piezoelectret microphones<br />
are their low equivalent noise level of about 26 dB�A� and the very small total harmonic distortion of less than 0.1% at 140 dB SPL.<br />
The piezoelectric activity of the PP films and the microphone sensitivities are stable at <strong>room</strong> temperature but start to decay above<br />
50 °C. Recently, directional piezoelectret microphones with various directional characteristics have been designed. Major advantages<br />
of piezoelectret microphones are their simple design, their low harmonic distortion, and their wide frequency range extending into the<br />
ultrasonic region.<br />
2:20<br />
4pEAa5. Expansion of frequency range for piezoelectric loudspeakers by new transducer construction. Juro Ohga �Shibaura<br />
Inst. of Technol., 3-7-5, Toyosu, Koto-ku, Tokyo 135-8548, Japan�<br />
Although simple construction of piezoelectric loudspeakers engenders various merits, expansion of its working frequency range to<br />
the very low region is difficult because the mechanically stiff characteristics of conventional piezoelectric ceramic diaphragms prevent<br />
their large amplitude operation. This paper proposes two sorts of new piezoelectric loudspeaker construction that are suitable for<br />
low-frequency signal radiation. One idea is the use of a tuck-shape diaphragm by a PVDF polymer film bimorph. It has large surface<br />
area with a very low resonant frequency. Resonant frequencies and sensitivity frequency characteristics are examined, and control<br />
methods of local diaphragm bending are discussed. The other idea is the use of continuous revolution of a piezoelectric ultrasonic<br />
motor. It produces a completely controlled large output force because its output mechanical impedance is much greater than that of<br />
any conventional transducer or motor. An ultrasonic motor, whose stator is connected to a direct-radiator loudspeaker cone by a rod<br />
and whose rotor is burdened by a heavy metal ring, rotates with a constant velocity. Modulation of the velocity by using an audio<br />
signal imparts a driving force to the diaphragm because the heavy ring tends to keep a constant velocity. Experimental models suggest<br />
that this construction is useful.<br />
3272 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3272
2:40<br />
4pEAa6. Modal array signal processing using circular microphone arrays applied to acoustic source detection and localization<br />
problems. Heinz Teutsch �Avaya Labs, 233 Mt. Airy Rd., Basking Ridge, NJ 07920, teutsch@avaya.com� and Walter Kellermann<br />
�Univ. of Erlangen-Nuremberg, Erlangen, Germany�<br />
Many applications of acoustic signal processing rely on estimates of several parameters present in the observed acoustic scene<br />
such as the number and location of acoustic sources. These parameters have been traditionally estimated by means of classical array<br />
signal processing �CASP� algorithms using microphone arrays. Algorithms for parameter estimation solely based on the paradigm of<br />
CASP often suffer from the narrowband assumption underlying the signal model. This restriction limits their usability when wideband<br />
signals, such as speech, are present in the wave field under observation. We investigate the parameter estimation problem by applying<br />
the notion of wave field decomposition using baffled circular microphone arrays. The obtained wave field representation is used as the<br />
basis for ‘‘modal array signal processing algorithms.’’ It is shown that by applying the notion of modal array signal processing, novel<br />
algorithms can be derived that have the potential to unambiguously detect and localize multiple simultaneously active wideband<br />
sources in the array’s full field-of-view. Performance evaluations by means of simulations, measurements, and real-time case studies<br />
are presented.<br />
FRIDAY AFTERNOON, 1 DECEMBER <strong>2006</strong> OAHU ROOM, 3:15 TO 6:00 P.M.<br />
Session 4pEAb<br />
Engineering Acoustics: Special Topics in Engineering Acoustics<br />
Timothy W. Leishman, Cochair<br />
Brigham Young Univ., Dept. of Physics and Astronomy, N247 ESC, Provo, UT 84602<br />
3:15<br />
4pEAb1. Enhanced voided piezoelectric polymer for underwater<br />
acoustic sensors. Juan Arvelo �Appl. Phys. Lab., Johns Hopkins Univ.,<br />
11100 Johns Hopkins Rd., Laurel, MD 20723-6099�, Ilene<br />
Busch-Vishniac, and James West �Johns Hopkins Univ., Baltimore, MD<br />
21218�<br />
A charged voided polymer has been shown to exhibit large piezoelectricity.<br />
This material consists of injected air bubbles into polypropylene.<br />
This sheet of voided material is then biaxially stretched to elongate the<br />
voids. After stretching this material, a strong electric field is applied to<br />
cause dielectric breakdown of the gas in the voids, creating electric<br />
charges that are trapped in the polymer frame. Since the sides of the voids<br />
have opposite charges, they form macroscopic dipoles. When an external<br />
force is applied to this material, the voids become narrower, causing stronger<br />
dipole strength. A simple model of this voided material was implemented<br />
to derive formulas to estimate its piezoelectric constant, electromechanical<br />
coupling factor, resonance frequency, and sensor sensitivity<br />
based on electrical and mechanical properties of the polymer and gas in<br />
the voids. These formulas and a survey of available polymers and gases<br />
yielded promising combinations that result in more sensitive voided materials<br />
that satisfy selected criteria. These criteria include high sensitivity<br />
and maximum service temperature, low dissipation factor, and high dynamic<br />
compressibility, but low hydrostatic compressibility. This talk will<br />
describe the model, derive the formulas, uncover measured properties of<br />
candidate polymers and gases, and show calculated sensitivity of selected<br />
polymer/gas combinations.<br />
Kiyofumi Inanaga, Cochair<br />
Sony Corp., Shinagawa Tec. 12-15-3, Tokyo 108-6201, Japan<br />
Contributed Papers<br />
3:30<br />
4pEAb2. Basic study on one-dimensional transducer array using<br />
hydrothermally synthesized lead zirconium titanete poly-crystalline<br />
film. Akito Endo, Tomohito Hasegawa, Norimichi Kawashima, Shinichi<br />
Takeuchi �1614, Kurogane-cho, Aoba-ku, Yokohama, Kanagawa,<br />
225-8502, Japan�, Mutsuo Ishikawa, and Minoru Kurosawa �Midori-ku,<br />
Yokohama, Kanagawa 226-8502, Japan�<br />
Recently, high-frequency miniature medical ultrasound probes with<br />
high resolution were actively developed. However, it is difficult to fabricate<br />
such tiny ultrasound probes using piezoelectric ceramic vibrator with<br />
thickness less than 100 �m. We deposited a PZT poly-crystalline film on<br />
a titanium substrate using the hydrothermal method and developed transducers<br />
using the PZT poly-crystalline film for ultrasound probes. In this<br />
study, we applied it to a miniature medical one-dimensional �1-D�-arraytype<br />
ultrasound probe with resonance frequency of 10 MHz. After sputtering<br />
of pure titanium on the surface of a hydroxyapatite substrate, the<br />
titanium film was etched using the photolithography method to form a 1-D<br />
titanium film electrode array with 75 �m element pitch, 40 �m element<br />
width, and 4 mm element length to scan an ultrasound beam electronically<br />
by sector scan mode using phased-array technique. Thereby we fabricated<br />
a miniature 1-D-array-type ultrasound probe. A transmitted ultrasound<br />
pulse from 10 MHz commercial ultrasound probe was received by this<br />
fabricated 1-D-array type ultrasound probe with hydrothermally synthesized<br />
PZT poly-crystalline film vibrators.<br />
3273 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3273<br />
4p FRI. PM
3:45<br />
4pEAb3. Analysis of a barrel-stave flextensional transducer using<br />
MAVART „model to analyze the vibrations and acoustic radiation of<br />
transducers… and ATILA „analysis of transducers by integration of<br />
LAplace equations… finite-element codes. Richard A. G. Fleming, Mark<br />
Kwiecinski, and Dennis F. Jones �Defence R&D Canada—Atlantic, P.O.<br />
Box 1012, Dartmouth, NS, Canada B2Y 3Z7,<br />
dennis.jones@drdc-rddc.gc.da�<br />
A small barrel-stave flextensional transducer, designed and tested at<br />
Defence Research and Development Canada—Atlantic, is a candidate<br />
sound source for underwater coastal surveillance and acoustic communications<br />
applications. This high-power transducer has an outside diameter,<br />
length, and mass of 5.7 cm, 12.7 cm, and 1.1 kg, respectively. The measured<br />
fundamental flexural resonance frequency was 1.8 kHz with a transmitting<br />
voltage response of 118 dB/1�Pa-m/V and an omnidirectional<br />
radiation pattern. Two finite-element models were developed for this transducer<br />
using the finite-element codes MAVART �Model to Analyze the<br />
Vibrations and Acoustic Radiation of Transducers� and ATILA �Analysis<br />
of Transducers by Integration of LAplace equations�. Comparisons are<br />
made between the calibration measurements and the model predictions.<br />
�Work supported in part by Sensor Technology Limited.�<br />
4:00<br />
4pEAb4. Thermal behavior of high-power active devices with the<br />
ATILA „analysis of transducers by integration of LAplace equations…<br />
finite-element code. Jean-Claude Debus �Institut Superieur de<br />
l’Electronique et du Numerique, 41 Bv Vauban, 59046 Lille, Cedex<br />
France�, John BlottmanIII, and Stephen Butler �Naval Undersea Warfare<br />
Ctr. Div. Newport, RI 02841�<br />
Many active devices using piezoelectric ceramics are driven with very<br />
high power densities and long pulse lengths. Due to mechanical and dielectric<br />
losses in the materials, this produces heat, causing a temperature<br />
rise in the devices, which may lead to their mechanical failure. The thermal<br />
issues have been shown to be the limiting device design criteria over<br />
electric field and mechanical stress limits, yet the effect of the temperature<br />
on performance is generally not considered in the numerical models used<br />
during the design stage. A coupled electro-mechanical thermal analysis is<br />
implemented in the ATILA code. For a steady-state or transient solution, a<br />
thermal behavior is weakly coupled to the electromechanical response.<br />
The method may take advantage of the order-of-magnitude-greater time<br />
constant for thermal effects compared to mechanical behavior. A two-step<br />
analysis is performed whereby the electromechanical behavior is first<br />
computed, and the resulting dissipated power is then applied as a heat<br />
generator to determine the resulting temperature of the device. A highdrive,<br />
31-mode, free flooded ring transducer and a sonar projector serve as<br />
validation of the numerical model. The approach addresses both the transient<br />
thermal response and the steady temperature profile that results from<br />
the high-power, high-duty-cycle drive.<br />
4:15<br />
4pEAb5. Development of multichannel optical sensor and<br />
visualization of vibration distribution. Jun Hasegawa and Kenji<br />
Kobayashi �Faculty of Eng., Takushoku Univ., 815-1 Tatemachi,<br />
Hachioji-shi, Tokyo 193-0985 Japan, jhase@es.takushoku-u.ac.jp�<br />
A multi-channel optical sensor system was developed to measure vibrations<br />
simultaneously with high spatial resolution. As sensor elements,<br />
optical displacement sensor units were developed not to disturb the natural<br />
vibration. Each sensor unit, which consists of the optical fiber bundle and<br />
focusing lens, can detect the displacement of the object as the variation of<br />
the reflected light power. The sensor unit has a displacement resolution of<br />
10 nm, a dynamic range of more than 90 dB, and a frequency band width<br />
of up to 80 kHz. Up to 64 sensor units can be arrayed as one sensor head,<br />
which realizes the simultaneous measurement of vibration distribution<br />
with the high spatial resolution of 4 mm. A calibrating function under the<br />
measurement circumstances was developed. Under calibration mode, the<br />
sensor array head is moved by a linear actuator, while the vibration of the<br />
object is stopped. Thus the calibrated data of each sensor unit can be<br />
obtained for the displacement magnitude. Measured vibration distributions<br />
can be monitored as the three-dimensional animations. With the system<br />
developed, several actuators for vibratory micro-injection were measured,<br />
and the system could reveal their detailed vibration distributions and could<br />
detect the existence of a failure portion of some actuator.<br />
4:30<br />
4pEAb6. Prediction of howling for a sound system in an acoustical<br />
environment with both reverberant and direct sounds. Hideki<br />
Akiyama and Juro Ohga �Shibaura Inst. of Technol., 3-7-5 Toyosu,<br />
Koto-ku, Tokyo 135-8548, Japan, m106003@shibaura-it.ac.jp�<br />
Prediction of howling is a key technology for a howling suppression<br />
design for a sound system with a loudspeaker and microphone. A howling<br />
occurrence prediction method for a sound system in a reverberant <strong>room</strong><br />
has already been presented �J. Ohga, J. Sakaguchi, ‘‘Prediction of howling<br />
of a sound system in a reverberant <strong>room</strong>,’’ W. C. Sabine Centennial Symposium<br />
�ASA, New York, 1994�, 2aAAd4�. It is apparently useful for<br />
ordinary public address systems whose distances of loudspeakers from<br />
microphones are large. However, this result was not perfect because the<br />
direct sound component is not negligible in hands-free telephones or teleconference<br />
systems whose loudspeakers and microphones are set close to<br />
each other. This report gives a quantitative howling occurrence prediction<br />
method for a sound system in an acoustical environment with both reverberant<br />
and direct sounds. The following design parameters are obtained:<br />
�1� the increase of howling occurrence level from the power average<br />
value, �2� the level occurrence probability, and �3� the critical level chart<br />
given by an equation as a function of direct and reverberant sounds ratio.<br />
Prediction results for particular examples are compared with calculations<br />
of sound-field transfer functions. Results confirmed that it is practical.<br />
4:45<br />
4pEAb7. Effect of background noise on dialogue in telephony. Koichi<br />
Amamoto and Juro Ohga �Sibaura Inst. of Technol., 3-7-5 Toyosu,<br />
Koto-ku, Tokyo, 135-8548, Japan, m106006@shibaura-it.ac.jp�<br />
Recent developments of mobile telephones include new sorts of impairments<br />
against speech. Conventional evaluation method for impairments<br />
by a talker and a few listeners cannot apply to these new ones,<br />
because they are brought by long signal delay. The effect of it cannot<br />
discriminate by ‘‘one-sided’’ test. This research relates to a speech quality<br />
evaluation by conversation between two persons. Variation of conversation<br />
stream is observed by addition of pink noise of various levels to a dialogue<br />
by microphones and earphones. Length of sentences and frequency of<br />
repeats are quantified and their meanings are discussed<br />
5:00<br />
4pEAb8. Best practices for auditory alarm design in space<br />
applications. Durand Begault and Martine Godfroy �Human Systems<br />
Integration Div., NASA Ames Res. Ctr., Moffett Field, CA 94035�<br />
This presentation reviews current knowledge in the design of auditory<br />
caution and warning signals, and sets criteria for development of ‘‘best<br />
practices’’ for designing new signals for NASA’s Crew Exploration Vehicle<br />
�CEV� and other future spacecraft, as well as for extra-vehicular<br />
operations. A design approach is presented that is based upon crossdisciplinary<br />
examination of psychoacoustic research, human factors experience,<br />
aerospace practices, and acoustical engineering requirements. Existing<br />
alarms currently in use with the NASA Space Shuttle flight deck are<br />
analyzed and then alternative designs are proposed that are compliant with<br />
ISO 7731 �‘‘Danger signals for work places Auditory Danger Signals’’�<br />
and that correspond to suggested methods in the literature to insure discrimination<br />
and audibility. Parallel analyses are shown for a sampling of<br />
medical equipment used in surgical, periop, and ICU contexts. Future<br />
development of auditory sonification techniques into the design of alarms<br />
will allow auditory signals to be extremely subtle, yet extremely useful in<br />
subtly indicating trends or root causes of failures. �Work funded by<br />
NASA’s Space Human Factors Engineering Project.�<br />
3274 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3274
5:15<br />
4pEAb9. Acoustic signal analysis for forensic applications. Durand<br />
Begault and Christopher Peltier �Audio Forensic Ctr., Charles M. Salter<br />
Assoc., 130 Sutter St., Ste. 500, San Francisco, CA 94104,<br />
durand.begault@cmsalter.com�<br />
Acoustical analysis of audio signals is important in many legal contexts<br />
for determining the authenticity, originality, and continuity of recorded<br />
media; determining the circumstances of events in question that<br />
may have been recorded; for determining the audibility of signals; and for<br />
identification or elimination of talkers as a match to an unknown exemplar.<br />
Recorded media are analyzed in forensic applications using both familiar<br />
techniques �waveform and spectral analyses� and more novel methods<br />
�e.g., ferro fluid development of media; specialized tape heads with nonstandard<br />
reproduction characteristics; crystal microscopy; detection and<br />
matching to power grid frequencies�. Audibility analyses frequently require<br />
careful reconstructive field measurements and criteria in excess of<br />
normally accepted standards. Voice identification-elimination protocols<br />
must account for examiner bias and exemplar quality and can be described<br />
using a receiver operator curve �ROC� model. This presentation gives an<br />
overview of these techniques and their comparative advantages for specific<br />
forensic applications.<br />
5:30<br />
4pEAb10. Without low-pass filter for the 1-bit digital amplifier.<br />
Kiyoshi Masuda �Coroprate Res. and Development Group, SHARP<br />
Corp., 2613-1 Ichinomoto-cho, Tenri-shi, Nara, Japan� and Yoshio<br />
Yamasaki �Waseda Univerity, Okubo, Shinzyuku-ku, Tokyo, Japan�<br />
SHARP collaborated with Waseda University from 1990 for 1-bit digital<br />
technology. SHARP had started to receive an order for the 1-bit digital<br />
amplifier ‘‘SM-SX100’’ on 20 August 1999. Until today, we have introduced<br />
the 1-bit digital amplifier for audio, flat panel TV �LCD TV�, and<br />
PC. These 1-bit amplifiers provided low-pass filter for the final stage,<br />
which is provided after 1-bit digital switching. We have to achieve more<br />
good sound and reduce deterioration of this low-pass filter. We have introduced<br />
new 1-bit digital amplifier without this low-pass filter beginning<br />
this April. This means we controlled the 1-bit digital signal to directly<br />
operate the speaker. We have proved a better effect for sound to compare<br />
the new 1-bit digital amplifier with the PWM switching amplifier, the<br />
A-class amplifier and the 1-bit digital amplifier with low-pass filter. If we<br />
do not measure any improvement for this new 1-bit digital amplifier, it has<br />
large radiation noise. We had achieved a reduction to the limit level of<br />
FCC, Denanhou, etc.<br />
5:45<br />
4pEAb11. Force-frequency effect of thickness mode langasite<br />
resonators. Haifeng Zhang �W317.4 Nebraska Hall, Univ. of Nebraska,<br />
Lincoln, NE 68588-0526, hfzhang@bigred.unl.edu�, Joseph A. Turner,<br />
Jiashi Yang �Univ. of Nebraska, Lincoln, NE 68588-0526�, and John A.<br />
Kosinski �U.S. Army CECOM, Fort Monmouth, NJ 07703-5211�<br />
Langasite resonators are of recent interest for a variety of applications<br />
because of their good temperature behavior, good piezoelectric coupling,<br />
low acoustic loss, and high Q factor. The force-frequency effect describes<br />
the shift in resonant frequency a resonator experiences due to the application<br />
of a mechanical load. A clear understanding of this effect is essential<br />
for many design applications such as pressure sensors. In this presentation,<br />
the frequency shift is analyzed theoretically and numerically for thin, circular<br />
langasite plates subjected to a diametrical force. The results are<br />
compared with experimental measurements of the same system for a variety<br />
of langasite resonators with various material orientations. In addition,<br />
the sensitivity of force-frequency effect is analyzed with respect to the<br />
nonlinear material constants. A comparison between the force-frequency<br />
effect of langasite and quartz resonators is also made. Finally, the application<br />
of such measurements for determining third-order elastic constants<br />
is discussed. �Work supported by ARO.�<br />
FRIDAY AFTERNOON, 1 DECEMBER <strong>2006</strong> IAO NEEDLE/AKAKA FALLS ROOM, 1:20 TO 4:25 P.M.<br />
Session 4pMU<br />
Musical Acoustics and Psychological and Physiological Acoustics: Acoustic Correlates of Timbre in Musical<br />
Instruments<br />
James W. Beauchamp, Cochair<br />
Univ. of Illinois Urbana-Champaign, School of Music, Dept. of Electrical and Computer Engineering, 1002 Eliot Dr.,<br />
Urbana, IL 61801<br />
Mashashi Yamada, Cochair<br />
Kanazawa Inst. of Technology, Dept. of Media Informatics, 3-1 Yatsukaho, Hakusan, Ishikawa 924-0838, Japan<br />
Invited Papers<br />
1:20<br />
4pMU1. A meta-analysis of acoustic correlates of timbre dimensions. Stephen McAdams, Bruno Giordano �CIRMMT, Schulich<br />
School of Music, McGill Univ., 555 Sherbrooke St. West, Montreal, QC, Canada H3A 1E3�, Patrick Susini, Geoffroy Peeters<br />
�STMS-IRCAM-CNRS, F-75004 Paris, France�, and Vincent Rioux �Maison des Arts Urbains, F-75020 Paris, France�<br />
A meta-analysis of ten published timbre spaces was conducted using multidimensional scaling analyses �CLASCAL� of dissimilarity<br />
ratings on recorded, resynthesized, or synthesized musical instrument tones. A set of signal descriptors derived from the tones<br />
was drawn from a large set developed at IRCAM, including parameters derived from the long-term amplitude spectrum �slope,<br />
centroid, spread, deviation, skewness, kurtosis�, from the waveform and amplitude envelope �attack time, fluctuation, roughness�, and<br />
from variations in the short-term amplitude spectrum �flux�. Relations among all descriptors across the 128 sounds were used to<br />
determine families of related descriptors and to reduce the number of descriptors tested as predictors. Subsequently multiple correlations<br />
between descriptors and the positions of timbres along perceptual dimensions determined by the CLASCAL analyses were<br />
3275 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3275<br />
4p FRI. PM
computed. The aim was �1� to select the subset of acoustic descriptors �or their linear combinations� that provided the most generalizable<br />
prediction of timbral relations and �2� to provide a signal-based model of timbral description for musical instrument tones.<br />
Four primary classes of descriptors emerge: spectral centroid, spectral spread, spectral deviation, and temporal envelope �effective<br />
duration/attack time�. �Work supported by CRC, CFI, NSERC, CUIDADO European Project.�<br />
1:40<br />
4pMU2. Perceptual acoustics of consonance and disssonance in multitimbral triads. Roger Kendall �Music Cognition and<br />
Acoust. Lab., Program in Systematic Musicology, UCLA, 405 Hilgard Ave., Los Angeles, CA 90095, kendall@ucla.edu� and Pantelis<br />
Vassilakis �DePaul Univ., Chicago, IL 60614�<br />
Most studies of consonance and dissonance assume a singular spectrum for the constituent intervals of a dyad. Recently, the<br />
principal author conducted experiments evaluating triads consisting of digitally mixed combinations drawn from the MUMS singlenote<br />
natural-instrument recordings. Results indicated that the main effect of ratings for consonance and dissonance correlated well<br />
with studies using artificial signals. However, interaction effects suggested perceptual differences related to the timbral differences<br />
across combinations. The present experiment evaluates perceptual and acoustical variables of the ten possible triadic combinations<br />
created with C4 as the lower and the ten with C5 as the upper notes. UCLA wind ensemble performers on oboe, flute, and clarinet,<br />
combinations designed to span timbral space, were digitally recorded. Analyses include perceptual ratings of consonance and dissonance,<br />
similarity, as well as acoustical analysis of roughness using a recently developed model. Since natural performances of any type<br />
vary in fundamental frequency, additional experiments will employ emulated oboe, flute, and clarinet �using the Kontakt Silver<br />
synthesizer in Sibelius 4� as well as purely synthetic stimuli, in order to ascertain the relationship of time-variant spectral properties<br />
to consonance, dissonance, and perceived similarity.<br />
2:00<br />
4pMU3. Multidimensional scaling analysis of centroid- and attackÕdecay-normalized musical instrument sounds. James W.<br />
Beauchamp �School of Music and Dept. of Elect. & Comput. Eng., Univ. of Illinois at Urbana–Champaign, Urbana, IL 61801,<br />
jwbeauch@uiuc.edu�, Andrew B. Horner �Hong Kong Univ. of Sci. & Technol., Kowloon, Hong Kong�, Hans-Friedrich Koehn, and<br />
Mert Bay �Univ. of Illinois at Urbana-Champaign, Urbana, IL 61801�<br />
Ten sustained musical instrument tones �bassoon, cello, clarinet, flute, horn, oboe, recorder, alto saxophone, trumpet, and violin�<br />
were spectrally analyzed and then equalized for duration, attack and decay time, fundamental frequency, number of harmonics,<br />
average spectral centroid, and presentation loudness. The tones were resynthesized both with time-varying harmonic amplitudes and<br />
frequencies �dynamic case� and fixed amplitudes and frequencies �static case�. Tone triads were presented to ten musically experienced<br />
listeners whose tasks were to specify the most dissimilar and most similar pairs in each triad. Based on the resulting dissimilarity<br />
matrix, multidimensional scaling �MDS� was used to position the instruments in two- and three-dimensional metric spaces. Two<br />
measures of instrument amplitude spectra were found to correlate strongly with MDS dimensions. For both the static- and dynamiccase<br />
2-D solutions, the ratio of even-to-odd rms amplitudes correlated strongly with one of the dimensions. For the dynamic case,<br />
spectral centroid variation correlated strongly with the second dimension. Also, 2-D solution instrument groupings agreed well with<br />
groupings based on coefficients of the first two components of a principle components analysis representing 90% of the instruments’<br />
spectral variance. �This work was supported by the Hong Kong Research Grants Council’s CERG Project 613505.�<br />
2:20<br />
4pMU4. Sound synthesis based on a new micro timbre notation. Naotoshi Osaka �School of Eng., Tokyo Denki Univ., 2-2,<br />
Kanda-Nishikicho, Chiyoda-ku, Tokyo, 101-8457, Japan, osaka@im.dendai.ac.jp�, Takayuki Baba, Nobuhiko Kitawaki, and Takeshi<br />
Yamada �Univ. of Tsukuba, Japan�<br />
Timbre has become a major musical factor in contemporary and computer music. However, sufficient timbre theory has not yet<br />
been established. The author is challenging to create new timbre theory for music composition. The first step of its construction is to<br />
make the timbre descriptive. A micro timbre is defined, which is a perceptual impression of a sound with approximately 50 to 100-ms<br />
duration, and describe sound as a micro timbre sequence. This can be used as a new notation system in place of common music<br />
notation. In dictation, micro timbre sequence and correspondent duration sequence are perceptually recorded. When synthesizing from<br />
this notation, sounds corresponding to the notation systems are either physically synthesized or searched for in a large sound database<br />
to generate sound data for a given duration. Two sequential sound data instances are first represented in sinusoidal representations and<br />
then are concatenated using a morphing technique. Sounds generated by a stream of water and similar sounds are described using the<br />
method as examples. Then scripts describing electronic sounds are introduced and explained. The ability to record, transmit to others,<br />
and resynthesize timbre is one of the useful functions of the theory.<br />
2:40<br />
4pMU5. Timbre representation for automatic classification of musical instruments. Bozena Kostek �Gdansk Univ. of Technol.,<br />
Narutowicza 11/12, PL-80-952 Gdansk, Poland�<br />
Human communication includes the capability of recognition. This is particularly true of auditory communication. Music information<br />
retrieval �MIR� turns out to be particularly challenging, since many problems remain still unsolved. Topics that should be<br />
included within the scope of MIR are automatic classification of musical instruments/phrases/styles, music representation and indexing,<br />
estimating musical similarity using both perceptual and musicological criteria, recognizing music using audio and/or semantic<br />
description, language modeling for music, auditory scene analysis, and others. Many features of music content description are based<br />
on perceptual phenomena and cognition. However, it can easily be observed that most of the low-level descriptors used, for example,<br />
in musical instrument classification are more data- than human-oriented. This is because the idea behind these features is to have data<br />
3276 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3276
defined and linked in such a way as to be able to use it for more effective automatic discovery, integration, and reuse in various<br />
applications. The ambitious task is, however, to provide seamless meaning to low- and high-level descriptors such as timbre descriptors<br />
and linking them together. In such a way data can be processed and shared by both systems and people. This paper presents a<br />
study related to timbre representation of musical instrument sounds.<br />
3:00–3:15 Break<br />
3:15<br />
4pMU6. An attempt to construct a quantitative scale of musical brightness for short melodies implementing timbral<br />
brightness. Masashi Yamada �Kanazawa Inst. of Technol., 3-1 Yatsukaho, Hakusan, Ishikawa 924-0838, Japan,<br />
m-yamada@neptune.kanazawa-it.ac.jp�<br />
It is known that a major tune is brighter than a minor one, and that music played in a faster tempo and a higher register is brighter<br />
than a slower and lower one. However, it has not been clarified how these factors quantitatively determine the musical brightness. On<br />
the other hand, it has been clarified that the timbral brightness of a tone corresponds well to the spectral centroid. In the present study,<br />
major and minor scales and two short melodies were played with pure tones, and listeners evaluated their musical brightness. For each<br />
performance, the spectral centroid was calculated for the overall-term spectrum during the performance on the transformed frequency<br />
scale of the ERB rate. The results showed that the musical brightness of the ascending scale increases proportionally as the spectral<br />
centroid shown in the ERB rate increases. Using this, a quantitative scale of musical brightness, BM, was constructed. The results also<br />
showed that the difference in the musical brightness between major and minor scales corresponded to the transposition of approximately<br />
5 ERB rate, and doubling the speed corresponded to the upper shift of the centroid in approximately 2.5 ERB rate.<br />
3:35<br />
4pMU7. Subjective congruency between a sound effect and a switching pattern of a visual image. Shinichiro Iwamiya,<br />
Motonori Arita, and Sun Su �Dept. of Acoust. Design, Kyushu Univ., 4-9-1, Shiobaru, Minami-ku, Fukuoka 185-8540, Japan�<br />
The relationship between the transformation of a visual image and the pitch pattern of sound can create formal congruency<br />
between sounds and moving pictures. This effect by switching patterns of enlarging and reducing images that were combined with<br />
ascending and descending pitch scales was examined. Rating experiments showed two congruent patterns of the combination of<br />
switching and scale patterns; one was a combination of an ascending pitch scale and an enlarging image pattern, and the other a<br />
combination of a descending pitch scale and a reducing image pattern. These forms of matching might be based on a Doppler illusion.<br />
An additional pair of congruent patterns for combinations of switching and scale patterns was also found: one was a combination of<br />
an ascending pitch scale and a sliding movement from left to right, and the other a combination of a descending pitch scale and a<br />
sliding movement from right to left. These forms of matching might be based on the correspondence of a progressive sensation.<br />
Further, the formal congruency between a pitch pattern and the formal transportation can contribute to integrating auditory and visual<br />
information and to making audio-visual products more impressive.<br />
3:55<br />
4pMU8. SRA: An online tool for spectral and roughness analysis of<br />
sound signals. Pantelis Vassilakis �School of Music, ITD, Libraries,<br />
DePaul Univ., 2350 N. Kenmore Ave., JTR 207, Chicago, IL 60614�<br />
SRA performs spectral and roughness analysis on user-submitted 250to<br />
1000-ms-long portions of sound files �.wav/.aif formats�. Spectral<br />
analysis incorporates an improved STFT algorithm �K. Fitz and L. Haken,<br />
J. Aud. Eng. Soc. 50�11�, 879–893 �2002�� and automates spectral peakpicking<br />
using the Loris open source C�� class library �Fitz and Haken<br />
�CERL Sound Group��. Users can manipulate three spectral analysis/peakpicking<br />
parameters: analysis bandwidth, spectral-amplitude normalization,<br />
and spectral-amplitude threshold. Instructions describe the parameters in<br />
detail and suggest settings appropriate to the submitted files and questions<br />
of interest. The spectral values obtained from the analysis enter a roughness<br />
estimation model �P. N. Vassilakis, Sele. Rep. in Ethnomusicol. 12,<br />
119–144 �2005��, outputting roughness values for each individual sinepair<br />
in the file’s spectrum and for the entire file. The roughness model<br />
quantifies the dependence of roughness on a sine-pair’s �a� intensity �combined<br />
amplitude of the sines�, �b� amplitude fluctuation degree �amplitude<br />
difference of the sines�, �c� amplitude fluctuation rate �frequency difference<br />
of the sines�, and �d� register �lower sine frequency�. Presentation of<br />
the roughness estimation model and the online tool will be followed by a<br />
discussion of research studies employing it and an outline of future possible<br />
applications. �Work supported by DePaul University and Eastern<br />
Washington University. Programmed by K. Fitz.�<br />
Contributed Papers<br />
4:10<br />
4pMU9. Further spectral correlations of timbral adjectives used by<br />
musicians. Alastair C. Disley, David M. Howard, and Andrew D. Hunt<br />
�Dept. of Electron., Univ. of York, Heslington, York, YO10 5DD, UK�<br />
As part of a project to develop a synthesis interface which nontechnical<br />
musicians should find intuitive, the adjectives musicians use to describe<br />
timbre have been studied in a large-scale listening test covering the<br />
timbre space of Western orchestral instruments. These were refined in<br />
previous work by the authors �A. C. Disley et al. ‘‘Spectral correlations of<br />
timbral adjectives used by musicians,’’ J. Acoust. Soc. Am. 119, 3333,<br />
�<strong>2006</strong>�� to a set of ten words which had good common understanding and<br />
discrimination between the samples �bright, clear, dull, gentle, harsh, nasal,<br />
percussive, ringing, thin, and warm�. To help explore potential relationships<br />
between these adjectives and spectral features, 20 listeners participated<br />
in a further listening experiment, comparing samples in pairs to<br />
produce dissimilarity data. Multidimensional scaling produced dimensions<br />
which were compared with a large number of spectral and time-domain<br />
analyses of the stimuli, suggesting a number of significantly correlated<br />
spectral cues with some of the adjectives. These results are compared with<br />
previous studies by the authors and others, showing both similarities and<br />
differences, suggesting that collective consideration of timbral adjectives<br />
is more likely to result in simultaneously applicable theories of acoustic<br />
correlates than individual consideration of words.<br />
3277 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3277<br />
4p FRI. PM
FRIDAY AFTERNOON, 1 DECEMBER <strong>2006</strong> MAUI ROOM, 1:00 TO 2:00 P.M.<br />
Session 4pNSa<br />
Noise and Architectural Acoustics: Soundscapes and Cultural Perception II<br />
Brigitte Schulte-Fortkamp, Cochair<br />
Technical Univ. Berlin, Inst. of Technical Acoustics, Secr TA 7, Einsteinufer 25, 10587 Berlin, Germany<br />
Bennett M. Brooks, Cochair<br />
Brooks Acoustics Corp., 27 Hartford Turnpike, Vernon, CT 06066<br />
1:00<br />
4pNSa1. Mapping soundscapes in urban quiet areas. Gaetano Licitra<br />
�ARPAT, Tuscany Regional Agency for Environ. Protection, Via N.<br />
Porpora, 22-50144, Firenze, Italy�, Gianluca Memoli �Memolix, Environ.<br />
Consultants, 56127 Pisa, Italy�, Mauro Cerchiai, and Luca Nencini<br />
�ARPAT, 56127 Pisa, Italy�<br />
Innovative action plans in noise-polluted environments require the description<br />
of the existing soundscape in terms of suitable indicators. The<br />
role of these indicators, giving the ‘‘fingerprint’’ of a fixed soundscape,<br />
would be not only to measure the improvement in the sound quality after<br />
the action taken, but also to guide the designer in the process, providing a<br />
reference benchmark. One of the open questions on new indicators is the<br />
way they relate to existing ones and to people’s perception. The present<br />
work will describe a ‘‘Sonic Garden’’ in Florence, using both the ‘‘slope’’<br />
indicator �constructed from the LA eq time history and related in previous<br />
studies to people’s perception� and classical psychoacoustical parameters<br />
�level, spectral structure, and perceived characteristics such as loudness,<br />
sharpness, fluctuation, and roughness�. The latter parameters will be acquired<br />
using a binaural technique.<br />
Contributed Papers<br />
1:30–2:00<br />
Panel Discussion<br />
1:15<br />
4pNSa2. A questionnaire survey of the attitude of Japanese and<br />
foreign residents in Japan to sound masking devices for toilets.<br />
Miwako Ueda and Shin-ichiro Iwamiya �Grad. School of Design, Kyushu<br />
Univ., Iwamiya Lab. 4-9-1, Shiobaru, Minami-ku, Fukuoka 815-8540<br />
Japan, amaria@white.livedoor.com�<br />
Unique sound masking devices for toilets can be used in women’s<br />
rest<strong>room</strong>s in Japan. Such devices function to produce the sound of flushing<br />
water without actual flushing. To mask the sound of bodily functions,<br />
women tended to flush the toilet continuously while using it, thereby wasting<br />
a large amount of water. In the circumstances, sound masking devices<br />
have been introduced to public toilets. We have recently conducted a questionnaire<br />
survey to clarify the attitude of people toward such sound masking<br />
devices for toilets. The results of the survey showed that many Japanese<br />
women know such devices and often use them, that foreign women<br />
currently living in Japan also know that such devices exist, and that some<br />
Japanese men have heard of such devices but never used them. Many<br />
Japanese women are quite embarrassed at the thought that someone else<br />
can hear them while they are on the toilet. Many noted the necessity of<br />
such devices and required a wide range of setting for toilets in public<br />
spaces. However, they are not satisfied with the sound quality of the playback<br />
toilet flush sounds of currently available devices. The above results<br />
suggest that the sound quality of such devices should be improved.<br />
FRIDAY AFTERNOON, 1 DECEMBER <strong>2006</strong> MAUI ROOM, 2:10 TO 4:35 P.M.<br />
Session 4pNSb<br />
Noise, Physical Acoustics, and Structural Acoustics and Vibration: Acoustics of Sports<br />
Joseph Pope, Cochair<br />
Pope Engineering Company, P.O. Box 590236, Newton, MA 02459-0002<br />
Kenji Kurakata, Cochair<br />
AIST, 1-1-1 Higashi, Tsukuba, Ibaraki 305-8566, Japan<br />
Chair’s Introduction—2:10<br />
Invited Papers<br />
2:15<br />
4pNSb1. A review of the vibration and sounds of the crack of the bat and player auditory clues. Robert Collier �Thayer School<br />
of Eng., 8000 Cummings Hall, Hanover, NH 03755�<br />
The purpose of this paper is to review the state-of-the-art in the acoustics of baseball. As is well known, the crack of the bat is an<br />
important phenomenon of solid wood bats and metal bats. Each has a very different sound signature. At the 148th meeting of the ASA<br />
in 2004, the author and coauthors Ken Kaliski and James Sherwood presented the results of laboratory and field tests, which showed<br />
3278 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3278
that the spectral characteristics of radiated sound are dependent on the ball-bat impact location and resultant bat vibration of both solid<br />
wood and tubular metal bats. These results will be reviewed together with those of other investigators in the context of player auditory<br />
clues and the player’s response in game situations.<br />
2:35<br />
4pNSb2. Measurements of the impact sound of golf clubs and risk of hearing impairment. Kenji Kurakata �Natl. Inst. of Adv.<br />
Industrial Sci. and Technol. �AIST�, 1-1-1 Higashi, Tsukuba, Ibaraki, 305-8566 Japan, kurakata-k@aist.go.jp�<br />
The ball-impact sounds of golf clubs with metal heads and of a club with a wood head were measured to investigate their different<br />
acoustic properties. Hitting was executed using either a swing machine or a human player. Results of these analyses showed that the<br />
metal-head clubs generated sounds around 100 dB (L pA,Fmax). This level was 5–15 dB higher than that of the wood-head club. The<br />
sounds of the metal-head clubs had greater power in the high-frequency region of 4 kHz and above compared to the wood-head club,<br />
which particularly increased the overall sound levels. These results suggest that it would be desirable to develop a metal head with<br />
pleasant sound qualities, keeping the sound level lower to minimize hearing damage. Some of these measurement data were published<br />
in Japanese in a previous paper �K. Kurakata, J. INCE/J 26, 60–63�2002��.<br />
2:55<br />
4pNSb3. New underwater sound system for synchronized swimming: The 9th International Swimming Federation<br />
Championships. Takayuki Watanabe, Shinji Kishinaga �YAMAHA Ctr. for Adv. Sound Technologies, 203 Matsunokijima, Iwata,<br />
Shizuoka, 438-0192 Japan�, Tokuzo Fukamachi �YAMAHA Motor Marine Operations, Arai, Hamana, 438-8501 Japan�, and Osamu<br />
Maeda �YAMAHA Motor Adv. Technol. Res. Div., Iwata, 438-8501 Japan�<br />
There have been concerns about the differences between underwater sound fields in a temporary fiberglass-reinforced plastic<br />
�FRP� pool and in a conventional reinforced concrete �RC� pool. A temporary FRP pool was to be used for competitions at the World<br />
Swimming Championships in Fukuoka. We considered three items as key factors for a swimming pool used for synchronized<br />
swimming: �1� the sound source itself �output level, fluctuations in frequency characteristics�; �2� the effect of materials used in pool<br />
construction upon sound source installation conditions; and �3� the effect of the mth mode low-frequency cutoff in ‘‘shallow water.’’<br />
To improve basic problems related to the first factor, we developed a new actuator-driven underwater sound system �YALAS�, which<br />
can eliminate the effect of installation conditions for underwater speakers in the FRP pool. This new underwater system has now seen<br />
practical use in competitions. The report summarizes this new underwater sound system and compares the system with conventional<br />
systems in terms of its acoustic characteristics. The system can offer music with sufficient audibility in water. We gained a good<br />
reputation with competitors because the system showed superior performance to conventional systems in sound volume and quality,<br />
and in uniformity of sound distribution.<br />
3:15<br />
4pNSb4. Acoustics of the Great Ball Court at Chichen Itza, Mexico. David Lubman �14301 Middletown Ln., Westminster, CA<br />
92683�<br />
The ball game has played a central role in Mayan religion and culture for 5000 years. Thousands of ball courts have been<br />
discovered. The Great Ball Court �GBC� at Chichen Itza is a late development and is architecturally unique. Two remarkable<br />
acoustical features were noticed during excavation in the 1920s, but never explained or interpreted. A whispering gallery permits voice<br />
communication between temples located about 460 feet �140 m� apart. A profound flutter echo is heard between the two massive<br />
parallel walls of the playing field, about 270 ft �82 m� long, 28 ft �8.5 m� high, and 119 ft �36 m� apart. Until recently, most<br />
archaeologists dismissed acoustical features at Mayan sites as unintended artifacts. That is now changing. Stimulated by archaeological<br />
acoustic studies and reports since 1999, eminent Mayanists Stephen Houston and Karl Taube have reinterpreted certain Mayan<br />
glyphs as vibrant sounds and ballcourt echoes, and have famously called for a new archaeology of the senses, especially hearing, sight,<br />
and smell �Cambridge Archaeol. J. 10 �2� 261–294 �2000��. By interpreting architectural, psychoacoustic, and cognitive features of the<br />
GBC in the context of ancient Mayan culture, this paper speculates that acoustical effects at the GBC may be original design features.<br />
3:35<br />
4pNSb5. Sleep disturbance caused by shooting sounds. Joos Vos<br />
�TNO Human Factors, P.O. Box 23, 3769 ZG Soesterberg, The<br />
Netherlands, joos.vos@tno.nl�<br />
In the present study relations between the sound level of shooting<br />
sounds and the probability of behaviorally confirmed noise-induced awakening<br />
reactions were determined. The sounds were presented by means of<br />
loudspeakers in the bed<strong>room</strong>s of 30 volunteers. The shooting sounds had<br />
been produced by a small and a medium-large firearm, and the stimuli<br />
consisted of individual bangs or volleys of ten isolated or partly overlapping<br />
impulses. Aircraft sound was included as a reference source. The<br />
sounds were presented during a 6-h period that started 75 min after the<br />
beginning of the sleeping period. The time period between the various<br />
stimuli varied between 12 and 18 min, with a mean of 15 min. To cope<br />
with at least a relevant portion of habituation effects, each subject participated<br />
in 18 nights to be completed within 4 weeks. Preliminary results are<br />
presented both for the awakening reactions described above, and for vari-<br />
Contributed Papers<br />
ous other dependent variables collected with the help of an actimeter or<br />
determined by means of subjective rating scales. �Work supported by the<br />
Dutch Ministry of Defense.�<br />
3:50<br />
4pNSb6. Sound inside a gymnasium. Sergio Beristain �ESIME, IPN,<br />
IMA., P.O. Box 12-1022, Narvarte, 03001, Mexico City, Mexico�<br />
A new gymnasium for a sports club was designed taking acoustic<br />
confort into consideration, in order to accomodate sports practice, sports<br />
events with the public, or musical and drama presentations, taking advatage<br />
of its large capacity for the public and performers. The floor plan<br />
included <strong>room</strong> enough for a basketball court with public space on one<br />
side, where grades for about 200 people will be permanently installed.<br />
Walls were treated in a way that is useful for the sports practice �hard<br />
surfaces�, with hidden absorption material to reduce the usual reverberant<br />
3279 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3279<br />
4p FRI. PM
field inside the court, and to allow for sound events with only the addition<br />
ofamat�to protect the floor woodwork� and extra grades and the sound<br />
reinforcement system.<br />
4:05<br />
4pNSb7. Occupational and recreational noise exposures at stock car<br />
racing circuits. Chucri A. Kardous, Thais Morata, and Luann E. Van<br />
Campen �Natl. Inst. for Occupational Safety Health, 4676 Columbia<br />
Pkwy., Cincinnati, OH 45226, ckardous@cdc.gov�<br />
Noise in stock car racing is accepted as a normal occurrence but the<br />
exposure levels associated with the sport have not been adequately characterized.<br />
Researchers from the National Institute for Occupational Safety<br />
and Health �NIOSH� conducted an exploratory assessment of noise exposures<br />
to drivers, pit crew, team staff, and spectators at three stock car<br />
racing events. Area measurements were made during race preparation,<br />
practice, qualification, and competition. Personal dosimetry measurements<br />
were conducted on drivers, crew members, infield staff, and spectators.<br />
Findings showed time-weighted averages �TWA� that ranged from 94<br />
decibels A-weighted �dBA� for spectators to 114 dBA for car drivers. Peak<br />
sound-pressure levels exceeded the maximum allowable limit of 140 decibels<br />
�dB� during race competitions. Personal exposure measurements exceeded<br />
the NIOSH recommended exposure limit �REL� of 85 dBA as an<br />
8-h TWA in less than a minute for one driver during practice, within 2 min<br />
for pit crew and infield staff, and 7 to 10 min for spectators during the<br />
race. Hearing protection use was variable and intermittent among crew,<br />
staff, and spectators. Among drivers and crew, there was greater concern<br />
for communication performance than for hearing protection.<br />
4:20<br />
4pNSb8. Sports acoustics: Using sound from resonant shells,<br />
vibrating cylinders, strumming shafts, and water impact to evaluate<br />
athletic performance. David G. Browning �Dept. of Phys., Univ. of<br />
Rhode Island, Kingston, RI 02881, decibeldb@aol.com� and Peter M.<br />
Scheifele �Univ. of Connecticut, Storrs, CT 06269�<br />
The sound from equipment used and/or specific acts during athletic<br />
competition, such as hitting a baseball with an aluminum bat, carries beyond<br />
the playing field and can provide a nonobtrusive method to evaluate<br />
athletic performance—such as where on the bat the ball was hit. Standardized<br />
equipment guarantees repeatability, for example, every volleyball<br />
resonates at the same frequency. Each major sport can have unique noise<br />
interference which in some circumstances can be overwhelming, and the<br />
distance from the sound source can vary significantly during a game. Still,<br />
it will be shown that useful performance information can be obtained<br />
under realistic conditions for at least the following sports: volleyball, softball,<br />
baseball, golf, swimming and diving, soccer, and football.<br />
FRIDAY AFTERNOON, 1 DECEMBER <strong>2006</strong> WAIANAE ROOM, 1:30 TO 6:20 P.M.<br />
Session 4pPA<br />
Physical Acoustics and Biomedical UltrasoundÕBioresponse to Vibration: Sound Propagation in<br />
Inhomogeneous Media II<br />
Takahiko Otani, Cochair<br />
Doshisha Univ., Lab. of Ultrasonic Electronics, Kyotonabe-shi, Kyoto 610-0321, Japan<br />
James G. Miller, Cochair<br />
Washington Univ., Dept. of Physics, 1 Brookings Dr., St. Louis, MO 63130<br />
Invited Paper<br />
1:30<br />
4pPA1. Observables and prediction modeling in the presence of ultra-wideband heterogeneity. John J. McCoy �The Catholic<br />
Univ. of America, Washington, DC <strong>2006</strong>4�<br />
Underlying virtually all propagation and scattering models is an intuitive understanding; the acoustic field is observable using a<br />
device of a sufficiently small size to obtain a sufficiently dense set of discrete measurements. This assumes the field variation cuts off<br />
at an inner length scale larger than the device size, assuring that no information is lost to the inherent spatial averaging in any<br />
measurement. This understanding is faulty in the presence of environment heterogeneity observed on an extreme range of length<br />
scales. The reason is that all physical devices have finite accuracy, which limits their ability to capture variation on scales significantly<br />
larger than their size, in the presence of variation on intermediate scales. A more refined understanding of the ability to observe a field<br />
requires multiple devices, an unbounded hierarchy in the limit, to obtain multiple dense sets of discrete ‘‘observables.’’ This, then,<br />
suggests a different class of prediction models for environments with ultra-wideband heterogeneity, expressed in multiple sets of<br />
discrete variables, each set describing field variation in a limited subband. A framework for formulating these prediction models and<br />
their application to a scenario for which environment heterogeneity has no inner scale cutoff is presented.<br />
3280 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3280
1:50<br />
4pPA2. Parameters arising from the Burridge-Keller formulation for<br />
poroelastic media, especially for granular media and marine<br />
sediments. Allan D. Pierce, William M. Carey, and Paul E. Barbone<br />
�Boston Univ., Boston, MA 02215�<br />
It was previously shown �Pierce et al. �<strong>2006</strong>�� that the Burridge-Keller<br />
formulation �J. Acoust. Soc. Am. �1981�� rigorously justifies the lowfrequency<br />
version ��1956a�� of Biot’s equations. Implementation involves<br />
two microscale problems: �1� incompressible viscous flow driven in a<br />
highly irregular space with rigid walls by a uniformly and externally applied<br />
apparent pressure distribution and �2� elastostatic deformation of an<br />
intricate elastic web caused by the joint influence of a distributed constant<br />
body force and by uniform tractions �including an external pressure� on<br />
the web’s exposed surface. Microscale averages produce the Biot ‘‘constants.’’<br />
Theoretical devices of applied mechanics and mathematics yield<br />
estimates of these and related parameters. In particular, it is shown that<br />
Wood’s equation is a reasonable first approximation for the sound speed in<br />
sediments in the low-frequency limit. The formulation also yields an estimation<br />
for the sound speed in the high-frequency limit, when the viscous<br />
boundary layers become thin. The well-known result that the attenuation<br />
varies as f 1/2 in the high-frequency limit also results without the necessity<br />
of Biot’s heuristic patching theory. Various heuristic approximations due<br />
to Gassmann, to Geertsma and Smit, and to Stoll and Bryant are analytically<br />
and numerically assessed.<br />
2:05<br />
4pPA3. Sound propagation in the mixtures of liquid and solid<br />
aggregate; similarities at micro- and nanoscales.. Hasson M. Tavossi<br />
�Dept. of Physical and Environ. Sci., Mesa State College, 1100 North Ave.,<br />
Grand Junction, CO 81504�<br />
Sound propagation phenomena in certain liquid and solid aggregate<br />
mixtures, at micrometer scales, in some cases resemble the wave propagation<br />
behaviors of materials observed at nanometer and atomic scales.<br />
For example, it can be shown that the sound wave dispersion, attenuation,<br />
and cutoff-frequency effects depend on the same structural parameters as<br />
those observed at nano or atomic levels and are similar at both scales.<br />
Therefore, to investigate theoretical models of wave and matter interactions<br />
it is more convenient to use, as experimental tools, the readily analyzable<br />
models of wave propagation, in mixtures of solid and liquid, constructed<br />
at micrometer scales. Theoretical findings on sound propagation<br />
in the mixtures of liquid and solid particles at micrometer scales will be<br />
discussed. These results show the resemblance to the behavior of acoustic<br />
phonons, the lattice thermal vibrations of crystalline structures, at radically<br />
different scales. Experimental data on wave dispersion, attenuation, bandpass,<br />
and cutoff frequency effects, measured for sound propagation, in<br />
inhomogeneous materials consisting of mixtures of solid and liquid will be<br />
presented, showing the similarities of wave propagation behaviors at<br />
micro- and nanoscales.<br />
2:20<br />
4pPA4. Nonlinear surface waves in soil. Evgenia A. Zabolotskaya,<br />
Yurii A. Ilinskii, and Mark F. Hamilton �Appl. Res. Labs., Univ. of Texas,<br />
P.O. Box 8029, Austin, TX 78713-8029�<br />
Nonlinear effects in surface waves propagating in soil are investigated<br />
theoretically. Analytic solutions are derived for the second harmonics and<br />
difference frequency waves generated by a bifrequency primary wave<br />
propagating at moderate amplitude. The soil is modeled as an isotropic<br />
solid. As such, its elastic properties are described by five elastic constants,<br />
two at second order in the strain energy density �the shear and bulk<br />
moduli� and three at third order. Nonlinear propagation of the surface<br />
waves is based on a theory developed previously �Zabolotskaya, J. Acoust.<br />
Soc. Am. 91, 2569–2575 �1992��. Elements of the nonlinearity matrix<br />
associated with the interacting spectral components are expressed in terms<br />
of the five elastic constants. It was found convenient to express the nonlinearity<br />
matrix for soil as a function of a nonlinearity parameter corre-<br />
Contributed Papers<br />
sponding to B/A for liquids, particularly for saturated soils exhibiting<br />
liquidlike properties. This nonlinearity parameter can vary by several orders<br />
of magnitude. For soils with shear wave speeds less than 20% of the<br />
compressional wave speeds, the nonlinearity of surfaces waves is found to<br />
be independent of the third-order elastic constants and dependent only on<br />
the shear modulus. �Work supported by ONR.�<br />
2:35<br />
4pPA5. The measurement of the hysteretic nonlinearity parameter of<br />
a field soil by the phase shift method: A long-term survey. Zhiqu Lu<br />
�Natl. Ctr. for Physical Acoust., The Univ. of Mississippi, Univ., MS<br />
38677�<br />
Soil properties significantly affect the performance of the acoustic<br />
landmine detection. The climate and seasonal changes cause the variations<br />
of soil properties and smear landmine signature over time. On the other<br />
hand, soil is a complicated granular material that exhibits strong nonlinear<br />
acoustic behaviors. To understand the weather and seasonal effects on<br />
nonlinear acoustic behaviors of soils, a phase shift method is used to<br />
measure the hysteretic nonlinearity parameter of a field soil. The technique<br />
is based on measuring the variation of phase difference between two transducers,<br />
i.e., the phase shift, induced by changing sound level. The hysteretic<br />
nonlinear parameter can be extracted from the measured phase shift as<br />
a function of sound level or dynamic strain. In a long-term survey, the<br />
nonlinearity parameter, sound speed, and environmental conditions such<br />
as temperature, moisture, soil water potential, and rainfall precipitation are<br />
measured. It is found that the nonlinearity parameter is much more sensitive<br />
than sound speed to the climate change. Soil water potential is the<br />
predominant factor that affects the nonlinearity parameter and sound speed<br />
of the shallow field soil.<br />
2:50<br />
4pPA6. Nonlinear acoustic landmine detection: Comparison of soil<br />
nonlinearity with soil-interface nonlinearity. Murray S. Korman,<br />
Kathleen E. Pauls, Sean A. Genis �Dept. of Phys., U. S. Naval Acad.,<br />
Annapolis, MD 21402�, and James M. Sabatier �Univ. of Mississippi,<br />
Univ., MS 38677�<br />
To model the soil-top plate interface in nonlinear acoustic landmine<br />
detection, the soil-plate oscillator was developed �J. Acoust. Soc. Am. 116,<br />
3354–3369 �2004��. A Lexan plate �2.39 mm thick, 18.5 cm diameter� is<br />
clamped at an inside diameter of 11.8 cm between two metal flanges. Dry<br />
sifted masonry sand �2-cm layer� is placed over the plate. Turning curves<br />
experiments are performed by driving a loudspeaker �located over the<br />
sand� by a swept sinusoid. The acceleration versus frequency is measured<br />
near resonance on a swept spectrum analyzer using an accelerometer centered<br />
on the surface. The corresponding backbone curve exhibits a linear<br />
decrease in resonant frequency f versus increasing acceleration, where a<br />
��a o( f � f o)/f o . Define a nonlinear parameter ��1/a o . When the elastic<br />
plate is replaced by a ‘‘rigid’’ plate, � decreased from 0.128 to 0.070<br />
(s 2 /m�, while f o increased from 191 to 466 Hz. When a cylindrical drumlike<br />
mine simulant �rigid walls, thin acrylic top-plate� wasburied2cm<br />
deep in a concrete sand box, ‘‘on the mine’’ results yielded ��0.30 (s 2 /m�<br />
with f o�147 Hz, while ‘‘off the mine,’’ ��0.03 (s 2 /m� at f o�147 Hz.<br />
�Work supported by ONR.�<br />
3:05<br />
4pPA7. Causality conditions and signal propagation in bubbly water.<br />
Gregory J. Orris, Dalcio K. Dacol, and Michael Nicholas �Naval Res.<br />
Lab., 4555 Overlook Ave. SW, Washington, DC 20375�<br />
Acoustic propagation through subsurface bubble clouds in the ocean<br />
can exhibit signal travel times with enormous variations depending on the<br />
acoustic signal frequency, bubble size distribution, and void fraction. Recent<br />
theories have predicted large variations in phase speeds and attenuation<br />
that have been largely validated for frequencies well below and well<br />
above bubble resonance. However, great care must be exercised when<br />
3281 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3281<br />
4p FRI. PM
theoretically treating signal propagation at frequencies near resonance,<br />
termed the ‘‘Anomalous Absorption Regime’’ nearly 100 years ago in the<br />
pioneering work of Sommerfeld �A. Sommerfeld, Physik. Z. 8, 841<br />
�1975�� while investigating aspects of electromagnetic causality. We will<br />
discuss similarities between acoustic propagation in bubbly media and<br />
electromagnetic propagation in the presence of a conducting medium. We<br />
show that the signal travel time is dependent on the behavior of the dispersion<br />
formula in the complex frequency plane and place limits on the<br />
range of validity of these formulas, leading naturally to the necessary<br />
modifications to the current dispersion formulas to bring them into compliance<br />
with causality. Finally, we present theoretical results for the velocity<br />
of signals for a representative environment of experimental work carried<br />
out at the Naval Research Laboratory. �Work supported by the ONR.�<br />
3:20<br />
4pPA8. Measurements of the attenuation and sound speed in bubbly<br />
salt water. Gregory J. Orris, Dalcio K. Dacol, and Michael Nicholas<br />
�Naval Res. Lab., 4555 Overlook Ave. SW, Washington, DC 20375�<br />
Bubble clouds were injected from below the surface of a 144-cubicmeter<br />
water tank, wherein hydrophones were placed at varying distances<br />
from an acoustic source. Measurements were made over a wide range of<br />
frequencies to verify and validate the theoretical predictions of the relevant<br />
dispersion formula. This work was undertaken under a variety of<br />
conditions by varying the relevant environmental parameters: void fraction,<br />
temperature, and salinity. Void fractions were varied from roughly<br />
0.02% to 0.1%. Temperatures ranged from 9 °C to 18 °C, and the salinity<br />
was varied from zero to approximately 10% of typical oceanic values.<br />
Particular attention was paid to tracking the phase of the transmitted signal<br />
as the frequency progressed toward resonance starting from 100 kHz. This<br />
yielded phase-speed measurements in an essentially free-field environment<br />
using a modified version of phase spectral analysis. Time-of-flight measurements<br />
gave signal velocities, while the received energy yielded the<br />
attenuation. Results are compared to theoretical calculations, leading to<br />
the conclusion that current theoretical dispersion formula requires modification.<br />
�This work supported by the ONR.�<br />
3:35<br />
4pPA9. Efficient computation of 3-D acoustical scattering from<br />
multiple arbitrarily shaped objects using the boundary element<br />
methodÕfast multipole method „BEMÕFMM…. Nail A. Gumerov and<br />
Ramani Duraiswami �Perceptual Interfaces and Reality Lab., Inst. for<br />
Adv. Comput. Studies, Univ. of Maryland, College Park, MD 20742�<br />
Many applications require computation of acoustic fields in systems<br />
consisting of a large number of scatterers, which may have complex shape.<br />
Despite the boundary element method being a well-known technique for<br />
solution of the boundary value problems for the Helmholtz equation, its<br />
capabilities are usually limited by the memory and speed of computers,<br />
and conventional methods can be applicable to relatively small �up to<br />
order of 10 000 boundary elements� problems. We developed and implemented<br />
an efficient computational technique, based on an iterative solver<br />
employing generalized minimal residual method in combination with<br />
matrix-vector multiplication speeded up with the fast multipole method.<br />
We demonstrate that this technique has O(N) memory and computational<br />
complexity and enables solution of problems with thousands of scatterers<br />
�millions of boundary elements� on a desktop PC. The test problems<br />
solved are of moderate frequency �up to kD�150, where k is the wavenumber<br />
and D is the size of the computational domain�. Solution of large<br />
scale scattering problems was tested by comparison with the FMM-based<br />
T-matrix method applicable for simple shape objects reported earlier<br />
�Gumerov and Duraiswami, J. Acoust. Soc. Am., 117�4�, 1744–1761<br />
�2005��, visualization, and physical interpretation of the results.<br />
3:50<br />
4pPA10. Fast acoustic integral-equation solver for complex<br />
inhomogeneous media. Elizabeth Bleszynski, Marek Bleszynski, and<br />
Thomas Jaroszewicz �Monopole Res., 739 Calle Sequoia, Thousand<br />
Oaks, CA 91360, elizabeth@monopoleresearch.com�<br />
We describe elements and representative applications of an integralequation<br />
solver for large-scale computations in acoustic wave propagation<br />
problems. In the solver construction we used elements of our previously<br />
developed fast integral-equation solver for Maxwell’s equations. In comparison<br />
with the conventional integral equation approach �method of moments�,<br />
our solver achieves significant reduction of execution time and<br />
memory through the FFT-based matrix compression. One particular aspect<br />
of the solver we discuss, pertinent to its high efficiency and accuracy, is an<br />
efficient treatment of problems associated with subwavelength discretization.<br />
We illustrate the approach and its application on the example of a<br />
numerical simulation of acoustic wave propagation through the human<br />
head. �Work was supported by a grant from AFOSR.�<br />
4:05<br />
4pPA11. Models for acoustic scattering in high contrast media. Max<br />
Denis, Charles Thompson, and Kavitha Chandra �Univ. Massachusetts<br />
Lowell, One University Ave., Lowell, MA 01854�<br />
In this work a numerical method for evaluating backscatter from a<br />
three-dimensional medium having high acoustic contrast is presented. The<br />
solution is sought in terms of a perturbation expansion in the contrast<br />
amplitude. It is shown that limitations of the regular perturbation expansion<br />
can be overcome by recasting the perturbation sequence as a rational<br />
fraction using Padé approximants. The resulting solution allows for an<br />
accurate representation of the pressure and allows for the poles in the<br />
frequency response to be modeled. The determination of the pulse-echo<br />
response for a high-contrast medium is discussed and presented.<br />
4:20<br />
4pPA12. Multiple scattering and visco-thermal effects. Aroune<br />
Duclos, Denis Lafarge, Vincent Pagneux �Laboratoire d’Acoustique de<br />
l’Universite du Maine, Ave. Olivier Messiaen, 72085 Le Mans, France�,<br />
and Andrea Cortis �Lawrence Berkeley Natl. Lab., Berkeley, CA 94720�<br />
For modeling sound propagation in a rigid-framed fluid-saturated porous<br />
material it is customary to use frequency-dependent density and compressibility<br />
functions. These functions, which describe ‘‘temporal’’ dispersion<br />
effects due to inertial/viscous and thermal effects, can be computed<br />
by FEM in simple geometries and give complete information about the<br />
long-wavelength properties of the medium. When the wavelength is reduced,<br />
new effects due to scattering must be considered. To study this, we<br />
consider solving the sound propagation problem in a 2-D ‘‘phononic crystal’’<br />
made of an infinite square lattice of solid cylinders embedded in a<br />
fluid. An exact multiple-scattering solution is first developed for an ideal<br />
saturating fluid and then generalized to the case of visco-thermal fluid, by<br />
using the concept of visco-thermal admittances. The condition to use this<br />
concept is that the viscous and thermal penetration depths are small compared<br />
to the cylinder radius. We validate our results in the longwavelength<br />
regime by direct comparisons with FEM data �A. Cortis, ‘‘Dynamic<br />
parameters of porous media,’’ Ph.D. dissertation �Delft U.P., Delft,<br />
�2002��. When frequency increases, differences appear between the longwavelength<br />
solution and the exact multiple-scattering solution, which<br />
could be interpreted in terms of ‘‘spatial’’ dispersion effects.<br />
4:35<br />
4pPA13. Effective parameters of periodic and random distributions of<br />
rigid cylinders in air. Daniel Torrent and José Sánchez-Dehesa �Wave<br />
Phenomena Group, Nanophotonics Technol. Ctr., Polytechnic Univ. of<br />
Valencia, C/Camino de vera s/n., E-46022 Valencia, Spain�<br />
The scattering of sound by finite-size clusters consisting of twodimensional<br />
distributions �periodic and random� of rigid cylinders in air is<br />
theoretically studied in the low-frequency limit �homogenization�. Analyti-<br />
3282 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3282
cal expressions for the effective density and sound speed obtained in the<br />
framework of multiple scattering will be reported. For the case of circularshaped<br />
cluster, we have theoretically analyzed the homogenization as a<br />
function of the filling fraction, the type of arrangement of the cylinders in<br />
the cluster �hexagonal and square lattice�, and the number of cylinders in<br />
the cluster. When the number of cylinders in the cluster is small we found<br />
that for certain ‘‘magic numbers’’ their effective parameters �sound speed<br />
and density� are the same as those of the corresponding infinite array.<br />
�Work supported by MEC of Spain.�<br />
4:50<br />
4pPA14. The application of k-space acoustic propagation models to<br />
biomedical photoacoustics. Benjamin T. Cox �Dept. of Med. Phys. and<br />
Bioengineering, Univ. College London, Gower St., London, WC1E 6BT,<br />
UK, bencox@mpb.ucl.ac.uk�, Simon. R. Arridge, and Paul C. Beard<br />
�Univ. College London, Gower St., London, WC1E 6BT, UK�<br />
k�space models for broadband acoustic pulse propagation differ from<br />
pseudo�spectral time domain �PSTD� models in their treatment of the<br />
time step. By replacing a finite�difference scheme with a propagator,<br />
exact for homogeneous media, larger time steps can be taken without loss<br />
of accuracy or stability and without introducing dispersion. Three<br />
k�space models for modeling photoacoustically generated �PA� pulses are<br />
described here. A very simple, exact, model of PA propagation in a homogeneous<br />
fluid is used to introduce the k�space propagator, and two models<br />
of propagation in heterogeneous media, originally designed for modeling<br />
scattering in soft tissue, are adapted for use in photoacoustics �Mast<br />
et al., IEEE Trans. UFFC 48, 341�354 �2001�; Tabei et al., J. Acoust.<br />
Soc. Am. 111,53�63 �2002��. Our motivation for describing these models<br />
comes from biomedical PA imaging, in which one of the current limitations<br />
is the assumption that soft tissue has a uniform sound speed. Efficient,<br />
accurate, and simple�to�encode forward models such as these are<br />
very useful for studying the effects of the heterogeneities encountered in<br />
practice. They may also be useful in designing PA imaging schemes that<br />
can account for acoustic heterogeneities. �This work was funded by the<br />
EPSRC, UK�<br />
5:05<br />
4pPA15. Emergence of the acoustic Green’s function from thermal<br />
noise. Oleg A. Godin �CIRES, Univ. of Colorado and NOAA, Earth<br />
System Res. Lab., 325 Broadway, Boulder, CO 80305,<br />
oleg.godin@noaa.gov�<br />
Recently proposed applications of noise cross-correlation measurements<br />
to passive remote sensing range from ultrasonics and acoustic<br />
oceanography to helioseismology and geophysics, at wave frequencies<br />
that differ by more than ten orders of magnitude. At the heart of these<br />
applications is the possibility to retrieve an estimate of a deterministic<br />
Green’s function from long-range correlations of diffuse noise fields. Apparently,<br />
S. M. Rytov �A Theory of Electrical Fluctuations and Thermal<br />
Radiation �USSR Academy of Sciences, Moscow, 1953�� was the first to<br />
establish theoretically a simple relation between the Green’s function and<br />
the two-point correlation function of fluctuations of wave fields generated<br />
by random sources. He used reciprocity considerations to analyze fluctuations<br />
of electromagnetic fields. In this paper, an acoustic counterpart of<br />
Rytov’s approach is applied to derive exact and asymptotic relations between<br />
respective acoustic Green’s functions and cross-correlation of thermal<br />
noise in inhomogeneous fluid, solid, and fluid-solid media. Parameters<br />
of the media are assumed to be time independent, but can be arbitrary<br />
functions of spatial coordinates. Theoretical results obtained are compared<br />
to those previously reported in the literature.<br />
5:20<br />
4pPA16. Simulation of elastic wave scattering in living tissue at the<br />
cellular level. Timothy E. Doyle and Keith H. Warnick �Dept. of Phys.,<br />
Utah State Univ., 4415 Old Main Hill, Logan, UT 84322-4415,<br />
timdoyle@cc.usu.edu�<br />
Elastic wave scattering in biological tissue has been simulated at the<br />
cellular level by incorporating a first-order approximation of the cell structure<br />
and multiple scattering between cells. The cells were modeled with a<br />
concentric spherical shell-core structure embedded in a medium, with the<br />
core, shell, and medium representing the cell nucleus, the cell cytoplasm,<br />
and the extracellular matrix, respectively. Using vector multipole expansions<br />
and boundary conditions, scattering solutions were derived for a<br />
single cell with either solid or fluid properties for each of the cell components.<br />
Multiple scattering between cells was simulated using addition<br />
theorems to translate the multipole fields from cell to cell and using an<br />
iterative process to refine the scattering solutions. Backscattering simulations<br />
of single cells demonstrated that changes in the nuclear diameter had<br />
the greatest effect on the frequency spectra as compared to changes in cell<br />
size, density, and shear modulus. Wave field images and spectra from<br />
clusters of up to several hundred cells were also simulated, and they exhibited<br />
phenomena such as wave field enhancement at the cell membrane<br />
and nuclear envelope due to the scattering processes. Relevant applications<br />
for these models include ultrasonic tissue characterization and<br />
ultrasound-mediated gene transfection and drug delivery.<br />
5:35<br />
4pPA17. Acoustic analog of electronic Bloch oscillations and Zener<br />
tunneling. José Sánchez-Dehesa, Helios Sanchis-Alepuz, Yu. A.<br />
Kosevich, and Daniel Torrent �Wave Phenomena Group, Polytechnic<br />
Univ. of Valencia, C/Camino de Vera s.n., E-46022 Valencia, Spain�<br />
The observation of Bloch oscillations in sound propagation through a<br />
multilayer of two different fluidlike components is predicted. In order to<br />
obtain the equivalent to the acoustic analog of a Wannier-Stark ladder �E.<br />
E. Mendez et al., Phys. Rev. Lett. 60, 2426–2429 �1988��, a set of cavities<br />
with increasing thickness is employed. Bloch oscillations were theoretically<br />
predicted as time-resolved oscillations in transmission in direct analogy<br />
to electronic Bloch oscillations in semiconductor superlattices �J.<br />
Feldmann et al., Phys. Rev. B 46, R7252–R7255 �1992��. Finally, an experimental<br />
setup is proposed to observe the phenomenon by using arrays<br />
of cylindrical rods in air, which acoustically behaves as a fluidlike system<br />
with effective sound velocity and density �D. Torrent et al., Phys. Rev.<br />
Lett. 96, 204302 �<strong>2006</strong>��. For the proposed system, Bloch oscillations and<br />
Zener tunneling are confirmed by using multiple scattering simulations.<br />
�Work supported by MEC of Spain.�<br />
5:50<br />
4pPA18. Comparison of time reversal acoustic and prefiltering<br />
methods of focusing of tone burst signals. Bok Kyoung Choi �Korea<br />
Ocean Res. and Development Inst., Sangrok-gu, 426-744, Korea�,<br />
Alexander Sutin, and Armen Sarvazyan �Artann Labs., Inc., West<br />
Trenton, NJ 08618�<br />
The concept of time reversal acoustics �TRA� provides an elegant<br />
possibility of both temporal and spatial concentration of acoustic energy in<br />
highly inhomogeneous media. TRA-based focusing is typically used for<br />
generation of short acoustic pulses, however, in some medical and industrial<br />
applications, longer pulses are required. TRA focusing of longer signals<br />
leads to an increase of side lobes in temporal and spatial domains.<br />
Another method for focusing, known as prefiltering, is based on measurements<br />
of the impulse response, which relates the signal at the TRA transmitter<br />
to that at the focusing point. After evaluating the impulse response,<br />
the excitation signal may be calculated to generate the desired waveform<br />
in the focus point. This method allows signal generation with any desired<br />
form including long tone-burst signals. Experiments on comparison TRA<br />
and prefiltering methods of ultrasound focusing were conducted in the<br />
frequency band of 200–1000 kHz. In the experiments, focused acoustic<br />
pulses with various forms and duration were generated: triangular, rectan-<br />
3283 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3283<br />
4p FRI. PM
gular, and amplitude-modulated tone burst signals. The prefiltering modes<br />
provide better temporal compression of the focused signal, and the signal<br />
energy outside the main pulse in the prefiltering mode was shown to be<br />
much lower than that in standard TRA focusing.<br />
6:05<br />
4pPA19. Modeling quasi-one-dimensional sound propagation in ducts<br />
having two propagation media using a cross-sectional averaging<br />
theory. Donald Bliss and Lisa Burton �Dept. of Mech. Eng. and Mater.<br />
Sci., Duke Univ., Durham, NC 27708�<br />
Sound propagation of quasi-one-dimensional waves through a uniform<br />
duct partially filled with porous material has been studied theoretically and<br />
experimentally. The porous material makes the effective propagation wave<br />
number in the duct complex. A fairly simple theory based on cross-<br />
sectional averaging is derived and tested and found to work extremely<br />
well up to fairly high frequency. Interestingly, the basic theory depends<br />
only on the ratio of cross-sectional areas and the properties of the individual<br />
propagation media, but not on the specific configuration of material<br />
in a cross section. A higher order correction is developed to achieve excellent<br />
accuracy to very high frequency. This correction includes a coefficient<br />
that does depend on the specific cross-sectional configuration. Results<br />
are compared to exact solutions for layered and annular<br />
configurations, and also to experimental measurements with open cell<br />
foam as the porous material. An interesting application is to use measured<br />
wave numbers to predict the complex effective density and sound speed of<br />
porous media samples partially filling the duct. Other applications include<br />
fairly simple improved predictions of the behavior of sound in ducts lined<br />
with, or partially filled with, bulk reacting absorbing material.<br />
FRIDAY AFTERNOON, 1 DECEMBER <strong>2006</strong> WAIALUA ROOM, 1:30 TO 4:20 P.M.<br />
Session 4pPP<br />
Psychological and Physiological Acoustics: Auditory Physiology<br />
G. Christopher Stecker, Cochair<br />
Univ. of Washington, Dept. of Speech and Hearing Science, 1417 NE 42nd St., Seattle, WA 98105<br />
Shigeto Furukawa, Cochair<br />
NTT Communication Science Labs., Human and Information Science Lab., 3-1 Morinosato-wakamiya, Atsugi-shi,<br />
Kanagawa-ken 243-0198, Japan<br />
1:35<br />
4pPP1. Transmission of bone-conducted sound measured acoustically<br />
and psycho-acoustically. Sabine Reinfeldt �Signals & Systems,<br />
Chalmers Univ. of Technol., SE-412 96 Goteborg, Sweden�, Stefan<br />
Stenfelt �Linkoping Univ., SE-581 83 Linkoping, Sweden�, and Bo<br />
Hakansson �Chalmers Univ. of Technol., SE-412 96 Goteborg, Sweden�<br />
The transcranial transmission is important in the bone-conducted �BC�<br />
audiometry where the BC hearing thresholds depend on the stimulation<br />
position. It is also important for fitting of BC hearing aids; the transcranial<br />
transmission determines the amount of the sound that reaches the contralateral<br />
cochlea. Previous reported transcranial transmission results seem<br />
to depend on the method used. Here, a comparison between the transcranial<br />
transmission measured with BC hearing thresholds and ECSP is performed<br />
for both open and occluded ear canal. A BC transducer provided<br />
stimulation at both mastoids and the forehead. The ECSP was measured<br />
with a probe microphone and the BC hearing thresholds were obtained<br />
while masking the nontest ear. The transcranial transmission was determined<br />
as the BC hearing threshold or the ECSP for contralateral stimulation<br />
relative ipsilateral stimulation. The transmission from the forehead<br />
was calculated in a similar way. The transcranial transmission was similar<br />
for BC hearing thresholds and ECSP above 800 Hz; this indicates that the<br />
ECSP can be used as an estimator of the relative hearing perception by<br />
BC. The transcranial transmission results are also similar to vibration measurements<br />
of the cochleae made in earlier studies. Hence, vibration measurements<br />
of the cochleae can also estimate relative BC hearing.<br />
Chair’s Introduction—1:30<br />
Contributed Papers<br />
1:50<br />
4pPP2. Customization of head-related transfer functions using<br />
principal components analysis in the time domain. Ki H. Shin and<br />
Youngjin Park �Dept. of Mech. Eng., Korea Adv. Inst. of Sci. and<br />
Technol. �KAIST�, Sci.e Town, Daejeon, 305-701, Republic of Korea�<br />
Pinna responses were separated from HRIRs �head-related impulse<br />
responses� of 45 subjects in the CIPIC HRTF �head-related transfer function�<br />
database and modeled as linear combinations of five basic temporal<br />
shapes �basis functions� by PCA �principal components analysis� accounting<br />
for more than 90% of the variance in the original pinna responses per<br />
each selected elevation angle in the median plane. By adjusting the weight<br />
of each basis function computed for a specific height to replace the pinna<br />
response in the KEMAR HRIR at the same height with the resulting pinna<br />
response and listening to the filtered stimuli over headphones, four subjects<br />
were able to create a set of median HRIRs that outperformed the<br />
KEMAR HRIRs in producing elevation effects in the median plane. Since<br />
the monoaural spectral features due to the pinna are strongly dependent on<br />
elevation instead of azimuth, similar elevation effects could also be generated<br />
at different azimuthal positions simply by inserting the customized<br />
median pinna responses into the KEMAR HRIRs at other azimuths and<br />
varying the ITD �interaural time difference� according to the direction as<br />
well as the size of the subject’s own head.<br />
3284 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3284
2:05<br />
4pPP3. An electrophysiological measure of basilar membrane<br />
nonlinearity in humans. Christopher J. Plack �Dept. of Psych.,<br />
Lancaster Univ., Lancaster, LA1 4YF, England� and Ananthanarayan<br />
Krishnan �Purdue Univ., West Lafayette, IN 47906�<br />
A behavioral measure of the basilar membrane response can be obtained<br />
by comparing the growth in forward masking for maskers at, and<br />
well below, the signal frequency. Since the off-frequency masker is assumed<br />
to be processed linearly at the signal place, the difference in masking<br />
growth with level is thought to reflect the compressive response to the<br />
on-frequency masker. The present experiment used an electrophysiological<br />
analog of this technique, based on measurements of the latency of wave V<br />
of the auditory brainstem response elicited by a 4-kHz, 4-ms pure tone,<br />
presented at 65 dB SPL. Responses were obtained in quiet and in the<br />
presence of either an on-frequency or an off-frequency �1.8 kHz� puretone<br />
forward masker. Wave V latency increased with masker level, although<br />
the increase was greater for the off-frequency masker than for the<br />
on-frequency masker, consistent with a more compressive response to the<br />
latter. Response functions generated from the data showed the characteristic<br />
shape, with a nearly linear response at lower levels, and 5:1 compression<br />
at higher levels. However, the breakpoint between the linear region<br />
and the compressive region was at about 60 dB SPL, higher than expected<br />
on the basis of previous physiological and psychophysical measures.<br />
2:20<br />
4pPP4. Possible involvement of the spiral limbus in chinchilla<br />
cochlear mechanics. William S. Rhode �Dept. of Physiol., 1300<br />
University Ave., Madison, WI 53706, rhode@physiology.wisc.edu�<br />
Differences between cochlear mechanical tuning curves and those of<br />
auditory nerve fibers �ANFs� exist. In particular, mechanical transfer functions<br />
exhibit a high-frequency plateau; ANFs frequency threshold curves<br />
�FTCs� do not. ANF-FTCs may have a low-frequency slope due to a<br />
velocity forcing function operating on inner hair cells at low frequencies.<br />
Neither basilar membrane velocity nor displacement adequately explain<br />
the entire ANF tuning curve. A displacement sensitive interferometer was<br />
used to study basilar membrane and spiral limbus mechanics in the 6-kHz<br />
region of the chinchilla cochlea. The spiral limbus vibrates at the same<br />
phase as the basilar membrane nearly up to the location’s characteristic<br />
frequency. In the plateau region, the limbus appears to vibrate 0 to 20 dB<br />
less than the basilar membrane. The basilar membrane/limbus amplitude<br />
transfer function has a low-frequency slope of �3 dB/oct at low frequencies<br />
and is �10 dB lower than the basilar membrane amplitude at 1 kHz.<br />
It appears that spiral limbus vibration may contribute to the excitation of<br />
the cilia of the inner hair cells. �Work supported by NIDCD grant R01<br />
DC001910.�<br />
2:35<br />
4pPP5. The effects of antioxidants on cochlear mechanics. Barbara<br />
Acker-Mills, Melinda Hill, and Angeline Ebuen �U.S. Army Aeromedical<br />
Res. Lab., P.O. Box 620577, Fort Rucker, AL 36362-0577,<br />
barbara.acker@us.army.mil�<br />
Several studies have evaluated the effect of N-acetylcysteine �NAC�<br />
on temporary threshold shifts �TTSs� in humans. Work at USAARL found<br />
that NAC did not reduce TTSs compared to a placebo, but suppressed<br />
otoacoustic emissions �OAEs� more than a placebo, indicating that NAC<br />
may reduce outer hair cell activity. Kramer et al. �JAAA, 17�4�, �<strong>2006</strong>��<br />
found similar results, where NAC did not affect thresholds in people who<br />
had TTSs from exposure to loud music. However, OAEs also did not<br />
differ between NAC and placebo. Toppilla et al. �XXII Barany Society<br />
Meeting �2002�� measured thresholds and balance in people exposed to<br />
loud music and found that while NAC did not affect TTS, it reduced<br />
noise-induced balance impairment. The current study administered NAC<br />
and measured cochlear microphonics, compound action potentials, and<br />
OAEs to evaluate cochlear function. The vestibular myogenic potential<br />
was measured to assess the effect of NAC on the saccule. The results<br />
provide a comprehensive analysis of the effect of NAC on the auditory<br />
system and one component of the vestibular system. �Work supported by<br />
the U.S. Army ILIR program. The work is that of the authors and is not<br />
necessarily endorsed by the U.S. Army or the Department of Defense.�<br />
2:50–3:05 Break<br />
3:05<br />
4pPP6. Time characteristics of distortion product otoacoustic<br />
emissions recovery function after moderate sound exposure. Miguel<br />
Angel Aranda de Toro, Rodrigo Ordoñez, and Dorte Hammersho”i �Dept.<br />
of Acoust., Aalborg Univ., Fredrik Bajers Vej 7 B5, DK-9220, Aalborg,<br />
Denmark, maat@acoustics.aau.dk�<br />
Exposure to sound of moderate level temporarily attenuates the amplitude<br />
of distortion product otoacoustic emissions �DPOAEs�. These<br />
changes are similar to the changes observed in absolute hearing thresholds<br />
after similar sound exposures. To be able to assess changes over time<br />
across a broad frequency range, a detailed model of the recovery time<br />
characteristics is necessary. In the present study, the methodological aspects<br />
needed in order to monitor changes in DPOAEs from human subjects<br />
measured with high time resolution are presented. The issues treated<br />
are �1� time resolution of the measurements, �2� number of frequency<br />
points required, and �3� effects in fine structures, are they affected with the<br />
exposure? �Work supported by the Danish Research Council for Technology<br />
and Production.�<br />
3:20<br />
4pPP7. Probability characteristics of neural coincidence detectors in<br />
the brainstem. Ram Krips and Miriam Furst �Dept. of Elec. Eng.<br />
Systems, Faculty of Eng., Tel Aviv Univ., Tel Aviv 69978, Israel,<br />
mira@eng.tau.ac.il�<br />
Auditory neural activity in the periphery is usually described as nonhomogeneous<br />
Poisson process �NHPP�, characterized as either EE or EI,<br />
which operates as a coincidence detector. In order to apply a general<br />
probabilistic analysis of those brainstem nuclei activity, the stochastic<br />
properties of the axons that exit the EE and EI nuclei is essential. An<br />
analytical analysis of the probability characteristics of the output of an EE<br />
nucleus �EEout� will be presented. Assuming that an EE nucleus receives<br />
inputs from two neurons, each behaves as an NHPP with instantaneous<br />
rates �1 and �2, and an output spike is generated if both spike at a coincidence<br />
window (�) which is relatively small �this matches biological<br />
findings�. Then EEout is also a NHPP with instantaneous rate r(t)<br />
t<br />
t<br />
�� 1(t)� t���<br />
2(�) d��� 2(t)� t���<br />
1(�) d�. A similar derivation was applied<br />
for an EI nucleus. We found also that the output activity is NHPP for<br />
a relatively small coincidence window (�). The obtained IR is r(t)<br />
t<br />
�� e(t)�1�� t���<br />
i(�) d��, where �E and �I are the excitatory and inhibitory<br />
input IRs. On the other hand, for larger �, the output activity is not a<br />
Poisson process. Those derivations enable theoretical analysis and performance<br />
evaluation of large neural networks, which perform binaural tasks<br />
as ITD and ILD.<br />
3:35<br />
4pPP8. Musical expertise and concurrent sound segregation.<br />
Benjamin Rich Zendel and Claude Alain �Rotman Res. Inst., Baycrest<br />
Ctr. & Dept. of Psych., Univ. of Toronto, Toronto, ON M6A 2E1, Canada�<br />
There is growing evidence suggesting that musical training can improve<br />
performance in various auditory perceptual tasks. These improvements<br />
can be paralleled by changes in scalp recorded auditory evoked<br />
potentials �AEPs�. The present study examined whether musical training<br />
modulates the ability to segregate concurrent auditory objects using behavioral<br />
measures and AEPs. Expert musicians and nonmusicians were<br />
presented with complex sounds comprised of six harmonics �220, 440, 660<br />
Hz, etc.�. The third harmonic was either tuned or mistuned by 1%–16% of<br />
its original value. Mistuning a component of a harmonic complex results<br />
in the precept of a second auditory object. Stimuli were presented passively<br />
�no response� and actively �participants responded by indicating if<br />
they heard one sound or two sounds�. Behaviorally both musicians and<br />
3285 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3285<br />
4p FRI. PM
nonmusicians perceived a second auditory object at similar levels of mistuning.<br />
In both groups, complex sounds generated N1 and P2 waves at<br />
fronto-central scalp regions. The perception of concurrent auditory objects<br />
was paralleled by an increased negativity around 150 ms post-stimulus<br />
onset. This increased negativity is referred to as object-related negativity<br />
�ORN�. Small differences between musicians and nonmusicians were<br />
noted in the ORN. The implication of these results will be discussed in<br />
terms of current auditory scene analysis theory.<br />
3:50<br />
4pPP9. Study of auditory temporal processing of language and music<br />
in stroke patients. Li Hsieh and Tamara Baubie �Dept. of Commun.<br />
Sci. and Disord., Wayne State Univ., 207 Rackham Hall, 60 Farnsworth,<br />
Detroit, MI 48202�<br />
This study focused on functional lateralization of temporal processing<br />
of language and music for nine stroke patients and nine normal controls.<br />
Subjects were asked to discriminate short versus long sounds in ABX<br />
tasks, and then to reproduce sequences of short and long sounds they<br />
heard. The reproduction tasks consisted of sequences of 3, 4, and 5<br />
sounds. The linguistic stimuli were nonsense CVC syllables, and the musical<br />
stimuli were computer-generated musical notes with the timbres of<br />
French horn and trombone. Both linguistic and music stimuli were controlled<br />
for frequency and intensity, and varied only for duration �i.e., 500<br />
and 750 ms�. Our findings are consistent with previous studies; the left<br />
hemisphere specializes in linguistics, while the right hemisphere in music.<br />
Moreover, both hemispheres appeared to work closely together in processing<br />
temporal information. Both left- and right-hemisphere-damaged patients<br />
performed worse than controls. Most subjects performed better with<br />
music than language. Patients with left posterior lesions performed worse<br />
compared to patients with left or right anterior lesions, particularly with<br />
linguistic stimuli. Patients with right anterior lesions not only involved<br />
with temporal processing in music, but also in linguistic stimuli. Our study<br />
provided additional information regarding transient temporal processing in<br />
language and music.<br />
4:05<br />
4pPP10. Human bioacoustic biology: Acoustically anomalous vocal<br />
patterns used to detect biometric expressions relating to structural<br />
integrity and states of health. Sharry Edwards �Sound Health Res.<br />
Inst., P.O. Box 267, Sauber Res. Ctr., Albany, OH 45710�<br />
Computerized analyses of acoustically anomalous vocal patterns are<br />
being used as biomarkers for predictive, prediagnostic, and efficient management<br />
of individual biological form and function. To date, biometrically<br />
distinct vocal data have resulted in outcomes that would be considered<br />
improbable by contemporary medical standards. For instance, independent<br />
EMG conclusions confirmed the regeneration of nerve tissue for a multiple<br />
sclerosis patient who used acoustic bioinformation to guide his primary<br />
therapy. Another study monitored the amounts of synthetic labor hormones<br />
present during induced labor. False labor costs amount to millions of dollars<br />
each year in insurance and hospital resources. The use of noninvasive,<br />
possibly remote, vocal profiling could ameliorate such costs. Anomalous<br />
vocal acoustics are being investigated by many health-related organizations<br />
including Pfizer Pharmaceuticals and the Institute of Automatic Control<br />
Engineering in Taiwan. Complementary research studying molecular<br />
frequencies of cellular chemistry is being done by James Gimjewski,<br />
Ph.D., UCLA, Department of Chemistry and Biochemistry. Known as<br />
BioAcoustic Biology, this research modality has the potential to advance<br />
current health care standards for biological function, disease processes,<br />
and metabolism. Organizations such as the Acoustical Society of America<br />
are considering standards for technically defining human bioacoustics.<br />
This paper would expand that language.<br />
FRIDAY AFTERNOON, 1 DECEMBER <strong>2006</strong> HONOLULU ROOM, 1:00 TO 3:20 P.M.<br />
Session 4pSAa<br />
Structural Acoustics and Vibration: Vehicle Interior Noise and Vibration<br />
Courtney B. Burroughs, Cochair<br />
Noise Control Engineering, Inc. 1241 Smithfield St., State College, PA 16801<br />
Hiroaki Morimura, Cochair<br />
Japan Science and Technology Agency, Hosei Univ., Dept. of Mechanical Engineering, 3-7-2 Kajino-cho, Kogamei-city,<br />
Tokyo, Japan<br />
Invited Papers<br />
1:00<br />
4pSAa1. Analytical model for flow-excited interior cavity resonance and its application to the Stratospheric Observatory for<br />
Infrared Astronomy „SOFIA…. Jerry H. Ginsberg �G. W. Woodruff School of Mech. Eng., Georgia Inst. of Technol., Atlanta, GA<br />
30332-0405�<br />
The Stratospheric Observatory for Infrared Astronomy �SOFIA� is a joint effort between NASA and the German Space Agency<br />
that has installed a 20 000-kg telescope in a modified 747-SP. The modifications entailed constructing bulkheads, one of which is used<br />
to provide the active mount for the telescope, and a door that rotates to open as much as one-quarter of the fuselage circumference to<br />
the atmosphere. This configuration represents a large compartment that can exhibit acoustic resonances at low frequencies. Concern<br />
arose that a Rossiter mode, which is an aerodynamic resonance in which vortices shed from the leading edge of a gap form a coherent<br />
standing pattern at certain speeds, would create a strong acoustic source for acoustic and structural modes, whose frequencies might<br />
coincide. A model consisting of a two-dimensional hard-walled waveguide having a Rossiter mode source and an elastic plate at one<br />
end was analyzed in order to understand these issues. Unlike conventional analyses of interior cabin noise, in which vibrating walls<br />
are the acoustic source, the elastic plate represents a compliant boundary that couples with the acoustic modes. The results lead to<br />
some general insights to the the possibility of a ‘‘triple resonance,’’ as well as the role of structural compliance for cavities that are<br />
excited by turbulent external flows.<br />
3286 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3286
1:20<br />
4pSAa2. Energy finite energy analysis for shipboard noise. Raymond Fischer, Leo Boroditsky �Noise Control Eng. Inc., 799<br />
Middlesex Tnpk, Billerica, MA 01821�, Layton Gilfroy �DRDC-Atlantic�, and David Brennan �Martec Ltd.�<br />
Machinery-induced habitability noise is difficult to model efficiently and accurately. The potential of energy finite-element analysis<br />
�EFEA� is compared to other prediction tools such as statistical energy analysis �SEA�. This paper will explore the benefits and costs<br />
of EFEA with respect to SEA for acoustic modeling. The focus will be on issues relating to structural modeling for EFEA purposes.<br />
EFEA techniques will be evaluated to see if they possess the capabilities of verified SEA approaches for predicting habitability and<br />
radiated noise, where it is necessary to account for the impact of diverse marine constructions and sources such as the lack of<br />
machinery source information with respect to force or moment inputs or the finite impedance of machinery foundations. The effort<br />
proposed herein will provide the necessary engineering to research and identify salient features of EFEA that are potentially applicable<br />
for the detailed analysis of the acoustic environment and response of surface ships to various excitation sources. The paper will also<br />
address the pros and cons of SEA versus energy-finite element analysis �EFEA� methods used to predict the habitability noise of<br />
surface ship platforms. �This work is supported by an Office of Navy Research contract.�<br />
1:40<br />
4pSAa3. Spectral-based multicoordinate substructuring model for vehicle noise, vibration, and harshness refinement. Teik C.<br />
Lim �Mech., Industrial and Nuclear Eng., 598 Rhodes Hall, Box 210072, Univ. of Cincinnati, Cincinnati, OH 45221,<br />
teik.lim@uc.edu�<br />
The success of vehicle NVH �noise, vibration, and harshness� refinement often depends on the ability to identify and understand<br />
noise and vibration transmission paths within the mid-frequency range, i.e., 200–1000 Hz, throughout the assembled structure. Due to<br />
the complexity of the dynamics in this frequency range, most modal or finite element-based methods do not possess the fidelity<br />
needed. To address this gap, a multicoordinate substructuring theory applying measured structural-acoustic and vibration spectra is<br />
applied. Three forms of substructuring formulation, namely the nondiagonal, block-diagonal, and purely diagonal coupling cases, are<br />
developed. The performances of these approaches are studied numerically, and the net effects of these coupling formulations on the<br />
predicted joint and free substructure dynamic characteristics, and system response, are determined. Conditions for applying the<br />
simpler coupler that can simplify the testing process and overcome computational deficiencies are also derived. When the measured<br />
data is noise contaminated, the singular value decomposition �SVD� algorithm is found to be quite helpful. Using an actual vehicle,<br />
a comprehensive analysis of the measured and predicted vehicle system responses is performed. The results are employed to develop<br />
an understanding of the primary controlling factors and transfer paths and to cascade system requirements to the substructure level.<br />
2:00<br />
4pSAa4. Practical application of digital simulation with physical test in vehicle virtual. Toshiro Abe �ESTECH Corp., 89-1<br />
Yamashita-cho, Naka-ku, Yokohama-shi, Kanagawa-ken, Japan 231-0023, toshiro.abe@estech.co.jp�<br />
In the current vehicle design and development program, the Virtual Product Development process �hereafter, VPD process� is the<br />
innovation for automotive industry, which improves product quality and shortens time to market. In general, valid CAE applications<br />
are the key component of the VPD process as well as physical tests being indispensable to create valid simulation technologies. This<br />
presentation explains how physical-test-based CAE leverages the VPD process. In particular, the presentation is based on how<br />
physical testing supports the VPD process and why the ESTECH philosophy is that ‘‘The essence of CAE lies in its synergy with<br />
Testing;’’ a philosophy that differentiates the company from the competition. To demonstrate these activities, case studies based on<br />
automotive dynamic and real time simulation will be presented. In the case studies, vehicle body NVH and brake noise analysis will<br />
be used to show the interaction between physical test and computer simulation. Finally, practical application of the VPD process in<br />
one of the leading Japanese automotive companies will be shown, where the effectiveness of the front loading in the actual vehicle<br />
development program and the actual deployment of VPD process to the Functional Digital Vehicle project in the powertrain design are<br />
to be presented.<br />
2:20<br />
4pSAa5. Active noise control simulations using minimization of<br />
energy density in a mock helicopter cabin. Jared Thomas, Stephan P.<br />
Lovstedt, Jonathan Blotter, Scott D. Sommerfeldt �Brigham Young Univ.,<br />
Provo, UT 84602�, Norman S. Serrano, and Allan Egelston �Silver State<br />
Helicopters, Provo, UT 84601�<br />
Helicopter cabin noise is dominated by low-frequency tonal noise,<br />
making it an ideal candidate for active noise control. Previous work in<br />
active control of cabin noise suggests an energy density approach to be a<br />
good solution �B. Faber and S.D. Sommerfeldt, Global Control in a Mock<br />
Tractor Cabin Using Energy Density, Proc. ACTIVE 04, Sept. 2004.�<br />
Simulations for active noise control using energy density minimization<br />
have been made using recorded data from a Robinson R44 helicopter.<br />
Initial computer models show substantial noise reductions in excess of 6<br />
dB at the error sensor are possible. Performance results for computer<br />
models and simulations in a mock cab for different control arrangements<br />
will be compared.<br />
Contributed Papers<br />
2:35<br />
4pSAa6. Modeling airborne interior noise in full vehicles using<br />
statistical energy analysis. Arnaud Charpentier and Phil Shorter �ESI,<br />
12555 High Bluff Dr., Ste. 250, San Diego, CA 92130,<br />
arnaud.charpentier@esi-group-na.com�<br />
SEA is particularly well suited for predicting airborne noise in vehicles.<br />
The acoustic sources found in such environment are typically spatially<br />
distributed around the vehicle and can be well represented with SEA<br />
diffuse acoustic loads. Multiple transmission paths contribute to interior<br />
noise levels including �1� mass law transmission through trimmed panels,<br />
�2� resonant radiation from vibrating structures, and �3� flanking paths<br />
through gaskets, seals, and holes. All these transmission mechanisms may<br />
be modeled using SEA techniques. Finally, interior trim �including carpet,<br />
headliner, seats� is a key contributor to the acoustic performance of modern<br />
vehicles. The vehicle sound package has a significant impact on both<br />
the strength of the transmissions paths into the vehicle as well as the<br />
acoustic absorption in the cabin. Both these effects can be accounted for<br />
with SEA through detailed models of the trim. SEA models of full vehicles<br />
3287 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3287<br />
4p FRI. PM
are usually validated against experimental results at both component and<br />
system levels. The models can then be confidently used to �a� rank key<br />
design parameters governing interior levels and �b� quickly evaluate the<br />
impact of potential design changes. Example vehicle models and correlation<br />
results are presented here.<br />
2:50<br />
4pSAa7. Radiation from vibrating panels at high frequency including<br />
an inquiry into the role of edges and drive points. Donald Bliss<br />
�Mech. Eng. and Mater. Sci., Duke Univ., Durham, NC 27708�<br />
In the high frequency limit, a vibrating finite panel is shown to have<br />
broadband power and directivity characteristics that can expressed analytically<br />
by a limited set of parameters. Two-dimensional problems with subsonic<br />
structural wave speed are considered. Three basic directivity patterns<br />
are identified, associated with right and left traveling waves and the correlation<br />
between them. The role of boundary conditions at the panel edge<br />
is illustrated, as are the effects of types of forcing. Overall, relatively<br />
simple broadband behaviors are revealed. The analytical characterization<br />
of the radiation is shown to be particularly straightforward in the high<br />
frequency broadband limit. Interestingly, the radiated mean-square pressures<br />
are independent of the panel length, indicating the radiation is associated<br />
with the edges and the drive point. However, the radiation patterns<br />
cannot be explained in terms of simple volumetric sources placed just at<br />
the edges and the drive point, showing that the often-stated idea of uncan-<br />
celed volumetric sources at these locations is not correct except under very<br />
restricted circumstances. A correct physical interpretation of the radiation<br />
is provided both in physical space and in terms of spatial Fourier transforms.<br />
3:05<br />
4pSAa8. Some effect of trim and body panels on the low-frequency<br />
interior noise in vehicles. Andrzej Pietrzyk �Volvo Car Corp., Noise<br />
and Vib. Ctr., 405 31 Gothenburg, Sweden�<br />
Structure borne noise dominates the interior noise in vehicles at low<br />
frequencies. One of the basic vinroacoustic characteristics of the trimmed<br />
body is the noise transfer function, i.e., the acoustic response at a selected<br />
position in the passenger compartment, e.g., driver’s ear due to a mechanical<br />
excitation at a selected body mount. Detailed CAE models based on<br />
the FE method are today available for calculating this characteristic at low<br />
frequencies, corresponding to the engine idling and road excitation. However,<br />
the accuracy of CAE predictions of interior noise is still considered<br />
insufficient for the so-called analytical sign off, i.e., zero-prototypes vision.<br />
The current paper describes some investigations into the contribution<br />
of individual body panels to the overall interior noise. Also, the effect of<br />
selected interior trim items on the area investigated. Relative errors of<br />
prediction at different trim levels and different frequency ranges are discussed.<br />
Both experimental and CAE results are provided. The aim is to<br />
better understand the way the interior noise in vehicles is created, and how<br />
it can be controlled.<br />
FRIDAY AFTERNOON, 1 DECEMBER <strong>2006</strong> HONOLULU ROOM, 3:45 TO 5:45 P.M.<br />
Session 4pSAb<br />
Structural Acoustics and Vibration: General Vibration and Measurement Technology<br />
Peter C. Herdic, Cochair<br />
Naval Research Lab., Physical Acoustics Branch, Code 7136, 4555 Overlook Ave., SW, Washington, DC 20375<br />
Naoya Kojima, Cochair<br />
Yamaguchi Univ., Dept. of Mechanical Engineering, 2-16-1 Tokiwadai, Ube, Yamaguchi 755-8611, Japan<br />
3:45<br />
4pSAb1. Direct computation of degenerate elastodynamic solution of<br />
elastic wave propagation in a thick plate. Jamal Ghorieshi �Eng.<br />
Dept., Wilkes Univ., Wilkes-Barre, PA 18766�<br />
The limiting form of elastodynamic solutions as frequency tends to<br />
zero leads to the elastostatic eigenvalue equations. However, this limiting<br />
procedure is not convenient. It is cumbersome when applied to the solutions<br />
obtained using Stokes potentials and, in the case of utilizing Lames<br />
potentials, it does not produce static solutions that are a function of position<br />
alone. In this paper it is shown that the exact solutions of elastostatic<br />
problems can, in general, be obtained in a straightforward manner by the<br />
use of harmonic potentials without recourse to any special limiting form of<br />
analysis. This method is applied to an infinite, elastic thick plate with<br />
traction-free parallel surfaces and the elastostatic eigenvalue equation. It is<br />
shown that the problem can be solved exactly in terms of harmonic functions,<br />
one of which is a scalar and the other one is a vector. It is noted that<br />
results are in agreement with the published solutions.<br />
Contributed Papers<br />
4:00<br />
4pSAb2. Wave propagation characteristics of an infinite fluid-loaded<br />
periodic plate. Abhijit Sarkar and Venkata R. Sonti �Indian Inst. of Sci.<br />
Mech. Eng., IISc., Bangalore-560012, India�<br />
A 1-D infinite periodic plate with simple supports placed along equidistant<br />
parallel lines is considered using the finite-element method. The<br />
plate is loaded with a finite-height fluid column covered on the top with a<br />
rigid plate. Results show a relation between the propagation constant of<br />
the fluid-loaded structure with its in vacuo counterpart. Since the acoustic<br />
medium is an additional wave carrier, the attenuation bands corresponding<br />
to the in vacuo structure turn out to be propagating. However, the presence<br />
of the fluid can also bring about attenuation regions within the in vacuo<br />
propagation bands. Primary propagation constants bring additional waves<br />
called space harmonics with them. Hence, a localized coincidence effect is<br />
seen where a particular harmonic falls below or above the acoustic wave<br />
number, leading to propagation or a mass loading effect. Occasionally, a<br />
complete attenuation band is created. This is verified by decomposing the<br />
3288 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3288
single span displacement profile into the space harmonics and also by<br />
computing the frequency response function �FRF� for a finite fluid-loaded<br />
periodic plate and observing the huge antiresonance dip in frequency in<br />
the exact same frequency band where an attenuation band was predicted<br />
for the infinite structure.<br />
4:15<br />
4pSAb3. The dynamic response of a plate subjected to various edge<br />
excitations. Baruch Karp �Faculty of Aerosp. Eng., Technion Israel Inst.<br />
of Technol., Haifa 32000, Israel�<br />
Plane strain response of a semi-infinite, elastic plate to harmonic edge<br />
excitation is investigated analytically. The exact solution to this problem is<br />
obtained as a series expansion of the Rayleigh-Lamb modes of the plate.<br />
The variation of energy partition among the propagating modes with frequency<br />
of the edge excitation was found for load and displacement �symmetrical�<br />
perturbations of uniform and cosine form. The biorthogonality<br />
relation was employed in deriving the relative amplitudes of each mode to<br />
the given perturbation. The emphasis here is on the sensitivity of the<br />
far-field response, represented by propagating waves, to the details of the<br />
excitation. Within the frequency range investigated it was found that the<br />
plate’s response is remarkably insensitive to whether the excitation is load<br />
or displacement type. The two types of edge excitation distributions, on<br />
the other hand, result in different patterns of energy partition above the<br />
first cutoff frequency with similar energy partition only within limited<br />
range of frequencies. The effect of the nature of the excitation on the<br />
dynamic response of the plate and a possible implication to dynamic<br />
equivalence is discussed.<br />
4:30<br />
4pSAb4. Measurement of structural intensity using patch near-field<br />
acoustical holography. Kenji Saijyou �Fifth Res. Ctr., TRDI, Japan<br />
Defense Agency, 3-13-1 Nagase, Yokosuka City, Kanagawa Prefecture,<br />
Japan, 239-0826�<br />
Measurement of power flow in a structure, called the structural intensity<br />
�SI�, is essential for vibration control and noise reduction. The nearfield<br />
acoustical holography �NAH�-based measurement method is suitable<br />
to analyze the interrelationship between SI and acoustic intensity �AI�<br />
because NAH-based methods provide measurements of SI and AI simultaneously.<br />
Use of NAH requires the measurement of a pressure field over<br />
a complete surface located exterior to the structure. Therefore, if the measurement<br />
aperture is smaller than the structure, reconstructed results from<br />
the pressure on the finite aperture are seriously contaminated. This finite<br />
aperture effect prevents implementation of this NAH-based method on an<br />
actual large-scale structure such as a ship. Patch NAH and regularization<br />
method for SI measurement are applied to overcome this difficulty. This<br />
method enables implementation of the NAH-based SI measurement<br />
method in a large-scale structure. The effectiveness of this method is demonstrated<br />
by experiment.<br />
4:45<br />
4pSAb5. Flexural component and extensional component of vibration<br />
energy in shell structure. Taito Ogushi, Manabu Yahara, Masato<br />
Mikami, and Naoya Kojima �Dept. of Mech. Eng., Yamaguchi Univ.,<br />
2-16-1 Tokiwadai, Ube, Yamaguchi 755-8611, Japan,<br />
j008ve@yamaguchi-u.ac.jp�<br />
In this research, the behavior of the flexural component and the extensional<br />
component of vibration intensity and their transmission in curved<br />
shells are presented. L-shaped shell model was employed as an analysis<br />
model of FEM. As FEM analysis methods, both the frequency response<br />
analysis and the transitional response analysis were employed. The flexural<br />
component and the extensional component of vibration intensity �VI�<br />
were calculated by the results of FEM analysis. In the flexural component<br />
of the VI, the vibration energy supplied in the flat part decreased at the<br />
boundary from the flat part to the curved part and VI vectors flew in<br />
circumferential direction in the curved part. In the extensional component<br />
of the VI, the vibration energy appeared at the boundary from the flat part<br />
to the curved part and most VI vectors flew parallel to the shell axis in the<br />
curved part. The total vibration energy of the flexural component and the<br />
extensional component was conserved. So, the vibration energy transformed<br />
to each other between the flexural component and the extensional<br />
component in L-shaped shell.<br />
5:00<br />
4pSAb6. SeismicÕacoustic detection of ground and air traffic for<br />
unattended ground sensor technology. Peter C. Herdic �Naval Res.<br />
Lab., Physical Acoust. Branch, Washington, DC 20375 and SFA Inc.,<br />
Crofton, MD�, Brian H. Houston, Phillip A. Frank, and Robert M. Baden<br />
�Naval Res. Lab., Washington, DC 20375�<br />
Human footfall and vehicle traffic create surface waves in soil media<br />
that can easily be detected by seismic sensors. Field measurement data<br />
have been acquired with a triaxial geophone at several experimental sites.<br />
The in-plane-surface wave components dominate the response and decay<br />
at a rate of approximately 1/R, where R is distance. This decay rate is due<br />
to the combined effect of spreading �1/sqrt(R)) and damping losses in the<br />
soil. Further, the detection range is dependent upon the level of environmental<br />
noise, soil compliance, moisture content, and topography. Human<br />
detection was achieved in rural environments at distances up to �30–40<br />
m, and vehicle detection was possible at much greater distances. Seismic<br />
signals due to aircraft are small when compared to the acoustic signature.<br />
Ground-based microphone measurements clearly show the blade passage<br />
frequency tones of propeller airplanes and the broader band signature of<br />
turbojet aircraft. Time- and frequency-domain signal-processing methods<br />
for the detection and identification will also be introduced. These experimental<br />
results will be discussed with particular emphasis placed on wave<br />
phenomenon, detection and identification algorithms, and the related physics.<br />
5:15<br />
4pSAb7. Modeling of acoustic and elastic wave phenomena using<br />
plane wave basis functions. Tomi Huttunen, Jari P. Kaipio �Dept. of<br />
Phys., Univ. of Kuopio, P.O. Box 1627, FI-70211 Kuopio, Finland�, and<br />
Peter Monk �Univ. of Delaware, Newark, DE 19716�<br />
When simulating acoustic, elastodynamic, or coupled fluid-solid vibration<br />
problems using standard finite element �FE� techniques, several elements<br />
per wavelength are needed to obtain a tolerable accuracy for engineering<br />
purposes. At high wave numbers, the requirement of dense meshes<br />
may lead to an overwhelmingly large computational burden, which significantly<br />
limits the feasibility of FE methods for the modeling of wave<br />
phenomena. A promising technique for reducing the computational complexity<br />
is to use plane wave basis functions opposed to the low-order<br />
polynomials that are used in conventional FE methods. A possible method<br />
for utilizing the plane wave basis is the ultra-weak variational formulation<br />
�UWVF�. The UWVF method can be used for acoustic Helmholtz problems,<br />
elastodynamic Navier problems, or fluid-solid systems characterized<br />
by a coupled Helmholtz-Navier model. A comparison of the UWVF technique<br />
with a low-order FE method shows reduced computational complexity<br />
and improved accuracy.<br />
3289 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3289<br />
4p FRI. PM
Contributed Poster Paper<br />
Poster paper 4pSAb8 will be on display from 1:00 p.m. to 5:45 p.m. The author will be at the poster from 5:30 p.m. to 5:45 p.m.<br />
4pSAb8. Application of vibration energy flow to evaluation of<br />
thickness. Akihiko Higashi �Dept. of Maritime Sci. and Technol., Japan<br />
Coast Guard Acad., 5-1 Wakaba-cho, Kure, Hiroshima, 737-8512, Japan�<br />
In this study, the possibility of the useful application of the vibration<br />
energy flow is investigated. The vibration energy flow means the propagation<br />
of the vibration energy of the flexural waves. The vibration energy<br />
flow is expressed by the structural intensity. Here, it is easy to input the<br />
flexural waves in the thin plates and beam elements. Then, large structures<br />
such as ships use many of these thin plates and beam elements. But the<br />
usual methods of the evaluation and the inspection of these large structures<br />
are inefficient. Then, we investigated the possibility of the evaluation of<br />
the changed thickness of the structure by using the vibration energy flow<br />
analysis. As the result of analysis, the structural intensity suddenly<br />
changes at the position of the changed thickness. The changed quantity of<br />
the structural intensity corresponds to the change quantity of the thickness.<br />
Then, the evaluation method of the thickness of the structure is proposed.<br />
As a result, it was found that the change of the structural intensity indicates<br />
the change of the thickness. And the evaluation of the change of<br />
thickness of beams could be possible by using the proposed method.<br />
FRIDAY AFTERNOON, 1 DECEMBER <strong>2006</strong> MOLOKAI ROOM, 1:00 TO 4:00 P.M.<br />
Session 4pSC<br />
Speech Communication: Variation in Production and Perception of Speech „Poster Session…<br />
Heriberto Avelino, Cochair<br />
Univ. of California—Berkeley, Dept. of Linguistics, 1203 Dwinelle Hall, Berkeley, CA 94720-2650<br />
Haruo Kubozono, Cochair<br />
Kobe Univ., Dept. of Linguistics, Faculty of Letters, Nada-ku, Kobe 657-8501, Japan<br />
Contributed Papers<br />
All posters will be on display from 1:00 p.m. to 4:00 p.m. To allow contributors an opportunity to see other posters, contributors of<br />
odd-numbered papers will be at their posters from 1:00 p.m. to 2:30 p.m. and contributors of even-numbered papers will be at their<br />
posters from 2:30 p.m. to 4:00 p.m.<br />
4pSC1. Cross-language perception of voice and affect. Christer Gobl,<br />
Irena Yanushevskaya, and Ailbhe N. Chasaide �Phonet. and Speech Lab.,<br />
School of Linguistic, Speech and Commun. Sci., Trinity College Dublin,<br />
Dublin 2, Ireland, yanushei@tcd.ie�<br />
The paper reports on a cross-language study of how voice quality and<br />
f 0 combine in the signaling of affect. Speakers of Irish-English and Japanese<br />
participated in perception tests. The stimuli consisted of a short utterance<br />
where f 0 and voice source parameters were varied using the LFmodel<br />
implementation of the KLSyn88a formant synthesizer, and were of<br />
three types: �1� VQ only involving voice quality variations and a neutral<br />
f 0 contour; �2� f 0 only, with different affect-related f 0 contours and<br />
modal voice; �3� VQ� f 0 stimuli, where the voice qualities of �1� combine<br />
with specific f 0 contours from �2�. Overall, stimuli involving voice quality<br />
variation were consistently associated with affect. In �2� only stimuli with<br />
high f 0 yielded high affective ratings. Striking differences emerge between<br />
the ratings obtained from the two language groups. The results<br />
show that not only were some affects consistently perceived by one language<br />
group and not the other, but also that specific voice qualities and<br />
pitch contours were associated with very different affects across the two<br />
groups. The results have important implications for expressive speech synthesis,<br />
indicating that language/culture-specific differences need to be considered.<br />
�This work is supported by the EU-funded Network of Excellence<br />
on Emotion, HUMAINE.�<br />
4pSC2. An articulatory study of coronal consonants in Arrernte.<br />
Marija Tabain �LaTrobe Univ., Melbourne, Australia�, Richard Beare<br />
�Monash Univ., Melbourne, Australia�, Catherine Best �Univ. of Western<br />
Sydney, Sydney, Australia�, and Louis Goldstein �Haskins Labs., CT�<br />
This paper presents electro-magnetic articulography �EMA� data on<br />
the four coronal stops of Arrernte, an Australian language. The stops are:<br />
the lamino-dental ‘‘th,’’ the apico-alveolar ‘‘t,’’ the apico-postalveolar �or<br />
‘‘retroflex’’� ‘‘rt,’’ and the lamino-palatal ‘‘ty.’’ Jaw, tongue tip �TT�, and<br />
tongue body �TB� data were collected for two female speakers of the<br />
language. Results for the first speaker show a fronted tongue position for<br />
the laminal consonants, with the TT reflecting a similar location for both<br />
the dental and the palatal. However, for the palatal, the TB position is<br />
much higher, whereas for the dental, the TB is very low. For the apical<br />
consonants, the TT is not as far forward, and the TB is not quite as high as<br />
for the lamino-palatal. For both TT and TB, apico-postalveolar is further<br />
back than apico-alveolar. For the second speaker, the TT sensor failed, but<br />
in line with the first speaker, the TB sensor showed a higher position for<br />
the palatal. The other stops were lower and more forward, with the postalveolar<br />
TB position higher than the laminal or alveolar stop position. For<br />
both speakers, the jaw position is lowest for the postalveolar. �Work supported<br />
by Australian Research Council and NIH: NIDCD.�<br />
3290 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3290
4pSC3. Symbolic phonetic features for pronunciation modeling.<br />
Rebecca A. Bates, a� , Mari Ostendorf �Dept. of Elec. Eng., Univ. of<br />
Washington, Box 352500, Seattle, WA 98195�, and Richard A. Wright<br />
�Univ. of Washington, Seattle, WA 98195�<br />
A significant source of variation in spontaneous speech is due to intraspeaker<br />
pronunciation changes, often realized as small feature changes,<br />
e.g., nasalized vowels or affricated stops, rather than full phone transformations.<br />
Previous computational modeling of pronunciation variation has<br />
typically involved transformations from one phone to another, partly because<br />
most speech processing systems use phone-based units. Here, a<br />
phonetic-feature-based prediction model is presented where phones are<br />
represented by a vector of symbolic features that can be on, off, unspecified,<br />
or unused. Feature interaction is examined using different groupings<br />
of possibly dependent features, and a hierarchical grouping with conditional<br />
dependencies led to the best results. Feature-based models are<br />
shown to be more efficient than phone-based models, in the sense of<br />
requiring fewer parameters to predict variation while giving smaller distance<br />
and perplexity values when comparing predictions to the handlabeled<br />
reference. A parsimonious model is better suited to incorporating<br />
new conditioning factors, and this work investigates high-level information<br />
sources, including both text �syntax, discourse� and prosody cues.<br />
Detailed results are under review with Speech Communication. �This research<br />
was supported in part by the NSF, Award No. IIS-9618926, an Intel<br />
Ph.D. Fellowship, and by a faculty improvement grant from Minnesota<br />
State University Mankato.� a� Currently at Minnesota State University,<br />
Mankato.<br />
4pSC4. Acoustic phonetic variability and auditory word recognition<br />
by dyslexic and nondyslexic children. Patricia Keating, Kuniko Nielsen<br />
�Phonet. Lab., Linguist., UCLA, Los Angeles, CA 90095-1543,<br />
keating@humnet.ucla.edu�, Frank Manis, and Jennifer Bruno �USC, Los<br />
Angeles, CA 90089�<br />
The hypothesis that dyslexia involves degraded phonological representations<br />
predicts impairments in behaviors that rely on these representations,<br />
such as auditory word recognition. Normal adult listeners recognize<br />
different pronunciations of a word as instances of the same lexical item,<br />
but more slowly and less accurately; dylexics should be even more impaired<br />
by acoustic phonetic variability. Children with and without dyslexia<br />
performed a word recognition task: on each trial, a child hears a target<br />
word, then eight probes �matching the target or not�, responding yes/no to<br />
each probe. On some trials, probes are spoken by multiple talkers who<br />
differ in age, sex, speech style, etc.; on some trials the match probes also<br />
differ from the target in final stop consonant allophone. Responses are<br />
scored for accuracy and speed. Research questions include: Do all children<br />
demonstrate less accurate/slower recognition of words spoken by multiple<br />
talkers versus by one talker? Do all children demonstrate less accurate/<br />
slower recognition of words spoken with different allophones? Do dyslexic<br />
children demonstrate less accurate/slower recognition than nondyslexic<br />
children and, if so, for all trials, only for multiple talker trials, and/or<br />
only for different allophone trials; for all dyslexic children, or only those<br />
with particular phonological impairments? �Work supported by NIH.�<br />
4pSC5. Intertalker differences in intelligibility of cochlear-implant<br />
simulated speech. Tessa Bent, Adam B. Buchwald, and David B. Pisoni<br />
�Indiana Univ., Dept. of Psychol. and Brain Sci., 1101 E. 10th St.,<br />
Bloomington, IN 47405, tbent@indiana.edu�<br />
Are the acoustic-phonetic factors that promote highly intelligible<br />
speech invariant across different listener populations? Two approaches<br />
have been taken to investigate intelligibility variation for a variety of<br />
listener populations including hearing-impaired listeners, second language<br />
learners, and listeners with cochlear implants: studies on how speaking<br />
style affects intelligibility and other research on how inherent differences<br />
among talkers influence intelligibility. Taking the second approach, we<br />
compared intertalker differences in intelligibility for normal-hearing listeners<br />
under cochlear implant �CI� simulation (n�120) and in quiet (n<br />
�200). Stimuli consisted of 20 native English talkers’ productions of 100<br />
sentences. These recordings were processed to simulate listening with an<br />
eight-channel CI. Both clear and CI-processed tokens were presented to<br />
listeners in a sentence transcription task. Results showed that the most<br />
intelligible talkers in quiet were not the most intelligible talkers under CI<br />
simulation. Furthermore, listeners demonstrated perceptual learning with<br />
the CI-simulated speech but showed little learning in the quiet. Some of<br />
the acoustic-phonetic properties that were correlated with intelligibility<br />
differed between the CI-simulated speech and the speech in the quiet.<br />
These results suggest that the intertalker variations that result in highly<br />
intelligible speech observed in earlier studies are dependent on listener<br />
characteristics. �Work supported by NIDCD.�<br />
4pSC6. The effect of phonological neighborhood density and word<br />
frequency on vowel production and perception in clear speech. Rajka<br />
Smiljanic, Josh Viau, and Ann Bradlow �Dept. of Linguist., Northwestern<br />
Univ., 2016 Sheridan Rd., Evanston, IL 60208�<br />
Previous research showed that phonological neighborhood density and<br />
word frequency influence word recognition �Luce and Pisoni, 1998� and<br />
vowel production �Wright, 2002; Munson and Solomon, 2004; Munson, to<br />
appear�, suggesting an interaction of lexical and phonetic factors in speech<br />
production and perception Here, we explore whether hyperarticulated,<br />
intelligibility-enhancing clear speech shows similar sensitivity to lexicallevel<br />
structure. Nine American English talkers �five females, four males�<br />
produced 40 monosyllabic easy �frequent words with few lexical neighbors�<br />
and hard �infrequent words with many lexical neighbors� words in<br />
conversational and clear speech. Twenty-four subjects participated in a<br />
word-in-noise listening test. Results revealed a large effect of style on<br />
intelligibility and vowel production: words were more intelligible and<br />
vowels were longer and more dispersed in clear compared to conversational<br />
speech. Moreover, the female talkers produced larger vowel spaces<br />
than male talkers in both speaking styles. Vowels in hard words were<br />
marginally more dispersed than vowels in easy words in both speaking<br />
styles. However, within both speaking styles, easy and hard words were<br />
equally intelligible and of approximately equal duration. These results<br />
showed that phonetic properties of vowels were enhanced equally in clear<br />
speech regardless of their lexical properties.<br />
4pSC7. Phoneme dependency of accuracy rates in familiar and<br />
unknown speaker identification. Kanae Amino, Takayuki Arai �Dept.<br />
of Elec. and Electron. Eng., Sophia Univ., 7-1 Kioi-cho, Chiyoda-ku,<br />
Tokyo, 102-8554 Japan, amino-k@sophia.ac.jp�, and Tsutomu Sugawara<br />
�Sophia Univ., Chiyoda-ku, Tokyo, 102-8554 Japan�<br />
For perceptual speaker identification, the identification accuracy depends<br />
on the speech contents presented to subjects. Our previous studies<br />
have shown that stimuli containing nasals are effective for identifying<br />
familiar speakers �Amino et al., Acoust. Sci. Tech. 27�4� �<strong>2006</strong>��. We have<br />
also presented the possibility that the interspeaker spectral distances reflect<br />
perceptual speaker similarities. In the present study, we conducted an experiment<br />
in which four unknown speakers were identified by 15 subjects.<br />
The stimuli were identical to those used in the previous study, in which ten<br />
speakers were identified by familiar listeners, although the speakers were<br />
fewer this time. Nine consonants in the CV structure were used as stimuli.<br />
The consonants were /d/, /t/, /z/, /s/, /r/, /j/, /m/, /n/, and /nj/; the vowel was<br />
restricted to /a/ for all CV syllables to simplify the experiment. The results<br />
showed that the nasals /n/ and /nj/ obtained higher scores. Tendencies in<br />
the differences among consonants were on the same order as those of the<br />
3291 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3291<br />
4p FRI. PM
previous experiment, but the average scores were lower than those for<br />
familiar listeners. �Work supported by Grant-in-Aid for JSPS Fellows 17-<br />
6901.�<br />
4pSC8. Speech style and stereotypical character in Japanese. Akiko<br />
Nakagawa �Grad. School of Cultural Studies and Human Sci., Kobe<br />
Univ., 1-2-1 Tsurukabuto, Nada-ku, Kobe 657-8501, Japan,<br />
akiko.nakagawa@atr.jp� and Hiroko Sawada �Kyoto Univ., Kyoto<br />
606-8501, Japan�<br />
This study shows that ‘‘stereotypical character’’ is necessary to understand<br />
Japanese speech communication in addition to existing conceptions<br />
such as emotion, communicative strategy, register, and so on. Stereotypical<br />
character is here defined as a complex entity, consisting of information<br />
about gender, age, social status, physical features, characteristics, and<br />
speech style. The necessity of stereotypical character was shown through<br />
an auditory experiment involving a total of 70 speech sounds comprised of<br />
15–19 short phrases �mean duration 1.4� selected from recordings of spontaneous<br />
speech of four adult female speakers of Japanese. Ten participants<br />
were asked to listen to these speech sounds randomly, and to classify them<br />
into four speakers. Each of the resulting auditory-perceptual categories<br />
was found to contain speech sounds from more than one speaker. Further<br />
analyses of these results suggested that the participants classified the<br />
speech sounds not according to invariant speaker characteristics but according<br />
to virtual stereotypical characters that are common in Japanese<br />
society. Therefore, such changeable speaker characteristics as ‘‘busybody’’<br />
‘‘thoughtful,’’ ‘‘high-handed,’’ and so on, can be elicited through speech<br />
sounds by Japanese speakers. �This work was partially supported by the<br />
Ministry of Education, Science, Sport, and Culture, Grant-in-Aid for Scientific<br />
Research �A�, 1620<strong>2006</strong>.�<br />
4pSC9. Perceived vocal age and its acoustic correlates. Hiroshi Kido<br />
�Dept. of Commun. Eng., Tohoku Inst. of Technol., Taihaku-ku, Sendai,<br />
Japan 989-8577, kidoh@tohtech.ac.jp� and Hideki Kasuya �Intl. Univ. of<br />
Health and Welfare, Otawara, Japan 324-8501�<br />
This study investigates relationships between perceived and chronological<br />
age of talkers and acoustic correlates of the perceived age. Most of<br />
the past studies were primarily concerned with the instability of the vocalfold<br />
vibration extracted from sustained vowels. This study focuses on the<br />
dynamic nature of sentence utterances. Talkers included 115 healthy men,<br />
aged 20–60 years, who read a short sentence in Japanese. Listeners consisted<br />
of 70 men and women, aged 20–40 years, who made direct estimations<br />
of age. The results showed a strong correlation (r�0.66) between<br />
the perceived and chronological age as well as the tendency toward overestimating<br />
the ages of younger talkers and underestimating those of older<br />
talkers, supporting past investigations �e.g., R. Huntley et al., J. Voice 1,<br />
49–52 �1987��. Acoustic parameters considered were median of the fundamental<br />
frequency (F0� contour, F0 range, declination of F0 contour,<br />
spectral tilt, median of the boundary frequencies above which irregularities<br />
dominate, and speaking rate. From both statistical graphical modeling<br />
and regression tree analysis, the speaking rate, F0 declination, and spectral<br />
tilt were found to be dominant acoustic correlates to the perceived age.<br />
�Work supported partly by a Grant-in-Aid for Scientific Research, JSPS<br />
�16300061�.�<br />
4pSC10. A cross-linguistic study of informational masking: English<br />
versus Chinese. Bruce A. Schneider, Liang Li, Meredyth Daneman<br />
�Dept. of Psych., Univ. of Toronto at Mississauga, Mississauga, ON, L5L<br />
1C6 Canada, bschneid@utm.utoronto.ca�, Xihong Wu, Zhigang Yang,<br />
Jing Chen, and Ying Huang �Peking Univ., Beijing, China 10087�<br />
The amount of release from informational masking in monolingual<br />
English �Toronto, Canada�, and Chinese �Beijing, China� listeners was<br />
measured using the paradigm developed by Freyman et al. �J. Acoust.<br />
Soc. Am. 106, 3578–3588�. Specifically, psychometric functions relating<br />
percent-correct word recognition to signal-to-noise ratio were determined<br />
under two conditions: �1� masker and target perceived as originating from<br />
the same position in space; �2� masker and target perceived as originating<br />
from different locations. The amount of release from masking due to spatial<br />
separation was the same for English and Chinese listeners when the<br />
masker was speech-spectrum noise or cross linguistic �Chinese speech<br />
masking English target sentences for English listeners or English speech<br />
masking Chinese target sentences for Chinese listeners�. However, there<br />
was a greater release from masking for same-language masking of English<br />
�English speech masking English target sentences� than for same-language<br />
masking of Chinese �Chinese speech masking Chinese target sentences�. It<br />
will be argued that the differences in same-language masking between<br />
English and Chinese listeners reflect structural differences between English<br />
and Mandarin Chinese. �Work supported by China NSF and CIHR.�<br />
4pSC11. Cross-linguistic differences in speech perception. Keith<br />
Johnson and Molly Babel �UC Berkeley, 1203 Dwinelle Hall, Berkeley,<br />
CA 94720-2650�<br />
This research explores language-specific perception of speech sounds.<br />
This paper discusses two experiments: experiment 1 is a speeded forcedchoice<br />
AX discrimination task and experiment 2 is a similarity rating task.<br />
Experiment 1 was intended to investigate the basic auditory perception of<br />
the listeners. It was predicted that listeners’ native languages would not<br />
influence responses in experiment 1. Experiment 2 asked subjects to rate<br />
the similarity between two tokens on a five-point equal interval scale; the<br />
purpose of this experiment was to explore listeners’ subjective impression<br />
of speech sounds. In experiment 2 it was predicted that listeners’ language<br />
would affect their responses. The same stimuli were used in both experiments.<br />
The stimuli consisted of vowel-fricative-vowel sequences produced<br />
by a trained phonetician. Six fricatives were used: /f, th, s, sh, x, h/. These<br />
fricatives were embedded in three vowel environments: /a_a/, /i_i/, and<br />
/u_u/. Tokens were presented to listeners over headphones with a 100-ms<br />
interval. Independent groups of 15 native Dutch and English listeners<br />
participated in each of the two experiments. Results suggest that listeners’<br />
language influenced responses in both experiments, albeit the result was<br />
larger in experiment 2. �Work supported by NIH.�<br />
4pSC12. Neural coding of perceptual interference at the preattentive<br />
level. Yang Zhang �Dept. of Speech-Lang.-Hearing Sci., Univ. of<br />
Minnesota, Minneapolis, MN 55455�, Patricia Kuhl, Toshiaki Imada<br />
�Univ. of Washington, Seattle, WA 98195�, Toshiaki Imada, and Masaki<br />
Kawakatsu �Tokyo Denki Univ., Inzai-shi, Chiba 270-1382, Japan�<br />
Language acquisition involves neural commitment to languagespecific<br />
auditory patterns, which may interfere with second language<br />
learning. This magnetoencephalography study tested whether perceptual<br />
interference could occur at the preattentive level. Auditory mismatch field<br />
�MMF� responses were recorded from ten American and ten Japanese<br />
adult subjects in the passive oddball paradigm. The subjects read selfchosen<br />
books and ignored the sounds. Three pairs of synthetic /ra-la/ syllables<br />
were used: one cross-category pair varied only in the third formant<br />
�F3�, and the other two within-category pairs varied only in the second<br />
formant �F2�. ANOVA results showed a main effect of acoustic dimension<br />
with significant interaction with subject groups �p�0.01). As reported<br />
earlier, American listeners showed larger but later MMF responses for the<br />
F3 change. By contrast, Japanese listeners showed larger and earlier<br />
MMFs than Americans for changes in F2. Moreover, Japanese listeners<br />
had larger and earlier MMF responses for the changes in F2 as against<br />
changes in F3, which was more prominent in the right hemisphere than in<br />
the left. These results provided further support for the hypothesis that<br />
language experience produces neural networks dedicated to the statistical<br />
properties of incoming speech experienced in infancy, which later interfere<br />
with second language acquisition.<br />
3292 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3292
4pSC13. Russian and Spanish listener’s perception of the English<br />
tenseÕlax vowel contrast: Contributions of native language allophony<br />
and individual experience. Maria V. Kondaurova �Program in<br />
Linguist., Purdue Univ., West Lafayette, IN 47907� and Alexander L.<br />
Francis �Purdue Univ., West Lafayette, IN 47906�<br />
We examined the influence of listeners native phonology on the perception<br />
of American English tense and lax front unrounded vowels ��i� and<br />
�(��. These vowels are distinguishable according to both spectral quality<br />
and duration. Nineteen Russian, 18 Spanish, and 16 American English<br />
listeners identified stimuli from a beat-bit continuum varying in nine spectral<br />
and nine duration steps. English listeners relied predominantly on<br />
spectral quality when identifying these vowels, but also showed some<br />
reliance on duration. Russian and Spanish speakers relied entirely on duration.<br />
Three additional tests examined listeners allophonic use of vowel<br />
duration in their native languages. Duration was found to be equally important<br />
for the perception of lexical stress for all three language groups.<br />
However, the use of duration as a cue to postvocalic consonant voicing<br />
differed due to phonotactic differences across the three languages. Group<br />
results suggest that non-native perception of the English tense/lax vowel<br />
contrast is governed by language-independent psychoacoustic factors<br />
and/or individual experience. Individual results show large variability<br />
within all three language groups, supporting the hypothesis that individual<br />
differences in perceptual sensitivity as well as the more frequently cited<br />
factors of second language education and experience play an important<br />
role in cross-language perception.<br />
4pSC14. An analysis of acoustic deviation manner in spontaneous<br />
speech. Norimichi Hosogai, Kanae Okita, Takuya Aida, and Shigeki<br />
Okawa �Chiba Inst. of Technol., 2-17-1 Tsudanuma, Narashino, Chiba<br />
275-0016, Japan�<br />
Natural speech typically contains various phenomena deviated from<br />
the formal mode such as read speech. It is well known that those paralinguistic<br />
phenomena have an important role to give the human emotions and<br />
the state of the speakers in speech communication. This study attempts to<br />
extract the deviation as an acoustic ‘‘vagueness,’’ defined by temporal and<br />
dynamical acoustic features of speech. Especially the change of the vagueness<br />
during a certain period of speech, such as a 10-minute presentation, is<br />
focused. As the acoustic features, it used �i� modulation spectrum and �ii�<br />
syllable speed, which may have relations to the speech clarity and the<br />
tempo. For the experiments, 70 academic presentation speech data in the<br />
Corpus of Spontaneous Japanese �CSJ� are used. As the experimental results,<br />
significant properties in the patterns of the modulation spectrum and<br />
the syllable speed are obtained as a difference of the beginning and the<br />
ending periods of the presentation. This result will contribute to a humanlike<br />
speech dialog system.<br />
4pSC15. Nondurational cues for durational contrast in Japanese.<br />
Kaori Idemaru �Dept. of East Asian Lang. and Lit., Univ. of Oregon, 1248<br />
Univ. of Oregon, Eugene, OR 97403� and Susan G Guion �Univ. of<br />
Oregon, 1290 Eugene, OR 97403�<br />
This study explores potential secondary cues to a durational contrast<br />
by examining short and long stop consonants in Japanese. Durational contrasts<br />
are susceptible to considerable variability in temporal dimensions<br />
caused by changes in speaking rate. In this study, the proposal is examined<br />
that multiple acoustic features covary with the stop length distinction and<br />
that these features may aid in accessing the percept intended by the<br />
speaker, even when the primary cue, closure duration, is unreliable. The<br />
results support the proposal, revealing the presence of multiple acoustic<br />
features covarying with the short versus long contrast. Not only are there<br />
durational correlates to this contrast—the preceding vowel is longer and<br />
the following vowel is shorter for geminates than singletons—but there<br />
also are nondurational features covarying with this contrast. Greater fundamental<br />
frequency and intensity drops are found from the preceding to<br />
the following vowel for the geminate than the singleton stops. These results<br />
suggest the possibility that systematic variation of these acoustic<br />
features is used in the perceptual categorization of the contrast in addition<br />
to the primary cue of closure duration. Moreover, the nondurational correlates<br />
are promising candidates for speech-rate resistant features.<br />
4pSC16. Different motor strategies for increasing speaking rate: Data<br />
and modeling. Majid Zandipour, Joseph Perkell, Mark Tiede, Frank<br />
Guenther �M.I.T., Res. Lab Electron., Speech Commun. Group, 50 Vassar<br />
St, Cambridge, MA 02139, majidz@speech.mit.edu�, Kiyoshi Honda<br />
�ATR Human Information Processing Res. Lab., Kyoto 619-0288, Japan�,<br />
and Emi Murano �Univ. Maryland Dental School, Baltimore, MD, 21209�<br />
Different motor strategies for increasing speaking rate: data and modeling<br />
EMG, kinematic and acoustic signals were recorded from two male<br />
subjects as they pronounced multiple repetitions of simple nonsense utterances.<br />
The resulting data indicate that the two subjects employed different<br />
motor strategies to increase speaking rate. When speaking faster, S1 significantly<br />
increased the size of the articulatory target region for his tongue<br />
movements, increased the speed of the tongue movements and the rate of<br />
EMG rise somewhat, while decreasing the movement duration significantly<br />
and movement distance slightly. In contrast, at the fast rate, S2 had<br />
the same size articulatory target region and rate of EMG rise as at the<br />
normal rate, but decreased the speed, distance, and duration of tongue<br />
movement slightly. Each subject had similar dispersions of acoustic targets<br />
in F1–F2 space at fast versus normal rates, but both shifted target centroids<br />
toward the center of the vowel space at the fast rate. Simulations<br />
with a biomechanical model of the vocal tract show how modulations of<br />
motor commands may account for such effects of speaking rate on EMG,<br />
kinematics, and acoustic outputs. �Work supported by NIDCD, NIH.�<br />
4pSC17. Effect of speaking rate on individual talker differences in<br />
voice-onset-time. Rachel M. Theodore, Joanne L. Miller, and David<br />
DeSteno �Dept. of Psych., 125 NI, Northeastern Univ., 360 Huntington<br />
Ave., Boston, MA, 02115-5000, r.theodore@neu.edu�<br />
Recent findings indicate that individual talkers systematically differ in<br />
phonetically relevant properties of speech. One such property is voiceonset-time<br />
�VOT� in word-initial voiceless stop consonants: at a given rate<br />
of speech, some talkers have longer VOTs than others. It is also known<br />
that for any given talker, VOT increases as speaking rate slows. We examined<br />
whether the pattern of individual differences in VOT holds across<br />
variation in rate. For example, if a given talker has relatively short VOTs<br />
at one rate, does that talker also have relatively short VOTs at a different<br />
rate? Numerous tokens of /ti/ were elicited from ten talkers across a range<br />
of rates using a magnitude-production procedure. VOT and syllable duration<br />
�a metric of speaking rate� were measured for each token. As expected,<br />
VOT increased as syllable duration increased �i.e., rate slowed� for<br />
each talker. However, the slopes as well as the intercepts of the functions<br />
relating VOT to syllable duration differed significantly across talkers. As a<br />
consequence, a talker with relatively short VOTs at one rate could have<br />
relatively long VOTs at another rate. Thus the pattern of individual talker<br />
differences in VOT is rate dependent. �Work supported by NIH/NIDCD.�<br />
4pSC18. Variation in vowel production. Joseph Perkell, Majid<br />
Zandipour, Satrajit Ghosh, Lucie Menard �Speech Commun. Group, Res.<br />
Lab. of Electron., Rm. 36-511, M.I.T., Cambridge, MA 02139�, Harlan<br />
Lane, Mark Tiede, and Frank Guenther �M.I.T., Cambridge, MA 02139�<br />
Acoustic and articulatory recordings were made of vowel productions<br />
by young adult speakers of American English—ten females and ten<br />
males—to investigate effects of speaker and speaking condition on measures<br />
of contrast and dispersion. The vowels in the words teat, tit, tet, tat,<br />
tot, and toot were embedded in two-syllable ‘‘compound words’’ consisting<br />
of two CVC syllables, in which each of the two syllables comprised a<br />
real word, the consonants were /p/, /t/ or /k/, the two adjoining consonants<br />
were always the same, the first syllable was unstressed and the second<br />
stressed. Variations of phonetic context and stress were used to induce<br />
3293 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3293<br />
4p FRI. PM
dispersion around each vowel centroid. The compound words were embedded<br />
in a carrier phrase and were spoken in normal, clear, and fast<br />
conditions. Initial analyses of F1 andF2 on 15 speakers have shown<br />
significant effects of speaker, speaking condition �and also vowel, stress,<br />
and context� on vowel contrast, and dispersion around means. Generally,<br />
dispersions increased and contrasts diminished going from clear to normal<br />
to fast conditions. Results of additional analyses will be reported. �Work<br />
supported by NIDCD, NIH.�<br />
4pSC19. Region, gender, and vowel quality: A word to the wise<br />
hearing scientist. Richard Wright �Dept. of Linguist., Univ. of<br />
Washington, Box 354340, Seattle, WA 98195-4340,<br />
rawright@u.washington.edu�, Stephanie Bor, and Pamela Souza �Univ. of<br />
Washington, Seattle, WA 98105�<br />
Sociophonetic research has established effects of regional accent and<br />
gender on spoken vowels. Many gender differences are due to sociolinguistic<br />
factors and thus vary by region. The implications for researchers<br />
and clinicians are important: gender variation must be controlled for according<br />
to the region of the listener and talker population. Moreover,<br />
speech perception stimuli used in research and in clinical applications<br />
have limited regional application. This poster illustrates these factors using<br />
the Pacific Northwest regional accent. The data, collected for a previous<br />
study on hearing aid processing, consist of three repetitions of eight vowels<br />
produced in real-word /h_d/ �or /_d/� contexts by six males and six<br />
females ranging in age from 19 to 60. Formants were measured using an<br />
LPC with an accompanying FFT and spectrogram for verification. The<br />
results revealed vowel-specific differences in the male and female speech<br />
over and above those typically associated with physiologic predictions,<br />
and different again from those observed in past studies from different<br />
regions. Taken as a whole, these data suggest that speech and hearing<br />
researchers should take care in selecting stimuli for general-use speech<br />
perception tests. �Work supported by NIDCD training grant �#DC00033�<br />
and NIH RO1 �1 RO1 DC006014�.�<br />
4pSC20. Acoustic characteristics of vowels in three regional dialects<br />
of American English. Ewa Jacewicz, Robert Allen Fox, Yolanda Holt<br />
�Speech Acoust. and Percept. Labs., Dept. of Speech and Hearing Sci., The<br />
Ohio State Univ., Columbus, OH 43210�, and Joseph Salmons �Univ. of<br />
Wisconsin—Madison, Madison, WI�<br />
Most of the comparative sociophonetic studies of regional dialect<br />
variation have focused on individual vowel differences across dialects as<br />
well as speaker variables. The present work seeks to define basic acoustic<br />
characteristics of entire vowel systems for three different regional variants<br />
of American English spoken in southeastern Wisconsin �affected by the<br />
Northern Cities Shift�, western North Carolina �affected by the Southern<br />
Vowel Shift�, and central Ohio �not considered to be affected currently by<br />
any vowel shift�. Three groups of speakers �men and women� aged 20–29<br />
years were recorded from each geographic area defined by two to three<br />
counties �creating a highly homogeneous set of speakers�. Acoustic measures<br />
for the set of 14 monophthongs and diphthongs in /h_d/ context<br />
included vowel space area for each speaker, global spectral rate of change<br />
for diphthongized vowels �defined over the first three formant slopes�, the<br />
amount of frequency change for F1 and F2 at two temporal points located<br />
close to vowel onset and offset �vector length�, and vowel duration. These<br />
measures will establish both systemic and vowel inherent characteristics<br />
across the three dialects, serving as a basis for future examination of<br />
conditioning factors on vowels in chain shifts. Dialectal differences will be<br />
discussed. �Work supported by NIH NIDCD R01 DC006871.�<br />
4pSC21. The rhythmic characterization of two varieties of<br />
Portuguese. Verna Stockmal, Emilia Alonso Marks, Audra Woods, and<br />
Z. S. Bond �Ohio Univ., Athens, OH 45701�<br />
As spoken in Europe, Portuguese is said to be stress-timed, while<br />
Brazilian Portuguese appears to display characteristics of both stress and<br />
syllable timing �P. A. Barbosa, D.E.L.T.A. 16, 369–402 �2000��. Weemployed<br />
the Ramus et al. metric, based on acoustic-phonetic measurements<br />
�Ramus et al., Cognition 73, 265–292 �1999��, to investigate the possibility<br />
of distinguishing between the two varieties of the language. Five native<br />
speakers of European Portuguese and five native speakers of Brazilian<br />
Portuguese recorded the same short prose passage taken from a magazine.<br />
The talkers were asked to read at a normal, comfortable rate. The reading<br />
time of the passage averaged 60 s, with considerable differences among<br />
the talkers. From the vocalic and consonantal intervals, the Ramus metrics,<br />
percent vocalic interval and standard deviation of consonantal and<br />
vocalic interval, were calculated. The five talkers of the two language<br />
varieties differed on the values of these parameters. The values of %V and<br />
SD-V showed overlap between the two talker groups, while the BP talkers<br />
tended to show lower values for SD-C. Apparently, the rhythmic characterization<br />
of the two varieties of the language is not clearly categorical, but<br />
rather ranges along a continuum.<br />
4pSC22. Indexical cues to talker sex and individual talker identity<br />
extracted from vowels produced in sentence-length utterances.<br />
Michael J. Berkowitz �Dept. of Psych., 301 Wilson Hall, Vanderbilt<br />
Univ., 111 21st Ave. South, Nashville, TN 37203,<br />
michael.j.berkowitz@vanderbilt.edu�, Jo-Anne Bachorowski �Vanderbilt<br />
Univ., Nashville, TN 37203�, and Michael J. Owren �Georgia State Univ.,<br />
Atlanta, GA 30302�<br />
The purpose of this study was to replicate and extend a previous study<br />
of indexical cuing �J.-A. Bachorowski and M. J. Owren, J. Acoust. Soc.<br />
Am. 106, 1054–1063 �1999�� by including more vowel sounds spoken in<br />
more diverse contexts. Specific combinations of acoustic parameters that<br />
should represent talker sex and individual talker identity were identified<br />
using predictions based on known sources of variation in vocal<br />
production-related anatomy. This study utilized 100 recordings of<br />
sentence-length utterances, produced by each of 43 male and 44 female<br />
undergraduates, as well as 22 stock-phrase recordings produced by these<br />
same participants. One of five vowel sounds �/æ, }, i,., u/� was isolated<br />
from each sentence and analyzed for F 0, F1, F2, F3, F4, vowel duration,<br />
jitter, shimmer, and harmonicity. Classification by talker sex was nearly<br />
perfect using a combination of cues related to both vocal-fold and vocaltract<br />
anatomy. The accuracy of classification by individual identity depended<br />
strongly on cues relating to vocal tract-variation within sex.<br />
4pSC23. Utterance-final position and projection of femininity in<br />
Japanese. Mie Hiramoto and Victoria Anderson �Dept. of Linguist.,<br />
Univ. of Hawaii, 1890 East-West Rd., Honolulu, HI 96822�<br />
Japanese female speakers frequent use of suprasegmental features such<br />
as higher pitch, longer duration, wider pitch range, and more instances of<br />
rising intonation vis-a-vis male speakers, is recognized as Japanese womens<br />
language �JWL� prosody. However, these features normally co-occur<br />
with gender-specific sentence-final particles �SFPs� like the strongly feminine<br />
‘‘kashira.’’ In this study, we examined the use of pitch and duration in<br />
utterances without SFPs, to determine whether JWL prosody is a function<br />
of SFPs or of utterance-final position. Eight male and eight female native<br />
Japanese speakers were instructed to read prepared sentences as though<br />
auditioning for a masculine theater role and then as though auditioning for<br />
a feminine role. Results indicate that utterance-final position is the projection<br />
point of JWL prosody even in the absence of SFPs. The data used for<br />
this study show high pitch, wide pitch range, long duration, and rising<br />
intonation at utterance-final positions when produced �by both men and<br />
women� in the feminine gender role. Conversely, in the masculine gender<br />
3294 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3294
ole, both men and women avoid the use of such prosodic features, and<br />
may even avoid using rising intonation in interrogative sentences, where<br />
the tonal grammar calls for it.<br />
4pSC24. Attitudinal correlate of final rise-fall intonation in Japanese.<br />
Toshiyuki Sadanobu �Kobe Univ., Tsurukabuto 1-2-1, Nada, Kobe,<br />
657-8501, Japan�<br />
Abrupt rise and subsequent fall intonation is common at the end of<br />
intonation units in Japanese, but its attitudinal correlate has not been fully<br />
elucidated yet. This intonation appears in the literature of the 1960’s as<br />
politicians’ way of speech, and nowadays not only politicians but many<br />
speakers including older generations often use it. However, this intonation<br />
is stigmatized as childish, and many people devalue it as an unintelligent<br />
way of speaking by young people. Where does this great gap between<br />
reality and image of this intonation come from? This presentation addresses<br />
this problem by focusing on natural conversation of Japanese daily<br />
life. The conclusions are as follows: �i� Rise-fall intonation often appears<br />
when the speaker talks about high-level knowledge, whereas it disappears<br />
when the speaker talks about their personal experience. �ii� Rise-fall intonation<br />
at the end of an intonation conveys the speaker’s being so occupied<br />
with speaking that intonation unit. The childish image comes from the<br />
speaker’s unawareness of their overall speech because of being occupied<br />
with local process. �Work supported by the Ministry of Education, Science,<br />
Sport, and Culture, Grant-in-Aid for Scientific Research �A�,<br />
1620<strong>2006</strong>, and by the Ministry of Internal Affairs and Communications,<br />
SCOPE 041307003.�<br />
4pSC25. Vowel devoicing in Japanese infant- and adult-directed<br />
speech. Laurel Fais, Janet Werker �Dept. of Psych., Univ. of BC, 2136<br />
West Mall, Vancouver, BC V6T 1Z4 Canada, jwlab@psych.ubc.ca�,<br />
Sachiyo Kajikawa, and Shigeaki Amano �NTT Commun. Sci. Labs.,<br />
Seika-cho, Soraku-gun, Kyoto 619-0237 Japan�<br />
It is well known that parents make systematic changes in the way they<br />
speak to infants; they use higher pitch overall, more pronounced pitch<br />
contours, more extreme point vowels, and simplified morphology and syntax<br />
�Andruski and Kuhl, 1996; Fernald et al., 1989�. Yet, they also preserve<br />
information crucial to the infants ability to acquire the phonology of<br />
the native language �e.g., phonemic length information, Werker et al.,<br />
<strong>2006</strong>�. The question examined in this paper is whether information other<br />
than phonemic segmental information is also preserved, namely, information<br />
concerning the phonological process of vowel devoicing. Devoicing<br />
of high vowels between voiceless consonants and word-finally after a<br />
voiceless consonant is a regular and well-attested phonological process in<br />
Japanese �Shibatani, 1990�. A corpus of speech by Japanese mothers addressed<br />
to their infants and addressed to another adult was examined, and<br />
the degree and frequency with which they apply vowel devoicing in each<br />
type of speech was analyzed. Rates of vowel devoicing in speech to adults<br />
and infants are compared, accommodations made to infants and to<br />
hearing-impaired children are discussed �Imaizumi et al., 1995�, and the<br />
implications of these accommodations for the acquisition of vowel devoicing<br />
by Japanese infants are explored.<br />
4pSC26. Language and gender differences in speech overlaps in<br />
conversation. Jiahong Yuan, Mark Liberman, and Christopher Cieri<br />
�Univ. of Pennsylvania, Philadelphia, PA 19104�<br />
Language and gender differences in speech overlaps in conversation<br />
were investigated, using the LDC CallHome telephone speech corpora of<br />
six languages: Arabic, English, German, Japanese, Mandarin, and Spanish.<br />
To automatically obtain the speech overlaps between two sides in a conversation,<br />
each side was segmented into pause and speaking segments, and<br />
the overlap segments during which both sides were speaking were time<br />
stamped. Two types of speech overlaps are distinguished: �1� One side<br />
takes over the turn before the other side finishes �turn-taking type�. �2�<br />
One side speaks in the middle of the other side’s turn �backchannel type�.<br />
It was found that Japanese conversations have more short �less than 500<br />
ms� turn-taking type of overlap segments than the other languages. The<br />
average number of such segments per 10 min of conversation for Japanese<br />
was 62.6 whereas the average numbers for the other languages ranged<br />
from 37.9 to 43.3. There were, however, no significant differences among<br />
the languages on the backchannel type of overlaps. Cross-linguistically,<br />
conversations between two females contained more speech overlaps �both<br />
types� than those between a male and a female or between two males, and<br />
there was no significant difference between the latter two.<br />
4pSC27. An acoustic study of laringeal contrast in three American<br />
Indian Languages. Heriberto Avelino �Dept. of Linguist., UC Berkeley,<br />
Berkeley, CA 94720-2650�<br />
A contrast between modal and nonmodal phonation is commonly<br />
found in American Indian languages. The use of laryngealized voice has<br />
been reported in a number of languages from different linguistic families.<br />
This paper investigates the acoustics of laryngealized phonation in three<br />
indigenous languages spoken in Mexico, Yalalag Zapotec, Yucatec Maya,<br />
and Mixe. These languages differ in terms of the use of other features<br />
controlled by action of the larynx, i.e., tone. In Zapotec there is a contrast<br />
between high, low, and falling tones; Maya has phonemic high and low<br />
tones, whereas Mixe does not present phonemic pitch. The results show<br />
that the production of phonemic laryngeal vowels differs from language to<br />
language. The data suggest that the specific implementation of laryngealization<br />
depends in part on the relationship with contrastive tone. The patterns<br />
of the languages investigated provide new evidence of the possible<br />
synchronization of phonation throughout the vowel. With this evidence, a<br />
typology of modal/nonmodal phonation in phonation-synchronizing languages<br />
is proposed.<br />
4pSC28. The comparison between Thai and Japanese temporal<br />
control characteristics using segmental duration models.<br />
Chatchawarn Hansakunbuntheung and Yoshinori Sagisaka �GITI, Waseda<br />
Univ., 29-7 Bldg. 1-3-10, Nishi-Waseda, Shinjuku-ku, Tokyo 169-0051,<br />
Japan, chatchawarnh@fuji.waseda.jp�<br />
This paper compares the temporal control characteristics between Thai<br />
and Japanese read speech data using segmental duration models. The same<br />
and the different control characteristics have been observed from phone<br />
level to sentence level. The language-dependent and language-independent<br />
control factors have also been observed. In phone and neighboring phone<br />
level, different characteristics are found. Japanese vowel durations are<br />
mainly compensated by only adjacent preceding and following phones,<br />
which results from mora timing. Unlike Japanese, Thai vowel durations<br />
are affected by two succeding phones. It can be guessed that the differences<br />
come from syllabic structures. In word level, most content words<br />
tend to have longer phone durations while function words have shorter<br />
ones. In phrase level, both languages express duration lengtening of<br />
syllable/mora at the phrase initial and final. For language-specific factors,<br />
Thai tones express small alteration on phone duration. The comparisions<br />
explore the duration characteristics of the languages and give more understanding<br />
to be used in speech synthesis and second-language learning<br />
research. �Work supported in part by Waseda Univ. RISE research project<br />
of ‘‘Analysis and modeling of human mechanism in speech and language<br />
processing’’ and Grant-in-Aid for Scientific Research A-2, No. 16200016<br />
of JSPS.�<br />
4pSC29. Articulatory settings of French and English monolinguals<br />
and bilinguals. Ian L. Wilson �Univ. of Aizu, Tsuruga, Ikki-machi,<br />
Aizu-Wakamatsu City, Fukushima, 965-8580, Japan, wilson@u-aizu.ac.jp�<br />
and Bryan Gick �Univ. of BC, Vancouver, BC V6T1Z1 Canada�<br />
This study investigated articulatory setting �AS�, a language’s underlying<br />
posture of the articulators. Interspeech posture �ISP� of the articulators<br />
�their position when motionless during interutterance pauses� was<br />
used as a measure of AS in Canadian English and Quebecois French.<br />
3295 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3295<br />
4p FRI. PM
Optotrak and ultrasound imaging were used to test whether ISP is language<br />
specific in monolingual and bilingual speakers. Results show significant<br />
differences in ISP across the monolingual groups, with English<br />
exhibiting a higher tongue tip, more protruded upper and lower lips, and<br />
narrower horizontal lip aperture. Results also show that upper and lower<br />
lip protrusion are greater for the English ISP than for the French ISP, in all<br />
bilinguals who were perceived as native speakers of both languages, but in<br />
no other bilinguals. Tongue tip height results mirror those of the monolingual<br />
groups, for half of the bilinguals perceived as native speakers of both<br />
languages. Finally, results show that there is no unique bilingual-mode<br />
ISP, but instead one that is equivalent to the monolingual-mode ISP of a<br />
bilingual’s currently most-used language. This research empirically confirms<br />
centuries of noninstrumental evidence for AS, and for bilinguals it<br />
suggests that differences between monolingual and bilingual modes do not<br />
hold at the phonetic level.<br />
4pSC30. Temporal and spectral variability of vowels within and<br />
across languages with small vowel inventories: Russian, Japanese,<br />
and Spanish. Franzo F. LawII �Speech Acoust. and Percept. Lab., City<br />
Univ. of New York—Grad. Ctr., 365 Fifth Ave., New York, NY<br />
10016-4309, flaw@gc.cuny.edu�, Yana D. Gilichinskaya, Kikuyo Ito,<br />
Miwako Hisagi, Shari Berkowitz, Mieko N. Sperbeck, Marisa<br />
Monteleone, and Winifred Strange �City Univ. of New York—Grad. Ctr.,<br />
New York, NY 10016-4309�<br />
Variability of vowels in three languages with small vowel inventories<br />
�Russian, Japanese, and Spanish� was explored. Three male speakers of<br />
each language produced vowels in two-syllable nonsense words �VCa� in<br />
isolation and three-syllable nonsense words �gaC1VC2a� embedded within<br />
carrier sentences in three contexts: bilabial stops in normal rate sentences<br />
and alveolar stops in both normal and rapid rate sentences. Dependent<br />
variables were syllable duration and formant frequency at syllable midpoint.<br />
Results showed very little variation across consonant and rate conditions<br />
in formants for /i/ in Russian and Japanese. Japanese short /u, o, a/<br />
showed fronting �F2 increases� in alveolar context, which was more pronounced<br />
in rapid sentences. Fronting of Japanese long vowels was less<br />
pronounced. Japanese long/short vowel ratios varied with speaking style<br />
�isolation versus sentences� and speaking rate. All Russian vowels except<br />
/i/ were fronted in alveolar context, but showed little change in either<br />
spectrum or duration with speaking rate. Spanish showed a strong effect of<br />
consonantal context: front vowels were backed in bilabial context and<br />
back vowels were fronted in alveolar context, also more pronounced in<br />
rapid sentences. Results will be compared to female productions of the<br />
same languages, as well as American English production patterns.<br />
4pSC31. Does infant-directed speech in Tagalog resemble infantdirected<br />
speech in American English? Emmylou Garza-Prisby, Shiri<br />
Katz-Gershon, and Jean Andruski �Aud. & Speech-Lang. Pathol. Dept.,<br />
Wayne State Univ., 207 Rackham, 60 Farnsworth St., Detroit, MI 48202�<br />
This study compared the speech of a Filipino mother in infant- and<br />
adult-directed speech in order to investigate whether the mother used the<br />
acoustic features typically found in the infant-directed speech of American<br />
English-speaking mothers. Little acoustic documentation is available on<br />
the acoustic features of Tagalog, and no acoustic comparison of speech<br />
registers has so far been conducted. Impressionistically, Tagalog-speaking<br />
mothers’ do not appear to use the features typically found in American<br />
mothers speech to their young infants. The mother was recorded talking to<br />
her infant and to another adult native speaker of Tagalog. Recordings were<br />
made in the mother’s home and visits occurred during the first 6 months of<br />
the infant’s life. Specific acoustic features analyzed include �a� vowel<br />
duration, �b� vowel format frequencies, �c� vowel triangle size, �d� rate of<br />
speech, �e� fundamental frequency, and �f� F0 range. Morphological and<br />
syntactic features were also analyzed, including �g� mean length of utterance<br />
and �h� sentence length. Results support a need for further study of<br />
speech registers in Filipino mothers.<br />
4pSC32. Restricting relativized faithfulness and local conjunction in<br />
optimality theory. Haruka Fukazawa �Keio Univ., 4-1-1 Hiyoshi,<br />
Kohoku-ku, Yokohama, Japan�<br />
Within the framework of optimality theory �OT�, the two mechanisms,<br />
relativized faithfulness and local conjunction, play inevitable roles, especially<br />
when a simple constraint ranking fails to account for the data. However,<br />
their domain of application are too unrestricted and even overlapping<br />
each other. For instance, there are some cases which could be explained<br />
both by the ranking with relativized faithfulness and by the one with local<br />
conjunction. Syllable-final neutralization in German and geminate devoicing<br />
in Japanese loanwords are of interest in this context. The present paper<br />
proposes formal restrictions mostly on the local conjunction mechanism:<br />
the soundness of constraint combination, the number of constraints involved<br />
in a conjunction, and the local domain of conjunction. They not<br />
only can simplify the analysis but also give a more principled account for<br />
the overlapping cases. In fact, relativized faithfulness approach wins over<br />
local conjunction approach both in German neutralization and in Japanese<br />
loanwords. It is desirable for the universal grammar to be more restricted.<br />
Removing an overlap of theoretical devices is an important step toward<br />
the goal.<br />
3296 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3296
FRIDAY AFTERNOON, 1 DECEMBER <strong>2006</strong> KAUAI ROOM, 1:00 TO 2:55 P.M.<br />
Session 4pUWa<br />
Underwater Acoustics: Session in Honor of Leonid Brekhovskikh II<br />
William A. Kuperman, Cochair<br />
Scripps Inst. of Oceanography, Univ. of California, San Diego, Marine Physical Lab., La Jolla, CA 92093-0238<br />
Oleg A. Godin, Cochair<br />
NOAA, Earth System Research Lab., 325 Broadway, Boulder, CO 80305-3328<br />
Invited Papers<br />
1:00<br />
4pUWa1. Underwater noise as source of information on conditions and dynamics of ocean environments. Alexander V.<br />
Furduev �N. N. Andreyev Acoust. Inst., 4 Shvernika St., Moscow 117036, Russia�<br />
Leonid Brekhovskikh wrote in his book The Ocean and the Human �1987�: ‘‘Ocean noise is as important oceanographic parameter<br />
as temperature, current, and wind.’’ Brekhovskikh created and headed the Laboratory of Acoustic Methods of Ocean Research in 1956.<br />
One of the scientific directions of the Laboratory was investigation of underwater noise both as interference for sound reception and<br />
a source of environmental information. Long-term studies on the unique acoustic research vessels created under the initiative of<br />
Brekhovskikh resulted in numerous important findings, including ambient noise spectra and envelopes of acoustic fluctuations, depth<br />
dependence of noise directivity, and mechanisms of ambient noise generation. Brekhovskikh was always eager to find practical<br />
applications of scientific results. Different methods of ensuring noise immunity of hydroacoustic arrays were developed under his<br />
supervision. Passive methods of acoustic navigation based on use of natural noise were suggested. Techniques for underwater acoustic<br />
monitoring of the ocean based either on ambient noise analysis or reemission of noise from a point away from the receiving system<br />
have been developed. The success of the team of scientists headed by Brekhovskikh was determined by the creative atmosphere<br />
around him: there was neither competition nor commercial interests. The common goal was knowledge of the ocean.<br />
1:20<br />
4pUWa2. Distributed acoustic sensing in shallow water. Henrik Schmidt �Ctr. for Ocean Eng., MIT, Cambridge, MA 02139�<br />
The significance of Leonid Brekhovskikh to the ocean acoustics is undisputed. He was pioneering not only in terms of fundamental<br />
understanding of the ocean acoustic waveguide, but also the development of efficient and numerically stable approaches to<br />
propagation of sound in a stratified ocean. As such he has been an inspiration to a whole generation of model developers, leading to<br />
today’s suite of extremely powerful wave theory models, capable of accurately representing the complexity of the shallow-water ocean<br />
waveguide physics. The availability of these computational tools have in turn led to major advances in adaptive, model-based<br />
signal-processing techniques. However, such computationally intensive approaches are not necessarily optimal for the next generation<br />
of acoustic sensing systems. Thus, ocean observation in general is currently experiencing a paradigm shift away from platform-centric<br />
sensing concepts toward distributed sensing systems, made possible by recent advances in underwater robotics. In addition to a fully<br />
autonomous capability, the latency and limited bandwidth of underwater communication make on-board processing essential for such<br />
systems to be operationally feasible. In addition, the reduced sensing capability of the smaller physical apertures may be compensated<br />
by using mobility and artificial intelligence to dynamically adapt the sonar configuration to the environment and the tactical situation,<br />
and by exploiting multiplatform collaborative sensing. The development of such integrated sensing and control concepts for detection,<br />
classification, and localization requires extensive use of artificial intelligence incorporating a fundamental understanding of the ocean<br />
acoustic waveguide. No other sources in literature provide this with the clarity and depth that is the trademark of Academician<br />
Brekovskikh’s articles and classical textbooks. �Work supported by ONR.�<br />
1:40<br />
4pUWa3. When the shear modulus approaches zero: Fluids don’t<br />
bend and Scholte leaves the <strong>room</strong>. Robert I. Odom �Appl. Phys. Lab.,<br />
Univ. of Washington, 1013 NE 40th St., Seattle, WA 98105-6698�, Donna<br />
L. G. Sylvester, and Caitlin P. McHugh �Seattle Univ., Seattle, WA<br />
98122-1090�<br />
The 4�4 linear system of differential equations describing the propagation<br />
of the displacements and tractions in an elastic layered medium<br />
becomes singular as the shear modulus of the elastic medium approaches<br />
zero. There are a number of approximate ways to handle this singularity in<br />
Contributed Papers<br />
order to impart numerical stability to the computation of the elastic waves<br />
in a layered medium. For example, one can impose an irrotational constraint<br />
on the displacements or introduce a massive elastic interface �MEI�.<br />
Both of these ways of handling the weak shear strength are approximate,<br />
but avoid the need for singular perturbation theory �Gilbert, 1998�. Scholte<br />
waves are interface waves that propagate along the interface between an<br />
elastic solid and a fluid. They have nodes near or on the interface and<br />
decay exponentially into the bounding media. Scholte waves do not occur<br />
at the boundary between fluids. As the shear speed in the bounding elastic<br />
medium approaches zero, the Scholte waves disappear from the spectrum.<br />
We investigate this disappearance by applying singular perturbation theory<br />
3297 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3297<br />
4p FRI. PM
to the coupled fluid-elastic system. Among other things, we will address<br />
the rate in wave-number space at which the Scholte waves disappear from<br />
the spectrum.<br />
1:55<br />
4pUWa4. Measurement of the plane-wave reflection coefficient of the<br />
ocean bottom and the legacy of Leonid Brekhovskikh. George V.<br />
Frisk �Florida Atlantic Univ., 101 N. Beach Rd., Dania Beach, FL 33004�<br />
and Luiz L. Souza �Bose Corp., Framingham, MA 01701�<br />
Leonid Brekhovskikh’s classic text �Waves in Layered Media �Academic,<br />
New York, 1980�� inspired the development of several techniques<br />
for measuring the plane-wave reflection coefficient of the ocean bottom.<br />
Specifically, his application of the geometrical acoustics approximation to<br />
the problem of reflection of a spherical wave from a horizontally stratified<br />
medium provided the theoretical foundation for evaluating the strengths<br />
and weaknesses of various measurement methods. The most popular<br />
method assumes that the reflected field also consists of a spherical wave<br />
multiplied by the reflection coefficient evaluated at the specular angle.<br />
However, Brekhovskikh’s work showed that this interpretation is confined<br />
to a limited range of angles and bottom structures and, if applied improperly,<br />
can lead to unphysical results such as negative bottom loss. This<br />
paper describes a technique which circumvents these deficiencies. It consists<br />
of measuring the pressure field magnitude and phase versus range due<br />
to a cw point source and Hankel transforming these data to obtain the<br />
depth-dependent Green’s function versus horizontal wavenumber. The reflection<br />
coefficient is then obtained from the Green’s function using the<br />
analytical relationship between the two quantities. The method is demonstrated<br />
using 220-Hz data obtained in a near-bottom geometry in the Icelandic<br />
Basin. �Work supported by ONR.�<br />
2:10<br />
4pUWa5. Field from a point source above a layered half-space; theory<br />
and observations on reflection from the seabed. Charles W. Holland<br />
�Appl. Res. Lab., The Penn State Univ., State College, PA�<br />
L. M. Brekovskikh’s book Waves in Layered Media has provided decades<br />
of graduate students and researchers alike with a comprehensive and<br />
enormously useful reference. One topic from that work, reflection from a<br />
point source above a plane-layered medium �the seabed�, is discussed<br />
here. Both theoretical underpinnings and observations of reflection from a<br />
homogeneous halfspace, a transition layer with smoothly varying density<br />
and velocity profiles, and discrete layered media are considered for various<br />
shallow water sediment fabrics. �Work supported by the Office of<br />
Naval Research and NATO Undersea Research Centre.�<br />
2:25<br />
4pUWa6. Plane-wave model and experimental measurements of the<br />
directivity of a Fabry-Perot, polymer film, ultrasound sensor.<br />
Benjamin T. Cox and Paul C. Beard �Dept. of Med. Phys. and<br />
Bioengineering, Univ. College London, Gower St., London, WC1E 6BT,<br />
UK, bencox@mpb.ucl.ac.uk�<br />
Optical detection of ultrasound is popular due to the small element<br />
sizes that can be achieved. One method exploits the thickness change of a<br />
Fabry-Perot �FP� interferometer caused by the passage of an acoustic wave<br />
to modulate a laser beam. This detection method can have greater sensitivity<br />
than piezoelectric detectors for sub-millimeter element sizes. The<br />
directivity of FP sensors and the smallest achievable effective element size<br />
are examined here. A plane-wave model of the frequency-dependent directional<br />
response of the sensor, based on Brekhovskikh’s work on elastic<br />
waves in layered media, is described and validated against experimental<br />
directivity measurements made over a frequency range of 15 MHz and<br />
from normal incidence to 80 deg. In terms of applications, the model may<br />
be used to provide a noise-free response function that can be deconvolved<br />
from sound field measurements in order to improve accuracy in highfrequency<br />
metrology and imaging applications, or, for example, as a predictive<br />
tool to improve sensor design. Here, the smallest achievable effective<br />
element radius was investigated by comparing the directivity with that<br />
of a rigid circular pressure transducer, and found to be �0.9d, where d is<br />
the thickness of the FP interferometer. �Funding was provided by the<br />
EPSRC, UK�<br />
2:40<br />
4pUWa7. The interference head wave and its parametric dependence.<br />
Jee Woong Choi and Peter H. Dahl �Appl. Phys. Lab., Univ. of<br />
Washington�<br />
The interference head wave that can exist in the presence of a soundspeed<br />
gradient in the sediment, is a precursor arrival representing a transition<br />
between the first-order head wave and the zeroth-order refracted<br />
wave. Using a parabolic equation �PE� simulation, Choi and Dahl �J.<br />
Acoust. Soc. Am. 119, 2660–2668 �<strong>2006</strong>�� showed how the small shift in<br />
the dominant frequency of the interference head wave behaves as a function<br />
of the nondimensional parameter zeta, which itself is a function of<br />
center frequency, gradient, and range. For example, it was shown that the<br />
maximum frequency shift occurring in the vicinity of zeta equals 2. In this<br />
work, we investigate the amplitude and additional spectral properties of<br />
the interference head wave and analyze the cause of the frequency shift<br />
phenomenon using the ray theory. The limitation on the application of ray<br />
method also will be discussed. Finally, the conclusion will be verified by<br />
the time-dependent simulation using the RAM PE algorithm. �Work supported<br />
by the ONR.�<br />
3298 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3298
FRIDAY AFTERNOON, 1 DECEMBER <strong>2006</strong> KAUAI ROOM, 3:10 TO 4:25 P.M.<br />
Session 4pUWb<br />
Underwater Acoustics: Session in Honor of Fredrick Fisher<br />
William A. Kuperman, Chair<br />
Scripps Inst. of Oceanography, Univ. of California, San Diego, Marine Physical Lab., La Jolla, CA 92093-0238<br />
Chair’s Introduction—3:10<br />
Invited Papers<br />
3:15<br />
4pUWb1. FLIP „Floating Instrument Platform…: A major Fred Fisher contribution to ocean science. Fred N. Spiess, Robert<br />
Pinkel, William S. Hodgkiss, John A. Hildebrand �Marine Physical Lab., Scripps Inst. of Oceanogr., UCSD 0205, 9500 Gilman Drive;<br />
La Jolla, CA 92093-0205, fspiess@ucsd.edu�, and Gerald L. D’Spain �UCSD, La Jolla, CA 92093-0205�<br />
Frederick H. Fisher, a loyal and zealous member of the Acoustical Society of America, was an imaginative and effective developer<br />
of new techniques for research in both laboratory and seagoing acoustics. Most notable among his contributions was his work in<br />
bringing into being and enhancing the usefulness of the spar buoy laboratory, FLIP, from its inception in 1960. Not only did Fred use<br />
FLIP in his own research, its existence and many of its ancillary capabilities constituted a base for the seagoing research of others. The<br />
authors of this paper have benefited from FLIP’s unique capabilities, starting with long-range sound propagation studies in the 1960’s<br />
and 1970’s. FLIP’s stability and deep draft structure provided the platform for development of acoustic Doppler techniques for the<br />
measurement of ocean currents. Most recently, FLIP has been involved in studies of marine mammal vocalizations and use of<br />
multielement arrays to investigate details of shallow-water propagation. Fred’s initial studies of sonar bearing accuracy, for which<br />
FLIP’s construction was funded, and his dedication to advancing FLIP’s ability to contribute to ocean science, constitute a legacy that<br />
is being utilized today, more than 40 years after FLIP’s launching.<br />
3:35<br />
4pUWb2. Absorption of sound in seawater and ocean ambient noise, the scientific passions of Fred Fisher. John A. Hildebrand<br />
�Scripps Inst. of Oceanogr., UCSD, La Jolla, CA 92093-0205�<br />
Fred Fisher made seminal contributions to ocean acoustics in the understanding of the absorption of sound in seawater and ocean<br />
ambient noise. Laboratory data and long-range sound propagation data revealed excess acoustic absorption in seawater. Fred Fisher<br />
spent much of his scientific career, beginning with his Ph.D dissertation, teasing out the contributions of various seawater components<br />
to sound absorption, and his work on this topic set the standard for understanding and modeling these phenomena. Ambient noise is<br />
an important aspect of underwater signal detection and is the focus of recent concerns about disturbance of marine organisms. Fred<br />
Fisher made important contributions to ambient noise studies by conducting measurements of vertical directionality, thereby testing<br />
models for ambient noise production. The value of archival ambient noise data and recent increases in ambient noise will be discussed.<br />
3:55<br />
4pUWb3. Fred Fisher’s high-pressure work with eyewash and epsom<br />
salts. Christian de Moustier �Ctr. for Coastal & Ocean Mapping, Chase<br />
Ocean Eng. Lab, Univ. of New Hampshire, 24 Colovos Rd., Durham, NH<br />
03824�<br />
Starting in 1957 Fred Fisher led research programs devoted to highpressure<br />
measurements related to the physical chemistry of sound absorption<br />
in seawater due to magnesium sulfate and other salts. As he put it, he<br />
spent his professional lifetime squeezing epsom salt. His interest in the<br />
low-frequency anomalous sound absorption in the ocean below 1 kHz led<br />
to the discovery of boric acid as the cause of the low-frequency relaxation.<br />
This paper is a short review of Fred Fisher’s contributions to our knowledge<br />
of sound absorption in seawater, based in part on his carefully handwritten<br />
lecture notes and numerous low-pressure discussions.<br />
Contributed Papers<br />
4:10<br />
4pUWb4. Fred Fisher and research with acoustic vector sensors;<br />
Marine Physical Laboratory’s vertical array of directional acoustic<br />
receivers and ocean noise. Gerald L. D’Spain and William S. Hodgkiss<br />
�Marine Physical Lab, Scripps Inst. of Oceanogr., La Jollla, CA<br />
93940-0701�<br />
Fred Fisher had boundless enthusiasm for all topics acoustic. A chance<br />
encounter with him in the hallway usually led to a half-hour discussion of<br />
the latest research efforts at the lab and recent results he found exciting. In<br />
the 1980s, Fred became interested in the problem of identifying the physical<br />
phenomena forming the pedestal about the horizontal in vertical directionality<br />
measurements of the deep ocean’s low-frequency noise field. Two<br />
competing mechanisms had been proposed: downslope conversion of<br />
coastal shipping and noise from high latitude winds coupling into the deep<br />
3299 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3299<br />
4p FRI. PM
sound channel due to the shoaling of the sound channel axis. The relative<br />
contributions of these two mechanisms possibly could be separated if the<br />
azimuthal ambiguity of a vertical line array of hydrophones somehow<br />
could be broken. Therefore, Fred proposed to build a vertical array of<br />
‘‘DIFAR’’ sensors, which led to the design and construction of the Marine<br />
Physical Lab’s Vertical ‘‘DIFAR’’ Array. This talk will reminisce a bit<br />
about Fred as well as present some results from an ambient noise experiment<br />
conducted in 1992 on the continental shelf using the Vertical DIFAR<br />
Array co-deployed with MPL’s freely drifting vector sensors, the Swallow<br />
floats. �Work supported by ONR and ONT.�<br />
FRIDAY EVENING, 1 DECEMBER <strong>2006</strong> HAWAII BALLROOM, 7:00 TO 10:00 P.M.<br />
Awards Ceremony<br />
Anthony A. Atchley, President<br />
Acoustical Society of America<br />
Yôiti Suzuki, President<br />
Acoustical Society of Japan<br />
Acknowledgment of Honolulu Local Meeting Organizing Committees<br />
Presentation of Fellowship Certificates<br />
Anders Askenfeldt James A. McAteer<br />
Sergio Beristain David R. Palmer<br />
Philippe Blanc-Benon Marehalli G. Prasad<br />
David A. Conant Hiroshi Riquimaroux<br />
Andes C. Gade Peter A. Rona<br />
Anthony W. Gummer Mark V. Trevorrow<br />
Charles W. Holland Michael Vörlander<br />
Jody E. Kreiman Joos Vos<br />
Kevin D. LePage Ben T. Zinn<br />
Science Writing Award in Acoustics for Journalists to Radek Boschetty<br />
Science Writing Award for Professionals in Acoustics to Edwin Thomas<br />
Announcement of 2005 A. B. Wood Medal and Prize to Aaron Thode<br />
Distinguished Service Citation to Thomas D. Rossing<br />
Silver Medal in Noise to Alan H. Marsh<br />
Silver Medal in Physical Acoustics to Henry E. Bass<br />
Silver Medal in Psychological and Physiological Acoustics to William A. Yost<br />
Wallace Clement Sabine Award to William J. Cavanaugh<br />
Recognition of Acoustical Society of Japan meeting organizers<br />
Recognition of Acoustical Society of America meeting organizers<br />
3300 J. Acoust. Soc. Am., Vol. 120, No. 5, Pt. 2, November <strong>2006</strong> Fourth Joint Meeting: ASA and ASJ<br />
3300