Sign up for our monthly newsletter so you never miss the latest from InDepth!
by Reilly Fogarty
Header photo courtesy of DAN.
The pool of people who explore the ocean depths beyond 122 m/400 ft is small, and the group of people who do it regularly and need to reach those depths quickly is microscopic. This niche application coupled with a significant fire hazard make it easy to understand why exotic gases like hydrogen have escaped both common use and public interest.
Despite the obvious concerns, however, hydrogen has shown some capacity to ameliorate the effects of high pressure nervous syndrome (HPNS) in deep divers, improve function at extreme depth, reduce work of breathing, and present a possible alternative to helium in case of helium reserve depletion. Use of the gas has made it possible to dive deeper, get to depth faster, and stay there longer, but there is a substantial risk-versus-reward calculation to be made before considering its use. Interested in diving deep and dabbling in the cutting edge of diving research? Here’s what we know about using hydrogen as a diving gas.
High Pressure Nervous Syndrome
Records of sojourns into the use of hydrogen as a breathing gas go back to as early as the 18th century with the experiments of Antione Lavoisier, but the use of the gas came to a head during the heyday of deep diving research in the late 20th century. From the Atlantis dives at Duke University to the Hydra Missions and the evolution of the Compagnie Maritime d’Expertises (COMEX) tables, researchers and deep divers quickly found issues as they pushed to explore deeper depths. Racing past 184 m/600 ft, researchers discovered that divers would face several new and potentially deadly phenomena in their push to the bottom. Chief among these was a condition that came to be called high pressure nervous syndrome. The condition was first described as “helium tremors” by Russian researcher G. L. Zal’tsman in 1961 and Peter B. Bennett in 1965 (Zal’tsman’s research was not available outside of the Soviet Union until 1967).
Later this condition would be named and correlated with symptoms including tremors, headache, dizziness, fatigue, myoclonic jerking, muscular weakness, and euphoria. Gastrointestinal complaints, memory and cognitive deficits, psychomotor impairment, nightmares, and somnolence are also possible, and convulsions have been noted in animal models (Kangal, 2019). The mechanism of HPNS has not yet been proven, but there are several working theories.
The first of these involves the compression of lipid components of the cell membranes in the central nervous system (CNS). Some theorize that the compression of these tissues may affect the transmembrane proteins, ion channels, and surface receptors critical to the signaling pathways of the CNS (Talpalar, 2007). Many connect the use of hydrogen to mitigate the effects of HPNS to this model. Some research has shown that anesthetic gases can reduce the effects of HPNS, and this has been proposed to be driven by some pressure reversal effect of narcosis (Kot, 2012). Hydrogen is a narcotic gas, and this may be one component of its ability to reduce the onset or severity of HPNS. This mechanism, the compression of lipid components also appears to be the one that initially gave rise to the use of breathing gases as a way to ameliorate the effects of HPNS, with some groundwork for this foundation laid out by Peter B. Bennett, PhD, as early as 1989 (Bennett, 1989).
Other researchers have focused on neurotransmitters, including gamma-aminobutyric acid (GABA), dopamine, serotonin, acetylcholine, and others. These models have also shown promise, one study showing a GABA increase in the cortex diminishing HPNS signs in baboon models and another showing NMDA antagonists preventing convulsions in rats (Pearce, 1989, 1991). Similar studies have been conducted with a range of neurotransmitters and neuronal calcium ion channels with similar results. In short, there are several potential avenues for the specific mechanism of HPNS, and while the general mechanism is likely a combination of several models, none have yet been definitively proven.
What we know is that there is notable variation in HPNS onset among divers (Bennett, 1989), and onset appears to be the result of a combination of breathing gas, compression rate, and depth. Faster compression, lower nitrogen, or hydrogen content in helium/oxygen breathing gases, and deeper depths have been correlated with more rapid onset and more severe symptoms (Jain, 1994).
The Benefits of Hydrogen
The use of hydrogen as a diving gas doesn’t just stem from its ability to reduce the onset of HPNS—it’s an extraordinarily light gas that’s useful in reducing work of breathing at extreme depth and a potential replacement for helium when worldwide demand has led to a quickly dwindling reserve. One U.S. Navy paper went so far as to propose hydrogen replacement of helium due to helium scarcity leading to a predicted depletion of supply by the year 2000 (Dougherty, 1965). Thankfully, that mark has come and gone due to the discovery of several new helium sources, but it’s not an unrealistic concern when the demand for helium is so high in manufacturing, aerospace, and technology.
The benefits of hydrogen are notable, but the hazards are nothing to balk at. Little research has been done on the decompression or thermal properties of hydrogen in divers, it’s reported to be mildly narcotic, and it’s highly flammable. While oxygen is an oxidizer that can feed a fire, hydrogen is actively flammable—in the presence of sufficient oxygen and a source of ignition, it will combust in dramatic fashion. In practice, this combustibility is managed by reducing both the sources of ignition and the available oxygen. With the lower explosive limit of hydrogen being around 4% by volume, using less than this amount in normoxic environments effectively mitigates the fire risk but does little for deep divers. Instead, extremely hypoxic gases and high concentrations of hydrogen and helium have been used with great success. The COMEX Hydra VIII mission, for example, used a mixture of 49% hydrogen, 50.2% helium, and 0.8% oxygen to take divers to a maximum depth of 536 m/1,752 ft.
The decompression profiles used in these deep saturation dives appear to be effective as well. As early as 1992, COMEX researchers found that teams of divers on a hydrogen mixture at a depth of 210 m/686 ft performed tasks more efficiently, both cognitively and physically, than their counterparts on helium (Offshore-mag.com, 1996). The same experiment resulted in a bubble study that showed “no evidence of bubbles” in the divers following decompression—an exercise in small sample sizes perhaps, but with promising results. (Note: Researchers suspected that “biochemical decompression” might be involved, i.e., a process in which metabolism of H2 by intestinal microbes facilitates decompression—ed.) U.S. Navy research found similar results, indicating that hydrogen increased the capacity for physical effort as a result of a decrease in work of breathing at depth (Dougherty, 1965).
Functionally, the benefits of the gas are hard to dispute; work of breathing is a constantly growing area of concern for dives at all depths, HPNS is a constant concern, and minimizing decompression is a perpetual goal. For divers reaching extreme depths without the ability to perform saturation dives, diving to depths beyond 122m/400 ft is a repetitive gamble with no guarantee of success. Rapid compression combined with limited options for gas mixes result in the need to play a dive profile by “feel” with emergency plans in place to respond to HPNS onset, and more than one diver, likely including Sheck Exley (See: “Examining Early Technical Diving Deaths,” by Michael Menduno, InDepth 2.2) has lost their life to the condition.
The Hazards of Hydrogen
By the same token, hydrogen presents unique hazards that require careful consideration. Unexplored decompression profiles and limited research on long-term effects make the decision to dive with hydrogen difficult, and the significant risk of fire places divers in more danger than is typically acceptable. Add to this the limited applications (either hydrogen content below 4% in normoxic environments or oxygen content below 6% in high-hydrogen environments), and it quickly becomes apparent why hydrogen hasn’t yet hit the mainstream.
The presence of project divers in our community performing near saturation dives with trimix and makeshift in-water habitats skews favor away from hydrogen as well, making it a gas viable only for the deepest dives without the option for saturation. The case for hydrogen isn’t entirely up in smoke, however. Research showing significant decompression benefits or depletion of helium reserves may well push us toward helium’s more flammable cousin, but it’s unlikely you’ll see hydrogen at your favorite fill station any time soon.
- Ozgok Kangal, M.K., & Murphy-Lavoie, H.M., (2019, November 14). Diving, High Pressure Nervous Syndrome. (. In: StatPearls StatPearls Publishing.
- Talpalar, A.E., (2007, Nov 16-30). High pressure neurological syndrome. Rev Neurol., 45(10), 631-6.
- Kot, J., (2012). Extremely deep recreational dives: the risk for carbon dioxide (CO2) retention and high pressure neurological syndrome (HPNS). Int Marit Health, 63(1), 49-55.
- Bennett, P.B., (1989). Physiological limitations to underwater exploration and work. Comp Biochem Physiol A Comp Physiol., 93(1), 295-300.
- Pearce, P.C., Clarke, D., Doré, C.J., Halsey, M.J., Luff, N.P., & Maclean, C.J., (1989. March). Sodium valproate interactions with the HPNS: EEG and behavioral observations. Undersea Biomed Res. 16(2), 99-113.
- Pearce, P.C., Halsey, M.J., MacLean, C.J., Ward, E.M., Webster, M.T., Luff, N.P., Pearson, J., Charlett, A., & Meldrum, B.S., (1991, July). The effects of the competitive NMDA receptor antagonist CPP on the high pressure neurological syndrome in a primate model. Neuropharmacology,30(7), 787-96.
- Jain, K.K. (1994, July). High-pressure neurological syndrome (HPNS). Acta Neurol. Scand.,90(1), 45-50.
- Dougherty, J., (1965). The Use of Hydrogen As An Inert Gas During Diving: Pulmonary Function During Hydrogen-Oxygen Breathing At Pressures Equivalent to 200 Feet of Sea Water.
- Saturation diving tests support claims for hydrogen breathing mix, (1996). Offshore-mag.com.
John Clarke, retired scientific director of the U.S. Navy Experimental Diving Unit, is knowledgeable about hydrogen diving and has used that knowledge both in his blogs as well as in the undersea sci-fi thrillers, the Jason Parker Trilogy , of which he is the author.
Fahlman, A., Lin, W.C., Whitman, W.B., & Kayar, S.R., (2002, November). Modulation of decompression sickness risk in pigs with caffeine during H2 biochemical decompression. JApp. Physio. 93(5).
Kayar, S.R., & Fahlman, A. (2001). Decompression sickness risk reduced by native intestinal flora in pigs after H2 dives. Undersea Hyperb Med., 28(2), 89-97.
Fahlman, A., Tikuisis, P., Himm, J.F., Weathersby, P.K., & Kayar, S,R., (2001, December). On the likelihood of decompression sickness during H2 biochemical decompression in pigs. J Appl Physiol., 91(6), 2720-9.
Reilly Fogarty is a team leader for risk mitigation initiatives at Divers Alert Network (DAN). When not working on safety programs for DAN, he can be found running technical charters and teaching rebreather diving in Gloucester, Mass. Reilly is a USCG licensed captain whose professional background includes surgical and wilderness emergency medicine as well as dive shop management.
Capturing 3D Audio Underwater
Listen up divers! Audio engineer turned techie, Symeon Manias is bringing 3D audio to the underwater world to provide a more immersive and convincing auditory experience to listeners whether its video, film, or virtual reality. So, pump up the volume—here’s what you need to know.
by Symeon Manias
Header image by Peter Gaertner/Symeon Manias and edited by Amanda White
Spatial hearing provides sound related information about the world. Where does a sound come from? How far away is it? In what space does the sound reside? Humans are pretty good at localizing sound sources and making sense of the world of sounds. The human auditory system performs localization tasks either by solely relying on auditory cues or by fusing them with visual, dynamic, or experience-based cues. Auditory cues consist of cues obtained by one ear, which is termed monaural, or by using both ears, termed binaural.
Humans can detect sounds of frequency content between 20 Hz and 20 kHz. That translates to wavelengths between 1.7 cm and 17 m. This frequency range decreases with age as the internal parts of the ear degrade; however, exposure to loud noises can accelerate this degradation. The peripheral part of each of the human ears, which by no coincidence is common to most mammals, consists of three structures. The external part of the ear, the middle ear, and the inner ear. Figure 1 shows in detail all these parts of the ear along with their basic structure
The external ear consists mainly of the pinna, the concha, and the ear canal and ends at the eardrum. The pinna and concha, which are visible as the outer part of the ear, are responsible for modifying the frequency spectrum of sounds coming from different directions. The ear canal, on the other hand, is responsible for enhancing specific frequencies of most incoming sounds. The ear canal can be seen as a resonant tube with a resonant frequency that depends on the length and diameter of the tube. For most humans, the resonant frequency of the ear canal is between 2 kHz and 4 kHz. This means that all sounds that travel through the ear canal will be enhanced in that frequency range. Airborne sounds travel from the world through the pinna, the concha, the ear canal, and hit the eardrum, which is at the end of the ear canal. The eardrum separates the outer from the middle ear and acts as a transformer that converts sounds to mechanical vibration.
On the other side of the eardrum, in the middle ear, there are three ossicles, the malleus, the incus and the stapes. These bones are responsible for transmitting the sounds from the outer ear through vibration to the entrance of the inner ear at the oval window. This is where mechanical vibration is transformed into sound to then be received by the sound receptors in the middle ear.
One might wonder why we even need the middle ear if the incoming sound is first transformed at the eardrum into mechanical vibration and then back to sound at the oval window. If the middle ear was absent, it is very likely that incoming sound would be reflected back at the eardrum. Another reason is that the middle ear generally acts as a dynamics processor; it has a protection functionality. When a loud sound travels through the ear canal and hits the eardrum, the muscles that are attached to the middle ear ossicles can attenuate the mechanical vibration. This is a phenomenon known in psychoacoustics as an acoustics reflex. The middle ear also serves as a means to reduce the transmission of internally generated sounds that are transmitted through bone conduction, as well as letting air be transmitted on the back side of the eardrum through the Eustachian tube, which balances the pressure that is building on the eardrum.
The inner ear is a very complex structure which is responsible for two functions: one is the vestibular system, which is responsible for our orientation and balance in the three-dimensional world, and the other is the cochlea, which is dedicated to sound perception. The cochlea, which looks like a hard shell, is essentially a hard bone filled with fluid and consisting of rigid walls which continue like a spiral with decreasing diameter. Sounds enter the cochlea through the oval window and cause the liquid inside the cochlea to begin to ripple, and that’s where the magic happens.
There are two membranes inside the cochlea, the vestibular membrane, which separates the cochlear duct from the vestibular duct, and the basilar membrane. The basilar membrane is where the organ of Corti resides; this area consists of two sets of sensory receptors, named hair cells, and these are the inner and outer cells. The inner hair cells, which move as the incoming sound waves travel through the cochlear fluid, act as transducers that transform the motion into neural spike activity that is then sent to the auditory nerve.
The outer hair cells can influence the overall sensitivity of each perceived sound. The neural spikes that travel through the auditory nerve are then received by the brain and are interpreted as sounds that we know and can understand. Different sections of the basilar membrane inside the cochlea are responsible for interpreting different frequencies of sound. The section near the entrance consists of hair cells that are responsible for the interpretation of high pitched sounds and, as we move further inside the cochlea, detect progressively lower pitched sounds.
Sound Source Localization
A single ear is basically a frequency analyzer that can provide information to the brain about the pitch of a sound with some dynamics processing that can protect us from loud sounds and some basic filtering of a sound that comes from different directions. So what can we do with two ears that it is almost impossible to do with one? Two ears can provide information about another dimension—the direction—and consequently can give us an estimate of the location of a sound. This is termed spatial hearing. The main theories about how spatial hearing works developed around the idea of how we localize sounds with two ears and get information about the environment in which they reside.
The human auditory system deploys an assortment of localization cues to determine the location of a sound event. The underlying human localization theory is that the main auditory cues that assist with localization are the time and level difference between the two ears and the shape of the ear itself. The most prominent auditory cues use both ears for determining the direction of a sound source and are called binaural cues. These cues measure the time difference of a sound arriving in the two ears, named as interaural time difference (ITD), and cues that track the level difference between the two ears, named as interaural intensity difference (IID). Sound arrives at different times between the two ears and the time delay can be up to approximately 0.7 milliseconds (ms). ITD is meaningful for frequencies up to approximately 1.5 kHz, while ILD provides meaning for frequencies above 1.5 kHz. See Figure 2 for a visual explanation of the ITD and ILD.
ITD and IID alone cannot adequately describe the sound localization process. Imagine a sound coming from directly in front vs the same sound coming directly from the back. ITD and IID provide the exact same values for both ears since their distance from the sound is exactly the same. This is true for all sounds that are on the median plane, which is the plane defined from the points directly in front of us, then up and then back. This is where monaural cues play a role. These are generated by the pinna, diffraction, and reflections of our torso and head. The frequencies of sounds coming from our back or on top of our head will be attenuated differently than sounds coming from front. Another set of cues are dynamic cues which are generated by moving our head. These can show the effect of our head movement to ITD and IID. Last but not least, visual and experience-based cues play an important role as well. Our visual cues, when available, fuse information with auditory cues and resolve to more accurate source localization.
Localizing Sound Underwater
If, under certain conditions, humans can localize a sound source with accuracy as high as 1 degree in the horizontal plane, what makes it so difficult to understand where sounds are coming from when we are underwater? The story of sound perception becomes very different underwater. When we are submerged the localization abilities of the auditory systems degrade significantly. The factor that has the biggest impact when we enter the water is the speed of sound. The speed of sound in air is approximately 343 m/s, while underwater it is 1480 m/s and depends mainly on temperature, salinity etc. A speed of sound that is more than 4x higher than the speed of sound in air will result in sound arriving in our ear with much smaller time differences, which could possibly diminish or even eliminate completely our spatial hearing abilities.
The perception of time differences is not the only cue that is affected; the perception of interaural intensity differences is also affected by the fact that there is very little mechanical impedance mismatch between water and our head. In air, our head acts as an acoustic barrier that attenuates high frequencies due to shadowing effect. Underwater sound most likely travels directly through our head since the impedances of water and our head are very similar. and the head is acoustically transparent. These factors might suggest that underwater, humans could be considered as a one-eared mammal.
In principle, the auditory system performs very poorly underwater. On the other hand, there have been studies and early reports by divers that suggest that some primitive type of localization can be achieved underwater. In a very limited study that aimed to determine the minimum audible angle, divers were instructed to perform a left-right discrimination and were able to improve their ability when they were given feedback on their performance.
Another study was conducted where various underwater projectors were placed at 0, +/-45 and +/-90 degrees at ear level in the horizontal plane with 0 degrees being in front of the diver in . The stimuli consisted of pure tones and noise. The results indicated that divers could perform enough localization that it could not be written off as coincidence or chance. It is not entirely clear whether the underwater localization ability is something that humans can learn and adapt or if it’s a limitation of the human auditory system itself.
Assistive Technologies for Sound Recording
In a perfect world, humans could perform localization tasks underwater as accurately as they could in air. Until that day, should it ever come, we can use assistive technologies to sense the underwater sound environment. Assistive technologies can be exploited for two main applications of 3D audio underwater: localization of sound and sound recording. Sound sensing with microphones is a well developed topic, especially nowadays since the cost of building devices with multiple microphones has decreased dramatically. Multi-microphone devices can be found everywhere in our daily life from portable devices to cars. These devices use multiple microphones to make sense of the world and deliver to the user improved sensing abilities but also the ability to capture the sounds with great accuracy.
The aim of 3D audio is to provide an immersive, appealing, and convincing auditory experience to a listener. This can be achieved by delivering the appropriate spatial cues using a reproduction system that consists of either headphones or loudspeakers. There are quite a few techniques regarding 3D audio localization and 3D audio recording that one can exploit for an immersive experience. Traditionally, recording relies on a small number of sensors that are either played back directly to the loudspeaker or are mixed linearly and then fed to loudspeakers. The main idea here is that if we can approximate the binaural cues, we can trick the human auditory system into thinking that sounds come from specific directions and give the impression of being there.
One of the earliest approaches in the microphone domain was a stereophonic technique invented by Blumlein in Bell Labs . Although this technique was invented in the beginning of the last century, it took many decades to actually commercialize it. From there on, many different stereophonic techniques were utilized that aimed at approximating the desired time and level differences to be delivered to the listener. This was performed by changing the distance between the sensors or by placing directional sensors really close together, and delivering the directional cues by exploiting the directivity of the sensors. The transition from stereophonic to surround systems came right after by using multiple sensors for capturing sounds and multiple loudspeakers for playback. These are, for example, the traditional microphone arrangements we can see on top of an orchestra playing in a hall.
A more advanced technology that existed since the 1970s is called Ambisonics. It consists of a unified generalized approach to the recording, analysis, and reproduction of any recorded sound field. It utilized a standard microphone recording setup with four sensors placed almost in a coincident manner on the tetrahedral arrangement. This method provided great flexibility in terms of playback: once the audio signals are recorded they can be played back in any type of output setup, such as a stereophonic pair of loudspeakers, a surround setup, or plain headphones. The mathematical background of this technique is based on solving the acoustic wave equation, and for a non-expert this can become an inconvenience to use in practice.
Fortunately, there are tools that make this process straightforward for the non-expert. In practical terms, this technology is based on recording signals from various sensors and then combining them to obtain a new set of audio signals, called ambisonic signals. The requirement for acquiring optimal ambisonic signals is to record the signals with sensors placed on the surface of a sphere for practical and theoretical convenience. The process is based on sampling the sound field pressure over a surface or volume around the origin. The accuracy of the sampling depends on the array geometry, the number of sensors, and the distance between the sensors. The encoded ambisonic signals can then accurately describe the sound field.
By using ambisonic signals, one can approximate the pressure and particle velocity of the sound field at the point where the sensors are located. By estimating these two quantities we can then perform an energetic analysis of the sound field and estimate the direction of a sound by using the active intensity vector. The active intensity vector points to the direction of the energy flow in a sound field. The opposite of that direction will point exactly where the energy is coming from, therefore it will show the direction of arrival. A generic view of such a system is shown in Figure 3. For more technical details on how an three-dimensional sound processor works the reader is referred to [3,7].
For underwater audio-related applications, a similar sensor is used, namely the hydrophone. The signal processing principles that can be used underwater are very similar to the ones used with microphones. However, the applications are not as widely spread. Multiple hydrophones, namely hydrophone arrays, have been utilized for localization tasks, but many of them are based on classical theories that require the hydrophones to be very far apart from each other. Therefore, compact, wearable devices are uncommon. A compact hydrophone array that can be carried by a single diver can serve multiple purposes: it can operate as a localization device that can locate quite accurately where sounds are originating from, and it can also capture the sounds with spatial accuracy that can later be played back and give the experience of being there. This can be especially useful for people that do not have the chance to experience the underwater world at all or listen to the sounds of highly inaccessible places where diving is only for highly-skilled technical divers.
A purposefully built compact hydrophone array which can be operated by a diver and that can perform simultaneously a localization task but also record spatial audio that can be later played through a pair of headphones, is shown in a recently published article as a proof of concept. It is suggested that a simple set of four hydrophone sensors can be exploited in real-time to identify the direction of underwater sound objects with relatively high accuracy, both visually and aurally.
Underwater sound localization is not a new technology, but the ability to do such tasks with compact hydrophone arrays where the hydrophones are closely spaced together has the potential for many applications. The hydrophone array used in the study discussed above, was an open-body hydrophone array, consisting of four sensors placed at the vertices of a tetrahedral frame. A photo of the hydrophone array is shown in Figure 4. More technical details on the hydrophone array can be found in the following publication . The hydrophone array was used to provide a visualization of the sound field. The sound-field visualization is basically a sound activity map where the relative energy for many directions is depicted using a color gradient like in Figure 5. These experiments were performed in a diving pool.
The depth of the hydrophone array was fixed by attaching it to the end of a metal rod, which was made out of a three-piece aluminum tube. The hydrophone array signals were recorded and processed in real time with a set of virtual instrument plugins. The observer, a single action camera in this case, was able to clearly track the movements of the diver from the perspective of the array, both visually and aurally. The diver from the perspective of the array and the corresponding activity map are both shown in Figure 5. A video demonstrating the tracking and headphone rendering with a diver inside the swimming pool can be found here.
This example has been utilized as a proof of concept that compact arrays can be indeed exploited by divers simultaneously for sound visualization and accurate three-dimensional sound recording. In the examples shown, the sound arrives at the hydrophone array and then the user can visually see the location of sound and simultaneously listen to the acoustic field. The three-dimensional audio cues are delivered accurately and the sound sources are represented to the user in their correct spatial location so the user can accurately localize them.
3D Video Recording
The current state of the art in 3D video recording underwater usually captures the audio with the microphones built into the cameras. Although sounds are often recorded in this way, the recorded sound through built-in microphones inside a casing cannot capture underwater sounds accurately. The development of a relatively compact hydrophone array such as the one demonstrated could potentially enhance the audio-visual experience by providing accurate audio rendering of the recorded sound scene. Visuals and specific 3D video recordings can be pretty impressive by themselves, but the addition of an accompanying 3D audio component can potentially enhance the experience.
Visuals and specific 3D video recordings can be pretty impressive by themselves, but the addition of an accompanying 3D audio component can potentially enhance the experience.
Ambisonic technologies are a great candidate for such an application since the underwater sound field can be easily recorded with an ambisonic hydrophone array (such as the one shown in Figure 4) and can be mounted near or within the 3D camera system. Ambisonics aim at a complete reconstruction of the physical sound scene. Once the sound scene is recorded and transformed in the ambisonic format, a series of additional spatial transformations/manipulation can come handy. Ambisonics allow straightforward manipulation of the recorded sound scene, such as a rotation of the whole sound field to arbitrary directions, mirroring of the sound field to any axis, direction-dependent manipulation such as focusing on different directions or areas, or even warping of the whole sound field. The quality of the reproduced sound scene and accuracy of all the sound scene manipulation techniques highly depends on the number of sensors involved in the recording.
Some modern, state-of-the art techniques are based on the same recording principle, but they can provide an enhanced audio experience by using a low number of sensors that exploit human hearing abilities as well as advanced statistical signal processing techniques. These techniques aim to imitate the way human hearing works, and by processing the sounds in time, frequency, and space, they deliver the necessary spatial cues for an engaging aural experience, which is usually perceived as very close to reality. They exploit the limitations of the human auditory system and analyze only the properties of the sound scene that are meaningful to a human listener. Once the sound scene is recorded and analyzed into the frequency content of different sounds, it can then be manipulated and synthesized for playback .
-  Delikaris Manias, Symeon, Leo McCormack, Ilkka Huhtakallio, and Ville Pulkki. “Real-time underwater spatial audio: a feasibility study.” Audio Engineering Society, 2018.
-  Spatial Audio Real-Time Applications (SPARTA).
-  Gerzon, Michael A. “Periphony: With-height sound reproduction.” Journal of the audio engineering society, 1973.
-  Feinstein, Stephen H. “Acuity of the human sound localization response underwater.” The Journal of the Acoustical Society of America 1973.
-  Hollien, Harry. “Underwater sound localization in humans.” The Journal of the Acoustical Society of America, 1973
-  A. D. Blumlein. British patent specification 394,325 (improvements in and relating to sound-transmission, sound-recording and sound-reproducing systems). Journal of the Audio Engineering Society, 6(2):91–130, 1958.
-  Pulkki, Ville, Delikaris Manias Symeon, and Politis Archontis. Parametric time-frequency domain spatial audio , John Wiley & Sons, 2018.
Dr. Symeon Manias has worked in the field of spatial audio since 2008. He has received awards from the Audio Engineering Society, the Nokia Foundation, the Foundation of Aalto Science and Technology, and the International College of the European University of Brittany. He is an editor and author of the first parametric spatial audio book from Wiley and IEEE Press. Manias can be usually found diving through the kelp forests of Southern California.
Thank You to Our Sponsors
Zen and the Art of Mexican Cave Navigation
Do you know your arrows, cookies, REMS, TEMS, presence, personal and team markers? Singley or in combination? Jump protocols? If...
The GUE Pre-dive Sequence
As you’d expect, Global Underwater Explorers (GUE) has a standardized approach to prepare your equipment for the dive, and its...
Maintaining Unit Cohesion
If you’re going to go metric or already are, it’s time to get your units straight, explains British tech instructor...
How Two Tech Agencies Address Isobaric Counterdiffusion
NAUI is one of the few training agencies that offers specific protocols to address Isobaric Counterdiffusion (ICD). Here NAUITEC Instructor...