Blog

Workflow

Viewing posts tagged Workflow

Immersive Audio Podcast – Episode 13 Hyunkook Lee

Summary

In today’s episode, Oliver was joined via Skype by Dr. Hyunkook Lee, Senior Lecturer in Music Technology and Production and the leader of the Applied Psychoacoustics Lab (APL) at the University of Huddersfield. Hyunkook joined Huddersfield in 2010 and developed research in the area of 3D audio psychoacoustics as well as undergraduate modules such as Acoustics and Concert hall recording technique. In 2014 he established the APL, a research group studying the mechanism of human auditory perception and developing new audio algorithms for practical applications. He has undertaken a number of consultancy works for companies such as Samsung Electronics, Volvo Car and L-ISA.

Hyunkook is also an experienced recording and mixing engineer specialising in acoustic music.

Before joining Huddersfield, Dr Lee was a Senior Research Engineer at LG Electronics in South Korea, where he led a project to develop audio post-processing algorithms for LG mobile phones. He has also participated in MPEG audio codec standardisation activities, contributing to the developments of codecs such as SAOC and USAC. Hyunkook graduated from the music and sound recording (Tonmeister) course at the University of Surrey in 2002. During the course he spent a placement year as an assistant engineer at Metropolis studios in London. He gained his PhD from the same university in 2006.

His PhD research was concerned with the subjective effects and objective measurements of interchannel crosstalk in multichannel microphone techniques, and as a Senior Lecturer, he now spends his time tutoring and guiding aspiring students in the research of 3D sound and continues to further progress the academic understanding of the subject.

In this episode, Dr Hyunkook Lee talks to 1.618 Digital about a variety of topics under 3D Sound and Ambisonics: Psychoacoustics, microphone and recording techniques, and theories such as Phantom Image and Elevation Perception. He also shares with us his personal researching tips for audio engineering students, the importance of realising the value of your own research and believing in the work you do for eventual real-world applications.

Listen to Podcast

Shownotes

University of Surrey – Music & Sound Recording (Tonmeister): https://www.surrey.ac.uk/undergraduate/music-and-sound-recording-tonmeister

University of Huddersfield – Music Technology: https://www.hud.ac.uk/inspire/musictechnology/

LG: http://www.lg.com

MPEG: https://mpeg.chiariglione.org/

Fraunhofer: https://www.fraunhofer.de

Dolby: www.dolby.com

Aspen Music Festival: http://www.aspenmusicfestival.com/

University of Huddersfield – Applied Psychoacoustics Lab: https://research.hud.ac.uk/institutes-centres/mtprg/projects/apl/

Schoeps: https://www.schoeps.de/

Investigation on the Phantom Image Elevation Effect (Lee 2015): http://eprints.hud.ac.uk/id/eprint/26558/

Perceptual Band Allocation (PBA) for the Rendering of Vertical Image Spread with a Vertical 2D Loudspeaker Array (Lee 2016): http://eprints.hud.ac.uk/id/eprint/29791/

Alternative Weighting Filters for Multi-Track Program Loudness Measurement (Fenton & Lee 2017): http://www.aes.org/e-lib/browse.cfm?elib=19215

Applied Psychoacoustics Lab on Facebook: https://www.facebook.com/applied.psychoacoustics.lab/

Hyunkooklee.com: www.hyunkooklee.com

MARRS App: https://itunes.apple.com/us/app/marrs/id1295926126?mt=8

Immersive Audio Podcast – Episode 10 Jane Gauntlett & Tessa Ratuszynska (In My Shoes)

Summary

Today we have two special guests in studio, Jane Gauntlett & Tessa Ratuszynska. Jane is a writer for film and theatre. She experiments with new technologies, focusing on designing high-impact experiences with a strong narrative. In 2011 Jane founded the In My Shoes Project, an ever-expanding library of interactive experiences which use story, theatre, audio-visual technology, virtual reality and first person documentary to recreate real-life experiences.

Tessa is a documentary filmmaker, installation artist and creative producer for virtual reality and performance. She is a producer of new media documentary, with particular creative focus in virtual reality, 360 documentary and has a passion for the potential for affecting works of interactive non-fiction. In My Shoes: Intimacy was written and directed by Jane and Tessa was creative producer.

In this episode we speak about early adoption of VR, the advances in technology and keeping ahead of the curve. We discuss reaching out to younger audiences, the importance of not using technology for the sake of it and this new technology as a platform for immersive storytelling.

Listen to Podcast

Shownotes

Video goggles – Vuzix: https://www.vuzix.com/

Oculus: https://www.oculus.com/

HoloLens: https://www.microsoft.com/en-gb/hololens

In My Shoes: http://janegauntlett.com/inmyshoesproject/

Intimacy: http://janegauntlett.com/in-my-shoes-intimacy/

Immersive Audio Podcast – Episode 3 Michael Kelly

Summary

Michael Kelly is Senior Director, Research and Development at DTS and is responsible for spatial audio SDKs used in games, AR, VR and other media. Kelly is also co-chair of the AES Technical Committee on Audio for games and chairs the AES Audio for Games conferences. He has over 15 year’s experience in the games industry working on both the creative and technical aspects of games. Michael began his career as Lead Sound Designer at Revolution Software working on Broken Sword: The Sleeping Dragon. Following this, he was responsible for the implementation of high quality DSP effects on a number of platforms at Creative Labs and led 3rd-party audio tools and high-level library development for PS3 and Vita at Sony Computer Entertainment Europe. He has a PhD in Electronic Engineering from the University of York, specialising in 3-D audio and Audio Coding.

 

Listen to Podcast

 

Shownotes

Michael Kelly Linkedin: https://uk.linkedin.com/in/michael-kelly-7991715

Broken Sword: https://revolution.co.uk/

DTS: http://dts.com/

DTS:X: http://dts.com/dtsx

https://www.youtube.com/watch?v=qEbRNeOcf9c

DTS Headphone:X: http://dts.com/headphonex

Xperi: https://www.xperi.com/

Elite Dangerous: https://www.elitedangerous.com/

The International Consumer Electronics Show: https://www.ces.tech/

Audio Engineering Society: http://www.aes.org/

Unity: https://unity3d.com/

Unreal: https://www.unrealengine.com

What YouTube’s Heatmap Is Really Saying About 360 Video

YouTube recently announced a new analytics tool for 360-degree and virtual reality content creators: heatmaps that illustrate where viewers are actually looking. The new tool allows creators to see exactly what parts of their video are holding a viewer’s attention, and for how long.

YouTube has also released some enlightening early statistics on how – and this is important – viewers currently engage with immersive content.

“Surprisingly” (says YouTube), viewers spend 75% of their time focused on the front 90 degrees of an immersive video. Understandably, this figure has a lot of people questioning the point of VR content if the audience is only willing to engage with a quarter of it.

It’s an easy argument to make, but perhaps what these numbers are really saying is that VR content creators need to learn new ways to grab viewers attention in a 360º world?

Ever since moving pictures became something we watched for entertainment purposes, our eyes have been guided by camera angles to tell us where to look. For over a century that’s what the viewing audience has come to expect.

Virtual reality reminds us very much of the 2D world of film and television, but it’s an entirely different medium with its own set of rules that are still being written. Nothing is set in stone.

And camera angles? Well, those are up to the viewer to choose.

Content creators in the virtual reality space have the difficult task of catching the attention of an audience with over 100 years of collective viewing experience of looking straight ahead.

Does this make virtual reality a fad? A gimmick? No, of course not. It simply means that VR can’t rely on the same tools that have been used for film and television to engage an audience in a fully-immersive format.

That’s a lot of unlearning to do for content creators, and a lot of new learning to do as the format develops. It’s an exciting new frontier.

Back to YouTube’s statistics: the most popular VR videos had the audience looking behind them almost 20% of the time. Markers and animation are what the company suggests will help draw attention to other parts of the surrounding space. In our day to day lives our attention is constantly guided by signs, so it’s a helpful suggestion. But think about this: what’s the one sure thing that will make you stop whatever you’re looking at and focus your attention elsewhere?

Sound…

We are programmed to react to sound. In a split second we must figure out where that sound is coming from and what it means. It is as true in the virtual world as it is in the real world, which is why 1.618 Digital is passionate about high-quality spatialised sound.

Spatial audio can be an effective tool to lead or surprise your audience.  By being in the habit of looking in one direction at any given time, the viewer can easily miss out on what is happening behind or beside them. Through the creative implementation of sonic cues within an immersive environment content creators can control or suggest a narrative. Ultimately, this encourages the audience to engage with specific elements – or viewing angles – within the experience.

Virtual reality is an effective form of visual storytelling. What YouTube’s early heatmap data points to isn’t VR’s failure to engage its viewers. It’s the bigger picture of where audience attention currently is, and the gaps content creators need to fill to direct it elsewhere.

1.618 Digital Team

Discussing the importance of spatial audio on happyfinish.com

Importance of Spatial Audio In VR Content

Hearing is fundamental to our perception of the surrounding world.  Achieving this effect in virtual reality requires audio that sounds real and authentic.  Implementing spatial audio to create full immersion in 360° video or interactive VR requires capturing audio or a physical acoustic modeling of the space where the scene takes place.  An appropriate soundscape can provide the quickest path to immersion for just about any VR experience, and even removing the visual element, still enables us to sufficiently perceive the surrounding world – giving us a sense of space, time, and presence.  In contrast, the silent experiences, or the ones with incongruent sound would break the sense of presence and immersion, thus immediately removing the suspension of disbelief, and as a result substantially degrading the overall experience.

Spatial sound recording or let’s do it in post?

The use of conventional industry formats such as Mono (single channel) and Stereo (two channels) are a basic requirement, although they are limiting and no longer sufficient to offer full immersion in 360° videos or interactive VR experiences.  The use of spatial audio is the only way to create true three-dimensional audio, which utilises higher number of channels, be it capturing sound on location or through the means of sound design and mixing in post-production.  Depending on the nature of the project both methods are important.  Often to design the full sonic experience in VR, it requires spatial sound recording on set along with sound design and spatialisation of individual elements in post-production such as atmosphere, dialogue, foley, sound effects, and music.

Ambisonic format is the most effective method to capture location sound

There are a number of ways to capture the location sound.  However, the most effective method is to record in an ambisonic format which utilises four channels capturing the sound in all directions, along with discrete sound sources such as dialogue or any required diegetic sounds that are part of the scene.  The latter can then be positioned accordingly within the 3D soundfield by employing specialist spatialisation software within audio editing application or a game engine.  This approach enables VR audio content makers to work with an adequate resolution within the virtual space for positioning sound components across four, 16 or more virtual or physical channels.

Ambisonic sound offers a number of significant benefits that play a crucial role in making experiences as realistic as possible.

-Firstly, sound that was captured in all directions then enables the user to move their head and body while wearing any head-mounted display, and with a use of head-tracking system perceive their own dynamic position within the space in relation to the surrounding environment.

-Second, greater channel count offers more accuracy in positioning individual elements within the 3D space. This avoids everything coming from the same general direction as is common when listening to music, but lacking in realism when comes to creating a metaverse or offering your audience an authentic 360° video experience.

Why is this essential?

The considerations mentioned above are essential due the phenomenon described as a head-related transfer function (HRTF), which is a response that informs how our ears perceive sound from own position in space.  Collectively, head-related functions for both ears give a perception of binaural sound, enabling us to effectively identify a location and a distance of sound sources by constantly receiving sonic information to measure sound intensity and the time difference between sounds arriving to both ears.  We re-create this psychoacoustic process in post-production to then achieve a latency-free, real-time binaural rendering via a close approximation of personal HRTF.  It is essential to take human physiology into consideration when making audience fully immersed and enjoy their experience, be it a story, game or cognitive therapy etc.

The use of audio in marketing campaigns to guide your audience in 360° content

Unlike 2D content where a viewer can see the entire field of view in one direction, the 360° environment presents challenges as well as the opportunities for creative content makers when it comes to constructing the narrative.  Spatial audio can be an effective tool to lead or surprise your audience.  By looking in one direction at any given time, the viewer can easily miss out on what is behind them or sidewards, by implementing sonic cues within the space we can control or suggest a narrative.  By helping our audience to navigate through their point of view, we can ultimately guide them to and encourage them to engage with a specific element within the experience.

What is more important?

When combined effectively, fully integrated visual and sonic perception work in perfect harmony that enables us to see, hear, feel and appreciate the beauty and richness of our world.  Virtual reality already proved its effectiveness in video storytelling, gaming, educational training, social interaction and medical applications.  In order to make any of the above experiences successful, it requires a coherent approach of applying sound and visual content to make it as effective for its purpose as possible – more immersive, more authentic and as the result more engaging, more memorable, more empathetic, more fun and ultimately good enough to have a desire to come back and experience it again and again.

1.618 Team