Blog

3D audio

Viewing posts tagged 3D audio

Immersive Audio Podcast – Episode 4 Sally-Anne Kellaway

Summary

In today’s episode we are joined by Sally-Anne Kellaway via Skype. Sally is the Senior Audio Designer on the Microsoft Mixed Reality – Audio and AI team, and an industry leading VR Audio Evangelist.

She has extensive experience in sound recording, editing, implementation and testing for interactive media in standard and VR/AR and has an interest in implementing true 3D spatial hearing and binaural sound in video games and other interactive experiences. She is the founder of the Virtual Reality Content Creators Network of Australia (VRCC)  which is a not for profit community for virtual reality and associated industries. The VRCC advocates for innovation and connection for all professionals working in virtual, augmented and mixed reality industries in Australia.

Previously the Creative Director at OSSIC, the leading audio technology providers active in the VR space, she has a Masters in Acoustics with a focus on psychoacoustics and spatial audio from the University of Sydney. Sally-Anne is sought after by VR and AR developers, Software and Middleware developers, Conferences, Meetup Groups and Panels to develop and consult on sound design, implementation, tutorials and presentations. She is the Founder of the Women/NB in VR Group for Australia and co-organiser of the San Diego VR community group and is on the AES Technical Committee for Game Audio (Spatial Audio) and on the Diversity and Inclusion Working group.

Sally is a vocal advocate for VR, AR and MR as the future of audio, entertainment and industry and uses her education in Sound Design, Acoustics and Psychoacoustics to rise the tide for all audio professionals passionate about the future.

In this episode, Sally-anne speaks about head-tracking, the future of Audio AR, and diversity in the industry.

Listen To Podcast

Shownotes

Sally-anne Kellaway: http://soundsbysal.weebly.com

University of Sydney: https://sydney.edu.au/

The Virtual Barbershop: https://www.youtube.com/watch?v=IUDTlvagjJA

OSSIC: https://www.ossic.com/

Audio Engineering Society: http://www.aes.org/

Virtual Reality Content Creators of Australia: http://www.vrcc-aus.rocks/

The Wave VR: http://thewavevr.com/

Zero Latency VR: https://www.zerolatencyvr.com/

Immersive Audio Podcast – Episode 3 Michael Kelly

Summary

Michael Kelly is Senior Director, Research and Development at DTS and is responsible for spatial audio SDKs used in games, AR, VR and other media. Kelly is also co-chair of the AES Technical Committee on Audio for games and chairs the AES Audio for Games conferences. He has over 15 year’s experience in the games industry working on both the creative and technical aspects of games. Michael began his career as Lead Sound Designer at Revolution Software working on Broken Sword: The Sleeping Dragon. Following this, he was responsible for the implementation of high quality DSP effects on a number of platforms at Creative Labs and led 3rd-party audio tools and high-level library development for PS3 and Vita at Sony Computer Entertainment Europe. He has a PhD in Electronic Engineering from the University of York, specialising in 3-D audio and Audio Coding.

 

Listen to Podcast

 

Shownotes

Michael Kelly Linkedin: https://uk.linkedin.com/in/michael-kelly-7991715

Broken Sword: https://revolution.co.uk/

DTS: http://dts.com/

DTS:X: http://dts.com/dtsx

https://www.youtube.com/watch?v=qEbRNeOcf9c

DTS Headphone:X: http://dts.com/headphonex

Xperi: https://www.xperi.com/

Elite Dangerous: https://www.elitedangerous.com/

The International Consumer Electronics Show: https://www.ces.tech/

Audio Engineering Society: http://www.aes.org/

Unity: https://unity3d.com/

Unreal: https://www.unrealengine.com

Immersive Audio Podcast – Episode 2 Justin Paterson

Summary

Justin Paterson joined us in the studio for today’s highly informative episode. After leaving silicon chip design, Justin worked as a session drummer for a number of years in the 90’s playing on releases, TV and radio before moving to synthesis, production and composition. His research on many aspects of music production and audio technology has been published widely. He has written for, given tutorials to, and peer-reviewed for the AES in the UK, US and Europe. In 2014, He was awarded an AHRC grant to develop an App release-format for music albums with interactive playback, a system that is currently patent pending and undergoing further AHRC-funded commercial exploitation by the Warner Music Group. Justin also worked with the record label Ninja Tune and the band Coldcut on binaural 3-D (interactive) audio for VR. Justin also reviews books for Focal Press, and is author of The Drum Programming Handbook, and a chapter on 3-D music production in a forthcoming Routledge edition in the ‘ Perspective on Music Production’ series. He is also co-editor of a newly commissioned Routledge book on 3-D audio.

Justin is also co-chair of the Innovation in Music Conference series, and supervises PhDs related to music technology, both in LCM and externally. He is also a practising record producer with commercial releases, and has authored a number of funding proposals relating to 3-D production, that are currently under review.

As well as discussing the above, Justin explains the concepts of ambisonics, binaural audio and head related transfer function (HRTF). We talk about the academic community, immersive audio and the development of technology associated with 3-D audio. Justin speaks about his plan to create a 3D audio research network, to bring together academia and industry and define cross-industrially relevant research questions.

“Success is getting what you want, happiness is wanting what you have, and you have to be very careful not to get those two confused”

 

Listen To Podcast

 

Shownotes

Professor Justin Paterson: Associate Professor of Music Technology: https://www.uwl.ac.uk/users/justin-paterson

MA Advanced Music Technology: University of West London: http://www.uwl.ac.uk/course/advanced-music-technology-1/29473

The Interactive Album App: https://www.uwl.ac.uk/academic-schools/music/lcm-research/current-research-projects/interactive-album-app

Daisy and the Dark: Red Planet: http://www.daisyandthedark.com/redplanetapp

Innovation in Music Conference: http://www.musicinnovation.co.uk/

Mi.Mu Gloves: http://mimugloves.com/

Auro-3D: https://www.auro-3d.com/

The Drum Programming Handbook: The Complete Guide to Creating Great Rhythm Tracks: http://amzn.to/2HFvvuo

SADIE: Spatial Audio for Domestic Interactive Entertainment: https://www.york.ac.uk/sadie-project/

S3A: Future Spatial Audio: http://www.s3a-spatialaudio.org/wordpress/

The Ambisonic Motherlode: http://decoy.iki.fi/dsound/ambisonic/motherlode/index

1.618 To Join The Panel at The BVE 2018 Excel London

1.618 Digital will be taking part in a panel at The Storyteller Theatre at the BVE 2018 Expo Excel London. Oliver Kadel will be discussing the use of immersive audio in modern storytelling, covering everything from future technologies to next generation audio. We’d love to see you there!

The panel discussion takes place on the 1st of Match at 12.30. You can get your free ticket here: Ticket Registration

Check out what else is taking place at BVE 2018.

 

See you soon!

1.618 Team

Double Nomination at Raindance Film Festival

Awards season is upon us and here at 1.618 Digital we’re very excited to announce that two projects we’ve worked on have been nominated at the Raindance Film Festival.

First up, In My Shoes: Intimacy, by Jane Gauntlett and Visualise, has been nominated in the Best Sensual Experience category.

In My Shoes: Intimacy is a 360 experience which explores the power of human connection. Put aside your inhibitions, let these strangers guide you through their impromptu, unconventional & intense moments of intimacy. Intimacy is a first-person documentary designed for two people to experience three encounters from six very different perspectives. To read more about this project please check this case study or see the nomination here.

Next, Laphroaig by Darren Emerson and VR City has been nominated in the Best Branded Content category.

This film takes you on a journey into the heart of the historic Laphroaig whiskey distillery on the beautiful island of Islay. See the nomination here.

Huge congrats to everyone involved and fingers crossed for the awards announcement later today!

Good luck to all industry friends who also have been nominated.

1.618 Digital

What YouTube’s Heatmap Is Really Saying About 360 Video

YouTube recently announced a new analytics tool for 360-degree and virtual reality content creators: heatmaps that illustrate where viewers are actually looking. The new tool allows creators to see exactly what parts of their video are holding a viewer’s attention, and for how long.

YouTube has also released some enlightening early statistics on how – and this is important – viewers currently engage with immersive content.

“Surprisingly” (says YouTube), viewers spend 75% of their time focused on the front 90 degrees of an immersive video. Understandably, this figure has a lot of people questioning the point of VR content if the audience is only willing to engage with a quarter of it.

It’s an easy argument to make, but perhaps what these numbers are really saying is that VR content creators need to learn new ways to grab viewers attention in a 360º world?

Ever since moving pictures became something we watched for entertainment purposes, our eyes have been guided by camera angles to tell us where to look. For over a century that’s what the viewing audience has come to expect.

Virtual reality reminds us very much of the 2D world of film and television, but it’s an entirely different medium with its own set of rules that are still being written. Nothing is set in stone.

And camera angles? Well, those are up to the viewer to choose.

Content creators in the virtual reality space have the difficult task of catching the attention of an audience with over 100 years of collective viewing experience of looking straight ahead.

Does this make virtual reality a fad? A gimmick? No, of course not. It simply means that VR can’t rely on the same tools that have been used for film and television to engage an audience in a fully-immersive format.

That’s a lot of unlearning to do for content creators, and a lot of new learning to do as the format develops. It’s an exciting new frontier.

Back to YouTube’s statistics: the most popular VR videos had the audience looking behind them almost 20% of the time. Markers and animation are what the company suggests will help draw attention to other parts of the surrounding space. In our day to day lives our attention is constantly guided by signs, so it’s a helpful suggestion. But think about this: what’s the one sure thing that will make you stop whatever you’re looking at and focus your attention elsewhere?

Sound…

We are programmed to react to sound. In a split second we must figure out where that sound is coming from and what it means. It is as true in the virtual world as it is in the real world, which is why 1.618 Digital is passionate about high-quality spatialised sound.

Spatial audio can be an effective tool to lead or surprise your audience.  By being in the habit of looking in one direction at any given time, the viewer can easily miss out on what is happening behind or beside them. Through the creative implementation of sonic cues within an immersive environment content creators can control or suggest a narrative. Ultimately, this encourages the audience to engage with specific elements – or viewing angles – within the experience.

Virtual reality is an effective form of visual storytelling. What YouTube’s early heatmap data points to isn’t VR’s failure to engage its viewers. It’s the bigger picture of where audience attention currently is, and the gaps content creators need to fill to direct it elsewhere.

1.618 Digital Team

1.618 Digital to join the panel at the Immerse UK event

Welcome to the Real World – Immersive Audio for Augmented and Virtual Reality

1.618 will be taking part in a panel discussion with the UK’s leading Immersive Audio experts. Join us as we talk about the latest cutting-edge work in immersive audio research, development and content design.

The event will cover everything from audio AR/VR creative process, application workflows and the potential for product development. This guarantees to be a truly interactive event, with teams demonstrating some of the most exciting immersive audio technology and experiences in the field.

Other speakers include:

  • Jelle van Mourik (Sony) – Immersive Audio in Games
  • Gavin Kearney (University of York) – Immersive Audio Research
  • Gareth Llewellyn (Mixed Immersion) – Immersive Audio Production for VR
  • Chris Pike (BBC) – Immersive Audio in Broadcast

The panel discussion will take place at 5.45pm-6.45pm on Wednesday 24th May 2017. Click here to register and get your free ticket: Ticket Registration

For more upcoming events hosted by Innovate UK click here

See you there!
1.618 Digital Team

1.618 to join the panel at The BVE 2017 Excel London

1.618 will be taking part in a panel of 3D audio experts at the BVE 2017 Expo Excel London on 2nd March. Join us as we discuss the next generation of immersive audio, covering everything from future technologies to next generation audio. We’d love to see you there!

Discussion topics will include:

Insights into the current and future technologies that will be taking the industry by storm

The role of object-based audio – how this will improve the customer experience.

Binaural audio – creating a really immersive sound for your audience

Will next generation audio take off – how far away are we?

The panel discussion takes place at 12.30-13.30 on Thursday 2nd March. You can get your free ticket here: Ticket Registration 

Check out what else is taking place at BVE on Thursday 2nd March here

See you soon!

1.618 Team

 

Discussing the importance of spatial audio on happyfinish.com

Importance of Spatial Audio In VR Content

Hearing is fundamental to our perception of the surrounding world.  Achieving this effect in virtual reality requires audio that sounds real and authentic.  Implementing spatial audio to create full immersion in 360° video or interactive VR requires capturing audio or a physical acoustic modeling of the space where the scene takes place.  An appropriate soundscape can provide the quickest path to immersion for just about any VR experience, and even removing the visual element, still enables us to sufficiently perceive the surrounding world – giving us a sense of space, time, and presence.  In contrast, the silent experiences, or the ones with incongruent sound would break the sense of presence and immersion, thus immediately removing the suspension of disbelief, and as a result substantially degrading the overall experience.

Spatial sound recording or let’s do it in post?

The use of conventional industry formats such as Mono (single channel) and Stereo (two channels) are a basic requirement, although they are limiting and no longer sufficient to offer full immersion in 360° videos or interactive VR experiences.  The use of spatial audio is the only way to create true three-dimensional audio, which utilises higher number of channels, be it capturing sound on location or through the means of sound design and mixing in post-production.  Depending on the nature of the project both methods are important.  Often to design the full sonic experience in VR, it requires spatial sound recording on set along with sound design and spatialisation of individual elements in post-production such as atmosphere, dialogue, foley, sound effects, and music.

Ambisonic format is the most effective method to capture location sound

There are a number of ways to capture the location sound.  However, the most effective method is to record in an ambisonic format which utilises four channels capturing the sound in all directions, along with discrete sound sources such as dialogue or any required diegetic sounds that are part of the scene.  The latter can then be positioned accordingly within the 3D soundfield by employing specialist spatialisation software within audio editing application or a game engine.  This approach enables VR audio content makers to work with an adequate resolution within the virtual space for positioning sound components across four, 16 or more virtual or physical channels.

Ambisonic sound offers a number of significant benefits that play a crucial role in making experiences as realistic as possible.

-Firstly, sound that was captured in all directions then enables the user to move their head and body while wearing any head-mounted display, and with a use of head-tracking system perceive their own dynamic position within the space in relation to the surrounding environment.

-Second, greater channel count offers more accuracy in positioning individual elements within the 3D space. This avoids everything coming from the same general direction as is common when listening to music, but lacking in realism when comes to creating a metaverse or offering your audience an authentic 360° video experience.

Why is this essential?

The considerations mentioned above are essential due the phenomenon described as a head-related transfer function (HRTF), which is a response that informs how our ears perceive sound from own position in space.  Collectively, head-related functions for both ears give a perception of binaural sound, enabling us to effectively identify a location and a distance of sound sources by constantly receiving sonic information to measure sound intensity and the time difference between sounds arriving to both ears.  We re-create this psychoacoustic process in post-production to then achieve a latency-free, real-time binaural rendering via a close approximation of personal HRTF.  It is essential to take human physiology into consideration when making audience fully immersed and enjoy their experience, be it a story, game or cognitive therapy etc.

The use of audio in marketing campaigns to guide your audience in 360° content

Unlike 2D content where a viewer can see the entire field of view in one direction, the 360° environment presents challenges as well as the opportunities for creative content makers when it comes to constructing the narrative.  Spatial audio can be an effective tool to lead or surprise your audience.  By looking in one direction at any given time, the viewer can easily miss out on what is behind them or sidewards, by implementing sonic cues within the space we can control or suggest a narrative.  By helping our audience to navigate through their point of view, we can ultimately guide them to and encourage them to engage with a specific element within the experience.

What is more important?

When combined effectively, fully integrated visual and sonic perception work in perfect harmony that enables us to see, hear, feel and appreciate the beauty and richness of our world.  Virtual reality already proved its effectiveness in video storytelling, gaming, educational training, social interaction and medical applications.  In order to make any of the above experiences successful, it requires a coherent approach of applying sound and visual content to make it as effective for its purpose as possible – more immersive, more authentic and as the result more engaging, more memorable, more empathetic, more fun and ultimately good enough to have a desire to come back and experience it again and again.

1.618 Team