Senior Multimodal Researcher - Real-time Spatial Audio and Multimodal Interactive Computing

Apply now »

Date: Nov 7, 2024

Location: Bangalore, IN

Company: Dolby Laboratories, Inc.

Join the leader in entertainment innovation and help us design the future. At Dolby, science meets art, and high tech means more than computer code. As a member of the Dolby team, you’ll see and hear the results of your work everywhere, from movie theaters to smartphones. We continue to revolutionize how people create, deliver, and enjoy entertainment worldwide. To do that, we need the absolute best talent. We’re big enough to give you all the resources you need, and small enough so you can make a real difference and earn recognition for your work. We offer a collegial culture, challenging projects, and excellent compensation and benefits, not to mention a Flex Work approach that is truly flexible to support where, when, and how you do your best work.

 

The Advanced Technology Group (ATG) is the research division of the company. ATG’s mission is to look ahead, deliver insights, and innovate technological solutions that will fuel Dolby’s continued growth. Our researchers have a broad range of expertise related to computer science and electrical engineering, such as AI/ML, algorithms, digital signal processing, audio engineering, image processing, computer vision, human perception and cognition, data science & analytics, distributed systems, cloud, edge & mobile computing, computer networking, and IoT.
We are seeking talented Senior Multimodal Researchers to join the Perceptual and Interactive Multimedia Computing team in the Multimodal Experiences Lab.
We are a key research team within Dolby’s Advanced Technology Group, focused on creating cutting edge multimodal technologies that drive next generation experiences. We’re looking for skilled researchers who are excited to advance the state of the art in technologies of interest to Dolby as well as the human society at large, in particular, in the area of developing real-time spatial audio-multimodal and interactive algorithms and experiences. 
We welcome the opportunity to have you join our growing Bangalore Advanced Technology Research team.


Key Responsibilities:
•    Partner with ATG researchers across offices in the USA, Australia, China and Europe; develop solutions for novel XR/Spatial Media applications with a spatial audio emphasis
•    Develop prototypes of real-time audio-multimodal algorithms in head-mounted displays (HMDs) such as the Apple Vision Pro, Meta Quest and Varjo HMD, and instrument the prototypes to record user interactions.
•    Characterize 3rd party real-time algorithms and solutions, including spatial audio rendering solutions/tools such as those in PHASE, Unity, Unreal Engine, Wwise.
•    Develop and combine deep learning methodologies with perceptually relevant signal processing and metrics.


What you need to succeed


Competencies:
•    Technical depth: Necessary technical knowledge to create and implement new real-time spatial media experiences with an audio focus. Solid knowledge of audio and acoustics fundamentals and experiences with development tools such as Xcode and Swift for the Apple Ecosystem, Android dev tools for Meta Quest, or equivalent in relation to the Varjo HMD.
•    Explore new technologies: Openness to learn new skills, work with cutting-edge technologies, and innovate in new areas.
•    Invent & Innovate: Develop know-how, algorithms and software tools with both a short and long-term focus to further strengthen Dolby as a world leader in enhancing the sight and sound associated with digital content consumption. Influence and collaborate with business group partners put the technology into production.
•    Work with a sense of Urgency: Respond assertively to changing trends and new technologies and create new algorithms to capitalize on them. Take appropriate risks to be ahead of the competition and the market.
•    Collaborate: Collaborate with and influence peers in developing industry-leading technologies. Work with external trendsetters and technology drivers in academia and in partner enterprises.
 
Desired Background:
•    PhD in Human Computer Interfaces, Acoustics, Computer Science, Electrical and Computer Engineering, or similar fields
•    Proven ability to pursue new areas of multimodal research for Spatial Audio, real-time interactivity, AI, and signal analysis, and demonstrate results through projects, prototypes, patent filings, and papers in peer reviewed journals and conferences
•    High comfort level in creating real-time software
•    Solid knowledge on audio signal analysis, spatial analysis, content creation, and content generation 
•    Familiarity with Apple, Android or Windows ecosystem development environments and toolkits relevant for XR
•    Familiarity with real-time game engines such as Unreal Engine and Unity.
•    Excellent problem-solving and partnership skills
•    Excellent communication and presentation skills

 

*LI-SB1

Apply now »