Senior Multimodal AI Researcher
Date: Apr 16, 2026
Location: Atlanta, US
Company: Dolby Laboratories, Inc.
Join the leader in entertainment innovation and help us design the future. At Dolby, science meets art, and high tech means more than computer code. As a member of the Dolby team, you’ll see and hear the results of your work everywhere, from movie theaters to smartphones. We continue to revolutionize how people create, deliver, and enjoy entertainment worldwide. To do that, we need the absolute best talent. We’re big enough to give you all the resources you need, and small enough so you can make a real difference and earn recognition for your work. We offer a collegial culture, challenging projects, and excellent compensation and benefits, not to mention a Flex Work approach that is truly flexible to support where, when, and how you do your best work.
The Advanced Technology Group (ATG) is the research division of the company. ATG’s mission is to look ahead, deliver insights, and innovate technological solutions that will fuel Dolby’s continued growth. Our researchers have a broad range of expertise related to computer science and electrical engineering, such as AI/ML, algorithms, digital signal processing, audio engineering, image processing, computer vision, data science & analytics, distributed systems, cloud, edge & mobile computing, computer networking, and IoT.
We are seeking a talented Senior Multimodal AI Researcher to join the Multimodal Experiences Lab within Dolby’s Advanced Technology Group. Our team invents next-generation technologies that power immersive audio-visual experiences, combining machine learning with decades of leadership in perceptual signal processing.
In this role, you will develop next-generation multimodal generative AI systems for immersive audio-visual experiences. You will work at the intersection of deep learning, perceptual modeling, and media signal processing, building systems that generate, enhance, and render realistic audio-visual content at scale.
This role spans fundamental research through production deployment, with opportunities to publish, patent, and directly influence shipped technologies.
What you’ll do (Key Responsibilities)
- Design and develop novel multimodal AI models for immersive audio‑visual content creation, enhancement, and rendering
- Combine deep learning with perceptually grounded approaches (signal processing, human perception, and quality metrics)
- Build prototypes and evaluate systems using both objective metrics and human/perceptual assessment
- Partner with cross-functional teams to translate research into robust, production-ready solutions
- Contribute to Dolby’s innovation pipeline through publications, patents, and technology transfer
What you’ll need to succeed (How you’ll work)
- Strong foundation in ML/AI with the ability to develop new algorithms and model architectures
- Interest in perceptual media (audio, vision, and human perception) and curiosity about emerging approaches
- A collaborative, iterative mindset—comfortable working across research and engineering teams
- Clear communication skills: sharing ideas, reasoning about tradeoffs, and learning from feedback
What you bring (Required)
- PhD in Computer Science, Electrical Engineering, or related field or equivalent industry research experience
- Research experience in one or more of: multimodal AI, generative AI, computer vision, audio analysis, audiovisual analysis
- Demonstrated research impact through publications and/or patents in top venues (e.g., NeurIPS, ICLR, ICML, CVPR, ICCV) or equivalent industry impact
- Proficiency in Python and/or C/C++, with hands-on prototyping experience building and evaluating ML systems
- Strong cross-functional collaboration skills and the ability to communicate effectively with technical and non-technical partners
What’s a plus (Desired / Bonus)
- Experience with perceptual quality metrics for video and/or audio (objective + subjective evaluation)
- Experience working on real-time or production-scale media systems
- Experience with model compression/optimization and deployment of deep learning models
The Atlanta Area base salary range for this full-time position is $140,000-$170,000, which can vary if outside this location, plus bonus, benefits, and some roles may also include equity. Our salary ranges are determined by role, level, and location. Within the range, individual pay is determined by work location and additional factors, including job-related skills, competencies, experience, market demands, internal parity, and relevant education or training. Your recruiter can share more about the specific salary range and perks and benefits for your location during the hiring process.
Dolby will consider qualified applicants with criminal histories in a manner consistent with the requirements of San Francisco Police Code, Article 49, and Administrative Code, Article 12
Equal Employment Opportunity:
Dolby is proud to be an equal opportunity employer. Our success depends on the combined skills and talents of all our employees. We are committed to making employment decisions without regard to race, religious creed, color, age, sex, sexual orientation, gender identity, national origin, religion, marital status, family status, medical condition, disability, military service, pregnancy, childbirth and related medical conditions or any other classification protected by federal, state, and local laws and ordinances.
Nearest Major Market: Atlanta