| |
|
| |
From real-time multilingual voice agents that translate & dub with perfect cultural accuracy to emotionally intelligent audio systems that detect tone, intent, & even unspoken context, the frontier of AI is expanding into the auditory world.
But building speech & audio AI is not just a modeling challenge. It requires massive advances in high-quality audio data, multimodal fusion, on-device inference, & real-world evaluation at scale.
Join Centific for an evening exploring the emerging stack powering the next generation of voice & audio intelligence.
Dr Dinesh Manocha (University of Maryland - College Park) & contributor to Nvidia's Audio-Flamingo Speech Models, will be a speaker & panelist.
Dr. Victor Barres, PhD neuroscientist shaping the future of conversational AI at Sierra will also join the panel
Harshit Rajgarhria, Senior Manager of AI Research at Centific & ICML 2026 author whose MedMosaic benchmark is advancing medical audio AI will be panel moderator.
Other panelists TBA
Agenda:
6:00 - 7:00: Cocktail Hour & Networking
7:00 - 7:05: Welcome & Introductions
7:05 - 7:15: Dr. Manocha's talk "Flexible & Open Audio-LLM Models" - GAMMA Lab UMD
7:15 - 8:00: Panel Discussion
8:00 - 8:05: Q&A
8:05 - 9:00: Networking & Cocktails
About our host: Centific helps model labs & enterprises build, train, deploy, & govern intelligent systems through high-quality data, human expertise, & end-to-end platforms that turn complexity into scalable, real-world impact. Learn more about Centific at www.centific.com.
|
|
|
|
|
|
|
|