DeepMind
Our team is responsible for building the Productionization and Serving infrastructure for Gemini Audio Inference at Google DeepMind. We help land Gemini Audio capabilities into numerous clients including Astra, GeminiApp, YouTube, Search, Meet, Cloud, Geo, Assistant etc. Based on research Gemini model flavors, our tasks involve model latency/throughput optimizations, serializations, orchestration, evaluation & finally landing in production, at Google scale. Our team focuses deeply on Inference efficiencies for Gemini and its related components. We actively develop new infrastructure to make Gemini more accessible to new streaming use cases. Our team has deep collaboration with both the research team and production platform team, exposed with SOTA research work and their inference optimizations. Our team owns the infra to serve the Audio tokenization & Audio generation around the Gemini models. Join us if you are interested in having a direct impact on making Google's products better for our users in over 100 languages! Here are showcase videos that directly used the infra built from our team: Gemini Audio, Project Astra, Real Time Translation.
At Google DeepMind, we value diversity of experience, knowledge, backgrounds and perspectives and harness these qualities to create extraordinary impact. We are committed to equal employment opportunities regardless of sex, race, religion or belief, ethnic or national origin, disability, age, citizenship, marital, domestic or civil partnership status, sexual orientation, gender identity, pregnancy, or related condition (including breastfeeding) or any other basis as protected by applicable law.
Key skills identified from this job posting
Sign upto access all insights for this job
Website
deepmind.com
Company Size
1001-5000 employees
Location
Mountain View, CA
Industry
Software Publishers
Other opportunities you might be interested in
Actalent
Boeing
Infosys
Adobe
Adobe