Our team is responsible for building the Productionization and Serving infrastructure for Gemini Audio Inference at Google DeepMind. We help land Gemini Audio capabilities into numerous clients including Astra, GeminiApp, YouTube, Search, Meet, Cloud, Geo, Assistant etc. Based on research Gemini model flavors, our tasks involve model latency/throughput optimizations, serializations, orchestration, evaluation & finally landing in production, at Google scale. Our team focuses deeply on Inference efficiencies for Gemini and its related components. We actively develop new infrastructure to make Gemini more accessible to new streaming use cases. We have deep collaboration with both the research team and production platform team, exposed with SOTA research work and their inference optimizations. Our team owns the infrastructure to serve the Audio tokenization & Audio generation around the Gemini models. Join us if you are interested in having a direct impact on making Google's products better for our users in over 100 languages!
At Google DeepMind, we value diversity of experience, knowledge, backgrounds and perspectives and harness these qualities to create extraordinary impact. We are committed to equal employment opportunities regardless of sex, race, religion or belief, ethnic or national origin, disability, age, citizenship, marital, domestic or civil partnership status, sexual orientation, gender identity, pregnancy, or related condition. We prioritize safety and ethics in our work and support accommodations for individuals with disabilities or additional needs.
Key skills identified from this job posting
Sign upto access all insights for this job
Website
google.com
Location
Mountain View, NY
Industry
Web Search Portals and All Other Information Services
Other opportunities you might be interested in
Actalent
Boeing
Infosys
Adobe
Adobe