Google DeepMind
Research Scientist, World Modeling
Full-time
On-site, Mountain View, CA
$136k–$245k
Yesterday
Please mention that you found this job through
XR Jobs Board—it helps us grow!
About
Join an ambitious project to build generative models that simulate the physical world. We believe scaling pretraining on video and multimodal data is on the critical path to artificial general intelligence. World models will power numerous domains, such as visual reasoning and simulation, planning for embodied agents, and real-time interactive entertainment. The team will collaborate with and build on work from Gemini, Veo and Genie teams, and tackle critical new problems to scale world models to the highest levels of compute.
Key Responsibilities
- Implement core infrastructure and conduct research to build generative models of the physical world
- Solve essential problems to train world simulators at massive scale
- Develop metrics and scaling laws for physical intelligence
- Curate and annotate training data
- Enable real-time interactive generation
- Study integration of world models with multimodal language models
- Embrace the bitter lesson and seek simple methods that scale, with emphasis on strong systems and infrastructure
Areas of Focus
- Systems for training multimodal transformers at massive scale
- Infrastructure for large-scale video data pipelines and annotation
- Inference optimization and distillation for real-time generation
- Methods for native multimodal generation in language models
- Methods for ultra-long-context transformers
- Quantitative evals for physical accuracy and intelligence
- Scaling law science for video pretraining
Requirements
- Experience with large-scale transformer models and/or large-scale data pipelines
- MSc or PhD in computer science or machine learning, or equivalent industry experience
- Track record of releases, publications, and/or open source projects relating to video generation, world models, multimodal language models, or transformer architectures
- Strong systems and engineering skills in deep learning frameworks like JAX or PyTorch
Preferred Qualifications
- Experience building training codebases for large-scale video or multimodal transformers
- Expertise optimizing efficiency of distributed training systems and/or inference systems
About Google DeepMind
Artificial Intelligence could be one of humanity's most useful inventions. At Google DeepMind, we're a team of scientists, engineers, machine learning experts and more, working together to advance the state of the art in artificial intelligence. We use our technologies for widespread public benefit and scientific discovery, and collaborate with others on critical challenges, ensuring safety and ethics are the highest priority.
Equal Opportunity
At Google DeepMind, we value diversity of experience, knowledge, backgrounds and perspectives and harness these qualities to create extraordinary impact. We are committed to equal employment opportunities regardless of sex, race, religion or belief, ethnic or national origin, disability, age, citizenship, marital, domestic or civil partnership status, sexual orientation, gender identity, pregnancy, or related condition (including breastfeeding) or any other basis as protected by applicable law. If you have a disability or additional need that requires accommodation, please do not hesitate to let us know.
Please mention that you found this job through
XR Jobs Board—it helps us grow!