.png?1762423916)
Machine Learning Engineer , AI
Biohub is the first large-scale initiative bringing frontier AI models, massive compute, and frontier experimental capabilities under one roof. We're building a general-purpose system to accelerate scientific discovery, integrating frontier AI models, biological foundation models, and lab capabilities, with the ultimate goal of curing disease. Our technology powers scientists around the world, translating AI capabilities into tools that accelerate research everywhere.
Biohub operates one of the largest AI compute clusters dedicated to biology, spanning three frontier research institutes with some of the world's leading biologists. We're not a startup trying to find product-market fit, and we're not a pharma company optimizing a pipeline. We're building frontier AI for fundamental science, as open science, at a scale no one else is doing. This is a unique moment for scientific acceleration. The problems are among the hardest and most impactful problems you can choose to work on, and we move at a pace that meets this moment.
Our research spans:
- Frontier molecular modeling, from protein language models (e.g., ESM) to structure prediction (e.g., ESMFold) and beyond.
- Scaled biological foundation models trained on some of the largest GPU clusters dedicated to science
- Imaging foundation models trained across the world's largest microscopy datasets
- Reasoning and agentic systems that connect frontier LLMs with biological foundation models
- Mechanistic interpretability of biological foundation models: extracting new biological knowledge directly from model weights
- Scientific data at unprecedented scale: AI systems to collect, curate, and learn from some of the richest biological datasets ever assembled
Join our Team!
As an ML Engineer, you'll join some of the strongest infrastructure engineers in AI, building the systems that connect everything together. The infrastructure problems you solve directly determine what science becomes possible.
What You'll Do
- Work with high-dimensional scientific data formats and contribute to backend compatibility, format evaluation, and I/O performance benchmarking at petabyte scale.
- Define and shape the engineering patterns your team and collaborating researchers will build on for years; the abstractions you write today become the foundation others depend on at scale.
- Work at the intersection of AI systems and biological discovery, where the infrastructure problems you solve directly determine what science becomes possible.
- Deploy models to production and manage artifact tracking across models and datasets.
- Design and optimize GPU-native data loading pipelines for large-scale multi-dimensional tensor workloads, including profiling and resolving hardware utilization bottlenecks across multi-backend systems.
- Work on simplification and improvement of codebase abstractions to accelerate research momentum.
- Build and maintain primitives for pre-training infrastructure that ensure the reliability and continuity of large-scale training runs.
- Help cultivate best practices in MLOps, and think about the full ML lifecycle, including data, fine-tuning, deployment, reliability and monitoring.
- Possesses the ability to execute complex modifications to the research pipeline, such as fast data loading and distributed training.
- Handle DevOps responsibilities, focused on making all engineers and researchers more productive. This includes tasks like cluster monitoring, unit testing and integration testing of research codebase, and continuous integration.
- Collaborate with partner researchers and engineers to deploy our technology within external infrastructure.
What You'll Bring
- Hands-on experience with PyTorch, including custom training loops, distributed training, or low-level performance work.
- Familiarity with GPU-native data I/O tools and large-scale tensor formats (e.g. Zarr, HDF5, TensorStore, or similar).
- Experience with distributed computing frameworks such as Apache Spark, Dask, or Ray.
- Familiarity with containerization and orchestration tools such as Docker and Kubernetes.
- Experience building or working with AI agent frameworks is a plus.
- A track record of building systems that other engineers and researchers depend on. Not just running experiments, but shipping infrastructure that scales.
Compensation
The future anticipated Redwood City, CA, and New York City, NY base pay range for a role in this field is $150,000 to $350,000+ annually. Compensation ranges will vary based on job-related skills, level of experience, and knowledge. Actual placement in range is based on job-related skills and experience, as evaluated throughout the interview process.
Benefits for the Whole You
We’re thankful to have an incredible team behind our work. To honor their commitment, we offer a wide range of benefits to support the people who make all we do possible.
- Provides a generous employer match on employee 401(k) contributions to support planning for the future.
- Paid time off to volunteer at an organization of your choice.
- Funding for select family-forming benefits.
- Relocation support for employees who need assistance moving
Please note that applying to this opportunity does not guarantee that we will be in touch with you regarding our opportunities. Our recruiting team will contact you if your experience aligns with the skills we seek for future open positions. We will keep your interest on file, contact you as opportunities arise, and send you information about the exciting work we are doing at Biohub. You can opt out at any time!
#LI-Hybrid
Apply for this job
*
indicates a required field