About Us: We build power-efficient, low-precision foundation models designed to run from edge devices to large-scale deployments. We train models ranging from roughly 1B to 100B+ parameters across LLMs, diffusion models, and other modalities, with a strong emphasis on efficient training, inference, and real-world deployment under power and memory constraints.
Role Overview: We are seeking a Staff-level (or higher) multimodal expert to lead the development of multimodal capabilities that expand monetization and consumer reach for our edge-optimized models. This role focuses on building vision, speech, and other modality components that integrate tightly with our core models, while providing technical leadership across multimodal systems.
Responsibilities: You will design, build, and integrate multimodal components optimized for efficiency, quality, and deployability. Key responsibilities include:
Basic Qualifications: You have a strong background in multimodal ML systems and technical leadership, including:
Preferred Qualifications: You bring experience that directly supports consumer and edge-focused AI products, including:
Ideal Candidate Profile: You enjoy turning foundation models into usable products, understand how multimodal systems unlock consumer value, think deeply about efficiency and deployment constraints, and naturally take ownership of technical direction while helping other engineers grow.