Applied Machine Learning Scientist - Visual Generation Job at Apple in Seattle
Apple, Seattle, WA, US, 98127
Role Number: 200615083-3337
Summary
The VCV Science Team is pioneering the future of visual creativity through innovative AI. We're seeking an Applied ML Scientist who thrives at the intersection of computer vision and generative AI to help us build breakthrough experiences with image, video, and 3D generation technologies that will delight and inspire millions of users worldwide.
As part of the team, you will have the opportunity to incubate powerful research ideas, partner with teams across Apple to push the boundaries of what’s possible, and transform groundbreaking research into magical user experiences!
Description
As an Applied ML Scientist on our team, you will:
-
Design and develop pioneering generative models for visual content such as image, video, and 3D.
-
Stay at the forefront of artificial intelligence and machine learning advancements. Continuously exploring and evaluating new technologies and methodologies to enhance the technical capabilities of our team.
-
Push boundaries by rapidly prototyping ideas, testing hypotheses, and exploring novel approaches to visual generation challenges, including exploring new architectures ranging from diffusion, auto-regressive, multi-modal generation and/or hybrid approaches.
-
Partner with world-class researchers, engineers, and designers to transform prototypes into robust, production-ready features that will delight users while maintaining Apple's standards for quality and privacy.
-
Explore distillation and optimization techniques, that will enable the models to run efficiently, while maintaining quality.
-
Present your work directly to executive leadership and shape the roadmap for how AI will transform creative tools across Apple's ecosystem.
Minimum Qualifications
-
M.S. in Computer Science, Machine Learning, Computer vision or related field.
-
Solid understanding and experience with recent visual generative AI models, such as multi-modal LLMs, and diffusion-based large vision models.
-
Proficiency in Python, and experience in using ML toolkits, e.g., PyTorch, JAX etc.
-
Knowledge of and proficiency with ML-based product lifecycle, and methods for model training, statistical analysis and data science.
Preferred Qualifications
-
5+ years of industry or academic experience in CVML and an advanced degree (M.Sc./Ph.D.) in CS, CVML or similar field.
-
Hands-on experience with building, training from scratch and scaling innovative visual generation models, such as large image/video diffusion models, multi-modal LLMs, contrastive learning models, or other visual foundation models.
-
Strong software engineering skills to build scalable and robust infrastructure for deep learning data, modeling, and evaluation systems.
-
Comfort with researching current ML literature and math including optimization methods and modeling techniques.
-
Proven track record in research, innovation and/or delivering ML products, demonstrated through publications in top-tier journals or conferences, patents, or impactful software developments.
-
Strong collaboration skills in a multi-functional setting, and ability to leverage partnerships to drive outcomes and deliver impact at scale.
Apple is an equal opportunity employer that is committed to inclusion and diversity. We seek to promote equal opportunity for all applicants without regard to race, color, religion, sex, sexual orientation, gender identity, national origin, disability, Veteran status, or other legally protected characteristics. Learn more about your EEO rights as an applicant (https://www.eeoc.gov/sites/default/files/2023-06/22-088_EEOC_KnowYourRights6.12ScreenRdr.pdf) .