at Apple
Location
Seattle, United States of America
Compensation
$140k–$258k USD
Type
full time
Posted
3 weeks ago
Market range · company + function + seniority
p25 · target · p75 · n=515
Posted $258k · below the band
Tailor your résumé to this role in 30 seconds.
Free account · ATS keyword check · per-job bullet rewrite by Claude.
The Human-centered AI, Data Quality Operations team is looking for a Senior Applied Scientist to join our growing team. We are building the systems and methodologies that make AI evaluation trustworthy, and scalable — directly shaping how Apple develops and validates AI across products and services. In this role, you will develop novel, scalable quality control solutions, working closely with cross-functional teams to ensure the data powering our AI/ML systems meets the highest standards of accuracy, consistency, and relevance.
Your work will span two connected problem spaces. The first is the methodology and tooling that generates reliable ground truth and detects quality failures across human annotation and automated evaluation pipelines. The second is the autonomous QA agents that make those methodologies generalizable across teams and use cases. This role demands fluency across research thinking and engineering execution — you will prototype, validate, and ship. A strong point of view on when not to use a model or agent is as valued here as the ability to build one.
Design and implement scalable ground truth generation pipelines across varied task types, annotation modalities, and cold start conditions
Build and maintain calibration frameworks that keep LLM evaluators anchored to human judgment over time
Develop anomaly detection systems that surface evaluator drift, distribution shifts, and coverage gaps across human annotation and automated evaluation pipelines
Design, build, and deploy autonomous QA agents targeting specific facets of evaluation quality, architected for generalizability and self-service adoption across teams
Partner closely with cross-functional teams to ensure evaluation systems meet the highest standards of accuracy, consistency, and relevance
Communicate findings and recommendations clearly to both technical and non-technical stakeholders, including senior leadership
Contribute to a culture of technical excellence by sharing knowledge and best practices across the team
5+ years of industry experience in applied science or machine learning with demonstrated impact on shipped systems
Strong hands-on experience with Large Language Models including prompt engineering and applied use cases such as grading, validation, or classification
Strong working knowledge of evaluation methodology for generative AI, including LLM-as-a-judge design, meta-evaluation, and failure mode analysis
Familiarity with human-in-the-loop evaluation systems and the operational dynamics that affect data quality at scale
Hands-on experience designing ground truth generation pipelines across varied task types and annotation modalities
Proficiency in Python and relevant ML frameworks, with production experience building, deploying, and monitoring LLM-based pipelines and agents
MS or PhD in Computer Science, Machine Learning, Statistics, or a related quantitative field, or equivalent practical experience
PhD in Computer Science, Machine Learning, Statistics, or a related field
Experience designing agent architectures that are configurable and extensible by practitioners who did not build them
Hands-on experience building anomaly detection systems for evaluation quality, including drift detection, distribution analysis, and systematic bias identification
Strong communication skills with the ability to influence technical direction across cross-functional teams
Demonstrated passion for leveraging AI to improve work efficiency and scale
Apple Services Engineering (ASE) powers the AI and LLM features behind experiences that hundreds of millions of users love every day. As these systems increasingly rely on human-in-the-loop evaluation, the quality of our products is directly constrained by the quality of our evaluation systems. We believe that to build exceptional AI, you need exceptional mechanisms to validate the signals used to train and evaluate them.
At Apple, base pay is one part of our total compensation package and is determined within a range. This provides the opportunity to progress as you grow and develop within a role. The base pay range for this role is between $139,500 and $258,100, and your base pay will depend on your skills, qualifications, experience, and location.Apple is an equal opportunity employer that is committed to inclusion and diversity. We seek to promote equal opportunity for all applicants without regard to race, color, religion, sex, sexual orientation, gender identity, national origin, disability, Veteran status, or other legally protected characteristics. Learn more about your EEO rights as an applicant
At Apple, we believe accessibility is a fundamental human right. You’ll find that idea reflected in everything here — in our culture, our benefits and our digital tools. By welcoming as many perspectives as possible, we help you build a career where you feel like you belong.
Learn about accessibility in Apple’s workplace
Learn about reasonable accommodations for job applicants
Apple accepts applications to this posting on an ongoing basis.