ML Research Engineer Green Card Jobs
ML Research Engineer roles qualify for EB-2 green card sponsorship when the position requires an advanced degree in machine learning, computer science, or a related field. Employers file a PERM labor certification with DOL before sponsoring your I-140 petition, starting the path to permanent residency rather than a renewable work visa.
See All ML Research Engineer JobsOverview
Showing 5 of 367+ ML Research Engineer jobs


Have you applied for this role?


Have you applied for this role?


Have you applied for this role?


Have you applied for this role?


Have you applied for this role?
See all 367+ ML Research Engineer jobs
Sign up for free to unlock all listings, filter by visa type, and get alerts for new ML Research Engineer roles.
Get Access To All Jobs
INTRODUCTION
AI systems are only as trustworthy as the methods used to evaluate them. At Apple, where AI powers experiences for billions of people, getting evaluation right is not a support function - it is a foundational science. Our team, part of Apple Services Engineering, is building that scientific foundation: rigorous, scalable evaluation methodology for LLMs, agentic systems, and human-AI interaction.
What makes this team unusual is its interdisciplinary core. You will work alongside measurement scientists (psychometrics, validity theory), ML researchers, and platform engineers - bringing together ML research, statistical rigor, and production engineering. We are looking for an ML Research Engineer who can move fluidly across this landscape: someone who loves implementing the latest techniques in AI, has the engineering instincts to make them robust and scalable, and thrives at the intersection of research and production.
ROLE AND RESPONSIBILITIES
This is a combined research and engineering role, sitting with and between research/applied scientists and platform engineers. New evaluation research can be challenging to use at scale - that's where your skills in both machine learning and engineering come into play.
On the research side, you will partner with scientists to rapidly prototype their ideas, implement methods from recent papers, run large-scale experiments, and provide critical feedback grounded in your engineering experience. On the engineering side, you will work with platform engineers to bring those research prototypes into production - moving from Python packages on local machines to robust services deployed in the cloud.
-
Rapid Prototyping & Experimentation: Collaborate with research and applied scientists to translate evaluation research ideas into working prototypes - implementing methods from recent papers, building experimental pipelines, and iterating quickly to validate hypotheses in areas such as preference learning, LLM-as-judge calibration, and automated failure discovery.
-
Research-to-Production Bridge: Own the lifecycle of moving evaluation methods from research prototypes to production-ready systems. Refactor research code into robust, well-tested Python packages and partner with platform engineers to deploy them as scalable services, APIs, and SDK components.
-
Experiment Infrastructure: Design and maintain the infrastructure for running large-scale evaluation experiments - orchestrating LLM judge calls, managing datasets, tracking experiment results, and ensuring reproducibility across the team's research portfolio.
-
Technical Feedback & Collaboration: Serve as a critical technical partner to researchers, providing engineering perspective on feasibility, scalability, and system design. Identify opportunities where engineering improvements (parallelization, caching, smarter batching) can unlock new research directions or dramatically accelerate experimentation.
-
Scaling Evaluation Methods: Identify bottlenecks in evaluation workflows and engineer solutions to operate at Apple scale - optimizing for throughput, cost, and reliability when running evaluation methods across large model populations and diverse use cases.
-
Code Quality & Engineering Standards: Champion engineering best practices within the research workflow, including version control, automated testing, documentation, and CI/CD, raising the bar for code quality across the research-engineering boundary.
-
Cross-Functional Integration: Work across the research and platform engineering teams to ensure that evaluation methods integrate seamlessly with Apple's broader ML infrastructure, developer workflows, and internal tooling ecosystem.
BASIC QUALIFICATIONS
-
Bachelor's degree in Computer Science, Machine Learning, Software Engineering, or a closely related field (Master's preferred)
-
2+ years of hands-on experience in a role combining machine learning and software engineering (e.g., ML engineer, research engineer, or applied scientist with strong engineering output), or a Master's degree in Computer Science, Machine Learning, or a closely related field with relevant project experience
-
Strong proficiency in Python and the modern ML ecosystem (PyTorch, JAX, or TensorFlow), with demonstrated ability to implement complex methods from recent ML papers
-
Solid software engineering fundamentals: clean code design, version control, testing, debugging, and performance optimization
-
Experience working with large language models - whether fine-tuning, inference, prompting pipelines, or building LLM-powered applications
-
Demonstrated ability to work across the research-to-production spectrum: you have taken experimental or prototype code and made it robust, scalable, and usable by others
-
Practical experience with cloud-native development and deployment: containerization (Docker/Kubernetes), CI/CD pipelines, and distributed computing frameworks (e.g., Ray, Spark)
-
Strong communication skills and comfort working in interdisciplinary teams, with the ability to engage productively with both researchers and platform engineers
-
Comfort with ambiguity and new problem spaces - you thrive when building something that doesn't yet have a playbook
PREFERRED QUALIFICATIONS
-
Master's or Ph.D. in Computer Science, Machine Learning, or a related field
-
Experience with evaluation-specific methods or frameworks: LLM-as-judge approaches, reward modeling, RLHF, calibration techniques, benchmark design, or human evaluation methodology
-
Familiarity with modern evaluation tools and frameworks (e.g., DeepEval, Ragas, TruLens, LangSmith) and an understanding of how to implement and scale model-based evaluation workflows
-
Track record of contributing to research outputs - co-authored publications, open-source contributions, or internal research reports - even if research is not your primary role
-
Experience with the engineering challenges specific to generative AI and agentic systems: managing token economics, handling non-deterministic outputs, evaluating multi-turn agent trajectories and tool usage
-
Familiarity with statistical concepts relevant to evaluation: calibration, inter-rater reliability, scoring rules, or measurement validity
-
Experience in fast-moving, early-stage teams where you helped define technical direction and engineering culture from the ground up
PAY & BENEFITS
At Apple, base pay is one part of our total compensation package and is determined within a range. This provides the opportunity to progress as you grow and develop within a role. The base pay range for this role is between $171,600 and $258,100, and your base pay will depend on your skills, qualifications, experience, and location.
Apple employees also have the opportunity to become an Apple shareholder through participation in Apple's discretionary employee stock programs. Apple employees are eligible for discretionary restricted stock unit awards, and can purchase Apple stock at a discount if voluntarily participating in Apple's Employee Stock Purchase Plan. You'll also receive benefits including: Comprehensive medical and dental coverage, retirement benefits, a range of discounted products and free services, and for formal education related to advancing your career at Apple, reimbursement for certain educational expenses - including tuition. Additionally, this role might be eligible for discretionary bonuses or commission payments as well as relocation.
Apple is an equal opportunity employer that is committed to inclusion and diversity. We seek to promote equal opportunity for all applicants without regard to race, color, religion, sex, sexual orientation, gender identity, national origin, disability, Veteran status, or other legally protected characteristics. Learn more about your EEO rights as an applicant.
Note: Apple benefit, compensation and employee stock programs are subject to eligibility requirements and other terms of the applicable plan or program.

INTRODUCTION
AI systems are only as trustworthy as the methods used to evaluate them. At Apple, where AI powers experiences for billions of people, getting evaluation right is not a support function - it is a foundational science. Our team, part of Apple Services Engineering, is building that scientific foundation: rigorous, scalable evaluation methodology for LLMs, agentic systems, and human-AI interaction.
What makes this team unusual is its interdisciplinary core. You will work alongside measurement scientists (psychometrics, validity theory), ML researchers, and platform engineers - bringing together ML research, statistical rigor, and production engineering. We are looking for an ML Research Engineer who can move fluidly across this landscape: someone who loves implementing the latest techniques in AI, has the engineering instincts to make them robust and scalable, and thrives at the intersection of research and production.
ROLE AND RESPONSIBILITIES
This is a combined research and engineering role, sitting with and between research/applied scientists and platform engineers. New evaluation research can be challenging to use at scale - that's where your skills in both machine learning and engineering come into play.
On the research side, you will partner with scientists to rapidly prototype their ideas, implement methods from recent papers, run large-scale experiments, and provide critical feedback grounded in your engineering experience. On the engineering side, you will work with platform engineers to bring those research prototypes into production - moving from Python packages on local machines to robust services deployed in the cloud.
-
Rapid Prototyping & Experimentation: Collaborate with research and applied scientists to translate evaluation research ideas into working prototypes - implementing methods from recent papers, building experimental pipelines, and iterating quickly to validate hypotheses in areas such as preference learning, LLM-as-judge calibration, and automated failure discovery.
-
Research-to-Production Bridge: Own the lifecycle of moving evaluation methods from research prototypes to production-ready systems. Refactor research code into robust, well-tested Python packages and partner with platform engineers to deploy them as scalable services, APIs, and SDK components.
-
Experiment Infrastructure: Design and maintain the infrastructure for running large-scale evaluation experiments - orchestrating LLM judge calls, managing datasets, tracking experiment results, and ensuring reproducibility across the team's research portfolio.
-
Technical Feedback & Collaboration: Serve as a critical technical partner to researchers, providing engineering perspective on feasibility, scalability, and system design. Identify opportunities where engineering improvements (parallelization, caching, smarter batching) can unlock new research directions or dramatically accelerate experimentation.
-
Scaling Evaluation Methods: Identify bottlenecks in evaluation workflows and engineer solutions to operate at Apple scale - optimizing for throughput, cost, and reliability when running evaluation methods across large model populations and diverse use cases.
-
Code Quality & Engineering Standards: Champion engineering best practices within the research workflow, including version control, automated testing, documentation, and CI/CD, raising the bar for code quality across the research-engineering boundary.
-
Cross-Functional Integration: Work across the research and platform engineering teams to ensure that evaluation methods integrate seamlessly with Apple's broader ML infrastructure, developer workflows, and internal tooling ecosystem.
BASIC QUALIFICATIONS
-
Bachelor's degree in Computer Science, Machine Learning, Software Engineering, or a closely related field (Master's preferred)
-
2+ years of hands-on experience in a role combining machine learning and software engineering (e.g., ML engineer, research engineer, or applied scientist with strong engineering output), or a Master's degree in Computer Science, Machine Learning, or a closely related field with relevant project experience
-
Strong proficiency in Python and the modern ML ecosystem (PyTorch, JAX, or TensorFlow), with demonstrated ability to implement complex methods from recent ML papers
-
Solid software engineering fundamentals: clean code design, version control, testing, debugging, and performance optimization
-
Experience working with large language models - whether fine-tuning, inference, prompting pipelines, or building LLM-powered applications
-
Demonstrated ability to work across the research-to-production spectrum: you have taken experimental or prototype code and made it robust, scalable, and usable by others
-
Practical experience with cloud-native development and deployment: containerization (Docker/Kubernetes), CI/CD pipelines, and distributed computing frameworks (e.g., Ray, Spark)
-
Strong communication skills and comfort working in interdisciplinary teams, with the ability to engage productively with both researchers and platform engineers
-
Comfort with ambiguity and new problem spaces - you thrive when building something that doesn't yet have a playbook
PREFERRED QUALIFICATIONS
-
Master's or Ph.D. in Computer Science, Machine Learning, or a related field
-
Experience with evaluation-specific methods or frameworks: LLM-as-judge approaches, reward modeling, RLHF, calibration techniques, benchmark design, or human evaluation methodology
-
Familiarity with modern evaluation tools and frameworks (e.g., DeepEval, Ragas, TruLens, LangSmith) and an understanding of how to implement and scale model-based evaluation workflows
-
Track record of contributing to research outputs - co-authored publications, open-source contributions, or internal research reports - even if research is not your primary role
-
Experience with the engineering challenges specific to generative AI and agentic systems: managing token economics, handling non-deterministic outputs, evaluating multi-turn agent trajectories and tool usage
-
Familiarity with statistical concepts relevant to evaluation: calibration, inter-rater reliability, scoring rules, or measurement validity
-
Experience in fast-moving, early-stage teams where you helped define technical direction and engineering culture from the ground up
PAY & BENEFITS
At Apple, base pay is one part of our total compensation package and is determined within a range. This provides the opportunity to progress as you grow and develop within a role. The base pay range for this role is between $171,600 and $258,100, and your base pay will depend on your skills, qualifications, experience, and location.
Apple employees also have the opportunity to become an Apple shareholder through participation in Apple's discretionary employee stock programs. Apple employees are eligible for discretionary restricted stock unit awards, and can purchase Apple stock at a discount if voluntarily participating in Apple's Employee Stock Purchase Plan. You'll also receive benefits including: Comprehensive medical and dental coverage, retirement benefits, a range of discounted products and free services, and for formal education related to advancing your career at Apple, reimbursement for certain educational expenses - including tuition. Additionally, this role might be eligible for discretionary bonuses or commission payments as well as relocation.
Apple is an equal opportunity employer that is committed to inclusion and diversity. We seek to promote equal opportunity for all applicants without regard to race, color, religion, sex, sexual orientation, gender identity, national origin, disability, Veteran status, or other legally protected characteristics. Learn more about your EEO rights as an applicant.
Note: Apple benefit, compensation and employee stock programs are subject to eligibility requirements and other terms of the applicable plan or program.
See all 367+ ML Research Engineer jobs
Sign up for free to unlock all listings, filter by visa type, and get alerts for new ML Research Engineer roles.
Get Access To All JobsTips for Finding Green Card Sponsorship as a ML Research Engineer
Document your research contributions before applying
Publications, patents, conference presentations, and open-source contributions strengthen your PERM case and support an EB-2 classification. Compile these into a credentials portfolio before you reach the offer stage, since employers will need them for the I-140 petition.
Target labs with established PERM filing history
Research divisions at large tech companies and national labs file PERM applications regularly and have dedicated immigration counsel on retainer. These employers already understand the ML Research Engineer job duties DOL expects to see in the labor certification.
Use Migrate Mate to find green-card-sponsoring ML employers
Search Migrate Mate to surface employers with active EB-2 and EB-3 sponsorship history for research engineering roles. This filters out postings where sponsorship is stated as possible but has no track record behind it.
Understand how your country affects priority date timing
EB-2 and EB-3 visa backlogs vary significantly by birth country. If you're born in India or China, the wait for a current priority date can stretch years, so confirming your category and country situation early shapes how you negotiate offer timelines with employers.
Confirm the job description matches DOL specialty occupation criteria
PERM requires the posted role to specify a minimum of a bachelor's degree in a directly related field. Vague ML Research Engineer descriptions listing preferred qualifications instead of required ones can trigger a DOL audit or denial during labor certification.
Ask whether the employer uses concurrent I-485 filing
When a visa number is immediately available, some employers file the I-485 adjustment of status concurrently with the I-140, cutting overall processing time. Clarify this with the employer's immigration counsel before signing an offer, since USCIS processing windows affect your work authorization continuity.
ML Research Engineer jobs are hiring across the US. Find yours.
Find ML Research Engineer JobsML Research Engineer Green Card Sponsorship: Frequently Asked Questions
Does an ML Research Engineer role qualify for EB-2 or EB-3 green card sponsorship?
ML Research Engineer positions typically qualify for EB-2 because the role requires at minimum a master's degree or a bachelor's degree plus at least five years of progressive research experience in machine learning or a related field. Employers document this requirement in the PERM labor certification. EB-3 is available for candidates whose credentials or the specific job description fit the skilled worker category instead.
How does PERM green card sponsorship differ from H-1B sponsorship for this role?
H-1B is a temporary work visa subject to an annual lottery and a six-year maximum stay. PERM-based green card sponsorship leads to permanent residency with no renewal required. There is no lottery for EB-2 or EB-3 categories, though per-country visa backlogs can extend the timeline for nationals of India and China. The PERM process also requires the employer to conduct a supervised recruitment process with DOL before filing the I-140 petition.
What does the PERM labor certification process look like for an ML Research Engineer?
The employer files a PERM application with DOL after completing a supervised recruitment process that demonstrates no qualified U.S. workers were available for the role. For ML Research Engineer positions, the job description must specify a minimum degree requirement in a directly related field such as computer science or machine learning. DOL processing currently averages several months before certification, after which the employer files the I-140 immigrant petition with USCIS.
How can I find ML Research Engineer jobs where the employer will sponsor a green card?
Migrate Mate lets you search specifically for employers with verified EB-2 and EB-3 sponsorship history for research engineering roles, so you're not relying on job postings that claim sponsorship is possible without any track record. Filtering by sponsorship history saves significant time compared to applying broadly and discovering late in the process that an employer won't commit to PERM.
Can I use O*NET to verify that my ML Research Engineer role qualifies as a specialty occupation?
Yes. The O*NET occupation profile for machine learning or related research engineering titles lists the typical education and training requirements that DOL uses when reviewing PERM applications. If the O*NET profile for your specific job code shows a requirement of a bachelor's or higher degree in a directly related field, that supports the employer's specialty occupation argument in both the PERM labor certification and the I-140 petition.
See which ML Research Engineer employers are hiring and sponsoring visas right now.
Search ML Research Engineer Jobs