AI ML Engineering Visa Sponsorship Jobs in California
California leads the country in AI and ML engineering hiring, with major employers like Google, Meta, Apple, and dozens of well-funded startups concentrated in the San Francisco Bay Area, Los Angeles, and San Diego. International candidates will find more visa sponsorship activity here than anywhere else in the United States, particularly for roles in model development, MLOps, and applied research.
See All AI ML Engineering JobsOverview
Showing 5 of 2,272+ AI ML Engineering Jobs in California with Visa Sponsorship jobs


Have you applied for this role?


Have you applied for this role?


Have you applied for this role?


Have you applied for this role?


Have you applied for this role?
See all 2,272+ AI ML Engineering Jobs in California with Visa Sponsorship
Sign up for free to unlock all listings, filter by visa type, and get alerts for new AI ML Engineering Jobs in California with Visa Sponsorship.
Get Access To All Jobs
Job Description
WHO WE ARE
Goldman Sachs is a leading global investment banking, securities and investment management firm that provides a wide range of services worldwide to a substantial and diversified client base that includes corporations, financial institutions, governments and high net-worth individuals. Founded in 1869, it is one of the oldest and largest investment banking firms. The firm is headquartered in New York and maintains offices in London, Bangalore, Frankfurt, Tokyo, Hong Kong and other major financial centers around the world. We are committed to growing our distinctive Culture and holding to our core values which always place our client's interests first. These values are reflected in our Business Principles, which emphasize integrity, commitment to excellence, innovation and teamwork.
Business Unit Overview
Enterprise Technology Operations (ETO) is a Business Unit within Core Engineering focused on running scalable production management services with a mandate of operational excellence and operational risk reduction achieved through large scale automation, best-in-class engineering, and application of data science and machine learning. The Production Runtime Experience (PRX) team in ETO applies software engineering and machine learning to production management services, processes, and activities to streamline monitoring, alerting, automation, and workflows.
TEAM OVERVIEW
The Machine Learning and Artificial Intelligence team in PRX applies advanced ML and GenAI to reduce the risk and cost of operating the firm’s large-scale compute infrastructure and extensive application estate. Building on strengths in statistical modelling, anomaly detection, predictive modelling, and time-series forecasting, we leverage foundational LLM Models to orchestrate multi-agent systems for automated production management services. By unifying classical ML with agentic AI, we deliver reliable, explainable, and cost-efficient operations at scale.
ROLE AND RESPONSIBILITIES
In this role, you will be responsible for launching and implementing GenAI agentic solutions aimed at reducing the risk and cost of managing large-scale production environments with varying complexities. You will address various production runtime challenges by developing agentic AI solutions that can diagnose, reason, and take actions in production environments to improve productivity and address issues related to production support.
What You’ll Do:
- Build agentic AI systems: Design and implement tool-calling agents that combine retrieval, structured reasoning, and secure action execution (function calling, change orchestration, policy enforcement) following MCP protocol. Engineer robust guardrails for safety, compliance, and least-privilege access.
- Productionize LLMs: Build evaluation framework for open-source and foundational LLMs; implement retrieval pipelines, prompt synthesis, response validation, and self-correction loops tailored to production operations.
- Integrate with runtime ecosystems: Connect agents to observability, incident management, and deployment systems to enable automated diagnostics, runbook execution, remediation, and post-incident summarization with full traceability.
- Collaborate directly with users: Partner with production engineers, and application teams to translate production pain points into agentic AI roadmaps; define objective functions linked to reliability, risk reduction, and cost; and deliver auditable, business-aligned outcomes.
- Safety, reliability, and governance: Build validator models, adversarial prompts, and policy checks into the stack; enforce deterministic fallbacks, circuit breakers, and rollback strategies; instrument continuous evaluations for usefulness, correctness, and risk.
- Scale and performance: Optimize cost and latency via prompt engineering, context management, caching, model routing, and distillation; leverage batching, streaming, and parallel tool-calls to meet stringent SLOs under real-world load.
- Build a RAG pipeline: Curate domain-knowledge; build data-quality validation framework; establish feedback loops and milestone framework maintain knowledge freshness.
- Raise the bar: Drive design reviews, experiment rigor, and high-quality engineering practices; mentor peers on agent architectures, evaluation methodologies, and safe deployment patterns.
Qualifications
A Bachelor’s degree (Masters/ PhD preferred) in a computational field (Computer Science, Applied Mathematics, Engineering, or in a related quantitative discipline), with 7+ years of experience as an applied data scientist / machine learning engineer.
Essential Skills
- 7+ years of software development in one or more languages (Python, C/C++, Go, Java); strong hands-on experience building and maintaining large-scale Python applications preferred.
- 3+ years designing, architecting, testing, and launching production ML systems, including model deployment/serving, evaluation and monitoring, data processing pipelines, and model fine-tuning workflows.
- Practical experience with Large Language Models (LLMs): API integration, prompt engineering, finetuning/adaptation, and building applications using RAG and tool-using agents (vector retrieval, function calling, secure tool execution).
- Understanding of different LLMs, both commercial and open source, and their capabilities (e.g., OpenAI, Gemini, Llama, Qwen, Claude).
- Solid grasp of applied statistics, core ML concepts, algorithms, and data structures to deliver efficient and reliable solutions.
- Strong analytical problem-solving, ownership, and urgency; ability to communicate complex ideas simply and collaborate effectively across global teams with a focus on measurable business impact.
- Preferred: Proficiency building and operating on cloud infrastructure (ideally AWS), including containerized services (ECS/EKS), serverless (Lambda), data services (S3, DynamoDB, Redshift), orchestration (Step Functions), model serving (SageMaker), and infra-as-code (Terraform/CloudFormation).
YOUR CAREER
Goldman Sachs is a meritocracy where you will be given all the tools to advance your career. At Goldman Sachs, you will have access to excellent training programs designed to improve multiple facets of your skill portfolio. Our in-house training program, “Goldman Sachs University” offers a comprehensive series of courses that you will have access to as your career progresses. Goldman Sachs University has an impressive catalogue of courses which span technical, business and leadership skills.
Salary Range
The expected base salary for this New York, New York, United States-based position is $150000-$250000. In addition, you may be eligible for a discretionary bonus if you are an active employee as of fiscal year-end.
Benefits
Goldman Sachs is committed to providing our people with valuable and competitive benefits and wellness offerings, as it is a core part of providing a strong overall employee experience. A summary of these offerings, which are generally available to active, non-temporary, full-time and part-time US employees who work at least 20 hours per week, can be found here.

Job Description
WHO WE ARE
Goldman Sachs is a leading global investment banking, securities and investment management firm that provides a wide range of services worldwide to a substantial and diversified client base that includes corporations, financial institutions, governments and high net-worth individuals. Founded in 1869, it is one of the oldest and largest investment banking firms. The firm is headquartered in New York and maintains offices in London, Bangalore, Frankfurt, Tokyo, Hong Kong and other major financial centers around the world. We are committed to growing our distinctive Culture and holding to our core values which always place our client's interests first. These values are reflected in our Business Principles, which emphasize integrity, commitment to excellence, innovation and teamwork.
Business Unit Overview
Enterprise Technology Operations (ETO) is a Business Unit within Core Engineering focused on running scalable production management services with a mandate of operational excellence and operational risk reduction achieved through large scale automation, best-in-class engineering, and application of data science and machine learning. The Production Runtime Experience (PRX) team in ETO applies software engineering and machine learning to production management services, processes, and activities to streamline monitoring, alerting, automation, and workflows.
TEAM OVERVIEW
The Machine Learning and Artificial Intelligence team in PRX applies advanced ML and GenAI to reduce the risk and cost of operating the firm’s large-scale compute infrastructure and extensive application estate. Building on strengths in statistical modelling, anomaly detection, predictive modelling, and time-series forecasting, we leverage foundational LLM Models to orchestrate multi-agent systems for automated production management services. By unifying classical ML with agentic AI, we deliver reliable, explainable, and cost-efficient operations at scale.
ROLE AND RESPONSIBILITIES
In this role, you will be responsible for launching and implementing GenAI agentic solutions aimed at reducing the risk and cost of managing large-scale production environments with varying complexities. You will address various production runtime challenges by developing agentic AI solutions that can diagnose, reason, and take actions in production environments to improve productivity and address issues related to production support.
What You’ll Do:
- Build agentic AI systems: Design and implement tool-calling agents that combine retrieval, structured reasoning, and secure action execution (function calling, change orchestration, policy enforcement) following MCP protocol. Engineer robust guardrails for safety, compliance, and least-privilege access.
- Productionize LLMs: Build evaluation framework for open-source and foundational LLMs; implement retrieval pipelines, prompt synthesis, response validation, and self-correction loops tailored to production operations.
- Integrate with runtime ecosystems: Connect agents to observability, incident management, and deployment systems to enable automated diagnostics, runbook execution, remediation, and post-incident summarization with full traceability.
- Collaborate directly with users: Partner with production engineers, and application teams to translate production pain points into agentic AI roadmaps; define objective functions linked to reliability, risk reduction, and cost; and deliver auditable, business-aligned outcomes.
- Safety, reliability, and governance: Build validator models, adversarial prompts, and policy checks into the stack; enforce deterministic fallbacks, circuit breakers, and rollback strategies; instrument continuous evaluations for usefulness, correctness, and risk.
- Scale and performance: Optimize cost and latency via prompt engineering, context management, caching, model routing, and distillation; leverage batching, streaming, and parallel tool-calls to meet stringent SLOs under real-world load.
- Build a RAG pipeline: Curate domain-knowledge; build data-quality validation framework; establish feedback loops and milestone framework maintain knowledge freshness.
- Raise the bar: Drive design reviews, experiment rigor, and high-quality engineering practices; mentor peers on agent architectures, evaluation methodologies, and safe deployment patterns.
Qualifications
A Bachelor’s degree (Masters/ PhD preferred) in a computational field (Computer Science, Applied Mathematics, Engineering, or in a related quantitative discipline), with 7+ years of experience as an applied data scientist / machine learning engineer.
Essential Skills
- 7+ years of software development in one or more languages (Python, C/C++, Go, Java); strong hands-on experience building and maintaining large-scale Python applications preferred.
- 3+ years designing, architecting, testing, and launching production ML systems, including model deployment/serving, evaluation and monitoring, data processing pipelines, and model fine-tuning workflows.
- Practical experience with Large Language Models (LLMs): API integration, prompt engineering, finetuning/adaptation, and building applications using RAG and tool-using agents (vector retrieval, function calling, secure tool execution).
- Understanding of different LLMs, both commercial and open source, and their capabilities (e.g., OpenAI, Gemini, Llama, Qwen, Claude).
- Solid grasp of applied statistics, core ML concepts, algorithms, and data structures to deliver efficient and reliable solutions.
- Strong analytical problem-solving, ownership, and urgency; ability to communicate complex ideas simply and collaborate effectively across global teams with a focus on measurable business impact.
- Preferred: Proficiency building and operating on cloud infrastructure (ideally AWS), including containerized services (ECS/EKS), serverless (Lambda), data services (S3, DynamoDB, Redshift), orchestration (Step Functions), model serving (SageMaker), and infra-as-code (Terraform/CloudFormation).
YOUR CAREER
Goldman Sachs is a meritocracy where you will be given all the tools to advance your career. At Goldman Sachs, you will have access to excellent training programs designed to improve multiple facets of your skill portfolio. Our in-house training program, “Goldman Sachs University” offers a comprehensive series of courses that you will have access to as your career progresses. Goldman Sachs University has an impressive catalogue of courses which span technical, business and leadership skills.
Salary Range
The expected base salary for this New York, New York, United States-based position is $150000-$250000. In addition, you may be eligible for a discretionary bonus if you are an active employee as of fiscal year-end.
Benefits
Goldman Sachs is committed to providing our people with valuable and competitive benefits and wellness offerings, as it is a core part of providing a strong overall employee experience. A summary of these offerings, which are generally available to active, non-temporary, full-time and part-time US employees who work at least 20 hours per week, can be found here.
AI ML Engineering Job Roles in California
See all 2,272+ AI ML Engineering Jobs in California
Sign up for free to filter by visa type, set job alerts, and find employers with verified sponsorship history.
Search AI ML Engineering Jobs in CaliforniaAI ML Engineering Jobs in California: Frequently Asked Questions
Which companies sponsor visas for AI ML engineers in California?
The most active sponsors for AI ML engineering roles in California include Google, Meta, Apple, Microsoft, Amazon, and Nvidia, all of which have major California operations. Beyond big tech, well-funded AI-focused startups such as Anthropic, OpenAI, and Scale AI have also filed significant numbers of H-1B petitions for ML engineering positions. Enterprise companies in fintech, healthcare tech, and autonomous vehicles are additional sources of sponsorship.
Which visa types are most common for AI ML engineering roles in California?
The H-1B is by far the most common visa for AI ML engineers in California, given that these roles typically require a bachelor's degree or higher in computer science, mathematics, or a closely related field, satisfying the specialty occupation standard. Candidates with outstanding research records may qualify for the O-1A. Australians in the field often pursue the E-3 as an H-1B alternative. Some PhD-level researchers explore EB-1 or EB-2 NIW pathways for permanent residence.
Which cities in California have the most AI ML engineering sponsorship jobs?
The San Francisco Bay Area accounts for the largest share, spanning San Francisco, Mountain View, Sunnyvale, Palo Alto, and Menlo Park, where the density of AI-focused employers is unmatched. Los Angeles has grown significantly, particularly around Santa Monica and Culver City, driven by media-tech and autonomous vehicle companies. San Diego is a smaller but active market, with strengths in defense AI and biotech-adjacent machine learning roles.
How to find ai ml engineering visa sponsorship jobs in California?
Migrate Mate filters AI ML engineering jobs specifically by visa sponsorship status, so you're only seeing roles where the employer has indicated willingness to sponsor. For California, you can narrow results by city or region to focus on Bay Area, Los Angeles, or San Diego opportunities. Because the sponsorship filter removes the guesswork of cold-applying to companies that won't sponsor, it significantly shortens the job search timeline for international candidates in this field.
Are there state-specific considerations for AI ML engineers seeking sponsorship in California?
California's prevailing wage requirements under the H-1B program are among the highest in the country, reflecting the state's cost of living and competitive tech market. Employers must certify on the Labor Condition Application that the offered wage meets the prevailing wage for the specific role and location. California also has a dense university pipeline through UC Berkeley, Stanford, UCLA, and UCSD, which means many employers have established international hiring processes and are familiar with OPT-to-H-1B transitions.
What is the prevailing wage for sponsored ai ml engineering jobs in California?
U.S. employers sponsoring a visa must pay at least the prevailing wage, which is what workers in the same role, area, and experience level typically earn. The Department of Labor sets this rate to make sure companies aren't hiring foreign workers simply because they'd accept lower pay than a U.S. worker. It varies by job title, location, and experience. You can look up current prevailing wage rates for any occupation and location using the OFLC Wage Search page.
See which ai ml engineering employers are hiring and sponsoring visas in California right now.
Search AI ML Engineering Jobs in California