AI Architect Jobs at NVIDIA with Visa Sponsorship
NVIDIA's AI Architect roles sit at the intersection of large-scale GPU infrastructure, deep learning systems design, and enterprise deployment. NVIDIA has a consistent track record of sponsoring work visas across multiple categories for technical architecture roles, making it a realistic target for international candidates with the right background.
See All AI Architect at NVIDIA JobsOverview
Showing 5 of 30+ AI Architect Jobs at NVIDIA jobs


Have you applied for this role?


Have you applied for this role?


Have you applied for this role?


Have you applied for this role?


Have you applied for this role?
See all 30+ AI Architect Jobs at NVIDIA
Sign up for free to unlock all listings, filter by visa type, and get alerts for new AI Architect Jobs at NVIDIA.
Get Access To All Jobs
INTRODUCTION
An applied research team within NVIDIA’s Networking Systems & Software Architecture group is solving some of AI’s hardest infrastructure problems. The team builds systems-level software that moves data between GPUs, nodes, and storage at the speed modern AI demands—spanning low-level transport optimization, hardware-software co-design, and communication frameworks that plug directly into production AI stacks. The team's charter expands into emerging domains including quantum computing interconnects.
This Principal Architect role leads the research agenda and architectural direction for how NVIDIA’s AI systems communicate at scale—across GPUs, DPUs, NICs, and heterogeneous storage. It requires someone who defines project scope from scratch, publishes original work, and translates research breakthroughs into production-grade software that ships industry-wide!
What you will be doing:
- Setting the long-term technical vision for distributed AI communication systems—GPU-to-GPU, GPU-to-storage, and cross-node data movement.
- Conducting original research and prototyping next-generation networking solutions over RDMA, NVLink, and GPUDirect.
- Driving hardware-software co-optimization with GPU, DPU, NIC, and network switch. Investigating fundamental bottlenecks in communication runtimes for large-scale AI workloads (KV cache transfer, disaggregated prefill/decode, model parallelism).
- Integrating networking capabilities into AI serving stacks such as vLLM, SGLang, and TensorRT-LLM.
- Publishing findings, representing NVIDIA in industry forums and standards bodies, and mentoring senior engineers across the organization.
What we need to see:
- 15+ years in systems software and/or networking with deep expertise in high-performance networking (InfiniBand, RoCE, RDMA, NVLink), communication libraries (e.g. NIXL, NCCL, UCX, MPI, NVSHMEM), and GPU accelerated systems, with track record of defining and delivering complex, cross-team technical initiatives from research concept to production.
- MS, PhD or equivalent experience in Computer Science, Computer Engineering, Electrical Engineering, or a related field.
- Deep understanding of computer architecture, memory hierarchies, DMA engines, and OS-level networking.
- Understanding of ML systems concepts—transformer architectures, KV cache mechanics, model parallelism, or distributed training and inference patterns.
- Proficiency in programming languages such as C, C++, Rust and Python.
Ways to stand out from the crowd:
- Knowledge of ML inference frameworks (vLLM, SGLang, TensorRT-LLM) and their communication requirements.
- CUDA programming and NVIDIA GPU architecture expertise.
- Proved experience influencing product strategy and technical roadmap at a senior level.
- Major open-source contributions.
With competitive salaries and a comprehensive benefits package, NVIDIA is widely regarded as one of the most desirable technology employers in the world. Our teams are composed of some of the most forward-thinking and driven engineers in the industry, and we continue to grow rapidly. If you are a senior data engineer passionate about building large-scale, high-impact data platforms, we’d love to hear from you.
Your base salary will be determined based on your location, experience, and the pay of employees in similar positions. The base salary range is 272,000 USD - 431,250 USD.
You will also be eligible for equity and benefits.
Applications for this job will be accepted at least until April 27, 2026.
This posting is for an existing vacancy.
NVIDIA uses AI tools in its recruiting processes.
NVIDIA is committed to fostering a diverse work environment and proud to be an equal opportunity employer. As we highly value diversity in our current and future employees, we do not discriminate (including in our hiring and promotion practices) on the basis of race, religion, color, national origin, gender, gender expression, sexual orientation, age, marital status, veteran status, disability status or any other characteristic protected by law.

INTRODUCTION
An applied research team within NVIDIA’s Networking Systems & Software Architecture group is solving some of AI’s hardest infrastructure problems. The team builds systems-level software that moves data between GPUs, nodes, and storage at the speed modern AI demands—spanning low-level transport optimization, hardware-software co-design, and communication frameworks that plug directly into production AI stacks. The team's charter expands into emerging domains including quantum computing interconnects.
This Principal Architect role leads the research agenda and architectural direction for how NVIDIA’s AI systems communicate at scale—across GPUs, DPUs, NICs, and heterogeneous storage. It requires someone who defines project scope from scratch, publishes original work, and translates research breakthroughs into production-grade software that ships industry-wide!
What you will be doing:
- Setting the long-term technical vision for distributed AI communication systems—GPU-to-GPU, GPU-to-storage, and cross-node data movement.
- Conducting original research and prototyping next-generation networking solutions over RDMA, NVLink, and GPUDirect.
- Driving hardware-software co-optimization with GPU, DPU, NIC, and network switch. Investigating fundamental bottlenecks in communication runtimes for large-scale AI workloads (KV cache transfer, disaggregated prefill/decode, model parallelism).
- Integrating networking capabilities into AI serving stacks such as vLLM, SGLang, and TensorRT-LLM.
- Publishing findings, representing NVIDIA in industry forums and standards bodies, and mentoring senior engineers across the organization.
What we need to see:
- 15+ years in systems software and/or networking with deep expertise in high-performance networking (InfiniBand, RoCE, RDMA, NVLink), communication libraries (e.g. NIXL, NCCL, UCX, MPI, NVSHMEM), and GPU accelerated systems, with track record of defining and delivering complex, cross-team technical initiatives from research concept to production.
- MS, PhD or equivalent experience in Computer Science, Computer Engineering, Electrical Engineering, or a related field.
- Deep understanding of computer architecture, memory hierarchies, DMA engines, and OS-level networking.
- Understanding of ML systems concepts—transformer architectures, KV cache mechanics, model parallelism, or distributed training and inference patterns.
- Proficiency in programming languages such as C, C++, Rust and Python.
Ways to stand out from the crowd:
- Knowledge of ML inference frameworks (vLLM, SGLang, TensorRT-LLM) and their communication requirements.
- CUDA programming and NVIDIA GPU architecture expertise.
- Proved experience influencing product strategy and technical roadmap at a senior level.
- Major open-source contributions.
With competitive salaries and a comprehensive benefits package, NVIDIA is widely regarded as one of the most desirable technology employers in the world. Our teams are composed of some of the most forward-thinking and driven engineers in the industry, and we continue to grow rapidly. If you are a senior data engineer passionate about building large-scale, high-impact data platforms, we’d love to hear from you.
Your base salary will be determined based on your location, experience, and the pay of employees in similar positions. The base salary range is 272,000 USD - 431,250 USD.
You will also be eligible for equity and benefits.
Applications for this job will be accepted at least until April 27, 2026.
This posting is for an existing vacancy.
NVIDIA uses AI tools in its recruiting processes.
NVIDIA is committed to fostering a diverse work environment and proud to be an equal opportunity employer. As we highly value diversity in our current and future employees, we do not discriminate (including in our hiring and promotion practices) on the basis of race, religion, color, national origin, gender, gender expression, sexual orientation, age, marital status, veteran status, disability status or any other characteristic protected by law.
See all 30+ AI Architect at NVIDIA jobs
Sign up for free to unlock all listings, filter by visa type, and get alerts for new AI Architect at NVIDIA roles.
Get Access To All JobsTips for Finding AI Architect Jobs at NVIDIA Jobs
Align your portfolio to NVIDIA's deployment scale
NVIDIA hires AI Architects who have designed systems at GPU-cluster scale, not just trained models. Document architectures you've built involving distributed inference, CUDA optimization, or multi-node training pipelines before you apply.
Target NVIDIA's enterprise and cloud divisions specifically
NVIDIA's AI Architect openings are concentrated in teams supporting DGX systems, NVIDIA AI Enterprise, and cloud service partnerships. Roles tied to those product lines move through hiring more consistently than exploratory research positions.
Confirm your degree maps to the specialty occupation standard
USCIS requires a direct relationship between your degree field and the role. For AI Architect positions, a degree in computer science, electrical engineering, or a closely related discipline is the safest foundation. A general IT or business degree may require additional documentation.
Negotiate the start date around H-1B cap timing
If you're not already in H-1B status, an October 1 start date is required for cap-subject petitions. Discuss this constraint with your NVIDIA recruiter early so the offer letter and onboarding plan reflect the correct timeline from the start.
Browse AI Architect openings at NVIDIA through Migrate Mate
Filter verified visa-sponsoring openings by role and employer on Migrate Mate to find active AI Architect positions at NVIDIA before applying directly through NVIDIA's career portal.
Prepare for NVIDIA's technical interview structure before discussing visa status
NVIDIA's AI Architect interviews typically include system design rounds focused on inference optimization and deployment architecture. Clear the technical bar first. Visa logistics are handled by a dedicated immigration team after an offer is extended.
AI Architect at NVIDIA jobs are hiring across the US. Find yours.
Find AI Architect at NVIDIA JobsFrequently Asked Questions
Does NVIDIA sponsor H-1B visas for AI Architects?
Yes, NVIDIA sponsors H-1B visas for AI Architect roles. NVIDIA works with immigration counsel to file H-1B petitions for qualifying candidates, including both cap-subject registrations and cap-exempt transfers for candidates already holding H-1B status with another employer. If you're in the H-1B lottery pool, NVIDIA's size and resourcing means the process is well-structured once an offer is made.
How do I apply for AI Architect jobs at NVIDIA?
Apply directly through NVIDIA's careers portal, targeting roles in AI infrastructure, enterprise AI, or cloud solutions architecture. Tailor your resume to reflect GPU-scale system design and deployment experience. You can also browse verified AI Architect openings at NVIDIA on Migrate Mate, which surfaces roles from employers with confirmed sponsorship track records so you're not applying blind on sponsorship eligibility.
Which visa types does NVIDIA commonly use for AI Architect roles?
NVIDIA sponsors H-1B visas for most international AI Architect hires. Australian citizens can pursue the E-3 visa, which bypasses the H-1B lottery entirely and allows year-round applications. For candidates seeking permanent residence, NVIDIA supports EB-2 and EB-3 Green Card sponsorship, typically beginning the PERM labor certification process after an employee has established tenure in the role.
What qualifications does NVIDIA expect for AI Architect roles?
NVIDIA generally expects a bachelor's or master's degree in computer science, electrical engineering, or a related field, combined with hands-on experience designing large-scale AI systems. Practical depth in CUDA, TensorRT, distributed training frameworks like PyTorch or JAX, and deployment on NVIDIA hardware platforms carries significant weight. Candidates who have architected inference pipelines at production scale stand out in NVIDIA's hiring process.
How do I plan the timeline for an AI Architect role at NVIDIA if I need visa sponsorship?
If you need a new H-1B, the cap registration window runs in March for an October 1 start date, so factor roughly six to seven months between offer and day one. E-3 applicants typically move faster, with consular processing taking two to four weeks. For candidates transferring an existing H-1B, NVIDIA can file a portability petition and you can start shortly after filing, before USCIS adjudicates.
See which AI Architect at NVIDIA employers are hiring and sponsoring visas right now.
Search AI Architect at NVIDIA Jobs