Amazon Web Services OPT Eligible Jobs USA
Amazon Web Services hires OPT students across software engineering, cloud infrastructure, data science, and machine learning roles. AWS is a well-established employer for F-1 students on OPT, with a consistent track record of supporting the OPT-to-H-1B visa transition for high-performing hires.
See All Amazon Web Services JobsOverview
Showing 5 of 226+ Amazon Web Services OPT Eligible Jobs USA


Have you applied for this role?


Have you applied for this role?


Have you applied for this role?


Have you applied for this role?


Have you applied for this role?
See all 226+ Amazon Web Services OPT Eligible Jobs USA
Sign up for free to unlock all listings, filter by visa type, and get alerts for new Amazon Web Services OPT Eligible Jobs USA.
Get Access To All Jobs
DESCRIPTION
The Annapurna Labs team at Amazon Web Services (AWS) builds AWS Neuron, the software development kit used to accelerate deep learning and GenAI workloads on Amazon’s custom machine learning accelerators, Inferentia and Trainium.
The AWS Neuron SDK, developed by the Annapurna Labs team at AWS, is the backbone for accelerating deep learning and GenAI workloads on Amazon's Inferentia and Trainium ML accelerators. This comprehensive toolkit includes an ML compiler, runtime, and application framework that seamlessly integrates with popular ML frameworks like PyTorch and JAX enabling unparalleled ML inference and training performance.
The Inference Enablement and Acceleration team is at the forefront of running a wide range of models and supporting novel architecture alongside maximizing their performance for AWS's custom ML accelerators. Working across the stack from PyTorch till the hardware-software boundary, our engineers build systematic infrastructure, innovate new methods and create high-performance kernels for ML functions, ensuring every compute unit is fine tuned for optimal performance for our customers' demanding workloads. We combine deep hardware knowledge with ML expertise to push the boundaries of what's possible in AI acceleration.
As part of the broader Neuron organization, our team works across multiple technology layers - from frameworks and kernels and collaborate with compiler to runtime and collectives. We not only optimize current performance but also contribute to future architecture designs, working closely with customers to enable their models and ensure optimal performance. This role offers a unique opportunity to work at the intersection of machine learning, high-performance computing, and distributed architectures, where you'll help shape the future of AI acceleration technology.
You will architect and implement business critical features, and mentor a brilliant team of experienced engineers. We operate in spaces that are very large, yet our teams remain small and agile. There is no blueprint. We're inventing. We're experimenting. It is a very unique learning culture. The team works closely with customers on their model enablement, providing direct support and optimization expertise to ensure their machine learning workloads achieve optimal performance on AWS ML accelerators. The team collaborates with open source ecosystems to provide seamless integration and bring peak performance at scale for customers and developers.
This role is responsible for development, enablement and performance tuning of a wide variety of LLM model families, including massive scale large language models like the Llama family, DeepSeek and beyond. The Inference Enablement and Acceleration team works side by side with compiler engineers and runtime engineers to create, build and tune distributed inference solutions with Trainium and Inferentia. Experience optimizing inference performance for both latency and throughput on such large models across the stack from system level optimizations through to Pytorch or JAX is a must have.
Key job responsibilities
This role will help lead the efforts in building distributed inference support for Pytorch in the Neuron SDK. This role will tune these models to ensure highest performance and maximize the efficiency of them running on the customer AWS Trainium and Inferentia silicon and servers. Strong software development using Python, System level programming and ML knowledge are both critical to this role. Our engineers collaborate across compiler, runtime, framework, and hardware teams to optimize machine learning workloads for our global customer base. Working at the intersection of software, hardware, and machine learning systems, you'll bring expertise in low-level optimization, system architecture, and ML model acceleration. In this role, you will:
- Design, develop, and optimize machine learning models and frameworks for deployment on custom ML hardware accelerators.
- Participate in all stages of the ML system development lifecycle including distributed computing based architecture design, implementation, performance profiling, hardware-specific optimizations, testing and production deployment.
- Build infrastructure to systematically analyze and onboard multiple models with diverse architecture.
- Design and implement high-performance kernels and features for ML operations, leveraging the Neuron architecture and programming models.
- Analyze and optimize system-level performance across multiple generations of Neuron hardware.
- Conduct detailed performance analysis using profiling tools to identify and resolve bottlenecks.
- Implement optimizations such as fusion, sharding, tiling, and scheduling.
- Conduct comprehensive testing, including unit and end-to-end model testing with continuous deployment and releases through pipelines.
- Work directly with customers to enable and optimize their ML models on AWS accelerators.
- Collaborate across teams to develop innovative optimization techniques.
A day in the life
You will collaborate with a cross-functional team of applied scientists, system engineers, and product managers to deliver state-of-the-art inference capabilities for Generative AI applications. Your work will involve debugging performance issues, optimizing memory usage, and shaping the future of Neuron's inference stack across Amazon and the Open Source Community. As you design and code solutions to help our team drive efficiencies in software architecture, you’ll create metrics, implement automation and other improvements, and resolve the root cause of software defects.
You will also build high-impact solutions to deliver to our large customer base and participate in design discussions, code review, and communicate with internal and external stakeholders. You will work cross-functionally to help drive business decisions with your technical input. You will work in a startup-like development environment, where you’re always working on the most important initiative.
About the team
The Inference Enablement and Acceleration team fosters a builder’s culture where experimentation is encouraged, and impact is measurable. We emphasize collaboration, technical ownership, and continuous learning. Our team is dedicated to supporting new members. We have a broad mix of experience levels and tenures, and we’re building an environment that celebrates knowledge-sharing and mentorship. Our senior members enjoy one-on-one mentoring and thorough, but kind, code reviews. We care about your career growth and strive to assign projects that help our team members develop your engineering expertise so you feel empowered to take on more complex tasks in the future. Join us to solve some of the most interesting and impactful infrastructure challenges in AI/ML today.
BASIC QUALIFICATIONS
- 3+ years of non-internship professional software development experience
- Bachelor's degree in computer science or equivalent
- 3+ years of non-internship design or architecture (design patterns, reliability and scaling) of new and existing systems experience
- Fundamentals of Machine learning and LLMs, their architecture, training and inference lifecycles along with work experience on optimizations for improving the model execution.
- Software development experience in C++, Python (experience in at least one language is required).
- Strong understanding of system performance, memory management, and parallel computing principles.
- Proficiency in debugging, profiling, and implementing best software engineering practices in large-scale systems.
PREFERRED QUALIFICATIONS
- Familiarity with PyTorch, JIT compilation, and AOT tracing.
- Familiarity with CUDA kernels or equivalent ML or low-level kernels.
- Candidates with performant kernel development such as CUTLASS, FlashInfer etc., would be well suited.
- Familiar with syntax and tile-level semantics similar to Triton.
- Experience with online/offline inference serving with vLLM, SGLang, TensorRT or similar platforms in production environments.
- Deep understanding of computer architecture, operation systems level software and working knowledge of parallel computing.
Amazon is an equal opportunity employer and does not discriminate on the basis of protected veteran status, disability, or other legally protected status.
Our inclusive culture empowers Amazonians to deliver the best results for our customers. If you have a disability and need a workplace accommodation or adjustment during the application and hiring process, including support for the interview or onboarding process, please visit https://amazon.jobs/content/en/how-we-hire/accommodations for more information. If the country/region you’re applying in isn’t listed, please contact your Recruiting Partner.
The base salary range for this position is listed below. Your Amazon package will include sign-on payments and restricted stock units (RSUs). Final compensation will be determined based on factors including experience, qualifications, and location. Amazon also offers comprehensive benefits including health insurance (medical, dental, vision, prescription, Basic Life & AD&D insurance and option for Supplemental life plans, EAP, Mental Health Support, Medical Advice Line, Flexible Spending Accounts, Adoption and Surrogacy Reimbursement coverage), 401(k) matching, paid time off, and parental leave. Learn more about our benefits at https://amazon.jobs/en/benefits.
Location: USA, WA, Seattle - 143,700.00 - 194,400.00 USD annually
Job Roles at Amazon Web Services
See all 226+ Amazon Web Services Jobs
Sign up for free to unlock all listings, filter by visa type, and get alerts for new Amazon Web Services roles.
Get Access To All JobsTips for Finding Amazon Web Services OPT Eligible Jobs USA
Target AWS teams with OPT-friendly pipelines
AWS's cloud platform, AI/ML, and developer tools orgs hire OPT students most consistently. Focus your applications on these teams rather than applying broadly across all AWS divisions, which vary significantly in their willingness to hire OPT candidates.
Verify your SOC code matches your AWS role
Your OPT EAD is tied to your degree field, not your job title. Before accepting an offer, confirm your AWS role's SOC code aligns with your major using O*NET, since a mismatch could jeopardize your STEM OPT extension eligibility later.
Confirm AWS's E-Verify enrollment before your extension deadline
STEM OPT requires your employer to be enrolled in E-Verify. AWS is enrolled, but confirm this through E-Verify's employer search before your 24-month extension application, since USCIS denies STEM OPT extensions when employer enrollment lapses.
Search AWS's verified H-1B filings on Migrate Mate
To assess your long-term sponsorship odds, search AWS's Labor Condition Application filing history by job title on Migrate Mate. This shows which roles AWS has actively sponsored for H-1B, so you can prioritize positions with a documented transition track record.
Time your AWS internship to pre-position for H-1B
AWS internships converting to full-time roles before April 1 give your employer time to file an H-1B petition in the same cap season. Starting full-time after April significantly delays your first cap-subject H-1B eligibility by a full year.
Negotiate your AWS offer letter to name your position title precisely
Your offer letter's job title affects how AWS classifies your role in the LCA and future H-1B petition. Vague titles like 'engineer' create RFE risk. Push for a specific title, such as Software Development Engineer or Applied Scientist, that maps cleanly to a specialty occupation.
Amazon Web Services OPT Eligibility: Frequently Asked Questions
Does Amazon Web Services sponsor OPT visas?
AWS does not sponsor OPT itself. OPT is work authorization USCIS grants directly to F-1 students. What AWS does is hire students who already hold an OPT EAD. For STEM OPT extensions, AWS must be enrolled in E-Verify, which it is, allowing eligible employees to extend their work authorization by 24 months.
Which AWS roles and departments typically hire OPT students?
AWS most consistently hires OPT students in software engineering, cloud infrastructure, applied science, and data engineering. Teams building core AWS services and AI/ML tooling have historically been the most active. Business, operations, and non-technical roles are less predictable for OPT candidates due to specialty occupation requirements tied to your degree field.
How do I apply to AWS jobs as an OPT student?
Apply through AWS's careers portal and disclose your OPT status when asked about work authorization. Selecting 'yes' to sponsorship requirements triggers recruiter review, so be precise: you currently have OPT authorization but will need H-1B sponsorship in the future. Use Migrate Mate to identify AWS roles with strong H-1B filing histories before applying.
How do I know if my AWS role qualifies for a STEM OPT extension?
Your role must fall within a STEM-designated field as defined by your degree's CIP code and map to an eligible SOC occupation. Verify your AWS job title's SOC code against the STEM OPT designated degree list published by USCIS. If the SOC code doesn't match your CIP code, the extension will be denied regardless of your employer's status.
What is the typical timeline for AWS to transition OPT employees to H-1B?
AWS typically files H-1B petitions for OPT employees during the April cap season, covering the October 1 employment start. If you join AWS mid-year, your H-1B filing may be delayed until the following April. The H-1B lottery selection is random, so AWS cannot guarantee transition timing even for high-performing employees.
How does Amazon Web Services hire OPT students?
OPT is work authorization granted directly to F-1 students after graduation — no employer petition is required. Amazon Web Services can hire OPT students as soon as their EAD card is approved. STEM degree holders can extend OPT by 24 months when their employer is enrolled in E-Verify. Most companies that hire OPT students also support the transition to H-1B when the student's OPT period is ending. Check Amazon Web Services's individual postings on Migrate Mate to confirm OPT acceptance per role.