Senior ASIC Timing Engineer
Westford, MA jobs
NVIDIA has continuously reinvented itself over two decades. Our invention of the GPU in 1999 sparked the growth of the PC gaming market, redefined modern computer graphics, and revolutionized parallel computing. More recently, GPU deep learning ignited modern AI - the next era of computing. NVIDIA is a “learning machine” that constantly evolves by adapting to new opportunities which are hard to solve, that only we can pursue, and that matter to the world. This is our life's work, to amplify human inventiveness and intelligence.
NVIDIA is looking for best-in-class Senior ASIC Timing Design Engineers to join our outstanding Networking Silicon engineering team, developing the industry's best high speed communication devices, delivering the highest throughput and lowest latency! Come and take a part in crafting our groundbreaking and innovating chips, enjoy working in a meaningful, growing and professional environment where you make a significant impact in a technology-focused company.
What you will be doing:
You will drive physical design and timing of high-frequency and low-power DPUs and SoCs at block level, cluster level, and/or full chip level.
Analyze and optimize design constraints and synthesis parameters to achieve performance, power, and area targets.
Help in driving frontend and backend implementation from RTL to gds2, including synthesis, equivalence checking, floor-planning, timing constraints, timing and power convergence, and ECO implementation.
What we need to see:
Great teammate
BS (or equivalent experience) in Electrical or Computer Engineering
8+ years experience or MS (or equivalent experience) with 2 years experience in Synthesis and Timing.
Understanding of DFT logic and hands-on experience in design closure.
Expertise in analyzing and converging crosstalk delay, noise glitch, and electrical/manufacturing rules in deep-sub micron processes.
Knowledge in process variation effect modeling and experience in design convergence taking into account process variations.
Experience in critical path planning and crafting needed.
Power user of Static Timing tools like Synopsys PrimeTime or Cadence Tempus.
Solid experience in full-chip/sub-chip Static Timing Analysis (STA), timing constraints generation and management, and timing convergence.
Proficiency in Python, Tcl and Make for automation and scripting tasks.
NVIDIA is widely considered to be the leader of AI computing, and one of the technology world's most desirable employers. We have some of the most forward-thinking and hardworking people in the world working for us. If you're creative and autonomous, we want to hear from you.
Your base salary will be determined based on your location, experience, and the pay of employees in similar positions. The base salary range is 168,000 USD - 264,500 USD for Level 4, and 196,000 USD - 310,500 USD for Level 5.
You will also be eligible for equity and benefits.
Applications for this job will be accepted at least until December 5, 2025.NVIDIA is committed to fostering a diverse work environment and proud to be an equal opportunity employer. As we highly value diversity in our current and future employees, we do not discriminate (including in our hiring and promotion practices) on the basis of race, religion, color, national origin, gender, gender expression, sexual orientation, age, marital status, veteran status, disability status or any other characteristic protected by law.
Auto-ApplySenior Deep Learning Engineer
Santa Clara, CA jobs
We are now looking for a Senior Deep Learning Engineer! At NVIDIA, we are at the forefront of advancing the capabilities of artificial intelligence. We are seeking an ambitious and forward-thinking senior deep learning engineer to contribute to the development of next-generation inference optimizations and deliver industry-leading performance without compromising model quality. In this role, you will analyze and explore techniques to scale test-time compute and optimize low-latency inference. Your work will leverage cross-stack optimizations at the algorithmic and system level.
As NVIDIA makes significant strides in AI datacenters, our team holds a central role in maximizing the efficiency of our exponentially growing inference deployment needs and establishing a data-driven approach to algorithmic improvements, hardware design and system software development. We collaborate extensively with diverse teams at NVIDIA, spanning deep learning research and framework development teams, to silicon architecture. Thriving in such a high-impact, interdisciplinary environment necessitates not only technical proficiency but also a growth mindset and a pragmatic attitude - qualities that fuel our collective success in shaping the future of datacenter technology.
What You'll Be Doing:
* Keeping abreast of the latest advancements in generative AI research.
* Prototyping and analyzing emergent techniques in the test-time compute space such as output refinement, speculation, and retrieval. Identifying opportunities for algorithmic as well as system optimizations.
* Pioneering the development of innovative optimizations to enable high quality inferencing on NVIDIA GPUs.
* Collaborating closely with production teams to incorporate the latest advancements into cutting-edge software frameworks.
What We Need to See:
* Master's degree (or equivalent experience) in Computer Science, Artificial Intelligence, Applied Mathematics, or related fields.
* A strong foundation in deep learning, with a particular emphasis on generative models and inferencing.
* A track record of at least 5 years of relevant software development experience in modern deep learning frameworks such as PyTorch.
* Growth mindset and pragmatic attitude.
Ways to Stand Out From the Crowd:
* Published research or noteworthy contributions to the field of deep learning, particularly in areas such as inference-time compute, conditional compute, speculative decoding, etc.
* Experience with prototyping and/or deployment of emergent test time compute techniques.
* Experience with collaborating across algorithms, software and performance teams to deliver high quality solutions.
* Familiarity with computer architecture and how it relates to AI algorithms development.
NVIDIA is widely considered to be one of the technology world's most desirable employers. We have some of the most forward-thinking and hardworking people on the planet working for us. If you're creative and autonomous, we want to hear from you!
Your base salary will be determined based on your location, experience, and the pay of employees in similar positions. The base salary range is 148,000 USD - 235,750 USD for Level 3, and 184,000 USD - 287,500 USD for Level 4.
You will also be eligible for equity and benefits.
Applications for this job will be accepted at least until October 14, 2025.
NVIDIA is committed to fostering a diverse work environment and proud to be an equal opportunity employer. As we highly value diversity in our current and future employees, we do not discriminate (including in our hiring and promotion practices) on the basis of race, religion, color, national origin, gender, gender expression, sexual orientation, age, marital status, veteran status, disability status or any other characteristic protected by law.
Auto-ApplySenior DFT Engineer
Santa Clara, CA jobs
NVIDIA has continuously reinvented itself over two decades. Our invention of the GPU in 1999 sparked the growth of the PC gaming market, redefined modern computer graphics, and revolutionized parallel computing. More recently, GPU deep learning ignited modern AI - the next era of computing. NVIDIA is a "learning machine" that constantly evolves by adapting to new opportunities that are hard to solve, that only we can tackle, and that matter to the world. This is our life's work , to amplify human imagination and intelligence. Make the choice to join us today.
Design-for-Test Engineering at NVIDIA works on groundbreaking innovations involving crafting creative solutions for DFT architecture, verification and post-silicon validation on some of the industry's most complex semiconductor chips.
What you'll be doing:
* As a member in our team, you will own and work with cross functional teams, implementing state-of-the-art designs in test access mechanisms, IO BIST, memory BIST and scan compression.
* In addition, you will help develop and deploy DFT methodologies for our next generation products.
* You will also help mentor junior engineers on test designs and trade-offs including cost and quality.
What we need to see:
* BSEE (or equivalent experience) with 5+, MSEE with 3+ years of experience or PhD in DFT or related domains
* Demonstrated knowledge and expertise in defining scan test plans, BIST including memories and IOs, fault modeling, ATPG and fault simulation
* Excellent analytical skills in verification and validation of test patterns and logic on complex and multi-million gate designs using vendor tools
* Good exposure to cross functional areas including RTL & clocks design, STA, place-n-route and power, to ensure we are making the right trade-offs
* Experience in Silicon debug and bring-up on the ATE with an understanding of pattern formats, failure processing, and test program development
* Strong programming and scripting skills in Perl, Python or Tcl desired
* Extraordinary written and oral communication skills with the curiosity to work on rare challenges
Our technology has no boundaries! NVIDIA is building the world's most groundbreaking and state of the art compute platforms for the world to use. It's because of our work that scientists, researchers and engineers can advance their ideas. At its core, our visual computing technology not only enables an outstanding computing experience, but it is also energy efficient! We pioneered a supercharged form of computing loved by the most demanding computer users in the world - scientists, designers, artists, and gamers.
NVIDIA offers highly competitive salaries and a comprehensive benefits package. We have some of the most brilliant and talented people in the world working for us and, due to unprecedented growth, our world-class engineering teams are growing fast. If you're a creative and autonomous engineer with real passion for technology, we want to hear from you!
#LI-Hybrid
Your base salary will be determined based on your location, experience, and the pay of employees in similar positions. The base salary range is 136,000 USD - 212,750 USD for Level 3, and 168,000 USD - 264,500 USD for Level 4.
You will also be eligible for equity and benefits.
Applications for this job will be accepted at least until December 13, 2025.
NVIDIA is committed to fostering a diverse work environment and proud to be an equal opportunity employer. As we highly value diversity in our current and future employees, we do not discriminate (including in our hiring and promotion practices) on the basis of race, religion, color, national origin, gender, gender expression, sexual orientation, age, marital status, veteran status, disability status or any other characteristic protected by law.
Auto-ApplySenior Compiler Engineer - DL
Santa Clara, CA jobs
NVIDIA has been transforming computer graphics, PC gaming, and accelerated computing for more than 25 years. It's a unique legacy of innovation that's fueled by great technology-and amazing people. Today, we're tapping into the unlimited potential of AI to define the next era of computing. An era in which our GPU acts as the brains of computers, robots, and self-driving cars that can understand the world. Doing what's never been done before takes vision, innovation, and the world's best talent. As an NVIDIAN, you'll be immersed in a diverse, supportive environment where everyone is inspired to do their best work. Come join the team and see how you can make a lasting impact on the world
We are looking for a Deep Learning Compiler Engineer. NVIDIA is hiring software engineers for its Deep Learning Compiler (DLC) team. Academic and commercial groups around the world are using GPUs to power a revolution in deep learning, enabling breakthroughs in many areas, e.g. large language models, generative AIs, recommendation systems, image classification, speech recognition, etc. Our DLC has been the backbone of NVIDIA inference engine, spanning across data centers, personal devices, automotive, and robotics. The compiler must deliver leading inference performance, fast build time, reduced memory footprints, and ease of use in the forms of both Ahead-of-Tine and Just-in-Time. Join the team building the DLC which will be used by the entire deep learning community.
What you'll be doing:
Analyzing deep learning networks and developing compiler optimization algorithms.
Collaborating with members of the deep learning software framework teams and the hardware architecture teams to accelerate the next generation of deep learning software.
Scope of these efforts includes defining public APIs, performance optimizations and analysis, crafting and implementing compiler infrastructure techniques for neural networks, and other general software engineering work.
What we need to see:
Bachelors, Masters or Ph.D. in Computer Science, Computer Engineering, related field or equivalent experience
3+ years of relevant work or research experience in performance analysis and compiler optimizations.
Ability to work independently, define project goals and scope, and lead your own development efforts.
Excellent C/C++ and Python programming and software design skills, including debugging, performance analysis, and test design.
Strong interpersonal skills are required along with the ability to work in a dynamic product-oriented team.
Ways to stand out from the crowd:
Proficient in CPU and/or GPU architecture. CUDA or OpenCL programming experience.
Experiences in systems with constrained resources, such as embedded platforms, small memory size, and cross compilation.
Experience with the following technologies: MLIR, XLA, TVM, LLVM, deep learning models and algorithms, and deep learning frameworks, such as PyTorch.
GPU kernel generation with high performance and fast build time.
A track record of success in mentoring junior engineers and interns is a bonus.
With highly competitive salaries and a comprehensive benefits package, NVIDIA is widely considered to be one of the technology industry's most desirable employers. We have some of the most brilliant and hardworking people in the world working with us and our product lines are growing fast in some of the hottest state of the art fields such as Virtual Reality, Artificial Intelligence, Deep Learning and Autonomous Vehicles.
Your base salary will be determined based on your location, experience, and the pay of employees in similar positions. The base salary range is 148,000 USD - 235,750 USD for Level 3, and 184,000 USD - 287,500 USD for Level 4.
You will also be eligible for equity and benefits.
Applications for this job will be accepted at least until November 8, 2025.NVIDIA is committed to fostering a diverse work environment and proud to be an equal opportunity employer. As we highly value diversity in our current and future employees, we do not discriminate (including in our hiring and promotion practices) on the basis of race, religion, color, national origin, gender, gender expression, sexual orientation, age, marital status, veteran status, disability status or any other characteristic protected by law.#deeplearning
Auto-ApplySenior CI/CD Engineer
Santa Clara, CA jobs
As a Senior System CI/CD Engineer, your responsibility will be crucial in maintaining and improving our outstanding software development infrastructure. Collaborate with a team of hardworking engineers on ambitious projects that push boundaries, revolutionizing applications and sophisticated system software. If you're a creative and autonomous engineer with a real passion for technology, we want to hear from you.
What you'll be doing:
Crafting, implementing, and maintaining robust build systems to streamline the software development lifecycle.
Developing and improving CI/CD pipelines to ensure flawless code integration, testing, and deployment.
Analyzing, prioritizing, and categorizing bugs to determine root causes and successfully implement fixes.
Defining and managing branching strategies to support collaborative development and release processes.
Collaborating with system-level developers to optimize software performance, compatibility, and maintainability.
What we need to see:
8+ years of experience triaging complex system software bugs while prioritizing them effectively.
B.S. or higher degree in Computer Science/Engineering and related field (or equivalent experience).
Strong experience with developing cloud backends and databases in AWS, Azure, GCP.
Strong expertise in build systems such as Gradle, Maven, Make, or Bazel.
Hands-on experience with CI/CD tools, including Jenkins, GitLab CI, CircleCI, or Azure DevOps.
In-depth knowledge of Linux and Windows build systems.
Proven experience with system software and kernel development, including debugging and optimization.
Outstanding problem-solving and analytical skills with a strong inclination for prioritizing and resolving bugs.
Familiarity with version control systems and branching strategies like Perforce or trunk-based development.
Proficiency in quality assurance practices, benchmarks, and automated testing frameworks.
Ways to stand out from the crowd:
Familiarity with build systems and debugging complex issues.
We are widely considered to be one of the technology world's most desirable employers, and as a result have some of the most forward-thinking and hardworking people in the world working for us. If you're passionate, creative, and driven, we'd love to have you join the team. With competitive salaries and a generous benefits package, we are widely considered to be one of the technology world's most desirable employers. We have some of the most forward-thinking and hardworking people in the world working for us and, due to unprecedented growth, our exclusive engineering teams are rapidly growing.
Your base salary will be determined based on your location, experience, and the pay of employees in similar positions. The base salary range is 184,000 USD - 287,500 USD for Level 4, and 224,000 USD - 356,500 USD for Level 5.
You will also be eligible for equity and benefits.
Applications for this job will be accepted at least until December 7, 2025.NVIDIA is committed to fostering a diverse work environment and proud to be an equal opportunity employer. As we highly value diversity in our current and future employees, we do not discriminate (including in our hiring and promotion practices) on the basis of race, religion, color, national origin, gender, gender expression, sexual orientation, age, marital status, veteran status, disability status or any other characteristic protected by law.
Auto-ApplySenior DFT Engineer
Santa Clara, CA jobs
NVIDIA has continuously reinvented itself over two decades. Our invention of the GPU in 1999 sparked the growth of the PC gaming market, redefined modern computer graphics, and revolutionized parallel computing. More recently, GPU deep learning ignited modern AI - the next era of computing. NVIDIA is a “learning machine” that constantly evolves by adapting to new opportunities that are hard to solve, that only we can tackle, and that matter to the world. This is our life's work, to amplify human creativity and intelligence. Make the choice to join us today.
NVIDIA's DFX team is looking for an exceptional DFT Engineer to help shape the future of compute. As stewards of the entire Scan Test Lifecycle, we drive innovation for the most advanced silicon in the world-spanning 2.5D/3D AI data center platforms, Gaming and Enterprise GPUs, and complex SOCs powering Autonomous Machines, Robotics, and Industrial systems. You will innovate at scale, designing and prototyping breakthrough Test Architectures for reticle sized, multi-chiplet products-from RTL to verification to post-silicon ATE bring-up. Join a globally recognized team that consistently delivers breakthrough performance across multiple high-impact tape-outs each year.
What you'll be doing:
Develop and deploy Industry-leading test methodologies on NVIDIA's next-generation silicon platforms.
Collaborate with leading EDA vendors to shape tool capabilities that meet NVIDIA's ambitious design goals, and partner with internal CAD teams to drive scalable, automated solutions.
Co-architect novel DFT strategies alongside VLSI and Product Engineering teams to push the boundaries of silicon test innovation.
Own the full ATPG lifecycle-verification, coverage analysis, pattern generation, and ATE bring-up-across NVIDIA's full product portfolio.
Guide and mentor junior engineers, helping them navigate complex design trade-offs to achieve world-class quality and efficiency.
What we need to see:
MS/PhD or equivalent experience in Electrical Engineering or a related field
5+ years of hands-on experience in Design-For-Test (DFT)
Deep knowledge of DFT tools, methodologies, and test strategies for complex, large-scale designs
Strong experience with industry standard ATPG tools
Clear, effective communicator-strong written and verbal skills
Passion for mentoring and scaling technical excellence in a team
Ways to stand out from the crowd:
Experience with 2.5D/3D ICs, multi-chiplet architectures, or reticle-sized designs
Background in developing or enhancing EDA tool flows
Experience with Silicon testing and Automatic Test Equipment (ATE)
Expertise in using programming languages and AI for automation
Personal success stories in leading org wide changes
NVIDIA offers highly competitive salaries and a comprehensive benefits package. At NVIDIA, we work on the hardest problems, and some of the most brilliant and motivated people in their fields choose to work here. Our world-class engineering teams are growing, as we lead the AI revolution from the front. If you are ready to make an impact in this journey, we want to hear from you!
#LI-Hybrid
Your base salary will be determined based on your location, experience, and the pay of employees in similar positions. The base salary range is 168,000 USD - 264,500 USD for Level 4, and 196,000 USD - 310,500 USD for Level 5.
You will also be eligible for equity and benefits.
Applications for this job will be accepted at least until November 14, 2025.NVIDIA is committed to fostering a diverse work environment and proud to be an equal opportunity employer. As we highly value diversity in our current and future employees, we do not discriminate (including in our hiring and promotion practices) on the basis of race, religion, color, national origin, gender, gender expression, sexual orientation, age, marital status, veteran status, disability status or any other characteristic protected by law.
Auto-ApplySenior Reset and Boot ASIC Engineer
Santa Clara, CA jobs
NVIDIA is looking for a Senior Reset and Boot ASIC Engineer to join our System ASIC team! NVIDIA has continuously reinvented itself over two decades. Our invention of the GPU in 1999 fueled the growth of the PC gaming market, redefined modern computer graphics, and revolutionized parallel computing. More recently, GPU deep learning ignited modern AI - the next era of computing. NVIDIA is a “learning machine” that constantly evolves by adapting to new opportunities that are hard to pursue, that only we can tackle, and that matter to the world. This is our life's work, to amplify human creativity and intelligence. Make the choice to join us today.
With the System-ASIC team, you will contribute to designing multiple products that will be used in a variety of applications from consumer graphics to self-driving cars and the growing field of artificial intelligence. System-ASIC team works closely with System Architecture team on product definitions, implements new innovative flows to improve our chip yield and to address multiple sectors, defines Reset and Boot Architecture and implements them in complex GPU and Tegra chips. The team is also handling the architecture, design, and synthesis of multiple System-level modules.
What you'll be doing:
Be an integral part of the System ASIC Design team to help with the Micro-architecture definition for system-level functions, including Reset and Boot.
Collaborate with architects, ASIC designers, and verification engineers to design sophisticated system-level modules such as Floorsweep, In-silicon measurement, Reset and Boot controllers.
You will be responsible for the RTL design, logic synthesis, and timing analysis of several modules.
Integrate modules into the overall SOC design and work closely with other teams in the silicon bring-up process and ensure successful SOC level integration.
You will also use your expertise to identify and implement improvements in the current design flow and methodologies to improve efficiency and quality.
What we need to see:
BS / MS in Microelectronics/ Electronic Engineering/ Computer Science and related (or equivalent experience)
5+ years of SoC design/integration experience, including architecture and implementation of system-level functions like Reset or Chip Boot
Solid frontend ASIC design skills, including RTL design, asynchronous and synchronous Reset design, synthesis, timing analysis, and Spyglass/CDC/RDC checks
Excellent analytical and problem-solving skills and attention to detail.
Ways to stand out from the crowd:
Broad knowledge with GPU techniques, SOC architecture, Computer architecture, and ARM/RISC-V CPU
Experience with general interface protocols such as PCIE, CXL, UCIE, AMBA, and CHI
Familiar with OCP secure boot specification and physical security handling process
Possess design experience with cross-die/chipset communication in multi-dielet/chipset systems
Proficient in Perl/Python scripting
With competitive salaries and a generous benefits package, we are widely considered to be one of the technology world's most desirable employers. We have some of the most brilliant people in the world working for us and, due to unprecedented growth, our teams are rapidly growing. Are you passionate about becoming a part of a best-in-class team supporting the latest in GPU and AI technology? If so, we want to hear from you.
#LI-Hybrid
Your base salary will be determined based on your location, experience, and the pay of employees in similar positions. The base salary range is 136,000 USD - 212,750 USD for Level 3, and 168,000 USD - 264,500 USD for Level 4.
You will also be eligible for equity and benefits.
Applications for this job will be accepted at least until November 11, 2025.NVIDIA is committed to fostering a diverse work environment and proud to be an equal opportunity employer. As we highly value diversity in our current and future employees, we do not discriminate (including in our hiring and promotion practices) on the basis of race, religion, color, national origin, gender, gender expression, sexual orientation, age, marital status, veteran status, disability status or any other characteristic protected by law.
Auto-ApplySenior Math Libraries Engineer - Dense Linear Algebra
Santa Clara, CA jobs
We are looking for software engineers to join our development efforts in the area of dense linear algebra kernels for high-performance libraries such as cu SOLVER. Around the world, leading commercial and academic organizations are revolutionizing AI, data analytics, and scientific and engineering simulations, using data centers powered by GPUs and high-performance linear algebra libraries. Applications of these technologies include computer aided engineering (CAE), electronic design automation (EDA), quantum chemistry, autonomous vehicles, LLMs, computer vision, encryption, and countless others. Did you know our team develops the GPU accelerated libraries and SDKs that help make these possible?
In this role, you will work together with other developers on designing, developing, and optimizing kernels for various algorithms including triangular factorizations, eigenvalue decompositions and singular value decompositions. Ideal candidates will not only have experience developing accelerated computing kernels, but also be motivated to advance the state-of-the-art in a variety of accelerated computing domains. If this sounds exciting, we would love to meet you!
What you will be doing:
Designing, implementing and optimizing scalable high-performance numerical dense linear algebra software on GPUs
Providing technical leadership and guidance to library engineers, QA engineers, and interns working with you on projects
Working closely with product management and other internal and external partners to understand feature and performance requirements and contribute to the technical roadmaps of libraries
Finding and realizing opportunities to improve library quality, performance and maintainability through re-architecting and establishing innovative software development practices
What we need to see:
PhD or MSc degree in Computational Science and Engineering, Computer Science, Applied Mathematics, or related science or engineering field (or equivalent experience)
5+ years of overall experience in developing, debugging and optimizing high-performance numerical linear algebra software using C++ and parallel programming; ideally using CUDA, MPI, OpenMP, OpenACC, pthreads
Strong fundamentals in numerical methods such as computational linear algebra, linear system solvers, and methods for eigenvalue, singular value, and other decompositions
Experience developing dense linear algebra libraries such as BLAS, LAPACK; and their parallel counterparts like PBLAS and SCALAPACK
Strong collaboration, communication, and documentation habits
Ways to stand out from the crowd:
Good knowledge of CPU and/or GPU hardware architecture
Experience with adopting and advancing, software development practices such as CI/CD systems and project management tools such as JIRA.
Experience with working in a globally distributed organization
Strong background of large-scale computing technologies such as PDE solvers, eigenvalue solvers and time-domain simulation methods (e.g., CFD, FEA)
NVIDIA's invention of the GPU in 1999 sparked the growth of the PC gaming market, redefined modern computer graphics, and revolutionized parallel computing for science and engineering. More recently, GPU deep learning ignited modern AI - the next era of computing - with the GPU acting as the brain of computers, robots, and self-driving cars that can perceive and understand the world. Today, we are increasingly known as “the AI computing company.” We're looking to grow our company and build our teams with the smartest people in the world! Join us at the forefront of technological advancement.
#LI-Hybrid
Your base salary will be determined based on your location, experience, and the pay of employees in similar positions. The base salary range is 184,000 USD - 287,500 USD for Level 4, and 224,000 USD - 356,500 USD for Level 5.
You will also be eligible for equity and benefits.
Applications for this job will be accepted at least until July 29, 2025.NVIDIA is committed to fostering a diverse work environment and proud to be an equal opportunity employer. As we highly value diversity in our current and future employees, we do not discriminate (including in our hiring and promotion practices) on the basis of race, religion, color, national origin, gender, gender expression, sexual orientation, age, marital status, veteran status, disability status or any other characteristic protected by law.
Auto-ApplySenior Engineer, GenAI for Media and Gaming
Santa Clara, CA jobs
NVIDIA is seeking an outstanding engineer in conversational AI to join our AI for Games team. We are developing next-generation gaming solutions enriched with AI Assistants, Actors, and Agents, and we are looking for outstanding engineers to help us achieve this vision. If you are passionate about generative AI, language models, conversational pipelines, and their applications in games, this is the opportunity for you. Collaborating with other teams across the company, you will productize promising research and develop new features through your own work.
What you'll be doing:
Use AI to solve product challenges in gaming and other interactive experiences.
Build upon the latest research to create world-class conversational pipelines for AI assistants and agents.
Improve and fine-tune language models and retrieval-augmented generation solutions for accuracy and performance.
Build prototypes to demonstrate real-life applications of your ideas and to accelerate productization.
Collaborate with NVIDIA's internal and external teams, including AI/DL researchers, hardware architects, and software engineers.
Participate in technology transfers to and from teams across NVIDIA.
What we need to see:
PhD or Master's degree in Computer Science/Engineering, Machine Learning, AI, or related fields; or equivalent experience.
12+ years of work experience with last 5+ years focused on language models, AI assistants, and agents.
Proficiency in C, C++, and Python, with the ability to write high-performance production code.
Experience with GPU programming, CUDA, and system optimizations is a significant plus.
A track record of proven research excellence, demonstrated through presentations, demos, or publications at leading venues such as GDC, ICCV/ECCV, SIGGRAPH, or other research artifacts such as software projects or significant product development.
AI-powered machines can learn, reason, and interact with people, thanks to GPU deep learning. We offer competitive salaries and great benefits as a top tech employer with leading talent.
Your base salary will be determined based on your location, experience, and the pay of employees in similar positions. The base salary range is 224,000 USD - 356,500 USD for Level 5, and 272,000 USD - 425,500 USD for Level 6.
You will also be eligible for equity and benefits.
Applications for this job will be accepted at least until November 2, 2025.NVIDIA is committed to fostering a diverse work environment and proud to be an equal opportunity employer. As we highly value diversity in our current and future employees, we do not discriminate (including in our hiring and promotion practices) on the basis of race, religion, color, national origin, gender, gender expression, sexual orientation, age, marital status, veteran status, disability status or any other characteristic protected by law.
Auto-ApplySenior Algorithm Engineer, Map-Perception Fusion
Santa Clara, CA jobs
Are you passionate about the future of autonomous vehicles? Join our DRIVE Road Structure, Online Mapping, and Context Fusion team to help build the next generation of self-driving technology! In this role, you'll work alongside experienced engineers to develop advanced 3D world models that combine map and perception data, enabling vehicles to safely navigate complex roads and intersections in the U.S. and China. This is a fantastic opportunity to grow your skills and contribute to groundbreaking projects at the heart of our vision for autonomous driving.
What you'll be doing:
Contribute to Core Autonomous Driving Systems: Help develop and refine systems that integrate state-of-the-art perception and mapping technologies for use in diverse driving environments.
Build Real-Time 3D World Models: Assist in creating dynamic 3D models by combining data from sensors and external sources to support planning and navigation.
Support Mapless Driving Capabilities: Work on enriched bird's-eye-view (BEV) models to enable autonomous vehicles to navigate complex urban settings without relying solely on high-definition maps.
Enhance Scenario Understanding: Help build static obstacle and occupancy grids, and generate occlusion masks to improve the vehicle's understanding of its environment.
Collaborate Across Teams: Work closely with perception and planning teams, learning from experts and contributing to shared goals.
Maintain High Standards: Write high-quality code and follow automotive industry best practices for safety and reliability.
What we need to see:
Bachelor's or Master's degree in Computer Science, Engineering, or a related field (or equivalent experience).
5+ years of relevant software development experience, ideally with some exposure to robotics, autonomous vehicles, or related fields.
Eagerness to learn and grow in the field of robotics and autonomous driving.
Strong problem-solving and analytical skills.
Good communication and teamwork abilities, especially in a diverse and global environment.
Experience with embedded systems or automotive platforms is a plus.
Ways to Stand Out from the Crowd:
Familiarity with 3D geometry, computer vision, sensor calibration, or state estimation algorithms (e.g., Kalman or particle filters).
Experience working on real-time computer vision systems.
Skills in designing APIs or software frameworks.
At NVIDIA, we're dedicated to making self-driving vehicles a reality and believe this technology can save millions of lives. Join a team of innovative thinkers at one of the world's most respected technology companies. If you're motivated, curious, and ready to make a difference, we'd love to meet you! We believe that building self-driving vehicles will be a defining contribution of our generation (e.g. traffic accidents are responsible for ~1.25 million deaths per year world-wide). We have the funding and scale, but we need your help on our team. NVIDIA is widely considered to be one of the technology world's most desirable employers with some of the most forward-thinking people in the world working here. If you're entrepreneurial and autonomous, we want to hear from you!
Your base salary will be determined based on your location, experience, and the pay of employees in similar positions. The base salary range is 148,000 USD - 235,750 USD for Level 3, and 184,000 USD - 287,500 USD for Level 4.
You will also be eligible for equity and benefits.
Applications for this job will be accepted at least until December 9, 2025.NVIDIA is committed to fostering a diverse work environment and proud to be an equal opportunity employer. As we highly value diversity in our current and future employees, we do not discriminate (including in our hiring and promotion practices) on the basis of race, religion, color, national origin, gender, gender expression, sexual orientation, age, marital status, veteran status, disability status or any other characteristic protected by law.
Auto-ApplySenior Backend Compiler Engineer - GPU
Santa Clara, CA jobs
NVIDIA's invention of the GPU 1999 sparked the growth of the PC gaming market, redefined modern computer graphics, and revolutionized parallel computing. More recently, GPU deep learning ignited modern AI - the next era of computing - with the GPU acting as the brain of computers, robots, and self-driving cars that can perceive and understand the world. Today, we are increasingly known as “the AI computing company”.
We are searching for a Senior Backend Compiler Engineer with experience in LLVM code generation for an exciting and fun role in our GPU Software organization. Our Compiler team is responsible for constructing and emitting the highest performance GPU machine instructions for Graphics (OpenGL, Vulkan, DX) and Compute (CUDA, PTX, OpenCL, Fortran, C++). This team is comprised of worldwide leading compiler engineering experts who provide leading edge performance and capabilities for NVIDIA's current and future complex parallel SIMT architectures.
What you will be doing:
Guide the design and implementation of a new LLVM based backend JIT compiler targeting NVIDIA GPUs
Design and develop new compiler passes and optimizations required to meet the high performance and low compilation time requirements of the compiler
Work with global compiler, hardware and application teams to oversee improvements and problem resolutions
Apply and adapt the latest compiler technologies to produce premier, robust, supportable compiler and tools
Get a chance to influence and work on new GPU architectures
What we need to see:
B.S. or higher degree in Computer Science/Engineering or equivalent experience
5+ years of compiler development experience including 3+ years of experience in LLVM with a focus in code generation
Expertise in LLVM code generation including instruction scheduling, software pipelining, register allocation, GlobalISel, TableGen, LLVM IR, and Machine IR (MIR)
Proficient in compiler design and in performance analysis tools
Excellent hands-on C++ programming skills applied to industry standard C++ compilers and development frameworks
Strong background in software engineering principles with a focus on crafting robust and maintainable solutions to challenging problems
Good communication and documentation skills and self-motivated
Ways to stand out from the crowd:
Masters or PhD preferred, with 5+ years of relevant experience
Experience developing compiler optimizations and code generation for GPUs
Experience in developing CUDA, DirectX, OpenGL/Vulkan applications
You should have strong interpersonal skills and be capable of leading and influencing others in the department
With competitive salaries and a generous benefits package, we are widely considered to be one of the technology world's most desirable employers. We have some of the most forward-thinking and hardworking people in the world working for us and, due to unprecedented growth, our exclusive engineering teams are rapidly growing. If you're a creative and autonomous engineer with a real passion for technology, we want to hear from you.
Your base salary will be determined based on your location, experience, and the pay of employees in similar positions. The base salary range is 148,000 USD - 235,750 USD for Level 3, and 184,000 USD - 287,500 USD for Level 4.
You will also be eligible for equity and benefits.
Applications for this job will be accepted at least until December 15, 2025.NVIDIA is committed to fostering a diverse work environment and proud to be an equal opportunity employer. As we highly value diversity in our current and future employees, we do not discriminate (including in our hiring and promotion practices) on the basis of race, religion, color, national origin, gender, gender expression, sexual orientation, age, marital status, veteran status, disability status or any other characteristic protected by law.
Auto-ApplySenior Perception Engineer
Santa Clara, CA jobs
Intelligent machines powered by Artificial Intelligence computers that can learn, reason and interact with people are no longer science fiction. GPU Deep Learning has provided the foundation for machines to learn, perceive, reason and solve problems. Now, NVIDIA's GPU runs Deep Learning algorithms, simulating human intelligence, and acts as the brain of computers, robots and self-driving cars that can perceive and understand the world.
We are now looking for an extraordinary Senior Perception Engineer to develop and productize NVIDIA's autonomous driving solutions. As a member of our perception team, you will be driving E2E solutions for various components of Wait Condition Perception and Fine grained recognition tasks. You will be challenged to improve robustness and accuracy as well as efficiency of the solutions to fully enable autonomous driving anywhere and anytime.
What you'll be doing:
Designing end2end solutions for Perception and AV stack to enable wait condition and fine grained classification tasks in complex driving environments. Example perception signals include traffic light, traffic signs, roadmarks, texts, classes of dynamic objects, and vehicles' light signals (brake, turn, hazard etc).
Applied research and development of innovative deep learning models. Additionally, develop localization and tracking algorithms to improve output accuracy of detection and classification solutions under challenging and diverse scenarios.
Develop generalizable approaches to support ODD and Country/region expansion
Productize the developed perception solutions by meeting product requirements for safety, latency, and SW robustness.
Drive and prioritize data-driven development by working with large data collection and labeling teams to bring in high value data to improve perception system accuracy. Efforts will include data collection prioritization and planning, labeling prioritization, labeling efficiency optimization, so that value of data is maximized.
What we need to see:
2+ years of technical leadership demonstrating high technical and organizational complexity
12+ years of hands-on work experience in developing deep learning and algorithms to solve sophisticated real world problems, and proficiency in using deep learning frameworks (e.g., PyTorch).
Experience in data-driven development and collaboration with data and ground truth teams.
Strong programming skills in python and/or C++.
Outstanding communication and teamwork skills as we work as a tightly-knit team, always discussing and learning from each other.
BS/MS/PhD in CS, EE, sciences or related fields (or equivalent experience)
Ways to stand out from the crowd:
Proven expertise in developing perception solutions for autonomous driving or robotics using deep learning with cameras.
Hands-on experience in developing and deploying DNN-based solutions to embedded platforms for real time applications.
Proven expertise in deep learning backed up by technical publications in leading conferences/journals.
Good understanding of fundamentals of 3D computer vision, camera calibrations including intrinsic and extrinsic.
Experience with development in CUDA language. The ability to implement CUDA kernels as part of training or inference pipelines.
Your base salary will be determined based on your location, experience, and the pay of employees in similar positions. The base salary range is 224,000 USD - 356,500 USD.
You will also be eligible for equity and benefits.
Applications for this job will be accepted at least until November 21, 2025.NVIDIA is committed to fostering a diverse work environment and proud to be an equal opportunity employer. As we highly value diversity in our current and future employees, we do not discriminate (including in our hiring and promotion practices) on the basis of race, religion, color, national origin, gender, gender expression, sexual orientation, age, marital status, veteran status, disability status or any other characteristic protected by law.
Auto-ApplySenior Deep Learning Algorithm Engineer
Santa Clara, CA jobs
NVIDIA is looking for engineers for our core AI Frameworks (Megatron Core and NeMo Framework) team to design, develop and optimize diverse real world workloads. Megatron Core and NeMo Framework are open-source, scalable and cloud-native frameworks built for researchers and developers working on Large Language Models (LLM) and Multimodal (MM) foundation model pretraining and post-training. Our GenAI Frameworks provide end-to-end model training, including pretraining, reasoning, alignment, customization, evaluation, deployment and tooling to optimize performance and user experience.
In this critical role, you will expand Megatron Core and NeMo Framework's capabilities, enabling users to develop, train, and optimize models by designing and implementing the latest in distributed training algorithms, model parallel paradigms, model optimizations, defining robust APIs, meticulously analyzing and tuning performance, and expanding our toolkits and libraries to be more comprehensive and coherent. You will collaborate with internal partners, users, and members of the open source community to analyze, design, and implement highly optimized solutions.
What you'll be doing:
Develop algorithms for AI/DL, data analytics, machine learning, or scientific computing
Contribute and advance open source NeMo-RL, Megatron Core, NeMo Framework
Solve large-scale, end-to-end AI training and inference challenges, spanning the full model lifecycle from initial orchestration, data pre-processing, running of model training and tuning, to model deployment.
Work at the intersection of compter-architecture, libraries, frameworks, AI applications and the entire software stack.
Innovate and improve model architectures, distributed training algorithms, and model parallel paradigms.
Performance tuning and optimizations, model training and finetuning with mixed precision recipes on next-gen NVIDIA GPU architectures.
Research, prototype, and develop robust and scalable AI tools and pipelines.
What we need to see:
MS, PhD or equivalent experience in Computer Science, AI, Applied Math, or related fields.
5+ years of industry experience.
Experience with AI Frameworks (e.g. PyTorch, JAX, Ray), and/or inference and deployment environments (e.g. TRTLLM, vLLM, SGLang).
Proficient in Python programming, software design, debugging, performance analysis, test design and documentation.
Consistent record of working effectively across multiple engineering initiatives and improving AI libraries with new innovations.
Strong understanding of AI/Deep-Learning fundamentals and their practical applications.
Ways to stand out from the crowd:
Hands-on experience in large-scale AI training, with a deep understanding of core compute system concepts (such as latency/throughput bottlenecks, pipelining, and multiprocessing) and demonstrated excellence in related performance analysis and tuning.
Prior experience with Reinforcement Learning algorithms and compute patterns
Expertise in distributed computing, model parallelism, and mixed precision training
Prior experience with Generative AI techniques applied to LLM and Multi-Modal learning (Text, Image, and Video).
Knowledge of GPU/CPU architecture and related numerical software.
Your base salary will be determined based on your location, experience, and the pay of employees in similar positions. The base salary range is 184,000 USD - 287,500 USD for Level 4, and 224,000 USD - 356,500 USD for Level 5.
You will also be eligible for equity and benefits.
Applications for this job will be accepted at least until October 28, 2025.NVIDIA is committed to fostering a diverse work environment and proud to be an equal opportunity employer. As we highly value diversity in our current and future employees, we do not discriminate (including in our hiring and promotion practices) on the basis of race, religion, color, national origin, gender, gender expression, sexual orientation, age, marital status, veteran status, disability status or any other characteristic protected by law.#deeplearning
Auto-ApplySenior DFT Methodology Engineer
Santa Clara, CA jobs
NVIDIA has continuously reinvented itself over two decades. Our invention of the GPU in 1999 sparked the growth of the PC gaming market, redefined modern computer graphics, and revolutionized parallel computing. More recently, GPU deep learning ignited modern AI - the next era of computing. NVIDIA is a “learning machine” that constantly evolves by adapting to new opportunities that are hard to solve, that only we can tackle, and that matter to the world. This is our life's work , to amplify human imagination and intelligence. Make the choice to join us today.
DFX Methodology Group at NVIDIA works on groundbreaking innovations involving crafting creative solutions for cutting edge test techniques, in-system test architecture, as well as verification and post-silicon validation on some of the industry's most complex semiconductor chips.
What you'll be doing:
As a member in our team, you will work next generation test architectures. You will work with multi-functional teams, implementing brand-new designs in test access mechanisms, high-speed test interfaces, and in-system test architecture.
In addition, you will help develop and deploy In-System Test (IST) methodologies for our next generation products for scan architecture, ATPG, MBIST, and IOBIST applications.
You will also help mentor junior engineers on test designs and trade-offs including cost and quality.
What we need to see:
BSEE (or equivalent experience) with 5+ years, MSEE with 3+ years, or PhD with 2+ years of experience in DFT, system architecture, or RTL design.
Understanding of fundamental DFT topics, such as, fault modeling, ATPG and fault simulation. Tessent ATPG/SSN experience is a plus.
Experience in Silicon debug and bring-up on the ATE or SLT platforms, post silicon validation and characterization.
Excellent understanding of MBIST and IOBIST fundamentals.
Experience in architecting DFT access mechanisms in 3D stacked and dielet/chiplet based designs, and UCIe protocol.
Good understanding of RTL coding principles.
Knowledge of high-speed interface architectures such as PCIe, USB3, DDR is a plus.
Excellent analytical skills in verification and validation of logic on complex and multi-million gate designs using vendor tools.
Good exposure to multi-functional areas including RTL & clocks design, STA, place-n-route and power.
Strong programming and scripting skills in Perl, Python or Tcl desired.
Outstanding written and oral communication skills with the curiosity to work on rare challenges.
NVIDIA offers highly competitive salaries and a comprehensive benefits package. We have some of the most brilliant and talented people in the world working for us and, due to unprecedented growth, our world-class engineering teams are growing fast. If you're a creative and autonomous engineer with real passion for technology, we want to hear from you!
#LI-Hybrid
Your base salary will be determined based on your location, experience, and the pay of employees in similar positions. The base salary range is 136,000 USD - 212,750 USD for Level 3, and 168,000 USD - 264,500 USD for Level 4.
You will also be eligible for equity and benefits.
Applications for this job will be accepted at least until December 8, 2025.NVIDIA is committed to fostering a diverse work environment and proud to be an equal opportunity employer. As we highly value diversity in our current and future employees, we do not discriminate (including in our hiring and promotion practices) on the basis of race, religion, color, national origin, gender, gender expression, sexual orientation, age, marital status, veteran status, disability status or any other characteristic protected by law.
Auto-ApplySenior Timing Methodology Engineer
Santa Clara, CA jobs
NVIDIA has continuously reinvented itself over two decades. Our invention of the GPU in 1999 sparked the growth of the PC gaming market, redefined modern computer graphics, and revolutionized parallel computing. More recently, GPU deep learning ignited modern AI - the next era of computing. NVIDIA is a “learning machine” that constantly evolves by adapting to new opportunities that are hard to resolve, that only we can seek, and that matter to the world. This is our life's work, to amplify human inventiveness and intelligence.
We are seeking an innovative Senior Timing Methodology Engineer to help drive sign-off strategies for the world's leading GPUs and SoCs. This position is a broad opportunity to optimize performance, yield, and reliability through increasingly comprehensive modeling, informative analysis, and automation. This work will influence the entire next generation computing landscape through critical contributions across NVIDIA's many product lines ranging from consumer graphics to self-driving cars and the growing domain of artificial intelligence! We have crafted a team of highly motivated people whose mission is to push the frontiers of what is possible today and define the platform for the future of computing. If you are fascinated by the immense scale of precision, craftsmanship, and artistry required to make billions of transistors function on every die at technology nodes as deep as 5 nm and beyond, this is an ideal role.
What You'll Be Doing:
Improve and validate flows for Prime-Time , Prime-Shield and Tempus STA QoR metrics for sign-off flow, and tool for high-speed designs, with focus on CAD and automation.
Develop custom flows for validating QoR of ETM models, both of std cells and custom IPs.
Develop flows/recommendations on STA sign-off to model deep submicron physical effects aging, self-heating, thermal impact, IR drop etc.
Collaborate with technology leads, VLSI physical design, and timing engineers to define and deploy the most sophisticated strategies of signing off timing in design for world-class silicon performance.
Develop tools, and methodologies to improve design performance, predictability, and silicon reliability beyond what industry standard tools can offer.
Work on various aspects of STA, constraints, timing and power optimization.
What We Need To See:
MS (or equivalent experience) in Electrical or Computer Engineering with 3 years' experience in ASIC Design and Timing.
Good understanding of modeling circuits for sign-off
Good knowledge of extraction, device physics, STA methodology and EDA tools limitations. Good understanding of mathematics/physics fundamentals of electrical design.
Clear understanding of low power design techniques such as multi VT, Clock gating, Power gating, Block Activity Power, and Dynamic Voltage-Frequency Scaling (DVFS), CDC, signal/power integrity, etc.
Understanding of 3DIC, stacking, packing, self-heating and its impact on timing/STA closure.
Background with crosstalk, electro-migration, noise, OCV, timing margins. Familiarity with Clocking specs: jitter, IR drop, crosstalk, spice analysis.
Understanding of standard cells/memory/IO IP modeling and its usage in the ASIC flow. Hands-on experience in advanced CMOS technologies, design with FinFET technology 5nm/3nm/2nm and beyond.
Expertise in coding- TCL, Python. C++ is a plus. Familiarity with industry standard ASIC tools: PT, ICC, Redhawk, Tempus etc.
Strong communications skill and good standout colleague
With competitive salaries and a generous benefits package, NVIDIA is widely considered to be one of the technology world's most desirable employers. We welcome you join our team with some of the most hard-working people in the world working together to promote rapid growth. Are you passionate about becoming a part of a best-in-class team supporting the latest in GPU and AI technology? If so, we want to hear from you.
Your base salary will be determined based on your location, experience, and the pay of employees in similar positions. The base salary range is 136,000 USD - 212,750 USD for Level 3, and 168,000 USD - 264,500 USD for Level 4.
You will also be eligible for equity and benefits.
Applications for this job will be accepted at least until December 7, 2025.NVIDIA is committed to fostering a diverse work environment and proud to be an equal opportunity employer. As we highly value diversity in our current and future employees, we do not discriminate (including in our hiring and promotion practices) on the basis of race, religion, color, national origin, gender, gender expression, sexual orientation, age, marital status, veteran status, disability status or any other characteristic protected by law.
Auto-ApplySenior AI Workflow Engineer
Santa Clara, CA jobs
Today, NVIDIA is tapping into the unlimited potential of AI to define the next era of computing. An era in which our GPU acts as the brains of computers, robots, and self-driving cars that can understand the world. Doing what's never been done before takes vision, innovation, and the world's best talent. As an NVIDIAN, you'll be immersed in a diverse, encouraging environment where everyone is inspired to do their best work. Come join the team and see how we can make a lasting impact on the world.
NVIDIA is hiring senior software engineers in its Infrastructure, Planning and Process Team (IPP), to accelerate AI adoption across various engineering workflows within the company. IPP is a global organization within NVIDIA. The group works with various other teams within NVIDIA such as Graphics Processors, Mobile Processors, Deep Learning, Artificial Intelligence and Driverless Cars to cater to their infrastructure and software development workflow needs. As a senior engineer on AI Workflow, you will design and implement tools and software solutions that leverage Large Language Models and agentic AI to automate end to end software engineering workflows and enhance the productivity of engineers across NVIDIA.
What you'll be doing:
Design and implement AI-driven optimizations within software development workflows to enhance developer productivity, accelerate feedback loops, and improve release reliability.
Experience designing, developing, and deploying AI agents to automate and software development workflows and processes.
Continuously measure and report on the impact of AI interventions, demonstrating improvements in key metrics like cycle time, change failure rate, and mean time to recovery (MTTR).
Create and deploy predictive models to identify high-risk commits, forecast potential build failures, and flag changes that have a high probability of failures.
Conduct research on emerging technologies to recommend best practices and improvements
What we need to see:
BE (MS preferred) or equivalent experience in EE/CS with 10+ years of work experience.
Well versed with Large Language Mode (LLM), Machine Learning (ML), Agentic AI techniques.
Hands-on experience in using large language models (LLMs) and implementing AI for software engineering workflows.
Hands-on experience on Python/Java/Go with extensive python scripting experience.
Experience in working with SQL/NoSQL database systems such as MySQL, MongoDB or Elasticsearch.
Experience in Full stack development.. Proficient in front-end (e.g., React, Angular, Vue.js, HTML, CSS, JavaScript), back-end (e.g., Node.js, Python/Django/Flask, Ruby on Rails, Java/Spring, .NET) development, database management (SQL/NoSQL), and deployment/hosting (e.g., AWS, Azure, GCP).
Experience with tools for CI/CD setup such as Jenkins, Gitlab CI, Packer, Terraform, Artifactory, Ansible, Chef or similar tools.
Good understanding of distributed systems, understanding of microservice architecture and REST APIs.
Good to have knowledge of build tools like Make, Maven or Ant.
Ability to effectively work across organizational boundaries to enhance alignment and productivity between teams.
Ways to stand out from the crowd:
Proactively track AI tool and technology trends, build insights, and collaborate with development teams early to evangelize AI driven workflows NVIDIA adoption.
We have some of the most forward-thinking and versatile people in the world working for us and, due to unprecedented growth, our best-in-class engineering teams are rapidly growing. We are building a team that will truly change the world. If you are passionate about new technologies, care about software quality, and want to be part of the future of transportation and AI, I would love for you to join us.
Expertise in leveraging large language models (LLMs) and Agentic AI to automate complex workflows, with knowledge of retrieval-augmented generation(RAG) and fine-tuning LLMs on enterprise data.
Prior development of a large software project using service oriented architecture operating with real time constraints
We have some of the most forward-thinking and versatile people in the world working for us and, due to unprecedented growth, our best-in-class engineering teams are rapidly growing. We are building a team that will truly change the world. If you are passionate about new technologies, care about software quality, and want to be part of the future of transportation and AI, we would love for you to join us.
Your base salary will be determined based on your location, experience, and the pay of employees in similar positions. The base salary range is 184,000 USD - 287,500 USD for Level 4, and 224,000 USD - 356,500 USD for Level 5.
You will also be eligible for equity and benefits.
Applications for this job will be accepted at least until October 4, 2025.NVIDIA is committed to fostering a diverse work environment and proud to be an equal opportunity employer. As we highly value diversity in our current and future employees, we do not discriminate (including in our hiring and promotion practices) on the basis of race, religion, color, national origin, gender, gender expression, sexual orientation, age, marital status, veteran status, disability status or any other characteristic protected by law.
Auto-ApplySenior GenAI Algorithms Engineer - Post-Training Optimizations
Santa Clara, CA jobs
NVIDIA is at the forefront of the generative AI revolution! The Algorithmic Model Optimization Team specifically focuses on optimizing generative AI models such as large language models (LLM) and diffusion models for maximal inference efficiency using techniques ranging from quantization, speculative decoding, sparsity, knowledge distillation, pruning to neural architecture search, and streamlined deployment strategies with open-sourced inference frameworks. Seeking a Senior Deep Learning Algorithms Engineer to improve innovative LLMs, VLMs, and multi-modality models. In this role, you will design, implement, and productionize model optimization algorithms for inference and deployment on NVIDIA's latest hardware platforms. The focus is on ease of use, compute and memory efficiency, and achieving the best accuracy-performance tradeoffs through software-hardware co-design.
Your work will span multiple layers of the AI software stack-ranging from algorithm design to integration-within NVIDIA's ecosystem (TensorRT Model Optimizer, Megatron-LM, Megatron-Bridge, Nvidia-NeMo, NeMo-AutoModel, TensorRT-LLM) and open-source frameworks (PyTorch, Hugging Face, vLLM, SGLang). You may also dive deeper into GPU-level optimization, including custom kernel development with CUDA and Triton. This role offers a unique opportunity to work at the intersection of research and engineering, pushing the boundaries of large-scale AI optimization. We are looking for passionate engineers with strong foundations in both machine learning and software systems/architecture who are eager to make a broad impact across the AI stack.
What you'll be doing:
Design and build modular, scalable model optimization software platforms that deliver exceptional user experiences while supporting diverse AI models and optimization techniques to drive widespread adoption.
Explore, develop, and integrate innovative deep learning optimization algorithms (e.g., quantization, speculative decoding, sparsity) into NVIDIA's AI software stack, e.g., TensorRT Model Optimizer, NeMo/Megatron, and TensorRT-LLM.
Construct and curate large problem specific datasets for post-training, finetuning, and reinforcement learning.
Deploy optimized models into leading OSS inference frameworks and contribute specialized APIs, model-level optimizations, and new features tailored to the latest NVIDIA hardware capabilities.
Partner with NVIDIA teams to deliver model optimization solutions for customer use cases, ensuring optimal end-to-end workflows and balanced accuracy-performance trade-offs.
Drive continuous innovation in deep learning inference performance to strengthen NVIDIA platform integration and expand market adoption across the AI inference ecosystem.
What we need to see:
Master's, PhD, or equivalent experience in Computer Science, Artificial Intelligence, Applied Mathematics, or a related field.
5+ years of relevant work or research experience in deep learning.
Strong software design skills, including debugging, performance analysis, and test development.
Proficiency in Python, PyTorch, and modern ML frameworks/tools.
Proven foundation in algorithms and programming fundamentals.
Strong written and verbal communication skills, with the ability to work both independently and collaboratively in a fast-paced environment.
Ways to stand out from the crowd:
Contributions to PyTorch, Megatron-LM, NeMo, TensorRT-LLM, vLLM, SGLang, or other machine learning training and inference frameworks.
Hands-on training, fine-tuning, or reinforcement learning experience on LLM or VLM models with large-scale GPU clusters.
Proficient in GPU architectures and compilation stacks, adept at analyzing and debugging end-to-end performance.
Familiarity with NVIDIA's deep learning SDKs (e.g., NeMo, TensorRT, TensorRT-LLM).
Increasingly known as “the AI computing company” and widely considered to be one of the technology world's most desirable employers, NVIDIA offers highly competitive salaries and a comprehensive benefits package. Are you creative, motivated, and love a challenge? If so, we want to hear from you! Come, join our model optimization group, where you can help build real-time, cost-effective computing platforms driving our success in this exciting and rapidly-growing field.
Your base salary will be determined based on your location, experience, and the pay of employees in similar positions. The base salary range is 184,000 USD - 287,500 USD.
You will also be eligible for equity and benefits.
Applications for this job will be accepted at least until September 20, 2025.NVIDIA is committed to fostering a diverse work environment and proud to be an equal opportunity employer. As we highly value diversity in our current and future employees, we do not discriminate (including in our hiring and promotion practices) on the basis of race, religion, color, national origin, gender, gender expression, sexual orientation, age, marital status, veteran status, disability status or any other characteristic protected by law.
Auto-ApplySenior SoC Power Analysis Engineer
Santa Clara, CA jobs
We are now looking for a Senior SoC Power Analysis Engineer!
NVIDIA is seeking an exceptional silicon power analysis and optimization engineer to help us build power-efficient and performance-leading SOC's. This position offers the opportunity to have real impact in a dynamic, technology-focused company impacting product lines ranging from consumer graphics to self-driving cars, to the growing field of artificial intelligence. We have built a team of extraordinary people, whose mission is to push the frontiers of what is possible today and define the platform for the future of computing.
What you'll be doing:
Analyzing and optimizing Silicon power and performance across technical, product, and usage constraints for Nvidia SoCs used in data center, driving, robotics, multimedia, edge computing, and other embedded applications.
Working with teams throughout the company (HW, SW, Platform, Thermal, Operations, Marketing, etc...) to deliver outstanding power solutions.
Guiding HW, SW and Platform teams in evaluating and improving Perf/Watt and battery life of their implementations.
Creating power models of key SoC units to evaluate architectural tradeoffs in DL/ML (training/inference), CPU, GPU, and multimedia workloads.
What we need to see:
A Master's degree in Computer Engineering or Electrical Engineering or equivalent experience
8+ years of additional relevant work experience focused on computer architecture and SOC power-perf analysis and optimization.
Understanding of the many factors influencing power efficiency at HW, SW, system, and product levels.
Good understanding of HW-SW interactions for Deep learning, Multimedia, or Productivity use cases
Strong interpersonal and teamwork skills.
A drive to continuously learn and expand architectural breadth and depth.
Ways to stand out from the crowd:
Experience with modeling and optimization of power and/or performance.
Python scripting and SW programming
Experience with prototyping power-optimization ideas on silicon
NVIDIA is widely considered to be one of the technology world's most desirable employers. We have some of the most forward-thinking and hardworking people in the world working for us. If you're creative, autonomous and love a challenge, we want to hear from you!
Your base salary will be determined based on your location, experience, and the pay of employees in similar positions. The base salary range is 184,000 USD - 287,500 USD for Level 4, and 224,000 USD - 356,500 USD for Level 5.
You will also be eligible for equity and benefits.
Applications for this job will be accepted at least until November 23, 2025.NVIDIA is committed to fostering a diverse work environment and proud to be an equal opportunity employer. As we highly value diversity in our current and future employees, we do not discriminate (including in our hiring and promotion practices) on the basis of race, religion, color, national origin, gender, gender expression, sexual orientation, age, marital status, veteran status, disability status or any other characteristic protected by law.
Auto-ApplySenior Math Libraries Engineer, Post Quantum Cryptography
Santa Clara, CA jobs
At NVIDIA, we're solving the world's most ambitious problems with our unique approach to accelerated computing. We're looking for a passionate and energized software engineer to join the NVIDIA Cryptography team working on ground-breaking cryptography solutions. In this role, you will contribute to the development and integration of cryptographic algorithms and low-level mathematical primitives within cu PQC Software Development Kit (SDK), focusing on Post-Quantum Cryptography (PQC), and Privacy-Enhancing Technologies (PETs). Our SDK is focused on security software stacks targeting a range of processors from edge computing to data center class GPUs. These high-performance libraries provide accelerated cryptographic building blocks as well as NIST standard PQC algorithms.
The advent of quantum computing requires major changes in the cryptography landscape, driving significant growth in the market for Post-Quantum Cryptography (PQC) and Privacy-Enhancing Technologies (PETs). As these technologies become increasingly essential for securing sensitive data and ensuring privacy across various industries, the demand for robust and efficient solutions is higher than ever. You will play a crucial role in designing, developing, and optimizing cryptographic algorithms and low-level mathematical primitives for high-performance execution on NVIDIA GPUs, influencing the creation of APIs, documentation, and the overall developer experience. Does the idea of being at the heart of these projects and applying your knowledge to architect and optimize algorithms which make an impact around the world excite you? If yes, then come and join our team!
What you will be doing:
Develop and optimize scalable high-performance cryptographic primitives, algorithms, and building blocks on the latest GPU hardware architectures
Emphasize robust long term software architectures and designs that effectively use many generations of hardware
Work closely with internal(product management, engineering) and external partners to understand feature and performance requirements and deliver timely cu PQC releases.
What we need to see:
PhD or MSc's degree in Applied Mathematics, Computer Science, or related science or engineering field is preferred (or equivalent experience).
5+ years of experience designing and developing software for cryptography in low latency or high throughput environments
Strong mathematical foundations.
Advanced C++ skills, including modern design paradigms (e.g. template meta-programming, SFINAE, RAII, constexpr, etc).
Strong collaboration, communication, and documentation habits.
Ways to stand out from the crowd:
Experience developing libraries consumed by many users.
Experience with CUDA C++, and GPU computing.
Programming skills with contemporary automation setups for both building software (e.g. cmake) as well as testing (e.g. CI/CD, sanitizers).
Strong understanding of mathematical foundations and algorithms used in cryptography, including but not limited to finite field arithmetic, lattice-based cryptography, and cryptographic hash functions.
With competitive salaries and a generous benefits package, we are widely considered to be one of the technology world's most desirable employers. We have some of the most forward-thinking and hardworking people in the world working for us and, due to unprecedented growth, our exclusive engineering teams are rapidly growing. If you're a creative and autonomous engineer with a real passion for technology, we want to hear from you!
#LI-Hybrid
Your base salary will be determined based on your location, experience, and the pay of employees in similar positions. The base salary range is 184,000 USD - 287,500 USD for Level 4, and 224,000 USD - 356,500 USD for Level 5.
You will also be eligible for equity and benefits.
Applications for this job will be accepted at least until August 25, 2025.NVIDIA is committed to fostering a diverse work environment and proud to be an equal opportunity employer. As we highly value diversity in our current and future employees, we do not discriminate (including in our hiring and promotion practices) on the basis of race, religion, color, national origin, gender, gender expression, sexual orientation, age, marital status, veteran status, disability status or any other characteristic protected by law.
Auto-ApplySenior GenAI Algorithms Engineer
Santa Clara, CA jobs
We are now looking for a Senior Gen AI Algorithms Engineer! NVIDIA is seeking engineers to design, develop and optimize Artificial Intelligence solutions to diverse real-world problems. If you have a strong understanding of deep learning and in particular large language models and their multimodal variants, then this role may be a great fit for you! Collaborate and interact with internal partners, users, and members of the open source community to analyze, define and implement highly optimized AI algorithms. The scope of these efforts includes a combination of implementing new algorithms, performance/accuracy tuning and analysis, defining APIs, and analyzing functionality coverage to build larger, coherent toolsets and libraries. The ability to work in a multifaceted, product-centric environment with excellent interpersonal skills are required, to be successful in this role.
What you'll be doing:
Contribute to the cutting-edge open source NeMo framework
Develop and maintain SOTA GenAI models (e.g., large language models (LLMs), multimodal LLMs)
Tackle large-scale distributed systems capable of performing end-to-end AI training and inference-deployment (data fetching, pre-processing, orchestrate and run model training and tuning, model serving)
Analyze, influence, and improve AI/DL libraries, frameworks and APIs according to good engineering practices
Research, prototype, and develop effective tools and infrastructure pipelines
Publish innovative results on Github and scientific publications
What we need to see:
A PhD or Master's Degree (or equivalent experience) and 5+ years of industry experience in Computer Science, AI, Applied Math, or related field
Strong mathematical fundamentals and AI/DL algorithms skills or experience
Excellent programming, debugging, performance analysis, test design and documentation skills
Experience with AI/DL Frameworks (e.g. PyTorch, JAX)
Excellent Python programming skills
Ways to stand out from the crowd:
Prior experience with Generative AI techniques applied to LLMs and multimodal variants (Image, Video, Speech etc.)
Exposure to large-scale AI training, understanding of the compute system concepts (latency/throughput bottlenecks, pipelining, multiprocessing etc) and related performance analysis and tuning
Hands-on experience with inference and deployment environments would be an asset (e.g. TRT, ONNX, Triton)
Knowledge of GPU/CPU architecture and related numerical software
Your base salary will be determined based on your location, experience, and the pay of employees in similar positions. The base salary range is 148,000 USD - 235,750 USD for Level 3, and 184,000 USD - 287,500 USD for Level 4.
You will also be eligible for equity and benefits.
Applications for this job will be accepted at least until December 14, 2025.NVIDIA is committed to fostering a diverse work environment and proud to be an equal opportunity employer. As we highly value diversity in our current and future employees, we do not discriminate (including in our hiring and promotion practices) on the basis of race, religion, color, national origin, gender, gender expression, sexual orientation, age, marital status, veteran status, disability status or any other characteristic protected by law.
Auto-Apply