Do you have expertise in CUDA kernel optimization, C++ systems programming, or compiler infrastructure? Join NVIDIA's nvFuser team to build the next-generation fusion compiler that automatically optimizes deep learning models for workloads scaling to thousands of GPUs! We're looking for engineers who excel at parallel programming and systems-level performance work and want to directly impact the future of AI compilation.
The Deep Learning Frameworks Team @ NVIDIA is responsible for building nvFuser, an advanced compiler that sits at the intersection of compiler technology and high-performance computing. You'll work closely with the PyTorch Core team and collaborate with Lightning-AI/Thunder, which integrates nvFuser to accelerate PyTorch workloads. We collaborate with hardware architects, framework maintainers, and optimization experts to create compiler infrastructure that advances GPU performance, developing manual optimization techniques into systematic, automated compiler optimizations.
What you'll be doing
As an nvFuser engineer, you'll work on exciting challenges in compiler technology and performance optimization! You'll design algorithms that generate highly optimized code from deep learning programs and build GPU-aware CPU runtime systems that coordinate kernel execution for maximum performance. Working directly with NVIDIA's hardware engineers, you'll master the latest GPU architectures while collaborating with optimization specialists to develop innovative techniques for emerging AI workloads. From debugging performance bottlenecks in thousand-GPU distributed systems to influencing next-generation hardware design, we push the boundaries of what's possible in AI compilation.
What we need to see
MS or PhD in Computer Science, Computer Engineering, Electrical Engineering, or related field (or equivalent experience).
4+ years advanced C++ programming with large codebase development, template meta-programming, and performance-critical code.
Strong parallel programming experience with multi-threading, OpenMP, CUDA, MPI, NCCL, NVSHMEM, or other parallel computing technologies.
Shown experience with low-level performance optimization and systematic bottleneck identification beyond basic profiling.
Performance analysis skills: experience analyzing high-level programs to identify performance bottlenecks and develop optimization strategies.
Collaborative problem-solving approach with adaptability in ambiguous situations, first-principles based thinking, and a sense of ownership.
Excellent verbal and written communication skills.
Ways to stand out from the crowd
Experience with HPC/Scientific Computing: CUDA optimization, GPU programming, numerical libraries (cuBLAS, NCCL), or distributed computing.
Compiler engineering background: LLVM, GCC, domain-specific language design, program analysis, or IR transformations and optimization passes.
Deep technical foundation in CPU/GPU architectures, numeric libraries, modular software design, or runtime systems.
Experience with large software projects, performance profiling, and demonstrated track record of rapid learning.
Expertise with distributed parallelism techniques, tensor operations, auto-tuning, or performance modeling.
You will also be eligible for equity and benefits. NVIDIA accepts applications on an ongoing basis.
Top Skills
Similar Jobs
What you need to know about the Los Angeles Tech Scene
Key Facts About Los Angeles Tech
- Number of Tech Workers: 375,800; 5.5% of overall workforce (2024 CompTIA survey)
- Major Tech Employers: Snap, Netflix, SpaceX, Disney, Google
- Key Industries: Artificial intelligence, adtech, media, software, game development
- Funding Landscape: $11.6 billion in venture capital funding in 2024 (Pitchbook)
- Notable Investors: Strong Ventures, Fifth Wall, Upfront Ventures, Mucker Capital, Kittyhawk Ventures
- Research Centers and Universities: California Institute of Technology, UCLA, University of Southern California, UC Irvine, Pepperdine, California Institute for Immunology and Immunotherapy, Center for Quantum Science and Engineering