Senior Research Engineer at NVIDIA
Join NVIDIA and help build the software that will define the future of generative AI. We are looking for a research engineer who is passionate about open‑source and excited to create our next‑generation post‑training software stack. You will work at the intersection of research and engineering, collaborating with the Post‑Training and Frameworks teams to invent, implement, and scale the core technologies behind our Nemotron models.
What You’ll Be Doing
- Work with applied researchers to design, implement, and test the next generation of RL and post‑training algorithms.
- Contribute to and advance open source by developing NeMo‑RL, Megatron Core, NeMo Framework, and other yet‑to‑be‑announced software.
- Be engaged as part of one team during Nemotron models post‑training.
- Solve large‑scale, end‑to‑end AI training and inference challenges, spanning the full model lifecycle from orchestration, data pre‑processing, model training and tuning, to deployment.
- Work at the intersection of computer architecture, libraries, frameworks, AI applications, and the entire software stack.
- Perform performance tuning and optimizations, model training with mixed‑precision recipes on next‑gen NVIDIA GPU architectures.
- Publish and present your results at academic and industry conferences.
What We Need To See
- BS, MS, or PhD in Computer Science, AI, Applied Math, or related fields (or equivalent experience).
- 3+ years of proven experience in machine learning, systems, distributed computing, or large‑scale model training.
- Experience with AI frameworks such as PyTorch or JAX.
- Experience with at least one inference and deployment environment such as vLLM, SGLang, or TRT‑LLM.
- Proficient in Python programming, software design, debugging, performance analysis, test design, and documentation.
- Strong understanding of AI/Deep‑Learning fundamentals and their practical applications.
Ways To Stand Out From The Crowd
- Contributions to open source deep learning libraries.
- Hands‑on experience in large‑scale AI training, with a deep understanding of core compute system concepts (e.g., latency/throughput bottlenecks, pipelining, multiprocessing) and demonstrated excellence in performance analysis and tuning.
- Expertise in distributed computing, model parallelism, and mixed‑precision training.
- Prior experience with Generative AI techniques applied to LLMs and Multi‑Modal learning (Text, Image, Video).
- Knowledge of GPU/CPU architecture and related numerical software.
NVIDIA is widely considered to be one of the technology world’s most desirable employers. We have some of the most forward‑thinking and hardworking people on the planet working with us. If you’re creative and autonomous, we want to hear from you!
#J-18808-Ljbffr
Kontaktperson:
Nvidia HR Team