Deep Learning Engineer - Distributed Task-Based Backends

September 15

Apply Now
Logo of NVIDIA

NVIDIA

GPU-accelerated computing β€’ artificial intelligence β€’ deep learning β€’ virtual reality β€’ gaming

10,000+

Description

β€’ Develop extensions to popular Deep Learning frameworks, that enable easy experimentation with various parallelization strategies! β€’ Develop compiler optimizations and parallelization heuristics to improve the performance of AI models at extreme scales β€’ Develop tools that enable performance debugging of AI models at large scales β€’ Study and tune Deep Learning training workloads at large scale, including important enterprise and academic models β€’ Support enterprise customers and partners to scale novel models using our platform β€’ Collaborate with Deep Learning software and hardware teams across NVIDIA, to drive development of future Deep Learning libraries β€’ Contribute to the development of runtime systems that underlay the foundation of all distributed GPU computing at NVIDIA

Requirements

β€’ BS, MS or PhD degree in Computer Science, Electrical Engineering or related field (or equivalent experience) β€’ 5+ years of relevant industry experience or equivalent academic experience after BS β€’ Proficient with Python and C++ programming β€’ Strong background with parallel and distributed programming, preferably on GPUs β€’ Hands-on development skills using Machine Learning frameworks (e.g. PyTorch, TensorFlow, Jax, MXNet, scikit-learn etc.) β€’ Understanding of Deep Learning training in distributed contexts (multi-GPU, multi-node)

Benefits

β€’ Eligible for equity and benefits.

Apply Now

Similar Jobs

September 7

Develop Rust backend systems for a sports analytics betting startup.

Built byΒ Lior Neu-ner. I'd love to hear your feedback β€” Get in touch via DM or lior@remoterocketship.com