PhD Research Intern, Large Language Models

October 22

Apply Now
Logo of NVIDIA

NVIDIA

GPU-accelerated computing • artificial intelligence • deep learning • virtual reality • gaming

10,000+

Description

• Investigate novel approaches to infuse theory-of-mind reasoning into the post- or pre-training phases of large language models. • Collaborate with other team members, teams, and/or external researchers. • Transfer your research to product groups to enable new products or types of products. • Opportunity to publish original research.

Requirements

• Currently pursuing a PhD Degree in Computer Science/Engineering, Electrical Engineering. • Research experience in at least one of the following areas: Large Language Models – training, alignment, and evaluation, Foundation Models, Multimodal Models/Agents, Vision-Language Models, Deep Learning, Model Compression, and Acceleration Techniques, Pruning, Quantization, NAS, Efficient Backbone Architecture, Distillation, Neural Architecture Search. • Strong research track record and publication record at top-tier conferences. • Excellent communication skills. • Excellent programming skills in some rapid prototyping environment such as Python; C++ and parallel programming (e.g., CUDA) is a plus. • Hands-on experience with large-scale model training is a plus. • Knowledge of common machine learning frameworks, such as PyTorch.

Benefits

• Eligible for Intern benefits.

Apply Now
Built by Lior Neu-ner. I'd love to hear your feedback — Get in touch via DM or lior@remoterocketship.com