ML Research Engineer - LLM Safety

August 27

Apply Now
Logo of Dynamo AI

Dynamo AI

11 - 50

πŸ’° $15.1M Series A on 2023-08

Description

β€’ Own an LLM vertical with a focus on a specific safety domain, technique, or use case (either from defense or red-team attack perspective) β€’ Generate high quality synthetic data, train LLMs, and conduct rigorous benchmarking. β€’ Deliver robust, scalable, and reproducible production code. β€’ Push the envelope by developing novel techniques and research that delivers the world’s most harmless and helpful models. Your research will directly empower our customers to more feasibly deploy safe and responsible LLMs. β€’ Co-author papers, patents, and presentations with our research team by integrating other members’ work with your vertical.

Requirements

β€’ Deep domain knowledge in LLM safety techniques. β€’ Extensive experience in designing, training, and implementing multiple different types of LLM models and architectures in the real world. Comfortability with leading end-to-end projects. β€’ Adaptability and flexibility. In both the academic and startup world, a new finding in the community may necessitate an abrupt shift in focus. You must be able to learn, implement, and extend state-of-the-art research. β€’ Preferred: past research or projects in either attacking or defending LLMs.

Apply Now

Similar Jobs

Built byΒ Lior Neu-ner. I'd love to hear your feedback β€” Get in touch via DM or lior@remoterocketship.com