Upvote
Downvote
Applied Researcher I
Share Job
- Suggest Revision
$230,000 a year
- We are committed to building world-class applied science and engineering teams and continue our industry leading capabilities with breakthrough product experiences and scalable, high-performance AI infrastructure.
- Experience building large deep learning models, whether on language, images, events, or graphs, as well as expertise in one or more of the following: training optimization, self-supervised learning, robustness, explainability, RLHF.
- PhD in Computer Science, Machine Learning, Computer Engineering, Applied Mathematics, Electrical Engineering or related fields
- LLM PhD focus on NLP or Masters with 5 years of industrial NLP research experience Multiple publications on topics related to the pre-training of large language models (e.g. technical reports of pre-trained LLMs, SSL techniques, model pre-training optimization) Member of team that has trained a large language model from scratch (10B + parameters, 500B+ tokens) Publications in deep learning theory Publications at ACL, NAACL and EMNLP, Neurips, ICML or ICLR
- Optimization (Training & Inference) PhD focused on topics related to optimizing training of very large deep learning models Multiple years of experience and/or publications on one of the following topics: Model Sparsification, Quantization, Training Parallelism/Partitioning Design, Gradient Checkpointing, Model Compression Experience optimizing training for a 10B+ model Deep knowledge of deep learning algorithmic and/or optimizer design Experience with compiler design
Active Job
Updated TodaySimilar Job
Relevance
Active