Location: Boston, New York, us
Job Summary:
Job Duties:
- Build large-scale distributed fine-tuning and training infrastructure.
- Deploy LLMs on GPU instances for real-time use.
- Collaborate with AI specialists on service design and deployment.
- Present research in journal club on LLMs and Generative AI.
- Develop, monitor, and maintain LLM models and services.
Required Skills:
- Language Models (NLP).
- Distributed training frameworks (PyTorch, TensorFlow).
- CUDA (plus).
- Code simplicity and performance.
Required Experiences:
- BS/MS/PhD in Computer Science or related field.
- Experience with large-scale systems and datasets.
- Production data pipelines (plus).
Job URLs: