- Design and optimize prompting strategies to enhance the output quality and relevance of large language models (LLMs).
- Develop adaptive and context-aware prompts tailored to specific use cases, ensuring AI outputs meet user needs.
- Build and maintain robust text processing pipelines, including tasks like tokenization, normalization, and data augmentation.
- Implement and fine-tune RAG systems to improve the relevance and accuracy of AI-generated content by integrating external knowledge sources.
- Fine-tune pre-trained LLMs on custom datasets, optimizing them for domain-specific tasks and improving their accuracy and effectiveness.
- Deploy and manage local instances of LLMs, ensuring they are optimized for speed, resource efficiency, and data privacy.
- Collaborate with cross-functional teams to integrate local LLMs into existing infrastructure and workflows, providing scalable AI solutions.
- Bachelor’s or Master’s degree in Computer Science, Engineering, Mathematics, or a related field. Advanced degrees or specialized training in AI/ML are a plus.
- At least 4 years of hands-on experience working with natural language processing (NLP) techniques and large language models (LLMs).
- Strong programming skills in languages such as Python or R.
- Deep understanding of neural networks, particularly generative models.
- Familiarity with cloud platforms (AWS, GCP, Azure) and distributed computing environments are a plus.
- Ability to work with large datasets and experience with data manipulation tools.
- Ability to collaborate and work in a cross-functional/multi-national team environment, work directly to client for analysis and collect requirements.
- Knowledge of version control systems (e.g., Git) and collaborative tools.
- Expertise in TensorFlow, PyTorch, and Hugging Face.
- Experience with tools like OpenAI API, Keras, or Scikit-learn.
- Understanding of architectures like GPT, BERT, and Vision Transformers (ViT).
- Expertise in training and tuning diffusion models, such as DALL·E or Stable Diffusion.
TBD