Negotiable
Outside
Remote
USA
Summary: The AI Engineer role focuses on building, hosting, and managing Large Language Models (LLMs). The ideal candidate will possess a deep understanding of LLM training, deployment, and management, with hands-on experience in fine-tuning these models. Responsibilities include managing model performance and reliability in production environments while collaborating across various teams. This position requires strong technical skills in Python and cloud-native environments.
Key Responsibilities:
- Building, hosting, and managing LLMs.
- Hands-on experience building or fine-tuning LLMs.
- Understanding of model deployment and hosting pipelines (API-based serving, GPU utilization, scaling).
- Ability to manage and monitor model performance and reliability in production.
- Familiar with vector databases, tokenization, embeddings, and inference optimization.
- Strong grounding in Python and/or modern software engineering practices.
- Experience working in cloud-native environments and with containerization (Docker, Kubernetes).
- Ability to work in fast-paced, experimental environments where proof-of-concepts and iteration cycles are common.
- Strong communication and documentation skills capable of collaborating across engineering, data, and product teams.
Key Skills:
- Deep understanding of LLMs.
- Hands-on experience with model fine-tuning.
- Knowledge of deployment and hosting pipelines.
- Experience with vector databases and inference optimization.
- Strong Python programming skills.
- Experience with cloud-native environments and containerization.
- Strong communication and documentation skills.
Salary (Rate): undetermined
City: undetermined
Country: USA
Working Arrangements: remote
IR35 Status: outside IR35
Seniority Level: undetermined
Industry: IT
Focus:
- Building, Hosting, and Managing LLMs
- We re looking for an AI Engineer with a deep understanding of how Large Language Models (LLMs) are trained, deployed, and managed.
Key Expectations:
- Hands-on experience building or fine-tuning LLMs.
- Understanding of model deployment and hosting pipelines (API-based serving, GPU utilization, scaling).
- Ability to manage and monitor model performance and reliability in production.
- Familiar with vector databases, tokenization, embeddings, and inference optimization.
General Requirements for All Roles
- Strong grounding in Python and/or modern software engineering practices.
- Experience working in cloud-native environments and with containerization (Docker, Kubernetes).
- Ability to work in fast-paced, experimental environments where proof-of-concepts and iteration cycles are common.
- Strong communication and documentation skills capable of collaborating across engineering, data, and product teams.