AI Engineer – LLM Infrastructure & Hosting
AI Engineer LLM Infrastructure & Hosting
Location : Remote
Focus:
- Building, Hosting, and Managing LLMs
- We’re looking for an AI Engineer with a deep understanding of how Large Language Models (LLMs) are trained, deployed, and managed.
Key Expectations:
- Hands-on experience building or fine-tuning LLMs.
- Understanding of model deployment and hosting pipelines (API-based serving, GPU utilization, scaling).
- Ability to manage and monitor model performance and reliability in production.
- Familiar with vector databases, tokenization, embeddings, and inference optimization.
General Requirements for All Roles
-
Strong grounding in Python and/or modern software engineering practices.
-
Experience working in cloud-native environments and with containerization (Docker, Kubernetes).
-
Ability to work in fast-paced, experimental environments where proof-of-concepts and iteration cycles are common.
-
Strong communication and documentation skills – capable of collaborating across engineering, data, and product teams.