Failed to check availability
Retry
Original
Simplified
- On-Premise Open-Source LLM Hosting and Deployment:
- Design and implement infrastructure that supports large-scale LLM hosting, ensuring security, scalability, and resource optimization.
- Develop best practices for the deployment and lifecycle management of open-source LLMs, from testing to production monitoring.
- Multi-Modal LLM Operations:
- Collaborate with NLP and data science teams to ensure multi-modal models are trained, optimised, and deployed for cross-functional use cases.- Implement strategies for efficient data handling, storage, and processing of multi-modal inputs and outputs.
- Data Preparation for AI Models:
- MLOps Pipeline Development and Management:
- Automate workflows to streamline the development and deployment of ML and LLM models, with a focus on reducing latency and computational overhead.- Set up CI/CD pipelines for continuous integration and delivery of ML models.
- Model Monitoring, Optimization, and Maintenance:
- Collaboration and Documentation:
- 5-10 years of experience in MLOps, Data Engineering, or similar roles with a focus on ML model deployment and operationalization.
- Strong knowledge of MLOps, LLM deployment, and open-source model hosting, with hands-on experience in models like Llama and Mistral.
- Expertise in multi-modal LLMs and handling diverse data types (text, image, audio).
- Proficiency in Python, SQL, and ML/LLM frameworks (e.g., Hugging Face Transformers, TensorFlow, PyTorch).
- Familiarity with containerization and orchestration tools (Docker, Kubernetes) and cloud platforms (AWS, GCP, Azure) as well as on-premise infrastructure.
- Experience with CI/CD tools (Azure DevOps, GitLab, or similar).
- Strong problem-solving skills and the ability to adapt to a fast-paced, dynamic environment.
Similar Jobs