AI Engineer, LLM / LMM Deployment, Q Team CoE
HTX Home Team Science Technology Agency
Job Description
The Home Team Science and Technology Agency (HTX) is a statutory board under the Ministry of Home Affairs (MHA) which aims to pioneer innovation solutions and develop world class science and technology capabilities to transform and empower the Home Team in delivering safety and security for Singapore.
The HTxAI Products team is committed to empower our Home Team Department (HTD)'s flagship projects through the seamless integration of AI technologies. From identifying opportunities to delivery, launch, and continuous enhancement, HTxAI aims to push boundaries and setting new standards in public safety.
We're looking for an AI Engineer working on LLM/LMM Operations to join our team working at the intersection of large language models and production systems. You'll help build and maintain the systems that powers our model fine-tuning workflows, RAG pipelines and deployment processes while partnering closely with application teams.
You'll be joining a growing team where you'll have significant influence on our technical direction and practices as we scale our ML capabilities. We support continuous learning through conference attendance (such as NeurIPS, AAAI, WSDM) and relevant courses to help you stay current with the rapidly evolving ML landscape.
What You'll DoBuild and maintain RAG systems that reliably serve our applicationsEstablish fine-tuning pipelines for domain-specific modelsDesign evaluation frameworks and monitoring systems to track model performance in productionDeploy and optimize models in production environmentsCollaborate with application teams to translate their needs into robust ML solutionsBuild and maintain critical components of our ML systems from development through deployment
Our InfrastructureThe majority of our work runs on-prem where we maintain access to the latest GPUs in the market. This setup gives our team the freedom to experiment and iterate quickly on R&D projects without cloud cost constraints or API rate limits. You'll work directly with this hardware to improve training runs, benchmark different approaches, and push the boundaries of what's possible with current models.We primarily work with open-source frameworks and in-house trained models, giving you the flexibility to choose the right tools for each challenge.
While cloud experience is valuable and we do use cloud services for certain workloads, expect to spend most of your time working with our on-prem infrastructure.
What We're Looking For2+ years of hands-on experience with ML operations, ideally working with LLMs/LMMs or similar systemsStrong Python skills and experience with ML frameworks (PyTorch, Transformers)Experience deploying models using Docker, Kubernetes, or cloud ML platformsPractice building or maintaining production ML systemsUnderstanding of the operational challenges in keeping ML systems running reliably
Nice to HaveExperience with vector databases, embedding models, and retrieval systemsFamiliarity with model evaluation methods and metricsExperience with experiment tracking tools (MLflow, Weights & Biases, etc.)Knowledge of CI/CD practices for ML workflows
You'll Thrive Here IfYou care about the details but don't lose sight of the bigger pictureYou can debug a failing deployment pipeline and think strategically about what metrics matterYou communicate well with both technical and non-technical teammatesYou're comfortable working in an environment where requirements evolve as we learn
- ** All new hires are appointed on a two-year contract in the first instance and will be assessed and considered for permanent tenure over time, based on performance. As part of the shortlisting process for this role, you may be required to complete a medical declaration and/or undergo further assessment. All applicants will be updated on the status of their applications within 4 weeks upon closing of the advertisement.