Posted:3 days ago|
Platform:
On-site
Part Time
Date Opened
Job Type
Industry
Work Experience
City
State/Province
Country
Zip/Postal Code
XenonStack is the fastest-growing data and AI foundry for agentic systems, enabling people and organizations to gain real-time and intelligent business insights.
Agentic Systems for AI Agents: akira.ai
Vision AI Platform: xenonstack.ai
Inference AI Infrastructure for Agentic Systems: nexastack.ai
We are looking for an LLM Reliability & Evaluation Engineer to design, implement, and maintain rigorous evaluation frameworks for Large Language Models (LLMs) powering our Agentic AI systems. This role will ensure models meet high standards of accuracy, safety, and performance across enterprise and regulated industry use cases.
If you’re passionate about model trustworthiness, benchmarking, and Responsible AI practices, and want to shape how AI agents behave in mission-critical workflows, this is the role for you.
Design, implement, and maintain evaluation pipelines for LLM-based applications and agentic workflows.
Define and track key performance indicators (accuracy, latency, cost, reliability) for deployed models.
Develop automated test suites, benchmark datasets, and stress-testing scenarios for LLMs.
Collaborate with data scientists, ML engineers, and product teams to integrate evaluation into the model lifecycle.
Assess bias, fairness, and safety risks in LLM outputs and recommend mitigations.
Validate model alignment to enterprise use case requirements and regulatory standards.
Conduct A/B tests, prompt performance analysis, and long-context reliability checks.
Document evaluation methodologies and maintain transparent reporting for internal and client use.
Stay updated on state-of-the-art LLM evaluation techniques, frameworks, and metrics.
Must-Have:
3–5 years of experience in AI/ML engineering, applied research, or QA for ML systems.
Strong understanding of LLM architectures, prompt engineering, and agentic workflows.
Proficiency in Python and ML frameworks (PyTorch, TensorFlow, Hugging Face).
Experience with dataset curation, evaluation metrics (BLEU, ROUGE, BERTScore, factuality scores), and performance profiling.
Familiarity with Responsible AI principles, fairness auditing, and bias detection methods.
Good-to-Have:
Experience with LangChain, LangGraph, or similar agent frameworks.
Knowledge of model monitoring tools (Weights & Biases, MLflow, Arize AI, TruLens).
Familiarity with multi-turn conversation evaluation and human-in-the-loop testing.
Exposure to regulated industry AI governance (finance, healthcare, etc.).
Continuous Learning & Growth
Certification sponsorships and advanced training in AI evaluation, safety, and optimization.
Access to cutting-edge AI systems and enterprise-scale evaluation environments.
Recognition & Rewards
Performance incentives and awards for innovation in model reliability.
Fast-track opportunities to AI Governance or Model Ops leadership roles.
Work Benefits & Well-Being
Comprehensive medical insurance and project-based allowances.
Cab facilities for women employees and additional perks for special projects.
We foster a culture of cultivation with bold, human-centric leadership principles. We value obsession and deep work in every initiative, and we are on a mission to reshape how enterprises adopt AI + Human Intelligence systems.
Product Values:
Obsessed with Adoption – Making AI accessible and enterprise-ready.
Obsessed with Simplicity – Turning complexity into seamless, intuitive AI experiences.
Be a part of our vision to accelerate the world’s transition to AI + Human Intelligence.
XenonStack
Upload Resume
Drag or click to upload
Your data is secure with us, protected by advanced encryption.
Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.
We have sent an OTP to your contact. Please enter it below to verify.
Practice Python coding challenges to boost your skills
Start Practicing Python Now4.0 - 8.0 Lacs P.A.
4.0 - 8.0 Lacs P.A.