Home
Jobs
Companies
Resume

20 Llmops Jobs

Filter
Filter Interviews
Min: 0 years
Max: 25 years
Min: ₹0
Max: ₹10000000
Setup a job Alert
JobPe aggregates results for easy application access, but you actually apply on the job portal directly.

8.0 - 10.0 years

6 - 10 Lacs

Noida

Work from Office

Naukri logo

Position Summary LLMOps(Large language model operations) Engineer will play a pivotal role in building and maintaining the infrastructure and pipelines for our cutting-edge Generative AI applications, establishing efficient and scalable systems for LLM research, evaluation, training, and fine-tuning. Engineer will be responsible for managing and optimizing large language models (LLMs) across various platforms This position is uniquely tailored for those who excel in crafting pipelines, cloud infrastructure, environments, and workflows. Your expertise in automating and streamlining the ML lifecycle will be instrumental in ensuring the efficiency, scalability, and reliability of our Generative AI models and associated platform. LLMOps engineers expertise will ensure the smooth deployment, maintenance, and performance of these AI platforms and powerful large language models. You will follow Site Reliability Engineering & MLOps principles and will be encouraged to contribute your own best practices and ideas to our ways of working. Reporting to the Head of Cloud Native operations, you will be an experienced thought leader, and comfortable engaging senior managers and technologists. You will engage with clients, display technical leadership, and guide the creation of efficient and complex products/solutions. Key Responsibilities Technical & Architectural Leadership Contribute to the technical delivery of projects, ensuring a high quality of work that adheres to best practices, brings innovative approaches and meets client expectations. Project types include following (but not limited to): Solution architecture, Proof of concepts (PoCs), MVP, design, develop, and implementation of ML/LLM pipelines for generative AI models, data management & preparation for fine tuning, training, deployment, and monitoring. Automate ML tasks across the model lifecycle. Contribute to HCL thought leadership across the Cloud Native domain with an expert understanding of advanced AI solutions using Large Language Models (LLM) & Natural Language Processing (NLP) techniques and partner technologies. Collaborate with cross-functional teams to integrate LLM and NLP technologies into existing systems. Ensure the highest levels of governance and compliance are maintained in all ML and LLM operations. Stay abreast of the latest developments in ML and LLM technologies and methodologies, integrating these innovations to enhance operational efficiency and model effectiveness. Collaborate with global peers from partner ecosystems on joint technical projects. This partner ecosystem includes Google, Microsoft, Nvidia, AWS, IBM, Red Hat, Intel, Cisco, and Dell VMware etc. Service Delivery Provide a technical hands-on contribution. Create scalable infra to support enterprise loads (distributed GPU compute, foundation models, orchestrating across multiple cloud vendors, etc.) Ensuring the reliable and efficient platform operations. Apply data science, machine learning, deep learning, and natural language processing methods to analyse, process, and improve the models data and performance. Understanding of Explainability & Biased Detection concepts. Create and optimize prompts and queries for retrieval augmented generation and prompt engineering techniques to enhance the models capabilities and user experience w.r.t Operations & associated platforms. Client-facing influence and guidance, engaging in consultative client discussions and performing a Trusted Advisor role. Provide effective support to HCL Sales and Delivery teams. Support sales pursuits and enable HCL revenue growth. Define the modernization strategy for client platform and associated IT practices, create solution architecture and provide oversight of the client journey. Innovation & Initiative Always maintain hands-on technical credibility, keep in front of the industry, and be prepared to show and lead the way forward to others. Engage in technical innovation and support HCLs position as an industry leader. Actively contribute to HCL sponsorship of leading industry bodies such as the CNCF and Linux Foundation. Contribute to thought leadership by writing Whitepapers, blogs, and speaking at industry events. Be a trusted, knowledgeable internal innovator driving success across our global workforce. Client Relationships Advise on best practices related to platform & Operations engineering and cloud native operations, run client briefings and workshops, and engage technical leaders in a strategic dialogue. Develop and maintain strong relationships with client stakeholders. Perform a Trusted Advisor role. Contribute to technical projects with a strong focus on technical excellence and on-time delivery. Mandatory Skills & Experience Expertise in designing and optimizing machine-learning operations, with a preference for LLMOps. Proficient in Data Science, Machine Learning, Python, SQL, Linux/Unix shell scripting. Experience on Large Language Models and Natural Language Processing (NLP), and experience with researching, training, and fine-tuning LLMs. Contribute towards fine-tune Transformer models for optimal performance in NLP tasks, if required. Implement and maintain automated testing and deployment processes for machine learning models w.r.t LLMOps. Implement version control, CI/CD pipelines, and containerization techniques to streamline ML and LLM workflows. Develop and maintain robust monitoring and alerting systems for generative AI models ensuring proactive identification and resolution of issues. Research or engineering experience in deep learning with one or more of the following: generative models, segmentation, object detection, classification, model optimisations. Experience implementing RAG frameworks as part of available-ready products. Experience in setting up the infrastructure for the latest technology such as Kubernetes, Serverless, Containers, Microservices etc. Experience in scripting programming to automate deployments and testing, worked on tools like Terraform and Ansible. Scripting languages like Python, bash, YAML etc. Experience on CI/CD opensource and enterprise tool sets such as Argo CD, Jenkins. Experience with the GitHub/DevOps Lifecycle Experience in at least one of the Observability solutions (Prometheus, EFK stacks, ELK stacks, Grafana, Dynatrace, AppDynamics) Experience in at-least one of the clouds for example - Azure/AWS/GCP Significant experience on microservices-based, container-based or similar modern approaches of applications and workloads. You have exemplary verbal and written communication skills (English). Able to interact and influence at the highest level, you will be a confident presenter and speaker, able to command the respect of your audience. Desired Skills & Experience Bachelor level technical degree or equivalent experience; Computer Science, Data Science, or Engineering background preferred; masters degree desired. Experience in LLMOps or related areas, such as DevOps, data engineering, or ML infrastructure. Hands-on experience in deploying and managing machine learning and large language model pipelines in cloud platforms (e.g., AWS, Azure) for ML workloads. Familiar with data science, machine learning, deep learning, and natural language processing concepts, tools, and libraries such as Python, TensorFlow, PyTorch, NLTK etc. Experience in using retrieval augmented generation and prompt engineering techniques to improve the models quality and diversity to improve operations efficiency. Proven experience in developing and fine-tuning Language Models (LLMs). Stay up-to-date with the latest advancements in Generative AI, conduct research, and explore innovative techniques to improve model quality and efficiency. The perfect candidate will already be working within a System Integrator, Consulting or Enterprise organisation with 8+ years of experience in a technical role within the Cloud domain. Deep understanding of core practices including SRE, Agile, Scrum, XP and Domain Driven Design. Familiarity with the CNCF open-source community. Enjoy working in a fast-paced environment using the latest technologies, love Labs dynamic and high-energy atmosphere, and want to build your career with an industry leader.

Posted 6 days ago

Apply

12.0 - 16.0 years

40 - 50 Lacs

Pune, Chennai, Bengaluru

Hybrid

Naukri logo

AI Ops Senior Architect 12 -17 Years Work Location - Pune/ Bengaluru/Hyderabad/Chennai/ Gurugram Tredence is Data science, engineering, and analytics consulting company that partners with some of the leading global Retail, CPG, Industrial and Telecom companies. We deliver business impact by enabling last mile adoption of insights by uniting our strengths in business analytics, data science and data engineering. Headquartered in the San Francisco Bay Area, we partner with clients in US, Canada, and Europe. Bangalore is our largest Centre of Excellence with skilled analytics and technology teams serving our growing base of Fortune 500 clients. JOB DESCRIPTION At Tredence, you will lead the evolution of Industrializing AI ” solutions for our clients by implementing ML/LLM/GenAI & Agent Ops best practices. You will lead the Architecture , Design & development of large scale ML/LLMOps platforms for our clients. You’ll build and maintain tools for deployment, monitoring, and operations. You’ll be a trusted advisor to our clients in ML/GenAI/Agent Ops space & coach to the ML engineering practitioners to build effective solutions to Industrialize AI solutions THE IDEAL CANDIDATE WILL BE RESPONSIBLE FOR AI Ops Strategy, Innovation, Research and Technical Standards 1. Conduct research and experiment with emerging AI Ops technologies and trends. Create POV’s, POC’s & present Proof of Technology to use latest tools, Technologies & services from Hyper scalers focussed on ML, GenAI & Agent Ops 2. Define and propose new technical standards and best practices for the organization's AI Ops environment. 3. Lead the evaluation and adoption of innovative MLOps solutions to address critical business challenges. 4. Conduct meet ups, attend & present in Industry events, conferences, etc 5. Ideate & develop accelerators to strengthen service offerings of AI Ops practice Solution Design & Architectural Development 6. Lead Design & architecture of scalable model training & deployment pipelines for large-scale deployments 7. Architect & Design large scale ML & GenAI Ops platforms 8. Collaborate with Data science & GenAI practice to define and implement strategies of AI solutions for model explainability and interpretability 9. Mentor and guide senior architects in crafting cutting-edge AI Ops solutions 10. Lead architecture reviews and identify opportunities for significant optimizations and improvements. Documentation and Best Practices 11. Develop and maintain comprehensive documentation of AIOps architectures designs and best practices. 12. Lead the development and delivery of training materials and workshops on AIOps tools and techniques. 13. Actively participate in sharing knowledge and expertise with the MLOps team through internal presentations and code reviews. Qualifications and Skills: 1. Bachelor’s or Master’s degree in Computer Science, Data Science, or a related field with minimum 12 years of experience 2. Proven experience in architecting & developing AIOps solutions – to streamline Machine Learning & GenAI development lifecycle 3. Proven experience as an AI Ops Architect – ML & GenAI in architecting & design of ML & GenAI platforms 4. Hands on experience in Model deployment strategies, Designing ML & GenAI model pipelines to scale in production, Model Observability techniques used to monitor performance of ML & LLM’s 5. Strong coding skills with experience in implementing best coding practices Technical Skills & Expertise Python, PySpark, PyTorch ,Java, Micro Services, API’s LLMOps – Vector DB, RAG, LLM Orchestration tools, LLM Observability, LLM Guardrails, Responsible AI MLOps - MLFlow, ML/DL libraries, Model & Data Drift Detection libraries & techniques Real Time & Batch Streaming Container Orchestration Platforms Cloud platforms – Azure/ AWS/ GCP, Data Platforms – Databricks/ Snowflake Nice to Have: Understanding of Agent Ops Exposure to Databricks platform You can expect to – Work with world’s biggest Retailers, CPG’s, HealthCare, Banking & Manufacturing customers and help them solve some of their most critical problems Create multi-million Dollar business opportunities by leveraging impact mindset, cutting edge solutions and industry best practices. Work in a diverse environment that keeps evolving Hone your entrepreneurial skills as you contribute to growth of the organization

Posted 1 week ago

Apply

10.0 - 14.0 years

15 - 20 Lacs

Noida

Work from Office

Naukri logo

Position Summary The Principal AI Architect is responsible for leading the design and implementation of advanced AI solutions and strategic architecture. Working closely with technology leaders from across our global client community, you will be their senior Trusted Advisor for their AI-enabled transformation journey This role demands deep understanding of AI and related technologies, running in Edge, on-prem and Public Cloud environments. Acting at the forefront of our industry you will be fully conversant with Generative AI, and its impact at both the individual employee and strategic organisational level. The ideal candidate will be an established thought-leader, with solid architectural and engineering credentials. Working ahead of industry trends, deeply passionate about technology-enabled business transformation and demonstrating a strong innovation-led posture. As a thought leader, you will interact frequently with CxO level clients, AI industry leaders, provide expert opinions, and contribute to HCLs strategic vision. Key Responsibilities Technical & Engineering Leadership Design comprehensive AI solution and technology architecture, integrating latest AI technology developments into world-class solutions. Lead high-level architectural discussions with clients, providing expert guidance on best practices for AI implementations across AI PC, Edge, Data Centre and Public Cloud environments. Ensure solutions align with modern best practices across the full spectrum of platforms and environments. Deep understanding across GPU/NPU, Cognitive Infrastructure, Application and Copilot/agent domains. Contribute to HCLs thought leadership in the AI & Cloud domains with a deep understanding of opensource technologies (e.g., Kubernetes, OPEA) and partner technologies. Collaborate on joint technical projects with global partners, including Google, Microsoft, AWS, NVIDIA, IBM, Red Hat, Intel, and Dell. Service Delivery & Innovation Architect innovative AI solutions from ideation to MVP, rapidly enabling genuine business value. Optimize AI and cloud architectures to meet client requirements, balancing efficiency, accuracy and effectiveness. Assess and review existing complex solutions and recommend architectural improvements to transform applications with latest AI technologies. Drive the adoption of cutting-edge GenAI technologies spearheading initiatives that push the boundaries of AI integration across the full spectrum of environments. Thought Leadership and Client Engagement Provide expert architectural and strategy guidance to clients on incorporating Generative AI into their business and technology landscape. Conduct workshops, briefings, and strategic dialogues to educate clients on AI benefits and applications, establishing strong, trust-based relationships. Act as a trusted advisor, contributing to technical projects with a strong focus on technical excellence and on-time delivery. Author whitepapers, blogs, and speak at industry events, maintaining a visible presence as a thought leader in AI and associated technologies. Collaboration and Customer Engagement Engage with multiple customers simultaneously, providing high-impact consultative relationships. Work closely with internal teams and global partners to ensure seamless collaboration and knowledge sharing across projects. Maintain a hands-on technical credibility, staying ahead of industry trends and mentoring others in the organization. Mandatory Skills & Experience Experience: 10+ years architecture design • 7+ years in software engineering. Technologies: Professional-level expertise in Public Cloud environments (AWS, Azure, Google Cloud). Demonstrable coding proficiency with Python, Java or Go languages. AI Expertise: Advanced machine learning algorithms, GenAI models (e.g., GPT, BERT, DALL-E, GEMINI), NLP techniques. Working familiarity with Copilot solutions, in both software engineering and office productivity domains. Communication: Exemplary verbal and written communication skills. Project Methodologies: Agile and Scrum project management. Desired Skills & Experience Knowledge of GenAI operations (LLMOps), experience Governing AI models in production environments. Proficiency in data engineering for AI, including data preprocessing, feature engineering, and pipeline creation. Expertise in AI model fine-tuning and evaluation, with a focus on improving performance for specialized tasks. Copilot design, engineering and extensions. Knowledgeable about Responsible AI, including governance and ethics. Bias mitigation, with experience in implementing strategies to ensure fair and unbiased AI solutions. Deep Learning Frameworks (TensorFlow, PyTorch) Innovation and Emerging Technology Trends Strategic AI Vision and Road mapping Enthusiastic about working in a fast-paced environment using the latest technologies, and passionate about HCLs dynamic and high-energy Lab culture. Verifiable Certification Recognized Professional certification from Google, Microsoft or AWS in an AI and/or Cloud-related domain. Soft Skills and Behavioural Competencies Exemplary communication and leadership skills, capable of inspiring teams and making strategic decisions that align with business goals. Demonstrates a strong customer orientation, innovative problem-solving abilities, and effective cross-cultural collaboration. Expert at driving organizational change and fostering a culture of innovation.

Posted 1 week ago

Apply

10.0 - 14.0 years

18 - 20 Lacs

Noida

Work from Office

Naukri logo

Position Summary This is a highly visible role that requires a perfect combination of deep technical credibility, strategic acumen and demonstrable leadership competency. You will be the ultimate Trusted Advisor, capable of engaging business and technology leaders within the worlds largest enterprises, and guiding their strategic AI-enabled journey. The Country Leader, AI Architecture, is responsible for leading the Labs Architectural services within the region. You will need to provide hands-on technical leadership, whilst managing a small team of senior AI architects and consultants. Operating in a fast-moving, highly innovative environment, collaborating with senior Sales and Technical leaders. You will have business responsibility for the provision of innovation-led Labs services. focusing on the design and implementation of advanced AI solutions enabling genuine transformational outcomes. This hands-on leadership role demands deep understanding of AI and related technologies, running in Edge, onprem and Public Cloud environments. Acting at the forefront of our industry you will be fully conversant with Generative AI, and its impact at both the individual employee and strategic organisational level. The ideal candidate will be an established thought-leader in the AI domain, with solid architectural and engineering credentials that you maintain at the highest level. Working ahead of industry trends, deeply passionate about AI-enabled business transformation and demonstrating a strong innovation-led posture. As a thought leader, you will interact frequently with CxO level clients, industry leaders, provide expert opinions, and contribute to HCLs strategic vision. Key Responsibilities Technical & Engineering Leadership Act as ultimate Design Authority for sophisticated AI solutions and related technology architecture. Lead high-level architectural discussions with clients, providing expert guidance on best practices for AI implementations across AI PC, Edge, Data Centre and Public Cloud environments. Ensure solutions align with modern best practices across the full spectrum of platforms and environments. Deep understanding across GPU/NPU, Cognitive Infrastructure, Application and Copilot/agent domains. Contribute to HCLTech thought leadership in the AI & Cloud domains with a deep understanding of open-source (e.g., Kubernetes, OPEA) and partner technologies. Collaborate on joint technical projects with global partners, including Google, Microsoft, AWS, NVIDIA, IBM, Red Hat, Intel, and Dell. Service Delivery & Innovation Design innovative AI solutions from ideation to MVP, rapidly enabling genuine business value. Optimize AI and cloud architectures to meet client requirements, balancing efficiency, accuracy and effectiveness. Assess and review existing complex solutions and recommend architectural improvements to transform applications with latest AI technologies. Drive the adoption of cutting-edge GenAI technologies spearheading initiatives that push the boundaries of AI capability across the full spectrum of environments. Thought Leadership and Client Engagement Provide expert architectural and strategy guidance to clients on incorporating Generative AI into their business and technology landscape. Conduct workshops, briefings, and strategic dialogues to educate clients on AI benefits and applications, establishing strong, trust-based relationships. Act as a trusted advisor, contributing to technical projects with a strong focus on technical excellence and on-time delivery. Author whitepapers, blogs, and speak at industry events, maintaining a visible presence as a thought leader in AI and associated technologies. Collaboration and Customer Engagement Engage with multiple customers simultaneously, building high-impact consultative relationships. Work closely with internal teams and global partners to ensure seamless collaboration and knowledge sharing across projects. Maintain hands-on technical credibility, staying ahead of industry trends and mentoring others in the organization. Management and Leadership Demonstrable track record building and managing small Architectural or Engineering teams. Support career growth and professional development of the team. Enrich and enable world-class technical excellence across the team; supported by a culture of collaboration, respect, diversity, inclusion and deep trustful relationships. Mandatory Skills & Experience Management & leadership : Demonstrable track record building and leading Architectural or Engineering teams. Proven ability to combine strategic business and commercial skills, performing at the highest-level in senior client relationships. Experience: 10+ years architecture design 10+ years software engineering. 5+ years in a senior Team Leader or similar management position. Significant client-facing engagement within a GSI, system integrator, professional services or technology organization. Technologies: Professional-level expertise in Public Cloud environments (AWS, Azure, Google Cloud). Demonstrable coding proficiency with Python, Java or Go languages. AI Expertise: Advanced machine learning algorithms, GenAI models (e.g., GPT, BERT, DALL-E, GEMINI), NLP techniques. Working familiarity with Copilot solutions, in both software engineering and office productivity domains. Business Expertise: Extensive track record performing a lead technical role in a sales, business-development or other commercial environment. Negotiating and consultative skills; experience leading the complete engagement lifecycle. Communication: Experienced public speaker, with an ability to connect with senior business leaders. Project Methodologies: Agile and Scrum project management. Desired Skills & Experience Knowledge of GenAI operations (LLMOps), experience Governing AI models in production environments. Proficiency in data engineering for AI, including data preprocessing, feature engineering, and pipeline creation. Expertise in AI model fine-tuning and evaluation, with a focus on improving performance for specialized tasks. Copilot design, engineering and extensions. Knowledgeable about Responsible AI, including governance and ethics. Bias mitigation, with experience in implementing strategies to ensure fair and unbiased AI solutions. Deep Learning Frameworks (TensorFlow, PyTorch) Innovation and Emerging Technology Trends Strategic AI Vision and Road mapping Enthusiastic about working in a fast-paced environment using the latest technologies, and passionate about HCLs dynamic and high-energy Lab culture. Verifiable Certification Recognized Professional certification from Google, Microsoft or AWS in an AI and/or Cloudrelated domain. Soft Skills and Behavioural Competencies Exemplary communication and leadership skills, capable of inspiring teams and making strategic decisions that align with business goals. Demonstrates a strong customer orientation, innovative problem-solving abilities, and effective cross-cultural collaboration. Expert at driving organizational change and fostering a culture of innovation.

Posted 1 week ago

Apply

5.0 - 10.0 years

15 - 20 Lacs

Bengaluru

Work from Office

Naukri logo

Develop and deploy ML pipelines using MLOps tools, build FastAPI-based APIs, support LLMOps and real-time inferencing, collaborate with DS/DevOps teams, ensure performance and CI/CD compliance in AI infrastructure projects. Required Candidate profile Experienced Python developer with 4–8 years in MLOps, FastAPI, and AI/ML system deployment. Exposure to LLMOps, GenAI models, containerized environments, and strong collaboration across ML lifecycle

Posted 1 week ago

Apply

12.0 - 18.0 years

35 - 40 Lacs

Chennai

Work from Office

Naukri logo

Tech stack required: Programming languages: Python Public Cloud: AzureFrameworks: Vector Databases such as Milvus, Qdrant/ ChromaDB, or usage of CosmosDB or MongoDB as Vector stores. Knowledge of AI Orchestration, AI evaluation and Observability Tools. Knowledge of Guardrails strategy for LLM. Knowledge on Arize or any other ML/LLM observability tool. Experience: Experience in building functional platforms using ML, CV, LLM platforms. Experience in evaluating and monitoring AI platforms in production Nice to have requirements to the candidate Excellent communication skills, both written and verbal. Strong problem-solving and critical-thinking abilities. Effective leadership and mentoring skills. Ability to collaborate with cross-functional teams and stakeholders. Strong attention to detail and a commitment to delivering high-quality solutions. Adaptability and willingness to learn new technologies. Time management and organizational skills to handle multiple projects and priorities.

Posted 2 weeks ago

Apply

5.0 - 8.0 years

15 - 25 Lacs

Hyderabad, Pune, Bengaluru

Work from Office

Naukri logo

Role: Gen AI Engineer Exp: 5 to 8 yrs. Loc: Bangalore, Pune, Hyderabad NP: Immediate joiners, who can join in 30 days. Required Skills: Python, Large Language Models (LLM), Machine Learning (ML), Generative AI

Posted 2 weeks ago

Apply

10.0 - 12.0 years

0 Lacs

Bengaluru / Bangalore, Karnataka, India

On-site

Foundit logo

Job Description: Oracle Cloud Infrastructure (OCI) is a pioneering force in cloud technology, merging the agility of startups with the robustness of an enterprise software leader. Within OCI, the Oracle Generative AI Service team spearheads innovative solutions at the convergence of artificial intelligence and cloud infrastructure. As part of this team, you'll contribute to large-scale cloud solutions utilizing cutting-edge machine learning technologies, aimed at addressing complex global challenges. Join us to create innovative solutions using top-notch machine learning technologies to solve global challenges. We're looking for an experienced Principal Applied Data Scientist to join our OCI Gen-AI Solutions team for strategic customers. In this role, you'll collaborate with applied scientists and product managers to design, develop, and deploy tailored Gen-AI solutions with an emphasis on Large Language Models (LLMs), Agents, MPC and Retrieval Augmented Generation (RAG) with large OpenSearch clusters. As part of the OCI Gen AI and Data Solutions for strategic customers team, you will be responsible for developing innovative Gen AI and data services for our strategic customers.As a Principal Applied Data Scientist, you'll lead the development of advanced Gen AI solutions using the latest ML technologies combined with Oracle's cloud expertise. Your work will significantly impact sectors like financial services, telecom, healthcare, and code generation by creating distributed, scalable, high-performance solutions for strategic customers. Work directly with key customers and accompany them on their Gen AI journey - understanding their requirements, help them envision and design and build the right solutions and work together with their ML engineering to remove blockers. You will dive deep into model structure to optimize model performance and scalability. You will build state of art solutions with brand new technologies in this fast-evolving area. You will configure large scale OpenSearch clusters, setting up ingestion pipelines to get the data into the OpenSearch. You will diagnose, troubleshoot, and resolve issues in AI model training and serving. You may also perform other duties as assigned. Build re-usable solution patterns and reference solutions / showcases that can apply across multiple customers. Be an enthusiastic, self-motivated, and a great collaborator. Be our product evangelist - engage directly with customers and partners, participate and present in external events and conferences, etc. Qualifications and experience Bachelors or master's in computer science or equivalent technical field with 10+ years of experience Able to optimally communicate technical ideas verbally and in writing (technical proposals, design specs, architecture diagrams and presentations). Demonstrated experience in designing and implementing scalable AI models and solutions for production,relevant professional experience as end-to-end solutions engineer or architect (data engineering, data science and ML engineering is a plus), with evidence of close collaborations with PM and Dev teams. Experience with OpenSearch, Vector databases, PostgreSQL and Kafka Streaming. Practical experience with setting up and finetuning large OpenSearch Clusters. Experience in setting up data ingestion pipelines with OpenSearch. Experience with search algorithms, indexing, optimizing latency and response times. Practical experience with the latest technologies in LLM and generative AI, such as parameter-efficient fine-tuning, instruction fine-tuning, and advanced prompt engineering techniques like Tree-of-Thoughts. Familiarity with Agents and Agent frameworks and Model Predictive Control (MPC) Hands-on experience with emerging LLM frameworks and plugins, such as LangChain, LlamaIndex, VectorStores and Retrievers, LLM Cache, LLMOps (MLFlow), LMQL, Guidance, etc. Strong publication record, including as a lead author or reviewer, in top-tier journals or conferences. Ability and passion to mentor and develop junior machine learning engineers. Proficient in Python and shell scripting tools. Preferred Qualifications : Masters or Bachelor's in related field with 5+ years relevant experience Experience with RAG based solutions architecture. Familiarity in OpenSearch and Vector stores as a knowledge store Knowledge of LLM and experience delivering, Generative AI And Agent models are a significant plus. Familiarity and experience with the latest advancements in computer vision and multimodal modeling is a plus. Experience with semantic search, multi-modal search and conversational search. Experience in working on a public cloud environment, and in-depth knowledge of IaaS/PaaS industry and competitive capabilities.Experience with popular model training and serving frameworks like KServe, KubeFlow, Triton etc. Experience with LLM fine-tuning, especially the latest parameter efficient fine-tuning technologies and multi-task serving technologies. Deep technical understanding of Machine Learning, Deep Learning architectures like Transformers, training methods, and optimizers. Experience with deep learning frameworks (such as PyTorch, JAX, or TensorFlow) and deep learning architectures (especially Transformers). Experience in diagnosing, fixing, and resolving issues in AI model training and serving. Career Level - IC4

Posted 3 weeks ago

Apply

10.0 - 18.0 years

30 - 45 Lacs

Hyderabad, Pune, Bengaluru

Work from Office

Naukri logo

Role - Senior Data Scientist / Senior Gen AI Engineer Exp Range - 8 to 18 yrs Position - Permanent Fulltime Company - Data Analytics & AIML MNC Location - Hyderabad, Pune, Bangalore (Relocation accepted) About the Role: We are seeking a Software Engineer with expertise in Generative AI and Microsoft technologies to design, develop, and deploy AI-powered solutions using the Microsoft ecosystem. You will work with cross-functional teams to build scalable applications leveraging generative AI models and Azure services. Skills Required: Experience with Large Language Models (LLMs) like GPT, LLaMA, Claude, etc. Proficiency in Python for building and fine-tuning AI/ML models Familiarity with LangChain , LLMOps , or RAG (Retrieval-Augmented Generation) pipelines Experience with Vector Databases (e.g. FAISS, Pinecone, Weaviate) Knowledge of Prompt Engineering and model evaluation techniques Exposure to cloud platforms (Azure, AWS or GCP) for deploying GenAI solutions Preferred Skills: Experience with Azure OpenAI , Databricks or Microsoft Fabric Hands-on with Hugging Face Transformers , OpenAI APIs or custom model training

Posted 3 weeks ago

Apply

0.0 years

3 - 6 Lacs

Delhi, Delhi, IN

On-site

Internshala logo

About the job: Key responsibilities: 1. Build AI-driven tools and products using APIs (OpenAI, Gemini, etc.) 2. Design and fine-tune prompts for various use cases. 3. Integrate vector databases (Pinecone, ChromaDB) for retrieval-augmented generation (RAG) 4. Use tools like LangChain or LlamaIndex for multi-step worklows 5. Collaborate with designers, content teams, and founders to turn ideas into polished tools Who can apply: Only those candidates can apply who: are Computer Science Engineering students Salary: ₹ 3,20,000 - 6,50,000 /year Experience: 0 year(s) Deadline: 2025-06-22 23:59:59 Skills required: Natural Language Processing (NLP), Deep Learning, Prompt Engineering, ChatGPT, Claude, Gemini, LLMOps and Model fine-tuning Other Requirements: 1. Degree Btech - AI/Ml, others ( who has done AI/ML projects) 2. Strong understanding of LLM APIs (OpenAI, Claude, Gemini, etc.) 3. REST API integration and deployment knowledge 4. GitHub portfolio with working AI tools or integrations About Company: Stirring Minds is a premier startup ecosystem in India, dedicated to helping businesses launch, scale, and succeed. As a leading incubator, we provide funding, co-working spaces, and mentorship to support the growth of innovative companies. In addition to our incubator services, we also host the largest startup event in the country known as Startup Summit Live, bringing together entrepreneurs and industry leaders to connect, learn, and collaborate. Our community-driven approach extends beyond our event and incubator offerings, as we work to create communities of like-minded individuals who can support and learn from one another. We have been recognized by top media outlets both in India and internationally, including the BBC, The Guardian, Entrepreneur, and Business Insider. Our goal is to provide a comprehensive ecosystem for startups and help turn their ideas into reality.

Posted 3 weeks ago

Apply

0.0 years

3 - 4 Lacs

IN

Remote

Internshala logo

About the job: Key responsibilities: 1. Design and develop scalable backend systems and APIs for our AI-powered SaaS platform using Python and Node.js 2. Build and maintain cloud infrastructure on AWS, including configuration and management of S3, DynamoDB, SNS, EC2, and CloudWatch services 3. Implement and optimize data processing pipelines for machine learning model deployment and integration 4. Collaborate with data scientists to integrate AI models into production systems and ensure efficient model serving 5. Deploy and monitor applications using DevOps practices and LLMOps for large language model implementations 6. Create robust API endpoints that connect our frontend applications with AI functionalities 7. Design and implement efficient database schemas and queries optimized for AI applications 8. Develop and maintain secure authentication and authorization systems for our platform 9. Write clean, maintainable, and well-tested code following best practices 10. Troubleshoot and resolve complex technical issues in production environments Additional candidate preferences: 1. Computer Science or related Engineering degree preferred 2. Experience with containerization technologies like Docker 3. Familiarity with AI model serving platforms Who can apply: Only those candidates can apply who: are Computer Science Engineering students Salary: ₹ 3,10,000 - 4,60,000 /year Experience: 0 year(s) Deadline: 2025-06-16 23:59:59 Other perks: 5 days a week Skills required: Python, Node.js, Artificial intelligence, DevOps, Amazon EC2, Amazon S3, Amazon CloudWatch, Amazon SNS, Amazon DynamoDB and LLMOps Other Requirements: 1. Computer Science or related Engineering degree preferred 2. Experience with containerization technologies like Docker 3. Familiarity with AI model serving platforms and ML workflows About Company: Smartify is a marketplace for automation companies and also India's leading home automation store. We are trying to reduce the knowledge-execution gap and encourage early-adopters in the IoT space to launch their products and get to the mainstream market.

Posted 1 month ago

Apply

1 - 6 years

7 - 14 Lacs

Hyderabad

Work from Office

Naukri logo

Position - AI Engineer As an AI Engineer, you will design, implement, and optimize machine learning models and AI systems to solve complex problems. You will work closely with cross-functional teams to integrate AI solutions into our products and services, ensuring scalability and efficiency. Key Responsibilities: Application Development: Design and develop AI-powered applications using state-of-the-art LLM models and generative AI techniques. Implement scalable solutions that integrate LLM-powered tools into existing workflows or standalone products. Model Optimization: Fine-tune pre-trained LLM models to meet specific application requirements. Optimize model performance for real-time and high-throughput environments. LLMOps Implementation: Develop and maintain pipelines for model deployment, monitoring, and retraining. Set up robust systems for model performance monitoring and diagnostics. Ensure reliable operations through analytics and insights into model behavior. Vector Databases and Data Management: Utilize vector databases for efficient storage and retrieval of embeddings. Integrate databases with LLM applications to enhance query and recommendation systems. Collaboration and Innovation: Work closely with cross-functional teams, including product managers, data scientists, and software engineers. Stay up-to-date with advancements in generative AI and LLM technologies to drive innovation. Skills and Experience 3+ years of experience in AI/ML development, with a focus on generative AI and LLMs. Proficiency in programming languages such as Python and frameworks like PyTorch or TensorFlow. Hands-on experience in fine-tuning and deploying LLM models (e.g., GPT, BERT, etc.). Familiarity with LLMOps practices, including pipeline automation, monitoring, and analytics. Experience with vector databases (e.g., Pinecone, Weaviate, or similar). Strong knowledge of natural language processing (NLP) and machine learning principles. You should certainly apply if: Understanding of MLOps principles and cloud platforms (AWS, GCP, Azure). Familiarity with prompt engineering and reinforcement learning from human feedback (RLHF). Experience in building real-time applications powered by generative AI. Knowledge of distributed systems and scalable architectures.

Posted 1 month ago

Apply

1.0 years

3 - 5 Lacs

IN

Remote

Internshala logo

About the job: We empower the people who build the world. Taiy .AI is the world's largest infrastructure construction data-mesh technology and the first AI platform for the global infrastructure construction industry. Our clients include some of the largest construction firms, suppliers, and the government. About The Team: We are looking for a Python Engineer to help support and lead our data engineering ops. Key Responsibilities: 1. Developing and executing processes for monitoring data sanity, checking for data availability and reliability. 2. Understanding the business drivers and building insights through data. 3. Partner with stakeholders at all levels to establish current and ongoing data support and reporting needs. 4. Ensure continuous data accuracy and recognize data discrepancies in systems that require immediate attention/escalation. 5. Become an expert in the company's data warehouse and other data storage tools, understanding the definition, context, and proper use of all attributes and metrics. 6. Creating dashboards based on business requirements. 7. Distributed systems, Scala, cloud, Caching, CI/CD (Continuous integration and deployment), Distributed logging, Data pipeline, Recommendation Engine, Data at Rest Encryption What To Bring: 1. Graduate/Post Graduate degree in Computer Science or Engineering. 2. 1-3 years of hands-on experience with AWS Open Search v1.0 or Elastic Search 7.9 3. 3+ years of work experience on Scala 4. Must be able to drive, design, code, review the work, and assist the teams 5. Good problem-solving skills 6. Good oral and written communication in English 7. Should be open to/have experience of working in a fast-paced delivery environment 8. Strong understanding of object-oriented design, data structures, algorithms, profiling, and optimization. 9. Good to have experience on Elasticsearch and Spark-Elasticsearch 10. Knowledge of Garbage Collection and experience in GC tuning. 11. Knowledge of algorithms like sorting, heap/stack, queue, search, etc. 12. Experience with Git and build tools like Gradle/Maven/SBT 13. Should be able to write complex queries independently. 14. Strong knowledge of programming languages like Python, Scala, etc. 15. Ability to work independently and take ownership of things. 16. An analytical mindset and strong attention to detail. 17. Good verbal & written communication skills for coordinating across teams. Who can apply: Only those candidates can apply who: have minimum 1 years of experience are Computer Science Engineering students Salary: ₹ 3,00,000 - 5,00,000 /year Experience: 1 year(s) Deadline: 2025-06-05 23:59:59 Other perks: 5 days a week Skills required: Python, Selenium, Machine Learning, REST API, Data Extraction, Data Engineering and LLMOps About Company: Taiyo is a Silicon Valley startup that aggregates, predicts, and visualizes the world's data so customers don't have to. We are a globally-distributed team with a focus on the infrastructure vertical. The Taiyo team was founded by an interdisciplinary group of experts from Stanford University's AI Institute, World Bank, International Monetary Fund, and UC Berkeley.

Posted 1 month ago

Apply

7 - 12 years

20 - 35 Lacs

Bengaluru

Work from Office

Naukri logo

Location: Bangalore / Hybrid Department: Data & AI Company: Resolve Tech Solutions / Juno Labs About Juno Labs: Juno Labs is at the forefront of AI-driven cloud solutions, helping businesses unlock the power of data with scalable, intelligent, and high-performance architectures. We specialize in building next-gen data platforms, leveraging cloud technologies, AI/ML, vector databases, and advanced frameworks to drive real-time insights and intelligent decision-making. Job Description: We are looking for an experienced MLOps Engineer to join our Data & AI team. This role will focus on building, deploying, and optimizing end-to-end machine learning systems with an emphasis on LLMOps (Large Language Models operationalization). The ideal candidate will have strong expertise in MLOps , LLMOps , and DevOps , with hands-on experience managing and deploying large-scale models, particularly LLMs , in both cloud and on-premise environments. The role involves not only building robust MLOps pipelines but also self-hosting models, optimizing GPU usage, and performing quantization to reduce the cost of deployment. Key Responsibilities: Design and implement scalable MLOps pipelines to deploy, monitor, and manage machine learning models, with a particular focus on LLMOps . Integrate, fine-tune, and optimize Hugging Face models (e.g., Transformers , BART , GPT-2/3 ) for diverse NLP tasks such as text generation , text classification , and NER , and deploy them for production-scale systems. Use LangChain to build sophisticated LLM-driven applications , enabling seamless model workflows for NLP and decision-making tasks. Optimize and manage LLMOps pipelines for large-scale models using technologies such as OpenAI API , Amazon Bedrock , DeepSpeed , and Hugging Face Hub . Develop and scale self-hosted LLM solutions (e.g., fine-tuning and serving models on-premises or in a hybrid cloud environment) to meet performance, reliability, and cost-effectiveness goals. Leverage cloud-native tools such as Amazon SageMaker , Vertex AI , GCP , AWS for scaling large language models, and ensure their optimization in distributed cloud environments. Use GPU-based optimization for large-scale model training and deployment, ensuring high performance and efficient resource allocation in the cloud or on- premises environments. Deploy models via containerized solutions using Docker , Kubernetes , and Helm , allowing for seamless scaling and management in both cloud and on- premise infrastructures. Implement model quantization and pruning techniques to reduce the resource footprint of deployed models while maintaining high performance. Monitor model performance in production using Prometheus , Grafana , ELK Stack , and other observability tools to track metrics such as inference latency, accuracy, and throughput. Automate the end-to-end workflow of model development and deployment via CI/CD pipelines with tools like GitLab CI , Jenkins , and CircleCI . Integrate vector databases (e.g., Pinecone , FAISS , Milvus ) for efficient storage, retrieval, and querying of model-generated embeddings. Stay up to date with the latest advancements in MLOps , LLMOps , and machine learning technologies, ensuring the adoption of best practices in model development, deployment, and optimization. Required Skills & Qualifications: Bachelors or Masters degree in Computer Science, Engineering, or a related field. 5+ years of experience in MLOps , LLMOps , DevOps , or related roles, with a focus on deploying and managing machine learning models in production environments. Experience with cloud platforms such as AWS , GCP , Azure , and services like Amazon SageMaker , Vertex AI , TensorFlow Serving , DeepSpeed , and Amazon Bedrock . Expertise in Hugging Face models and the Transformers library, including model fine-tuning , deployment , and optimizing NLP models for large-scale production. Experience with LangChain for building and deploying LLM-based applications that handle dynamic and real-time tasks. Strong experience with self-hosting LLMs in cloud or on-premises environments using GPU-based infrastructure for training and inference (e.g., NVIDIA GPUs , CUDA ). Expertise in GPU utilization and optimization for large-scale model training, inference, and cost-effective deployment. Hands-on experience in model quantization techniques to reduce the memory footprint and inference time, such as TensorFlow Lite , ONNX , or DeepSpeed . Familiarity with distributed ML frameworks like Kubeflow , Ray , Dask , MLflow , for managing end-to-end ML workflows and large-scale model training and evaluation. Proficiency with containerization and orchestration tools such as Kubernetes , Docker , Helm , and Terraform for infrastructure automation. Knowledge of vector databases like Pinecone , Milvus , or FAISS to facilitate fast and scalable retrieval of model-generated embeddings. Expertise in setting up and managing CI/CD pipelines for model training, validation, testing, and deployment with tools like Jenkins , GitLab CI , and CircleCI . Strong programming skills in Python , Bash , and Shell scripting . Solid understanding of monitoring and logging tools such as Prometheus , Grafana , and ELK Stack to ensure high system performance, error detection, and model health tracking. Preferred Qualifications: Proven experience in deploying and managing large-scale LLMs like GPT-3 , BERT , T5 , and BLOOM in production environments using cloud-native solutions and on-premises hosting. Deep expertise in quantization , model compression , and pruning to optimize deployed models for lower latency and reduced resource consumption. Strong understanding of NLP tasks and deep learning concepts such as transformers, attention mechanisms, and pretrained model fine-tuning. Experience with Kedro for building reproducible ML pipelines with a focus on data engineering, workflow orchestration, and modularity. Familiarity with Apache Spark and Hadoop for handling big data processing needs, especially in real-time AI workloads . Familiarity with advanced data engineering pipelines and data lakes for the effective management of large datasets required for training LLMs. Why Join Us: Work with cutting-edge technologies in AI , MLOps , and LLMOps , including self- hosting and optimizing large-scale language models. Be part of an innovative, fast-growing team working on the future of AI-driven cloud solutions. Flexibility in work style with a hybrid work environment that promotes work-life balance. Competitive salary and benefits package, with opportunities for personal and professional growth.

Posted 3 months ago

Apply

10 - 15 years

32 - 37 Lacs

Bengaluru

Work from Office

Naukri logo

Responsibilities: Customer & Architecture Interact with customers to define problem statements and technical requirements Design and architect AI/ML cloud solutions using AWS/GCP Create system architecture diagrams and establish best practices for ML systems Define data architecture and model serving strategies Technical Leadership & Development Guide team in building Python-based APIs and AI/ML models Design and implement ML pipeline automation Establish model deployment and monitoring architectures Review technical designs and approve implementation approaches Lead architectural decision-making and technical direction Project & Team Management Overall responsibility for project/program delivery Manage and mentor AI/ML engineers Lead sprint planning and technical grooming sessions Coordinate with cross-functional teams Balance resource allocation and technical debt Qualifications Must have at least 10 years of IT experience with 5+ years experience in designing, developing, deploying, and operationalizing AI/ML solutions. Experience in GenAI and Conversational AI LLMs, frameworks like Lang chain, Llama Index etc, Vector DBs like Pinecone, Elasticsearch, etc. Deep understanding and experience with popular Deep Learning and Machine Learning techniques and algorithms such as LLMs, Neural Networks, Convolutional Neural Networks (CNN), Recurrent Neural Networks (RNN), LSTM, Transformers, Time Series Forecasting, Segmentation, etc. Experience in Data Science and Machine Learning, preferably with experience in NLP, Generative AI, LLMs, MLOps/LLMOps, Optimization techniques, and AI solution Architecture. Experience with ML Pipelines both On-Prem and Cloud (like SageMaker) Expert in Python as well as open-source frameworks such as PyTorch, TensorFlow, NumPy, etc. Experience in designing cloud-native solutions using AI and other services from Azure, GCP, AWS, etc. Ability to lead Data Engineers and ML engineers to provide guidance on technical architecture and best practices.

Posted 3 months ago

Apply

6 - 8 years

8 - 12 Lacs

Chennai, Hyderabad

Work from Office

Naukri logo

What youll be doing... In this role as a Sr Engr in the Tech Strategy & Planning team for TSGI-Cloud and Enterprise Architecture team, you'll be managing multiple programs designed to help GTS move towards its strategic objectives. You will be a key contributor in planning and building Data, Platforms and Emerging Technology North star, and ensuring the smooth and successful execution of initiatives by coordinating with cross-functional teams. Your expertise will help us solve complex problems and find unique solutions to optimize the technology landscape of our organization. Your responsibilities include but are not limited to: Technology Road-mapping: Develop and maintain a technology roadmap that aligns with the organization's strategic goals and evolving industry trends. Strategic Planning: Collaborate with leadership to identify opportunities where technology can be a competitive advantage and contribute to the development of the company's overall strategic plan. Innovation and Research: Stay updated on emerging technologies, assess their relevance to the business, and propose innovative solutions. Vendor and Partner Management: Evaluate and manage relationships with technology vendors and partners to ensure they support the organization's strategic objectives. Evaluate the Gen AI tools in partnership with AI&D team Develop scalable prototypes of LLM and NLP Modules and systems which are critical to the companys product lines. Apply state-of-the-art LLM techniques to understand a large amount of unstructured data and translate them to a meaningful and structured data Design, develop and evaluate predictive LLM models that are on par with industry standards & define metrics that measure success and customer value delivery Work closely with process experts to analyze and design solutions based on business requirements. What were looking for... Youre analytical and great at quickly grasping challenging concepts. As a strong written and verbal communicator, you deliver complex messages vividly to technical and business audiences alike. Youre no stranger to a fast-paced environment and tight deadlines, and you adapt to changing priorities and balance multiple projects with ease. You take pride in your work and get a lot of satisfaction from meeting and exceeding the expectations of your customers Youll need to have: Bachelors degree or four or more years of experience. Minimum 6 years of experience in one or more of Data Science, LLM, Gen AI Established experience delivering information management solutions to large numbers of end users Experience in building LLM solutions to business problems across support, sales, digital, chat, voice etc., Familiarity with Gen AI Models (OpenAI, Gemini AI, Vertex.AI, Mistral, LlaMa, etc.,) and finetune based on domain specific needs Experience in text processing, Vector databases and embedding models Experience in NLP, Transformers and Neural Networks Hands-on experience in using Langchain and good exposure to LLMOps Strong independent and creative research skills necessary to keep up with the latest trends in advanced analytics and ML Research, recommend and implement best practices of LLM and NLP systems that can scale Identify new process opportunities, while quickly assessing feasibility Excellent written and verbal communication skills able to effectively communicate technical details to a non-technical audience as well as produce clear and concise written documentation Even better if you have one or more of the following: Familiarity with Graph DB Concepts Familiarity with one or more Data Platforms - CloudEra, Snowflake, DataBricks etc. Experience in one or more Big data and ETL technologies - Informatica, Talend, Teradata, Hadoop etc. Experience in one or more BI platforms - Tableau, Looker, Thoughtspot etc. Masters degree from an accredited college or university preferred.

Posted 3 months ago

Apply

3 - 8 years

12 - 22 Lacs

Chennai, Bengaluru, Hyderabad

Hybrid

Naukri logo

Project Description: Grid Dynamics aims building enterprise generative AI framework to deliver innovative, scalable and efficient AI-driven solutions across business functions.Due to constant scaling of digital capabilities the platform requires enhancements to incorporate cutting-edge generative AI features and meet emerging business demands. Platform should onboard brand new capabilities like Similarity Search (image,video and voice);Ontology and entity managment;voice and file mgmt [text to speech & vice-versa, metadata tagging, multi-media file support];Advanced RAG ; Multi-Modal capabilities Responsibilities: As an LLMOps Engineer, you will be responsible for providing expertise on overseeing the complete lifecycle management of large language models (LLM). This includes the development of strategies for deployment, continuous integration and delivery (CI/CD) processes, performance tuning, and ensuring high availability of our LLM services. You will collaborate closely with data scientists, AI/ML engineers, and IT teams to define and align LLM operations with business goals, ensuring a seamless and efficient operating model. In this role, you will: Define and disseminate LLMOps best practices. Evaluate and compare different LLMOps tools to incorporate the best practices. Stay updated on industry trends and advancements in LLM technologies and operational methodologies. Participate in architecture design/validation sessions for the Generative AI use cases with entities. Contribute to the development and expansion of GenAI use cases, including standard processes, framework, templates, libraries, and best practices around GenAI. Design, implement, and oversee the infrastructure required for the efficient operation of large language models in collaboration with client entities. Provide expertise and guidance to client entities in the development and scaling of GenAI use cases, including standard processes, framework, templates, libraries, and best practices around GenAI Serve as the expert and representative on LLMops Practices, including: (1) Developing and maintaining CI/CD pipelines for LLM deployment and updates. (2) Monitoring LLM performance, identifying and resolving bottlenecks, and implementing optimizations. (3) Ensuring the security of LLM operations through comprehensive risk assessments and the implementation of robust security measures. Collaborate with data and IT teams to facilitate data collection, preparation, and model training processes. Practical experience with training, tuning, utilizing LLMs/SLMs. Strong experience with GenAI/LLM frameworks and techniques, like guardrails, Langchain, etc. Knowledge of LLM security and observability principles. Experience of using Azure cloud services for ML Experience of using Azure cloud services for ML Min requirements: Programming languages: Python Public Cloud: Azure Frameworks: K8s, Terraform, Arize or any other ML/LLM observability tool Experience: Experience with public services like Open AI, Anthropic and similar, experience deploying open source LLMs will be a plus Tools: LangSmith/LangChain,guardrails Would be a plus: Knowledge of LLMOps best practices. Experience with monitoring/logging for production models (e.g. Prometheus, Grafana, ELK stack) We offer: Opportunity to work on bleeding-edge projects Work with a highly motivated and dedicated team Competitive salary Flexible schedule Benefits package - medical insurance, sports Corporate social events Professional development opportunities Well-equipped office

Posted 3 months ago

Apply

3 - 8 years

15 - 25 Lacs

Chennai, Bengaluru, Hyderabad

Hybrid

Naukri logo

Project Description: Grid Dynamics wants to build a centralized, observable and secure platform for their ML, Computer Vision, LLM and SLM models. Grid Dynamics wants to onboard a vast number of AI agents, able to cover multiple required skills, ensuring a certain level of control and security in regards to their usage and availability. The observable platform must be vendor-agnostic, easy to extend to multiple type of AI applications and flexible in terms of technologies, frameworks and data types. This project is focused on establishing a centralized LLMOps capability where every ML, CV, AI-enabled application is monitored, observed, secured and provides logs of every activity. The solution consists of key building blocks such monitor every step in a RAG, Multimodal RAG or Agentic Platform, track performances and provide curated datasets for potential fine-tuning. Alignment with business scenarios, PepVigil provides also certain guardrails that allow or block interactions user-to-agent, agent-to-agent or agent-to-user. Also, Guardrails will enable predefined workflows, aimed to give more control over the series of LLM chains. Details on Tech Stack Job Qualifications and Skill Sets Advanced degree in Data Science, Computer Science, Statistics, or a related field Setting up Agent Mesh (LangSmith) Setting up Agent communication protocols (JSON/XML etc) Setting up message queues, CI/CD pipelines (Azure Queue Storage, Azure DevOps) Setting up integrations Langgrah, LangFuse Knowledge on Observability tool Arize-Phoenix tools Managing Agent Registry, Integrating with AgentAuth framework like Composio Setting up AgentCompute (Sandpack, E2BDev, Assistant APIs) Integration with IAM (Azure IAM, OKTA) Performing/Configuring Dynamic Orchestration and agent permissions Tech Stack Required: ML MLOPs Agent (Agent / Agent Mesh) LangFuse, LanChain, LangGraph Deployments (Docker, Jenkins, Kubernetes) Cloud Platforms: Azure/AWS/GCP We offer: Opportunity to work on bleeding-edge projects Work with a highly motivated and dedicated team Competitive salary Flexible schedule Benefits package - medical insurance, sports Corporate social events Professional development opportunities Well-equipped office

Posted 3 months ago

Apply

5 - 10 years

16 - 25 Lacs

Bengaluru

Work from Office

Naukri logo

Job Description Role & Responsibilities: Develop and deploy Generative AI solutions using LLMs and related technologies. Fine-tune large language models (LLM fine-tuning) to improve performance. Work on Retrieval-Augmented Generation (RAG) methodologies to optimize AI outputs. Implement Prompt Engineering techniques for effective model interaction. Utilize Python, TensorFlow, and PyTorch for AI model development and optimization. Work with cloud platforms like Azure, AWS, or GCP to deploy and scale AI models. Collaborate with cross-functional teams to integrate AI solutions into business applications. Maintain AI/ML pipelines and workflows using LLMOps best practices. Preferred Candidate Profile: 5 to 8 years of experience in AI, ML, or Data Science roles. 2 to 3 years of experience specifically in GenAI solution development, LLM fine-tuning, RAG, or LLMOps. Strong background in Natural Language Processing (NLP) and Transformer models. Hands-on experience with cloud-based AI solutions (Azure, AWS, GCP). Excellent problem-solving skills and ability to work in a fast-paced environment. Perks and Benefits: Competitive salary and performance-based incentives. Work with cutting-edge AI technologies and industry leaders. Opportunities for career growth and learning. Health insurance and other standard benefits.

Posted 3 months ago

Apply

5 - 8 years

25 Lacs

Gurgaon

Remote

Naukri logo

Job Summary : We are hiring a DevOps/ MLOps/ LLMOps Specialist to optimize our AI/ML infrastructure, streamline deployments, and enhance collaboration across teams. The ideal candidate will manage Kubernetes clusters, implement Infrastructure as Code (IaC), and develop automated ML workflows for efficient model deployment and monitoring. What you'll do : - Manage and optimize Kubernetes clusters with GPU integration. - Implement IaC (Terraform, Ansible) for scalable infrastructure. - Design CI/CD pipelines for ML model deployment and monitoring. - Automate cloud infrastructure provisioning and resource optimization. - Collaborate with data scientists, developers, and infrastructure teams. Who you are : - 5-8 years of experience in DevOps, MLOps, or LLMOps. - Expertise in Kubernetes, Terraform, Ansible, and cloud platforms (AWS/ GCP/ Azure). - Strong scripting skills (Python, Bash) for automation. - Hands-on experience with CI/CD, GitOps, and ML model deployment. - Qualification: B.Tech (CS & IT) / M.Tech (CS & IT) / MCA.

Posted 3 months ago

Apply
cta

Start Your Job Search Today

Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.

Job Application AI Bot

Job Application AI Bot

Apply to 20+ Portals in one click

Download Now

Download the Mobile App

Instantly access job listings, apply easily, and track applications.

Featured Companies