Jobs
Interviews

44 Llmops Jobs - Page 2

Setup a job Alert
JobPe aggregates results for easy application access, but you actually apply on the job portal directly.

10.0 - 12.0 years

0 - 33 Lacs

Mumbai, Maharashtra, India

On-site

Job Description Summary role description: Hiring for a Solution Architect for an InsurTech platform provider, Life and Health Insurance. Company description: Our client is a VC-funded InsurTech platform company, providing software platforms for Life Insurance and Health Insurance companies across the globe. Leveraging their domain expertise, regulatory knowledge and technology experience, they architect innovative products and disrupt the Insurance value chain from Customer Acquisition to Engagement. Their products serve customers across the APAC region. Role details: Title / Designation : Solutions Architect Location: Pune/Mumbai Work Mode: Work from office Role & responsibilities: Define and evolve AI/ML architecture roadmap for FWA, IDP, and Agentic AI frameworks. Lead technical presentations and solution design sessions with customers. Design scalable architectures for multi-agent systems and autonomous decision-making. Drive innovation by evaluating emerging AI/ML technologies, especially AI agents. Architect cloud-native platforms supporting the complete AI/ML lifecycle. Provide technical leadership across product development and customer implementation. Collaborate with data scientists, engineers, and business stakeholders. Stay at the forefront of AI/ML innovations, particularly autonomous agents and LLMs. Establish and enforce technical standards and architectural guidelines. Candidate requirements: 10+ years in software architecture/system design in insurance domain, with 5+ years in AI/ML systems/platforms. Proven experience delivering large-scale AI/ML solutions, preferably with autonomous agents. Experience with cloud-native architectures (AWS, Azure, GCP), containerization (Docker, Kubernetes), and microservices. Deep expertise in AI/ML system architecture (model serving, MLOps/LLMOps pipelines, distributed computing). Strong understanding of Agentic AI, multi-agent systems, and LLMs (including LoRA, PEFT fine-tuning). Bachelor's or Master's in CS, SE, Data Science, or related technical field. Exceptional technical leadership and communication skills. Selection process: Interview with Senior Solution Architect Interview with CTO HR Discussion Check Your Resume for Match Upload your resume and our tool will compare it to the requirements for this job like recruiters do.

Posted 1 month ago

Apply

8.0 - 13.0 years

15 - 25 Lacs

Noida, Hyderabad, Bengaluru

Work from Office

Job Description : As an LLMOps Engineer, you will play a crucial role in the deployment, maintenance, and optimization of large language models (LLMs). Your responsibilities will span the entire lifecycle of LLMs, from initial deployment to ongoing operations, ensuring optimal performance, scalability, and reliability. Key Responsibilities : LLM Deployment and Integration - Deploy and integrate large language models into production environments, ensuring seamless integration with existing systems and applications. Infrastructure Planning and Scaling - Collaborate with cross-functional teams to plan and design the infrastructure required for LLM deployment. Implement scalable solutions to accommodate growing data volumes and user loads. Automation of Deployment Processes - Develop and maintain automation scripts and tools for efficient deployment, scaling, and versioning of LLMs. Streamline deployment processes to minimize downtime. Continuous Monitoring and Alerting - Implement monitoring systems to track LLM performance metrics. Set up alerts for potential issues and respond promptly to ensure uninterrupted service. Performance Monitoring and Optimization - Monitor the performance of LLMs in real-time, conduct regular assessments, and implement optimizations to enhance efficiency and responsiveness. Fault Tolerance and Disaster Recovery - Design and implement fault-tolerant systems for LLMs, incorporating strategies such as redundancy, sharding, and replication. Develop and maintain disaster recovery plans. Security Measures Implementation - Implement robust security measures to safeguard LLMs and associated data. Ensure compliance with data security regulations and industry standards. Collaboration with NLP Engineers and Data Scientists - Collaborate with NLP (Natural Language Processing) engineers and data scientists to understand model requirements and implement necessary infrastructure adjustments. Skills & Tools Infrastructure as Code (IaC) - Experience with IaC tools such as Terraform or Ansible for automating infrastructure provisioning. Containerization and Orchestration - Proficiency in containerization technologies (e.g., Docker) and orchestration tools (e.g., Kubernetes) for managing LLM deployments. Cloud Platforms - Familiarity with cloud platforms such as AWS (Bedrock), Azure, or GCP, and experience in deploying and managing applications in a cloud environment. Monitoring and Logging Tools - Knowledge of monitoring tools (e.g., Prometheus, Grafana) and logging systems (e.g., ELK stack) for real-time performance monitoring and analysis. Security Measures - Understanding of security guardrails using tools like LLM Guard and familiarity of how to mask / redact / obfuscate sensitive data, protect the input and output of toxic and harmful content to / from LLMs and understand the performance implications of the same. Scripting and Automation - Proficient in scripting languages such as Python, Shell, or similar, and experience in automating deployment and maintenance processes.

Posted 1 month ago

Apply

2.0 - 7.0 years

30 - 40 Lacs

Hyderabad, Bengaluru

Hybrid

Role: AIML Engineer Experience: 2 to 8 Yrs Notice period: 0- 10 Days Job Location: Hyderabad and Bangalore. (Hybrid) Job Description: Strong proficiency in Python for AI/ML development. Hands-on experience with OpenAI, GPT models, and LangChain or LlamaIndex. Deep understanding of Retrieval-Augmented Generation (RAG) concepts and implementations. Familiarity with vector databases like FAISS, Pinecone, or Weaviate. Good knowledge of data engineering principles (ETL, data modeling, batch/streaming pipelines). Experience with cloud platforms (Azure, AWS, or GCP) for deploying AI/ML solutions. Proficient in tools like Pandas, NumPy, Scikit-learn, and MLflow. Exposure to LLMOps, model versioning, and monitoring is a plus.

Posted 1 month ago

Apply

1.0 years

6 Lacs

IN

Remote

About the job: As a Full Stack GenAI Developer at MeetMinutes, you will be responsible for creating cutting-edge AI solutions using Python, Generative AI Development, LangChain, LLM evaluation, LLMOps, JavaScript, React, Amazon Web Services (AWS), Google Cloud Platforms (GCP), Docker, Machine Learning, Natural Language Processing (NLP), PostgreSQL, REST API, FastAPI, GitHub, System Design, and Prompt Engineering. Key responsibilities: 1. Developing and implementing AI algorithms and models to enhance the functionality of our platform. 2. Integrating AI technologies and features into our existing systems to improve user experience. 3. Collaborating with the engineering team to optimize system performance and scalability. 4. Building and maintaining RESTful APIs for seamless communication between different components. 5. Monitoring and troubleshooting any issues related to AI functionality and recommending solutions. 6. Contributing to the overall architecture and design of our AI-driven products. 7. Staying updated on the latest trends and advancements in AI and actively participating in knowledge sharing within the team. If you are passionate about using AI to revolutionize meeting productivity and collaboration, and have a strong background in full-stack development and AI technologies, we'd love to have you join our team! Who can apply: Only those candidates can apply who: have minimum 1 years of experience are Computer Science Engineering students Salary: ₹ 6,50,000 /year Experience: 1 year(s) Deadline: 2025-07-13 23:59:59 Other perks: 5 days a week Skills required: JavaScript, Python, Machine Learning, PostgreSQL, Docker, GitHub, React, REST API, Amazon Web Services (AWS), Natural Language Processing (NLP), Google Cloud Platforms (GCP), FastAPI, Generative AI Development, LangChain, Prompt Engineering, System Design, LLMOps and LLM evaluation About Company: A platform for the future of AI meeting productivity. It is a tool for professionals having conversations in mixed Indian languages and need workflow automations. It has support for Google Meet, MS Teams, and many more meeting platforms. It is a startup recognized by the Ministry of IT and won several accolades working with businesses, IT firms, agencies, listed companies, and SMBs.

Posted 1 month ago

Apply

1.0 - 4.0 years

3 - 14 Lacs

Bengaluru / Bangalore, Karnataka, India

On-site

Perform performance evaluations of the LLM models, ImplementLLMOpsprocesses to run the end-to-end lifecycle of LLMs Deploy,monitor, andmaintainMachine Learning models and build AI Products in production environments, ensuringoptimalperformance and reliability. Ensure high code quality, performance, and reliability through rigorous testing, code reviews, and adherence to software development best practices. Drive innovation by researching and incorporatingstate-of-the-artmachine learning techniques, tools, and frameworks into the platform. Effective communication, listening, interpersonal, influencing, and alignment driving skills; able to convey important messages in a clear and compelling manner Mentor team members,providetechnical guidance, and foster a culture of collaboration, innovation, and continuous learning. What do you need to bring Qualifications Masters / bachelor s in computer science, Computer engineering, Machine Learning, Data Mining, Information Systems, or related disciplines, with technicalexpertisein one or more of the above-mentioned areas or equivalent practical experience. Strong background in deep learning techniques, particularly in NLP and Vision Expertisein applying LLMs, prompt design, and fine-tuning methods Strongproficiencyin machine learning concepts, algorithms, and techniques, with hands-on experience in developing and deploying machine learning models. Expert in multiple Programming/scripting languages,i.e.,Python, Java, Scala, SQL, NoSQL (like HBase, Redis, Aerospike) Good understanding of distributed systems, data streaming, complex event Processing,NoSQLsolutions for creating and managing data integration pipelines for batch and Real Time data needs. Expertisein machine learning libraries/frameworks such as TensorFlow,PyTorch, scikit-learn,etc. Experience with cloud platforms (e.g., AWS, Azure, GCP) and containerization technologies (e.g., Docker, Kubernetes). Experience in Azure is a plus Stay up to date with the latest advancements in AI/ML technology and industry trends andleveragethis knowledge to enhance the platforms capabilities Strong communication, listening, interpersonal, influencing, and alignment driving skills; able to convey important messages in a clear and compelling manner Expertisein Big Data technologies such as Hadoop, Spark, HBase, Kafka. Preferred Prior experience in Content Understanding, enrichment, entity resolution or knowledge graph Experience developing Gen AI applications/services for sophisticated business use cases andlarge amountsof unstructured data. Strong background in MLOps and experimentation frameworks

Posted 1 month ago

Apply

9.0 - 14.0 years

35 - 50 Lacs

Hyderabad, Pune, Bengaluru

Work from Office

Role - Senior Data Scientist / Senior Gen AI Engineer Exp Range - 8 to 18 yrs Position - Permanent Fulltime Company - Data Analytics & AIML MNC Location - Hyderabad, Pune, Bangalore (Relocation accepted) About the Role: We are seeking a Software Engineer with expertise in Generative AI and Microsoft technologies to design, develop, and deploy AI-powered solutions using the Microsoft ecosystem. You will work with cross-functional teams to build scalable applications leveraging generative AI models and Azure services. Skills Required: Experience with Large Language Models (LLMs) like GPT, LLaMA, Claude, etc. Proficiency in Python for building and fine-tuning AI/ML models Familiarity with LangChain , LLMOps , or RAG (Retrieval-Augmented Generation) pipelines Experience with Vector Databases (e.g. FAISS, Pinecone, Weaviate) Knowledge of Prompt Engineering and model evaluation techniques Exposure to cloud platforms (Azure, AWS or GCP) for deploying GenAI solutions Preferred Skills: Experience with Azure OpenAI , Databricks or Microsoft Fabric Hands-on with Hugging Face Transformers , OpenAI APIs or custom model training

Posted 1 month ago

Apply

8.0 - 10.0 years

6 - 10 Lacs

Noida

Work from Office

Position Summary LLMOps(Large language model operations) Engineer will play a pivotal role in building and maintaining the infrastructure and pipelines for our cutting-edge Generative AI applications, establishing efficient and scalable systems for LLM research, evaluation, training, and fine-tuning. Engineer will be responsible for managing and optimizing large language models (LLMs) across various platforms This position is uniquely tailored for those who excel in crafting pipelines, cloud infrastructure, environments, and workflows. Your expertise in automating and streamlining the ML lifecycle will be instrumental in ensuring the efficiency, scalability, and reliability of our Generative AI models and associated platform. LLMOps engineers expertise will ensure the smooth deployment, maintenance, and performance of these AI platforms and powerful large language models. You will follow Site Reliability Engineering & MLOps principles and will be encouraged to contribute your own best practices and ideas to our ways of working. Reporting to the Head of Cloud Native operations, you will be an experienced thought leader, and comfortable engaging senior managers and technologists. You will engage with clients, display technical leadership, and guide the creation of efficient and complex products/solutions. Key Responsibilities Technical & Architectural Leadership Contribute to the technical delivery of projects, ensuring a high quality of work that adheres to best practices, brings innovative approaches and meets client expectations. Project types include following (but not limited to): Solution architecture, Proof of concepts (PoCs), MVP, design, develop, and implementation of ML/LLM pipelines for generative AI models, data management & preparation for fine tuning, training, deployment, and monitoring. Automate ML tasks across the model lifecycle. Contribute to HCL thought leadership across the Cloud Native domain with an expert understanding of advanced AI solutions using Large Language Models (LLM) & Natural Language Processing (NLP) techniques and partner technologies. Collaborate with cross-functional teams to integrate LLM and NLP technologies into existing systems. Ensure the highest levels of governance and compliance are maintained in all ML and LLM operations. Stay abreast of the latest developments in ML and LLM technologies and methodologies, integrating these innovations to enhance operational efficiency and model effectiveness. Collaborate with global peers from partner ecosystems on joint technical projects. This partner ecosystem includes Google, Microsoft, Nvidia, AWS, IBM, Red Hat, Intel, Cisco, and Dell VMware etc. Service Delivery Provide a technical hands-on contribution. Create scalable infra to support enterprise loads (distributed GPU compute, foundation models, orchestrating across multiple cloud vendors, etc.) Ensuring the reliable and efficient platform operations. Apply data science, machine learning, deep learning, and natural language processing methods to analyse, process, and improve the models data and performance. Understanding of Explainability & Biased Detection concepts. Create and optimize prompts and queries for retrieval augmented generation and prompt engineering techniques to enhance the models capabilities and user experience w.r.t Operations & associated platforms. Client-facing influence and guidance, engaging in consultative client discussions and performing a Trusted Advisor role. Provide effective support to HCL Sales and Delivery teams. Support sales pursuits and enable HCL revenue growth. Define the modernization strategy for client platform and associated IT practices, create solution architecture and provide oversight of the client journey. Innovation & Initiative Always maintain hands-on technical credibility, keep in front of the industry, and be prepared to show and lead the way forward to others. Engage in technical innovation and support HCLs position as an industry leader. Actively contribute to HCL sponsorship of leading industry bodies such as the CNCF and Linux Foundation. Contribute to thought leadership by writing Whitepapers, blogs, and speaking at industry events. Be a trusted, knowledgeable internal innovator driving success across our global workforce. Client Relationships Advise on best practices related to platform & Operations engineering and cloud native operations, run client briefings and workshops, and engage technical leaders in a strategic dialogue. Develop and maintain strong relationships with client stakeholders. Perform a Trusted Advisor role. Contribute to technical projects with a strong focus on technical excellence and on-time delivery. Mandatory Skills & Experience Expertise in designing and optimizing machine-learning operations, with a preference for LLMOps. Proficient in Data Science, Machine Learning, Python, SQL, Linux/Unix shell scripting. Experience on Large Language Models and Natural Language Processing (NLP), and experience with researching, training, and fine-tuning LLMs. Contribute towards fine-tune Transformer models for optimal performance in NLP tasks, if required. Implement and maintain automated testing and deployment processes for machine learning models w.r.t LLMOps. Implement version control, CI/CD pipelines, and containerization techniques to streamline ML and LLM workflows. Develop and maintain robust monitoring and alerting systems for generative AI models ensuring proactive identification and resolution of issues. Research or engineering experience in deep learning with one or more of the following: generative models, segmentation, object detection, classification, model optimisations. Experience implementing RAG frameworks as part of available-ready products. Experience in setting up the infrastructure for the latest technology such as Kubernetes, Serverless, Containers, Microservices etc. Experience in scripting programming to automate deployments and testing, worked on tools like Terraform and Ansible. Scripting languages like Python, bash, YAML etc. Experience on CI/CD opensource and enterprise tool sets such as Argo CD, Jenkins. Experience with the GitHub/DevOps Lifecycle Experience in at least one of the Observability solutions (Prometheus, EFK stacks, ELK stacks, Grafana, Dynatrace, AppDynamics) Experience in at-least one of the clouds for example - Azure/AWS/GCP Significant experience on microservices-based, container-based or similar modern approaches of applications and workloads. You have exemplary verbal and written communication skills (English). Able to interact and influence at the highest level, you will be a confident presenter and speaker, able to command the respect of your audience. Desired Skills & Experience Bachelor level technical degree or equivalent experience; Computer Science, Data Science, or Engineering background preferred; masters degree desired. Experience in LLMOps or related areas, such as DevOps, data engineering, or ML infrastructure. Hands-on experience in deploying and managing machine learning and large language model pipelines in cloud platforms (e.g., AWS, Azure) for ML workloads. Familiar with data science, machine learning, deep learning, and natural language processing concepts, tools, and libraries such as Python, TensorFlow, PyTorch, NLTK etc. Experience in using retrieval augmented generation and prompt engineering techniques to improve the models quality and diversity to improve operations efficiency. Proven experience in developing and fine-tuning Language Models (LLMs). Stay up-to-date with the latest advancements in Generative AI, conduct research, and explore innovative techniques to improve model quality and efficiency. The perfect candidate will already be working within a System Integrator, Consulting or Enterprise organisation with 8+ years of experience in a technical role within the Cloud domain. Deep understanding of core practices including SRE, Agile, Scrum, XP and Domain Driven Design. Familiarity with the CNCF open-source community. Enjoy working in a fast-paced environment using the latest technologies, love Labs dynamic and high-energy atmosphere, and want to build your career with an industry leader.

Posted 1 month ago

Apply

12.0 - 16.0 years

40 - 50 Lacs

Pune, Chennai, Bengaluru

Hybrid

AI Ops Senior Architect 12 -17 Years Work Location - Pune/ Bengaluru/Hyderabad/Chennai/ Gurugram Tredence is Data science, engineering, and analytics consulting company that partners with some of the leading global Retail, CPG, Industrial and Telecom companies. We deliver business impact by enabling last mile adoption of insights by uniting our strengths in business analytics, data science and data engineering. Headquartered in the San Francisco Bay Area, we partner with clients in US, Canada, and Europe. Bangalore is our largest Centre of Excellence with skilled analytics and technology teams serving our growing base of Fortune 500 clients. JOB DESCRIPTION At Tredence, you will lead the evolution of Industrializing AI ” solutions for our clients by implementing ML/LLM/GenAI & Agent Ops best practices. You will lead the Architecture , Design & development of large scale ML/LLMOps platforms for our clients. You’ll build and maintain tools for deployment, monitoring, and operations. You’ll be a trusted advisor to our clients in ML/GenAI/Agent Ops space & coach to the ML engineering practitioners to build effective solutions to Industrialize AI solutions THE IDEAL CANDIDATE WILL BE RESPONSIBLE FOR AI Ops Strategy, Innovation, Research and Technical Standards 1. Conduct research and experiment with emerging AI Ops technologies and trends. Create POV’s, POC’s & present Proof of Technology to use latest tools, Technologies & services from Hyper scalers focussed on ML, GenAI & Agent Ops 2. Define and propose new technical standards and best practices for the organization's AI Ops environment. 3. Lead the evaluation and adoption of innovative MLOps solutions to address critical business challenges. 4. Conduct meet ups, attend & present in Industry events, conferences, etc 5. Ideate & develop accelerators to strengthen service offerings of AI Ops practice Solution Design & Architectural Development 6. Lead Design & architecture of scalable model training & deployment pipelines for large-scale deployments 7. Architect & Design large scale ML & GenAI Ops platforms 8. Collaborate with Data science & GenAI practice to define and implement strategies of AI solutions for model explainability and interpretability 9. Mentor and guide senior architects in crafting cutting-edge AI Ops solutions 10. Lead architecture reviews and identify opportunities for significant optimizations and improvements. Documentation and Best Practices 11. Develop and maintain comprehensive documentation of AIOps architectures designs and best practices. 12. Lead the development and delivery of training materials and workshops on AIOps tools and techniques. 13. Actively participate in sharing knowledge and expertise with the MLOps team through internal presentations and code reviews. Qualifications and Skills: 1. Bachelor’s or Master’s degree in Computer Science, Data Science, or a related field with minimum 12 years of experience 2. Proven experience in architecting & developing AIOps solutions – to streamline Machine Learning & GenAI development lifecycle 3. Proven experience as an AI Ops Architect – ML & GenAI in architecting & design of ML & GenAI platforms 4. Hands on experience in Model deployment strategies, Designing ML & GenAI model pipelines to scale in production, Model Observability techniques used to monitor performance of ML & LLM’s 5. Strong coding skills with experience in implementing best coding practices Technical Skills & Expertise Python, PySpark, PyTorch ,Java, Micro Services, API’s LLMOps – Vector DB, RAG, LLM Orchestration tools, LLM Observability, LLM Guardrails, Responsible AI MLOps - MLFlow, ML/DL libraries, Model & Data Drift Detection libraries & techniques Real Time & Batch Streaming Container Orchestration Platforms Cloud platforms – Azure/ AWS/ GCP, Data Platforms – Databricks/ Snowflake Nice to Have: Understanding of Agent Ops Exposure to Databricks platform You can expect to – Work with world’s biggest Retailers, CPG’s, HealthCare, Banking & Manufacturing customers and help them solve some of their most critical problems Create multi-million Dollar business opportunities by leveraging impact mindset, cutting edge solutions and industry best practices. Work in a diverse environment that keeps evolving Hone your entrepreneurial skills as you contribute to growth of the organization

Posted 1 month ago

Apply

10.0 - 14.0 years

15 - 20 Lacs

Noida

Work from Office

Position Summary The Principal AI Architect is responsible for leading the design and implementation of advanced AI solutions and strategic architecture. Working closely with technology leaders from across our global client community, you will be their senior Trusted Advisor for their AI-enabled transformation journey This role demands deep understanding of AI and related technologies, running in Edge, on-prem and Public Cloud environments. Acting at the forefront of our industry you will be fully conversant with Generative AI, and its impact at both the individual employee and strategic organisational level. The ideal candidate will be an established thought-leader, with solid architectural and engineering credentials. Working ahead of industry trends, deeply passionate about technology-enabled business transformation and demonstrating a strong innovation-led posture. As a thought leader, you will interact frequently with CxO level clients, AI industry leaders, provide expert opinions, and contribute to HCLs strategic vision. Key Responsibilities Technical & Engineering Leadership Design comprehensive AI solution and technology architecture, integrating latest AI technology developments into world-class solutions. Lead high-level architectural discussions with clients, providing expert guidance on best practices for AI implementations across AI PC, Edge, Data Centre and Public Cloud environments. Ensure solutions align with modern best practices across the full spectrum of platforms and environments. Deep understanding across GPU/NPU, Cognitive Infrastructure, Application and Copilot/agent domains. Contribute to HCLs thought leadership in the AI & Cloud domains with a deep understanding of opensource technologies (e.g., Kubernetes, OPEA) and partner technologies. Collaborate on joint technical projects with global partners, including Google, Microsoft, AWS, NVIDIA, IBM, Red Hat, Intel, and Dell. Service Delivery & Innovation Architect innovative AI solutions from ideation to MVP, rapidly enabling genuine business value. Optimize AI and cloud architectures to meet client requirements, balancing efficiency, accuracy and effectiveness. Assess and review existing complex solutions and recommend architectural improvements to transform applications with latest AI technologies. Drive the adoption of cutting-edge GenAI technologies spearheading initiatives that push the boundaries of AI integration across the full spectrum of environments. Thought Leadership and Client Engagement Provide expert architectural and strategy guidance to clients on incorporating Generative AI into their business and technology landscape. Conduct workshops, briefings, and strategic dialogues to educate clients on AI benefits and applications, establishing strong, trust-based relationships. Act as a trusted advisor, contributing to technical projects with a strong focus on technical excellence and on-time delivery. Author whitepapers, blogs, and speak at industry events, maintaining a visible presence as a thought leader in AI and associated technologies. Collaboration and Customer Engagement Engage with multiple customers simultaneously, providing high-impact consultative relationships. Work closely with internal teams and global partners to ensure seamless collaboration and knowledge sharing across projects. Maintain a hands-on technical credibility, staying ahead of industry trends and mentoring others in the organization. Mandatory Skills & Experience Experience: 10+ years architecture design • 7+ years in software engineering. Technologies: Professional-level expertise in Public Cloud environments (AWS, Azure, Google Cloud). Demonstrable coding proficiency with Python, Java or Go languages. AI Expertise: Advanced machine learning algorithms, GenAI models (e.g., GPT, BERT, DALL-E, GEMINI), NLP techniques. Working familiarity with Copilot solutions, in both software engineering and office productivity domains. Communication: Exemplary verbal and written communication skills. Project Methodologies: Agile and Scrum project management. Desired Skills & Experience Knowledge of GenAI operations (LLMOps), experience Governing AI models in production environments. Proficiency in data engineering for AI, including data preprocessing, feature engineering, and pipeline creation. Expertise in AI model fine-tuning and evaluation, with a focus on improving performance for specialized tasks. Copilot design, engineering and extensions. Knowledgeable about Responsible AI, including governance and ethics. Bias mitigation, with experience in implementing strategies to ensure fair and unbiased AI solutions. Deep Learning Frameworks (TensorFlow, PyTorch) Innovation and Emerging Technology Trends Strategic AI Vision and Road mapping Enthusiastic about working in a fast-paced environment using the latest technologies, and passionate about HCLs dynamic and high-energy Lab culture. Verifiable Certification Recognized Professional certification from Google, Microsoft or AWS in an AI and/or Cloud-related domain. Soft Skills and Behavioural Competencies Exemplary communication and leadership skills, capable of inspiring teams and making strategic decisions that align with business goals. Demonstrates a strong customer orientation, innovative problem-solving abilities, and effective cross-cultural collaboration. Expert at driving organizational change and fostering a culture of innovation.

Posted 1 month ago

Apply

10.0 - 14.0 years

18 - 20 Lacs

Noida

Work from Office

Position Summary This is a highly visible role that requires a perfect combination of deep technical credibility, strategic acumen and demonstrable leadership competency. You will be the ultimate Trusted Advisor, capable of engaging business and technology leaders within the worlds largest enterprises, and guiding their strategic AI-enabled journey. The Country Leader, AI Architecture, is responsible for leading the Labs Architectural services within the region. You will need to provide hands-on technical leadership, whilst managing a small team of senior AI architects and consultants. Operating in a fast-moving, highly innovative environment, collaborating with senior Sales and Technical leaders. You will have business responsibility for the provision of innovation-led Labs services. focusing on the design and implementation of advanced AI solutions enabling genuine transformational outcomes. This hands-on leadership role demands deep understanding of AI and related technologies, running in Edge, onprem and Public Cloud environments. Acting at the forefront of our industry you will be fully conversant with Generative AI, and its impact at both the individual employee and strategic organisational level. The ideal candidate will be an established thought-leader in the AI domain, with solid architectural and engineering credentials that you maintain at the highest level. Working ahead of industry trends, deeply passionate about AI-enabled business transformation and demonstrating a strong innovation-led posture. As a thought leader, you will interact frequently with CxO level clients, industry leaders, provide expert opinions, and contribute to HCLs strategic vision. Key Responsibilities Technical & Engineering Leadership Act as ultimate Design Authority for sophisticated AI solutions and related technology architecture. Lead high-level architectural discussions with clients, providing expert guidance on best practices for AI implementations across AI PC, Edge, Data Centre and Public Cloud environments. Ensure solutions align with modern best practices across the full spectrum of platforms and environments. Deep understanding across GPU/NPU, Cognitive Infrastructure, Application and Copilot/agent domains. Contribute to HCLTech thought leadership in the AI & Cloud domains with a deep understanding of open-source (e.g., Kubernetes, OPEA) and partner technologies. Collaborate on joint technical projects with global partners, including Google, Microsoft, AWS, NVIDIA, IBM, Red Hat, Intel, and Dell. Service Delivery & Innovation Design innovative AI solutions from ideation to MVP, rapidly enabling genuine business value. Optimize AI and cloud architectures to meet client requirements, balancing efficiency, accuracy and effectiveness. Assess and review existing complex solutions and recommend architectural improvements to transform applications with latest AI technologies. Drive the adoption of cutting-edge GenAI technologies spearheading initiatives that push the boundaries of AI capability across the full spectrum of environments. Thought Leadership and Client Engagement Provide expert architectural and strategy guidance to clients on incorporating Generative AI into their business and technology landscape. Conduct workshops, briefings, and strategic dialogues to educate clients on AI benefits and applications, establishing strong, trust-based relationships. Act as a trusted advisor, contributing to technical projects with a strong focus on technical excellence and on-time delivery. Author whitepapers, blogs, and speak at industry events, maintaining a visible presence as a thought leader in AI and associated technologies. Collaboration and Customer Engagement Engage with multiple customers simultaneously, building high-impact consultative relationships. Work closely with internal teams and global partners to ensure seamless collaboration and knowledge sharing across projects. Maintain hands-on technical credibility, staying ahead of industry trends and mentoring others in the organization. Management and Leadership Demonstrable track record building and managing small Architectural or Engineering teams. Support career growth and professional development of the team. Enrich and enable world-class technical excellence across the team; supported by a culture of collaboration, respect, diversity, inclusion and deep trustful relationships. Mandatory Skills & Experience Management & leadership : Demonstrable track record building and leading Architectural or Engineering teams. Proven ability to combine strategic business and commercial skills, performing at the highest-level in senior client relationships. Experience: 10+ years architecture design 10+ years software engineering. 5+ years in a senior Team Leader or similar management position. Significant client-facing engagement within a GSI, system integrator, professional services or technology organization. Technologies: Professional-level expertise in Public Cloud environments (AWS, Azure, Google Cloud). Demonstrable coding proficiency with Python, Java or Go languages. AI Expertise: Advanced machine learning algorithms, GenAI models (e.g., GPT, BERT, DALL-E, GEMINI), NLP techniques. Working familiarity with Copilot solutions, in both software engineering and office productivity domains. Business Expertise: Extensive track record performing a lead technical role in a sales, business-development or other commercial environment. Negotiating and consultative skills; experience leading the complete engagement lifecycle. Communication: Experienced public speaker, with an ability to connect with senior business leaders. Project Methodologies: Agile and Scrum project management. Desired Skills & Experience Knowledge of GenAI operations (LLMOps), experience Governing AI models in production environments. Proficiency in data engineering for AI, including data preprocessing, feature engineering, and pipeline creation. Expertise in AI model fine-tuning and evaluation, with a focus on improving performance for specialized tasks. Copilot design, engineering and extensions. Knowledgeable about Responsible AI, including governance and ethics. Bias mitigation, with experience in implementing strategies to ensure fair and unbiased AI solutions. Deep Learning Frameworks (TensorFlow, PyTorch) Innovation and Emerging Technology Trends Strategic AI Vision and Road mapping Enthusiastic about working in a fast-paced environment using the latest technologies, and passionate about HCLs dynamic and high-energy Lab culture. Verifiable Certification Recognized Professional certification from Google, Microsoft or AWS in an AI and/or Cloudrelated domain. Soft Skills and Behavioural Competencies Exemplary communication and leadership skills, capable of inspiring teams and making strategic decisions that align with business goals. Demonstrates a strong customer orientation, innovative problem-solving abilities, and effective cross-cultural collaboration. Expert at driving organizational change and fostering a culture of innovation.

Posted 1 month ago

Apply

5.0 - 10.0 years

15 - 20 Lacs

Bengaluru

Work from Office

Develop and deploy ML pipelines using MLOps tools, build FastAPI-based APIs, support LLMOps and real-time inferencing, collaborate with DS/DevOps teams, ensure performance and CI/CD compliance in AI infrastructure projects. Required Candidate profile Experienced Python developer with 4–8 years in MLOps, FastAPI, and AI/ML system deployment. Exposure to LLMOps, GenAI models, containerized environments, and strong collaboration across ML lifecycle

Posted 1 month ago

Apply

12.0 - 18.0 years

35 - 40 Lacs

Chennai

Work from Office

Tech stack required: Programming languages: Python Public Cloud: AzureFrameworks: Vector Databases such as Milvus, Qdrant/ ChromaDB, or usage of CosmosDB or MongoDB as Vector stores. Knowledge of AI Orchestration, AI evaluation and Observability Tools. Knowledge of Guardrails strategy for LLM. Knowledge on Arize or any other ML/LLM observability tool. Experience: Experience in building functional platforms using ML, CV, LLM platforms. Experience in evaluating and monitoring AI platforms in production Nice to have requirements to the candidate Excellent communication skills, both written and verbal. Strong problem-solving and critical-thinking abilities. Effective leadership and mentoring skills. Ability to collaborate with cross-functional teams and stakeholders. Strong attention to detail and a commitment to delivering high-quality solutions. Adaptability and willingness to learn new technologies. Time management and organizational skills to handle multiple projects and priorities.

Posted 2 months ago

Apply

5.0 - 8.0 years

15 - 25 Lacs

Hyderabad, Pune, Bengaluru

Work from Office

Role: Gen AI Engineer Exp: 5 to 8 yrs. Loc: Bangalore, Pune, Hyderabad NP: Immediate joiners, who can join in 30 days. Required Skills: Python, Large Language Models (LLM), Machine Learning (ML), Generative AI

Posted 2 months ago

Apply

10.0 - 12.0 years

0 Lacs

Bengaluru / Bangalore, Karnataka, India

On-site

Job Description: Oracle Cloud Infrastructure (OCI) is a pioneering force in cloud technology, merging the agility of startups with the robustness of an enterprise software leader. Within OCI, the Oracle Generative AI Service team spearheads innovative solutions at the convergence of artificial intelligence and cloud infrastructure. As part of this team, you'll contribute to large-scale cloud solutions utilizing cutting-edge machine learning technologies, aimed at addressing complex global challenges. Join us to create innovative solutions using top-notch machine learning technologies to solve global challenges. We're looking for an experienced Principal Applied Data Scientist to join our OCI Gen-AI Solutions team for strategic customers. In this role, you'll collaborate with applied scientists and product managers to design, develop, and deploy tailored Gen-AI solutions with an emphasis on Large Language Models (LLMs), Agents, MPC and Retrieval Augmented Generation (RAG) with large OpenSearch clusters. As part of the OCI Gen AI and Data Solutions for strategic customers team, you will be responsible for developing innovative Gen AI and data services for our strategic customers.As a Principal Applied Data Scientist, you'll lead the development of advanced Gen AI solutions using the latest ML technologies combined with Oracle's cloud expertise. Your work will significantly impact sectors like financial services, telecom, healthcare, and code generation by creating distributed, scalable, high-performance solutions for strategic customers. Work directly with key customers and accompany them on their Gen AI journey - understanding their requirements, help them envision and design and build the right solutions and work together with their ML engineering to remove blockers. You will dive deep into model structure to optimize model performance and scalability. You will build state of art solutions with brand new technologies in this fast-evolving area. You will configure large scale OpenSearch clusters, setting up ingestion pipelines to get the data into the OpenSearch. You will diagnose, troubleshoot, and resolve issues in AI model training and serving. You may also perform other duties as assigned. Build re-usable solution patterns and reference solutions / showcases that can apply across multiple customers. Be an enthusiastic, self-motivated, and a great collaborator. Be our product evangelist - engage directly with customers and partners, participate and present in external events and conferences, etc. Qualifications and experience Bachelors or master's in computer science or equivalent technical field with 10+ years of experience Able to optimally communicate technical ideas verbally and in writing (technical proposals, design specs, architecture diagrams and presentations). Demonstrated experience in designing and implementing scalable AI models and solutions for production,relevant professional experience as end-to-end solutions engineer or architect (data engineering, data science and ML engineering is a plus), with evidence of close collaborations with PM and Dev teams. Experience with OpenSearch, Vector databases, PostgreSQL and Kafka Streaming. Practical experience with setting up and finetuning large OpenSearch Clusters. Experience in setting up data ingestion pipelines with OpenSearch. Experience with search algorithms, indexing, optimizing latency and response times. Practical experience with the latest technologies in LLM and generative AI, such as parameter-efficient fine-tuning, instruction fine-tuning, and advanced prompt engineering techniques like Tree-of-Thoughts. Familiarity with Agents and Agent frameworks and Model Predictive Control (MPC) Hands-on experience with emerging LLM frameworks and plugins, such as LangChain, LlamaIndex, VectorStores and Retrievers, LLM Cache, LLMOps (MLFlow), LMQL, Guidance, etc. Strong publication record, including as a lead author or reviewer, in top-tier journals or conferences. Ability and passion to mentor and develop junior machine learning engineers. Proficient in Python and shell scripting tools. Preferred Qualifications : Masters or Bachelor's in related field with 5+ years relevant experience Experience with RAG based solutions architecture. Familiarity in OpenSearch and Vector stores as a knowledge store Knowledge of LLM and experience delivering, Generative AI And Agent models are a significant plus. Familiarity and experience with the latest advancements in computer vision and multimodal modeling is a plus. Experience with semantic search, multi-modal search and conversational search. Experience in working on a public cloud environment, and in-depth knowledge of IaaS/PaaS industry and competitive capabilities.Experience with popular model training and serving frameworks like KServe, KubeFlow, Triton etc. Experience with LLM fine-tuning, especially the latest parameter efficient fine-tuning technologies and multi-task serving technologies. Deep technical understanding of Machine Learning, Deep Learning architectures like Transformers, training methods, and optimizers. Experience with deep learning frameworks (such as PyTorch, JAX, or TensorFlow) and deep learning architectures (especially Transformers). Experience in diagnosing, fixing, and resolving issues in AI model training and serving. Career Level - IC4

Posted 2 months ago

Apply

10.0 - 18.0 years

30 - 45 Lacs

Hyderabad, Pune, Bengaluru

Work from Office

Role - Senior Data Scientist / Senior Gen AI Engineer Exp Range - 8 to 18 yrs Position - Permanent Fulltime Company - Data Analytics & AIML MNC Location - Hyderabad, Pune, Bangalore (Relocation accepted) About the Role: We are seeking a Software Engineer with expertise in Generative AI and Microsoft technologies to design, develop, and deploy AI-powered solutions using the Microsoft ecosystem. You will work with cross-functional teams to build scalable applications leveraging generative AI models and Azure services. Skills Required: Experience with Large Language Models (LLMs) like GPT, LLaMA, Claude, etc. Proficiency in Python for building and fine-tuning AI/ML models Familiarity with LangChain , LLMOps , or RAG (Retrieval-Augmented Generation) pipelines Experience with Vector Databases (e.g. FAISS, Pinecone, Weaviate) Knowledge of Prompt Engineering and model evaluation techniques Exposure to cloud platforms (Azure, AWS or GCP) for deploying GenAI solutions Preferred Skills: Experience with Azure OpenAI , Databricks or Microsoft Fabric Hands-on with Hugging Face Transformers , OpenAI APIs or custom model training

Posted 2 months ago

Apply

0.0 years

3 - 6 Lacs

Delhi, Delhi, IN

On-site

About the job: Key responsibilities: 1. Build AI-driven tools and products using APIs (OpenAI, Gemini, etc.) 2. Design and fine-tune prompts for various use cases. 3. Integrate vector databases (Pinecone, ChromaDB) for retrieval-augmented generation (RAG) 4. Use tools like LangChain or LlamaIndex for multi-step worklows 5. Collaborate with designers, content teams, and founders to turn ideas into polished tools Who can apply: Only those candidates can apply who: are Computer Science Engineering students Salary: ₹ 3,20,000 - 6,50,000 /year Experience: 0 year(s) Deadline: 2025-06-22 23:59:59 Skills required: Natural Language Processing (NLP), Deep Learning, Prompt Engineering, ChatGPT, Claude, Gemini, LLMOps and Model fine-tuning Other Requirements: 1. Degree Btech - AI/Ml, others ( who has done AI/ML projects) 2. Strong understanding of LLM APIs (OpenAI, Claude, Gemini, etc.) 3. REST API integration and deployment knowledge 4. GitHub portfolio with working AI tools or integrations About Company: Stirring Minds is a premier startup ecosystem in India, dedicated to helping businesses launch, scale, and succeed. As a leading incubator, we provide funding, co-working spaces, and mentorship to support the growth of innovative companies. In addition to our incubator services, we also host the largest startup event in the country known as Startup Summit Live, bringing together entrepreneurs and industry leaders to connect, learn, and collaborate. Our community-driven approach extends beyond our event and incubator offerings, as we work to create communities of like-minded individuals who can support and learn from one another. We have been recognized by top media outlets both in India and internationally, including the BBC, The Guardian, Entrepreneur, and Business Insider. Our goal is to provide a comprehensive ecosystem for startups and help turn their ideas into reality.

Posted 2 months ago

Apply

0.0 years

3 - 4 Lacs

IN

Remote

About the job: Key responsibilities: 1. Design and develop scalable backend systems and APIs for our AI-powered SaaS platform using Python and Node.js 2. Build and maintain cloud infrastructure on AWS, including configuration and management of S3, DynamoDB, SNS, EC2, and CloudWatch services 3. Implement and optimize data processing pipelines for machine learning model deployment and integration 4. Collaborate with data scientists to integrate AI models into production systems and ensure efficient model serving 5. Deploy and monitor applications using DevOps practices and LLMOps for large language model implementations 6. Create robust API endpoints that connect our frontend applications with AI functionalities 7. Design and implement efficient database schemas and queries optimized for AI applications 8. Develop and maintain secure authentication and authorization systems for our platform 9. Write clean, maintainable, and well-tested code following best practices 10. Troubleshoot and resolve complex technical issues in production environments Additional candidate preferences: 1. Computer Science or related Engineering degree preferred 2. Experience with containerization technologies like Docker 3. Familiarity with AI model serving platforms Who can apply: Only those candidates can apply who: are Computer Science Engineering students Salary: ₹ 3,10,000 - 4,60,000 /year Experience: 0 year(s) Deadline: 2025-06-16 23:59:59 Other perks: 5 days a week Skills required: Python, Node.js, Artificial intelligence, DevOps, Amazon EC2, Amazon S3, Amazon CloudWatch, Amazon SNS, Amazon DynamoDB and LLMOps Other Requirements: 1. Computer Science or related Engineering degree preferred 2. Experience with containerization technologies like Docker 3. Familiarity with AI model serving platforms and ML workflows About Company: Smartify is a marketplace for automation companies and also India's leading home automation store. We are trying to reduce the knowledge-execution gap and encourage early-adopters in the IoT space to launch their products and get to the mainstream market.

Posted 2 months ago

Apply

1 - 6 years

7 - 14 Lacs

Hyderabad

Work from Office

Position - AI Engineer As an AI Engineer, you will design, implement, and optimize machine learning models and AI systems to solve complex problems. You will work closely with cross-functional teams to integrate AI solutions into our products and services, ensuring scalability and efficiency. Key Responsibilities: Application Development: Design and develop AI-powered applications using state-of-the-art LLM models and generative AI techniques. Implement scalable solutions that integrate LLM-powered tools into existing workflows or standalone products. Model Optimization: Fine-tune pre-trained LLM models to meet specific application requirements. Optimize model performance for real-time and high-throughput environments. LLMOps Implementation: Develop and maintain pipelines for model deployment, monitoring, and retraining. Set up robust systems for model performance monitoring and diagnostics. Ensure reliable operations through analytics and insights into model behavior. Vector Databases and Data Management: Utilize vector databases for efficient storage and retrieval of embeddings. Integrate databases with LLM applications to enhance query and recommendation systems. Collaboration and Innovation: Work closely with cross-functional teams, including product managers, data scientists, and software engineers. Stay up-to-date with advancements in generative AI and LLM technologies to drive innovation. Skills and Experience 3+ years of experience in AI/ML development, with a focus on generative AI and LLMs. Proficiency in programming languages such as Python and frameworks like PyTorch or TensorFlow. Hands-on experience in fine-tuning and deploying LLM models (e.g., GPT, BERT, etc.). Familiarity with LLMOps practices, including pipeline automation, monitoring, and analytics. Experience with vector databases (e.g., Pinecone, Weaviate, or similar). Strong knowledge of natural language processing (NLP) and machine learning principles. You should certainly apply if: Understanding of MLOps principles and cloud platforms (AWS, GCP, Azure). Familiarity with prompt engineering and reinforcement learning from human feedback (RLHF). Experience in building real-time applications powered by generative AI. Knowledge of distributed systems and scalable architectures.

Posted 2 months ago

Apply

1.0 years

3 - 5 Lacs

IN

Remote

About the job: We empower the people who build the world. Taiy .AI is the world's largest infrastructure construction data-mesh technology and the first AI platform for the global infrastructure construction industry. Our clients include some of the largest construction firms, suppliers, and the government. About The Team: We are looking for a Python Engineer to help support and lead our data engineering ops. Key Responsibilities: 1. Developing and executing processes for monitoring data sanity, checking for data availability and reliability. 2. Understanding the business drivers and building insights through data. 3. Partner with stakeholders at all levels to establish current and ongoing data support and reporting needs. 4. Ensure continuous data accuracy and recognize data discrepancies in systems that require immediate attention/escalation. 5. Become an expert in the company's data warehouse and other data storage tools, understanding the definition, context, and proper use of all attributes and metrics. 6. Creating dashboards based on business requirements. 7. Distributed systems, Scala, cloud, Caching, CI/CD (Continuous integration and deployment), Distributed logging, Data pipeline, Recommendation Engine, Data at Rest Encryption What To Bring: 1. Graduate/Post Graduate degree in Computer Science or Engineering. 2. 1-3 years of hands-on experience with AWS Open Search v1.0 or Elastic Search 7.9 3. 3+ years of work experience on Scala 4. Must be able to drive, design, code, review the work, and assist the teams 5. Good problem-solving skills 6. Good oral and written communication in English 7. Should be open to/have experience of working in a fast-paced delivery environment 8. Strong understanding of object-oriented design, data structures, algorithms, profiling, and optimization. 9. Good to have experience on Elasticsearch and Spark-Elasticsearch 10. Knowledge of Garbage Collection and experience in GC tuning. 11. Knowledge of algorithms like sorting, heap/stack, queue, search, etc. 12. Experience with Git and build tools like Gradle/Maven/SBT 13. Should be able to write complex queries independently. 14. Strong knowledge of programming languages like Python, Scala, etc. 15. Ability to work independently and take ownership of things. 16. An analytical mindset and strong attention to detail. 17. Good verbal & written communication skills for coordinating across teams. Who can apply: Only those candidates can apply who: have minimum 1 years of experience are Computer Science Engineering students Salary: ₹ 3,00,000 - 5,00,000 /year Experience: 1 year(s) Deadline: 2025-06-05 23:59:59 Other perks: 5 days a week Skills required: Python, Selenium, Machine Learning, REST API, Data Extraction, Data Engineering and LLMOps About Company: Taiyo is a Silicon Valley startup that aggregates, predicts, and visualizes the world's data so customers don't have to. We are a globally-distributed team with a focus on the infrastructure vertical. The Taiyo team was founded by an interdisciplinary group of experts from Stanford University's AI Institute, World Bank, International Monetary Fund, and UC Berkeley.

Posted 2 months ago

Apply
Page 2 of 2
cta

Start Your Job Search Today

Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.

Job Application AI Bot

Job Application AI Bot

Apply to 20+ Portals in one click

Download Now

Download the Mobile App

Instantly access job listings, apply easily, and track applications.

Featured Companies