Get alerts for new jobs matching your selected skills, preferred locations, and experience range. Manage Job Alerts
1.0 - 5.0 years
0 Lacs
navi mumbai, maharashtra
On-site
As a Senior Software Engineer specializing in Data Science, you will be responsible for leveraging your 3.5 years of relevant experience in Python programming to design, develop, and maintain Python-based applications with a focus on Machine Learning (ML) and Artificial Intelligence (AI). With a minimum of 1+ years of experience in machine learning and natural language processing, you will have the opportunity to build and optimize ML/NLP models and seamlessly integrate them into production systems. Your role will also involve implementing and fine-tuning Large Language Models (LLMs) and Generative AI (GenAI) solutions for various business applications, drawing from your minimum 1.5 years of experience with LLMs and GenAI. Additionally, your familiarity with deploying ML models and Python applications on cloud platforms such as Azure or AWS will be essential, along with your knowledge of database management to support ML applications and data pipelines. Collaboration with cross-functional teams to deliver effective solutions, participation in code reviews, and providing feedback will be integral parts of your responsibilities. Your expertise in troubleshooting and resolving technical issues in production environments will ensure the smooth functioning of the applications you develop. Moreover, staying updated with emerging trends in ML, NLP, and GenAI will further enhance your contributions to the team. This full-time role based in Airoli, Navi Mumbai, requires you to work in-office during the hours of 11 AM to 8 PM. As an immediate joiner, you will have the opportunity to apply your skills and experience towards impactful projects while growing professionally within the organization.,
Posted 2 days ago
6.0 - 10.0 years
0 Lacs
karnataka
On-site
Data Science is all about breaking new ground to enable businesses to answer their most urgent questions. Pioneering massively parallel data-intensive analytic processing, the mission is to develop a whole new approach to generating meaning and value from petabyte-scale data sets and shape brand new methodologies, tools, statistical methods, and models. In collaboration with leading academics, industry experts, and highly skilled engineers, the goal is to equip customers to generate sophisticated new insights from the biggest of big data. Join the team to do the best work of your career and make a profound social impact as an Advisor on the Data Science Team in Bangalore. As a Data Science Advisor, you will contribute to the business strategy and influence decision-making based on information gained from deep dive analysis. You will produce actionable and compelling recommendations by interpreting insights from complex data sets. Designing processes to consolidate and examine unstructured data to generate actionable insights will be part of your responsibilities. Additionally, you will partner with business leaders, engineers, and industry experts to construct predictive models, algorithms, and probability engines. You will: - Partner with internal and external teams to understand customer requirements and develop proposals. - Conduct interactions with external customers to gather project requirements, provide status updates, and share analytical insights. - Implement preliminary data exploration and data preparation steps for model development/validation. - Apply a broad range of techniques and theories from statistics, machine learning, and business intelligence to deliver actionable business insights. - Solution, build, deploy, and set up monitoring for models. Qualifications: - 6+ years of related experience with proficiency in NLP, Machine Learning, Computer Vision, and GenAI. - Working experience in data visualization (e.g., Power BI, matplotlib, plotly). - Hands-on experience with CNN, LSTM, YOLO, and database skills including SQL, Postgres SQL, PGVector, and ChromaDB. - Proven experience in MLOps and LLMOps, with a strong understanding of ML lifecycle management. - Expertise with large language models (LLMs), prompt engineering, fine-tuning, and integrating LLMs into applications for natural language processing (NLP) tasks. Desirable Skills: - Strong product/technology/industry knowledge and familiarity with streaming/messaging frameworks (e.g., Kafka, RabbitMQ, ZeroMQ). - Experience with cloud platforms (e.g., AWS, Azure, GCP). - Experience with web technologies and frameworks (e.g., HTTP/REST/GraphQL, Flask, Django). - Skilled in programming languages like Java or JavaScript. Dell Technologies is committed to providing equal employment opportunities for all employees and creating a work environment free of discrimination and harassment. If you are looking for an opportunity to grow your career with advanced technology and some of the best minds in the industry, this role might be the perfect fit for you. Join Dell Technologies to build a future that works for everyone because Progress Takes All of Us.,
Posted 2 days ago
4.0 - 8.0 years
0 Lacs
karnataka
On-site
As a Technical Manager, you will lead and manage a team of software engineers to ensure high performance and quality delivery. Your responsibilities will include designing, developing, and maintaining scalable and robust Python applications. You will architect and implement cloud solutions on AWS and Azure, adhering to best practices in security, scalability, and cost-efficiency. Collaborating with cross-functional teams, you will define, design, and ship new features while mentoring and guiding team members in their technical and professional growth. In this role, you will implement DevOps practices to streamline CI/CD pipelines and automate deployment processes. Developing and maintaining APIs using FastAPI and GraphQL will be part of your tasks, ensuring that the team follows best practices in coding, testing, and documentation. You will also oversee database design, optimization, and maintenance, driving productivity within the team by implementing efficient workflows and leveraging code assist tools. To be successful in this position, you should possess a Bachelor's or Master's degree in Computer Science, Engineering, or a related field. Proven experience as a Technical Manager or in a similar role, leading large teams, is required. Strong proficiency in Python programming, extensive experience with AWS or Azure cloud services, and a solid understanding of DevOps practices and tools are essential. Experience with FastAPI and GraphQL, database systems (SQL and NoSQL), design patterns, and microservices architecture is also necessary. Additionally, familiarity with MLOps for deploying and managing machine learning models, LLMOps for large language model operations, Databricks for big data processing and analytics, and excellent problem-solving skills with attention to detail are desired. Strong communication and leadership skills, along with the ability to drive productivity and enhance team efficiency using code assist tools, will be valuable assets in this role.,
Posted 2 days ago
9.0 - 14.0 years
18 - 22 Lacs
Hyderabad, Chennai, Bengaluru
Work from Office
Job Description: As an LLMOps Engineer, you will play a crucial role in the deployment, maintenance, and optimization of large language models (LLMs). Your responsibilities will span the entire lifecycle of LLMs, from initial deployment to ongoing operations, ensuring optimal performance, scalability, and reliability. 5+ years of relevant experience in data engineering roles. Key Responsibilities: LLM Deployment and Integration - Deploy and integrate large language models into production environments, ensuring seamless integration with existing systems and applications. Infrastructure Planning and Scaling - Collaborate with cross-functional teams to plan and design the infrastructure required for LLM deployment. Implement scalable solutions to accommodate growing data volumes and user loads. Automation of Deployment Processes - Develop and maintain automation scripts and tools for efficient deployment, scaling, and versioning of LLMs. Streamline deployment processes to minimize downtime. Continuous Monitoring and Alerting - Implement monitoring systems to track LLM performance metrics. Set up alerts for potential issues and respond promptly to ensure uninterrupted service. Performance Monitoring and Optimization - Monitor the performance of LLMs in real-time, conduct regular assessments, and implement optimizations to enhance efficiency and responsiveness. Fault Tolerance and Disaster Recovery - Design and implement fault-tolerant systems for LLMs, incorporating strategies such as redundancy, sharding, and replication. Develop and maintain disaster recovery plans. Security Measures Implementation - Implement robust security measures to safeguard LLMs and associated data. Ensure compliance with data security regulations and industry standards. Collaboration with NLP Engineers and Data Scientists - Collaborate with NLP (Natural Language Processing) engineers and data scientists to understand model requirements and implement necessary infrastructure adjustments. Skills & Tools Infrastructure as Code (IaC) - Experience with IaC tools such as Terraform or Ansible for automating infrastructure provisioning. Containerization and Orchestration - Proficiency in containerization technologies (e.g., Docker) and orchestration tools (e.g., Kubernetes) for managing LLM deployments. Cloud Platforms - Familiarity with cloud platforms such as AWS (Bedrock), Azure, or GCP, and experience in deploying and managing applications in a cloud environment. Monitoring and Logging Tools - Knowledge of monitoring tools (e.g., Prometheus, Grafana) and logging systems (e.g., ELK stack) for real-time performance monitoring and analysis. Security Measures - Understanding of security guardrails using tools like LLM Guard and familiarity of how to mask / redact / obfuscate sensitive data, protect the input and output of toxic and harmful content to / from LLMs and understand the performance implications of the same. Scripting and Automation - Proficient in scripting languages such as Python, Shell, or similar, and experience in automating deployment and maintenance processes.
Posted 3 days ago
1.0 - 5.0 years
0 Lacs
maharashtra
On-site
As a Senior Software Engineer specializing in Data Science, you are expected to bring 3 to 6 years of experience to the role. Your expertise will be crucial in driving our projects forward and achieving our goals. Working at our location in Airoli, Navi Mumbai, you will adhere to office hours from 11 AM to 8 PM across 5 days a week. Your primary responsibility will be to leverage your 3 years of experience as a Data Scientist and a minimum of 3 years of proficiency in Python programming to design, develop, and maintain Python-based applications with a focus on ML and AI. Additionally, you should possess at least 1+ years of experience in ML & NLP, 1.5 years of experience with LLMs and generative AI (GenAI), and 2 years of experience with any database. Your role will also involve deploying ML models/Python applications on Azure or AWS, and familiarity with LLMOps will be considered a plus. In this position, you will design and optimize ML/NLP models and integrate them into production systems. You will be responsible for implementing and fine-tuning LLMs and GenAI solutions for business applications, managing database structures to support ML applications and data pipelines, and deploying and monitoring ML models in cloud environments such as AWS/Azure. Collaboration is key in our work environment, and you will be expected to work closely with cross-functional teams to deliver effective solutions. Your ability to write clean, efficient, and well-documented code will be essential. Participation in code reviews, providing feedback, troubleshooting technical issues in production, and staying updated with emerging trends in ML, NLP, and GenAI are all vital aspects of this role.,
Posted 4 days ago
2.0 - 10.0 years
0 Lacs
thane, maharashtra
On-site
The ideal candidate should have a minimum of 5-10 years of experience in data science with a strong proficiency in Python programming. Additionally, the candidate should have at least 4 years of experience in Machine Learning/Natural Language Processing (NLP), with a minimum of 1.5 years in Large Language Models (LLM) & Generative AI (GenAI) technologies. Proficiency in database management for at least 2 years is required, along with 1+ years of experience in cloud deployment of ML models/Python applications on Azure or AWS. Familiarity with LLMOps is considered a plus. As a part of the role, the candidate will be responsible for leading the development, deployment, and optimization of ML, NLP, and LLM/GenAI solutions. This includes the development and maintenance of Python-based ML pipelines, as well as managing associated databases. The candidate will also be tasked with deploying AI models on Azure/AWS, ensuring performance and scalability. Additionally, the candidate will be expected to stay updated on advancements in ML, NLP, and GenAI technologies and provide mentorship to team members. Implementing LLMOps best practices for efficient model lifecycle management will also be a key responsibility.,
Posted 1 week ago
0.0 - 4.0 years
0 Lacs
lucknow, uttar pradesh
On-site
We are seeking a dynamic Web Development intern to join our team at Terrain Bhuexpert Innovation Pvt Ltd! If you are enthusiastic about Node.js, JavaScript, React, Microservices, Artificial Intelligence, and LLMOps, this is an excellent opportunity for you to acquire hands-on experience in a fast-paced environment. Your responsibilities will include collaborating with our development team to develop and maintain cutting-edge web applications using Node.js, React, and Microservices. You will apply your knowledge of Artificial Intelligence to improve user experience and drive innovation within our projects. Additionally, you will assist in optimizing our infrastructure and processes through the adoption of LLMOps practices. Writing clean, efficient, and well-documented code to ensure the scalability and reliability of our applications will be a key aspect of your role. Staying updated on industry trends and technologies to bring fresh ideas and insights to our projects is essential. You will participate in code reviews and offer constructive feedback to enhance the overall quality of our codebase. Taking ownership of assigned tasks and projects, showcasing a strong work ethic, and demonstrating a willingness to learn and grow in a supportive team environment are crucial. If you are eager to advance your web development career and leverage your skills to create a meaningful impact, apply now and embark on this exciting journey of innovation and growth with us! About Company: Bhuexpert is developing India's premier managed real estate marketplace, where technology intersects with trust. Our mission is to digitize and automate the entire property buying journey, from AI-powered property recommendations and dynamic pricing to e-KYC, digital signatures, and secure document storage. We are not merely constructing a platform; we are establishing the technological backbone of a trillion-dollar industry. If you are passionate about solving intricate problems and aspire to have your code influence real lives, Bhuexpert is the perfect place for you.,
Posted 1 week ago
1.0 - 5.0 years
0 Lacs
hyderabad, telangana
On-site
You will be working full-time at Soothsayer Analytics in Hyderabad as a Generative AI/LLM Engineer. Your primary responsibility will be to design, develop, and deploy generative AI models using cutting-edge technologies like Azure OpenAI GPT-4, GPT-4 Vision, or GPT-4 Omni. You should have a strong background in building and deploying AI models, with a focus on leveraging technologies such as Retrieval-Augmented Generation (RAG) and working with Vector Databases. While experience in fine-tuning large language models (LLMs) is beneficial, it is not mandatory. You are expected to have a general understanding of training or fine-tuning deep learning models and be able to quickly learn and implement advanced techniques. Your key responsibilities will include designing, developing, and deploying generative AI models using GPT-4 variants, implementing and optimizing RAG techniques, building and managing AI services using Python frameworks like LangChain or LlamaIndex, and developing APIs with FastAPI or Quart for efficient integration. You will focus on scalability, performance, and optimization of AI solutions across cloud environments, particularly with Azure and AWS. Working with Vector Databases is mandatory, and experience with Graph Databases is optional. You will also utilize Cosmos DB and SQL for data storage and management solutions, apply MLOps or LLMOps practices, and manage Azure Pipelines for continuous integration and deployment. Additionally, you will continuously research and adopt the latest advancements in AI technologies. To qualify for this role, you should have a Bachelor's or Master's degree in Computer Science, AI, Data Science, or a related field, with at least 1+ years of experience in Generative AI/LLM technologies and 5+ years of experience in related fields. Proficiency in Python and experience with frameworks like LangChain, LlamaIndex, FastAPI, or Quart is required. Expertise in RAG and experience with Vector Databases are mandatory. Knowledge of Cosmos DB and SQL is also essential. Experience with fine-tuning LLMs and Graph Databases is beneficial but not mandatory. You should have proven experience in MLOps, LLMOps, or DevOps, a strong understanding of CI/CD processes, automation, and pipeline management, and familiarity with containers, Docker, or Kubernetes. Experience with cloud platforms, particularly Azure or AWS, and cloud-native AI services is desirable. Strong problem-solving abilities and a proactive approach to learning new AI trends and best practices quickly are necessary for this role.,
Posted 1 week ago
6.0 - 10.0 years
20 - 35 Lacs
Hyderabad, Pune, Bengaluru
Work from Office
Role: Gen AI Engineer Exp: 5 to 8 yrs. Loc: Bangalore, Pune, Hyderabad NP: Immediate joiners, who can join in 30 days. Required Skills: Python, Large Language Models (LLM), Machine Learning (ML), Generative AI Required Candidate profile Looking for candidates working in Data Analytics companies preferably.
Posted 1 week ago
5.0 - 9.0 years
0 Lacs
maharashtra
On-site
The Implementation Technical Architect role focuses on designing, developing, and deploying cutting-edge Generative AI (GenAI) solutions using the latest Large Language Models (LLMs) and frameworks. Your responsibilities include creating scalable and modular architecture for GenAI applications, leading Python development for GenAI applications, building tools for automated data curation, integrating solutions with cloud platforms like Azure, GCP, and AWS, applying advanced fine-tuning techniques to optimize LLM performance, establishing LLMOps pipelines, ensuring ethical AI practices, implementing Reinforcement Learning with Human Feedback and Retrieval-Augmented Generation techniques, collaborating with front-end developers, and more. Key Responsibilities: - Design and Architecture: Create scalable and modular architecture for GenAI applications using frameworks like Autogen, Crew.ai, LangGraph, LlamaIndex, and LangChain. - Python Development: Lead the development of Python-based GenAI applications, ensuring high-quality, maintainable, and efficient code. - Data Curation Automation: Build tools and pipelines for automated data curation, preprocessing, and augmentation to support LLM training and fine-tuning. - Cloud Integration: Design and implement solutions leveraging Azure, GCP, and AWS LLM ecosystems, ensuring seamless integration with existing cloud infrastructure. - Fine-Tuning Expertise: Apply advanced fine-tuning techniques such as PEFT, QLoRA, and LoRA to optimize LLM performance for specific use cases. - LLMOps Implementation: Establish and manage LLMOps pipelines for continuous integration, deployment, and monitoring of LLM-based applications. - Responsible AI: Ensure ethical AI practices by implementing Responsible AI principles, including fairness, transparency, and accountability. - RLHF and RAG: Implement Reinforcement Learning with Human Feedback (RLHF) and Retrieval-Augmented Generation (RAG) techniques to enhance model performance. - Modular RAG Design: Develop and optimize Modular RAG architectures for complex GenAI applications. - Open Source Collaboration: Leverage Hugging Face and other open-source platforms for model development, fine-tuning, and deployment. - Front-End Integration: Collaborate with front-end developers to integrate GenAI capabilities into user-friendly interfaces. Required Skills: - Python Programming: Deep expertise in Python for building GenAI applications and automation tools. - LLM Frameworks: Proficiency in frameworks like Autogen, Crew.ai, LangGraph, LlamaIndex, and LangChain. - Large-Scale Data Handling & Architecture: Design and implement architectures for handling large-scale structured and unstructured data. - Multi-Modal LLM Applications: Familiarity with text chat completion, vision, and speech models. - Fine-tune SLM(Small Language Model) for domain specific data and use cases. - Prompt injection fallback and RCE tools such as Pyrit and HAX toolkit etc. - Anti-hallucination and anti-gibberish tools such as Bleu etc. - Cloud Platforms: Extensive experience with Azure, GCP, and AWS LLM ecosystems and APIs. - Fine-Tuning Techniques: Mastery of PEFT, QLoRA, LoRA, and other fine-tuning methods. - LLMOps: Strong knowledge of LLMOps practices for model deployment, monitoring, and management. - Responsible AI: Expertise in implementing ethical AI practices and ensuring compliance with regulations. - RLHF and RAG: Advanced skills in Reinforcement Learning with Human Feedback and Retrieval-Augmented Generation. - Modular RAG: Deep understanding of Modular RAG architectures and their implementation. - Hugging Face: Proficiency in using Hugging Face and similar open-source platforms for model development. - Front-End Integration: Knowledge of front-end technologies to enable seamless integration of GenAI capabilities. - SDLC and DevSecOps: Strong understanding of secure software development lifecycle and DevSecOps practices for LLMs.,
Posted 2 weeks ago
4.0 - 8.0 years
0 Lacs
maharashtra
On-site
At PwC, our data and analytics team focuses on utilizing data to drive insights and support informed business decisions. We leverage advanced analytics techniques to assist clients in optimizing their operations and achieving strategic goals. As a data analysis professional at PwC, your role will involve utilizing advanced analytical methods to extract insights from large datasets, enabling data-driven decision-making. Your expertise in data manipulation, visualization, and statistical modeling will be pivotal in helping clients solve complex business challenges. PwC US - Acceleration Center is currently seeking a highly skilled MLOps/LLMOps Engineer to play a critical role in deploying, scaling, and maintaining Generative AI models. This position requires close collaboration with data scientists, ML/GenAI engineers, and DevOps teams to ensure the seamless integration and operation of GenAI models within production environments at PwC and for our clients. The ideal candidate will possess a strong background in MLOps practices and a keen interest in Generative AI technologies. With a preference for candidates with 4+ years of hands-on experience, core qualifications for this role include: - 3+ years of experience developing and deploying AI models in production environments, alongside 1 year of working on proofs of concept and prototypes. - Proficiency in software development, including building and maintaining scalable, distributed systems. - Strong programming skills in languages such as Python and familiarity with ML frameworks like TensorFlow and PyTorch. - Knowledge of containerization and orchestration tools like Docker and Kubernetes. - Understanding of cloud platforms such as AWS, GCP, and Azure, including their ML/AI service offerings. - Experience with continuous integration and delivery tools like Jenkins, GitLab CI/CD, or CircleCI. - Familiarity with infrastructure as code tools like Terraform or CloudFormation. Key Responsibilities: - Develop and implement MLOps strategies tailored for Generative AI models to ensure robustness, scalability, and reliability. - Design and manage CI/CD pipelines specialized for ML workflows, including deploying generative models like GANs, VAEs, and Transformers. - Monitor and optimize AI model performance in production, utilizing tools for continuous validation, retraining, and A/B testing. - Collaborate with data scientists and ML researchers to translate model requirements into scalable operational frameworks. - Implement best practices for version control, containerization, and orchestration using industry-standard tools. - Ensure compliance with data privacy regulations and company policies during model deployment. - Troubleshoot and resolve issues related to ML model serving, data anomalies, and infrastructure performance. - Stay updated with the latest MLOps and Generative AI developments to enhance AI capabilities. Project Delivery: - Design and implement scalable deployment pipelines for ML/GenAI models to transition them from development to production environments. - Oversee the setup of cloud infrastructure and automated data ingestion pipelines to meet GenAI workload requirements. - Create detailed documentation for deployment pipelines, monitoring setups, and operational procedures. Client Engagement: - Collaborate with clients to understand their business needs and design ML/LLMOps solutions. - Present technical approaches and results to technical and non-technical stakeholders. - Conduct training sessions and workshops for client teams. - Create comprehensive documentation and user guides for clients. Innovation And Knowledge Sharing: - Stay updated with the latest trends in MLOps/LLMOps and Generative AI. - Develop internal tools and frameworks to accelerate model development and deployment. - Mentor junior team members and contribute to technical publications. Professional And Educational Background: - Any graduate / BE / B.Tech / MCA / M.Sc / M.E / M.Tech / Masters Degree / MBA,
Posted 2 weeks ago
5.0 - 9.0 years
0 Lacs
kochi, kerala
On-site
As a highly skilled Senior Machine Learning Engineer, you will leverage your expertise in Deep Learning, Large Language Models (LLMs), and MLOps/LLMOps to design, optimize, and deploy cutting-edge AI solutions. Your responsibilities will include developing and scaling deep learning models, fine-tuning LLMs (e.g., GPT, Llama), and implementing robust deployment pipelines for production environments. You will be responsible for designing, training, fine-tuning, and optimizing deep learning models (CNNs, RNNs, Transformers) for various applications such as NLP, computer vision, or multimodal tasks. Additionally, you will fine-tune and adapt LLMs for domain-specific tasks like text generation, summarization, and semantic similarity. Experimenting with RLHF (Reinforcement Learning from Human Feedback) and alignment techniques will also be part of your role. In the realm of Deployment & Scalability (MLOps/LLMOps), you will build and maintain end-to-end ML pipelines for training, evaluation, and deployment. Deploying LLMs and deep learning models in production environments using frameworks like FastAPI, vLLM, or TensorRT is crucial. You will optimize models for low-latency, high-throughput inference and implement CI/CD workflows for ML systems using tools like MLflow and Kubeflow. Monitoring & Optimization will involve setting up logging, monitoring, and alerting for model performance metrics such as drift, latency, and accuracy. Collaborating with DevOps teams to ensure scalability, security, and cost-efficiency of deployed models will also be part of your responsibilities. The ideal candidate will possess 5-7 years of hands-on experience in Deep Learning, NLP, and LLMs. Strong proficiency in Python, PyTorch, TensorFlow, Hugging Face Transformers, and LLM frameworks is essential. Experience with model deployment tools like Docker, Kubernetes, and FastAPI, along with knowledge of MLOps/LLMOps best practices and familiarity with cloud platforms (AWS, GCP, Azure) are required qualifications. Preferred qualifications include contributions to open-source LLM projects, showcasing your commitment to advancing the field of machine learning.,
Posted 2 weeks ago
4.0 - 8.0 years
0 Lacs
karnataka
On-site
At PwC, we focus on leveraging data to drive insights and make informed business decisions in the field of data and analytics. Our team utilises advanced analytics techniques to help clients optimize their operations and achieve their strategic goals. As a Data Analyst at PwC, you will play a crucial role in utilizing advanced analytical techniques to extract insights from large datasets and facilitate data-driven decision-making. Your responsibilities will include leveraging skills in data manipulation, visualization, and statistical modeling to assist clients in solving complex business problems. We are currently seeking a highly skilled MLOps/LLMOps Engineer to join PwC US - Acceleration Center. In this role, you will be responsible for the deployment, scaling, and maintenance of Generative AI models. Working closely with data scientists, ML/GenAI engineers, and DevOps teams, you will ensure seamless integration and operation of GenAI models within production environments at PwC and our clients. The ideal candidate will have a strong background in MLOps practices, coupled with experience and interest in Generative AI technologies. As a candidate, you should have a minimum of 4+ years of hands-on experience. Core qualifications for this role include 3+ years of experience developing and deploying AI models in production environments, with a year of experience in developing proofs of concept and prototypes. Additionally, a strong background in software development, proficiency in programming languages like Python, knowledge of ML frameworks and libraries, familiarity with containerization and orchestration tools, and experience with cloud platforms and CI/CD tools are essential. Key responsibilities of the role involve developing and implementing MLOps strategies tailored for Generative AI models, designing and managing CI/CD pipelines specialized for ML workflows, monitoring and optimizing the performance of AI models in production, collaborating with data scientists and ML researchers, and ensuring compliance with data privacy regulations. You will also be responsible for troubleshooting and resolving issues related to ML model serving, data anomalies, and infrastructure performance. The successful candidate will be proficient in MLOps tools such as MLflow, Kubeflow, Airflow, or similar, have expertise in generative AI frameworks, containerization technologies, MLOps and LLMOps practices, and cloud-based AI services. Nice-to-have qualifications include experience with advanced GenAI applications, familiarity with experiment tracking tools, knowledge of high-performance computing techniques, and contributions to open-source MLOps or GenAI projects. In addition to technical skills, the role requires project delivery capabilities such as designing scalable deployment pipelines for ML/GenAI models, overseeing cloud infrastructure setup, and creating detailed documentation for deployment pipelines. Client engagement is another essential aspect, involving collaboration with clients to understand their business needs, presenting technical approaches and results, conducting training sessions, and creating user guides for clients. To stay ahead in the field, you will need to stay updated with the latest trends in MLOps/LLMOps and Generative AI, apply this knowledge to improve existing systems and processes, develop internal tools and frameworks, mentor junior team members, and contribute to technical publications. The ideal candidate for this position should hold any graduate/BE/B.Tech/MCA/M.Sc/M.E/M.Tech/Masters Degree/MBA. Join us at PwC and be part of a dynamic team driving innovation in data and analytics!,
Posted 2 weeks ago
5.0 - 7.0 years
30 - 45 Lacs
Mumbai, Delhi / NCR, Bengaluru
Work from Office
About the Role We are seeking a highly skilled and experienced Senior AI Engineer to lead the design, development, and implementation of robust and scalable pipelines and backend systems for our Generative AI applications. In this role, you will be responsible for orchestrating the flow of data, integrating AI services, developing RAG pipelines, working with LLMs, and ensuring the smooth operation of the backend infrastructure that powers our Generative AI solutions. You will also be expected to apply modern LLMOps practices, handle schema-constrained generation, optimize cost and latency trade-offs, mitigate hallucinations, and ensure robust safety, personalization, and observability across GenAI systems. Responsibilities Generative AI Pipeline Development Design and implement scalable and modular pipelines for data ingestion, transformation, and orchestration across GenAI workloads. Manage data and model flow across LLMs, embedding services, vector stores, SQL sources, and APIs. Build CI/CD pipelines with integrated prompt regression testing and version control. Use orchestration frameworks like LangChain or LangGraph for tool routing and multi-hop workflows. Monitor system performance using tools like Langfuse or Prometheus. Data and Document Ingestion Develop systems to ingest unstructured (PDF, OCR) and structured (SQL, APIs) data. Apply preprocessing pipelines for text, images, and code. Ensure data integrity, format consistency, and security across sources. AI Service Integration Integrate external and internal LLM APIs (OpenAI, Claude, Mistral, Qwen, etc.). Build internal APIs for smooth backend-AI communication. Optimize performance through fallback routing to classical or smaller models based on latency or cost budgets. Use schema-constrained prompting and output filters to suppress hallucinations and maintain factual accuracy. Retrieval-Augmented Generation (RAG) Pipelines Build hybrid RAG pipelines using vector similarity (FAISS/Qdrant) and structured data (SQL/API). Design custom retrieval strategies for multi-modal or multi-source documents. Apply post-retrieval ranking using DPO or feedback-based techniques. Improve contextual relevance through re-ranking, chunk merging, and scoring logic. LLM Integration and Optimization Manage prompt engineering, model interaction, and tuning workflows. Implement LLMOps best practices: prompt versioning, output validation, caching (KV store), and fallback design. Optimize generation using temperature tuning, token limits, and speculative decoding. Integrate observability and cost-monitoring into LLM workflows. Backend Services Ownership Design and maintain scalable backend services supporting GenAI applications. Implement monitoring, logging, and performance tracing. Build RBAC (Role-Based Access Control) and multi-tenant personalization. Support containerization (Docker, Kubernetes) and autoscaling infrastructure for production. Required Skills and Qualifications Education Bachelors or Masters in Computer Science, Artificial Intelligence, Machine Learning, or related field. Experience 5+ years of experience in AI/ML engineering with end-to-end pipeline development. Hands-on experience building and deploying LLM/RAG systems in production. Strong experience with public cloud platforms (AWS, Azure, or GCP). Technical Skills Proficient in Python and libraries such as Transformers, SentenceTransformers, PyTorch. Deep understanding of GenAI infrastructure, LLM APIs, and toolchains like LangChain/LangGraph. Experience with RESTful API development and version control using Git. Knowledge of vector DBs (Qdrant, FAISS, Weaviate) and similarity-based retrieval. Familiarity with Docker, Kubernetes, and scalable microservice design. Experience with observability tools like Prometheus, Grafana, or Langfuse. Generative AI Specific Skills Knowledge of LLMs, VAEs, Diffusion Models, GANs. Experience building structured + unstructured RAG pipelines. Prompt engineering with safety controls, schema enforcement, and hallucination mitigation. Experience with prompt testing, caching strategies, output filtering, and fallback logic. Familiarity with DPO, RLHF, or other feedback-based fine-tuning methods. Soft Skills Strong analytical, problem-solving, and debugging skills. Excellent collaboration with cross-functional teams: product, QA, and DevOps. Ability to work in fast-paced, agile environments and deliver production-grade solutions. Clear communication and strong documentation practices. Preferred Qualifications Experience with OCR, document parsing, and layout-aware chunking. Hands-on with MLOps and LLMOps tools for Generative AI. Contributions to open-source GenAI or AI infrastructure projects. Knowledge of GenAI governance, ethical deployment, and usage controls. Experience with hallucination suppression frameworks like Guardrails.ai, Rebuff, or Constitutional AI. Shift Time: 2:30 PM to 11:30 PM IST Location-Remote,Delhi NCR,Bangalore,Chennai,Pune,Kolkata,Ahmedabad,Mumbai,Hyderabad
Posted 3 weeks ago
2.0 - 7.0 years
6 - 14 Lacs
Chennai
Hybrid
We're looking for experienced GenAI Engineers in the following areas: - Architect and implement RAG pipelines (Retrieval-Augmented Generation) - Prompt Engineering - Open-source and proprietary LLMs (like LLaMA, Mistral, Claude, GPT-4o, etc.) - Building AI Agents, MCP, A2A Tech Stack [any other similar to this is also ok]: - LangChain, LangFlow, LangGraph, LangSmith, Dify, AWS bedrock, sageworks, GenAI, - Vector DBs [Postgres, Pinecone, etc] - Agents [A2A] SDK [google, others] and any Agentic AI platform - CrewAI, Cursor, Windsurf, etc
Posted 3 weeks ago
5.0 - 10.0 years
22 - 30 Lacs
Pune
Hybrid
We are looking for a Machine Learning Engineer with expertise in MLOps (Machine Learning Operations) or LLMOps (Large Language Model Operations) to design, deploy, and maintain scalable AI/ML systems. You will work on automating ML workflows, optimizing model deployment, and managing large-scale AI applications, including LLMs (Large Language Models) , ensuring they run efficiently in production. Key Responsibilities: Design and implement end-to-end MLOps pipelines for training, validation, deployment, monitoring, and retraining of ML models. Optimize and fine-tune large language models (LLMs) for various applications, ensuring performance and efficiency. Develop CI/CD pipelines for ML models to automate deployment and monitoring in production. Monitor model performance, detect drift , and implement automated retraining mechanisms. Work with cloud platforms ( AWS, GCP, Azure ) and containerization technologies ( Docker, Kubernetes ) for scalable deployments. Implement best practices in data engineering , feature stores, and model versioning. Collaborate with data scientists, engineers, and product teams to integrate ML models into production applications. Ensure compliance with security, privacy, and ethical AI standards in ML deployments. Optimize inference performance and cost of LLMs using quantization, pruning, and distillation techniques . Deploy LLM-based APIs and services, integrating them with real-time and batch processing pipelines. Key Requirements: Technical Skills: Strong programming skills in Python, with experience in ML frameworks ( TensorFlow, PyTorch, Hugging Face, JAX ). Experience with MLOps tools (MLflow, Kubeflow, Vertex AI, SageMaker, Airflow). Deep understanding of LLM architectures , prompt engineering, and fine-tuning. Hands-on experience with containerization (Docker, Kubernetes) and orchestration tools . Proficiency in cloud services (AWS/GCP/Azure) for ML model training and deployment. Experience with monitoring ML models (Prometheus, Grafana, Evidently AI). Knowledge of feature stores (Feast, Tecton) and data pipelines (Kafka, Apache Beam). Strong background in distributed computing (Spark, Ray, Dask) . Soft Skills: Strong problem-solving and debugging skills. Ability to work in cross-functional teams and communicate complex ML concepts to stakeholders. Passion for staying updated with the latest ML and LLM research & technologies . Preferred Qualifications: Experience with LLM fine-tuning , Reinforcement Learning with Human Feedback ( RLHF ), or LoRA/PEFT techniques . Knowledge of vector databases (FAISS, Pinecone, Weaviate) for retrieval-augmented generation ( RAG ). Familiarity with LangChain, LlamaIndex , and other LLMOps-specific frameworks. Experience deploying LLMs in production (ChatGPT, LLaMA, Falcon, Mistral, Claude, etc.) .
Posted 1 month ago
4.0 - 5.0 years
4 - 5 Lacs
Mumbai, Maharashtra, India
On-site
MSCI Services is looking for a Senior AI Engineer to join our AI team. In this role, you will design and develop highly scalable, robust systems that drive our AI initiatives and data operations. Success depends on strong software engineering skills, familiarity with large-scale distributed systems, and expertise in AI technologies. Our ideal candidate has a proven ability to build reliable platforms (rather than standalone applications) and to iterate effectively over multiple release cycles. Throughout your work, you'll apply best practices in software engineering and system architecture. If you're passionate about delivering high-impact AI solutions in a dynamic environment, this is an exciting opportunity to have a substantial influence on our AI capabilities through your expertise. Responsibilities: Design and implement scalable distributed systems . Architect solutions that can handle large volumes of data for real-time and batch processing. Design and develop efficient AI pipelines with automation and reliability across the platform. Integrate agentic workflows and AI agents into data extraction processes, and enable systems to perform multi-step reasoning and tool usage to improve accuracy and efficiency of data extraction. Deploy, monitor, and maintain LLM-based extraction systems in production , ensuring reliability and scalability. Set up appropriate monitoring, logging, and evaluation metrics to track performance, and perform continual tuning and improvements based on human-in-the-loop feedback. Conduct applied research and experimentation with the latest generative AI models and techniques to enhance extraction capabilities. Prototype new approaches and iterate quickly to integrate successful methods into the production pipeline. Collaborate with cross-functional teams (data engineers, product managers, domain experts) to gather requirements and align AI solutions with business needs. Qualifications: Experience in applied AI or machine learning engineering , with a track record of building and deploying AI solutions (especially in NLP). Hands-on experience with using Generative AI models and APIs/frameworks (e.g., OpenAI GPT-4, Google Gemini). Ability to build Agentic AI systems where LLMs interact with tools or perform multi-step workflows. Proficiency in Python (preferred) and experience deploying machine learning models or pipelines at scale. Good understanding of embeddings, LLM models , and experience with retrieval-augmented generation (RAG) workflows to incorporate external knowledge into LLM-based systems. Knowledge of LLMOps and cloud services (Azure, GCP, or similar) for deploying and managing AI solutions. Experience with containerization, orchestration, and monitoring of ML models in a production cloud environment. Excellent collaboration and communication skills , with the ability to work effectively in a team, translate complex technical concepts to non-technical stakeholders, and document work clearly.
Posted 1 month ago
1.0 - 4.0 years
1 - 4 Lacs
Pune, Maharashtra, India
On-site
MSCI Services is looking for a Senior AI Engineer to join our AI team. In this role, you will design and develop highly scalable, robust systems that drive our AI initiatives and data operations. Success depends on strong software engineering skills, familiarity with large-scale distributed systems, and expertise in AI technologies. Our ideal candidate has a proven ability to build reliable platforms (rather than standalone applications) and to iterate effectively over multiple release cycles. Throughout your work, you'll apply best practices in software engineering and system architecture. If you're passionate about delivering high-impact AI solutions in a dynamic environment, this is an exciting opportunity to have a substantial influence on our AI capabilities through your expertise. Responsibilities: Design and implement scalable distributed systems . Architect solutions that can handle large volumes of data for real-time and batch processing. Design and develop efficient AI pipelines with automation and reliability across the platform. Integrate agentic workflows and AI agents into data extraction processes, and enable systems to perform multi-step reasoning and tool usage to improve accuracy and efficiency of data extraction. Deploy, monitor, and maintain LLM-based extraction systems in production , ensuring reliability and scalability. Set up appropriate monitoring, logging, and evaluation metrics to track performance, and perform continual tuning and improvements based on human-in-the-loop feedback. Conduct applied research and experimentation with the latest generative AI models and techniques to enhance extraction capabilities. Prototype new approaches and iterate quickly to integrate successful methods into the production pipeline. Collaborate with cross-functional teams (data engineers, product managers, domain experts) to gather requirements and align AI solutions with business needs. Qualifications: Experience in applied AI or machine learning engineering , with a track record of building and deploying AI solutions (especially in NLP). Hands-on experience with using Generative AI models and APIs/frameworks (e.g., OpenAI GPT-4, Google Gemini). Ability to build Agentic AI systems where LLMs interact with tools or perform multi-step workflows. Proficiency in Python (preferred) and experience deploying machine learning models or pipelines at scale. Good understanding of embeddings, LLM models , and experience with retrieval-augmented generation (RAG) workflows to incorporate external knowledge into LLM-based systems. Knowledge of LLMOps and cloud services (Azure, GCP, or similar) for deploying and managing AI solutions. Experience with containerization, orchestration, and monitoring of ML models in a production cloud environment. Excellent collaboration and communication skills , with the ability to work effectively in a team, translate complex technical concepts to non-technical stakeholders, and document work clearly.
Posted 1 month ago
8.0 - 12.0 years
1 - 2 Lacs
Bengaluru, Karnataka, India
On-site
The Implementation Technical Architect will be responsible for designing, developing, and deploying cutting-edge Generative AI (GenAI) solutions using the latest Large Language Models (LLMs) and frameworks. This role requires deep expertise in Python programming, cloud platforms (Azure, GCP, AWS), and advanced AI techniques such as fine-tuning, LLMOps, and Responsible AI. The architect will lead the development of scalable, secure, and efficient GenAI applications, ensuring alignment with business goals and technical requirements. Key Responsibilities: Design and Architecture: Create scalable and modular architecture for GenAI applications using frameworks like Autogen, Crew.ai, LangGraph, LlamaIndex, and LangChain. Python Development: Lead the development of Python-based GenAI applications, ensuring high-quality, maintainable, and efficient code. Data Curation Automation: Build tools and pipelines for automated data curation, preprocessing, and augmentation to support LLM training and fine-tuning. Cloud Integration: Design and implement solutions leveraging Azure, GCP, and AWS LLM ecosystems, ensuring seamless integration with existing cloud infrastructure. Fine-Tuning Expertise: Apply advanced fine-tuning techniques such as PEFT, QLoRA, and LoRA to optimize LLM performance for specific use cases. LLMOps Implementation: Establish and manage LLMOps pipelines for continuous integration, deployment, and monitoring of LLM-based applications. Responsible AI: Ensure ethical AI practices by implementing Responsible AI principles, including fairness, transparency, and accountability. RLHF and RAG: Implement Reinforcement Learning with Human Feedback (RLHF) and Retrieval-Augmented Generation (RAG) techniques to enhance model performance. Modular RAG Design: Develop and optimize Modular RAG architectures for complex GenAI applications. Open-Source Collaboration: Leverage Hugging Face and other open-source platforms for model development, fine-tuning, and deployment. Front-End Integration: Collaborate with front-end developers to integrate GenAI capabilities into user-friendly interfaces. SDLC and DevSecOps: Implement secure software development lifecycle (SDLC) and DevSecOps practices tailored to LLM-based projects. Technical Documentation: Create detailed design artifacts, technical specifications, and architecture diagrams for complex projects. Stakeholder Collaboration: Work closely with cross-functional teams, including data scientists, engineers, and product managers, to deliver high-impact solutions. Mentorship and Leadership: Guide and mentor junior developers and engineers, fostering a culture of innovation and technical excellence. Required Skills: Python Programming: Deep expertise in Python for building GenAI applications and automation tools. LLM Frameworks: Proficiency in frameworks like Autogen, Crew.ai, LangGraph, LlamaIndex, and LangChain. Large-Scale Data Handling & Architecture: Design and implement architectures for handling large-scale structured and unstructured data. Multi-Modal LLM Applications: Familiarity with text chat completion, vision, and speech models. Fine-tune SLM(Small Language Model) for domain specific data and use cases. Prompt injection fallback and RCE tools such as Pyrit and HAX toolkit etc. Anti-hallucination and anti-gibberish tools such as Bleu etc. Cloud Platforms: Extensive experience with Azure, GCP, and AWS LLM ecosystems and APIs. Fine-Tuning Techniques: Mastery of PEFT, QLoRA, LoRA, and other fine-tuning methods. LLMOps: Strong knowledge of LLMOps practices for model deployment, monitoring, and management. Responsible AI: Expertise in implementing ethical AI practices and ensuring compliance with regulations. RLHF and RAG: Advanced skills in Reinforcement Learning with Human Feedback and Retrieval-Augmented Generation. Modular RAG: Deep understanding of Modular RAG architectures and their implementation. Hugging Face: Proficiency in using Hugging Face and similar open-source platforms for model development. Front-End Integration: Knowledge of front-end technologies to enable seamless integration of GenAI capabilities. SDLC and DevSecOps: Strong understanding of secure software development lifecycle and DevSecOps practices for LLMs. Data Curation: Expertise in building automated data curation and preprocessing pipelines. API Development: Experience in designing and implementing APIs for GenAI applications. Technical Documentation: Ability to create clear and comprehensive design artifacts and technical documentation. Leadership and Mentorship: Proven ability to lead teams, mentor junior developers, and drive technical innovation.
Posted 1 month ago
8.0 - 12.0 years
1 - 2 Lacs
Hyderabad, Telangana, India
On-site
The Implementation Technical Architect will be responsible for designing, developing, and deploying cutting-edge Generative AI (GenAI) solutions using the latest Large Language Models (LLMs) and frameworks. This role requires deep expertise in Python programming, cloud platforms (Azure, GCP, AWS), and advanced AI techniques such as fine-tuning, LLMOps, and Responsible AI. The architect will lead the development of scalable, secure, and efficient GenAI applications, ensuring alignment with business goals and technical requirements. Key Responsibilities: Design and Architecture: Create scalable and modular architecture for GenAI applications using frameworks like Autogen, Crew.ai, LangGraph, LlamaIndex, and LangChain. Python Development: Lead the development of Python-based GenAI applications, ensuring high-quality, maintainable, and efficient code. Data Curation Automation: Build tools and pipelines for automated data curation, preprocessing, and augmentation to support LLM training and fine-tuning. Cloud Integration: Design and implement solutions leveraging Azure, GCP, and AWS LLM ecosystems, ensuring seamless integration with existing cloud infrastructure. Fine-Tuning Expertise: Apply advanced fine-tuning techniques such as PEFT, QLoRA, and LoRA to optimize LLM performance for specific use cases. LLMOps Implementation: Establish and manage LLMOps pipelines for continuous integration, deployment, and monitoring of LLM-based applications. Responsible AI: Ensure ethical AI practices by implementing Responsible AI principles, including fairness, transparency, and accountability. RLHF and RAG: Implement Reinforcement Learning with Human Feedback (RLHF) and Retrieval-Augmented Generation (RAG) techniques to enhance model performance. Modular RAG Design: Develop and optimize Modular RAG architectures for complex GenAI applications. Open-Source Collaboration: Leverage Hugging Face and other open-source platforms for model development, fine-tuning, and deployment. Front-End Integration: Collaborate with front-end developers to integrate GenAI capabilities into user-friendly interfaces. SDLC and DevSecOps: Implement secure software development lifecycle (SDLC) and DevSecOps practices tailored to LLM-based projects. Technical Documentation: Create detailed design artifacts, technical specifications, and architecture diagrams for complex projects. Stakeholder Collaboration: Work closely with cross-functional teams, including data scientists, engineers, and product managers, to deliver high-impact solutions. Mentorship and Leadership: Guide and mentor junior developers and engineers, fostering a culture of innovation and technical excellence. Required Skills: Python Programming: Deep expertise in Python for building GenAI applications and automation tools. LLM Frameworks: Proficiency in frameworks like Autogen, Crew.ai, LangGraph, LlamaIndex, and LangChain. Large-Scale Data Handling & Architecture: Design and implement architectures for handling large-scale structured and unstructured data. Multi-Modal LLM Applications: Familiarity with text chat completion, vision, and speech models. Fine-tune SLM(Small Language Model) for domain specific data and use cases. Prompt injection fallback and RCE tools such as Pyrit and HAX toolkit etc. Anti-hallucination and anti-gibberish tools such as Bleu etc. Cloud Platforms: Extensive experience with Azure, GCP, and AWS LLM ecosystems and APIs. Fine-Tuning Techniques: Mastery of PEFT, QLoRA, LoRA, and other fine-tuning methods. LLMOps: Strong knowledge of LLMOps practices for model deployment, monitoring, and management. Responsible AI: Expertise in implementing ethical AI practices and ensuring compliance with regulations. RLHF and RAG: Advanced skills in Reinforcement Learning with Human Feedback and Retrieval-Augmented Generation. Modular RAG: Deep understanding of Modular RAG architectures and their implementation. Hugging Face: Proficiency in using Hugging Face and similar open-source platforms for model development. Front-End Integration: Knowledge of front-end technologies to enable seamless integration of GenAI capabilities. SDLC and DevSecOps: Strong understanding of secure software development lifecycle and DevSecOps practices for LLMs. Data Curation: Expertise in building automated data curation and preprocessing pipelines. API Development: Experience in designing and implementing APIs for GenAI applications. Technical Documentation: Ability to create clear and comprehensive design artifacts and technical documentation. Leadership and Mentorship: Proven ability to lead teams, mentor junior developers, and drive technical innovation.
Posted 1 month ago
7.0 - 12.0 years
15 - 25 Lacs
Bengaluru
Work from Office
Role: Python with Gen AI Developer Exp: 7+yrs Budget: Max 28 LPA Location: Bangalore (Hybrid) Must Have: Python libraries and frameworks,Generative AI, Prompt Engineering. Immediate to 30 Days. Ragul 8428065584
Posted 1 month ago
5.0 - 10.0 years
30 - 45 Lacs
Hyderabad, Bengaluru, Delhi / NCR
Work from Office
About the Role We are seeking a highly skilled and experienced Senior AI Engineer to lead the design, development, and implementation of robust and scalable pipelines and backend systems for our Generative AI applications. In this role, you will be responsible for orchestrating the flow of data, integrating AI services, developing RAG pipelines, working with LLMs, and ensuring the smooth operation of the backend infrastructure that powers our Generative AI solutions. You will also be expected to apply modern LLMOps practices, handle schema-constrained generation, optimize cost and latency trade-offs, mitigate hallucinations, and ensure robust safety, personalization, and observability across GenAI systems. Responsibilities Generative AI Pipeline Development Design and implement scalable and modular pipelines for data ingestion, transformation, and orchestration across GenAI workloads. Manage data and model flow across LLMs, embedding services, vector stores, SQL sources, and APIs. Build CI/CD pipelines with integrated prompt regression testing and version control. Use orchestration frameworks like LangChain or LangGraph for tool routing and multi-hop workflows. Monitor system performance using tools like Langfuse or Prometheus. Data and Document Ingestion Develop systems to ingest unstructured (PDF, OCR) and structured (SQL, APIs) data. Apply preprocessing pipelines for text, images, and code. Ensure data integrity, format consistency, and security across sources. AI Service Integration Integrate external and internal LLM APIs (OpenAI, Claude, Mistral, Qwen, etc.). Build internal APIs for smooth backend-AI communication. Optimize performance through fallback routing to classical or smaller models based on latency or cost budgets. Use schema-constrained prompting and output filters to suppress hallucinations and maintain factual accuracy. Retrieval-Augmented Generation (RAG) Pipelines Build hybrid RAG pipelines using vector similarity (FAISS/Qdrant) and structured data (SQL/API). Design custom retrieval strategies for multi-modal or multi-source documents. Apply post-retrieval ranking using DPO or feedback-based techniques. Improve contextual relevance through re-ranking, chunk merging, and scoring logic. LLM Integration and Optimization Manage prompt engineering, model interaction, and tuning workflows. Implement LLMOps best practices: prompt versioning, output validation, caching (KV store), and fallback design. Optimize generation using temperature tuning, token limits, and speculative decoding. Integrate observability and cost-monitoring into LLM workflows. Backend Services Ownership Design and maintain scalable backend services supporting GenAI applications. Implement monitoring, logging, and performance tracing. Build RBAC (Role-Based Access Control) and multi-tenant personalization. Support containerization (Docker, Kubernetes) and autoscaling infrastructure for production. Required Skills and Qualifications Education Bachelors or Masters in Computer Science, Artificial Intelligence, Machine Learning, or related field. Experience 5+ years of experience in AI/ML engineering with end-to-end pipeline development. Hands-on experience building and deploying LLM/RAG systems in production. Strong experience with public cloud platforms (AWS, Azure, or GCP). Technical Skills Proficient in Python and libraries such as Transformers, SentenceTransformers, PyTorch. Deep understanding of GenAI infrastructure, LLM APIs, and toolchains like LangChain/LangGraph. Experience with RESTful API development and version control using Git. Knowledge of vector DBs (Qdrant, FAISS, Weaviate) and similarity-based retrieval. Familiarity with Docker, Kubernetes, and scalable microservice design. Experience with observability tools like Prometheus, Grafana, or Langfuse. Generative AI Specific Skills Knowledge of LLMs, VAEs, Diffusion Models, GANs. Experience building structured + unstructured RAG pipelines. Prompt engineering with safety controls, schema enforcement, and hallucination mitigation. Experience with prompt testing, caching strategies, output filtering, and fallback logic. Familiarity with DPO, RLHF, or other feedback-based fine-tuning methods. Soft Skills Strong analytical, problem-solving, and debugging skills. Excellent collaboration with cross-functional teams: product, QA, and DevOps. Ability to work in fast-paced, agile environments and deliver production-grade solutions. Clear communication and strong documentation practices. Preferred Qualifications Experience with OCR, document parsing, and layout-aware chunking. Hands-on with MLOps and LLMOps tools for Generative AI. Contributions to open-source GenAI or AI infrastructure projects. Knowledge of GenAI governance, ethical deployment, and usage controls. Experience with hallucination suppression frameworks like Guardrails.ai, Rebuff, or Constitutional AI. Experience and Shift Shift Time: 2:30 PM to 11:30 PM IST Location: Remote- Bengaluru,Hyderabad,Delhi / NCR,Chennai,Pune,Kolkata,Ahmedabad,Mumbai
Posted 1 month ago
8.0 - 10.0 years
10 - 15 Lacs
Noida
Work from Office
Mandatory Skills & Experience : - Expertise in designing and optimizing machine-learning operations, with a preference for LLM Ops. - Proficient in Data Science, Machine Learning, Python, SQL, Linux/Unix shell scripting. - Experience on Large Language Models and Natural Language Processing (NLP), and experience with researching, training, and fine-tuning LLMs. - Contribute towards fine-tune Transformer models for optimal performance in NLP tasks, if required. - Implement and maintain automated testing and deployment processes for machine learning models w.r.t LLMOps. - Implement version control, CI/CD pipelines, and containerization techniques to streamline ML and LLM workflows. - Develop and maintain robust monitoring and alerting systems for generative AI models ensuring proactive identification and resolution of issues. - Research or engineering experience in deep learning with one or more of the following : generative models, segmentation, object detection, classification, model optimisations. - Experience implementing RAG frameworks as part of available-ready products. - Experience in setting up the infrastructure for the latest technology such as Kubernetes, Serverless, Containers, Microservices etc. - Experience in scripting/programming to automate deployments and testing, working on tools like Terraform and Ansible. - Scripting languages like Python, bash, YAML etc. - Experience on CI/CD opensource and enterprise tool sets such as Argo CD, and Jenkins (others like Jenkins X, Circle CI, Argo CD, Tekton, Travis, Concourse an advantage). - Experience with the GitHub/DevOps Lifecycle. - Experience in Observability solutions (Prometheus, EFK stacks, ELK stacks, Grafana, Dynatrace, AppDynamics). - Experience in at-least one of the clouds for example Azure/AWS/GCP. - Significant experience on microservices-based, container-based or similar modern approaches of applications and workloads. - You have exemplary verbal and written communication skills (English). - Able to interact and influence at the highest level, you will be a confident presenter and speaker, able to command the respect of your audience. Desired Skills & Experience : - Bachelor level technical degree or equivalent experience; Computer Science, Data Science, or Engineering background preferred; Master's Degree desired. - Experience in LLM Ops or related areas, such as DevOps, data engineering, or ML infrastructure. - Hands-on experience in deploying and managing machine learning and large language model pipelines in cloud platforms (i.e., AWS, Azure) for ML workloads. - Familiar with data science, machine learning, deep learning, and natural language processing concepts, tools, and libraries such as Python, TensorFlow, PyTorch, NLTK etc. - Experience in using retrieval augmented generation and prompt engineering techniques to improve the model's quality and diversity to improve operations efficiency. - Proven experience in developing and fine-tuning Language Models (LLMs). - Stay up-to-date with the latest advancements in Generative AI, conduct research, and explore innovative techniques to improve model quality and efficiency. - The perfect candidate will already be working within a System Integrator, Consulting or Enterprise organisation with 8+ years of experience in a technical role within the Cloud domain. - Deep understanding of core practices including SRE, Agile, Scrum, XP and Domain Driven Design. - Familiarity with the CNCF open-source community. - Enjoy working in a fast-paced and dynamic environment using the latest technologies.
Posted 1 month ago
10.0 - 12.0 years
0 - 33 Lacs
Mumbai, Maharashtra, India
On-site
Job Description Summary role description: Hiring for a Solution Architect for an InsurTech platform provider, Life and Health Insurance. Company description: Our client is a VC-funded InsurTech platform company, providing software platforms for Life Insurance and Health Insurance companies across the globe. Leveraging their domain expertise, regulatory knowledge and technology experience, they architect innovative products and disrupt the Insurance value chain from Customer Acquisition to Engagement. Their products serve customers across the APAC region. Role details: Title / Designation : Solutions Architect Location: Pune/Mumbai Work Mode: Work from office Role & responsibilities: Define and evolve AI/ML architecture roadmap for FWA, IDP, and Agentic AI frameworks. Lead technical presentations and solution design sessions with customers. Design scalable architectures for multi-agent systems and autonomous decision-making. Drive innovation by evaluating emerging AI/ML technologies, especially AI agents. Architect cloud-native platforms supporting the complete AI/ML lifecycle. Provide technical leadership across product development and customer implementation. Collaborate with data scientists, engineers, and business stakeholders. Stay at the forefront of AI/ML innovations, particularly autonomous agents and LLMs. Establish and enforce technical standards and architectural guidelines. Candidate requirements: 10+ years in software architecture/system design in insurance domain, with 5+ years in AI/ML systems/platforms. Proven experience delivering large-scale AI/ML solutions, preferably with autonomous agents. Experience with cloud-native architectures (AWS, Azure, GCP), containerization (Docker, Kubernetes), and microservices. Deep expertise in AI/ML system architecture (model serving, MLOps/LLMOps pipelines, distributed computing). Strong understanding of Agentic AI, multi-agent systems, and LLMs (including LoRA, PEFT fine-tuning). Bachelor's or Master's in CS, SE, Data Science, or related technical field. Exceptional technical leadership and communication skills. Selection process: Interview with Senior Solution Architect Interview with CTO HR Discussion Check Your Resume for Match Upload your resume and our tool will compare it to the requirements for this job like recruiters do.
Posted 1 month ago
8.0 - 13.0 years
15 - 25 Lacs
Noida, Hyderabad, Bengaluru
Work from Office
Job Description : As an LLMOps Engineer, you will play a crucial role in the deployment, maintenance, and optimization of large language models (LLMs). Your responsibilities will span the entire lifecycle of LLMs, from initial deployment to ongoing operations, ensuring optimal performance, scalability, and reliability. Key Responsibilities : LLM Deployment and Integration - Deploy and integrate large language models into production environments, ensuring seamless integration with existing systems and applications. Infrastructure Planning and Scaling - Collaborate with cross-functional teams to plan and design the infrastructure required for LLM deployment. Implement scalable solutions to accommodate growing data volumes and user loads. Automation of Deployment Processes - Develop and maintain automation scripts and tools for efficient deployment, scaling, and versioning of LLMs. Streamline deployment processes to minimize downtime. Continuous Monitoring and Alerting - Implement monitoring systems to track LLM performance metrics. Set up alerts for potential issues and respond promptly to ensure uninterrupted service. Performance Monitoring and Optimization - Monitor the performance of LLMs in real-time, conduct regular assessments, and implement optimizations to enhance efficiency and responsiveness. Fault Tolerance and Disaster Recovery - Design and implement fault-tolerant systems for LLMs, incorporating strategies such as redundancy, sharding, and replication. Develop and maintain disaster recovery plans. Security Measures Implementation - Implement robust security measures to safeguard LLMs and associated data. Ensure compliance with data security regulations and industry standards. Collaboration with NLP Engineers and Data Scientists - Collaborate with NLP (Natural Language Processing) engineers and data scientists to understand model requirements and implement necessary infrastructure adjustments. Skills & Tools Infrastructure as Code (IaC) - Experience with IaC tools such as Terraform or Ansible for automating infrastructure provisioning. Containerization and Orchestration - Proficiency in containerization technologies (e.g., Docker) and orchestration tools (e.g., Kubernetes) for managing LLM deployments. Cloud Platforms - Familiarity with cloud platforms such as AWS (Bedrock), Azure, or GCP, and experience in deploying and managing applications in a cloud environment. Monitoring and Logging Tools - Knowledge of monitoring tools (e.g., Prometheus, Grafana) and logging systems (e.g., ELK stack) for real-time performance monitoring and analysis. Security Measures - Understanding of security guardrails using tools like LLM Guard and familiarity of how to mask / redact / obfuscate sensitive data, protect the input and output of toxic and harmful content to / from LLMs and understand the performance implications of the same. Scripting and Automation - Proficient in scripting languages such as Python, Shell, or similar, and experience in automating deployment and maintenance processes.
Posted 1 month ago
Upload Resume
Drag or click to upload
Your data is secure with us, protected by advanced encryption.
Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.
We have sent an OTP to your contact. Please enter it below to verify.
Accenture
39581 Jobs | Dublin
Wipro
19070 Jobs | Bengaluru
Accenture in India
14409 Jobs | Dublin 2
EY
14248 Jobs | London
Uplers
10536 Jobs | Ahmedabad
Amazon
10262 Jobs | Seattle,WA
IBM
9120 Jobs | Armonk
Oracle
8925 Jobs | Redwood City
Capgemini
7500 Jobs | Paris,France
Virtusa
7132 Jobs | Southborough