Home
Jobs
Companies
Resume

18 Pinecone Jobs

Filter
Filter Interviews
Min: 0 years
Max: 25 years
Min: ₹0
Max: ₹10000000
Setup a job Alert
JobPe aggregates results for easy application access, but you actually apply on the job portal directly.

3.0 - 6.0 years

11 - 20 Lacs

Jaipur, Jodhpur

Work from Office

Naukri logo

Key Responsibilities Design, develop, and deploy AI solutions that span both traditional ML models and GenAI-based systems . Build machine learning pipelines using algorithms like linear/logistic regression, decision trees, SVMs, random forests, XGBoost , clustering (K-means, DBSCAN), and time series forecasting . Analyze datasets to derive meaningful insights and build predictive models to solve business problems. Work on GenAI applications using LLMs, including prompt engineering, fine-tuning, and retrieval-augmented generation (RAG). Develop and integrate LLM-based features using frameworks like LangChain, Hugging Face Transformers, or OpenAI API. Collaborate with data, product, and engineering teams to define and implement AI-driven functionalities . Apply statistical modeling and inference techniques for feature selection, model evaluation, and data exploration. Optimize performance of ML and GenAI models through hyperparameter tuning, cross-validation , and error analysis. Design and maintain data pipelines and ML workflows using tools like Airflow, DVC , or MLflow . Deploy models into production with appropriate MLOps practices, ensuring monitoring, retraining , and version control . Research and evaluate advancements in both traditional ML and LLM-based AI . Skills & Experience 3-6 years of experience in AI/ML using Python. Proficient in machine learning algorithms (classification, regression, clustering, dimensionality reduction, ensemble methods). Hands-on experience with GenAI/LLM applications (prompt design, RAG, fine-tuning, etc.). Familiarity with data preprocessing, feature engineering , and working with structured and unstructured data . Proficient in Python ML libraries : Scikit-learn, XGBoost, LightGBM, Pandas, NumPy. Experience with deep learning frameworks : PyTorch or TensorFlow. Familiarity with vector databases (e.g., FAISS, Pinecone) and LLM orchestration tools (LangChain, Hugging Face). Experience in model evaluation techniques , including AUC-ROC, precision-recall, RMSE, etc. Familiarity with cloud AI services (AWS SageMaker, GCP AI Platform, or Azure ML). Solid understanding of MLOps tools : MLflow, Docker, Git, CI/CD pipelines. Strong analytical, communication, and collaboration skills.

Posted 13 hours ago

Apply

1.0 - 3.0 years

3 - 5 Lacs

New Delhi, Chennai, Bengaluru

Hybrid

Naukri logo

Your day at NTT DATA We are seeking an experienced Data Engineer to join our team in delivering cutting-edge Generative AI (GenAI) solutions to clients. The successful candidate will be responsible for designing, developing, and deploying data pipelines and architectures that support the training, fine-tuning, and deployment of LLMs for various industries. This role requires strong technical expertise in data engineering, problem-solving skills, and the ability to work effectively with clients and internal teams. What youll be doing Key Responsibilities: Design, develop, and manage data pipelines and architectures to support GenAI model training, fine-tuning, and deployment Data Ingestion and Integration: Develop data ingestion frameworks to collect data from various sources, transform, and integrate it into a unified data platform for GenAI model training and deployment. GenAI Model Integration: Collaborate with data scientists to integrate GenAI models into production-ready applications, ensuring seamless model deployment, monitoring, and maintenance. Cloud Infrastructure Management: Design, implement, and manage cloud-based data infrastructure (e.g., AWS, GCP, Azure) to support large-scale GenAI workloads, ensuring cost-effectiveness, security, and compliance. Write scalable, readable, and maintainable code using object-oriented programming concepts in languages like Python, and utilize libraries like Hugging Face Transformers, PyTorch, or TensorFlow Performance Optimization: Optimize data pipelines, GenAI model performance, and infrastructure for scalability, efficiency, and cost-effectiveness. Data Security and Compliance: Ensure data security, privacy, and compliance with regulatory requirements (e.g., GDPR, HIPAA) across data pipelines and GenAI applications. Client Collaboration: Collaborate with clients to understand their GenAI needs, design solutions, and deliver high-quality data engineering services. Innovation and R&D: Stay up to date with the latest GenAI trends, technologies, and innovations, applying research and development skills to improve data engineering services. Knowledge Sharing: Share knowledge, best practices, and expertise with team members, contributing to the growth and development of the team. Bachelors degree in computer science, Engineering, or related fields (Masters recommended) Experience with vector databases (e.g., Pinecone, Weaviate, Faiss, Annoy) for efficient similarity search and storage of dense vectors in GenAI applications 5+ years of experience in data engineering, with a strong emphasis on cloud environments (AWS, GCP, Azure, or Cloud Native platforms) Proficiency in programming languages like SQL, Python, and PySpark Strong data architecture, data modeling, and data governance skills Experience with Big Data Platforms (Hadoop, Databricks, Hive, Kafka, Apache Iceberg), Data Warehouses (Teradata, Snowflake, BigQuery), and lakehouses (Delta Lake, Apache Hudi) Knowledge of DevOps practices, including Git workflows and CI/CD pipelines (Azure DevOps, Jenkins, GitHub Actions) Experience with GenAI frameworks and tools (e.g., TensorFlow, PyTorch, Keras) Nice to have: Experience with containerization and orchestration tools like Docker and Kubernetes Integrate vector databases and implement similarity search techniques, with a focus on GraphRAG is a plus Familiarity with API gateway and service mesh architectures Experience with low latency/streaming, batch, and micro-batch processing Familiarity with Linux-based operating systems and REST APIs

Posted 6 days ago

Apply

3.0 - 8.0 years

15 - 30 Lacs

Hyderabad, Chennai, Bengaluru

Hybrid

Naukri logo

Job Description: We are seeking a highly skilled and passionate AI/ML Engineer with strong expertise in Generative AI and Large Language Models (LLMs) . The ideal candidate will have hands-on experience in building, fine-tuning, and deploying agentic AI systems using modern GenAI frameworks. You will work on cutting-edge projects involving prompt engineering , RAG pipelines , and memory architectures such as vector databases. Responsibilities: Design and implement AI/ML solutions using modern LLM architectures and agentic AI concepts . Build and optimize intelligent agents using frameworks such as LangChain, AutoGen, CrewAI , or Semantic Kernel . Develop and fine-tune generative AI models with Transformers , HuggingFace , OpenAI API , etc. Implement and enhance Retrieval-Augmented Generation (RAG) pipelines and memory systems like vector databases (e.g., FAISS, Pinecone). Write high-performance Python code to support experimentation, model integration, and API interactions. Collaborate cross-functionally with product, design, and engineering teams in an agile development environment. Deploy AI solutions on cloud platforms (AWS, Azure, or GCP) with a focus on scalability and performance. Stay updated with the latest advancements in the AI/ML/GenAI space. Required Experience: 3 to 8 years of experience in AI/ML , with at least 1 year in Generative AI / LLM-based projects . Proven expertise in Python programming and related libraries for ML/GenAI. Hands-on experience with one or more GenAI frameworks (LangChain, AutoGen, etc.). Solid understanding of prompt engineering , RAG , vector DBs , and agent-based systems . Cloud deployment experience (AWS, Azure, or GCP) is a must. Strong analytical and problem-solving skills.

Posted 1 week ago

Apply

7.0 - 10.0 years

20 - 30 Lacs

Bangalore Rural, Bengaluru

Work from Office

Naukri logo

"We're Hiring For Generative Ai Engineer Role at Bangalore Location" Position: Generative Ai Engineer Experience: 7+ Years Location: Bangalore Responsibilities Develop and deploy scalable big data and AI solutions using Databricks and Azure. Implement RAG pipelines and integrate GenAI APIs for document and image retrieval. Perform finetuning of LLM models and manage prompt engineering workflows. Ensure the end-to-end solution is optimized for performance and scalability. Collaborate with software and ML teams for integration of AI capabilities. Deploy GenAI projects in production in Azure and Databricks. (Must) Required Qualifications Bachelors or Masters degree in Computer Science or related field. 7+ years of experience in big data and AI development. Strong experience with Databricks, Pyspark, and Python. Experience in deploying GenAI projects in production. Preferred Qualifications Familiarity with RAG architecture and document-based retrieval systems. Experience with Azure OpenAI, LangChain, Pinecone or similar tools. Experience with deploying LLM, VLM Models in Cloud Experience with MLOps tools such as MLFlow or similar tools. Technologies Used Python, Pyspark, Databricks, Azure ML, LangChain, OpenAI API, Pinecone, MLFlow More information: +91 73597 10155 | rushit@tekpillar.com

Posted 1 week ago

Apply

10.0 - 20.0 years

20 - 30 Lacs

Bengaluru

Work from Office

Naukri logo

Job Title: ML Prompt Engineer Location - Bangalore Hybrid . Job Description: Principle Developer - ML/Prompt Engineer Technologies: Amazon Bedrock, RAG Models, Java, Python, C or C++, AWS Lambda Responsibilities: Responsible for developing, deploying, and maintaining a Retrieval Augmented Generation (RAG) model in Amazon Bedrock, our cloud-based platform for building and scaling generative AI applications. Design and implement a RAG model that can generate natural language responses, commands, and actions based on user queries and context, using the Anthropic Claude model as the backbone. Integrate the RAG model with Amazon Bedrock, our platform that offers a choice of high-performing foundation models from leading AI companies and Amazon via a single API, along with a broad set of capabilities to build generative AI applications with security, privacy, and responsible AI. Optimize the RAG model for performance, scalability, and reliability, using best practices and robust engineering methodologies. Design, test, and optimize prompts to improve performance, accuracy, and alignment of large language models across diverse use cases. Develop and maintain reusable prompt templates, chains, and libraries to support scalable and consistent GenAI applications. Skills/Qualifications: Experience in programming with at least one software language, such as Java, Python, or C/C++. Experience in working with generative AI tools, models, and frameworks, such as Anthropic, OpenAI, Hugging Face, TensorFlow, PyTorch, or Jupyter. Experience in working with RAG models or similar architectures, such as RAG, Ragna, or Pinecone. Experience in working with Amazon Bedrock or similar platforms, such as AWS Lambda, Amazon SageMaker, or Amazon Comprehend. Ability to design, iterate, and optimize prompts for various LLM use cases (e.g., summarization, classification, translation, Q&A, and agent workflows). Deep understanding of prompt engineering techniques (zero-shot, few-shot, chain-of-thought, etc.) and their effect on model behavior. Familiarity with prompt evaluation strategies, including manual review, automatic metrics, and A/B testing frameworks. Experience building prompt libraries, reusable templates, and structured prompt workflows for scalable GenAI applications. Ability to debug and refine prompts to improve accuracy, safety, and alignment with business objectives. Awareness of prompt injection risks and experience implementing mitigation strategies. Familiarity with prompt tuning, parameter-efficient fine-tuning (PEFT), and prompt chaining methods. Familiarity with continuous deployment and DevOps tools preferred. Experience with Git preferred Experience working in agile/scrum environments Successful track record interfacing and communicating effectively across cross-functional teams. Good communication, analytical and presentation skills, problem-solving skills and learning attitude

Posted 2 weeks ago

Apply

10.0 - 20.0 years

37 - 45 Lacs

Chandigarh

Remote

Naukri logo

Job Title: AI/ML and Chatbot Lead Experience Level: 10+ Years (Lead/Architect level) Location: Remote Employment Type: Full-time No. of Positions: 1 Job Overview: We are seeking a visionary and hands-on AI/ML and Chatbot Lead to spearhead the design, development, and deployment of enterprise-wide Conversational and Generative AI solutions. This role will establish and scale our AI Lab function, define chatbot and multimodal AI strategies, and deliver intelligent automation solutions that enhance user engagement and operational efficiency. Key Responsibilities Define and lead the enterprise-wide strategy for Conversational AI, Multimodal AI, and Large Language Models (LLMs). Build an AI/Chatbot Lab , creating a roadmap and driving innovations across in-app, generative, and conversational AI. Architect scalable AI/ML systems including presentation, orchestration, AI, and data layers. Collaborate with business stakeholders to assess needs, conduct ROI analyses, and deliver impactful AI use cases. Identify and implement agentic AI capabilities and SaaS optimization opportunities. Deliver POCs, pilots, and MVPs owning the design, development, and deployment lifecycle. Lead, mentor, and scale a high-performing team of AI/ML engineers and chatbot developers . Build multi-turn, memory-aware conversations using frameworks like LangChain or Semantic Kernel . Integrate bots with platforms like Salesforce, NetSuite, Slack , and custom applications via APIs/webhooks. Implement and monitor chatbot KPIs using tools like Kibana , Grafana , and custom dashboards. Champion ethical AI , governance, and data privacy/security best practices. Must-Have Skills 10+ years in AI/ML; demonstrable success in chatbot, conversational AI , and generative AI implementations. Experience building and operationalizing an AI/Chatbot architecture framework used enterprise-wide. Expertise in: Python , LangChain, ElasticSearch, NLP (spaCy, NLTK, Hugging Face) LLMs (e.g., GPT, BERT), RAG, prompt engineering Chatbot platforms (Azure OpenAI, MS Bot Framework), CLU, CQA AI solution deployment and monitoring at scale Familiarity with: Machine learning algorithms, deep learning, reinforcement learning NLP techniques for NLU/NLG Cloud platforms ( AWS, Azure, GCP ), Docker , Kubernetes Vector DBs (Pinecone, Weaviate, Qdrant) Semantic search, knowledge graphs, intelligent document processing Strong grasp of AI governance , documentation, and compliance standards Excellent team leadership, communication, and documentation skills Good-to-Have Skills Experience with Glean , Perplexity.ai , Rasa , XGBoost Familiarity with Salesforce , NetSuite , and business domains like Customer Success Knowledge of RPA tools like UiPath and its AI Center Role & responsibilities Interested candidate can call at 7087707007

Posted 3 weeks ago

Apply

6.0 - 11.0 years

40 - 60 Lacs

Kolkata

Work from Office

Naukri logo

We're looking for an experienced AI/ML Technical Lead to architect and drive the development of our intelligent conversation engine. Youll lead model selection, integration, training workflows (RAG/fine-tuning), and scalable deployment of natural language and voice AI components. This is a foundational hire for a technically ambitious platform. Key Responsibilities AI System Architecture: Design the architecture of the AI-powered agent including LLM-based conversation workflows, voice bots, and follow-up orchestration. Model Integration & Prompt Engineering: Leverage APIs from OpenAI, Anthropic, or deploy open models (e.g., LLaMA 3, Mistral). Implement effective prompt strategies and retrieval-augmented generation (RAG) pipelines for contextual responses. Data Pipelines & Knowledge Management: Build secure data pipelines to ingest, embed, and serve tenant-specific knowledge bases (FAQs, scripts, product docs) using vector databases (e.g., Pinecone, Weaviate). Voice & Text Interfaces: Implement and optimize multimodal agents (text + voice) using ASR (e.g., Whisper), TTS (e.g., Polly), and NLP for automated qualification and call handling. Conversational Flow Orchestration: Design dynamic, stateful conversations that can take actions (e.g., book meetings, update CRM records) using tools like LangChain, Temporal, or n8n. Platform Scalability: Ensure models and agent workflows scale across tenants with strong data isolation, caching, and secure API access. Lead a Cross-Functional Team: Collaborate with backend, frontend, and DevOps engineers to ship intelligent, production-ready features. Monitoring & Feedback Loops: Define and monitor conversation analytics (drop-offs, booking rates, escalation triggers), and create pipelines to improve AI quality continuously. Qualifications Must-Haves: 5+ years of experience in ML/AI, with at least 2 years leading conversational AI or LLM projects. Strong background in NLP, dialog systems, or voice AI preferably with production experience. Experience with OpenAI, or open-source LLMs (e.g. LLaMA, Mistral, Falcon) and orchestration tools (LangChain, etc.). Proficiency with Python and ML frameworks (Hugging Face, PyTorch, TensorFlow). Experience deploying RAG pipelines, vector DBs (e.g. Pinecone, Weaviate), and managing LLM-agent logic. Familiarity with voice processing (ASR, TTS, IVR design). Solid understanding of API-based integration and microservices. Deep care for data privacy, multi-tenancy security, and ethical AI practices. Nice-to-Haves: Experience with CRM ecosystems (e.g. Salesforce, HubSpot) and how AI agents sync actions to CRMs. Knowledge of sales pipelines and marketing automation tools. Exposure to calendar integrations (Google Calendar API, Microsoft Graph). Knowledge of Twilio APIs (SMS, Voice, WhatsApp) and channel orchestration logic. Familiarity with Docker, Kubernetes, CI/CD, and scalable cloud infrastructure (AWS/GCP/Azure). What We Offer Founding team role with strong ownership and autonomy Opportunity to shape the future of AI-powered sales Flexible work environment Competitive salary Access to cutting-edge AI tools and training resources Post your resume and any relevant project links (GitHub, blog, portfolio) to career@sourcedeskglobal.com. Include a short note on your most interesting AI project or voicebot/conversational AI experience.

Posted 3 weeks ago

Apply

3.0 - 8.0 years

15 - 30 Lacs

Hyderabad, Chennai, Bengaluru

Hybrid

Naukri logo

Job Description: We are seeking a highly skilled and passionate AI/ML Engineer with strong expertise in Generative AI and Large Language Models (LLMs) . The ideal candidate will have hands-on experience in building, fine-tuning, and deploying agentic AI systems using modern GenAI frameworks. You will work on cutting-edge projects involving prompt engineering , RAG pipelines , and memory architectures such as vector databases. Responsibilities: Design and implement AI/ML solutions using modern LLM architectures and agentic AI concepts . Build and optimize intelligent agents using frameworks such as LangChain, AutoGen, CrewAI , or Semantic Kernel . Develop and fine-tune generative AI models with Transformers , HuggingFace , OpenAI API , etc. Implement and enhance Retrieval-Augmented Generation (RAG) pipelines and memory systems like vector databases (e.g., FAISS, Pinecone). Write high-performance Python code to support experimentation, model integration, and API interactions. Collaborate cross-functionally with product, design, and engineering teams in an agile development environment. Deploy AI solutions on cloud platforms (AWS, Azure, or GCP) with a focus on scalability and performance. Stay updated with the latest advancements in the AI/ML/GenAI space. Required Experience: 3 to 8 years of experience in AI/ML , with at least 1 year in Generative AI / LLM-based projects . Proven expertise in Python programming and related libraries for ML/GenAI. Hands-on experience with one or more GenAI frameworks (LangChain, AutoGen, etc.). Solid understanding of prompt engineering , RAG , vector DBs , and agent-based systems . Cloud deployment experience (AWS, Azure, or GCP) is a must. Strong analytical and problem-solving skills.

Posted 3 weeks ago

Apply

4.0 - 5.0 years

8 - 12 Lacs

Vadodara

Hybrid

Naukri logo

Job Type: Full Time Job Description: We are seeking an experienced AI Engineer with 4-5 years of hands-on experience in designing and implementing AI solutions. The ideal candidate should have a strong foundation in developing AI/ML-based solutions, including expertise in Computer Vision (OpenCV). Additionally, proficiency in developing, fine-tuning, and deploying Large Language Models (LLMs) is essential. As an AI Engineer, candidate will work on cutting-edge AI applications, using LLMs like GPT, LLaMA, or custom fine-tuned models to build intelligent, scalable, and impactful solutions. candidate will collaborate closely with Product, Data Science, and Engineering teams to define, develop, and optimize AI/ML models for real-world business applications. Key Responsibilities: Research, design, and develop AI/ML solutions for real-world business applications, RAG is must. Collaborate with Product & Data Science teams to define core AI/ML platform features. Analyze business requirements and identify pre-trained models that align with use cases. Work with multi-agent AI frameworks like LangChain, LangGraph, and LlamaIndex. Train and fine-tune LLMs (GPT, LLaMA, Gemini, etc.) for domain-specific tasks. Implement Retrieval-Augmented Generation (RAG) workflows and optimize LLM inference. Develop NLP-based GenAI applications, including chatbots, document automation, and AI agents. Preprocess, clean, and analyze large datasets to train and improve AI models. Optimize LLM inference speed, memory efficiency, and resource utilization. Deploy AI models in cloud environments (AWS, Azure, GCP) or on-premises infrastructure. Develop APIs, pipelines, and frameworks for integrating AI solutions into products. Conduct performance evaluations and fine-tune models for accuracy, latency, and scalability. Stay updated with advancements in AI, ML, and GenAI technologies. Required Skills & Experience: AI & Machine Learning: Strong experience in developing & deploying AI/ML models. Generative AI & LLMs: Expertise in LLM pretraining, fine-tuning, and optimization. NLP & Computer Vision: Hands-on experience in NLP, Transformers, OpenCV, YOLO, R-CNN. AI Agents & Multi-Agent Frameworks: Experience with LangChain, LangGraph, LlamaIndex. Deep Learning & Frameworks: Proficiency in TensorFlow, PyTorch, Keras. Cloud & Infrastructure: Strong knowledge of AWS, Azure, or GCP for AI deployment. Model Optimization: Experience in LLM inference optimization for speed & memory efficiency. Programming & Development: Proficiency in Python and experience in API development. Statistical & ML Techniques: Knowledge of Regression, Classification, Clustering, SVMs, Decision Trees, Neural Networks. Debugging & Performance Tuning: Strong skills in unit testing, debugging, and model evaluation. Hands-on experience with Vector Databases (FAISS, ChromaDB, Weaviate, Pinecone). Good to Have: Experience with multi-modal AI (text, image, video, speech processing). Familiarity with containerization (Docker, Kubernetes) and model serving (FastAPI, Flask, Triton).

Posted 3 weeks ago

Apply

8.0 - 13.0 years

14 - 24 Lacs

Pune, Ahmedabad

Hybrid

Naukri logo

Senior Technical Architect Machine Learning Solutions We are looking for a Senior Technical Architect with deep expertise in Machine Learning (ML), Artificial Intelligence (AI) , and scalable ML system design . This role will focus on leading the end-to-end architecture of advanced ML-driven platforms, delivering impactful, production-grade AI solutions across the enterprise. Key Responsibilities Lead the architecture and design of enterprise-grade ML platforms , including data pipelines, model training pipelines, model inference services, and monitoring frameworks. Architect and optimize ML lifecycle management systems (MLOps) to support scalable, reproducible, and secure deployment of ML models in production. Design and implement retrieval-augmented generation (RAG) systems, vector databases , semantic search , and LLM orchestration frameworks (e.g., LangChain, Autogen). Define and enforce best practices in model development, versioning, CI/CD pipelines , model drift detection, retraining, and rollback mechanisms. Build robust pipelines for data ingestion, preprocessing, feature engineering , and model training at scale , using batch and real-time streaming architectures. Architect multi-modal ML solutions involving NLP, computer vision, time-series, or structured data use cases. Collaborate with data scientists, ML engineers, DevOps, and product teams to convert research prototypes into scalable production services . Implement observability for ML models including custom metrics, performance monitoring, and explainability (XAI) tooling. Evaluate and integrate third-party LLMs (e.g., OpenAI, Claude, Cohere) or open-source models (e.g., LLaMA, Mistral) as part of intelligent application design. Create architectural blueprints and reference implementations for LLM APIs, model hosting, fine-tuning, and embedding pipelines . Guide the selection of compute frameworks (GPUs, TPUs), model serving frameworks (e.g., TorchServe, Triton, BentoML) , and scalable inference strategies (batch, real-time, streaming). Drive AI governance and responsible AI practices including auditability, compliance, bias mitigation, and data protection. Stay up to date on the latest developments in ML frameworks, foundation models, model compression, distillation, and efficient inference . 14. Ability to coach and lead technical teams , fostering growth, knowledge sharing, and technical excellence in AI/ML domains. Experience managing the technical roadmap for AI-powered products , documentations ensuring timely delivery, performance optimization, and stakeholder alignment. Required Qualifications Bachelors or Master’s degree in Computer Science, Artificial Intelligence, Data Science, or a related field. 8+ years of experience in software architecture , with 5+ years focused specifically on machine learning systems and 2 years in leading team. Proven expertise in designing and deploying ML systems at scale , across cloud and hybrid environments. Strong hands-on experience with ML frameworks (e.g., PyTorch, TensorFlow, Hugging Face, Scikit-learn). Experience with vector databases (e.g., FAISS, Pinecone, Weaviate, Qdrant) and embedding models (e.g., SBERT, OpenAI, Cohere). Demonstrated proficiency in MLOps tools and platforms : MLflow, Kubeflow, SageMaker, Vertex AI, DataBricks, Airflow, etc. In-depth knowledge of cloud AI/ML services on AWS, Azure, or GCP – including certification(s) in one or more platforms. Experience with containerization and orchestration (Docker, Kubernetes) for model packaging and deployment. Ability to design LLM-based systems , including hybrid models (open-source + proprietary), fine-tuning strategies, and prompt engineering. Solid understanding of security, compliance , and AI risk management in ML deployments. Preferred Skills Experience with AutoML , hyperparameter tuning, model selection, and experiment tracking. Knowledge of LLM tuning techniques : LoRA, PEFT, quantization, distillation, and RLHF. Knowledge of privacy-preserving ML techniques , federated learning, and homomorphic encryption Familiarity with zero-shot, few-shot learning , and retrieval-enhanced inference pipelines. Contributions to open-source ML tools or libraries. Experience deploying AI copilots, agents, or assistants using orchestration frameworks.

Posted 3 weeks ago

Apply

2 - 5 years

8 - 12 Lacs

Pune

Work from Office

Naukri logo

About the job: The Red Hat, Experience Engineering (XE) team is looking for a skilled Python Developer with 2+ years of experience to join our Software Engineering team. In this role, the ideal candidate should have a strong background in Python development, a deep understanding of LLMs, and the ability to debug and optimize AI applications. Your work will directly impact our product development, helping us drive innovation and improve the customer experience. What will you do? Develop and maintain Python-based applications, integrating LLMs and AI-powered solutions. Collaborate with cross-functional teams (product managers, software engineers, and data teams) to understand requirements and translate them into data-driven solutions. Assist in the development, testing, and optimization of AI-driven features. Optimize performance and scalability of applications utilizing LLMs. Debug and resolve Python application errors, ensuring stability and efficiency. Conduct exploratory data analysis and data cleaning to prepare raw data for modelling. Optimize and maintain data storage and retrieval systems for model input/output. Research and experiment with new LLM advancements and AI tools to improve existing applications. Document workflows, model architectures, and code to ensure reproducibility and knowledge sharing across the team. What will you bring? Bachelor's degree in Computer Science, Software Engineering, or a related field with 2+ years of relevant experience. Strong proficiency in Python, including experience with frameworks like FastAPI/ Flask, or Django. Understanding of fundamental AI/ML concepts, algorithms, techniques and implementation of workflows. Familiarity with DevOps/MLOps practices and tools for managing the AI/ML lifecycle in production environments. Understanding of LLM training processes and data requirements. Experience in LLM fine-tuning, RAG and prompt engineering. Hands-on experience with LLMs (e.g., OpenAI GPT, Llama, or other transformer models) and their integration into applications(e.g. LangChain or Llama Stack). Familiarity with REST APIs, data structures, and algorithms. Strong problem-solving skills with the ability to analyze and debug complex issues. Experience with Git, CI/CD pipelines, and Agile methodologies. Experience working with cloud-based environments (AWS, GCP, or Azure) is a plus. Knowledge of vector databases (e.g., Pinecone, FAISS, ChromaDB) is a plus.

Posted 2 months ago

Apply

6 - 11 years

20 - 30 Lacs

Mumbai Suburbs, Mumbai, Mumbai (All Areas)

Work from Office

Naukri logo

Position Overview: We seek a skilled and innovative AI Engineer with background experience in Python, LangChain, AI, ML, and Data Science principles to design, develop, and deploy Agentic AI Agents / Vertical LLM Agents. The ideal candidate will possess extensive experience with LangChain, data science workflows, prompt engineering, retrieval-augmented generation (RAG), and LLM fine-tuning. You will be working to integrate structured and unstructured data into scalable knowledge bases and evaluate systems for continuous improvement. The role involves developing solutions for use cases primarily for UK-based clients and solving industry-specific challenges with cutting-edge AI technologies. Job Type: Full-Time Location: Powai, Mumbai Salary Range: Competitive, based on experience Working Hours: 10:30 am to 7:30 pm Indian Standard Time Days of work: Monday to Friday Key Responsibilities: 1. Knowledge Base Development and Integration Define Knowledge Base Scope: Collaborate with domain experts to identify industry-specific requirements and tasks. Assess and select appropriate structured and unstructured data sources. Data Curation and Organization: Collect and preprocess data from authoritative sources (e.g., research papers, databases, manuals). Structure unstructured data using techniques like knowledge graphs. Implement data cleaning workflows to ensure high-quality input. Knowledge Integration: Embed knowledge bases into LLM workflows using tools like Pinecone, Weaviate, or Milvus. 2. LLM Fine-Tuning Fine-tune LLMs using frameworks such as Hugging Face Transformers or OpenAI APIs. Use domain-specific datasets to adapt base models to specialized industries. Apply transfer learning techniques to enhance model performance for niche applications. Monitor and improve fine-tuned models using validation metrics and feedback loops. 3. Prompt Engineering Design, test, and optimize prompts for industry-specific tasks. Implement contextual prompting strategies to enhance accuracy and relevance. Iterate on prompt designs based on system evaluations and user feedback. 4. Retrieval-Augmented Generation (RAG) Implement RAG workflows to integrate external knowledge bases with LLMs. Develop and optimize embedding-based retrieval systems using vector databases. Combine retrieved knowledge with user queries to generate accurate and context-aware responses. 5. System Integration Build APIs and middleware to interface between the LLM, knowledge base, and user-facing applications. Develop scalable and efficient query-routing mechanisms for hybrid retrieval and generation tasks. Ensure seamless deployment of LLM-powered applications. 6. Validation and Testing Evaluate model responses against domain-specific benchmarks and ground truths. Collaborate with domain experts to refine system outputs. Conduct user testing and gather feedback to improve system performance iteratively. 7. Maintenance and Updates Implement strategies to keep the knowledge base current with periodic updates. Develop monitoring tools to track system performance and identify areas for improvement. Address ethical, regulatory, and privacy considerations (e.g., GDPR, HIPAA compliance). Qualifications: Technical Skills Programming: Strong knowledge of Python and frameworks like Flask, FastAPI, or LangChain for API development. Data Preprocessing: Familiarity with preprocessing pipelines for structured and unstructured data. LLM Proficiency: Experience with LLM platforms such as OpenAI GPT, Hugging Face Transformers, or similar. Knowledge Base Management: Hands-on experience with vector databases (e.g., Pinecone, Milvus, Weaviate) and relational databases (e.g., PostgreSQL, MySQL). Fine-Tuning Expertise: Proficiency in adapting LLMs for specialized domains using domain-specific datasets. RAG Implementation: Practical experience with retrieval-augmented generation workflows. Prompt Engineering: Ability to craft and optimize prompts for complex, context-driven tasks. Soft Skills Strong problem-solving skills and attention to detail. Ability to collaborate effectively with cross-functional teams, including domain experts. Excellent communication and documentation skills. Experience 6+ years of experience in AI/ML roles, focusing on LLM agent development and deployment in recent years. 2 + years would creating AI solution with langChain, experience. Demonstrated experience in designing domain-specific AI systems. Hands-on experience in integrating structured/unstructured data into AI models.

Posted 2 months ago

Apply

4 - 9 years

0 - 1 Lacs

Chennai

Work from Office

Naukri logo

Dear Professionals! We are seeking a skilled AI Developer to design, develop, and implement intelligent solutions that enhance business processes. The ideal candidate will leverage machine learning models, natural language processing (NLP), and deep learning techniques to build impactful AI-driven applications. Key Responsibilities Develop, train, and deploy machine learning models for various business solutions. Collaborate with cross-functional teams to define AI project goals and technical requirements. Design algorithms for data processing, feature engineering, and model evaluation. Implement and optimize NLP, computer vision, and predictive analytics models. Craft and refine production-grade prompting strategies for large language models (LLMs), ensuring reliability and efficiency. Build and maintain LLM pipelines using LangChain, integrating state-of-the-art models like GPT, Claude, and Gemini. Develop comprehensive frameworks for LLM performance metrics, quality assessments, and cost optimization. Design and implement GenAI applications, including LLM agents and Retrieval-Augmented Generation (RAG). Optimize similarity-based retrieval systems using modern vector databases like Weaviate and Pinecone. Skills & Qualifications Strong proficiency in Python, with a focus on GenAI best practices and frameworks. Expertise in machine learning algorithms, data modeling, and model evaluation. Experience with NLP techniques, computer vision, or generative AI. Deep knowledge of LLMs, prompt engineering, and GenAI technologies. Proficiency in data analysis tools like Pandas and NumPy. Hands-on experience with vector databases such as Weaviate or Pinecone. Familiarity with cloud platforms (AWS, Azure, GCP) for AI deployment. Strong problem-solving skills and critical-thinking abilities. Experience with AI model fairness, bias detection, and adversarial testing. Excellent communication skills to translate business needs into technical solutions. Preferred Qualifications Bachelors or Masters degree in Computer Science, AI, or a related field. Experience with MLOps practices for model deployment and maintenance. Strong understanding of data pipelines, APIs, and cloud infrastructure. Advanced degree in Computer Science, Machine Learning, or a related field (preferred). interested professionals kindly share your updated resume to hr@wee4techsolutions.com with Subject as GENAIML Developer - Contract/Freelancing

Posted 2 months ago

Apply

4 - 9 years

15 - 25 Lacs

Pune

Hybrid

Naukri logo

Job Title: AI/ML Engineer Location: Pune (Hybrid) Join our team as a AI/ML Engineer and be a part of our cutting-edge Generative AI lab. As an AI Engineer, you will play a pivotal role in developing and implementing core AI functionalities to drive value throughout the consulting lifecycle. Your responsibilities will include innovating information extraction, synthesis, and analysis of both structured and unstructured data to derive valuable quantitative insights. Responsibilities: Develop and own AI functionalities across the lifecycle, from initial proof of concept to production-ready pipelines. Research and experiment with new and emerging AI capabilities to establish performance metrics and suitability for proof of concept. Enhance developed pipelines into production-level modules for seamless integration into user experiences. Collaborate with the product team to ensure seamless deployment and functionality of AI solutions. Stay abreast of industry trends and technological advancements to continuously improve AI capabilities. Requirements: Experience in data science, specializing in building and deploying AI/ML applications. Proficiency in Python is a must, along with experience in ML libraries like Scikit-Learn and SQL. Familiarity with cloud-based platforms such as Azure and vector databases like pinecone is preferred. Deep understanding of statistical and machine learning models, including data preparation, model selection, hyperparameter tuning, and performance evaluation. Experience in processing unstructured natural language data and text extraction. Knowledge of NLP techniques, including text processing, TF-IDF, semantic parsing, and sentiment analysis, is a plus. Previous exposure to Generative AI and large language models will be advantageous. Strong analytical and problem-solving skills, with the ability to work independently with guidance. An advanced degree in Statistics, Computer Science, Physics, or Mathematics is preferred. Benefits: Unlimited Leave Policy Flexible working Hours Supportive team for Personal and Professional Growth Joining our team means being part of a founding team where you'll have ample opportunities for personal and professional growth.

Posted 3 months ago

Apply

3 - 6 years

20 - 35 Lacs

Bengaluru

Remote

Naukri logo

Python LLM Engineer (WFH) Experience: 3 - 5 Years Salary: INR 20,00,000-35,00,000 / year Preferred Notice Period : Within 15 days Shift : 10:30 AM to 7:30 PM IST Opportunity Type: Remote Placement Type: Permanent (*Note: This is a requirement for one of Uplers' Clients.) Must have skills required : API, Communication, LangChain, LLMs, Pinecone/ Weaviate/ FAISS/ ChromaDB, rag, AWS, Python Good to have skills : CI/CD, multimodal AI, Prompt engineering, Reinforcement Learning, Voice AI Platformance (One of Uplers' Clients) is Looking for: Python LLM Engineer who is passionate about their work, eager to learn and grow, and who is committed to delivering exceptional results. If you are a team player with a positive attitude and a desire to make a difference, then we want to hear from you. Role Overview Description We are seeking a highly skilled Python LLM Engineer to join our AI team. The ideal candidate should have deep expertise in large language models (LLMs), experience in building Retrieval-Augmented Generation (RAG) systems, and a strong background in AI-driven applications. This role requires hands-on experience with LangChain, multimodal AI, vector databases, agentic AI, and cloud-based AI infrastructure, particularly AWS and AWS Bedrock. Python will be the primary development language for this role. Key Responsibilities: Design, develop, and optimize applications leveraging LLMs using LangChain and other frameworks. Build and fine-tune Retrieval-Augmented Generation (RAG) based AI systems for efficient information retrieval. Implement and integrate major LLM APIs such as OpenAI, Anthropic, Google Gemini, and Mistral. Develop and optimize AI-driven voice applications and conversational agents using Python. Research and apply the latest advancements in AI, multimodal models, and vector databases. Architect and deploy scalable AI applications using AWS services, including AWS Bedrock. Design and implement vector search solutions using Pinecone, Weaviate, FAISS, or similar technologies. Develop agentic AI products that leverage autonomous decision-making and multi-agent coordination. Write efficient and scalable backend services in Python for AI-powered applications. Develop and optimize AI model fine-tuning and inference pipelines in Python. Implement end-to-end MLOps pipelines for model training, deployment, and monitoring using Python-based tools. Optimize LLM inference for performance and cost efficiency using Python frameworks. Ensure the security, scalability, and reliability of AI systems deployed in cloud environments. Required Skills and Experience: Strong experience with Large Language Models (LLMs) and their APIs (OpenAI, Anthropic, Cohere, Google Gemini, Mistral, etc.). Proficiency in LangChain and experience in developing modular AI pipelines. Deep knowledge of Retrieval-Augmented Generation (RAG) and its implementation. Experience with voice AI technologies, ASR (Automatic Speech Recognition), and TTS (Text-to-Speech), using Python-based frameworks. Familiarity with multimodal AI models (text, image, audio, and video processing) and Python libraries such as OpenCV, PIL, and SpeechRecognition. Hands-on experience with vector databases (Pinecone, Weaviate, FAISS, ChromaDB, etc.). Strong background in developing agentic AI products and autonomous AI workflows. Expertise in Python for AI/ML development, including libraries like TensorFlow, PyTorch, Hugging Face, FastAPI, and LangChain. Experience with AWS cloud services, including AWS Bedrock, Lambda, S3, and API Gateway, with Python-based implementations. Strong understanding of AI infrastructure, model deployment, and cloud scalability. Preferred Qualifications: Experience in reinforcement learning and self-improving AI agents. Exposure to prompt engineering, chain-of-thought prompting, and function calling. Prior experience in building production-grade AI applications in enterprise environments. Familiarity with CI/CD pipelines for AI model deployment and monitoring, using Python-based tools such as DVC, MLflow, and Airflow. Why Join Us? Work with cutting-edge AI technologies and build next-gen AI products. Be part of a highly technical and innovative AI-driven team. Competitive salary, stock options, and benefits. Opportunity to shape the future of AI-driven applications and agentic AI systems. Engagement Type: Direct-hire on the TBD payroll on behalf of platformance Job Type: Permanent Location: Remote Working time: 10:30 AM to 7:30 PM IST Interview Process - The HR team will conduct an initial culture fit assessment before technical rounds. Initial Technical Discussion: Live discussion to assess core competencies. Technical Assignment: Candidates will be given 4 days to complete a hands-on coding test. Final Interview (Optional): Review of the coding test and further technical discussion if required. How to Apply? Easy 3-Step Process: Step 1: Click On Apply! And Register or Login on our portal Step 2: Upload updated Resume & Complete the Screening Form Step 3: Increase your chances to get shortlisted & meet the client for the Interview! About Our Client: Platformance is a Growth Technology Platform that helps brands connect with customers using a pay-per-outcome model. Platformance is a growth technology platform built to help advertisers achieve measurable business outcomes, not just marketing results. Our mission is to simplify the complexities of digital advertising while ensuring every campaign delivers tangible results. About Uplers: Our goal is to make hiring reliable, simple, and fast. Our role will be to help all our talents find and apply for relevant product and engineering job opportunities and progress in their career. ( Note: There are many more opportunities apart from this on the portal.) So, if you are ready for a new challenge, a great work environment, and an opportunity to take your career to the next level, don't hesitate to apply today. We are waiting for you!

Posted 3 months ago

Apply

3 - 5 years

5 - 15 Lacs

Pune

Work from Office

Naukri logo

Job Description We are seeking a skilled Generative AI Engineer with a strong background in Python to join our dynamic team. In this role, you will integrate backend development expertise with the latest advancements in AI to create impactful solutions. If you excel in a fast-paced environment and enjoy tackling complex challenges, we encourage you to apply. Key Responsibilities: Generative AI Development Develop and implement generative AI models using frameworks like LangChain or Llama-Index. Apply prompt engineering techniques to design effective queries and ensure optimal LLM responses for diverse use cases. Master advanced LLM functionalities, including prompt optimization, hyperparameter tuning, and response caching. Implement Retrieval-Augmented Generation (RAG) workows by integrating vector databases like Pinecone, Weaviate, Supabase, or PGVector for ecient similarity searches. Work with embeddings and build solutions that leverage similarity search for personalized query resolution. Explore and process multimodal data, including image and video understanding and generation. Integrate observability tools for monitoring and evaluating LLM performance to ensure system reliability. Backend Engineering Build and maintain scalable backend systems using Python frameworks such as FastAPI, Django, or Flask. Design and implement RESTful APIs for seamless communication between systems and services. Optimize database performance with relational databases (PostgreSQL, MySQL) and integrate vector databases (Pinecone, PGVector, Weaviate, Supabase) for advanced AI workows. Implement asynchronous programming and adhere to clean code principles for maintainable, high-quality code. Seamlessly integrate third-party SDKs and APIs, ensuring robust interoperability with external systems. Develop backend pipelines for handling multimodal data processing, and supporting text, image, and video workows. Manage and schedule background tasks with tools like Celery, cron jobs, or equivalent job queuing systems. Leverage containerization tools such as Docker for ecient and reproducible deployments. Ensure security and scalability of backend systems with adherence to industry best practices. Qualications: Essential : Strong Programming Skills: Prociency in Python and experience with backend frameworks like FastAPI, Django, or Flask. Generative AI Expertise: Knowledge of frameworks like LangChain, Llama-Index, or similar tools, with experience in prompt engineering and Retrieval-Augmented Generation (RAG). Data Management: Hands-on experience with relational databases (PostgreSQL, MySQL) and vector databases (Pinecone, Weaviate, Supabase, PGVector) for embeddings and similarity search. Machine Learning Knowledge: Familiarity with LLMs, embeddings, and multimodal AI applications involving text, images, or video. Deployment Experience: Prociency in deploying AI models in production environments using Docker and managing pipelines for scalability and reliability. Testing and Debugging: Strong skills in writing and managing unit and integration tests (e.g., Pytest), along with application debugging and performance optimization. Asynchronous Programming: Understanding of asynchronous programming concepts for handling concurrent tasks eciently. Preferred: Cloud Prociency: Familiarity with platforms like AWS, GCP, or Azure, including serverless applications and VM setups. Frontend Basics: Understanding of HTML, CSS, and optionally JavaScript frameworks like Angular or React for better collaboration with frontend teams. Observability and Monitoring: Experience with observability tools to track and evaluate LLM performance in real-time. Cutting-Edge Tech: Awareness of trends in generative AI, including multimodal AI applications and advanced agentic workows. Security Practices: Knowledge of secure coding practices and backend system hardening. Certications: Relevant certications in AI, machine learning, or cloud technologies are a plus.

Posted 3 months ago

Apply

2 - 5 years

8 - 12 Lacs

Pune

Work from Office

Naukri logo

About the job: The Red Hat, Experience Engineering (XE) team is looking for a skilled Python Developer with 2+ years of experience to join our Software Engineering team. In this role, the ideal candidate should have a strong background in Python development, a deep understanding of LLMs, and the ability to debug and optimize AI applications. Your work will directly impact our product development, helping us drive innovation and improve the customer experience. What will you do? Develop and maintain Python-based applications, integrating LLMs and AI-powered solutions. Collaborate with cross-functional teams (product managers, software engineers, and data teams) to understand requirements and translate them into data-driven solutions. Assist in the development, testing, and optimization of AI-driven features. Optimize performance and scalability of applications utilizing LLMs. Debug and resolve Python application errors, ensuring stability and efficiency. Conduct exploratory data analysis and data cleaning to prepare raw data for modelling. Optimize and maintain data storage and retrieval systems for model input/output. Research and experiment with new LLM advancements and AI tools to improve existing applications. Document workflows, model architectures, and code to ensure reproducibility and knowledge sharing across the team. What will you bring? Bachelor's degree in Computer Science, Software Engineering, or a related field with 2+ years of relevant experience. Strong proficiency in Python, including experience with frameworks like FastAPI/ Flask, or Django. Understanding of fundamental AI/ML concepts, algorithms, techniques and implementation of workflows. Familiarity with DevOps/MLOps practices and tools for managing the AI/ML lifecycle in production environments. Understanding of LLM training processes and data requirements. Experience in LLM fine-tuning, RAG and prompt engineering. Hands-on experience with LLMs (e.g., OpenAI GPT, Llama, or other transformer models) and their integration into applications(e.g. LangChain or Llama Stack). Familiarity with REST APIs, data structures, and algorithms. Strong problem-solving skills with the ability to analyze and debug complex issues. Experience with Git, CI/CD pipelines, and Agile methodologies. Experience working with cloud-based environments (AWS, GCP, or Azure) is a plus. Knowledge of vector databases (e.g., Pinecone, FAISS, ChromaDB) is a plus.

Posted 1 month ago

Apply

5 - 10 years

25 - 30 Lacs

Mumbai, Navi Mumbai, Chennai

Work from Office

Naukri logo

We are looking for an AI Engineer (Senior Software Engineer). Interested candidates email me resumes on mayura.joshi@lionbridge.com OR WhatsApp on 9987538863 Responsibilities: Design, develop, and optimize AI solutions using LLMs (e.g., GPT-4, LLaMA, Falcon) and RAG frameworks. Implement and fine-tune models to improve response relevance and contextual accuracy. Develop pipelines for data retrieval, indexing, and augmentation to improve knowledge grounding. Work with vector databases (e.g., Pinecone, FAISS, Weaviate) to enhance retrieval capabilities. Integrate AI models with enterprise applications and APIs. Optimize model inference for performance and scalability. Collaborate with data scientists, ML engineers, and software developers to align AI models with business objectives. Ensure ethical AI implementation, addressing bias, explainability, and data security. Stay updated with the latest advancements in generative AI, deep learning, and RAG techniques. Requirements: 8+ years experience in software development according to development standards. Strong experience in training and deploying LLMs using frameworks like Hugging Face Transformers, OpenAI API, or LangChain. Proficiency in Retrieval-Augmented Generation (RAG) techniques and vector search methodologies. Hands-on experience with vector databases such as FAISS, Pinecone, ChromaDB, or Weaviate. Solid understanding of NLP, deep learning, and transformer architectures. Proficiency in Python and ML libraries (TensorFlow, PyTorch, LangChain, etc.). Experience with cloud platforms (AWS, GCP, Azure) and MLOps workflows. Familiarity with containerization (Docker, Kubernetes) for scalable AI deployments. Strong problem-solving and debugging skills. Excellent communication and teamwork abilities Bachelors or Masters degree in computer science, AI, Machine Learning, or a related field.

Posted 1 month ago

Apply
cta

Start Your Job Search Today

Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.

Job Application AI Bot

Job Application AI Bot

Apply to 20+ Portals in one click

Download Now

Download the Mobile App

Instantly access job listings, apply easily, and track applications.

Featured Companies