Get alerts for new jobs matching your selected skills, preferred locations, and experience range. Manage Job Alerts
1.0 - 5.0 years
0 Lacs
pune, maharashtra
On-site
As an AI/ML Agent Developer at EXL, you will play a crucial role in designing, developing, and deploying intelligent agents within a multi-agent orchestration framework. Your primary responsibility will be to create autonomous agents that utilize LLMs, reinforcement learning, prompt engineering, and decision-making strategies to handle complex data and workflow tasks. Working closely with cross-functional teams, you will help in implementing AI solutions across various use cases such as annotation, data quality, knowledge graph construction, and enterprise automation. Your key responsibilities will include designing and implementing modular AI agents capable of autonomous decision-making, engineering prompt strategies for task-specific agent workflows, integrating ML models into agent behavior pipelines, contributing to multi-agent orchestration logic, collaborating with different teams for deployment at scale, and monitoring and refining agent performance. To be successful in this role, you are required to have a Bachelors or Masters degree in Computer Science, AI/ML, Data Science, or a related field along with at least 3 years of experience in developing AI/ML systems. You should also possess proficiency in Python and ML libraries, experience with LLM frameworks, a strong grasp of NLP, reinforcement learning, and decision systems, as well as knowledge of cloud environments and CI/CD for AI systems. Preferred skills include familiarity with multi-agent frameworks, experience in building autonomous AI applications for data governance, annotation, or knowledge extraction, background in human-in-the-loop systems, and understanding of vector databases and semantic search. By joining our team, you will have the opportunity to work at the forefront of AI orchestration and intelligent agents, collaborate with a high-performing team driving innovation in enterprise AI platforms, and shape the future of AI-based automation in real-world domains like healthcare, finance, and unstructured data.,
Posted 1 month ago
5.0 - 9.0 years
0 Lacs
ahmedabad, gujarat
On-site
The Technical Lead for AI/ML plays a critical role in driving the technical direction and execution of artificial intelligence and machine learning projects within the organization. This position combines strong technical expertise in AI/ML with leadership skills to guide a team of engineers and data scientists in developing innovative AI/ML solutions. You will collaborate closely with cross-functional teams to translate business requirements into technical designs and drive the successful implementation, testing and deployment of AI/ML models and systems. Responsibilities 1. AI/ML Model Training and Development Lead the development of new AI Models from scratch and fine-tuning existing base models with new data. Implement and execute model evaluation experiments to test and pick the best model for a certain task. Build high throughput and automated data pipelines to allow Data analysis and model training. Lead feature engineering to select the most relevant features in datasets to train models on. Pick the right set of evaluation metrics to test model candidates for deployment. Use MLOps principles to allow continuous model tuning and performance improvements. 2. AI/ML Model Deployment Lead the deployment of AI/ML models into production environments, ensuring scalability, reliability, and performance. Implement best practices for model versioning, monitoring, and maintenance to ensure ongoing model accuracy and effectiveness. Collaborate with DevOps and infrastructure teams to integrate AI/ML components into CI/CD pipelines and automated deployment processes. Implement CI/CD practices for AI/ML development, including automated testing, code review processes, and continuous integration pipelines. Automate deployment processes for AI/ML models using tools such as Jenkins, GitLab CI/CD, or similar platforms. 3. Technology Expertise Demonstrate deep expertise in AI/ML technologies, including TensorFlow, PyTorch, Keras, NumPy, Pandas and familiarity with platforms such as OpenAI, Hugging Face, Perplexity AI and Anthropic. Stay current with advancements in AI/ML research and technologies, evaluating their applicability to the organization's needs and projects. 4. Architecture and Design Design and implement architectures around AI/ML solutions, including data pipelines, model serving infrastructure, and integration with existing systems. Collaborate with data engineers to ensure the availability, quality, and reliability of data sources for AI/ML model training and deployment. 5. Python Development Utilize Python programming for AI/ML model development, automation scripts, and development of supporting tools and utilities. Collaborate with software engineering teams to integrate AI/ML capabilities into software applications and services. Requirements Bachelor's or Master's degree in Computer Science, Engineering, Mathematics, or related field. Extensive experience (5+ years) in AI/ML development, with a focus on deploying models into production environments. Strong proficiency in AI/ML frameworks such as TensorFlow, PyTorch, Keras, NumPy, Pandas and familiarity with platforms such as OpenAI, Hugging Face, Perplexity AI and Anthropic. Experience building architectures around AI/ML solutions, including data pipelines, model serving infrastructure, and integration with existing systems. Hands-on experience with CI/CD practices and tools, with a strong understanding of software development lifecycle processes. Proficiency in Python programming and experience with relevant libraries and frameworks for AI/ML development. Experience of Python Pandas and similar languages is a must. Worked on pre-processing pipelines ensuring security compliances standards are met. Excellent communication skills and the ability to collaborate effectively with cross-functional teams. Strong problem-solving abilities and a passion for innovation and continuous improvement in AI/ML deployment practices.,
Posted 1 month ago
5.0 - 7.0 years
0 Lacs
Pune, Maharashtra, India
On-site
STAND 8 provides end-to-end IT solutions to enterprise partners across the United States and with offices in Los Angeles, New York, New Jersey, Atlanta, and more including internationally in Mexico and India. We are seeking a Senior AI Engineer / Data Engineer to join our engineering team and help build the future of AI-powered business solutions. In this role, you&aposll be developing intelligent systems that leverage advanced large language models (LLMs), real-time AI interactions, and cutting-edge retrieval architectures. Your work will directly contribute to products that are reshaping how businesses operate-particularly in recruitment, data extraction, and intelligent decision-making. This is an exciting opportunity for someone who thrives in building production-grade AI systems and working across the full stack of modern AI technologies. Responsibilities Design, build, and optimize AI-powered systems using multi-modal architectures (text, voice, visual). Integrate and deploy LLM APIs from providers such as OpenAI, Anthropic, and AWS Bedrock. Build and maintain RAG (Retrieval-Augmented Generation) systems with hybrid search, re-ranking, and knowledge graphs. Develop real-time AI features using streaming analytics and voice interaction tools (e.g., ElevenLabs). Build APIs and pipelines using FastAPI or similar frameworks to support AI workflows. Process and analyze unstructured documents with layout and semantic understanding. Implement predictive models that power intelligent business recommendations. Deploy and maintain scalable solutions using AWS services (EC2, S3, RDS, Lambda, Bedrock, etc.). Use Docker for containerization and manage CI/CD workflows and version control via Git. Debug, monitor, and optimize performance for large-scale data pipelines. Collaborate cross-functionally with product, data, and engineering teams. Qualifications 5+ years of experience in AI/ML or data engineering with Python in production environments. Hands-on experience with LLM APIs and frameworks such as OpenAI, Anthropic, Bedrock, or LangChain. Production experience using vector databases like PGVector, Weaviate, FAISS, or Pinecone. Strong understanding of NLP, document extraction, and text processing. Proficiency in AWS cloud services including Bedrock, EC2, S3, Lambda, and monitoring tools. Experience with FastAPI or similar frameworks for building AI/ML APIs. Familiarity with embedding models, prompt engineering, and RAG systems. Asynchronous programming knowledge for high-throughput pipelines. Experience with Docker, Git workflows, CI/CD pipelines, and testing best practices. Preferred Background in HRTech or ATS integrations (e.g., Greenhouse, Workday, Bullhorn). Experience working with knowledge graphs (e.g., Neo4j) for semantic relationships. Real-time AI systems (e.g., WebRTC, OpenAI Realtime API) and voice AI tools (e.g., ElevenLabs). Advanced Python development skills using design patterns and clean architecture. Large-scale data processing experience (1-2M+ records) with cost optimization techniques for LLMs. Event-driven architecture experience using AWS SQS, SNS, or EventBridge. Hands-on experience with fine-tuning, evaluating, and deploying foundation models. Show more Show less
Posted 1 month ago
3.0 - 7.0 years
0 Lacs
noida, uttar pradesh
On-site
As a skilled developer, you will be responsible for various aspects of web development, including designing, developing, and maintaining scalable RESTful APIs using frameworks such as Django, Flask, or FastAPI. You will also be involved in building and deploying modern web applications that prioritize performance, modularity, and security. Furthermore, your role will entail implementing API authentication and authorization mechanisms, such as OAuth2 JWT or session-based approaches, to ensure secure access. Collaboration with frontend teams is crucial as you will work closely with them to integrate APIs seamlessly and guarantee smooth end-to-end flows within the applications. Following continuous integration/continuous deployment (CI/CD) practices and utilizing Git-based deployment workflows with tools like GitHub, GitLab, Jenkins, or Docker will be part of your routine tasks. In addition to web development, you will be engaged in GenAI application development, focusing on building intelligent applications using tools like LangGraph and LangChain for orchestrating Large Language Model (LLM) workflows. Integration of various LLMs, including OpenAI, Anthropic, HuggingFace, or custom models, into production workflows will be a key responsibility. Your expertise will also be required in designing and optimizing Retrieval Augmented Generation (RAG) pipelines using vector databases like Pinecone, PGVector, etc. Furthermore, your role will involve database and backend integration tasks, where you will work with relational databases such as PostgreSQL and MySQL. Writing efficient and scalable queries for large-scale datasets will be essential, and any experience with cloud platforms like AWS, GCP, or Azure would be beneficial. To excel in this position, you are expected to have a minimum of 3 years of experience in application development using Python/Django, along with proficiency in developing and consuming RESTful APIs. Familiarity with tools like LangGraph, LangChain, and GenAI workflows, as well as experience with various LLMs such as OpenAI GPT-4, Claude, Llama, etc., will be advantageous. A solid understanding of software design principles, code modularity, and version control using Git is also required. If you are passionate about cutting-edge technologies and enjoy working in a collaborative environment, this opportunity with Finoit, as posted by Neetu Misri, could be the perfect fit for you.,
Posted 1 month ago
3.0 - 5.0 years
0 Lacs
Mumbai, Maharashtra, India
On-site
Responsibilities Design and fine-tune LLMs (Large Language Models) for BFSI use-cases: intelligent document processing, report generation, chatbots, advisory tools. Evaluate and apply prompt engineering, retrieval-augmented generation (RAG), and fine-tuning methods. Implement safeguards, red-teaming, and audit mechanisms for LLM usage in BFSI. Work with data privacy, legal, and compliance teams to align GenAI outputs with industry regulations. Collaborate with enterprise architects to integrate GenAI into existing digital platforms. Qualifications 35 years in AI/ML; 13 years hands-on in GenAI/LLM-based solutions. BFSI-specific experience in document processing, regulatory reporting, or virtual agents using GenAI is highly preferred. Exposure to prompt safety, model alignment, and RAG pipelines is critical. Essential Skills Tech Stack LLMs: GPT (OpenAI), Claude, LLaMA, Mistral, Falcon Tools: LangChain, LlamaIndex, Pinecone, Weaviate Frameworks: Transformers (Hugging Face), PEFT, DeepSpeed APIs: OpenAI, Cohere, Anthropic, Azure OpenAI Cloud: GCP GenAI Studio, GCP Vertex AI Others: Prompt engineering, RAG, vector databases, role-based guardrails Experience 35 years in AI/ML; 13 years hands-on in GenAI/LLM-based solutions. Show more Show less
Posted 1 month ago
4.0 - 12.0 years
0 Lacs
hyderabad, telangana
On-site
Join a high-performing engineering team that is pioneering the future of low-code and AI-powered development. The team is passionate about building scalable, intelligent systems that simplify complex business processes. By blending deep domain knowledge in low-code platforms with emerging capabilities in Generative AI, the team delivers innovative, human-centered solutions and encourages each member to bring out their best to build a world-class Pega Infinity product. The Pega Platform engineering team drives the product strategy and delivery with an innovation mindset to provide a low-code platform for AI-powered decision-making and workflow automation. If you are excited about shaping how businesses leverage automation and AI, this is the place for you. Pegasystems is offering exciting, multi-level opportunities to join our team as a Manager, Platform Engineering, Low-code Studios, and GenAI. As a Manager, you will work with a self-organizing team to develop and maintain best-of-breed software in the Pega ecosystem and at the intersection of technology and innovation. You will be part of a collaborative environment where your ideas are valued, your growth is prioritized, and your work directly contributes to real-world impact. Whether it's improving customer experiences, streamlining operations, or accelerating product development with GenAI, you will be solving meaningful problems that matter to our clients and our company. **What You'll Do At Pega** - Manage product engineering teams working on designing, developing, and deploying applications using leading low-code platforms and integrating Generative AI capabilities to enable smart automation, application generation, and enhanced user experiences. - Collaborate with designers, product managers, and other teams to deliver high-quality, performant solutions. - Participate in architecture discussions and help shape the technical direction of our platform. - Continuously evaluate new GenAI and low-code trends, bringing innovative ideas into development cycles. - Provide technical assistance to advance the capabilities of products and resolve complex issues. - Develop and maintain technical expertise in assigned areas of product functionality. **Who You Are** - You are a passionate leader with a deep interest in Pega and stay updated with the latest advancements in GenAI, including models, use cases, and industry trends. - Strong understanding of software engineering principles, with a focus on Pega Infinity, GenAI, and machine learning concepts. - You bring good practices around design patterns with you, which you can leverage to support our Pega platform with our "as-a-Service" evolution, particularly in the context of GenAI. - Passionate about helping teams grow by inspiring and mentoring engineers. - Experienced in integrating APIs, building reusable logic, and optimizing workflows for efficiency and scale. - Excellent verbal and written communication skills, with the ability to remain composed under pressure. - Bonus: Experience working with Pega platform (App Studio, Case Management, Decisioning) and knowledge of AI tooling (e.g., OpenAI, Anthropic, LangChain) and prompt engineering is a plus. **What You've Accomplished** - 8-12 years of overall experience, preferably in a product development company, and holds a Bachelor's or Master's degree in Computer Science, Engineering, or a related field. - Minimum 4+ years of experience managing product development teams. - Pega 8.x Certified CSA and CSSA preferably. - Strong problem-solving, debugging, and design skills. Participated in the design and development of Pega Infinity features. - Ability to apply analytical and investigative skills to resolve customer issues. - Customer-oriented, enthusiastic, courteous, assertive, and motivated to take charge of both customer engagement and problem resolution. - Rich experience of managing engineering teams in a product development company. - Ability to understand, identify, and resolve conflicts as necessary. - Working experience in an Agile/Scrum team environment. Pega Offers You - Gartner Analyst acclaimed technology leadership across our categories of products. - Continuous learning and development opportunities. - An innovative, inclusive, agile, flexible, and fun work environment.,
Posted 1 month ago
1.0 - 5.0 years
0 Lacs
ahmedabad, gujarat
On-site
You are a skilled Data Engineer / AI/ML Developer with a strong expertise in Python and AI/ML, responsible for building and scaling production-grade systems using LLMs. Your role involves working on real-world AI integrations, APIs, and cloud deployments within a fast-paced, growth-focused team. Your key responsibilities include building AI/LLM systems using OpenAI, Anthropic, etc., working with frameworks like LangChain, CrewAI, or AutoGen, developing APIs with Flask/FastAPI and async patterns, integrating databases (PostgreSQL/MySQL) and REST APIs, deploying on AWS, GCP, or Azure, writing clean, testable, production-grade Python code, and translating business needs into technical solutions. To excel in this role, you must have 2+ years of experience with LLM APIs and 1+ AI project in production, 3+ years of Python development experience, familiarity with function calling, prompt design (ReAct, CoT), cloud deployment (any platform), Git workflows, and testing. Additionally, you should possess strong debugging, error handling, and system design skills, along with the ability to work independently and collaborate cross-functionally. A preferred background would include a degree in CS or equivalent experience (bootcamp, self-taught, etc.), along with a portfolio showcasing live AI projects or GitHub repos. Knowledge of Vector DBs (Pinecone/Weaviate), Docker, and exposure to fintech would be considered as a bonus. Cloud or AI certifications, as well as experience in fintech or data-heavy domains, are nice-to-have skills for this role.,
Posted 1 month ago
8.0 - 12.0 years
0 Lacs
hyderabad, telangana
On-site
You are looking for a DevOps Technical Lead who will play a crucial role in leading the development of an Infrastructure Agent powered by Generative AI (GenAI) technology. In this role, you will be responsible for designing and implementing an intelligent Infra Agent that can handle provisioning, configuration, observability, and self-healing autonomously. Your key responsibilities will include leading the architecture and design of the Infra Agent, integrating various automation frameworks to enhance DevOps workflows, automating infrastructure provisioning and incident remediation, developing reusable components and frameworks using Infrastructure as Code (IaC) tools, and collaborating with AI/ML engineers and SREs to create intelligent infrastructure decision-making logic. You will also be expected to implement secure and scalable infrastructure on cloud platforms such as AWS, Azure, and GCP, continuously improve agent performance through feedback loops, telemetry, and model fine-tuning, drive DevSecOps best practices, compliance, and observability, as well as mentor DevOps engineers and work closely with cross-functional teams. To qualify for this role, you should hold a Bachelor's or Master's degree in Computer Science, Engineering, or a related field, along with at least 8 years of experience in DevOps, SRE, or Infrastructure Engineering. You must have proven experience in leading infrastructure automation projects, expertise with cloud platforms like AWS, Azure, GCP, and deep knowledge of tools such as Terraform, Kubernetes, Helm, Docker, Jenkins, and GitOps. Hands-on experience with LLMs/GenAI APIs, familiarity with automation frameworks, and proficiency in programming/scripting languages like Python, Go, or Bash are also required. Preferred qualifications for this role include experience in building or fine-tuning LLM-based agents, contributions to open-source GenAI or DevOps projects, understanding of MLOps pipelines and AI infrastructure, and certifications in DevOps, cloud, or AI technologies.,
Posted 1 month ago
5.0 - 9.0 years
0 Lacs
karnataka
On-site
DecisionX is pioneering a new category with the world's first Decision AI, an AI Super-Agent that assists high-growth teams in making smarter, faster decisions by transforming fragmented data into clear next steps. Whether it involves strategic decisions in the boardroom or operational decisions across various departments like Sales, Marketing, Product, and Engineering, down to the minutiae that drives daily operations, Decision AI serves as your invisible co-pilot, thinking alongside you, acting ahead of you, and evolving beyond you. We are seeking a dedicated and hands-on AI Engineer to join our Founding team. In this role, you will collaborate closely with leading AI experts to develop the intelligence layer of our exclusive "Agentic Number System." Key Responsibilities - Building, fine-tuning, and deploying AI/ML models for tasks such as segmentation, scoring, recommendation, and orchestration. - Developing and optimizing agent workflows using LLMs (OpenAI, Claude, Mistral, etc.) for contextual reasoning and task execution. - Creating vector-based memory systems utilizing tools like FAISS, Chroma, or Weaviate. - Working with APIs and connectors to incorporate third-party data sources (e.g., Salesforce, HubSpot, GSuite, Snowflake). - Designing pipelines that transform structured and unstructured signals into actionable insights. - Collaborating with GTM and product teams to define practical AI agent use cases. - Staying informed about the latest developments in LLMs, retrieval-augmented generation (RAG), and agent orchestration frameworks (e.g., CrewAI, AutoGen, LangGraph). Must Have Skills - 5-8 years of experience in AI/ML engineering or applied data science. - Proficient programming skills in Python, with expertise in LangChain, Pandas, NumPy, and Scikit-learn. - Experience with LLMs (OpenAI, Anthropic, etc.), prompt engineering, and RAG pipelines. - Familiarity with vector stores, embeddings, and semantic search. - Expertise in data wrangling, feature engineering, and model deployment. - Knowledge of MLOps tools such as MLflow, Weights & Biases, or equivalent. What you will get - Opportunity to shape the AI architecture of a high-ambition startup. - Close collaboration with a visionary founder and experienced product team. - Ownership, autonomy, and the thrill of building something from 0 to 1. - Early team equity and a fast growth trajectory.,
Posted 1 month ago
4.0 - 9.0 years
20 - 35 Lacs
Hyderabad, Chennai, Bengaluru
Work from Office
4+ years of experience in Java development with Spring Boot. Experience integrating AI/ML models into backend systems. Must have experience in any of the GenAI tools like Bedrock/ OpenAi/Tensorflow/Ai21/Anthropic/Cohere/Stability OR Scikit-Learn Strong understanding of RESTful API design and microservices. Familiarity with AI/ML tools and frameworks (e.g., Python, TensorFlow, Scikit-learn). Experience with cloud platforms (AWS, GCP, or Azure). Knowledge of containerization (Docker, Kubernetes) and event-driven architectures. Preferred Qualifications Experience with GenAI platforms (e.g., AWS Bedrock, OpenAI) Understanding of MLOps practices and model lifecycle management. Bachelors or Masters degree in Computer Science, Engineering, or related field.
Posted 1 month ago
5.0 - 10.0 years
8 - 18 Lacs
Hyderabad, Pune
Work from Office
5–8 years of experience in AI/ML or automation engineering Exp with CrewAI or other LLM orchestration frameworks like LangChain, AutoGen, or Semantic Kernel.Proficiency in Python.Understanding of LLMs OpenAI, Anthropic etc
Posted 1 month ago
5.0 - 10.0 years
15 - 25 Lacs
Hyderabad
Remote
Crew AI Engineer Remote Contractual-6 months Job Description : We are looking for people with strong python skills, knowledge of multi agent frameworks like Crew AI, knowledge of RAG concepts mandatory. Good conceptual knowledge of LLM concepts Langraph and Langchain Required Skills : 5-8 years of experience in AI/ML or automation engineering. Strong hands-on experience with CrewAI or other LLM orchestration frameworks like LangChain, AutoGen, or Semantic Kernel. Proficiency in Python, including experience with async programming and API integration. Deep understanding of LLMs (OpenAI, Anthropic, Mistral, etc.) and prompt engineering. Familiarity with vector databases (e.g., Pinecone, FAISS, Chroma) and embeddings. Experience building and deploying production-ready agent-based solutions. Strong problem-solving skills and ability to translate business requirements into technical implementations.
Posted 1 month ago
2.0 - 6.0 years
0 Lacs
maharashtra
On-site
As a talented software developer at our company, you will have the opportunity to showcase your passion for coding and product building. Your problem-solving mindset and love for taking on new challenges will make you a valuable addition to our rockstar engineering team. You will be responsible for designing and developing robust, scalable, and secure backend architectures using Django. Your focus will be on backend development to ensure the smooth functioning of web applications and systems. Creating high-quality RESTful APIs to facilitate seamless communication between frontend, backend, and other services will also be a key part of your role. In addition, you will play a crucial role in designing, implementing, and maintaining database schemas to ensure data integrity, performance, and security. You will work on ensuring the scalability and reliability of our backend infrastructure on AWS, aiming for zero downtime of systems. Writing clean, maintainable, and efficient code while following industry standards and best practices will be essential. Collaboration is key in our team, and you will conduct code reviews, provide feedback to team members, and work closely with frontend developers, product managers, and designers to plan and optimize features. You will break down high-level business problems into smaller components and build efficient systems to address them. Staying updated with the latest technologies such as LLM frameworks and implementing them as needed will be part of your continuous learning process. Your skills in Python, Django, SQL/PostgreSQL databases, and AWS services will be put to good use as you optimize systems, identify bottlenecks, and resolve them to enhance efficiency. To qualify for this role, you should have a Bachelor's degree in Computer Science or a related field, along with at least 2 years of experience in full-stack web development with a focus on backend development. Proficiency in Python, Django, and experience with Django Rest Framework are required. Strong problem-solving skills, excellent communication, and collaboration abilities are also essential for success in this position.,
Posted 1 month ago
3.0 - 7.0 years
0 Lacs
pune, maharashtra
On-site
As a Data Analytics focused Senior Software Engineer at PubMatic, you will be responsible for developing advanced AI agents to enhance data analytics capabilities. Your expertise in building and optimizing AI agents, along with strong skills in Hadoop, Spark, Scala, Kafka, Spark Streaming, and cloud-based solutions, will play a crucial role in improving data-driven insights and analytical workflows. Your key responsibilities will include building and implementing a highly scalable big data platform to process terabytes of data, developing backend services using Java, REST APIs, JDBC, and AWS, and building and maintaining Big Data pipelines using technologies like Spark, Hadoop, Kafka, and Snowflake. Additionally, you will design and implement real-time data processing workflows, develop GenAI-powered agents for analytics and data enrichment, and integrate LLMs into existing services for query understanding and decision support. You will work closely with cross-functional teams to enhance the availability and scalability of large data platforms and PubMatic software functionality. Participating in Agile/Scrum processes, discussing software features with product managers, and providing customer support over email or JIRA will also be part of your role. We are looking for candidates with three plus years of coding experience in Java and backend development, solid computer science fundamentals, expertise in developing software engineering best practices, hands-on experience with Big Data tools, and proven expertise in building GenAI applications. The ability to lead feature development, debug distributed systems, and learn new technologies quickly are essential. Strong interpersonal and communication skills, including technical communications, are highly valued. To qualify for this role, you should have a bachelor's degree in engineering (CS/IT) or an equivalent degree from well-known Institutes/Universities. PubMatic employees globally have returned to our offices via a hybrid work schedule to maximize collaboration, innovation, and productivity. Our benefits package includes paternity/maternity leave, healthcare insurance, broadband reimbursement, and office perks like healthy snacks, drinks, and catered lunches. About PubMatic: PubMatic is a leading digital advertising platform that provides transparent advertising solutions to publishers, media buyers, commerce companies, and data owners. Our vision is to enable content creators to run a profitable advertising business and invest back into the multi-screen and multi-format content that consumers demand.,
Posted 1 month ago
2.0 - 6.0 years
0 Lacs
hyderabad, telangana
On-site
The position of Gen AI / AI Engineer (Mid-Level) in Hyderabad requires a candidate with at least 2 years of experience in developing and deploying AI/ML solutions, focusing particularly on Generative AI and RAG (Retrieval-Augmented Generation) architectures. As an AI Engineer, you will be involved in working with cutting-edge models, fine-tuning, and integrating systems to create scalable AI products. Your responsibilities will include fine-tuning and optimizing Large Language Models (LLMs) and generative models for both text and image data. You will also be responsible for designing and implementing RAG pipelines that incorporate knowledge bases, developing APIs, microservices, and scalable AI pipelines, collaborating with MLOps teams for production deployment, and staying updated on the latest AI trends to apply innovative techniques effectively. The ideal candidate should possess expertise in Python programming language, PyTorch or TensorFlow frameworks, Generative AI tools such as OpenAI, Anthropic, LLaMA, RAG architecture, Natural Language Processing (NLP), Computer Vision, API development, MLOps practices, and experience working with Cloud Platforms like AWS, Azure, or GCP. Qualifications for this role include a Bachelor's degree in Computer Science or a related field, a minimum of 2 years of experience in AI/ML with a focus on GenAI and RAG, and hands-on experience in deploying models at a production level. If you are enthusiastic about contributing to the development of next-generation AI solutions, we encourage you to apply for this exciting opportunity.,
Posted 2 months ago
6.0 - 10.0 years
0 Lacs
karnataka
On-site
As a Full Stack-Backend Engineer at an AI startup based in Abu Dhabi, you will be an integral part of a fast-scaling company with a strong revenue stream, developer traction, and global growth. Your primary responsibility will be to contribute to the development of the next-generation agent by building scalable APIs and services for agent-based applications. You will have the opportunity to lead engineering efforts in the region from Day One. Your key responsibilities will include developing RAG pipelines using LangChain or custom tools, working with distributed systems and databases, and optionally contributing to UI development with React. You will be expected to own features end-to-end and collaborate effectively with global teams. To excel in this role, you should possess strong Python skills or a similar backend language, with an experience range of 6-8 years. Experience with distributed systems, APIs, and databases is essential, along with familiarity with LLMs, RAG, and agent frameworks. Frontend experience with React is a bonus, but not mandatory. An independent and ownership-driven mindset will be crucial for success in this position. The tech environment you will be working in includes Backend technologies such as Python and FastAPI, infrastructure tools like AWS, Docker, and Kubernetes, and AI Stack components like LangChain, Pinecone/Weaviate, OpenAI, and Anthropic. DevOps practices involve GitHub Actions and CI/CD pipelines, with a focus on developer tools, AI agents, and real-time search APIs. Joining this role will allow you to work with a well-funded startup that is making a real-world impact with over $3M+ ARR, 200K+ users, and 5M+ downloads. You will have the opportunity to build solutions across AI, search, and distributed infrastructure, lead development strategies within a global team, and be part of the thriving tech and innovation ecosystem in Abu Dhabi.,
Posted 2 months ago
6.0 - 10.0 years
8 - 12 Lacs
Bengaluru
Work from Office
What Were Looking For PM at a Consumer Startup (Pre-Series C) Youve built from scratch. Worked in lean teams. And you are a builder at heart. Launched & Scaled Something Real Youve taken something from v1 to real scaleDAUs, downloads, paid users. Your resume shows outcomes, not fluff. Growth DNA Youve worked on GTM, SEO, CRO, onboarding, referrals, or pricing. Youve shipped experiments and scaled what worked. Hands-On AI Application Youve integrated LLMs via APIs into actual products. Not prompt engineering, not AI curious. Youve built and shipped. Tools like OpenAI, Anthropic, LangChain, etc. Strong with Data You understand that insights are everything . You dont wait for dashboardsyou go hunting for truth in the numbers. Why Leap Work directly on 0-to-1s in a fast-growing, profitable company High ownership, fast decisions, no red tape A real shot at building products that scale to millions Work with founders and a high-velocity team that deeply cares Be at the front of building AI-native consumer experiences This is not a traditional PM job.
Posted 2 months ago
3.0 - 7.0 years
0 Lacs
chennai, tamil nadu
On-site
You are seeking a hands-on backend expert to elevate your FastAPI-based platform to the next level by developing production-grade model-inference services, agentic AI workflows, and seamless integration with third-party LLMs and NLP tooling. In this role, you will be responsible for various key areas: 1. Core Backend Enhancements: - Building APIs - Strengthening security with OAuth2/JWT, rate-limiting, SecretManager, and enhancing observability through structured logging and tracing - Adding CI/CD, test automation, health checks, and SLO dashboards 2. Awesome UI Interfaces: - Developing UI interfaces using React.js/Next.js, Redact/Context, and various CSS frameworks like Tailwind, MUI, Custom-CSS, and Shadcn 3. LLM & Agentic Services: - Designing micro/mini-services to host and route to platforms such as OpenAI, Anthropic, local HF models, embeddings & RAG pipelines - Implementing autonomous/recursive agents that orchestrate multi-step chains for Tools, Memory, and Planning 4. Model-Inference Infrastructure: - Setting up GPU/CPU inference servers behind an API gateway - Optimizing throughput with techniques like batching, streaming, quantization, and caching using tools like Redis and pgvector 5. NLP & Data Services: - Managing the NLP stack with Transformers for classification, extraction, and embedding generation - Building data pipelines to combine aggregated business metrics with model telemetry for analytics You will be working with a tech stack that includes Python, FastAPI, Starlette, Pydantic, Async SQLAlchemy, Postgres, Docker, Kubernetes, AWS/GCP, Redis, RabbitMQ, Celery, Prometheus, Grafana, OpenTelemetry, and more. Experience in building production Python REST APIs, SQL schema design in Postgres, async patterns & concurrency, UI application development, RAG, LLM/embedding workflows, cloud container orchestration, and CI/CD pipelines is essential for this role. Additionally, experience with streaming protocols, NGINX Ingress, SaaS security hardening, data privacy, event-sourced data models, and other related technologies would be advantageous. This role offers the opportunity to work on evolving products, tackle real challenges, and lead the scaling of AI services while working closely with the founder to shape the future of the platform. If you are looking for meaningful ownership and the chance to solve forward-looking problems, this role could be the right fit for you.,
Posted 2 months ago
2.0 - 6.0 years
0 Lacs
hyderabad, telangana
On-site
The company is seeking a Gen AI / AI Engineer with at least 2 years of experience in the field of AI/ML solutions, specifically focusing on Generative AI and RAG (Retrieval-Augmented Generation) architectures. The role involves working on cutting-edge models, fine-tuning, and integrating systems to develop scalable AI products. Responsibilities include fine-tuning and optimizing Large Language Models (LLMs) and generative models for text and image, designing and implementing RAG pipelines with knowledge bases, developing APIs, microservices, and scalable AI pipelines, collaborating with MLOps teams for production deployment, and staying updated on the latest AI trends to apply innovative techniques. The ideal candidate should possess skills in Python, PyTorch/TensorFlow, Generative AI (OpenAI, Anthropic, LLaMA), RAG architecture, Natural Language Processing (NLP), Computer Vision, API development, MLOps, and have experience working with Cloud Platforms such as AWS/Azure/GCP. Qualifications include a Bachelors degree in Computer Science or a related field, at least 2 years of experience in AI/ML with a focus on GenAI & RAG, and hands-on experience with production-grade model deployment. If you are enthusiastic about contributing to the development of next-generation AI solutions, we encourage you to apply for this exciting opportunity now!,
Posted 2 months ago
4.0 - 5.0 years
12 - 14 Lacs
Hyderabad, Pune
Work from Office
Were Hiring | Python Developer – Generative AI Location: Pune/Hyderabad Experience: 4–5 Years Join Us at the Cutting Edge of AI Innovation! Are you passionate about building products powered by Generative AI and LLMs? Do you enjoy working with cutting-edge tools like OpenAI, LangChain, and vector databases to bring AI ideas to life? We’re looking for a Python Developer (GenAI) who is ready to innovate, collaborate, and help us scale real-world AI solutions from prototype to production. Key Responsibilities: • Build scalable Python applications leveraging LLMs, prompt engineering, and RAG pipelines • Develop REST APIs using FastAPI/Flask for GenAI applications • Collaborate across teams to deploy LLM-powered tools (chatbots, assistants, summarizers) • Work with frameworks like LangChain, Hugging Face, and vector DBs like Pinecone, FAISS • Write clean, modular code with performance optimization in mind Skills We’re Looking For: • 4–5 years in Python development • Proven experience building LLM-based applications • Familiarity with OpenAI, Anthropic, Hugging Face, LangChain • Strong knowledge of REST APIs, microservices, and prompt engineering • Exposure to Git, CI/CD workflows, and agile environments Nice-to-Have: • Cloud exposure (AWS, Azure, GCP) • Experience with Docker/Kubernetes • Understanding of RAG pipelines, semantic search, or fine-tuning models Interested or know someone who fits? Drop your resume or referrals in the comments or send us a message! Thanks Dan - Dan@therxcloud.com
Posted 2 months ago
4.0 - 9.0 years
20 - 25 Lacs
Pune
Remote
Hiring AI Engineers | Multiple Roles | Hybrid/ WFH | Immediate| Apply Today Watch JD for more details: https://youtu.be/-YzVGtZ01SQ Unlock Your Future with Expert Recruitment! Explore Active Open Roles: Job's Portfolio : www.youtube.com/@vikranthodage7217
Posted 2 months ago
3.0 - 6.0 years
12 - 18 Lacs
Bengaluru
Work from Office
You will work closely with data scientists, product teams, and developers to fine-tune LLM behavior for diverse applications including chatbots, code generation, creative writing, and data extraction.
Posted 2 months ago
3.0 - 5.0 years
20 - 35 Lacs
Pune
Hybrid
Role Overview :- Monitor, evaluate, and optimize AI/LLM workflows in production environments. Ensure reliable, efficient, and high-quality AI system performance by building out an LLM Ops platform that is self-serve for the engineering and data science departments. Key Responsibilities:- Collaborate with data scientists and software engineers to integrate an LLM Ops platform (Opik by CometML) for existing AI workflows Identify valuable performance metrics (accuracy, quality, etc) for AI workflows and create on-going sampling evaluation processes using the LLM Ops platform that alert when metrics drop below thresholds Cross-team collaboration to create datasets and benchmarks for new AI workflows Run experiments on datasets and optimize performance via model changes and prompt adjustments Debug and troubleshoot AI workflow issues Optimize inference costs and latency while maintaining accuracy and quality Develop automations for LLM Ops platform integration to empower data scientists and software engineers to self-serve integration with the AI workflows they build Requirements:- Strong Python programming skills Experience with generative AI models and tools (OpenAI, Anthropic, Bedrock, etc) Knowledge of fundamental statistical concepts and tools in data science such as: heuristic and non-heuristic measurements in NLP (BLEU, WER, sentiment analysis, LLM-as-judge, etc), standard deviation, sampling rate, and a high level understanding of how modern AI models work (knowledge cutoffs, context windows, temperature, etc) Familiarity with AWS Understanding of prompt engineering concepts People skills: you will be expected to frequently collaborate with other teams to help to perfect their AI workflows Experience Level 3-5 years of experience in LLM/AI Ops, MLOps, Data Science, or MLE
Posted 2 months ago
2.0 - 4.0 years
2 - 7 Lacs
Kolkata, West Bengal, India
On-site
Key Responsibilities: Fine-tune and deploy LLMs (e.g., GPT, LLaMA, Mistral) using frameworks like Hugging Face Transformers and LangChain. Build and optimize RAG pipelines with vector databases (e.g., Pinecone, FAISS, Weaviate). Engineer prompts for structured and reliable outputs across diverse use cases such as chatbots, summarization tools, and coding assistants. Implement scalable inference pipelines; optimize for latency, throughput, and cost using quantization, distillation, and other model optimization techniques. Collaborate with product, design, and engineering teams to integrate generative AI capabilities into user-facing features. Monitor and improve model performance, accuracy, safety, and compliance in production. Ensure responsible AI practices through content filtering, output sanitization, and ethical deployment. Required Skills: Proficiency in Python and familiarity with modern machine learning tools and libraries. Hands-on experience with LLM development using Hugging Face Transformers, LangChain, or LlamaIndex. Experience building and deploying RAG pipelines, including managing embeddings and vector search. Strong understanding of transformer architectures, tokenization, and prompt engineering techniques. Comfortable working with LLM APIs (e.g., OpenAI, Anthropic, Cohere) and serving models with FastAPI, Flask, or similar frameworks. Familiarity with deploying ML systems using Docker, Kubernetes, and cloud services (AWS, GCP, Azure). Experience with model evaluation, logging, and inference pipeline troubleshooting. Nice to Have: Exposure to multimodal models (e.g., text-to-image, video generation, TTS). Experience with reinforcement learning from human feedback (RLHF) or alignment techniques. Familiarity with open-source LLMs (e.g., Mistral, Mixtral, LLaMA, Falcon) and optimization tools (LoRA, quantization, PEFT). Knowledge of LangChain agents, tool integration, and memory management. Contributions to open-source GenAI projects, public demos, or blogs in the generative AI space. Basic proficiency in frontend development (e.g., React, Next.js) for rapid prototyping.
Posted 2 months ago
2.0 - 4.0 years
2 - 7 Lacs
Delhi, India
On-site
Key Responsibilities: Fine-tune and deploy LLMs (e.g., GPT, LLaMA, Mistral) using frameworks like Hugging Face Transformers and LangChain. Build and optimize RAG pipelines with vector databases (e.g., Pinecone, FAISS, Weaviate). Engineer prompts for structured and reliable outputs across diverse use cases such as chatbots, summarization tools, and coding assistants. Implement scalable inference pipelines; optimize for latency, throughput, and cost using quantization, distillation, and other model optimization techniques. Collaborate with product, design, and engineering teams to integrate generative AI capabilities into user-facing features. Monitor and improve model performance, accuracy, safety, and compliance in production. Ensure responsible AI practices through content filtering, output sanitization, and ethical deployment. Required Skills: Proficiency in Python and familiarity with modern machine learning tools and libraries. Hands-on experience with LLM development using Hugging Face Transformers, LangChain, or LlamaIndex. Experience building and deploying RAG pipelines, including managing embeddings and vector search. Strong understanding of transformer architectures, tokenization, and prompt engineering techniques. Comfortable working with LLM APIs (e.g., OpenAI, Anthropic, Cohere) and serving models with FastAPI, Flask, or similar frameworks. Familiarity with deploying ML systems using Docker, Kubernetes, and cloud services (AWS, GCP, Azure). Experience with model evaluation, logging, and inference pipeline troubleshooting. Nice to Have: Exposure to multimodal models (e.g., text-to-image, video generation, TTS). Experience with reinforcement learning from human feedback (RLHF) or alignment techniques. Familiarity with open-source LLMs (e.g., Mistral, Mixtral, LLaMA, Falcon) and optimization tools (LoRA, quantization, PEFT). Knowledge of LangChain agents, tool integration, and memory management. Contributions to open-source GenAI projects, public demos, or blogs in the generative AI space. Basic proficiency in frontend development (e.g., React, Next.js) for rapid prototyping.
Posted 2 months ago
Upload Resume
Drag or click to upload
Your data is secure with us, protected by advanced encryption.
Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.
We have sent an OTP to your contact. Please enter it below to verify.
Accenture
73564 Jobs | Dublin
Wipro
27625 Jobs | Bengaluru
Accenture in India
22690 Jobs | Dublin 2
EY
20638 Jobs | London
Uplers
15021 Jobs | Ahmedabad
Bajaj Finserv
14304 Jobs |
IBM
14148 Jobs | Armonk
Accenture services Pvt Ltd
13138 Jobs |
Capgemini
12942 Jobs | Paris,France
Amazon.com
12683 Jobs |