GenAI Engineer (with Python, LLM, RAG) - Immediate Joiners

4 - 8 years

0 Lacs

Posted:5 days ago| Platform: Shine logo

Apply

Work Mode

On-site

Job Type

Full Time

Job Description

We are looking for a highly motivated Mid-Level AI Engineer to join our growing AI team. Your main responsibility will be to develop intelligent applications using Python, Large Language Models (LLMs), and Retrieval-Augmented Generation (RAG) systems. Working closely with data scientists, backend engineers, and product teams, you will build and deploy AI-powered solutions that provide real-world value. Your key responsibilities will include designing, developing, and optimizing applications utilizing LLMs such as GPT, LLaMA, and Claude. You will also be tasked with implementing RAG pipelines to improve LLM performance using domain-specific knowledge bases and search tools. Developing and maintaining robust Python codebases for AI-driven solutions will be a crucial part of your role. Additionally, integrating vector databases like Pinecone, Weaviate, and FAISS, as well as embedding models for information retrieval, will be part of your daily tasks. You will work with APIs, frameworks like LangChain and Haystack, and various tools to create scalable AI workflows. Collaboration with product and design teams to define AI use cases and deliver impactful features will also be a significant aspect of your job. Conducting experiments to assess model performance, retrieval relevance, and system latency will be essential for continuous improvement. Staying up-to-date with the latest research and advancements in LLMs, RAG, and AI infrastructure is crucial for this role. To be successful in this position, you should have at least 3-5 years of experience in software engineering or AI/ML engineering, with a strong proficiency in Python. Experience working with LLMs such as OpenAI and Hugging Face Transformers is required, along with hands-on experience in RAG architecture and vector-based retrieval techniques. Familiarity with embedding models like SentenceTransformers and OpenAI embeddings is also necessary. Knowledge of API design, deployment, performance optimization, version control (e.g., Git), containerization (e.g., Docker), and cloud platforms (e.g., AWS, GCP, Azure) is expected. Preferred qualifications include experience with LangChain, Haystack, or similar LLM orchestration frameworks. Understanding NLP evaluation metrics, prompt engineering best practices, knowledge graphs, semantic search, and document parsing pipelines will be beneficial. Experience deploying models in production, monitoring system performance, and contributing to open-source AI/ML projects are considered advantageous for this role.,

Mock Interview

Practice Video Interview with JobPe AI

Start Python Interview
cta

Start Your Job Search Today

Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.

Job Application AI Bot

Job Application AI Bot

Apply to 20+ Portals in one click

Download Now

Download the Mobile App

Instantly access job listings, apply easily, and track applications.

coding practice

Enhance Your Python Skills

Practice Python coding challenges to boost your skills

Start Practicing Python Now
Grid Dynamics logo
Grid Dynamics

Information Technology and Services

Los Altos

RecommendedJobs for You