Jobs
Interviews

22 Vllm Jobs

Setup a job Alert
JobPe aggregates results for easy application access, but you actually apply on the job portal directly.

3.0 - 7.0 years

0 Lacs

vadodara, gujarat

On-site

Dharmakit Networks is a premium global IT solutions partner dedicated to innovation and success worldwide. Specializing in website development, SaaS, digital marketing, AI Solutions, and more, we help brands turn their ideas into high-impact digital products. Known for blending global standards with deep Indian insight, we are now stepping into our most exciting chapter yet. Project Ax1 is our next-generation Large Language Model (LLM), a powerful AI initiative designed to make intelligence accessible and impactful for Bharat and the world. Built by a team of AI experts, Dharmakit Networks is committed to developing cost-effective, high-performance AI tailored for India and beyond, enabling enterprises to unlock new opportunities and drive deeper connections. Join us in reshaping the future of AI, starting from India. As a GPU Infrastructure Engineer, you will be at the core of building, optimizing, and scaling the GPU and AI compute infrastructure that powers Project Ax1. Your responsibilities will include designing, deploying, and optimizing GPU infrastructure for large-scale AI workloads, managing GPU clusters across cloud (AWS, Azure, GCP) and on-prem setups, setting up and maintaining model CI/CD pipelines for efficient training and deployment, optimizing LLM inference using TensorRT, ONNX, Nvidia NVCF, and more. You will also be responsible for managing offline/edge deployments of AI models, building and tuning data pipelines to support real-time and batch processing, monitoring model and infra performance for availability, latency, and cost efficiency, and implementing logging, monitoring, and alerting using tools like Prometheus, Grafana, ELK, CloudWatch. Collaboration with AI Experts, ML Experts, backend Experts, and full-stack teams will be essential to ensure seamless model delivery. **Key Responsibilities:** - Design, deploy, and optimize GPU infrastructure for large-scale AI workloads. - Manage GPU clusters across cloud (AWS, Azure, GCP) and on-prem setups. - Set up and maintain model CI/CD pipelines for efficient training and deployment. - Optimize LLM inference using TensorRT, ONNX, Nvidia NVCF, etc. - Manage offline/edge deployments of AI models (e.g., CUDA, Lambda, containerized AI). - Build and tune data pipelines to support real-time and batch processing. - Monitor model and infra performance for availability, latency, and cost efficiency. - Implement logging, monitoring, and alerting using Prometheus, Grafana, ELK, CloudWatch. - Work closely with AI Experts, ML Experts, backend Experts, and full-stack teams to ensure seamless model delivery. **Must-Have Skills And Qualifications:** - Bachelors degree in Computer Science, Engineering, or related field. - Hands-on experience with Nvidia GPUs, CUDA, and deep learning model deployment. - Strong experience with AWS, Azure, or GCP GPU instance setup and scaling. - Proficiency in model CI/CD and automated ML workflows. - Experience with Terraform, Kubernetes, and Docker. - Familiarity with offline/edge AI, including quantization and optimization. - Logging & Monitoring using tools like Prometheus, Grafana, CloudWatch. - Experience with backend APIs, data processing workflows, and ML pipelines. - Experience with Git, collaboration in agile, cross-functional teams. - Strong analytical and debugging skills. - Excellent communication, teamwork, and problem-solving abilities. **Good To Have:** - Experience with Nvidia NVCF, DeepSpeed, vLLM, Hugging Face Triton. - Knowledge of FP16/INT8 quantization, pruning, and other optimization tricks. - Exposure to serverless AI inference (Lambda, SageMaker, Azure ML). - Contributions to open-source AI infrastructure projects or a strong GitHub portfolio showcasing ML model deployment expertise.,

Posted 2 weeks ago

Apply

4.0 - 6.0 years

0 Lacs

mumbai, maharashtra, india

On-site

Requirements 45 years of experience as an ML Engineer or Applied Scientist in production Strong Python skills (FastAPI/Flask), with PyTorch or TensorFlow Proficient in building data pipelines and deploying ML systems Experience with LLMs, embeddings, RAG systems, and vector DBs (PGVector/Postgres preferred) Hands-on with multiple LLM providers (OpenAI, Anthropic, Google, open-source) Knowledge of containerization and deployment (Docker, Kubernetes, AWS/GCP, CI/CD) Bonus: Familiarity with LangGraph/LangChain, vLLM, Ray, LLM eval tools, or event-driven systems Responsibilities Build and scale ML pipelines for Novas AI teammates Implement agentic AI workflows with LLMs and orchestration frameworks Design and optimize RAG pipelines for performance and accuracy Build connectors to integrate AI with external systems Fine-tune models for domains like finance, legal, and tax Monitor model performance, optimize for cost, latency, and reliability Lead best practices for scalable, production-grade ML Job Details Location: Mulund, Mumbai Interview process Round 1: Technical screening Round 2: ML System Design Round 3: Coding and Hands-on Evaluation Round 4: Cultural Fit and Founder Discussion Show more Show less

Posted 2 weeks ago

Apply

8.0 - 12.0 years

0 Lacs

karnataka

On-site

As an Engineering Manager specializing in Data & Insights at a funded Fintech company in Bangalore, India, you will play a pivotal role in reimagining financial products for the next billion users. You will be tasked with addressing the challenge of enabling access to financial services for the population that is currently underserved by traditional channels like insurance, credit, and investments. In this role, you will lead a high-performance team focused on building data-driven products that accelerate time-to-market and foster innovation in credit services. You will be responsible for driving end-to-end product delivery, making critical technical decisions, and nurturing your team of engineers. Your key responsibilities will include owning the delivery of data-centric products, from initial planning to deployment, ensuring high quality and alignment with stakeholders. You will lead technical innovation, particularly in developing scalable backend systems such as APIs, databases, and queues, as well as data platforms like Airflow, Redshift, and Spark. Embracing open-source technologies to create robust and cost-effective solutions will be a key aspect of your role. You will also explore and integrate cutting-edge AI tools and LLMs to push technological boundaries further. As an ideal candidate for this role, you should have 8-10 years of software engineering experience, with at least 6 years dedicated to building backend systems. You should have successfully transitioned into data platform roles, demonstrating proficiency in tools like Airflow, Redshift, and Spark. Your enthusiasm for open-source technologies and familiarity with Apache Spark, Airflow, and similar tools will be highly valued. Knowledge or experience with LLMs and AI, including tools like LangChain, HuggingFace, or vLLM/TensorRT, will be advantageous. Additionally, having at least 1 year of experience in team management, mentoring, and leading projects using Agile principles is desirable. Previous experience in Fintech and a passion for promoting financial inclusion will be considered a significant plus. In return, you will have the opportunity to work alongside founders who have contributed to building foundational technologies like UPI, GST, and Aadhaar. You will have access to unlimited learning resources such as books, conferences, and courses. The company offers top-tier benefits including health insurance, mental health support, and a modern office environment in Bangalore. Joining a diverse and merit-driven team that values craftsmanship in code will be a rewarding aspect of this role.,

Posted 2 weeks ago

Apply

5.0 - 9.0 years

0 Lacs

chennai, tamil nadu

On-site

As an engineer in this role, you will be responsible for building and optimizing high-throughput, low-latency LLM inference infrastructure. This will involve using open-source models such as Qwen, LLaMA, and Mixtral on multi-GPU systems like A100/H100. Your main areas of focus will include performance tuning, model hosting, routing logic, speculative decoding, and cost-efficiency tooling. To excel in this position, you must have deep experience with vLLM, tensor/pipe parallelism, and KV cache management. A strong understanding of CUDA-level inference bottlenecks, FlashAttention2, and quantization is essential. Additionally, familiarity with FP8, INT4, and speculative decoding (e.g., TwinPilots, PowerInfer) will be beneficial. You should also possess a proven ability to scale LLMs across multi-GPU nodes using TP, DDP, and inference routing techniques. Proficiency in Python at a systems level, experience with containerized deployments using tools like Docker, GCP, or AWS, and knowledge of load testing with Locust are required skills for this role. Any experience with any-to-any model routing (e.g., text2sql, speech2text) and exposure to LangGraph, Triton kernels, or custom inference engines will be considered a bonus. Moreover, if you have successfully tuned models for less than $0.50 per million token inference at scale, it will be an advantage. In return for your expertise and contributions, we offer a very competitive rate card for the best candidate who fits this role.,

Posted 2 weeks ago

Apply

8.0 - 10.0 years

0 Lacs

bengaluru, karnataka, india

Remote

Senior Manager - Senior Data Scientist (NLP & Generative AI) Location: PAN India / Remote Employment Type: Full-time About the Role We are seeking a highly experienced Senior data scientist with 8+ years of expertise in machine learning, focusing on NLP, Generative AI, and advanced LLM ecosystems. This role demands leadership in designing and deploying scalable AI systems leveraging the latest advancements such as Google ADK, Agent Engine, and Gemini LLM. You will spearhead building real-time inference pipelines and agentic AI solutions that power complex, multi-user applications with cutting-edge technology. Key Responsibilities Lead the architecture, development, and deployment of scalable machine learning and AI systems centered on real-time LLM inference for concurrent users. Design, implement, and manage agentic AI frameworks leveraging Google Adk, Langgraph or custom-built agents. Integrate foundation models (GPT, LLaMA, Claude, Gemini) and fine-tune them for domain-specific intelligent applications. Build robust MLOps pipelines for end-to-end lifecycle management of models-training, testing, deployment, and monitoring. Collaborate with DevOps teams to deploy scalable serving infrastructures using containerization (Docker), orchestration (Kubernetes), and cloud platforms. Drive innovation by adopting new AI capabilities and tools, such as Google Gemini, to enhance AI model performance and interaction quality. Partner cross-functionally to understand traffic patterns and design AI systems that handle real-world scale and complexity. Required Skills & Qualifications Bachelor's or Master's degree in Computer Science, AI, Machine Learning, or related fields. 7+ years in ML engineering, applied AI, or senior data scientist roles. Strong programming expertise in Python and frameworks including PyTorch, TensorFlow, Hugging Face Transformers. Deep experience with NLP, Transformer models, and generative AI techniques. Practical knowledge of LLM inference scaling with tools like vLLM, Groq, Triton Inference Server, and Google ADK. Hands-on experience deploying AI models to concurrent users with high throughput and low latency. Skilled in cloud environments (AWS, GCP, Azure) and container orchestration (Docker, Kubernetes). Familiarity with vector databases (FAISS, Pinecone, Weaviate) and retrieval-augmented generation (RAG). Experience with agentic AI using Adk, LangChain, Langgraph and Agent Engine Preferred Qualifications Experience with Google Gemini and other advanced LLM innovations. Contributions to open-source AI/ML projects or participation in applied AI research. Knowledge of hardware acceleration and GPU/TPU-based inference optimization. Exposure to event-driven architectures or streaming pipelines (Kafka, Redis).

Posted 2 weeks ago

Apply

0.0 years

0 Lacs

bengaluru, karnataka, india

On-site

Description By applying to this position, your application will be considered for all locations we hire for in the United States. Annapurna Labs designs silicon and software that accelerates innovation. Customers choose us to create cloud solutions that solve challenges that were unimaginable a short time agoeven yesterday. Our custom chips, accelerators, and software stacks enable us to take on technical challenges that have never been seen before, and deliver results that help our customers change the world. Role AWS Neuron is the complete software stack for the AWS Trainium (Trn1/Trn2) and Inferentia (Inf1/Inf2) our cloud-scale Machine Learning accelerators. This role is for a Machine Learning Engineer on one of our AWS Neuron teams: The ML Distributed Training team works side by side with chip architects, compiler engineers and runtime engineers to create, build and tune distributed training solutions with Trainium instances. Experience with training these large models using Python is a must. FSDP (Fully-Sharded Data Parallel), Deepspeed, Nemo and other distributed training libraries are central to this and extending all of this for the Neuron based system is key. ML?Frameworks partners with compiler, runtime, and research experts to make AWS?Trainium and?Inferentia feel native inside the tools builders already lovePyTorch, JAX, and the rapidly evolving vLLM ecosystem. By weaving Neuron?SDK deep into these frameworks, optimizing operators, and crafting targeted extensions, we unlock every teraflop of Annapurnas AI chips for both training and lightning?fast inference. Beyond kernels, we shape next?generation serving by upstreaming new features and driving scalable deployments with vLLM, Triton, and TensorRTturning breakthrough ideas into production?ready AI for millions of customers. The ML Inference team collaborates closely with hardware designers, software optimization experts, and systems engineers to develop and optimize high-performance inference solutions for Inferentia chips. Proficiency in deploying and optimizing ML models for inference using frameworks like TensorFlow, PyTorch, and ONNX is essential. The team focuses on techniques such as quantization, pruning, and model compression to enhance inference speed and efficiency. Adapting and extending popular inference libraries and tools for Neuron-based systems is a key aspect of their work. Key job responsibilities You&aposll join one of our core ML teams - Frameworks, Distributed Training, or Inference - to enhance machine learning capabilities on AWS&aposs specialized AI hardware. Your responsibilities will include improving PyTorch and JAX for distributed training on Trainium chips, optimizing ML models for efficient inference on Inferentia processors, and collaborating with compiler and runtime teams to maximize hardware performance. You&aposll also develop and integrate new features in ML frameworks to support AWS AI services. We seek candidates with strong programming skills, eagerness to learn complex systems, and basic ML knowledge. This role offers growth opportunities in ML infrastructure, bridging the gap between frameworks, distributed systems, and hardware acceleration. About The Team Annapurna Labs was a startup company acquired by AWS in 2015, and is now fully integrated. If AWS is an infrastructure company, then think Annapurna Labs as the infrastructure provider of AWS. Our org covers multiple disciplines including silicon engineering, hardware design and verification, software, and operations. AWS Nitro, ENA, EFA, Graviton and F1 EC2 Instances, AWS Neuron, Inferentia and Trainium ML Accelerators, and in storage with scalable NVMe, Basic Qualifications To qualify, applicants should have earned (or will earn) a Bachelors or Masters degree between December 2022 and September 2025. Working knowledge of C++ and Python Experience with ML frameworks, particularly PyTorch, Jax, and/or vLLM Understanding of parallel computing concepts and CUDA programming Preferred Qualifications Experience in using analytical tools, such as Tableau, Qlikview, QuickSight Experience in building and driving adoption of new tools Our inclusive culture empowers Amazonians to deliver the best results for our customers. If you have a disability and need a workplace accommodation or adjustment during the application and hiring process, including support for the interview or onboarding process, please visit https://amazon.jobs/content/en/how-we-hire/accommodations for more information. If the country/region youre applying in isnt listed, please contact your Recruiting Partner. Company - Annapurna Labs (U.S.) Inc. Job ID: A3029797 Show more Show less

Posted 3 weeks ago

Apply

5.0 - 7.0 years

0 Lacs

Noida, Uttar Pradesh, India

On-site

About the Role: We are seeking an experienced MLOps Engineer to lead the deployment, scaling, and performance optimization of open-source Generative AI models on cloud infrastructure. Youll work at the intersection of machine learning, DevOps, and cloud engineering to help productize and operationalize large-scale LLM and diffusion models. Key Responsibilities: Design and implement scalable deployment pipelines for open-source Gen AI models (LLMs, diffusion models, etc.). Fine-tune and optimize models using techniques like LoRA, quantization, distillation, etc. Manage inference workloads, latency optimization, and GPU utilization. Build CI/CD pipelines for model training, validation, and deployment. Integrate observability, logging, and alerting for model and infrastructure monitoring. Automate resource provisioning using Terraform, Helm, or similar tools on GCP/AWS/Azure. Ensure model versioning, reproducibility, and rollback using tools like MLflow, DVC, or Weights & Biases. Collaborate with data scientists, backend engineers, and DevOps teams to ensure smooth production rollouts. Required Skills & Qualifications: 5+ years of total experience in software engineering or cloud infrastructure. 3+ years in MLOps with direct experience in deploying large Gen AI models. Hands-on experience with open-source models (e.g., LLaMA, Mistral, Stable Diffusion, Falcon, etc.). Strong knowledge of Docker, Kubernetes, and cloud compute orchestration. Proficiency in Python and familiarity with model-serving frameworks (e.g., FastAPI, Triton Inference Server, Hugging Face Accelerate, vLLM). Experience with cloud platforms (GCP preferred, AWS or Azure acceptable). Familiarity with distributed training, checkpointing, and model parallelism. Good to Have: Experience with low-latency inference systems and token streaming architectures. Familiarity with cost optimization and scaling strategies for GPU-based workloads. Exposure to LLMOps tools (LangChain, BentoML, Ray Serve, etc.). Why Join Us: Opportunity to work on cutting-edge Gen AI applications across industries. Collaborative team with deep expertise in AI, cloud, and enterprise software. Flexible work environment with a focus on innovation and impact. Show more Show less

Posted 1 month ago

Apply

4.0 - 8.0 years

5 - 9 Lacs

Bengaluru, Karnataka, India

On-site

We are looking for a highly motivated and skilled AI Software architectto join our team. You will work with a team of Software Engineers to optimize DL models for inference and training, libraries, and applications for Instinct GPUs in both on-prem and Cloud environments. Candidates should be strong in Python and/or C++ and GPU programming. Candidates should also have experience analyzing and optimizing the performance of AI software and understand hardware bottlenecks and harness performance to hit close to roofline. Must be self-motivated and possess the ability to work well within a team environment. KEY QUALIFICATIONS: Strong programming skills in C++ and Python Strong development experience is at least one major DL framework such as vLLM, Pytorch or Tensorflow in inference and/or fine tuning and/or training on multi-node clusters Seeking solid experience in developing kernels, quantizing models and hyper parameter optimizations Experience developing software and system-level performance optimizations with a solid architecture understanding and roofline performance in GPUs MS with years of related experience or PhD with years of related experience in Computer Science or Computer Engineering or related equivalent. Experience with open-source software development including collaboration with community maintainers and submitting contributions is a plus Development experience in CK, Triton and other GPU programming a plus Publications in reputed peer-reviewed ML conferences/journals a plus Excellent analytical and problem-solving skills root-causing/addressing performance issues. Ability to work independently and as part of a team. Willingness to learn skills, tools, and methods to advance the quality, consistency, and timeliness of AMD AI products. PREFERRED EXPERIENCE: Expertise in profiling tools across the AI SW Stack (Torchprofiler, RocM profiler, Vtune, Nsight) Experience in implementing and optimizing parallel methods on GPU accelerators (NCCL/RCCL, OpenMP, MPI) Performance analysis skills for GPUs Experience providing clear and timely communication related to status and other key aspects of the project to leadership team.

Posted 1 month ago

Apply

1.0 - 6.0 years

4 - 8 Lacs

Hyderabad, Telangana, India

On-site

THE ROLE: As a Senior Software Developer, you will develop both GPU kernel-level optimization and distributed software efforts for large-scale AI workloads. This is a technical leadership role with direct influence over critical software components in AMD s AI stack. You ll architect and implement optimized compute kernels, guide software teams through the full product lifecycle, and work closely with internal and external partners to deploy scalable, high-performance solutions. THE PERSON: We re looking for a highly skilled, deep systems thinker who thrives in complex problem domains involving parallel computing, GPU architecture, and AI model execution. You are confident leading software architecture decisions and know how to translate business goals into robust, optimized software solutions. You re just as comfortable writing performance-critical code as you are guiding agile development teams across product lifecycles. Ideal candidates have a strong balance of low-level programming, distributed systems knowledge, and leadership experience paired with a passion for AI performance at scale. KEY RESPONSIBILITIES: GPU Kernel Optimization : Develop and optimize GPU kernels to accelerate inference and training of large machine learning models while ensuring numerical accuracy and runtime efficiency. Multi-GPU and Multi-Node Scaling:Architect and implement strategies for distributed training/inference across multi-GPU/multi-node environments using model/data parallelism techniques. Performance Profiling:Identify bottlenecks and performance limitations using profiling tools; propose and implement optimizations to improve hardware utilization. Parallel Computing : Design and implement multi-threaded and synchronized compute techniques for scalable execution on modern GPU architectures. Benchmarking & Testing:Build robust benchmarking and validation infrastructure to assess performance, reliability, and scalability of deployed software. Documentation & Best Practices:Produce technical documentation and share architectural patterns, code optimization tips, and reusable components. PREFERRED EXPERIENCE: Software Team Leadership Collaboration with customers and business units to define deliverables and roadmaps. Interfacing with executive leadership on program progress and strategic planning. Experience in production-level software deployment (e.g., upstreaming to open source, commercial rollouts). Software Architecture Deep experience with GPU kernel optimization in C++12/17/20 . Working knowledge of frameworks such as PyTorch, vLLM, Cutlass, Kokkos . Practical expertise in CPU/GPU architecture and system-level performance tuning. Proficiency in Python scripting and infrastructure automation. Application of software design patterns and industry-standard engineering practices. GPU & Low-Level Optimization Hands-on experience with CUDA and low-level GPU programming. Kernel optimization in assembly and tight loops for latency-sensitive code. Proficiency with performance profiling tools (Nsight, VTune, Perf, etc.). Experience with distributed computing strategies in AI environments (multi-GPU, NCCL, MPI). Strong debugging, problem-solving, and performance tuning skills in complex systems. ACADEMIC CREDENTIALS: Bachelor s or Master s degree in Computer Engineering, Electrical Engineering, Computer Science, or a related technical field. Advanced degrees or published work in HPC, GPU computing, or AI systems is a plus.

Posted 1 month ago

Apply

5.0 - 10.0 years

0 - 3 Lacs

Bengaluru, Mumbai (All Areas)

Hybrid

Role & responsibilities AI/ML Python developers with Devops 2-3 Deployment(Mandatory). Machine Learning Model Experience. Either AWS Services( Bedrock, sagemaker, EKS, Lambda) / Azure Services is Mandatory Candidates need to work on Gen AI Projects 5yrs - 8yrs - 17.5 LPA 8yrs + - 21 LPA Bangalore & Mumbai

Posted 1 month ago

Apply

3.0 - 5.0 years

8 - 18 Lacs

Bengaluru, Mumbai (All Areas)

Hybrid

Primary Responsibilities: Implement and manage AIOps platforms for intelligent monitoring, alerting, anomaly detection, and root cause analysis (RCA). Possess end-to-end knowledge of VLLM model hosting and inferencing. Advanced knowledge of public cloud platforms such as AWS and Azure. Build and maintain machine learning pipelines and models for predictive maintenance, anomaly detection, and noise reduction. Experience in production support and real-time issue handling. Design dashboards and visualizations to provide operational insights to stakeholders. Working knowledge of Bedrock, SageMaker, EKS, Lambda, etc. 1 to 2 years of experience with Jenkins and GoCD to make build/deploy pipelines. Hands-on experience with open-source and self-hosted model APIs using SDKs. Drive data-driven decisions by analyzing operational data and generating reports on system health, performance, and availability. • Basic knowledge of kserve and rayserve inferencing . • Good knowledge of high level scaling using Karpenter , Keda , System based vertical/horizontal scaling. • Strong knowledge on linux operating system or linux certified . • Previous experience with Helm chart deployments and Terraform template and module creation is highly recommended. Secondary Responsibilities: • Proven experience in AIOps and DevOps, with a strong background in cloud technologies (AWS, Azure, Google Cloud). • Proficiency in tools such as Kubeflow, Kserve, ONNX, and containerization technologies (Docker, Kubernetes). • Experience with enterprise-level infrastructure, including tools like terraform, helm, and On-Prem servers hosting. • Previous experience in fintech or AI based tech companies are highly desirable. • Demonstrates the ability to manage workloads effectively in a production environment. • Possesses excellent communication and collaboration skills, with a strong focus on cross-functional teamwork.

Posted 1 month ago

Apply

3.0 - 7.0 years

0 - 3 Lacs

Bengaluru, Mumbai (All Areas)

Hybrid

Role & responsibilities Primary Responsibilities: Implement and manage AIOps platforms for intelligent monitoring, alerting, anomaly detection, and root cause analysis (RCA). Possess end-to-end knowledge of VLLM model hosting and inferencing. Advanced knowledge of public cloud platforms such as AWS and Azure. Build and maintain machine learning pipelines and models for predictive maintenance, anomaly detection, and noise reduction. Experience in production support and real-time issue handling. Design dashboards and visualizations to provide operational insights to stakeholders. Working knowledge of Bedrock, SageMaker, EKS, Lambda, etc. 1 to 2 years of experience with Jenkins and GoCD to make build/deploy pipelines. Hands-on experience with open-source and self-hosted model APIs using SDKs. Drive data-driven decisions by analyzing operational data and generating reports on system health, performance, and availability. Basic knowledge of kserve and rayserve inferencing . Good knowledge of high level scaling using Karpenter , Keda , System based vertical/horizontal scaling. Strong knowledge on linux operating system or linux certified . Previous experience with Helm chart deployments and Terraform template and module creation is highly recommended. Secondary Responsibilities: Proven experience in AIOps and DevOps, with a strong background in cloud technologies (AWS, Azure, Google Cloud). Proficiency in tools such as Kubeflow, Kserve, ONNX, and containerization technologies (Docker, Kubernetes). Experience with enterprise-level infrastructure, including tools like terraform, helm, and On-Prem servers hosting. Previous experience in fintech or AI based tech companies are highly desirable. Demonstrates the ability to manage workloads effectively in a production environment. Possesses excellent communication and collaboration skills, with a strong focus on cross-functional teamwork.

Posted 1 month ago

Apply

2.0 - 6.0 years

0 Lacs

karnataka

On-site

Tata Electronics Pvt. Ltd. is a key global player in the electronics manufacturing industry, specializing in Electronics Manufacturing Services, Semiconductor Assembly & Test, Semiconductor Foundry, and Design Services. Established in 2020 by the Tata Group, the company's primary objective is to provide integrated solutions to global customers across the electronics and semiconductor value chain. We are looking for an AI Core Developer to join our R&D team in Bangalore. This role is centered around fundamental AI research, algorithm development, and model pre-training, focusing on innovation rather than application engineering. As an AI Core Developer, you will be involved in cutting-edge AI research, creating novel algorithms, and constructing foundation models from scratch. This position is ideal for individuals with a strong background in pre-training methodologies and algorithm development who aspire to contribute to core AI advancements. Your responsibilities will include developing and implementing innovative machine learning algorithms for various AI systems, designing pre-training pipelines for large models, prototyping new AI architectures, collaborating with research scientists and engineers, and contributing to technical publications. The ideal candidate should hold a Bachelor's or Master's degree in Computer Science, Machine Learning, or a related field, with 2-4 years of hands-on experience in AI/ML development. Proficiency in Python, C/C++, knowledge of deep learning frameworks such as PyTorch and TensorFlow, and experience with model pre-training are essential requirements. Strong mathematical skills, familiarity with transformer architectures and attention mechanisms, and understanding of distributed computing are also key competencies. Preferred qualifications include advanced experience in multimodal AI systems, research contributions to top-tier AI conferences, and expertise in specific AI domains like healthcare or finance. The position is based in Bangalore, India, with a hybrid work arrangement and occasional travel for conferences and collaborations.,

Posted 1 month ago

Apply

3.0 - 6.0 years

0 - 3 Lacs

Bengaluru, Mumbai (All Areas)

Work from Office

Role & responsibilities Implement and manage AIOps platforms for intelligent monitoring, alerting, anomaly detection, and root cause analysis (RCA). Possess end-to-end knowledge of VLLM model hosting and inferencing. Advanced knowledge of public cloud platforms such as AWS and Azure. Build and maintain machine learning pipelines and models for predictive maintenance, anomaly detection, and noise reduction. Experience in production support and real-time issue handling. Design dashboards and visualizations to provide operational insights to stakeholders. Working knowledge of Bedrock, SageMaker, EKS, Lambda, etc. 1 to 2 years of experience with Jenkins and GoCD to make build/deploy pipelines. Hands-on experience with open-source and self-hosted model APIs using SDKs. Drive data-driven decisions by analyzing operational data and generating reports on system health, performance, and availability. Basic knowledge of kserve and rayserve inferencing . Good knowledge of high level scaling using Karpenter , Keda , System based vertical/horizontal scaling. Strong knowledge on linux operating system or linux certified . Previous experience with Helm chart deployments and Terraform template and module creation is highly recommended. Secondary Responsibilities: Proven experience in AIOps and DevOps, with a strong background in cloud technologies (AWS, Azure, Google Cloud). Proficiency in tools such as Kubeflow, Kserve, ONNX, and containerization technologies (Docker, Kubernetes). Experience with enterprise-level infrastructure, including tools like terraform, helm, and On-Prem servers hosting. Previous experience in fintech or AI based tech companies are highly desirable. Demonstrates the ability to manage workloads effectively in a production environment. Possesses excellent communication and collaboration skills, with a strong focus on cross-functional teamwork.

Posted 1 month ago

Apply

3.0 - 7.0 years

0 Lacs

hyderabad, telangana

On-site

You will be responsible for designing, building, and deploying scalable NLP/ML models for real-world applications. Your role will involve fine-tuning and optimizing Large Language Models (LLMs) using techniques like LoRA, PEFT, or QLoRA. You will work with transformer-based architectures such as BERT, GPT, LLaMA, and T5, and develop GenAI applications using frameworks like LangChain, Hugging Face, OpenAI API, or RAG (Retrieval-Augmented Generation). Writing clean, efficient, and testable Python code will be a crucial part of your tasks. Collaboration with data scientists, software engineers, and stakeholders to define AI-driven solutions will also be an essential aspect of your work. Additionally, you will evaluate model performance and iterate rapidly based on user feedback and metrics. The ideal candidate should have a minimum of 3 years of experience in Python programming with a strong understanding of ML pipelines. A solid background and experience in NLP, including text preprocessing, embeddings, NER, and sentiment analysis, are required. Proficiency in ML libraries such as scikit-learn, PyTorch, TensorFlow, Hugging Face Transformers, and spaCy is essential. Experience with GenAI concepts, including prompt engineering, LLM fine-tuning, and vector databases like FAISS and ChromaDB, will be beneficial. Strong problem-solving and communication skills are highly valued, along with the ability to learn new tools and work both independently and collaboratively in a fast-paced environment. Attention to detail and accuracy is crucial for this role. Preferred skills include theoretical knowledge or experience in Data Engineering, Data Science, AI, ML, RPA, or related domains. Certification in Business Analysis or Project Management from a recognized institution is a plus. Experience in working with agile methodologies such as Scrum or Kanban is desirable. Additional experience in deep learning and transformer architectures and models, prompt engineering, training LLMs, and GenAI pipeline preparation will be advantageous. Practical experience in integrating LLM models like ChatGPT, Gemini, Claude, etc., with context-aware capabilities using RAG or fine-tuning models is a plus. Knowledge of model evaluation and alignment, as well as metrics to calculate model accuracy, is beneficial. Data curation from sources for RAG preprocessing and development of LLM pipelines is an added advantage. Proficiency in scalable deployment and logging tooling, including skills like Flask, Django, FastAPI, APIs, Docker containerization, and Kubeflow, is preferred. Familiarity with Lang Chain, LlamaIndex, vLLM, HuggingFace Transformers, LoRA, and a basic understanding of cost-to-performance tradeoffs will be beneficial for this role.,

Posted 2 months ago

Apply

5.0 - 9.0 years

0 Lacs

kochi, kerala

On-site

As a highly skilled Senior Machine Learning Engineer, you will leverage your expertise in Deep Learning, Large Language Models (LLMs), and MLOps/LLMOps to design, optimize, and deploy cutting-edge AI solutions. Your responsibilities will include developing and scaling deep learning models, fine-tuning LLMs (e.g., GPT, Llama), and implementing robust deployment pipelines for production environments. You will be responsible for designing, training, fine-tuning, and optimizing deep learning models (CNNs, RNNs, Transformers) for various applications such as NLP, computer vision, or multimodal tasks. Additionally, you will fine-tune and adapt LLMs for domain-specific tasks like text generation, summarization, and semantic similarity. Experimenting with RLHF (Reinforcement Learning from Human Feedback) and alignment techniques will also be part of your role. In the realm of Deployment & Scalability (MLOps/LLMOps), you will build and maintain end-to-end ML pipelines for training, evaluation, and deployment. Deploying LLMs and deep learning models in production environments using frameworks like FastAPI, vLLM, or TensorRT is crucial. You will optimize models for low-latency, high-throughput inference and implement CI/CD workflows for ML systems using tools like MLflow and Kubeflow. Monitoring & Optimization will involve setting up logging, monitoring, and alerting for model performance metrics such as drift, latency, and accuracy. Collaborating with DevOps teams to ensure scalability, security, and cost-efficiency of deployed models will also be part of your responsibilities. The ideal candidate will possess 5-7 years of hands-on experience in Deep Learning, NLP, and LLMs. Strong proficiency in Python, PyTorch, TensorFlow, Hugging Face Transformers, and LLM frameworks is essential. Experience with model deployment tools like Docker, Kubernetes, and FastAPI, along with knowledge of MLOps/LLMOps best practices and familiarity with cloud platforms (AWS, GCP, Azure) are required qualifications. Preferred qualifications include contributions to open-source LLM projects, showcasing your commitment to advancing the field of machine learning.,

Posted 2 months ago

Apply

12.0 - 14.0 years

0 Lacs

Hyderabad, Telangana, India

On-site

Our vision is to transform how the world uses information to enrich life for . Micron Technology is a world leader in innovating memory and storage solutions that accelerate the transformation of information into intelligence, inspiring the world to learn, communicate and advance faster than ever. Principal / Senior Systems Performance Engineer Micron Data Center and Client Workload Engineering in Hyderabad, India, is seeking a senior/principal engineer to join our dynamic team. The successful candidate will primarily contribute to the ML development, ML DevOps, HBM program in the data center by analyzing how AI/ML workloads perform on the latest MU-HBM, Micron main memory, expansion memory and near memory (HBM/LP) solutions, conduct competitive analysis, showcase the benefits that workloads see with MU-HBM's capacity / bandwidth / thermals, contribute to marketing collateral, and extract AI/ML workload traces to help optimize future HBM designs. Job Responsibilities: The Job Responsibilities include but are not limited to the following: Design, implement, and maintain scalable & reliable ML infrastructure and pipelines. Collaborate with data scientists and ML engineers to deploy machine learning models into production environments. Automate and optimize ML workflows, including data preprocessing, model training, evaluation, and deployment. Monitor and manage the performance, reliability, and scalability of ML systems. Troubleshoot and resolve issues related to ML infrastructure and deployments. Implement and manage distributed training and inference solutions to enhance model performance and scalability. Utilize DeepSpeed, TensorRT, vLLM for optimizing and accelerating AI inference and training processes. Understand key care abouts when it comes to ML models such as: transformer architectures, precision, quantization, distillation, attention span & KV cache, MoE, etc. Build workload memory access traces from AI models. Study system balance ratios for DRAM to HBM in terms of capacity and bandwidth to understand and model TCO. Study data movement between CPU, GPU and the associated memory subsystems (DDR, HBM) in heterogeneous system architectures via connectivity such as PCIe/NVLINK/Infinity Fabric to understand the bottlenecks in data movement for different workloads. Develop an automated testing framework through scripting. Customer engagements and conference presentations to showcase findings and develop whitepapers. Requirements: Strong programming skills in Python and familiarity with ML frameworks such as TensorFlow, PyTorch, or scikit-learn. Experience in data preparation: cleaning, splitting, and transforming data for training, validation, and testing. Proficiency in model training and development: creating and training machine learning models. Expertise in model evaluation: testing models to assess their performance. Skills in model deployment: launching server, live inference, batched inference Experience with AI inference and distributed training techniques. Strong foundation in GPU and CPU processor architecture Familiarity with and knowledge of server system memory (DRAM) Strong experience with benchmarking and performance analysis Strong software development skills using leading scripting, programming languages and technologies (Python, CUDA, C, C++) Familiarity with PCIe and NVLINK connectivity Preferred Qualifications: Experience in quickly building AI workflows: building pipelines and model workflows to design, deploy, and manage consistent model delivery. Ability to easily deploy models anywhere: using managed endpoints to deploy models and workflows across accessible CPU and GPU machines. Understanding of MLOps: the overarching concept covering the core tools, processes, and best practices for end-to-end machine learning system development and operations in production. Knowledge of GenAIOps: extending MLOps to develop and operationalize generative AI solutions, including the management of and interaction with a foundation model. Familiarity with LLMOps: focused specifically on developing and productionizing LLM-based solutions. Experience with RAGOps: focusing on the delivery and operation of RAGs, considered the ultimate reference architecture for generative AI and LLMs. Data management: collect, ingest, store, process, and label data for training and evaluation. Configure role-based access control dataset search, browsing, and exploration data provenance tracking, data logging, dataset versioning, metadata indexing, data quality validation, dataset cards, and dashboards for data visualization. Workflow and pipeline management: work with cloud resources or a local workstation connect data preparation, model training, model evaluation, model optimization, and model deployment steps into an end-to-end automated and scalable workflow combining data and compute. Model management: train, evaluate, and optimize models for production store and version models along with their model cards in a centralized model registry assess model risks, and ensure compliance with standards. Experiment management and observability: track and compare different machine learning model experiments, including changes in training data, models, and hyperparameters. Automatically search the space of possible model architectures and hyperparameters for a given model architecture analyze model performance during inference, monitor model inputs and outputs for concept drift. Synthetic data management: extend data management with a new native generative AI capability. Generate synthetic training data through domain randomization to increase transfer learning capabilities. Declaratively define and generate edge cases to evaluate, validate, and certify model accuracy and robustness. Embedding management: represent data samples of any modality as dense multi-dimensional embedding vectors generate, store, and version embeddings in a vector database. Visualize embeddings for improvised exploration. Find relevant contextual information through vector similarity search for RAGs. Education: Bachelor's or higher (with 12+ years of experience) in Computer Science or related field.

Posted 2 months ago

Apply

8.0 - 10.0 years

0 Lacs

Noida, Uttar Pradesh, India

Remote

Senior Manager - Senior Data Scientist (NLP & Generative AI) Location: PAN India / Remote Employment Type: Full-time About the Role We are seeking a highly experienced Senior data scientist with 8+ years of expertise in machine learning, focusing on NLP, Generative AI, and advanced LLM ecosystems. This role demands leadership in designing and deploying scalable AI systems leveraging the latest advancements such as Google ADK, Agent Engine, and Gemini LLM. You will spearhead building real-time inference pipelines and agentic AI solutions that power complex, multi-user applications with cutting-edge technology. Key Responsibilities Lead the architecture, development, and deployment of scalable machine learning and AI systems centered on real-time LLM inference for concurrent users. Design, implement, and manage agentic AI frameworks leveraging Google Adk, Langgraph or custom-built agents. Integrate foundation models (GPT, LLaMA, Claude, Gemini) and fine-tune them for domain-specific intelligent applications. Build robust MLOps pipelines for end-to-end lifecycle management of models-training, testing, deployment, and monitoring. Collaborate with DevOps teams to deploy scalable serving infrastructures using containerization (Docker), orchestration (Kubernetes), and cloud platforms. Drive innovation by adopting new AI capabilities and tools, such as Google Gemini, to enhance AI model performance and interaction quality. Partner cross-functionally to understand traffic patterns and design AI systems that handle real-world scale and complexity. Required Skills & Qualifications Bachelor's or Master's degree in Computer Science, AI, Machine Learning, or related fields. 7+ years in ML engineering, applied AI, or senior data scientist roles. Strong programming expertise in Python and frameworks including PyTorch, TensorFlow, Hugging Face Transformers. Deep experience with NLP, Transformer models, and generative AI techniques. Practical knowledge of LLM inference scaling with tools like vLLM, Groq, Triton Inference Server, and Google ADK. Hands-on experience deploying AI models to concurrent users with high throughput and low latency. Skilled in cloud environments (AWS, GCP, Azure) and container orchestration (Docker, Kubernetes). Familiarity with vector databases (FAISS, Pinecone, Weaviate) and retrieval-augmented generation (RAG). Experience with agentic AI using Adk, LangChain, Langgraph and Agent Engine Preferred Qualifications Experience with Google Gemini and other advanced LLM innovations. Contributions to open-source AI/ML projects or participation in applied AI research. Knowledge of hardware acceleration and GPU/TPU-based inference optimization. Exposure to event-driven architectures or streaming pipelines (Kafka, Redis).

Posted 2 months ago

Apply

11.0 - 20.0 years

40 - 50 Lacs

Pune, Chennai, Bengaluru

Hybrid

Senior xOps Specialist AIOps, MLOps & DataOps Architect Location: Chennai, Pune Employment Type: Fulltime - Hybrid Experience Required: 12-15 years Job Summary: We are seeking a Senior xOps Specialist to architect, implement, and optimize AI-driven operational frameworks across AIOps, MLOps, and DataOps. The ideal candidate will design and enhance intelligent automation, predictive analytics, and resilient pipelines for large-scale data engineering, AI/ML deployments, and IT operations. This role requires deep expertise in AI/ML automation, data-driven DevOps strategies, observability frameworks, and cloud-native orchestration. Key Responsibilities – Design & Architecture AIOps: AI-Driven IT Operations & Automation Architect AI-powered observability platforms, ensuring predictive incident detection and autonomous IT operations. Implement AI-driven root cause analysis (RCA) for proactive issue resolution and performance optimization. Design self-healing infrastructures leveraging machine learning models for anomaly detection and remediation workflows. Establish event-driven automation strategies, enabling autonomous infrastructure scaling and resilience engineering. MLOps: Machine Learning Lifecycle Optimization Architect end-to-end MLOps pipelines, ensuring automated model training, validation, deployment, and monitoring. Design CI/CD pipelines for ML models, embedding drift detection, continuous optimization, and model explainability. Implement feature engineering pipelines, leveraging data versioning, reproducibility, and intelligent retraining techniques. Ensure secure and scalable AI/ML environments, optimizing GPU-accelerated processing and cloud-native model serving. DataOps: Scalable Data Engineering & Pipelines Architect data processing frameworks, ensuring high-performance, real-time ingestion, transformation, and analytics. Build data observability platforms, enabling automated anomaly detection, data lineage tracking, and schema evolution. Design self-optimizing ETL pipelines, leveraging AI-driven workflows for data enrichment and transformation. Implement governance frameworks, ensuring data quality, security, and compliance with enterprise standards. Automation & API Integration Develop Python or Go-based automation scripts for AI model orchestration, data pipeline optimization, and IT workflows. Architect event-driven xOps frameworks, enabling intelligent orchestration for real-time workload management. Implement AI-powered recommendations, optimizing resource allocation, cost efficiency, and performance benchmarking. Cloud-Native & DevOps Integration Embed AI/ML observability principles within DevOps pipelines, ensuring continuous monitoring and retraining cycles. Architect cloud-native solutions optimized for Kubernetes, containerized environments, and scalable AI workloads. Establish AIOps-driven cloud infrastructure strategies, automating incident response and operational intelligence. Qualifications & Skills – xOps Expertise Deep expertise in AIOps, MLOps, and DataOps, designing AI-driven operational frameworks. Proficiency in automation scripting, leveraging Python, Go, and AI/ML orchestration tools. Strong knowledge of AI observability, ensuring resilient IT operations and predictive analytics. Extensive experience in cloud-native architectures, Kubernetes orchestration, and serverless AI workloads. Ability to troubleshoot complex AI/ML pipelines, ensuring optimal model performance and data integrity. Preferred Certifications (Optional): AWS Certified Machine Learning Specialist Google Cloud Professional Data Engineer Kubernetes Certified Administrator (CKA) DevOps Automation & AIOps Certification

Posted 2 months ago

Apply

3.0 - 5.0 years

16 - 20 Lacs

Noida

Work from Office

Position Title: AI/ML Engineer Company: Cyfuture India Pvt. Ltd. Industry: IT Services and IT Consulting Location: Sector 81, NSEZ, Noida (5 Days Work From Office) Website: www.cyfuture.com About Cyfuture Cyfuture is a trusted name in IT services and cloud infrastructure, offering state-of-the-art data center solutions and managed services across platforms like AWS, Azure, and VMWare. We are expanding rapidly in system integration and managed services, building strong alliances with global OEMs like VMWare, AWS, Azure, HP, Dell, Lenovo, and Palo Alto. Position Overview We are hiring an experienced AI/ML Engineer to lead and shape our AI/ML initiatives. The ideal candidate will have hands-on experience in machine learning and artificial intelligence, with strong leadership capabilities and a passion for delivering production-ready solutions. This role involves end-to-end ownership of AI/ML projects, from strategy development to deployment and optimization of large-scale systems. Key Responsibilities Lead and mentor a high-performing AI/ML team. Design and execute AI/ML strategies aligned with business goals. Collaborate with product and engineering teams to identify impactful AI opportunities. Build, train, fine-tune, and deploy ML models in production environments. Manage operations of LLMs and other AI models using modern cloud and MLOps tools. Implement scalable and automated ML pipelines (e.g., with Kubeflow or MLRun). Handle containerization and orchestration using Docker and Kubernetes. Optimize GPU/TPU resources for training and inference tasks. Develop efficient RAG pipelines with low latency and high retrieval accuracy. Automate CI/CD workflows for continuous integration and delivery of ML systems. Key Skills & Expertise 1. Cloud Computing & Deployment Proficiency in AWS, Google Cloud, or Azure for scalable model deployment. Familiarity with cloud-native services like AWS SageMaker, Google Vertex AI, or Azure ML. Expertise in Docker and Kubernetes for containerized deployments Experience with Infrastructure as Code (IaC) using tools like Terraform or CloudFormation. 2. Machine Learning & Deep Learning Strong command of frameworks: TensorFlow, PyTorch, Scikit-learn, XGBoost. Experience with MLOps tools for integration, monitoring, and automation. Expertise in pre-trained models, transfer learning, and designing custom architectures. 3. Programming & Software Engineering Strong skills in Python (NumPy, Pandas, Matplotlib, SciPy) for ML development. Backend/API development with FastAPI , Flask , or Django . Database handling with SQL and NoSQL (PostgreSQL, MongoDB, BigQuery). Familiarity with CI/CD pipelines (GitHub Actions, Jenkins). 4. Scalable AI Systems Proven ability to build AI-driven applications at scale. Handle large datasets, high-throughput requests, and real-time inference. Knowledge of distributed computing: Apache Spark, Dask, Ray . 5. Model Monitoring & Optimization Hands-on with model compression, quantization, and pruning . A/B testing and performance tracking in production. Knowledge of model retraining pipelines for continuous learning. 6. Resource Optimization Efficient use of compute resources: GPUs, TPUs, CPUs . Experience with serverless architectures to reduce cost. Auto-scaling and load balancing for high-traffic systems. 7. Problem-Solving & Collaboration Translate complex ML models into user-friendly applications. Work effectively with data scientists, engineers, and product teams. Write clear technical documentation and architecture reports . Udisha Parashar Senior Talent Acquisition Specialist Mob: +91- 9301895707 Email: udisha.parashar@cyfuture.com URL: www.cyfuture.com

Posted 3 months ago

Apply

17 - 27 years

100 - 200 Lacs

Bengaluru

Work from Office

Senior Software Technical Director / Software Technical Director Bangalore Founded in 2023,by Industry veterans HQ in California,US We are revolutionizing sustainable AI compute through intuitive software with composable silicon We are looking for a Software Technical Director with a strong technical foundation in systems software, Linux platforms, or machine learning compiler stacks to lead and grow a high-impact engineering team in Bangalore. You will be responsible for shaping the architecture, contributing to codebases, and managing execution across projects that sit at the intersection of systems programming, AI runtimes, and performance-critical software. Key Responsibilities: Technical Leadership: Lead the design and development of Linux platform software, firmware, or ML compilers and runtimes. Drive architecture decisions across compiler, runtime, or low-level platform components. Write production-grade C++ code and perform detailed code reviews. Guide performance analysis and debugging across the full stackfrom firmware and drivers to user-level runtime libraries. Collaborate with architects, silicon teams, and ML researchers to build future-proof software stacks. Team & Project Management: Mentor and coach junior and senior engineers to grow technical depth and autonomy. Own end-to-end project planning, execution, and delivery, ensuring high-quality output across sprints/releases. Facilitate strong cross-functional communication with hardware, product, and other software teams globally. Recruit and grow a top-tier engineering team in Bangalore, contributing to the hiring strategy and team culture. Required Qualifications: Bachelors or Master’s degree in Computer Science, Electrical Engineering, or related field. 18+ years of experience in systems software development with significant time spent in C++, including architectural and hands-on roles. Proven experience in either: Linux kernel, bootloaders, firmware, or low-level platform software, or Machine Learning compilers (e.g., MLIR, TVM, Glow) or runtimes (e.g., ONNX Runtime, TensorRT, vLLM). Excellent communication skills—written and verbal. Prior experience in project leadership or engineering management with direct reports. Highly Desirable: Understanding of AI/ML compute workloads, particularly Large Language Models (LLMs). Familiarity with performance profiling, bottleneck analysis, and compiler-level optimizations. Exposure to AI accelerators, systolic arrays, or vector SIMD programming. Why Join Us? Work at the forefront of AI systems software, shaping the future of ML compilers and runtimes. Collaborate with globally distributed teams in a fast-paced, innovation-driven environment. Build and lead a technically elite team from the ground up in a growth-stage organization. Contact: Uday Mulya Technologies muday_bhaskar@yahoo.com "Mining The Knowledge Community"

Posted 4 months ago

Apply

5.0 - 10.0 years

1 - 2 Lacs

chennai

Work from Office

Strong experience with vLLM, HuggingFace Transformers, LoRA/QLoRA, and cybersecurity data— MITRE ATT&CK, CVE/NVD, YARA rules, Snort/Suricata rules, STIX/TAXII, or malware datasets, Python, ML libraries (PyTorch, Transformers), and MLOps practices.

Posted Date not available

Apply
cta

Start Your Job Search Today

Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.

Job Application AI Bot

Job Application AI Bot

Apply to 20+ Portals in one click

Download Now

Download the Mobile App

Instantly access job listings, apply easily, and track applications.

Featured Companies