Get alerts for new jobs matching your selected skills, preferred locations, and experience range.
5.0 years
0 Lacs
India
On-site
This posting is for one of our International Clients. About the Role We’re creating a new certification: Inside Gemini: Gen AI Multimodal and Google Intelligence (Google DeepMind) . This course is designed for technical learners who want to understand and apply the capabilities of Google’s Gemini models and DeepMind technologies to build powerful, multimodal AI applications. We’re looking for a Subject Matter Expert (SME) who can help shape this course from the ground up. You’ll work closely with a team of learning experience designers, writers, and other collaborators to ensure the course is technically accurate, industry-relevant, and instructionally sound. Responsibilities As the SME, you’ll partner with learning experience designers and content developers to: Translate real-world Gemini and DeepMind applications into accessible, hands-on learning for technical professionals. Guide the creation of labs and projects that allow learners to build pipelines for image-text fusion, deploy Gemini APIs, and experiment with DeepMind’s reinforcement learning libraries. Contribute technical depth across activities, from high-level course structure down to example code, diagrams, voiceover scripts, and data pipelines. Ensure all content reflects current, accurate usage of Google’s multimodal tools and services. Be available during U.S. business hours to support project milestones, reviews, and content feedback. This role is an excellent fit for professionals with deep experience in AI/ML, Google Cloud, and a strong familiarity with multimodal systems and the DeepMind ecosystem. Essential Tools & Platforms A successful SME in this role will demonstrate fluency and hands-on experience with the following: Google Cloud Platform (GCP) Vertex AI (particularly Gemini integration, model tuning, and multimodal deployment) Cloud Functions, Cloud Run (for inference endpoints) BigQuery and Cloud Storage (for handling large image-text datasets) AI Platform Notebooks or Colab Pro Google DeepMind Technologies JAX and Haiku (for neural network modeling and research-grade experimentation) DeepMind Control Suite or DeepMind Lab (for reinforcement learning demonstrations) RLax or TF-Agents (for building and modifying RL pipelines) AI/ML & Multimodal Tooling Gemini APIs and SDKs (image-text fusion, prompt engineering, output formatting) TensorFlow 2.x and PyTorch (for model interoperability) Label Studio, Cloud Vision API (for annotation and image-text preprocessing) Data Science & MLOps DVC or MLflow (for dataset and model versioning) Apache Beam or Dataflow (for processing multimodal input streams) TensorBoard or Weights & Biases (for visualization) Content Authoring & Collaboration GitHub or Cloud Source Repositories Google Docs, Sheets, Slides Screen recording tools like Loom or OBS Studio Required skills and experience: Demonstrated hands-on experience building, deploying, and maintaining sophisticated AI powered applications using Gemini APIs/SDKs within the Google Cloud ecosystem, especially in Firebase Studio and VS Code. Proficiency in designing and implementing agent-like application patterns, including multi-turn conversational flows, state management, and complex prompting strategies (e.g., Chain-of Thought, few-shot, zero-shot). Experience integrating Gemini with Google Cloud services (Firestore, Cloud Functions, App Hosting) and external APIs for robust, production-ready solutions. Proven ability to engineer applications that process, integrate, and generate content across multiple modalities (text, images, audio, video, code) using Gemini’s native multimodal capabilities. Skilled in building and orchestrating pipelines for multimodal data handling, synchronization, and complex interaction patterns within application logic. Experience designing and implementing production-grade RAG systems, including integration with vector databases (e.g., Pinecone, ChromaDB) and engineering data pipelines for indexing and retrieval. Ability to manage agent state, memory, and persistence for multi-turn and long-running interactions. Proficiency leveraging AI-assisted coding features in Firebase Studio (chat, inline code, command execution) and using App Prototyping agents or frameworks like Genkit for rapid prototyping and structuring agentic logic. Strong command of modern development workflows, including Git/GitHub, code reviews, and collaborative development practices. Experience designing scalable, fault-tolerant deployment architectures for multimodal and agentic AI applications using Firebase App Hosting, Cloud Run, or similar serverless/cloud platforms. Advanced MLOps skills, including monitoring, logging, alerting, and versioning for generative AI systems and agents. Deep understanding of security best practices: prompt injection mitigation (across modalities), secure API key management, authentication/authorization, and data privacy. Demonstrated ability to engineer for responsible AI, including bias detection, fairness, transparency, and implementation of safety mechanisms in agentic and multimodal applications. Experience addressing ethical challenges in the deployment and operation of advanced AI systems. Proven success designing, reviewing, and delivering advanced, project-based curriculum and hands-on labs for experienced software developers and engineers. Ability to translate complex engineering concepts (RAG, multimodal integration, agentic patterns, MLOps, security, responsible AI) into clear, actionable learning materials and real world projects. 5+ years of professional experience in AI-powered application development, with a focus on generative and multimodal AI. Strong programming skills in Python and JavaScript/TypeScript; experience with modern frameworks and cloud-native development. Bachelor’s or Master’s degree in Computer Science, Data Engineering, AI, or a related technical field. Ability to explain advanced technical concepts (e.g., fusion transformers, multimodal embeddings, RAG workflows) to learners in an accessible way. Strong programming experience in Python and experience deploying machine learning pipelines Ability to work independently, take ownership of deliverables, and collaborate closely with designers and project managers Preferred: Experience with Google DeepMind tools (JAX, Haiku, RLax, DeepMind Control Suite/Lab) and reinforcement learning pipelines. Familiarity with open data formats (Delta, Parquet, Iceberg) and scalable data engineering practices. Prior contributions to open-source AI projects or technical community engagement. Show more Show less
Posted 22 hours ago
10.0 - 15.0 years
0 Lacs
India
On-site
About Us: At Articul8 AI, we relentlessly pursue excellence and create exceptional GenAI products that exceed customer expectations. We are a team of dedicated individuals who take pride in our work and strive for greatness in every aspect of our business. We believe in using our advantages to make a positive impact on the world and inspiring others to do the same Job Description: Articul8 AI is seeking a Data Scientist to design, develop, and deploy AI-driven solutions that solve real-world problems at scale. You will work on machine learning models, large language models (LLMs), and AI applications while optimizing performance for production environments. This role requires expertise in AI/ML frameworks, cloud platforms, and software engineering best practices. You will be developing and deploying advanced deep learning and generative AI models and algorithms to enhance existing products or to create new products that fulfill critical business needs. In this role, you will be working closely with Product Management and Engineering teams to build GenAI products at scale. You will be responsible for transforming business needs to technical requirements and for leveraging state of the art research to develop and deliver products. You will also support Engineering with testing and validation of the product. Design, develop, and deploy AI-driven solution in production that solve real-world problems at scale. Train, fine-tune, and optimize deep learning and LLM-based solutions to enhance existing products or to create new products. Evaluate and implement state-of-the-art AI/ML algorithms to improve model accuracy and efficiency to enhance and deliver product. Optimize models ensuring low latency and high availability for cloud and on-prem environments. Collaborate closely with engineering teams and product management to build GenAI products at scale. Work with large-scale datasets, ensuring data quality, preprocessing, and feature engineering. Develop APIs and microservices to serve AI models in production at scale. Handle large-scale datasets, preprocessing, and feature engineering to ensuring data quality. Responsible for transforming business needs to technical requirements to develop and deliver products. Stay up to date with the latest AI trends, research, and best practices. Ensure ethical AI practices, data privacy, and security compliance Required Qualifications Master’s Degree in Science, Technology, Engineering and Mathematics (STEM) or Statistics with 10 to 15 years of experience In-depth knowledge and experience with algorithms for time series analysis including data pre-processing, pattern recognition, clustering, modeling and anomaly detection. Strong expertise in Deep Learning, Machine Learning and Generative AI models (including Language, Vision, Audio and Multi-modal models) Exposure to one or more of the following domains - Optimization, Stochastic Processes, Estimation theory Experience in deploying deep learning models on multiple GPUs Experience in developing models and algorithms using ML frameworks like PyTorch, TensorFlow Strong programming skills in one or more of the following languages - Python, Golang Experience in building Docker images in creating scalable, efficient, and portable applications Experience in Kubernetes for container orchestration and writing YAML manifests to define how applications and services should be deployed Knowledge of cloud platforms at least one of AWS, Azure, GCP and its services for deployment of applications Strong verbal and written communications skills. Preferred Qualifications Ph.D, in Science, Technology, Engineering and Mathematics (STEM) or Statistics with 6 to 8 years of experience. Deep expertise and experience in training/fine-tuning large language models on large GPU clusters. Experience in parallel programming including data, model and tensor parallelisms with PyTorch and TensorFlow Deep experience in building and scaling machine learning / deep learning or GenAI applications with Docker and Kubernetes. Strong working experience with at least two cloud service providers (AWS, Azure GCP). Knowledge of CI/CD pipelines, MLOps like MLflow, Kubeflow, or TensorBoard. Deep expertise and experience in one or more of the following areas like finance, healthcare, engineering. Ability to transform business needs to technical requirements, define tasks, metrics and milestones. Ability to communicate technological challenges and achievements to various stakeholders. What We Offer: By joining our team, you become part of a community that embraces diversity, inclusiveness, and lifelong learning. We nurture curiosity and creativity, encouraging exploration beyond conventional wisdom. Through mentorship, knowledge exchange, and constructive feedback, we cultivate an environment that supports both personal and professional development. If you're ready to join a team that's changing the game, apply now to become a part of the Articul8 team. Join us on this adventure and help shape the future of Generative AI in the enterprise. Show more Show less
Posted 3 days ago
3.0 years
0 Lacs
Mumbai Metropolitan Region
On-site
Work Experience : 3+ years Salary: 21 LPA Location: Bengaluru Title : MLops Engineer Team Charter: The team in India comes with multi-disciplinary skillset, including but not limited to the following areas: Develop models and algorithms using Deep Learning and Computer Vision on the captured data to provide meaningful analysis to our customers. Some of the projects include – object detection, OCR, barcode scanning, stereovision, SLAM, 3D-reconstruction, action recognition etc. Develop integrated embedded systems for our drones – including embedded system platform development, camera and sensor integration, flight controller and motor control system development, etc. Architect and develop full stack software to interface between our solution and customer database and access – including database development, API development, UI/UX, storage, security and processing for data acquired by the drone. Integration and testing of various off the shelf sensors and other modules with drone and related software. Design algorithms related to autonomy and flight controls. Responsibilities: As a Machine Learning Ops (MLOps) engineer, you will be responsible for building and maintaining the next generation of Vimaan’s ML Platform and Infrastructure. MLOps will have a major contribution in making CV & ML offerings scalable across the company products. We are building all these data & model pipelines to scale Vimaan operations and MLOps Engineer will play a key role in enabling that. You will lead initiatives geared towards making the Computer Vision Engineers at Vimaan more productive. You will setup the infrastructure that powers the ML teams, thus simplifying the development and deployment cycles of ML models. You will help establish best practices for the ML pipeline and partner with other infrastructure ops teams to help champion them across the company. Build and maintain data pipelines - data ingestion, filtering, generating pre-populated annotations, etc. Build and maintain model pipelines - model monitoring, automated triggering of model (re)training, auto-deployment of models to producti on servers and edge devices. Own the cloud stack which comprises all ML resources. Establish standards and practices around MLOps, including governance, compliance, and data security. Collaborate on managing ML infrastructure costs. Qualifications: Deep quantitative/programming background with degree (Bachelors, Masters or Ph.D.) in a highly analytical discipline, like Statistics, Electrical,Electronics, Computer Science, Mathematics, Operations Research, etc. A minimum of 3 years of experience in managing machine learning projects end-to-end focused on MLOps. Experience with building RESTful APIs for monitoring build & production systems using automated monitoring of models and corresponding alarm tools. Experience with data versioning tools such as Data Version Control (DVC). Build and maintain data pipelines by using tools like Dagster, Airflow etc. Experience with containerizing and deploying ML models. Hands-on experience with autoML tools, experiment tracking, model management, version tracking & model training (MLflow, W&B, Neptune etc.), model hyperparameter optimization, model evaluation, and visualization (Tensorboard). Sound knowledge and experience with atleast one DL frameworks such as PyTorch, TensorFlow, Keras. Experience with container technologies (Docker, Kubernetes etc). Experience with cloud services. Working knowledge of SQL based databases. Hands on experience with Python scientific computing stack such as numpy, scipy, scikit-learn Familiarity with Linux and git. Detail oriented design, code debugging and problem-solving skills. Effective communication skills: discussing with peers and driving logic driven conclusions. Ability to perspicuously communicate complex technical/architectural problems and propose solutions for the same. How to stand out Prior experience in deploying ML & DL solutions as services Experience with multiple cloud services. Ability to collaborate effectively across functions in a fast-paced environment. Experience with technical documentation and presentation for effective dissemination of work. Engineering experience in distributed systems and data infrastructure. Show more Show less
Posted 5 days ago
2.0 years
0 Lacs
Mumbai, Maharashtra, India
On-site
About Neo Group: Neo is a new-age, focused Wealth and Asset Management platform in India, catering to HNIs, UHNIs and multi-family offices. Neo stands on its three pillars of unbiased advisory, transparency and cost-efficiency, to offer comprehensive, trustworthy solutions. Founded by Nitin Jain (ex-CEO of Edelweiss Wealth), Neo has amassed over USD 3 Billion (₹25,000 Cr.) of Assets Under Advice within a short span of 2 years since inception, including USD 360 Million (₹3,000 Cr.) Assets Under Management. We have recently partnered with Peak XV Partners via a USD 35 Million growth round. To know more, please visit: www.neo-group.in Position: Senior Data Scientist Location: Mumbai Experience: 4 - 8 years Job Description: You are a data pro with deep statistical knowledge and analytical aptitude. You know how to make sense of massive amounts of data and gather deep insights. You will use statistics, data mining, machine learning, and deep learning techniques to deliver data-driven insights for clients. You will dig deep to understand their challenges and create innovative yet practical solutions. Responsibilities: • Meeting with the business team to discuss user interface ideas and applications. • Selecting features, building and optimizing classifiers using machine learning techniques • Data mining using state-of-the-art methods • Doing ad-hoc analysis and presenting results in a clear manner • Optimize application for maximum speed and scalability • Assure that all user input is validated before submitting code • Collaborate with other team members and stakeholders • Taking ownership of features and accountability Requirements: • 4+ years’ experience in developing Data Models • Excellent understanding of machine learning techniques and algorithms, such as k-NN, Naive Bayes, SVM, Decision Forests, etc. • Excellent understanding of NLP and language processing • Proficient understanding of Python or PySpark • Good experience of Python and databases such as MongoDB or MySQL • Good applied statistics skills, such as distributions, statistical testing, regression, etc. • Build Acquisition Scorecard Models • Build Behaviour Scorecard Models • Created Threat Detection Models • Created risk profiling model or classification model • Build Threat/Fraud Triggers from various sources of data • Experience with Data Analysis Libraries - NumPy, Pandas, Statsmodels, Dask • Good understanding of Word2vec, RNNs, Transformers, Bert, Resnet, MobileNet, Unet, Mask-RCNN, Siamese Networks, GradCam, image augmentation techniques, GAN, Tensorboard • Ability to provide accurate estimates for tasks and detailed breakdowns for planning and managing sprints • Deployment - Flask, Tensorflow serving, Lambda functions, Docker is a plus • Previous Experience leading a DS team is a plus Personal Qualities: • An ability to perform well in a fast-paced environment • Excellent analytical and multitasking skills • Stays up-to-date on emerging technologies • Data-oriented personality Why join us? We will provide you with the opportunity to challenge yourself and learn new skills, as you become an integral part our growth story. We are group of ambitious people who believe in building a business environment around new age concepts, framework, and technologies built on a strong foundation of industry expertise. We promise you the prospect of being surrounded by smart, ambitious, motivated people, day-in and day-out. That’s the kind of work you can expect to do at Neo. Show more Show less
Posted 6 days ago
2.0 years
0 Lacs
Mumbai, Maharashtra, India
On-site
About Neo Group: Neo is a new-age, focused Wealth and Asset Management platform in India, catering to HNIs, UHNIs and multi-family offices. Neo stands on its three pillars of unbiased advisory, transparency and cost-efficiency, to offer comprehensive, trustworthy solutions. Founded by Nitin Jain (ex-CEO of Edelweiss Wealth), Neo has amassed over USD 3 Billion (₹25,000 Cr.) of Assets Under Advice within a short span of 2 years since inception, including USD 360 Million (₹3,000 Cr.) Assets Under Management. We have recently partnered with Peak XV Partners via a USD 35 Million growth round. To know more, please visit: www.neo-group.in Position: Data Scientist Location: Mumbai Experience: 2-5 years Job Description: You are a data pro with deep statistical knowledge and analytical aptitude. You know how to make sense of massive amounts of data and gather deep insights. You will use statistics, data mining, machine learning, and deep learning techniques to deliver data-driven insights for clients. You will dig deep to understand their challenges and create innovative yet practical solutions. Responsibilities: • Meeting with the business team to discuss user interface ideas and applications. • Selecting features, building and optimizing classifiers using machine learning techniques • Data mining using state-of-the-art methods • Doing ad-hoc analysis and presenting results in a clear manner • Optimize application for maximum speed and scalability • Assure that all user input is validated before submitting code • Collaborate with other team members and stakeholders • Taking ownership of features and accountability Requirements: • 2+ years’ experience in developing Data Models • Excellent understanding of machine learning techniques and algorithms, such as k-NN, Naive Bayes, SVM, Decision Forests, etc. • Excellent understanding of NLP and language processing • Proficient understanding of Python or PySpark • Basic understanding of Python and databases such as MongoDB or MySQL • Good applied statistics skills, such as distributions, statistical testing, regression, etc. • Build Acquisition Scorecard Models • Build Behaviour Scorecard Models • Created Threat Detection Models • Created risk profiling model or classification model • Build Threat/Fraud Triggers from various sources of data • Experience with Data Analysis Libraries - NumPy, Pandas, Statsmodels, Dask • Good understanding of Word2vec, RNNs, Transformers, Bert, Resnet, MobileNet, Unet, Mask-RCNN, Siamese Networks, GradCam, image augmentation techniques, GAN, Tensorboard • Ability to provide accurate estimates for tasks and detailed breakdowns for planning and managing sprints • Deployment - Flask, Tensorflow serving, Lambda functions, Docker is a plus • Previous Experience leading a DS team is a plus Personal Qualities: • An ability to perform well in a fast-paced environment • Excellent analytical and multitasking skills • Stays up-to-date on emerging technologies • Data-oriented personality Why join us? We will provide you with the opportunity to challenge yourself and learn new skills, as you become an integral part our growth story. We are group of ambitious people who believe in building a business environment around new age concepts, framework, and technologies built on a strong foundation of industry expertise. We promise you the prospect of being surrounded by smart, ambitious, motivated people, day-in and day-out. That’s the kind of work you can expect to do at Neo. Show more Show less
Posted 6 days ago
0 years
0 Lacs
Gurgaon, Haryana, India
On-site
At Nielsen, we are passionate about our work to power a better media future for all people by providing powerful insights that drive client decisions and deliver extraordinary results. Our talented, global workforce is dedicated to capturing audience engagement with content - wherever and whenever it’s consumed. Together, we are proudly rooted in our deep legacy as we stand at the forefront of the media revolution. When you join Nielsen, you will join a dynamic team committed to excellence, perseverance, and the ambition to make an impact together. We champion you, because when you succeed, we do too. We enable your best to power our future. ABOUT THIS JOB Nielsen Global Media uses cutting edge technology and industry leading data science to tackle some of the hardest problems in marketing science. We’re automating our models with artificial intelligence and machine learning to produce the same quality insights as a traditional white-glove consulting engagement at unparalleled speed and scale. Intelligence Studio is a horizontally scalable, cross-cloud technology agnostic platform built with trusted open source components like VS Code, Apache Airflow, Jupyterhub and MLFlow. It allows data scientists to focus on doing data science by taking care of essential concerns like data access, logging, configuration, resource negotiation, dependency management, orchestration, and testing. We’re looking for a Staff Software Engineer to help our talented, cross-functional team improve user workflows in Intelligence Studio. Ideal candidates will be hands-on technologists with experience in Python, Kubernetes, Distributed Systems, AWS or Azure cloud infrastructure. This position is a fantastic opportunity for an experienced engineer to work with creative engineers and cutting-edge technologies. RESPONSIBILITIES Build software and integrations in a cloud-based microservices environment (kubernetes) for big data applications with Spark, Ray, etc. Writing software in python, typescript, go, Java and Scala Work with stakeholders and technical leadership to design and build interfaces, workflows, and services that enhance the delivery of data science products Actively participate in team code reviews and enforce quality standards Work within a cross-functional team to author clear and purposeful epics/stories Promote and enforce best practices in development and operations Identify opportunities and weaknesses in the platform architecture Design and develop data visualization tooling using electron, jupyterhub, plotly, typescript and pandas to enhance data exploration workflows for data science Integrate data science visualization and diagnostic tooling like tensorboard, ray serving, spark history server into an existing distributed compute and development environment Build secure integrations with the kubernetes api allowing the management of user workloads in a shared environment with potentially sensitive data Understand and debug interactions between cloud networking components (ALBs, web api proxies) cluster ingression and security using kong and istio, python-based web servers and modern web transfer protocols like websockets and http3 A LITTLE BIT ABOUT YOU You are an experienced software engineer with a proven track record of quickly learning and implementing new technologies. You love technology and are excited to work on a high-performance team building a very ambitious product. You are looking for an opportunity to grow your career and your technical depth by diving into a project working on the current state of the art in big data and cloud technologies. QUALIFICATIONS Bachelor’s degree in Computer Science or a related technical field, or equivalent industry experience Typescript, Python, Kubernetes, Airflow, Electron, Jupyter, Pandas, Keras, ray, tensorflow, CUDA Apache Spark, Istio, Scala, Java, Go, kong, cloud software design, containerized microservices & distributed caching Experience with machine learning (RNNs, CNNs, random forest, LLMs) a plus Please be aware that job-seekers may be at risk of targeting by scammers seeking personal data or money. Nielsen recruiters will only contact you through official job boards, LinkedIn, or email with a nielsen.com domain. Be cautious of any outreach claiming to be from Nielsen via other messaging platforms or personal email addresses. Always verify that email communications come from an @ nielsen.com address. If you're unsure about the authenticity of a job offer or communication, please contact Nielsen directly through our official website or verified social media channels. Qualified applicants will receive consideration for employment without regard to race, color, religion, sex, sexual orientation, gender identity, national origin, disability, protected veteran status or other characteristics protected by law. Show more Show less
Posted 2 weeks ago
8.0 years
0 Lacs
India
Remote
🔍 We're Hiring! – ML Ops Engineer (Remote, India) 📍 Location: Remote (Within India) 💼 Employment Type: Full-Time / Contractor 📅 Start Date: Immediate 🕒 Working Hours: 1:30 PM IST – 10:30 PM IST (Aligned with US CST) 🚀 Join Madlabs Global LLC as we lead the charge in deploying cutting-edge ML and Generative AI solutions at scale! We’re looking for a highly skilled ML Ops Engineer to lead the development, deployment, and lifecycle management of AI/ML models in cloud-native (preferably GCP) environments . 💼 Key Responsibilities Build scalable ML pipelines: ingestion, preprocessing, training, and serving. Collaborate with Data Scientists to turn prototypes into production-ready systems. Deploy and optimize LLM-based applications (instruction-tuned, fine-tuned models). Own continuous learning pipelines: retraining, model drift detection, performance tuning. Automate workflows using CI/CD , MLFlow , orchestration tools. Leverage GCP services like Vertex AI, BigQuery, Dataflow, Pub/Sub, Cloud Functions. Use Docker & Kubernetes to containerize and orchestrate model deployments. Monitor model performance with Prometheus, TensorBoard, Grafana, etc. Ensure security, fairness, and compliance across ML systems. 🧠 Required Experience 8+ years in ML Engineering, MLOps, or AI Infrastructure roles. Strong coding skills in Python with frameworks like TensorFlow, PyTorch, Scikit-learn. Deep expertise in GCP-native ML stacks . Hands-on experience in Generative AI model deployment and model optimization . Proficiency in Docker, Kubernetes, Jenkins, GitLab CI/CD . Solid understanding of model monitoring, versioning, rollback, and governance. 🕘 Work Hours Fully remote (India-based) Must provide overlap with CST time zone – working hours: 1:30 PM IST to 10:30 PM IST 💬 Interested or want to learn more? 📞 Contact: +91 98868 11767 📧 Email: naveed@madlabsinfotech.com Apply now or DM us to explore this opportunity to work with a team pushing the boundaries of AI innovation! #Hiring #MLOps #MachineLearning #GenerativeAI #LLM #VertexAI #RemoteJobsIndia #DataEngineering #AIJobs #GCP #DevOpsForAI #MLDeployment #LinkedInJobs Show more Show less
Posted 3 weeks ago
8.0 years
0 Lacs
Pune, Maharashtra, India
On-site
Job Requisition ID # 25WD85491 Position Overview We are looking for an experienced Principal Software Engineer to join our platform team focusing on AI/ML Platform (AMP). This team builds and maintains central components to fast track the development of new ML/AI models such as model development studio, feature store, model serving and model observability. The ideal candidate would have a background in ML Ops, Data engineering and DevOps with the experience of building high scale deployment architectures and observability. As an important contributor to our engineering team, you will help shape the future of our AI/ML capabilities, delivering solutions that inspire value for our organization. You will report directly to an Engineering Manager, and you will be based in Pune. Responsibilities System design: You will design, implement and manage software systems for the AI/ML Platform and orchestrate the full ML development lifecycle for the partner teams Mentoring: Spreading your knowledge, sharing best practices and doing design reviews to step up the expertise at the team level Multi-cloud architecture: Define components which leverages strengths from multiple cloud platforms (e.g., AWS, Azure) to optimize performance, cost, and scalability AI/ML observability: You will build systems for monitoring performance of AI/ML models and find insights on the underlying data such as drift detection, data fairness/bias and anomalies ML Solution Deployment: You will develop tools for building and deploying ML artefacts in production environments and facilitating a smooth transition from development to deployment Big Data Management: Automate and orchestrate tasks related to managing big data transformation and processing and build large-scale data stores for ML artifacts Scalable Services: Design and implement low-latency, scalable prediction, and inference services to support the diverse needs of our users Cross-Functional Collaboration: Collaborate across diverse teams, including machine learning researchers, developers, product managers, software architects, and operations, fostering a collaborative and cohesive work environment End-to-end ownership: You will take the end-to-end ownership of the components and work with other engineers in the team including design, architecture, implementation, rollout and onboarding support to partner teams, production on-call support, testing/verification, investigations etc Minimum Qualifications Educational Background: Bachelor’s degree in Computer Science or equivalent practical experience Experience: Over 8 years of experience in software development and engineering, delivering production systems and services Prior experience of working with MLOps team at the intersection of the expertise across ML model deployments, DevOps and data engineering Hands-on skills: Ability to fluently translate the design into high quality code in golang, python, Java Knowledge of DevOps practices, containerization, orchestration tools such as CI/CD, Terraform, Docker, Kubernetes, Gitops Demonstrate knowledge of distributed data processing frameworks, orchestrators, and data lake architectures using technologies such as Spark, Airflow, iceberg/ parquet formats Prior collaborations with Data science teams to deploy their models, setting up ML observability for inference level monitoring Exposure for building RAG based applications by collaborating with other product teams, Data scientists/AI engineers Demonstrate creative problem-solving skills with the ability to break down problems into manageable components Knowledge of Amazon AWS and/or Azure cloud for solutioning large scale application deployments Excellent communication and collaboration skills, fostering teamwork and effective information exchange Preferred Qualifications Experience in integrating with third party vendors Experience in latency optimization with the ability to diagnose, tune, and enhance the efficiency of serving systems Familiarity with tools and frameworks for monitoring and managing the performance of AI/ML models in production (e.g., MLflow, Kubeflow, TensorBoard) Familiarity with distributed model training/inference pipelines using (KubeRay or equivalent) Exposure to leveraging GPU computing for AI/ML workloads, including experience with CUDA, OpenCL, or other GPU programming tools, to significantly enhance model training and inference performance Exposure to ML libraries such as PyTorch, TensorFlow, XGBoost, Pandas, and ScikitLearn Learn More About Autodesk Welcome to Autodesk! Amazing things are created every day with our software – from the greenest buildings and cleanest cars to the smartest factories and biggest hit movies. We help innovators turn their ideas into reality, transforming not only how things are made, but what can be made. We take great pride in our culture here at Autodesk – our Culture Code is at the core of everything we do. Our values and ways of working help our people thrive and realize their potential, which leads to even better outcomes for our customers. When you’re an Autodesker, you can be your whole, authentic self and do meaningful work that helps build a better future for all. Ready to shape the world and your future? Join us! Salary transparency Salary is one part of Autodesk’s competitive compensation package. Offers are based on the candidate’s experience and geographic location. In addition to base salaries, we also have a significant emphasis on discretionary annual cash bonuses, commissions for sales roles, stock or long-term incentive cash grants, and a comprehensive benefits package. Diversity & Belonging We take pride in cultivating a culture of belonging and an equitable workplace where everyone can thrive. Learn more here: https://www.autodesk.com/company/diversity-and-belonging Are you an existing contractor or consultant with Autodesk? Please search for open jobs and apply internally (not on this external site). Show more Show less
Posted 3 weeks ago
5 years
0 Lacs
Bengaluru, Karnataka, India
On-site
Company Qualcomm India Private Limited Job Area Engineering Group, Engineering Group > Systems Engineering General Summary As a leading technology innovator, Qualcomm pushes the boundaries of what's possible to enable next-generation experiences and drives digital transformation to help create a smarter, connected future for all. As a Qualcomm Systems Engineer, you will research, design, develop, simulate, and/or validate systems-level software, hardware, architecture, algorithms, and solutions that enables the development of cutting-edge technology. Qualcomm Systems Engineers collaborate across functional teams to meet and exceed system-level requirements and standards. Minimum Qualifications Bachelor's degree in Engineering, Information Systems, Computer Science, or related field and 8+ years of Systems Engineering or related work experience. OR Master's degree in Engineering, Information Systems, Computer Science, or related field and 7+ years of Systems Engineering or related work experience. OR PhD in Engineering, Information Systems, Computer Science, or related field and 6+ years of Systems Engineering or related work experience. Principal Engineer – Machine Learning We are looking for a Principal AI/ML Engineer with expertise in model inference , optimization , debugging , and hardware acceleration . This role will focus on building efficient AI inference systems, debugging deep learning models, optimizing AI workloads for low latency, and accelerating deployment across diverse hardware platforms. In addition to hands-on engineering, this role involves cutting-edge research in efficient deep learning, model compression, quantization, and AI hardware-aware optimization techniques . You will explore and implement state-of-the-art AI acceleration methods while collaborating with researchers, industry experts, and open-source communities to push the boundaries of AI performance. This is an exciting opportunity for someone passionate about both applied AI development and AI research , with a strong focus on real-world deployment, model interpretability, and high-performance inference . Education & Experience 20+ years of experience in AI/ML development, with at least 5 years in model inference, optimization, debugging, and Python-based AI deployment. Master’s or Ph.D. in Computer Science, Machine Learning, AI Leadership & Collaboration Lead a team of AI engineers in Python-based AI inference development. Collaborate with ML researchers, software engineers, and DevOps teams to deploy optimized AI solutions. Define and enforce best practices for debugging and optimizing AI models Key Responsibilities Model Optimization & Quantization Optimize deep learning models using quantization (INT8, INT4, mixed precision etc), pruning, and knowledge distillation. Implement Post-Training Quantization (PTQ) and Quantization-Aware Training (QAT) for deployment. Familiarity with TensorRT, ONNX Runtime, OpenVINO, TVM AI Hardware Acceleration & Deployment Optimize AI workloads for Qualcomm Hexagon DSP, GPUs (CUDA, Tensor Cores), TPUs, NPUs, FPGAs, Habana Gaudi, Apple Neural Engine. Leverage Python APIs for hardware-specific acceleration, including cuDNN, XLA, MLIR. Benchmark models on AI hardware architectures and debug performance issues AI Research & Innovation Conduct state-of-the-art research on AI inference efficiency, model compression, low-bit precision, sparse computing, and algorithmic acceleration. Explore new deep learning architectures (Sparse Transformers, Mixture of Experts, Flash Attention) for better inference performance. Contribute to open-source AI projects and publish findings in top-tier ML conferences (NeurIPS, ICML, CVPR). Collaborate with hardware vendors and AI research teams to optimize deep learning models for next-gen AI accelerators. Details Of Expertise Experience optimizing LLMs, LVMs, LMMs for inference Experience with deep learning frameworks: TensorFlow, PyTorch, JAX, ONNX. Advanced skills in model quantization, pruning, and compression. Proficiency in CUDA programming and Python GPU acceleration using cuPy, Numba, and TensorRT. Hands-on experience with ML inference runtimes (TensorRT, TVM, ONNX Runtime, OpenVINO) Experience working with RunTimes Delegates (TFLite, ONNX, Qualcomm) Strong expertise in Python programming, writing optimized and scalable AI code. Experience with debugging AI models, including examining computation graphs using Netron Viewer, TensorBoard, and ONNX Runtime Debugger. Strong debugging skills using profiling tools (PyTorch Profiler, TensorFlow Profiler, cProfile, Nsight Systems, perf, Py-Spy). Expertise in cloud-based AI inference (AWS Inferentia, Azure ML, GCP AI Platform, Habana Gaudi). Knowledge of hardware-aware optimizations (oneDNN, XLA, cuDNN, ROCm, MLIR, SparseML). Contributions to open-source community Publications in International forums / conferences / journals Applicants : Qualcomm is an equal opportunity employer. If you are an individual with a disability and need an accommodation during the application/hiring process, rest assured that Qualcomm is committed to providing an accessible process. You may e-mail disability-accomodations@qualcomm.com or call Qualcomm's toll-free number found here. Upon request, Qualcomm will provide reasonable accommodations to support individuals with disabilities to be able participate in the hiring process. Qualcomm is also committed to making our workplace accessible for individuals with disabilities. (Keep in mind that this email address is used to provide reasonable accommodations for individuals with disabilities. We will not respond here to requests for updates on applications or resume inquiries). Qualcomm expects its employees to abide by all applicable policies and procedures, including but not limited to security and other requirements regarding protection of Company confidential information and other confidential and/or proprietary information, to the extent those requirements are permissible under applicable law. To all Staffing and Recruiting Agencies : Our Careers Site is only for individuals seeking a job at Qualcomm. Staffing and recruiting agencies and individuals being represented by an agency are not authorized to use this site or to submit profiles, applications or resumes, and any such submissions will be considered unsolicited. Qualcomm does not accept unsolicited resumes or applications from agencies. Please do not forward resumes to our jobs alias, Qualcomm employees or any other company location. Qualcomm is not responsible for any fees related to unsolicited resumes/applications. If you would like more information about this role, please contact Qualcomm Careers. 3072372 Show more Show less
Posted 3 weeks ago
Upload Resume
Drag or click to upload
Your data is secure with us, protected by advanced encryption.
Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.
Accenture
36723 Jobs | Dublin
Wipro
11788 Jobs | Bengaluru
EY
8277 Jobs | London
IBM
6362 Jobs | Armonk
Amazon
6322 Jobs | Seattle,WA
Oracle
5543 Jobs | Redwood City
Capgemini
5131 Jobs | Paris,France
Uplers
4724 Jobs | Ahmedabad
Infosys
4329 Jobs | Bangalore,Karnataka
Accenture in India
4290 Jobs | Dublin 2