Jobs
Interviews

12 Tvm Jobs

Setup a job Alert
JobPe aggregates results for easy application access, but you actually apply on the job portal directly.

1.0 - 3.0 years

0 Lacs

bengaluru, karnataka, india

On-site

JOB TITLE: AI ASSOCIATE OPTIMIZATION Location: Bengaluru | Experience: 1-3 Years ABOUT NEWCOLD NewCold is a service provider in cold chain logistics with a focus on development and operation of large, highly automated cold stores. NewCold strives to be crucial in the cold chain of leading food companies, by offering advanced logistic services worldwide. NewCold is one of the fastest growing companies (over 2,000 employees) in the cold chain logistics and they are expanding teams to support this growth. They use the latest technology that empowers people, to handle food responsibly and guarantee food safety in a sustainable way. They challenge the industry, believe in long-term partnerships, and deliver solid investment opportunities that enable next generation logistic solutions. NewCold has leading market in-house expertise in designing, engineering, developing and operating state-of-the-art automated cold stores: a result of successful development and operation of over 15 automated warehouses across three continents. With the prospect of many new construction projects around the world in the very near future, this vacancy offers an interesting opportunity to join an internationally growing and ambitious organization. POSITION SUMMARY NewCold is seeking an AI Associate Optimization to enhance the performance and efficiency of our AI-powered solutions within our highly automated cold chain logistics network. This role focuses on optimizing machine learning models deployed in warehouse operations, ensuring low latency, high throughput, and accurate predictions for improved decision-making. You will be instrumental in bridging the gap between data science and real-world deployment, contributing to the continuous improvement of our automated systems. This position requires a strong understanding of model serving, containerization, and edge AI technologies. YOUR ROLE As an AI Associate Optimization, you will play a critical role in ensuring the reliability, scalability, and performance of AI models powering NewColds automated warehouse processes. You will be responsible for optimizing models for deployment across diverse infrastructure, including cloud and edge environments, directly impacting operational efficiency, cost reduction, and the overall effectiveness of our logistics solutions. Your work will contribute to maintaining NewColds competitive edge through cutting-edge AI implementation. KEY RESPONSIBILITIES Implement model optimization techniques such as quantization and knowledge distillation to reduce model size and improve inference speed for deployment on edge devices and cloud infrastructure. Develop and maintain CI/CD pipelines for automated model deployment and updates, ensuring seamless integration with existing systems. Benchmark and profile model performance (latency, throughput, memory usage) to identify bottlenecks and areas for improvement. Deploy and manage machine learning models using model serving frameworks like TensorFlow Serving, TorchServe, ONNX Runtime, or Triton Inference Server. Containerize AI models and applications using Docker and Podman for consistent and reproducible deployments. Collaborate with data scientists and software engineers to troubleshoot model performance issues and implement solutions. Monitor model performance in production and proactively address any degradation in accuracy or efficiency. Develop and maintain APIs/SDKs (REST, gRPC, FastAPI) for accessing and integrating AI models into various applications. Work with edge devices (NVIDIA Jetson, Coral TPU, ARM-based boards) and edge frameworks (TensorRT, OpenVINO, TFLite, TVM) to optimize models for low-power, real-time inference. WHAT WE ARE LOOKING FOR Bachelors or masters degree in computer science, Artificial Intelligence, Machine Learning or a related field 1-3 years of experience in a role focused on machine learning model optimization and deployment. Proficiency in Python and C++ programming languages. Hands-on experience with model serving frameworks (TensorFlow Serving, TorchServe, ONNX Runtime, Triton Inference Server). Experience with containerization technologies (Docker, Podman) and orchestration tools (Kubernetes, K3s, Edge orchestrators). Knowledge of model optimization techniques such as quantization and knowledge distillation. Familiarity with benchmarking and profiling tools for evaluating model performance. Strong analytical and problem-solving skills with a data-driven approach. Experience with CI/CD pipelines for ML deployment is highly desirable. Knowledge of edge devices (NVIDIAJetson, Coral TPU, ARM-based boards) and edge AI frameworks (TensorRT, OpenVINO, TFLite, TVM) is a significant plus. WHY JOIN US Opportunity to work on cutting-edge AI applications in a rapidly growing and innovative cold chain logistics company. Exposure to a wide range of AI technologies and challenges within a highly automated warehouse environment. Career growth potential within a dynamic and international organization. Collaborative and supportive team environment with opportunities for learning and development. Contribute to the development of next-generation logistics solutions that are shaping the future of the food supply chain. Show more Show less

Posted 1 week ago

Apply

1.0 - 10.0 years

0 Lacs

hyderabad, telangana

On-site

Qualcomm India Private Limited is seeking passionate individuals interested in accelerating machine learning and artificial intelligence on mobile devices for millions of users to join their Engineering Group. As a software engineer with a machine learning or compiler background, you will play a crucial role in building software platforms that empower users of Qualcomm's silicon to create optimized neural networks and machine learning algorithms. In this role, you will have the opportunity to construct and fine-tune machine learning frameworks, develop compilers and tools, and collaborate closely with Qualcomm's hardware and software engineers to ensure the efficient utilization of Qualcomm's silicon for machine learning applications. Minimum qualifications for this position include a Bachelor's degree in Engineering, Information Systems, Computer Science, or a related field, proficiency in programming languages such as C/C++, and 0 to 10 years of software engineering or related work experience. Preferred qualifications consist of experience in machine learning frameworks like MxNet, NNVM, TVM, Pytorch, Tensorflow, or Caffe, or expertise in compilers with a keen interest in machine learning. Deep knowledge of software engineering is also highly valued. Candidates must meet at least one of the following minimum qualifications: - Bachelor's degree in Engineering, Information Systems, Computer Science, or related field with 2+ years of Software Engineering experience - Master's degree in Engineering, Information Systems, Computer Science, or related field with 1+ year of Software Engineering experience - PhD in Engineering, Information Systems, Computer Science, or related field Additionally, applicants should have a minimum of 2 years of academic or work experience with programming languages like C, C++, Java, Python, etc. Qualcomm is an equal opportunity employer committed to providing accessible processes for individuals with disabilities. If you require accommodations during the application/hiring process, you may contact Qualcomm at disability-accommodations@qualcomm.com or through their toll-free number. It is expected that all Qualcomm employees adhere to company policies and procedures, including those concerning the protection of confidential information and proprietary data in compliance with applicable laws. For staffing and recruiting agencies, note that Qualcomm's Careers Site is exclusively for individuals seeking job opportunities at Qualcomm. Unsolicited submissions from agencies or individuals represented by an agency will not be considered. Qualcomm does not accept unsolicited resumes or applications from agencies. Please refrain from forwarding resumes to Qualcomm employees or any company location. For further information about this role, please reach out to Qualcomm Careers.,

Posted 2 weeks ago

Apply

20.0 - 22.0 years

0 Lacs

karnataka

On-site

Qualcomm India Private Limited is a leading technology innovator in the Engineering Group, specifically in Systems Engineering. As a Qualcomm Systems Engineer, you will be involved in researching, designing, developing, simulating, and validating systems-level software, hardware, architecture, algorithms, and solutions to drive the development of cutting-edge technology. Collaboration across functional teams is essential to meet and exceed system-level requirements and standards. To qualify for this role, you should possess a Bachelor's degree in Engineering, Information Systems, Computer Science, or related field with at least 8 years of experience in Systems Engineering. Alternatively, a Master's degree with 7+ years of experience or a Ph.D. with 6+ years of experience in the same field is also acceptable. Currently, Qualcomm is seeking a Principal AI/ML Engineer with expertise in model inference, optimization, debugging, and hardware acceleration. The role focuses on building efficient AI inference systems, debugging deep learning models, optimizing AI workloads for low latency, and accelerating deployment across various hardware platforms. In addition to hands-on engineering tasks, the role also involves cutting-edge research in efficient deep learning, model compression, quantization, and AI hardware-aware optimization techniques. The ideal candidate will collaborate with researchers, industry experts, and open-source communities to enhance AI performance continuously. The suitable candidate should have a minimum of 20 years of experience in AI/ML development, with a focus on model inference, optimization, debugging, and Python-based AI deployment. A Master's or Ph.D. in Computer Science, Machine Learning, or AI is preferred. Key Responsibilities of this role include Model Optimization & Quantization, AI Hardware Acceleration & Deployment, and AI Research & Innovation. The candidate should have expertise in optimizing deep learning models, familiarity with deep learning frameworks, proficiency in CUDA programming, and experience with various ML inference runtimes. Qualcomm encourages applicants from diverse backgrounds and is an equal opportunity employer. The company is committed to providing reasonable accommodations to individuals with disabilities during the hiring process. It is vital for all employees to adhere to applicable policies and procedures, including those related to confidentiality and security. Qualcomm does not accept unsolicited resumes or applications from staffing and recruiting agencies. For further information about this role, interested individuals may reach out to Qualcomm Careers.,

Posted 2 weeks ago

Apply

0.0 years

0 Lacs

bengaluru, karnataka, india

On-site

Job Description We are seeking a highly skilled AI/ML Validation Engineer with a strong foundation in machine learning, deep learning, and system-level validation. The ideal candidate will have hands-on experience with ML frameworks, profiling tools, and AI compute stacks, and will play a key role in validating end-to-end AI pipelines and ensuring software quality across diverse platforms. Responsibilities Design and execute validation plans for AI/ML compute stacks (HIP, CUDA, OpenCL, OpenVINO, ONNX Runtime, TensorFlow, PyTorch). Validate end-to-end AI pipelines including model conversion, inference runtimes, compilers/toolchains, kernel execution, and memory transfer. Profile ML workloads and optimize performance across platforms. Collaborate with global teams to ensure high-quality deliverables. Apply software development lifecycle practices to validation workflows. Document defects, validation results, and improvement recommendations. Primary Skills Strong knowledge of ML fundamentals, deep learning, LLMs, and recommender systems. Proficiency in Python programming. Experience with PyTorch, TensorFlow, ONNX Runtime. Familiarity with Ubuntu/Yocto Linux environments. Expertise in profiling tools and performance analysis. Strong understanding of software QA methodologies. Secondary Skills Experience with compilers/toolchains like TVM, Vitis AI, XDNA, XLA. Exposure to production-grade software validation workflows. Excellent problem-solving and communication skills. Educational Qualification Bachelor's or Master's degree in Computer Science, Electrical Engineering, or related field.

Posted 3 weeks ago

Apply

8.0 - 12.0 years

25 - 35 Lacs

Pune, Ahmedabad, Bengaluru

Work from Office

Role & responsibilities : Job Title / Designation: Solution Architect/Project Manager/Associate Director based on experience & expertise Business Unit : Embedded Engineering Services (EES) Industry Experience Range : 8+ years Job Location : Preferably Pune / Ahmedabad / Bangalore Shift : General Shift (Mon-Fri) Job Function, Roles & Responsibilities: Lead strategic initiatives and own the practice for Edge AI/ML, data pipelines, and intelligent embedded systems Define and build the competency roadmap for machine learning, deep learning, model deployment, and real-time inferencing on edge platforms Oversee data creation including data collection, dataset curation, annotation, cleaning, augmentation, and synthetic data generation Champion use cases involving sensor fusion, combining data from multiple sources (vision, IMU, radar, audio, etc.) to create robust, efficient, and context-aware edge intelligence solutions Drive edge analytics and on-device learning across verticals such as Industrial Automation, Medical Devices, Automotive, and Smart Consumer Electronics Collaborate with global customers to gather requirements, architect solutions, track project delivery, and ensure alignment with business objectives Support business development with presales solutioning, proposal writing, and effort estimation Drive internal capability building through mentoring, training, and competency development Preferred candidate profile: ________________________________________ Experience: 8+ years in embedded systems, AI/ML, and data engineering, with a strong focus on edge intelligence and real-time systems. At least 3 years in a technical leadership or strategic role. Prior experience in a product engineering services environment preferred. ________________________________________ Area of Expertise: Proven expertise in deploying ML/DL models on edge devices (NVIDIA Jetson, NXP i.MX, Qualcomm QCS, TI Sitara, etc.) Strong knowledge of data workflows: dataset generation, manual/automated annotation, data cleaning, augmentation, and synthetic data creation Deep understanding of sensor fusion techniques combining inputs from vision, audio, IMU, radar, LIDAR, and other sources to improve model accuracy and efficiency Experience in model optimization using TensorRT, ONNX, OpenVINO, TFLite, and TVM Hands-on with TensorFlow, PyTorch, scikit-learn, and signal/image processing techniques Proficient in designing for real-time inference on resource-constrained platforms Exposure to AI accelerators, NPUs, DSPs, and hybrid SoC environments; must have exposure to NVIDIA SoC & Tools Presales, account engagement, and solutioning experience with North American or European clients ________________________________________ Nice to Have: Cloud-edge integration using AWS Greengrass, Azure IoT Edge, GCP Edge TPU Understanding of AI regulatory/safety standards (ISO, IEC, FDA compliance for AI/ML in regulated industries) ________________________________________ Educational Criteria: BE/ME/B.Tech/M.Tech Electronics, Computer Science, AI/ML, Embedded Systems, or Data Science ________________________________________ Travel: Flexibility to travel globally with sales or delivery teams for customer meetings, workshops, and project deployments as needed. Interested and qualified candidate can directly reach Mr. Anup Sharma at 99099-75421 or anup.s@acldigital.com. (staffing partner can communicate over the email)

Posted 2 months ago

Apply

0.0 - 4.0 years

2 - 4 Lacs

Hyderabad, Telangana, India

On-site

Interested in accelerating machine learning and artificial intelligence on mobile devices for millions of users Come join our team. We are building software platforms that enable users of Qualcomms silicon to construct optimized neural networks and machine learning algorithms. We are looking for software engineers with a machine learning or compiler background who will help us build these software platforms. Required Skills & Experience Programming in C/C++ . 0 to 10 years of software engineering or related work experience . Preferred Qualifications Experience in machine learning frameworks such as MxNet/NNVM/TVM, PyTorch, TensorFlow, Caffe. OR experience in compilers with an interest in machine learning. Deep knowledge of software engineering . Minimum Qualifications Bachelor's degree in Engineering, Information Systems, Computer Science, or related field. Bachelor's degree in Engineering, Information Systems, Computer Science, or related field and 2+ years of Software Engineering or related work experience.OR Master's degree in Engineering, Information Systems, Computer Science, or related field and 1+ year of Software Engineering or related work experience. OR PhD in Engineering, Information Systems, Computer Science, or related field. 2+ years of academic or work experience with Programming Language such as C, C++, Java, Python, etc.

Posted 2 months ago

Apply

2.0 - 4.0 years

2 - 4 Lacs

Hyderabad, Telangana, India

On-site

Interested in accelerating machine learning and artificial intelligence on mobile devices for millions of users Come join our team. We are building software platforms that enable users of Qualcomms silicon to construct optimized neural networks and machine learning algorithms. We are looking for software engineers with a machine learning or compiler background who will help us build these software platforms. Key Responsibilities: In this role, you will construct and tune machine learning frameworks, build compilers and tools, and collaborate with Qualcomm hardware and software engineers to enable efficient usage of Qualcomms silicon for machine learning applications. Minimum Qualifications: Bachelors degree in Engineering, Information Systems, Computer Science, or related field. Programming in C/C++ 2 to 4 years of software engineering or related work experience Bachelor's degree in Engineering, Information Systems, Computer Science, or related field and 2+ years of Software Engineering or related work experience. OR Master's degree in Engineering, Information Systems, Computer Science, or related field and 1+ year of Software Engineering or related work experience. OR PhD in Engineering, Information Systems, Computer Science, or related field. 2+ years of academic or work experience with Programming Language such as C, C++, Java, Python, etc. Preferred Qualifications: Experience in machine learning frameworks such as MxNet/NNVM/TVM, Pytorch, Tensorflow, Caffe OR experience in compilers with an interest in machine learning Deep knowledge of software engineering

Posted 2 months ago

Apply

12.0 - 20.0 years

12 - 20 Lacs

Bengaluru, Karnataka, India

On-site

We are looking for a Principal AI/ML Engineer with expertise in model inference, optimization, debugging, and hardware acceleration . This role will focus on building efficient AI inference systems, debugging deep learning models, optimizing AI workloads for low latency, and accelerating deployment across diverse hardware platforms. In addition to hands-on engineering, this role involves cutting-edge research in efficient deep learning, model compression, quantization, and AI hardware-aware optimization techniques . You will explore and implement state-of-the-art AI acceleration methods while collaborating with researchers, industry experts, and open-source communities to push the boundaries of AI performance. This is an exciting opportunity for someone passionate about both applied AI development and AI research , with a strong focus on real-world deployment, model interpretability, and high-performance inference. Education & Experience: 20+ years of experience in AI/ML development, with at least 5 years in model inference, optimization, debugging, and Python-based AI deployment. Masters or Ph.D. in Computer Science, Machine Learning, AI. Leadership & Collaboration: Lead a team of AI engineers in Python-based AI inference development. Collaborate with ML researchers, software engineers, and DevOps teams to deploy optimized AI solutions. Define and enforce best practices for debugging and optimizing AI models. Key Responsibilities: Model Optimization & Quantization: Optimize deep learning models using quantization (INT8, INT4, mixed precision etc), pruning, and knowledge distillation. Implement Post-Training Quantization (PTQ) and Quantization-Aware Training (QAT) for deployment. Familiarity with TensorRT, ONNX Runtime, OpenVINO, TVM. AI Hardware Acceleration & Deployment: Optimize AI workloads for Qualcomm Hexagon DSP, GPUs (CUDA, Tensor Cores), TPUs, NPUs, FPGAs, Habana Gaudi, Apple Neural Engine. Leverage Python APIs for hardware-specific acceleration, including cuDNN, XLA, MLIR. Benchmark models on AI hardware architectures and debug performance issues. AI Research & Innovation: Conduct state-of-the-art research on AI inference efficiency, model compression, low-bit precision, sparse computing, and algorithmic acceleration. Explore new deep learning architectures (Sparse Transformers, Mixture of Experts, Flash Attention) for better inference performance. Contribute to open-source AI projects and publish findings in top-tier ML conferences (NeurIPS, ICML, CVPR). Collaborate with hardware vendors and AI research teams to optimize deep learning models for next-gen AI accelerators. Details of Expertise: Experience optimizing LLMs, LVMs, LMMs for inference. Experience with deep learning frameworks: TensorFlow, PyTorch, JAX, ONNX. Advanced skills in model quantization, pruning, and compression. Proficiency in CUDA programming and Python GPU acceleration using cuPy, Numba, and TensorRT. Hands-on experience with ML inference runtimes (TensorRT, TVM, ONNX Runtime, OpenVINO). Experience working with RunTimes Delegates (TFLite, ONNX, Qualcomm). Strong expertise in Python programming, writing optimized and scalable AI code. Experience with debugging AI models, including examining computation graphs using Netron Viewer, TensorBoard, and ONNX Runtime Debugger. Strong debugging skills using profiling tools (PyTorch Profiler, TensorFlow Profiler, cProfile, Nsight Systems, perf, Py-Spy). Expertise in cloud-based AI inference (AWS Inferentia, Azure ML, GCP AI Platform, Habana Gaudi). Knowledge of hardware-aware optimizations (oneDNN, XLA, cuDNN, ROCm, MLIR, SparseML). Contributions to open-source community. Publications in International forums conferences journals.

Posted 2 months ago

Apply

8.0 - 13.0 years

10 - 14 Lacs

Bengaluru

Work from Office

General Summary: As a leading technology innovator, Qualcomm pushes the boundaries of what's possible to enable next-generation experiences and drives digital transformation to help create a smarter, connected future for all. As a Qualcomm Systems Engineer, you will research, design, develop, simulate, and/or validate systems-level software, hardware, architecture, algorithms, and solutions that enables the development of cutting-edge technology. Qualcomm Systems Engineers collaborate across functional teams to meet and exceed system-level requirements and standards. Minimum Qualifications: Bachelor's degree in Engineering, Information Systems, Computer Science, or related field and 8+ years of Systems Engineering or related work experience. OR Master's degree in Engineering, Information Systems, Computer Science, or related field and 7+ years of Systems Engineering or related work experience. OR PhD in Engineering, Information Systems, Computer Science, or related field and 6+ years of Systems Engineering or related work experience. Principal Engineer Machine Learning We are looking for a Principal AI/ML Engineer with expertise in model inference , optimization , debugging , and hardware acceleration . This role will focus on building efficient AI inference systems, debugging deep learning models, optimizing AI workloads for low latency, and accelerating deployment across diverse hardware platforms. In addition to hands-on engineering, this role involves cutting-edge research in efficient deep learning, model compression, quantization, and AI hardware-aware optimization techniques . You will explore and implement state-of-the-art AI acceleration methods while collaborating with researchers, industry experts, and open-source communities to push the boundaries of AI performance. This is an exciting opportunity for someone passionate about both applied AI development and AI research , with a strong focus on real-world deployment, model interpretability, and high-performance inference . Education & Experience: 20+ years of experience in AI/ML development, with at least 5 years in model inference, optimization, debugging, and Python-based AI deployment. Masters or Ph.D. in Computer Science, Machine Learning, AI Leadership & Collaboration Lead a team of AI engineers in Python-based AI inference development . Collaborate with ML researchers, software engineers, and DevOps teams to deploy optimized AI solutions. Define and enforce best practices for debugging and optimizing AI models Key Responsibilities Model Optimization & Quantization Optimize deep learning models using quantization (INT8, INT4, mixed precision etc), pruning, and knowledge distillation . Implement Post-Training Quantization (PTQ) and Quantization-Aware Training (QAT) for deployment. Familiarity with TensorRT, ONNX Runtime, OpenVINO, TVM AI Hardware Acceleration & Deployment Optimize AI workloads for Qualcomm Hexagon DSP, GPUs (CUDA, Tensor Cores), TPUs, NPUs, FPGAs, Habana Gaudi, Apple Neural Engine . Leverage Python APIs for hardware-specific acceleration , including cuDNN, XLA, MLIR . Benchmark models on AI hardware architectures and debug performance issues AI Research & Innovation Conduct state-of-the-art research on AI inference efficiency, model compression, low-bit precision, sparse computing, and algorithmic acceleration . Explore new deep learning architectures (Sparse Transformers, Mixture of Experts, Flash Attention) for better inference performance . Contribute to open-source AI projects and publish findings in top-tier ML conferences (NeurIPS, ICML, CVPR). Collaborate with hardware vendors and AI research teams to optimize deep learning models for next-gen AI accelerators. Details of Expertise: Experience optimizing LLMs, LVMs, LMMs for inference Experience with deep learning frameworks : TensorFlow, PyTorch, JAX, ONNX. Advanced skills in model quantization, pruning, and compression . Proficiency in CUDA programming and Python GPU acceleration using cuPy, Numba, and TensorRT . Hands-on experience with ML inference runtimes (TensorRT, TVM, ONNX Runtime, OpenVINO) Experience working with RunTimes Delegates (TFLite, ONNX, Qualcomm) Strong expertise in Python programming , writing optimized and scalable AI code. Experience with debugging AI models , including examining computation graphs using Netron Viewer, TensorBoard, and ONNX Runtime Debugger . Strong debugging skills using profiling tools (PyTorch Profiler, TensorFlow Profiler, cProfile, Nsight Systems, perf, Py-Spy) . Expertise in cloud-based AI inference (AWS Inferentia, Azure ML, GCP AI Platform, Habana Gaudi). Knowledge of hardware-aware optimizations (oneDNN, XLA, cuDNN, ROCm, MLIR, SparseML). Contributions to open-source community Publications in International forums conferences journals

Posted 3 months ago

Apply

10.0 - 20.0 years

100 - 200 Lacs

Hyderabad

Hybrid

Compiler Lead Hyderabad Founded by highly respected Silicon Valley veterans - with its design centers established in Santa Clara, California. / Hyderabad/Bangalore A US based well-funded product-based startup looking for Highly talented Verification Engineers for the following roles. We are looking for a highly experienced systems engineer with deep expertise in compilers, machine learning infrastructure, and system-level performance optimization. This role is hands-on and research-driven, ideal for someone who thrives on solving low-level performance challenges and building core infrastructure that powers next-generation AI workloads. Key Responsibilities: Compiler Design & Optimization Develop and enhance compiler toolchains based on LLVM, MLIR, Open64, or Glow. Build and optimize intermediate representations, custom dialects, and code generation flows for AI accelerators. Implement transformations and optimizations for latency, memory usage, and compute efficiency. AI System Integration Work closely with hardware teams to co-design compilers targeting custom silicon. Integrate compiler backends with ML frameworks like PyTorch, TensorFlow, or ONNX. Build graph-level and kernel-level transformations for AI training and inference pipelines. Performance Tuning & System Analysis Conduct low-level profiling and performance tuning across compiler and runtime layers. Identify and eliminate bottlenecks across CPU/GPU/NPU workloads. Develop parallel programming solutions leveraging SIMD, multi-threading, and heterogeneous computing. Tooling & Infrastructure Develop tooling for performance analysis, debug, and test automation. Contribute to internal SDKs and devkits used by AI researchers and system engineers. Required Skills & Experience: Strong compiler development experience using LLVM, MLIR, Glow, or similar toolchains. Proficiency in C/C++, with solid command of Python for tooling and automation. In-depth understanding of compiler internals, including IR design, lowering, codegen, and scheduling. Deep knowledge of hardware-software co-design, particularly for AI/ML workloads. Experience with runtime systems, memory models, and performance modeling. Solid grasp of parallel and heterogeneous computing paradigms. Nice to Have: Experience working with custom AI hardware or edge inference platforms. Familiarity with quantization, scheduling for dataflow architectures, or compiler autotuning. Contributions to open-source compiler projects (e.g., LLVM, MLIR, TVM). Qualifications: Bachelors or Masters degree in Computer Science, Electrical Engineering, or a related field. 10–15 years of relevant hands-on experience in compilers, systems programming, or AI infrastructure. Contact: Uday Mulya Technologies muday_bhaskar@yahoo.com "Mining The Knowledge Community"

Posted 3 months ago

Apply

17 - 27 years

100 - 200 Lacs

Bengaluru

Work from Office

Senior Software Technical Director / Software Technical Director Bangalore Founded in 2023,by Industry veterans HQ in California,US We are revolutionizing sustainable AI compute through intuitive software with composable silicon We are looking for a Software Technical Director with a strong technical foundation in systems software, Linux platforms, or machine learning compiler stacks to lead and grow a high-impact engineering team in Bangalore. You will be responsible for shaping the architecture, contributing to codebases, and managing execution across projects that sit at the intersection of systems programming, AI runtimes, and performance-critical software. Key Responsibilities: Technical Leadership: Lead the design and development of Linux platform software, firmware, or ML compilers and runtimes. Drive architecture decisions across compiler, runtime, or low-level platform components. Write production-grade C++ code and perform detailed code reviews. Guide performance analysis and debugging across the full stackfrom firmware and drivers to user-level runtime libraries. Collaborate with architects, silicon teams, and ML researchers to build future-proof software stacks. Team & Project Management: Mentor and coach junior and senior engineers to grow technical depth and autonomy. Own end-to-end project planning, execution, and delivery, ensuring high-quality output across sprints/releases. Facilitate strong cross-functional communication with hardware, product, and other software teams globally. Recruit and grow a top-tier engineering team in Bangalore, contributing to the hiring strategy and team culture. Required Qualifications: Bachelors or Master’s degree in Computer Science, Electrical Engineering, or related field. 18+ years of experience in systems software development with significant time spent in C++, including architectural and hands-on roles. Proven experience in either: Linux kernel, bootloaders, firmware, or low-level platform software, or Machine Learning compilers (e.g., MLIR, TVM, Glow) or runtimes (e.g., ONNX Runtime, TensorRT, vLLM). Excellent communication skills—written and verbal. Prior experience in project leadership or engineering management with direct reports. Highly Desirable: Understanding of AI/ML compute workloads, particularly Large Language Models (LLMs). Familiarity with performance profiling, bottleneck analysis, and compiler-level optimizations. Exposure to AI accelerators, systolic arrays, or vector SIMD programming. Why Join Us? Work at the forefront of AI systems software, shaping the future of ML compilers and runtimes. Collaborate with globally distributed teams in a fast-paced, innovation-driven environment. Build and lead a technically elite team from the ground up in a growth-stage organization. Contact: Uday Mulya Technologies muday_bhaskar@yahoo.com "Mining The Knowledge Community"

Posted 4 months ago

Apply

2.0 - 7.0 years

25 - 40 Lacs

hyderabad

Work from Office

Strong fundamentals in C/C++ and OS internals Strong knowledge of and experience with LLVM, tvm especially middle-end compilers. Knowledge and experience with graphics/compute APIs such as OpenCL, cuda, Vulkan, OpenGL or DirectX An appreciation of multi-threaded and/or parallel computation and associated complexity Knowledge of GPGPU programming Knowledge of AI/ML workloads that are amenable to GPU execution This is what you are responsible for Design and develop the application and system software for a cutting-edge AI silicon Ability to work in a fast-paced environment collaborating with various teams locally and globally A passion for problem-solving and the ability to consider the bigger picture during feature development Define feature requirements with customers and teams, gathering feedback to improve products Ensure clear understanding of requirements, documenting implementation strategies Review and test features to prevent issues, conducting impact analysis to mitigate potential problems

Posted Date not available

Apply
cta

Start Your Job Search Today

Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.

Job Application AI Bot

Job Application AI Bot

Apply to 20+ Portals in one click

Download Now

Download the Mobile App

Instantly access job listings, apply easily, and track applications.

Featured Companies