Get alerts for new jobs matching your selected skills, preferred locations, and experience range.
0 years
0 Lacs
India
On-site
About the Role: We are seeking an experienced MLOps Engineer with a strong background in NVIDIA GPU-based containerization and scalable ML infrastructure ( Contractual - Assignment Basis) . You will work closely with data scientists, ML engineers, and DevOps teams to build, deploy, and maintain robust, high-performance machine learning pipelines using NVIDIA NGC containers, Docker, Kubernetes , and modern MLOps practices. Key Responsibilities: Design, develop, and maintain end-to-end MLOps pipelines for training, validation, deployment, and monitoring of ML models. Implement GPU-accelerated workflows using NVIDIA NGC containers, CUDA, and RAPIDS . Containerize ML workloads using Docker and deploy on Kubernetes (preferably with GPU support like NVIDIA device plugin for K8s) . Integrate model versioning, reproducibility, CI/CD, and automated model retraining using tools like MLflow, DVC, Kubeflow, or similar . Optimize model deployment for inference on NVIDIA hardware using TensorRT, Triton Inference Server , or ONNX Runtime-GPU . Manage cloud/on-prem GPU infrastructure and monitor resource utilization and model performance in production. Collaborate with data scientists to transition models from research to production-ready pipelines. Required Skills: Proficiency in Python and ML libraries (e.g., TensorFlow, PyTorch, Scikit-learn). Strong experience with Docker , Kubernetes , and NVIDIA GPU containerization (NGC, nvidia-docker) . Familiarity with NVIDIA Triton Inference Server , TensorRT , and CUDA . Experience with CI/CD for ML (GitHub Actions, GitLab CI, Jenkins, etc.). Deep understanding of ML lifecycle management , monitoring, and retraining. Experience working with cloud platforms (AWS/GCP/Azure) or on-prem GPU clusters. Preferred Qualifications: Experience with Kubeflow , Seldon Core , or similar orchestration tools. Exposure to Airflow , MLflow , Weights & Biases , or DVC . Knowledge of NVIDIA RAPIDS and distributed GPU workloads. MLOps certifications or NVIDIA Deep Learning Institute training (preferred but not mandatory). Show more Show less
Posted 3 weeks ago
7.0 years
0 Lacs
Gurugram, Haryana, India
On-site
About TwoSD (2SD Technologies Limited) TwoSD is the innovation engine of 2SD Technologies Limited , a global leader in product engineering, platform development, and advanced IT solutions. Backed by two decades of leadership in technology, our team brings together strategy, design, and data to craft transformative solutions for global clients. Our culture is built around cultivating talent, curiosity, and collaboration. Whether you're a career technologist, a self-taught coder, or a domain expert with a passion for real-world impact, TwoSD is where your journey accelerates. Join us and thrive. At 2SD Technologies, we push past the expected—with insight, integrity, and a passion for making things better. Role Overview We are hiring a Solution Architect with a proven track record in SaaS platform architecture , AI-driven solutions , and CRM/enterprise systems like Microsoft Dynamics 365 . This is a full-time position based in Gurugram, India , for professionals who thrive on solving complex problems across cloud, data, and application layers. You’ll design and orchestrate large-scale platforms that blend intelligence , automation , and multi-tenant scalability —powering real-time customer experiences, operational agility, and cross-system connectivity. Key Responsibilities Architect cloud-native SaaS solutions with scalability, modularity, and resilience at the core Design end-to-end technical architectures spanning CRM systems , custom apps , AI services , and data pipelines Lead technical discovery, solution workshops, and architecture governance with internal and client teams Drive the integration of Microsoft Dynamics 365 with other platforms including AI/ML services and business applications Create architectural blueprints and frameworks for microservices, event-driven systems, and intelligent automation Collaborate with engineers, data scientists, UX/UI designers, and DevOps teams to deliver platform excellence Oversee security, identity, compliance, and performance in high-scale environments Evaluate and introduce modern tools, frameworks, and architectural patterns for enterprise innovation Required Qualifications Bachelor’s degree in Computer Science, Engineering, or a related field (Master’s is a plus) 7+ years of experience in enterprise application architecture Hands-on expertise in Microsoft Dynamics 365 CE/CRM with complex integrations Experience architecting and delivering SaaS applications on cloud platforms (preferably AWS/Azure/GCP) Familiarity with LLM APIs , AI orchestration tools , or machine learning workflows Proven ability to lead multi-team and multi-technology architecture efforts Deep understanding of security , multi-tenancy , data privacy , and compliance standards Preferred Qualifications Microsoft Certified: Dynamics 365 + Azure/AWS Architect certifications Experience with AI platform components like OpenAI, LangChain, or Azure/AWS Services Experience designing or re-architecting legacy monoliths into cloud-native microservices Familiarity with DevOps and Infrastructure as Code (IaC) practices using Terraform or Bicep Experience integrating event-based systems using AWS, Azure Event Grid, Service Bus, or Kafka Exposure to enterprise observability tools and performance monitoring strategies Core Competencies Enterprise SaaS Architecture Cloud-Native Platform Design (Azure preferred) CRM + AI Integration Strategy End-to-End System Thinking Cross-Functional Collaboration & Mentorship Future-Proof Solution Design & Documentation Tools & Platforms CRM/ERP: Microsoft Dynamics 365 CE, Power Platform, Dataverse AI & Data: OpenAI, AWS SageMaker, AWS Bedrock, Azure Cognitive Services, LangChain, MLflow Cloud: Azure (App Services, API Management, Logic Apps, Functions, Cosmos DB) DevOps & IaC: GitHub Actions, Azure DevOps, Terraform, Bicep Integration: REST/GraphQL APIs, Azure Service Bus, Event Grid, Kafka Modeling & Docs: Lucidchart, Draw.io, ArchiMate, PlantUML Agile & Collaboration: Jira, Confluence, Slack, MS Teams Why Join TwoSD? At TwoSD , innovation isn’t a department—it’s a mindset. Here, your voice matters, your expertise is valued, and your growth is supported by a collaborative culture that blends mentorship with autonomy. With access to cutting-edge tools, meaningful projects, and a global knowledge network, you’ll do work that counts—and evolve with every challenge. Solution Architect – SaaS Platforms, AI Solutions & Enterprise CRM Position: Solution Architect Location: Gurugram, India (Onsite/Hybrid) Company: TwoSD (2SD Technologies Limited) Industry: Enterprise Software / CRM / Cloud Platforms Employment Type: Permanent Date Posted: 26 May 2025 How to Apply To apply, send your resume and technical portfolio or project overview to hr@2sdtechnologies.com or visit our LinkedIn careers page. Show more Show less
Posted 3 weeks ago
3.0 years
0 Lacs
Bengaluru, Karnataka, India
On-site
This role is for one of Weekday's clients Salary range: Rs 1000000 - Rs 1500000 (ie INR 10-15 LPA) Min Experience: 3 years Location: Bengaluru JobType: full-time Requirements About the Role We are seeking a passionate and skilled AI Engineer to join our innovative engineering team. In this role, you will play a pivotal part in designing, developing, and deploying cutting-edge artificial intelligence solutions with a focus on natural language processing (NLP) , computer vision , and machine learning models using TensorFlow and related frameworks. You will work on challenging projects that leverage large-scale data, deep learning, and advanced AI techniques, helping transform business problems into smart, automated, and scalable solutions. If you're someone who thrives in a fast-paced, tech-driven environment and loves solving real-world problems with AI, we'd love to hear from you. Key Responsibilities Design, develop, train, and deploy AI/ML models using frameworks such as TensorFlow, Keras, and PyTorch. Implement solutions across NLP, computer vision, and deep learning domains, using advanced techniques such as transformers, CNNs, LSTMs, OCR, image classification, and object detection. Collaborate closely with product managers, data scientists, and software engineers to identify use cases, define architecture, and integrate AI solutions into products. Optimize model performance for speed, accuracy, and scalability, using industry best practices in model tuning, validation, and A/B testing. Deploy AI models to cloud platforms such as AWS, GCP, and Azure, leveraging their native AI/ML services for efficient and reliable operation. Stay up to date with the latest AI research, trends, and technologies, and propose how they can be applied within the company's context. Ensure model explainability, reproducibility, and compliance with ethical AI standards. Contribute to the development of MLOps pipelines for managing model versioning, CI/CD for ML, and monitoring deployed models in production. Required Skills & Qualifications 3+ years of hands-on experience building and deploying AI/ML models in production environments. Proficiency in TensorFlow and deep learning workflows; experience with PyTorch is a plus. Strong foundation in natural language processing (e.g., NER, text classification, sentiment analysis, transformers) and computer vision (e.g., image processing, object recognition). Experience deploying and managing AI models on AWS, Google Cloud Platform (GCP), and Microsoft Azure. Skilled in Python and relevant libraries such as NumPy, Pandas, OpenCV, Scikit-learn, Hugging Face Transformers, etc. Familiarity with model deployment tools such as TensorFlow Serving, Docker, and Kubernetes. Experience working in cross-functional teams and agile environments. Bachelor's or Master's degree in Computer Science, Artificial Intelligence, Data Science, or related field. Preferred Qualifications Experience with MLOps tools and pipelines (MLflow, Kubeflow, SageMaker, etc.). Knowledge of data privacy and ethical AI practices. Exposure to edge AI or real-time inference systems. Show more Show less
Posted 3 weeks ago
8.0 years
0 Lacs
Bengaluru, Karnataka, India
On-site
The Role: We are looking for an experienced Engineering Manager to own the strategy and execution of complex data initiatives across the organization. In this role, you will drive the architecture and roadmap of large-scale data platforms, ensuring alignment with business goals. As an Engineering leader you will lead a team of high-performing engineers to build scalable, low-latency, and fault-tolerant systems. Your team will power advanced analytics, AI/ML, and decision-making across the organization. You will play a critical role in shaping Razorpay’s data architecture, fostering a high-performance culture, and delivering solutions that support millions of users and businesses Roles and Responsibilities: Lead with Vision: Define the strategy and roadmap for Razorpay’s data platform, ensuring it aligns with business and technology goals. Team Leadership: Build, mentor, and grow a high-performing engineering team by fostering a collaborative, open, and learning-oriented culture. Architect Scalable Systems: Design and implement fault-tolerant, high-availability distributed systems for data ingestion, processing, and analytics. Collaboration: Work closely with product, business, and data science teams to identify requirements, set priorities, and build solutions that matter. Hands-On Contribution: Step into coding when needed to solve complex problems or provide architectural guidance. Operational Excellence: Establish and advocate for best practices in software development, deployment, monitoring, and incident management. Foster Innovation: Drive initiatives to adopt new technologies and continuously improve Razorpay’s data platform capabilities. Ownership: Take end-to-end ownership of projects, ensuring quality, scalability, and timely delivery. Mandatory Qualifications: Experience: 8+ years of professional experience in software/data engineering, with at least 2+ years managing high-performing teams. Technical Expertise: Proficiency in building and operating large-scale distributed systems. Familiarity with ETL and Analytical Platform and event-driven architectures. Expertise in Python, Scala, Golang , and SQL , with deep knowledge of distributed data systems . Proven ability to design real-time and batch pipelines using Kafka, Spark, Airflow , and Flink . Extensive experience working with cloud platforms (AWS, GCP, Azure) and tools like Terraform, Kubernetes, and Docker . Strong knowledge of MLOps frameworks (MLflow, TFX) and data governance tools (DataHub, Amundsen). Leadership Skills: Proven ability to inspire and lead teams while fostering a culture of accountability and excellence . Exceptional leadership skills with a proven ability to mentor engineers, lead projects , and work cross-functionally. Experience in performance management, hiring, mentoring, and coaching engineers . Strong collaboration skills with cross-functional teams and stakeholders. Soft Skills: Exceptional communication, decision-making, and problem-solving abilities. Nice-to-Have: Experience with AI/ML platforms, financial systems data platforms. Show more Show less
Posted 3 weeks ago
5.0 years
0 Lacs
India
On-site
Flexera saves customers billions of dollars in wasted technology spend. A pioneer in Hybrid ITAM and FinOps, Flexera provides award-winning, data-oriented SaaS solutions for technology value optimization (TVO), enabling IT, finance, procurement and cloud teams to gain deep insights into cost optimization, compliance and risks for each business service. Flexera One solutions are built on a set of definitive customer, supplier and industry data, powered by our Technology Intelligence Platform, that enables organizations to visualize their Enterprise Technology Blueprint™ in hybrid environments—from on-premises to SaaS to containers to cloud. We’re transforming the software industry. We’re Flexera. With more than 50,000 customers across the world, we’re achieving that goal. But we know we can’t do any of that without our team. Ready to help us re-imagine the industry during a time of substantial growth and ambitious plans? Come and see why we’re consistently recognized by Gartner, Forrester and IDC as a category leader in the marketplace. Learn more at flexera.com Job Summary: We are seeking an accomplished and forward-thinking Senior Data Scientist to join our Data Intelligence team. The ideal candidate will drive the transformation of manual data curation processes into automated, AI/ML-powered workflows, ensuring higher efficiency and accuracy in how we manage and utilize data. This role involves collaborating with cross-functional teams—including product managers, operations, and data engineering—to implement innovative solutions that enhance data quality, reduce manual effort, and drive business value. Key Responsibilities: Identify manual data curation and processing workflows, then design and implement AI/ML-driven automation solutions to improve efficiency. Lead the end-to-end development of ML solutions for data cleaning, validation, and standardization. Utilize statistical, machine learning, and deep learning techniques to solve complex data intelligence challenges. Perform exploratory data analysis to uncover insights, refine model features, and guide decision-making. Collaborate with data engineering teams to ensure robust and efficient data pipelines that feed into AI/ML systems. Research and implement state-of-the-art methodologies in ML, deep learning, and related fields. Oversee iterative experimentation, including A/B testing and model optimization. Mentor junior data scientists, reviewing code, models, and workflows for best practices. Communicate findings effectively to both technical and non-technical audiences. Champion AI/ML initiatives across the organization and identify new areas for innovation. Required Skills and Qualifications: 5+ years of hands-on experience in Data Science or Machine Learning roles, with a strong focus on automating data-centric processes. Proficiency in Python or R and core data science libraries (e.g., NumPy, Pandas, scikit-learn). Expertise in at least one deep learning framework (e.g., TensorFlow, PyTorch, Keras). Strong foundation in statistics, mathematics, or related quantitative fields. Proven experience deploying ML models or automation solutions in cloud environments (AWS, Azure, or GCP). Demonstrated success in reducing manual workload through ML-driven data curation or process optimization. Familiarity with version control (Git) and CI/CD workflows in a data science context. Exceptional problem-solving and analytical skills. Effective communication skills for presenting complex insights to diverse stakeholders. Preferred Qualifications: Experience with large-scale data processing frameworks (e.g., Spark) and distributed computing. Familiarity with specialized domains like Natural Language Processing or Computer Vision. Knowledge of MLOps tools (MLflow, Databricks, Kubeflow) and best practices for continuous model delivery. Background in experimentation methodologies, A/B testing, and performance evaluation. Experience in data governance, data lineage, or data stewardship best practices. Publications in AI/ML conferences or journals. Flexera is proud to be an equal opportunity employer. Qualified applicants will be considered for open roles regardless of age, ancestry, color, family or medical care leave, gender identity or expression, genetic information, marital status, medical condition, national origin, physical or mental disability, political affiliation, protected veteran status, race, religion, sex (including pregnancy), sexual orientation, or any other characteristic protected by local/national laws, policies and/or regulations. Flexera understands the value that results from employing a diverse, equitable, and inclusive workforce. We recognize that equity necessitates acknowledging past exclusion and that inclusion requires intentional effort. Our DEI (Diversity, Equity, and Inclusion) council is the driving force behind our commitment to championing policies and practices that foster a welcoming environment for all. We encourage candidates requiring accommodations to please let us know by emailing careers@flexera.com. Show more Show less
Posted 3 weeks ago
0 years
0 Lacs
India
On-site
Flexera saves customers billions of dollars in wasted technology spend. A pioneer in Hybrid ITAM and FinOps, Flexera provides award-winning, data-oriented SaaS solutions for technology value optimization (TVO), enabling IT, finance, procurement and cloud teams to gain deep insights into cost optimization, compliance and risks for each business service. Flexera One solutions are built on a set of definitive customer, supplier and industry data, powered by our Technology Intelligence Platform, that enables organizations to visualize their Enterprise Technology Blueprint™ in hybrid environments—from on-premises to SaaS to containers to cloud. We’re transforming the software industry. We’re Flexera. With more than 50,000 customers across the world, we’re achieving that goal. But we know we can’t do any of that without our team. Ready to help us re-imagine the industry during a time of substantial growth and ambitious plans? Come and see why we’re consistently recognized by Gartner, Forrester and IDC as a category leader in the marketplace. Learn more at flexera.com Job Summary: We are seeking a skilled and motivated Senior Data Engineer to join our Automation, AI/ML team. In this role, you will work on designing, building, and maintaining data pipelines and infrastructure to support AI/ML initiatives, while contributing to the automation of key processes. This position requires expertise in data engineering, cloud technologies, and database systems, with a strong emphasis on scalability, performance, and innovation. Key Responsibilities: Identify and automate manual processes to improve efficiency and reduce operational overhead. Design, develop, and optimize scalable data pipelines to integrate data from multiple sources, including Oracle and SQL Server databases. Collaborate with data scientists and AI/ML engineers to ensure efficient access to high-quality data for training and inference models. Implement automation solutions for data ingestion, processing, and integration using modern tools and frameworks. Monitor, troubleshoot, and enhance data workflows to ensure performance, reliability, and scalability. Apply advanced data transformation techniques, including ETL/ELT processes, to prepare data for AI/ML use cases. Develop solutions to optimize storage and compute costs while ensuring data security and compliance. Required Skills and Qualifications: Experience in identifying, streamlining, and automating repetitive or manual processes. Proven experience as a Data Engineer, working with large-scale database systems (e.g., Oracle, SQL Server) and cloud platforms (AWS, Azure, Google Cloud). Expertise in building and maintaining data pipelines using tools like Apache Airflow, Talend, or Azure Data Factory. Strong programming skills in Python, Scala, or Java for data processing and automation tasks. Experience with data warehousing technologies such as Snowflake, Redshift, or Azure Synapse. Proficiency in SQL for data extraction, transformation, and analysis. Familiarity with tools such as Databricks, MLflow, or H2O.ai for integrating data engineering with AI/ML workflows. Experience with DevOps practices and tools, such as Jenkins, GitLab CI/CD, Docker, and Kubernetes. Knowledge of AI/ML concepts and their integration into data workflows. Strong problem-solving skills and attention to detail. Preferred Qualifications: Knowledge of security best practices, including data encryption and access control. Familiarity with big data technologies like Hadoop, Spark, or Kafka. Exposure to Databricks for data engineering and advanced analytics workflows. Flexera is proud to be an equal opportunity employer. Qualified applicants will be considered for open roles regardless of age, ancestry, color, family or medical care leave, gender identity or expression, genetic information, marital status, medical condition, national origin, physical or mental disability, political affiliation, protected veteran status, race, religion, sex (including pregnancy), sexual orientation, or any other characteristic protected by local/national laws, policies and/or regulations. Flexera understands the value that results from employing a diverse, equitable, and inclusive workforce. We recognize that equity necessitates acknowledging past exclusion and that inclusion requires intentional effort. Our DEI (Diversity, Equity, and Inclusion) council is the driving force behind our commitment to championing policies and practices that foster a welcoming environment for all. We encourage candidates requiring accommodations to please let us know by emailing careers@flexera.com. Show more Show less
Posted 3 weeks ago
7.0 years
0 Lacs
Hyderabad, Telangana, India
On-site
Blend is hiring a Lead Data Scientist (Generative AI) to spearhead the development of advanced AI-powered classification and matching systems on Databricks. You will contribute to flagship programs like the Diageo AI POC by building RAG pipelines, deploying agentic AI workflows, and scaling LLM-based solutions for high-precision entity matching and MDM modernization. Key Responsibilitie s Design and implement end-to-end AI pipelines for product classification, fuzzy matching, and deduplication using LLMs, RAG, and Databricks-native workflow s.Develop scalable, reproducible AI solutions within Databricks notebooks and job clusters, leveraging Delta Lake, MLflow, and Unity Catalo g.Engineer Retrieval-Augmented Generation (RAG) workflows using vector search and integrate with Python-based matching logi c.Build agent-based automation pipelines (rule-driven + GenAI agents) for anomaly detection, compliance validation, and harmonization logi c.Implement explainability, audit trails, and governance-first AI workflows aligned with enterprise-grade MDM need s.Collaborate with data engineers, BI teams, and product owners to integrate GenAI outputs into downstream system s.Contribute to modular system design and documentation for long-term scalability and maintainabilit y. Qualificati ons Bachelor’s/Master’s in Computer Science, Artificial Intelligence, or related fi eld.7+ years of overall Data Science experience with 2+ years in Generative AI / LLM-based applicati ons.Deep experience with Databricks ecosystem: Delta Lake, MLflow, DBFS, Databricks Jobs & Workfl ows.Strong Python and PySpark skills with ability to build scalable data pipelines and AI workflows in Databri cks.Experience with LLMs (e.g., OpenAI, LLaMA, Mistral) and frameworks like LangChain or LlamaIn dex.Working knowledge of vector databases (e.g., FAISS, Chroma) and prompt engineering for classification/retrie val.Exposure to MDM platforms (e.g., Stibo STEP) and familiarity with data harmonization challen ges.Experience with explainability frameworks (e.g., SHAP, LIME) and AI audit tool ing. Preferred S kills Knowledge of agentic AI architectures and multi-agent orchestr ation.Familiarity with Azure Data Hub and enterprise data ingestion frame works.Understanding of data governance, lineage, and regulatory compliance in AI sy stems. Thrive & Grow w ith Us: Competitiv e Salary: Your skills and contributions are highly valued here, and we make sure your salary reflects that, rewarding you fairly for the knowledge and experience you bring to th e table.Dynamic Caree r Growth: Our vibrant environment offers you the opportunity to grow rapidly, providing the right tools, mentorship, and experiences to fast-track your career.Id ea Tanks: Innovation lives here. Our "Idea Tanks" are your playground to pitch, experiment, and collaborate on ideas that can shape the future.Growt h Chats: Dive into our casual "Growth Chats" where you can learn from the best—whether it's over lunch or during a laid-back session with peers, it's the perfect space to grow your skills.Sn ack Zone: Stay fuelled and inspired! In our Snack Zone, you'll find a variety of snacks to keep your energy high and ideas flowing.Recognition & Rewards: We believe great work deserves to be recognized. Expect regular Hive-Fives, shoutouts and the chance to see your ideas come to life as part of our reward program.Fuel Your Growth Journey with Certif ications: We’re all about your growth groove! Level up your skills with our support as we cover the cost of your certifi cations. Show more Show less
Posted 3 weeks ago
8.0 years
0 Lacs
India
Remote
🔍 We're Hiring! – ML Ops Engineer (Remote, India) 📍 Location: Remote (Within India) 💼 Employment Type: Full-Time / Contractor 📅 Start Date: Immediate 🕒 Working Hours: 1:30 PM IST – 10:30 PM IST (Aligned with US CST) 🚀 Join Madlabs Global LLC as we lead the charge in deploying cutting-edge ML and Generative AI solutions at scale! We’re looking for a highly skilled ML Ops Engineer to lead the development, deployment, and lifecycle management of AI/ML models in cloud-native (preferably GCP) environments . 💼 Key Responsibilities Build scalable ML pipelines: ingestion, preprocessing, training, and serving. Collaborate with Data Scientists to turn prototypes into production-ready systems. Deploy and optimize LLM-based applications (instruction-tuned, fine-tuned models). Own continuous learning pipelines: retraining, model drift detection, performance tuning. Automate workflows using CI/CD , MLFlow , orchestration tools. Leverage GCP services like Vertex AI, BigQuery, Dataflow, Pub/Sub, Cloud Functions. Use Docker & Kubernetes to containerize and orchestrate model deployments. Monitor model performance with Prometheus, TensorBoard, Grafana, etc. Ensure security, fairness, and compliance across ML systems. 🧠 Required Experience 8+ years in ML Engineering, MLOps, or AI Infrastructure roles. Strong coding skills in Python with frameworks like TensorFlow, PyTorch, Scikit-learn. Deep expertise in GCP-native ML stacks . Hands-on experience in Generative AI model deployment and model optimization . Proficiency in Docker, Kubernetes, Jenkins, GitLab CI/CD . Solid understanding of model monitoring, versioning, rollback, and governance. 🕘 Work Hours Fully remote (India-based) Must provide overlap with CST time zone – working hours: 1:30 PM IST to 10:30 PM IST 💬 Interested or want to learn more? 📞 Contact: +91 98868 11767 📧 Email: naveed@madlabsinfotech.com Apply now or DM us to explore this opportunity to work with a team pushing the boundaries of AI innovation! #Hiring #MLOps #MachineLearning #GenerativeAI #LLM #VertexAI #RemoteJobsIndia #DataEngineering #AIJobs #GCP #DevOpsForAI #MLDeployment #LinkedInJobs Show more Show less
Posted 3 weeks ago
3.0 - 4.0 years
0 Lacs
Chennai, Tamil Nadu, India
On-site
Organizations everywhere struggle under the crushing costs and complexities of “solutions” that promise to simplify their lives. To create a better experience for their customers and employees. To help them grow. Software is a choice that can make or break a business. Create better or worse experiences. Propel or throttle growth. Business software has become a blocker instead of ways to get work done. There’s another option. Freshworks. With a fresh vision for how the world works. At Freshworks, we build uncomplicated service software that delivers exceptional customer and employee experiences. Our enterprise-grade solutions are powerful, yet easy to use, and quick to deliver results. Our people-first approach to AI eliminates friction, making employees more effective and organizations more productive. Over 72,000 companies, including Bridgestone, New Balance, Nucor, S&P Global, and Sony Music, trust Freshworks’ customer experience (CX) and employee experience (EX) software to fuel customer loyalty and service efficiency. And, over 4,500 Freshworks employees make this possible, all around the world. Fresh vision. Real impact. Come build it with us. Job Description We’re looking for a Jr AI Security Architect to join our growing Security Architecture team. This role will support the design, implementation, and protection of AI/ML systems, models, and datasets. The ideal candidate is passionate about the intersection of artificial intelligence and cybersecurity, and eager to contribute to building secure-by-design AI systems that protect users, data, and business integrity. Key Responsibilities Secure AI Model Development - Partner with AI/ML teams to embed security into the model development lifecycle, including during data collection, model training, evaluation, and deployment. - Contribute to threat modeling exercises for AI/ML pipelines to identify risks such as model poisoning, data leakage, or adversarial input attacks. - Support the evaluation and implementation of model explainability, fairness, and accountability techniques to address security and compliance concerns. - Develop and train internal models for security purposes Model Training & Dataset Security - Help design controls to ensure the integrity and confidentiality of training datasets, including the use of differential privacy, data validation pipelines, and access controls. - Assist in implementing secure storage and version control practices for datasets and model artifacts. - Evaluate training environments for exposure to risks such as unauthorized data access, insecure third-party libraries, or compromised containers. AI Infrastructure Hardening - Work with infrastructure and MLOps teams to secure AI platforms (e.g., MLFlow, Kubeflow, SageMaker, Vertex AI) including compute resources, APIs, CI/CD pipelines, and model registries. - Contribute to security reviews of AI-related deployments in cloud and on-prem environments. - Assist in automating security checks in AI pipelines, such as scanning for secrets, validating container images, and enforcing secure permissions. Secure AI Integration in Products - Participate in the review and assessment of AI/ML models embedded into customer-facing products to ensure they comply with internal security and responsible AI guidelines. - Help develop misuse detection and monitoring strategies to identify model abuse (e.g., prompt injection, data extraction, hallucination exploitation). - Support product security teams in designing guardrails and sandboxing techniques for generative AI features (e.g., chatbots, image generators, copilots). Knowledge Sharing & Enablement - Assist in creating internal training and security guidance for data scientists, engineers, and developers on secure AI practices. - Help maintain documentation, runbooks, and security checklists specific to AI/ML workloads. - Stay current on emerging AI security threats, industry trends, and tools; contribute to internal knowledge sharing. Qualifications - 3-4 years of experience in LLM and 7-10 years of experience in cybersecurity, machine learning, or related fields. - Familiarity with ML frameworks (e.g., PyTorch, TensorFlow) and MLOps tools (e.g., MLFlow, Airflow, Kubernetes). - Familiarity with AI models and Supplychain risks - Understanding of common AI/ML security threats and mitigations (e.g., model inversion, adversarial examples, data poisoning). - Experience working with cloud environments (AWS, GCP, Azure) and securing workloads. - Some knowledge of responsible AI principles, privacy-preserving ML, or AI compliance frameworks is a plus. Soft Skills - Strong communication skills to collaborate across engineering, data science, and product teams. - A continuous learning mindset and willingness to grow in both AI and security domains. - Problem-solving approach with a focus on practical, scalable solutions. Additional Information At Freshworks, we are creating a global workplace that enables everyone to find their true potential, purpose, and passion irrespective of their background, gender, race, sexual orientation, religion and ethnicity. We are committed to providing equal opportunity for all and believe that diversity in the workplace creates a more vibrant, richer work environment that advances the goals of our employees, communities and the business. Show more Show less
Posted 3 weeks ago
3.0 - 4.0 years
0 Lacs
Chennai, Tamil Nadu, India
On-site
Organizations everywhere struggle under the crushing costs and complexities of “solutions” that promise to simplify their lives. To create a better experience for their customers and employees. To help them grow. Software is a choice that can make or break a business. Create better or worse experiences. Propel or throttle growth. Business software has become a blocker instead of ways to get work done. There’s another option. Freshworks. With a fresh vision for how the world works. At Freshworks, we build uncomplicated service software that delivers exceptional customer and employee experiences. Our enterprise-grade solutions are powerful, yet easy to use, and quick to deliver results. Our people-first approach to AI eliminates friction, making employees more effective and organizations more productive. Over 72,000 companies, including Bridgestone, New Balance, Nucor, S&P Global, and Sony Music, trust Freshworks’ customer experience (CX) and employee experience (EX) software to fuel customer loyalty and service efficiency. And, over 4,500 Freshworks employees make this possible, all around the world. Fresh vision. Real impact. Come build it with us. Job Description We’re looking for a Jr AI Security Architect to join our growing Security Architecture team. This role will support the design, implementation, and protection of AI/ML systems, models, and datasets. The ideal candidate is passionate about the intersection of artificial intelligence and cybersecurity, and eager to contribute to building secure-by-design AI systems that protect users, data, and business integrity. Key Responsibilities Secure AI Model Development - Partner with AI/ML teams to embed security into the model development lifecycle, including during data collection, model training, evaluation, and deployment. - Contribute to threat modeling exercises for AI/ML pipelines to identify risks such as model poisoning, data leakage, or adversarial input attacks. - Support the evaluation and implementation of model explainability, fairness, and accountability techniques to address security and compliance concerns. - Develop and train internal models for security purposes Model Training & Dataset Security - Help design controls to ensure the integrity and confidentiality of training datasets, including the use of differential privacy, data validation pipelines, and access controls. - Assist in implementing secure storage and version control practices for datasets and model artifacts. - Evaluate training environments for exposure to risks such as unauthorized data access, insecure third-party libraries, or compromised containers. AI Infrastructure Hardening - Work with infrastructure and MLOps teams to secure AI platforms (e.g., MLFlow, Kubeflow, SageMaker, Vertex AI) including compute resources, APIs, CI/CD pipelines, and model registries. - Contribute to security reviews of AI-related deployments in cloud and on-prem environments. - Assist in automating security checks in AI pipelines, such as scanning for secrets, validating container images, and enforcing secure permissions. Secure AI Integration in Products - Participate in the review and assessment of AI/ML models embedded into customer-facing products to ensure they comply with internal security and responsible AI guidelines. - Help develop misuse detection and monitoring strategies to identify model abuse (e.g., prompt injection, data extraction, hallucination exploitation). - Support product security teams in designing guardrails and sandboxing techniques for generative AI features (e.g., chatbots, image generators, copilots). Knowledge Sharing & Enablement - Assist in creating internal training and security guidance for data scientists, engineers, and developers on secure AI practices. - Help maintain documentation, runbooks, and security checklists specific to AI/ML workloads. - Stay current on emerging AI security threats, industry trends, and tools; contribute to internal knowledge sharing. Qualifications - 3-4 years of experience in LLM and 7-10 years of experience in cybersecurity, machine learning, or related fields. - Familiarity with ML frameworks (e.g., PyTorch, TensorFlow) and MLOps tools (e.g., MLFlow, Airflow, Kubernetes). - Familiarity with AI models and Supplychain risks - Understanding of common AI/ML security threats and mitigations (e.g., model inversion, adversarial examples, data poisoning). - Experience working with cloud environments (AWS, GCP, Azure) and securing workloads. - Some knowledge of responsible AI principles, privacy-preserving ML, or AI compliance frameworks is a plus. Soft Skills - Strong communication skills to collaborate across engineering, data science, and product teams. - A continuous learning mindset and willingness to grow in both AI and security domains. - Problem-solving approach with a focus on practical, scalable solutions. Additional Information At Freshworks, we are creating a global workplace that enables everyone to find their true potential, purpose, and passion irrespective of their background, gender, race, sexual orientation, religion and ethnicity. We are committed to providing equal opportunity for all and believe that diversity in the workplace creates a more vibrant, richer work environment that advances the goals of our employees, communities and the business. Show more Show less
Posted 3 weeks ago
0 years
0 Lacs
Chennai, Tamil Nadu, India
On-site
In Norconsulting we are currently looking for an AI developer to join us in Chennai in a freelancer opportunity for a major Banking organization. Duration : long term Location : Chennai Rate : 110 USD/day (around 2200 USD per month) Type of assignment: Full-time (8h/day, Monday to Friday) SKILLS / EXPERIENCE REQUIRED AI Developer • Large Language Models (LLMs) & Prompt Engineering: Experience working with transformer-based models (e.g., GPT, BERT) and crafting effective prompts for tasks like summarization, text classification and document understanding. • Azure Document Intelligence: Hands-on experience with Azure AI Document Intelligence for extracting structured data from unstructured documents (invoices, forms, contracts). • Model Development & Evaluation: Strong foundation in ML algorithms, model evaluation metrics, and hyperparameter tuning using tools like Scikit-learn, XGBoost, or PyTorch. • MLOps (Machine Learning Operations): Proficient in building and managing ML pipelines using Azure ML, MLflow, and CI/CD tools for model training, deployment, and monitoring. • Azure Machine Learning (Azure ML): Experience with Azure ML Studio, automated ML, model registry, and deployment to endpoints or containers. • Azure Functions & Serverless AI: Building event-driven AI workflows using Azure Functions for real-time inference, data processing, and integration with other Azure services. • Programming Languages: Strong coding skills in Python (preferred), with knowledge of libraries like NumPy, Pandas, Scikit-learn, and Matplotlib. • Database & Data Lakes: Experience with SQL and NoSQL databases, and integration with data lakes for AI pipelines. • DevOps & Git Integration: Experience with Azure DevOps for version control, testing, and continuous integration of AI workflows. WBGJP00012309 Show more Show less
Posted 3 weeks ago
5.0 years
0 Lacs
Ahmedabad, Gujarat, India
Remote
Senior DevOps Engineer Experience: 5 - 9 Years Exp Salary : Competitive Preferred Notice Period : Within 30 Days Shift : 10:00AM to 7:00PM IST Opportunity Type: Onsite (Ahmedabad) Placement Type: Permanent (*Note: This is a requirement for one of Uplers' Clients) Must have skills required : Azure (Microsoft Azure), Docker/Terraform, TensorFlow, Python, AWS Good to have skills : Kubeflow, MLFlow Attri (One of Uplers' Clients) is Looking for: Senior DevOps Engineer who is passionate about their work, eager to learn and grow, and who is committed to delivering exceptional results. If you are a team player, with a positive attitude and a desire to make a difference, then we want to hear from you. Role Overview Description About Attri Attri is an AI organization that helps businesses initiate and accelerate their AI efforts. We offer the industry’s first end-to-end enterprise machine learning platform, empowering teams to focus on ML development rather than infrastructure. From ideation to execution, our global team of AI experts supports organizations in building scalable, state-of-the-art ML solutions. Our mission is to redefine businesses by harnessing cutting-edge technology and a unique, value-driven approach. With team members across continents, we celebrate diversity, curiosity, and innovation. We’re now looking for a Senior DevOps Engineer to join our fast-growing, remote-first team. If you're passionate about automation, scalable cloud systems, and supporting high-impact AI workloads, we’d love to connect. What You'll Do (Responsibilities): Design, implement, and manage scalable, secure, and high-performance cloud-native infrastructure across Azure. Build and maintain Infrastructure as Code (IaC) using Terraform or CloudFormation. Develop event-driven and serverless architectures using AWS Lambda, SQS, and SAM. Architect and manage containerized applications using Docker, Kubernetes, ECR, ECS, or AKS. Establish and optimize CI/CD pipelines using GitHub Actions, Jenkins, AWS CodeBuild & CodePipeline. Set up and manage monitoring, logging, and alerting using Prometheus + Grafana, Datadog, and centralized logging systems. Collaborate with ML Engineers and Data Engineers to support MLOps pipelines (Airflow, ML Pipelines) and Bedrock with Tensorflow or PyTorch. Implement and optimize ETL/data streaming pipelines using Kafka, EventBridge, and Event Hubs. Automate operations and system tasks using Python and Bash, along with Cloud CLIs and SDKs. Secure infrastructure using IAM/RBAC and follow best practices in secrets management and access control. Manage DNS and networking configurations using Cloudflare, VPC, and PrivateLink. Lead architecture implementation for scalable and secure systems, aligning with business and AI solution needs. Conduct cost optimization through budgeting, alerts, tagging, right-sizing resources, and leveraging spot instances. Contribute to backend development in Python (Web Frameworks), REST/Socket and gRPC design, and testing (unit/integration). Participate in incident response, performance tuning, and continuous system improvement. Good to Have: Hands-on experience with ML lifecycle tools like MLflow and Kubeflow Previous involvement in production-grade AI/ML projects or data-intensive systems Startup or high-growth tech company experience Qualifications: Bachelor’s degree in Computer Science, Information Technology, or a related field. 5+ years of hands-on experience in a DevOps, SRE, or Cloud Infrastructure role. Proven expertise in multi-cloud environments (AWS, Azure, GCP) and modern DevOps tooling. Strong communication and collaboration skills to work across engineering, data science, and product teams. Benefits: Competitive Salary 💸 Support for continual learning (free books and online courses) 📚 Leveling Up Opportunities 🌱 Diverse team environment 🌍 How to apply for this opportunity: Easy 3-Step Process: 1. Click On Apply! And Register or log in on our portal 2. Upload updated Resume & Complete the Screening Form 3. Increase your chances to get shortlisted & meet the client for the Interview! About Our Client: Attri, an AI organization, leads the way in enterprise AI, offering advanced solutions and services driven by AI agents and powered by Foundation Models. Our comprehensive suite of AI-enabled tools drives business impact, enhances quality, mitigates risk, and also helps unlock growth opportunities. About Uplers: Our goal is to make hiring and getting hired reliable, simple, and fast. Our role will be to help all our talents find and apply for relevant product and engineering job opportunities and progress in their career. (Note: There are many more opportunities apart from this on the portal.) So, if you are ready for a new challenge, a great work environment, and an opportunity to take your career to the next level, don't hesitate to apply today. We are waiting for you! Show more Show less
Posted 3 weeks ago
3.0 years
0 Lacs
Bengaluru, Karnataka, India
On-site
Keen on working directly with our clients and developing innovative ML and LLM Solutions? Want to be part of our APAC Center of Excellence team delivering some of the most innovative solutions in Data & AI? Ready to join a growing company that has won Microsoft Partner Of The Year for Data & AI ? Practical Information : Location: Delhi/Noida/Mumbai/Bangalore/Chennai, India | Reports to: Director of CoE Data & AI | Language Requirements: Professional English, written and verbal | Work Arrangement: Hybrid Join our CoE team as our new ML/LLM Engineer where you'll be driving the development, deployment, and optimization of Large Language Models and machine learning solutions for production environments, both on-premises and on Cloud platforms. You will master your role with key performance indicators (KPIs) that revolve around successful implementation and efficiency of these solutions . Other responsibilities will include: Collaborating closely with clients to understand their requirements and delivering custom solutions Extending prototypes and enhancing them into robust and scalable solutions Designing and fine-tuning large language models tailored for specific applications Efficiently managing large language models in production environment s to facilitate close to real-time solutions Developing and executing ML pipelines, deploying models to enhance accuracy across various process steps within our pipeline Your Competencies: 3+ years of experience in NLP or a similar role Understanding of data structures, data modeling, ML algorithms, and software architecture Experience coding in Python with ML frameworks (e.g., Keras or PyTorch) and libraries (e.g., scikit-learn) Practical experience with ML Cloud services on Azure or AWS (e.g., Azure ML, Amazon SageMaker, or MLFlow) About You: Being learning-oriented with a dedication to staying updated on cutting-edge advancements in large language models and NLP Demonstrating strong analytical and problem-solving skills Exhibiting excellent communication skills What's in it for you: Medical, and life insurance Hybrid workplace Internet & Mobile reimbursement Upskilling through certifications and training At Crayon, we are deeply committed to fostering a culture of diversity, equity, inclusion, and belonging (DEIB). We believe that diversity in all its forms strengthens our team and enhances innovation and effectiveness. We welcome applications from individuals of all backgrounds, regardless of race, colour, age, origin, religion, sexual orientation, gender (identity), genetic information, neurodiversity, disability, or any other basis protected by local laws and regulations. When filling vacancies, we prioritize equally qualified candidates who bring diverse backgrounds and experiences, helping to enrich our team dynamics and foster a vibrant, inclusive work environment. If you require any assistance or reasonable accommodation during the application process, please let us know. Apply to join an award-winning employer! Show more Show less
Posted 3 weeks ago
12.0 years
0 Lacs
Chennai, Tamil Nadu, India
On-site
Role Summary: As a Principal Software Engineer, you will be a pivotal technical leader, driving the architectural vision, design, and implementation of our most critical and complex software solutions. You will leverage your deep expertise across a broad range of technologies including .NET, React, Scala, cloud platforms (Azure preferred), data engineering(Databricks), and emerging AI/ML capabilities to build scalable, resilient, and high-performant systems. You will mentor fellow engineers, influence technical roadmaps across multiple teams, and champion engineering excellence and innovation, including the practical application of Generative AI tools. Key Responsibilities: Architectural Leadership & Design Lead the design, architecture, and development of highly scalable, available, and performant software solutions, utilizing modern technologies like .NET, React, Scala, and cloud-native services. Functionally decompose complex business and technical problems into simple, straight-forward, and extensible solutions. Ensure the quality of architecture and design of systems, considering system inter-dependencies, limitations, and long-term strategic goals. Defi ne and drive the technical strategy for data engineering initiatives, utilizing platforms like Azure Databricks for large-scale data processing, transformation, and analytics. Technical Execution & Innovation Possess expert knowledge in performance, scalability, enterprise system architecture, security best practices, and engineering best practices. Spearhead the integration and practical application of Generative AI tools and techniques to enhance developer productivity and build innovative product features. Develop and implement robust solutions on cloud platforms (Azure preferred),leveraging IaaS, PaaS, and serverless paradigms. Design and manage efficient SQL database schemas and optimize query performance for critical applications. Contribute to and guide MLOps practices for deploying, monitoring, and managing machine learning models in production (preferred). Stay current with industry trends, emerging technologies (especially in Cloud, Data, AI/ML, and distributed systems), and benchmark our technology against competing systems. Required Skills & Qualifications Experience: 12+ years of professional software engineering experience, with a proven track record of leading the design and delivery of large-scale, complex software systems. Technical Proficiency (Core): Programming Languages: Deep expertise in multiple languages, with significant experience in NET (C#) React (JavaScript/TypeScript) ,Scala and Python, Cloud Computing: Extensive hands-on experience with cloud platforms, with a strong preference for Microsoft Azure (e.g., Azure App Service, Azure Functions ,Azure SQL, Azure Data Factory, Azure Synapse, Azure Kubernetes Service).Familiarity with other clouds (AWS, GCP) is a plus. Databases: Strong proficiency in SQL databases (e.g., SQL Server, PostgreSQL),including schema design, query optimization, and performance tuning. Data Engineering: Proven experience in designing and implementing data pipelines, ETL/ELT processes, and big data solutions, specifically with Azure Databricks and Apache Spark. Architecture: Expertise in designing distributed systems, microservices architecture, event-driven architectures, and RESTful APIs. Software Design Principles: Solid understanding of OOP, SOLID principles, design patterns, and domain-driven design. Technical Proficiency (Emerging & Important): Generative AI: Demonstrable experience in utilizing Generative AI tools (e.g., GitHub Copilot, large language models via APIs) for code generation, problem-solving, and to enhance development workflows. Understanding of how to integrate AI capabilities into applications. DevOps & CI/CD: Familiarity with CI/CD pipelines (e.g., Azure DevOps, Jenkins ,GitHub Actions), Infrastructure as Code (e.g., Terraform, ARM Templates), and containerization (Docker, Kubernetes). Soft Skills: Exceptional problem-solving and analytical skills. Strong leadership and mentoring capabilities. Excellent communication and interpersonal skills, with the ability to articulate complex technical concepts to diverse audiences. Proven ability to influence without authority and drive consensus. Strategic thinker with a passion for innovation and continuous improvement. Education: Bachelor’s or Master’s degree in Computer Science, Software Engineering, or related technical fi eld, or equivalent practical experience. Preferred Skills & Qualifications Experience with MLOps principles, tools, and practices for managing the lifecycle of machine learning models (e.g., MLflow, Azure Machine Learning). Advanced Azure Services: Deep expertise in specific c Azure services like AzureCosmos DB, Azure Event Hubs/Kafka, Azure Cognitive Services, Azure Foundry. Security: Solid understanding of security best practices, threat modeling, and secure development lifecycle (SDL). Experience with identity and access management (IAM) concepts. NoSQL Databases: Experience with NoSQL databases (e.g., MongoDB, Cassandra, Cosmos DB). Performance Optimization: Expertise in optimizing high-traffic, real-time systems for latency and throughput. Contributions: Contributions to open-source projects, patents, or publications in relevant fi elds. Show more Show less
Posted 3 weeks ago
7.0 - 12.0 years
30 - 45 Lacs
Bengaluru
Work from Office
Build and deploy scalable ML models and MLOps pipelines in collaboration with data scientists Required Candidate profile 6–12 yrs in ML development, Python, model tuning, and enterprise AI deployment.
Posted 3 weeks ago
5.0 years
0 Lacs
Greater Bengaluru Area
On-site
What if the work you did every day could impact the lives of people you know? Or all of humanity? At Illumina, we are expanding access to genomic technology to realize health equity for billions of people around the world. Our efforts enable life-changing discoveries that are transforming human health through the early detection and diagnosis of diseases and new treatment options for patients. Working at Illumina means being part of something bigger than yourself. Every person, in every role, has the opportunity to make a difference. Surrounded by extraordinary people, inspiring leaders, and world changing projects, you will do more and become more than you ever thought possible. Position Summary We are seeking a highly skilled Senior Data Engineer Developer with 5+ years of experience to join our talented team in Bangalore. In this role, you will be responsible for designing, implementing, and optimizing data pipelines, ETL processes, and data integration solutions using Python, Spark, SQL, Snowflake, dbt, and other relevant technologies. Additionally, you will bring strong domain expertise in operations organizations, with a focus on supply chain and manufacturing functions. If you're a seasoned data engineer with a proven track record of delivering impactful data solutions in operations contexts, we want to hear from you. Responsibilities Lead the design, development, and optimization of data pipelines, ETL processes, and data integration solutions using Python, Spark, SQL, Snowflake, dbt, and other relevant technologies. Apply strong domain expertise in operations organizations, particularly in functions like supply chain and manufacturing, to understand data requirements and deliver tailored solutions. Utilize big data processing frameworks such as Apache Spark to process and analyze large volumes of operational data efficiently. Implement data transformations, aggregations, and business logic to support analytics, reporting, and operational decision-making. Leverage cloud-based data platforms such as Snowflake to store and manage structured and semi-structured operational data at scale. Utilize dbt (Data Build Tool) for data modeling, transformation, and documentation to ensure data consistency, quality, and integrity. Monitor and optimize data pipelines and ETL processes for performance, scalability, and reliability in operations contexts. Conduct data profiling, cleansing, and validation to ensure data quality and integrity across different operational data sets. Collaborate closely with cross-functional teams, including operations stakeholders, data scientists, and business analysts, to understand operational challenges and deliver actionable insights. Stay updated on emerging technologies and best practices in data engineering and operations management, contributing to continuous improvement and innovation within the organization. All listed requirements are deemed as essential functions to this position; however, business conditions may require reasonable accommodations for additional task and responsibilities. Preferred Experience/Education/Skills Bachelor's degree in Computer Science, Engineering, Operations Management, or related field. 5+ years of experience in data engineering, with proficiency in Python, Spark, SQL, Snowflake, dbt, and other relevant technologies. Strong domain expertise in operations organizations, particularly in functions like supply chain and manufacturing. Strong domain expertise in life sciences manufacturing equipment, with a deep understanding of industry-specific challenges, processes, and technologies. Experience with big data processing frameworks such as Apache Spark and cloud-based data platforms such as Snowflake. Hands-on experience with data modeling, ETL development, and data integration in operations contexts. Familiarity with dbt (Data Build Tool) for managing data transformation and modeling workflows. Familiarity with reporting and visualization tools like Tableau, Powerbi etc. Good understanding of advanced data engineering and data science practices and technologies like pypark, sagemaker, cloudera MLflow etc. Experience with SAP, SAP HANA and Teamcenter applications is a plus. Excellent problem-solving skills, analytical thinking, and attention to detail. Strong communication and interpersonal skills, with the ability to collaborate effectively with cross-functional teams and operations stakeholders. Eagerness to learn and adapt to new technologies and tools in a fast-paced environment. Illumina believes that everyone has the ability to make an impact, and we are proud to be an equal opportunity employer committed to providing employment opportunity regardless of sex, race, creed, color, gender, religion, marital status, domestic partner status, age, national origin or ancestry, physical or mental disability, medical condition, sexual orientation, pregnancy, military or veteran status, citizenship status, and genetic information. Show more Show less
Posted 3 weeks ago
3.0 years
0 Lacs
Bengaluru, Karnataka, India
Remote
P-928 At Databricks, we are passionate about enabling data teams to solve the world's toughest problems — from making the next mode of transportation a reality to accelerating the development of medical breakthroughs. We do this by building and running the world's best data and AI infrastructure platform so our customers can use deep data insights to improve their business. Founded by engineers — and customer obsessed — we leap at every opportunity to solve technical challenges, from designing next-gen UI/UX for interfacing with data to scaling our services and infrastructure across millions of virtual machines. And we're only getting started in Bengaluru , India ! As a software engineer with a backend focus, you will work with your team to build infrastructure for the Databricks platform at scale. The Impact You'll Have Our backend teams cover a diverse range of domains, from core compute fabric resource management to service platforms and infrastructure. For example, you might work on challenges such as: Supporting Databricks' growth by building foundational infrastructure platforms that enable seamless operation across numerous geographic regions and cloud providers. Implementing cloud-agnostic infrastructure abstractions to help Databricks engineers more efficiently manage and operate their services. Develop tools and processes that drive engineering efficiency at Databricks. We enhance the developer experience for Databricks engineers across various areas, including programming languages, linters, static analysis, IDEs, remote development environments, automated release pipelines, and test automation frameworks. Our current focus is on optimizing the Rust development experience across the organization. What We Look For BS (or higher) in Computer Science, or a related field 3+ years of production level experience in one of: Python, Java, Scala, C++, or similar language. Experience developing large-scale distributed systems from scratch Experience working on a SaaS platform or with Service-Oriented Architectures Proficiency in one or more backend languages such as Java, Scala, or Go. Hands-on experience in developing and operating backend systems. Ability to contribute effectively throughout all project phases, from initial design and development to implementation and ongoing operations, with guidance from senior team members. About Databricks Databricks is the data and AI company. More than 10,000 organizations worldwide — including Comcast, Condé Nast, Grammarly, and over 50% of the Fortune 500 — rely on the Databricks Data Intelligence Platform to unify and democratize data, analytics and AI. Databricks is headquartered in San Francisco, with offices around the globe and was founded by the original creators of Lakehouse, Apache Spark™, Delta Lake and MLflow. To learn more, follow Databricks on Twitter, LinkedIn and Facebook. Benefits At Databricks, we strive to provide comprehensive benefits and perks that meet the needs of all of our employees. For specific details on the benefits offered in your region, please visit https://www.mybenefitsnow.com/databricks. Our Commitment to Diversity and Inclusion At Databricks, we are committed to fostering a diverse and inclusive culture where everyone can excel. We take great care to ensure that our hiring practices are inclusive and meet equal employment opportunity standards. Individuals looking for employment at Databricks are considered without regard to age, color, disability, ethnicity, family or marital status, gender identity or expression, language, national origin, physical and mental ability, political affiliation, race, religion, sexual orientation, socio-economic status, veteran status, and other protected characteristics. Compliance If access to export-controlled technology or source code is required for performance of job duties, it is within Employer's discretion whether to apply for a U.S. government license for such positions, and Employer may decline to proceed with an applicant on this basis alone. Show more Show less
Posted 3 weeks ago
5.0 years
0 Lacs
Bengaluru, Karnataka, India
On-site
We’re looking for problem solvers, innovators, and dreamers who are searching for anything but business as usual. Like us, you’re a high performer who’s an expert at your craft, constantly challenging the status quo. You value inclusivity and want to join a culture that empowers you to show up as your authentic self. You know that success hinges on commitment, that our differences make us stronger, and that the finish line is always sweeter when the whole team crosses together. Senior Software Engineer We’re looking for problem solvers, innovators, and dreamers who are searching for anything but business as usual. Like us, you’re a high performer who’s an expert at your craft, constantly challenging the status quo. You value inclusivity and want to join a culture that empowers you to show up as your authentic self. You know that success hinges on commitment, that our differences make us stronger, and that the finish line is always sweeter when the whole team crosses together. Overview Why work for just any analytics company? At Alteryx, Inc., we are explorers, dreamers and innovators. We’re on a journey to build the best analytics platform in the world, but we can’t do it without people like you, leading the way. Forget the stereotypical tech companies of the past. Embrace the unconventional, exercise your imagination and help alter the future with Alteryx. Job Title: Senior Software Engineer - AI/ML Location: Bangalore (Hybrid) Department: Engineering / Data Science / AI Solutions Reports To: Engineering Manager / Technical Lead About Alteryx At Alteryx, we’re transforming the way businesses leverage data. Our AI/ML solutions empower teams to make data-driven decisions, and we’re seeking a Senior Software Engineer - AI/ML to join our engineering team in Bangalore. In this role, you’ll contribute to developing and deploying scalable AI/ML solutions, leveraging Python and React to build impactful applications. While experience with Scala is a plus, a strong understanding of AI/ML algorithms and cloud infrastructure (preferably GCP, but AWS experience is also welcome) is essential. Key Responsibilities Design, develop, and deploy scalable AI/ML models using Python and React. Collaborate with data scientists and ML engineers to integrate models into productionenvironments Build interactive and dynamic web applications using React to visualize AI/ML insights Develop and maintain data pipelines to support model training, evaluation, and deployment mplement best practices for building efficient, maintainable, and scalable machine learning solutions Design and optimize data processing systems using GCP AI/ML services (preferred) or AWS ML tools Collaborate with cross-functional teams to understand business requirements and translate them into technical solutions Drive code quality, conduct peer code reviews, and improve system performance Research and stay updated on emerging AI/ML frameworks, libraries, and trend Required Skills & Experience 5+ years of experience as a Software Engineer or ML Engineer Strong programming skills in Python (mandatory) with experience in building scalable backend systems Hands-on experience with React for developing dynamic and interactive UI components Solid understanding of AI/ML algorithms, such as regression models, decision trees, clustering, and neural networks Multiple-LLMs, GenAI - HuggingFace, LangChain, LangGraph etc Vector databases, RAG Cloud databases, Snowflake, data warehouses, data lake etc Experience in building, training, and deploying ML models in cloud environments like GCP (preferred) or AWS Familiarity with data pipelines, ETL processes, and model serving frameworks (e.g., MLFlow, Kubeflow, or Seldon) Strong grasp of software engineering best practices such as code reviews, version control, and CI/CD pipelines Excellent problem-solving skills and the ability to work independently or collaboratively in a fast-paced environment Preferred Skills Experience with Scala for data engineering or large-scale ML pipelines Familiarity with GCP AI Platform, Vertex AI, or Amazon SageMaker Knowledge of RESTful API development and microservices architecture Understanding of containerization and orchestration tools such as Docker and Kubernetes Why Join Us? Be part of a forward-thinking team that values innovation and collaboration. Work on impactful AI/ML projects that solve real-world business challenges. Enjoy a flexible work environment with opportunities for growth and development Access to cutting-edge tools, cloud platforms, and the latest advancements in AI/ML. Find yourself checking a lot of these boxes but doubting whether you should apply? At Alteryx, we support a growth mindset for our associates through all stages of their careers. If you meet some of the requirements and you share our values, we encourage you to apply. As part of our ongoing commitment to a diverse, equitable, and inclusive workplace, we’re invested in building teams with a wide variety of backgrounds, identities, and experiences. This position involves access to software/technology that is subject to U.S. export controls. Any job offer made will be contingent upon the applicant’s capacity to serve in compliance with U.S. export controls. Show more Show less
Posted 3 weeks ago
12.0 years
0 Lacs
Bengaluru, Karnataka, India
On-site
P-1348 At Databricks, we are passionate about enabling data teams to solve the world's toughest problems, from security threat detection to cancer drug development. We do this by building and running the world's best data and AI infrastructure platform, so our customers can focus on the high value challenges that are central to their own missions. Our engineering teams build technical products that fulfill real, important needs in the world. We always push the boundaries of data and AI technology, while simultaneously operating with the resilience, security and scale that is important to making customers successful on our platform. We develop and operate one of the largest scale software platforms. The fleet consists of millions of virtual machines, generating terabytes of logs and processing exabytes of data per day. At our scale, we observe cloud hardware, network, and operating system faults, and our software must gracefully shield our customers from any of the above. As a Staff Software Engineer working on the Data Platform team you will help build the Data Intelligence Platform for Databricks that will allow us to automate decision-making across the entire company. You will achieve this in collaboration with Databricks Product Teams, Data Science, Applied AI and many more. You will develop a variety of tools spanning logging, orchestration, data transformation, metric store, governance platforms, data consumption layers etc. You will do this using the latest, bleeding-edge Databricks product and other tools in the data ecosystem - the team also functions as a large, production, in-house customer that dog foods Databricks and guides the future direction of the product. The Impact You Will Have Design and run the Databricks metrics store that enables all business units and engineering teams to bring their detailed metrics into a common platform for sharing and aggregation, with high quality, introspection ability and query performance. Design and run the cross-company Data Intelligence Platform, which contains every business and product metric used to run Databricks. You’ll play a key role in developing the right balance of data protections and ease of shareability for the Data Intelligence Platform as we transition to a public company. Develop tooling and infrastructure to efficiently manage and run Databricks on Databricks at scale, across multiple clouds, geographies and deployment types. This includes CI/CD processes, test frameworks for pipelines and data quality, and infrastructure-as-code tooling. Design the base ETL framework used by all pipelines developed at the company. Partner with our engineering teams to provide leadership in developing the long-term vision and requirements for the Databricks product. Build reliable data pipelines and solve data problems using Databricks, our partner’s products and other OSS tools. Provide early feedback on the design and operations of these products. Establish conventions and create new APIs for telemetry, debug, feature and audit event log data, and evolve them as the product and underlying services change. Represent Databricks at academic and industrial conferences & events. What We Look For 12+ years of industry experience 6+ years of experience providing technical leadership on large projects similar to the ones described above - ETL frameworks, metrics stores, infrastructure management, data security. Experience building, shipping and operating reliable multi-geo data pipelines at scale. Experience working with and operating workflow or orchestration frameworks, including open source tools like Airflow and DBT or commercial enterprise tools. Experience with large-scale messaging systems like Kafka or RabbitMQ or commercial systems. Excellent cross-functional and communication skills, consensus builder. Passion for data infrastructure and for enabling others by making their data easier to access. About Databricks Databricks is the data and AI company. More than 10,000 organizations worldwide — including Comcast, Condé Nast, Grammarly, and over 50% of the Fortune 500 — rely on the Databricks Data Intelligence Platform to unify and democratize data, analytics and AI. Databricks is headquartered in San Francisco, with offices around the globe and was founded by the original creators of Lakehouse, Apache Spark™, Delta Lake and MLflow. To learn more, follow Databricks on Twitter, LinkedIn and Facebook. Benefits At Databricks, we strive to provide comprehensive benefits and perks that meet the needs of all of our employees. For specific details on the benefits offered in your region, please visit https://www.mybenefitsnow.com/databricks. Our Commitment to Diversity and Inclusion At Databricks, we are committed to fostering a diverse and inclusive culture where everyone can excel. We take great care to ensure that our hiring practices are inclusive and meet equal employment opportunity standards. Individuals looking for employment at Databricks are considered without regard to age, color, disability, ethnicity, family or marital status, gender identity or expression, language, national origin, physical and mental ability, political affiliation, race, religion, sexual orientation, socio-economic status, veteran status, and other protected characteristics. Compliance If access to export-controlled technology or source code is required for performance of job duties, it is within Employer's discretion whether to apply for a U.S. government license for such positions, and Employer may decline to proceed with an applicant on this basis alone. Show more Show less
Posted 3 weeks ago
5.0 years
0 Lacs
Bengaluru, Karnataka, India
On-site
P-1376 Job Overview We are seeking a highly skilled Frontline Senior Technical Solutions Engineer with over 5 years of experience to join our Platform Support team. This role is pivotal in delivering exceptional support for our Databricks Data Intelligence platform, addressing complex technical challenges, and ensuring the seamless operation of our data solutions. As a frontline engineer, you will be the primary point of contact for critical issues, working closely with both internal teams and customers to resolve high-impact problems and drive platform improvements. Key Responsibilities Frontline Support: Serve as the primary technical point of contact for escalated issues related to the Databricks Data Intelligence platform. Provide expert-level troubleshooting, diagnostics, and resolution for complex problems affecting system performance and reliability. Customer Interaction: Engage with customers directly to understand their technical issues and requirements. Provide timely, clear, and actionable solutions to ensure high levels of customer satisfaction. Incident Management: Lead the resolution of high-priority incidents, coordinating with various teams to address and mitigate issues swiftly. Conduct thorough root cause analyses and develop preventive measures to avoid recurrence. Collaboration: Work closely with engineering, product management, and DevOps teams to share insights, identify recurring issues, and drive improvements to the Databricks Data Intelligence platform. Documentation and Knowledge Sharing: Create and maintain detailed documentation on support procedures, known issues, and solutions. Contribute to internal knowledge bases and create training materials to assist other support engineers. Performance Monitoring: Monitor and analyze platform performance metrics to identify potential issues before they impact customers. Implement optimizations and enhancements to improve platform stability and efficiency. Platform Upgrades: Manage and oversee the deployment of Databricks Data Intelligence platform upgrades and patches, ensuring minimal disruption to services and maintaining system integrity. Innovation and Improvement: Stay abreast of industry trends and advancements in Databricks technology. Propose and drive initiatives to enhance platform capabilities and support processes. Customer Feedback: Collect and analyze customer feedback to drive continuous improvement in support processes and platform features. Qualifications Experience: Minimum of 5 years of hands-on experience in a technical support or engineering role related to Databricks Data Intelligence platform, cloud data platforms, or big data technologies. Technical Skills: A deep understanding of Databricks architecture and Apache Spark™, along with experience in cloud platforms like AWS, Azure, or GCP, is essential. Strong capabilities in designing and managing data pipelines, distributed computing are required. Proficiency in Unix/Linux administration, familiarity with DevOps practices, and skills in log analysis and monitoring tools are also crucial for effective troubleshooting and system optimization. Problem-Solving: Demonstrated ability to diagnose and resolve complex technical issues with a strong analytical and methodical approach. Communication: Exceptional verbal and written communication skills, with the ability to effectively convey technical information to both technical and non-technical stakeholders. Customer Focus: Proven experience in managing high-impact customer interactions and ensuring a positive customer experience. Collaboration: Ability to work effectively in a team environment, collaborating with engineering, product, and customer-facing teams. Education: Bachelor’s degree in Computer Science, Engineering, or a related field. Advanced degree or relevant certifications are highly desirable. Preferred Skills Experience with additional big data tools and technologies such as Hadoop, Kafka, or NoSQL databases. Familiarity with automation tools and CI/CD pipelines. Understanding of data governance and compliance requirements. Why Join Us? Innovative Environment: Work with cutting-edge technology in a fast-paced, innovative company. Career Growth: Opportunities for professional development and career advancement. Team Culture: Collaborate with a talented and motivated team dedicated to excellence and continuous improvement. About Databricks Databricks is the data and AI company. More than 10,000 organizations worldwide — including Comcast, Condé Nast, Grammarly, and over 50% of the Fortune 500 — rely on the Databricks Data Intelligence Platform to unify and democratize data, analytics and AI. Databricks is headquartered in San Francisco, with offices around the globe and was founded by the original creators of Lakehouse, Apache Spark™, Delta Lake and MLflow. To learn more, follow Databricks on Twitter, LinkedIn and Facebook. Benefits At Databricks, we strive to provide comprehensive benefits and perks that meet the needs of all of our employees. For specific details on the benefits offered in your region, please visit https://www.mybenefitsnow.com/databricks. Our Commitment to Diversity and Inclusion At Databricks, we are committed to fostering a diverse and inclusive culture where everyone can excel. We take great care to ensure that our hiring practices are inclusive and meet equal employment opportunity standards. Individuals looking for employment at Databricks are considered without regard to age, color, disability, ethnicity, family or marital status, gender identity or expression, language, national origin, physical and mental ability, political affiliation, race, religion, sexual orientation, socio-economic status, veteran status, and other protected characteristics. Compliance If access to export-controlled technology or source code is required for performance of job duties, it is within Employer's discretion whether to apply for a U.S. government license for such positions, and Employer may decline to proceed with an applicant on this basis alone. Show more Show less
Posted 3 weeks ago
6.0 years
0 Lacs
Bengaluru, Karnataka, India
On-site
P-1348 At Databricks, we are passionate about enabling data teams to solve the world's toughest problems, from security threat detection to cancer drug development. We do this by building and running the world's best data and AI infrastructure platform, so our customers can focus on the high value challenges that are central to their own missions. Our engineering teams build technical products that fulfill real, important needs in the world. We always push the boundaries of data and AI technology, while simultaneously operating with the resilience, security and scale that is important to making customers successful on our platform. We develop and operate one of the largest scale software platforms. The fleet consists of millions of virtual machines, generating terabytes of logs and processing exabytes of data per day. At our scale, we observe cloud hardware, network, and operating system faults, and our software must gracefully shield our customers from any of the above. As a Senior Software Engineer working on the Data Platform team you will help build the Data Intelligence Platform for Databricks that will allow us to automate decision-making across the entire company. You will achieve this in collaboration with Databricks Product Teams, Data Science, Applied AI and many more. You will develop a variety of tools spanning logging, orchestration, data transformation, metric store, governance platforms, data consumption layers etc. You will do this using the latest, bleeding-edge Databricks product and other tools in the data ecosystem - the team also functions as a large, production, in-house customer that dog foods Databricks and guides the future direction of the product. The Impact You Will Have Design and run the Databricks metrics store that enables all business units and engineering teams to bring their detailed metrics into a common platform for sharing and aggregation, with high quality, introspection ability and query performance. Design and run the cross-company Data Intelligence Platform, which contains every business and product metric used to run Databricks. You’ll play a key role in developing the right balance of data protections and ease of shareability for the Data Intelligence Platform as we transition to a public company. Develop tooling and infrastructure to efficiently manage and run Databricks on Databricks at scale, across multiple clouds, geographies and deployment types. This includes CI/CD processes, test frameworks for pipelines and data quality, and infrastructure-as-code tooling. Design the base ETL framework used by all pipelines developed at the company. Partner with our engineering teams to provide leadership in developing the long-term vision and requirements for the Databricks product. Build reliable data pipelines and solve data problems using Databricks, our partner’s products and other OSS tools. Provide early feedback on the design and operations of these products. Establish conventions and create new APIs for telemetry, debug, feature and audit event log data, and evolve them as the product and underlying services change. Represent Databricks at academic and industrial conferences & events. What We Look For 6+ years of industry experience 4+ years of experience providing technical leadership on large projects similar to the ones described above - ETL frameworks, metrics stores, infrastructure management, data security. Experience building, shipping and operating reliable multi-geo data pipelines at scale. Experience working with and operating workflow or orchestration frameworks, including open source tools like Airflow and DBT or commercial enterprise tools. Experience with large-scale messaging systems like Kafka or RabbitMQ or commercial systems. Excellent cross-functional and communication skills, consensus builder. Passion for data infrastructure and for enabling others by making their data easier to access. About Databricks Databricks is the data and AI company. More than 10,000 organizations worldwide — including Comcast, Condé Nast, Grammarly, and over 50% of the Fortune 500 — rely on the Databricks Data Intelligence Platform to unify and democratize data, analytics and AI. Databricks is headquartered in San Francisco, with offices around the globe and was founded by the original creators of Lakehouse, Apache Spark™, Delta Lake and MLflow. To learn more, follow Databricks on Twitter, LinkedIn and Facebook. Benefits At Databricks, we strive to provide comprehensive benefits and perks that meet the needs of all of our employees. For specific details on the benefits offered in your region, please visit https://www.mybenefitsnow.com/databricks. Our Commitment to Diversity and Inclusion At Databricks, we are committed to fostering a diverse and inclusive culture where everyone can excel. We take great care to ensure that our hiring practices are inclusive and meet equal employment opportunity standards. Individuals looking for employment at Databricks are considered without regard to age, color, disability, ethnicity, family or marital status, gender identity or expression, language, national origin, physical and mental ability, political affiliation, race, religion, sexual orientation, socio-economic status, veteran status, and other protected characteristics. Compliance If access to export-controlled technology or source code is required for performance of job duties, it is within Employer's discretion whether to apply for a U.S. government license for such positions, and Employer may decline to proceed with an applicant on this basis alone. Show more Show less
Posted 3 weeks ago
5.0 years
0 Lacs
Bengaluru, Karnataka, India
On-site
P-1395 Why Databricks? We’re growing fast and attracting the best talent in the world. Bricksters — as we call ourselves — are a special mix of smart, curious, quick thinkers. If you ask a Brickster what they love about working here, you’ll likely hear about our culture. We are seeking an experienced NOC Engineer to join our team. The successful candidate will be responsible for monitoring critical Databricks’ infrastructure and developing monitoring tools and alerting dashboards. They will also work closely with stakeholders to investigate and resolve incidents, perform root cause analysis, and propose solutions to increase the reliability and stability of the Databricks unified analytics platform. The Impact You Will Have Here Monitor critical infrastructure, triage alerts to proactively identify incidents, and work with stakeholders to resolve incidents. Investigate incidents and propose solutions to improve platform reliability and stability. Perform root cause analysis for recurring incidents and provide proactive solutions. Develop toolings or automate processes to improve platform monitoring and alerting. Contribute to software development efforts to improve overall service reliability and stability. Communicate effectively with internal stakeholders, including executive staff, to provide incident analysis. Participate in war rooms and temporary communication channels during outages. Demonstrate cross-functional leadership and establish ownership of incidents and outages. Multitask on several incidents and/or projects What are we looking for? Minimum of 5 years of experience as a NOC, SRE, or DevOps engineer Strong knowledge of cloud technologies such as Azure, AWS, and GCP Hands-on experience with monitoring, logging, and alerting tools such as ELK, Prometheus, Grafana, Pager Duty, etc. Experience with containers and orchestration technologies such as Docker and Kubernetes. Proficiency in automation and scripting Linux systems administration skills. Excellent communication skills. Willingness to learn Databricks products Bachelor's degree in Computer Science or a related field About Databricks Databricks is the data and AI company. More than 10,000 organizations worldwide — including Comcast, Condé Nast, Grammarly, and over 50% of the Fortune 500 — rely on the Databricks Data Intelligence Platform to unify and democratize data, analytics and AI. Databricks is headquartered in San Francisco, with offices around the globe and was founded by the original creators of Lakehouse, Apache Spark™, Delta Lake and MLflow. To learn more, follow Databricks on Twitter, LinkedIn and Facebook. Benefits At Databricks, we strive to provide comprehensive benefits and perks that meet the needs of all of our employees. For specific details on the benefits offered in your region, please visit https://www.mybenefitsnow.com/databricks. Our Commitment to Diversity and Inclusion At Databricks, we are committed to fostering a diverse and inclusive culture where everyone can excel. We take great care to ensure that our hiring practices are inclusive and meet equal employment opportunity standards. Individuals looking for employment at Databricks are considered without regard to age, color, disability, ethnicity, family or marital status, gender identity or expression, language, national origin, physical and mental ability, political affiliation, race, religion, sexual orientation, socio-economic status, veteran status, and other protected characteristics. Compliance If access to export-controlled technology or source code is required for performance of job duties, it is within Employer's discretion whether to apply for a U.S. government license for such positions, and Employer may decline to proceed with an applicant on this basis alone. Show more Show less
Posted 3 weeks ago
2.0 years
0 Lacs
Bengaluru, Karnataka, India
On-site
Mission Deliver accurate and timely responses to bricksters across the globe on people related queries. The Role As an Associate People Ops Specialist, you will play a valuable role within the People Operations team by delivering a consistent and positive employee experience throughout the employee lifecycle. You will report to the Sr. Manager, People Operations and will be responsible for providing global support to employees and managers as the first point of contact for HR related requests that could expand to other areas in the future. You will partner across the organization to address specific requests, while delivering exceptional customer service. This role requires a proactive approach to managing employee inquiries research, liaise with other teams while managing competing priorities, and a commitment to driving process efficiency continuously. Collaboration and teamwork are core values of this role. The Impact You Will Have Be the first contact for employee and manager inquiries lifecycle, ensuring timely and accurate responses to their questions and requests leveraging all available resources Assist employees with navigating internal resources, with a focus on promoting self service via Tier 0 options and BOT technology Escalate requests to other functions across the People team and reassign to different departments, with a focus on a seamless transition, clear communication and customer experience Partner with functional teams within the People group to maintain an updated matrix for escalation paths and resolver groups Support global employee lifecycle activities such as onboarding, benefits, compensation, documents, offboarding, people systems and other related processes Initiate Workday transactions including new hire data entry, basic data changes and document management, with a focus on data accuracy and efficiency Produce employee documents including offer modifications, employment verification letters, business visa letters and others while partnering with tech teams to automate similar processes Leverage data to identify opportunities for continuous process improvements to promote efficiency and scalability Create, adapt and maintain content for internal knowledge management to facilitate self service and knowledge dissemination across the People Operations team Execution of scheduled and ad-hoc tasks related to the employee lifecycle journey including several geographies and complexities Lead or co-lead projects and tasks as assigned to drive efficiency and experience for employees and managers What We Look For Bachelor's Degree or equivalent experience required 2+ year of HR shared services or generalist experience Experience working with various lifecycle processes across multiple regions (AMER, EMEA, APJ) Provide exceptional customer service Success with managing a high volume of work Adaptability to the fast paced environment of a growing company Analytical and detail oriented mindset able to spot trends, find causes and drive actions Active participant of continuous improvement initiatives Workday experience is a most, advanced knowledge of the tool is ideal Experience with automation projects for HR support in different platforms is highly desired Knowledge of AI basics, tools and process execution to deploy such solutions in an HR environment is a desired skill that will provide advantage as a selection parameter Continuous improvement methodology knowledge or certification is a plus Strong teamwork skills Solid communication skills written and verbal to interact with employees, partner teams and vendors About Databricks Databricks is the data and AI company. More than 10,000 organizations worldwide — including Comcast, Condé Nast, Grammarly, and over 50% of the Fortune 500 — rely on the Databricks Data Intelligence Platform to unify and democratize data, analytics and AI. Databricks is headquartered in San Francisco, with offices around the globe and was founded by the original creators of Lakehouse, Apache Spark™, Delta Lake and MLflow. To learn more, follow Databricks on Twitter, LinkedIn and Facebook. Benefits At Databricks, we strive to provide comprehensive benefits and perks that meet the needs of all of our employees. For specific details on the benefits offered in your region, please visit https://www.mybenefitsnow.com/databricks. Our Commitment to Diversity and Inclusion At Databricks, we are committed to fostering a diverse and inclusive culture where everyone can excel. We take great care to ensure that our hiring practices are inclusive and meet equal employment opportunity standards. Individuals looking for employment at Databricks are considered without regard to age, color, disability, ethnicity, family or marital status, gender identity or expression, language, national origin, physical and mental ability, political affiliation, race, religion, sexual orientation, socio-economic status, veteran status, and other protected characteristics. Compliance If access to export-controlled technology or source code is required for performance of job duties, it is within Employer's discretion whether to apply for a U.S. government license for such positions, and Employer may decline to proceed with an applicant on this basis alone. Show more Show less
Posted 3 weeks ago
6.0 years
0 Lacs
Bengaluru, Karnataka, India
Remote
P-926 At Databricks, we are passionate about enabling data teams to solve the world's toughest problems — from making the next mode of transportation a reality to accelerating the development of medical breakthroughs. We do this by building and running the world's best data and AI infrastructure platform so our customers can use deep data insights to improve their business. Founded by engineers — and customer obsessed — we leap at every opportunity to solve technical challenges, from designing next-gen UI/UX for interfacing with data to scaling our services and infrastructure across millions of virtual machines. And we're only getting started in Bengaluru , India ! As a software engineer with a backend focus, you will work with your team to build infrastructure for the Databricks platform at scale. The Impact You'll Have Our backend teams cover a diverse range of domains, from core compute fabric resource management to service platforms and infrastructure. For example, you might work on challenges such as: Supporting Databricks' growth by building foundational infrastructure platforms that enable seamless operation across numerous geographic regions and cloud providers. Implementing cloud-agnostic infrastructure abstractions to help Databricks engineers more efficiently manage and operate their services. Develop tools and processes that drive engineering efficiency at Databricks. We enhance the developer experience for Databricks engineers across various areas, including programming languages, linters, static analysis, IDEs, remote development environments, automated release pipelines, and test automation frameworks. Our current focus is on optimizing the Rust development experience across the organization. What We Look For BS (or higher) in Computer Science, or a related field 6+ years of production level experience in one of: Python, Java, Scala, C++, or similar language. Experience developing large-scale distributed systems from scratch Experience working on a SaaS platform or with Service-Oriented Architectures Proficiency in one or more backend languages such as Java, Scala, or Go. Hands-on experience in developing and operating backend systems. Ability to contribute effectively throughout all project phases, from initial design and development to implementation and ongoing operations, with guidance from senior team members. About Databricks Databricks is the data and AI company. More than 10,000 organizations worldwide — including Comcast, Condé Nast, Grammarly, and over 50% of the Fortune 500 — rely on the Databricks Data Intelligence Platform to unify and democratize data, analytics and AI. Databricks is headquartered in San Francisco, with offices around the globe and was founded by the original creators of Lakehouse, Apache Spark™, Delta Lake and MLflow. To learn more, follow Databricks on Twitter, LinkedIn and Facebook. Benefits At Databricks, we strive to provide comprehensive benefits and perks that meet the needs of all of our employees. For specific details on the benefits offered in your region, please visit https://www.mybenefitsnow.com/databricks. Our Commitment to Diversity and Inclusion At Databricks, we are committed to fostering a diverse and inclusive culture where everyone can excel. We take great care to ensure that our hiring practices are inclusive and meet equal employment opportunity standards. Individuals looking for employment at Databricks are considered without regard to age, color, disability, ethnicity, family or marital status, gender identity or expression, language, national origin, physical and mental ability, political affiliation, race, religion, sexual orientation, socio-economic status, veteran status, and other protected characteristics. Compliance If access to export-controlled technology or source code is required for performance of job duties, it is within Employer's discretion whether to apply for a U.S. government license for such positions, and Employer may decline to proceed with an applicant on this basis alone. Show more Show less
Posted 3 weeks ago
8.0 years
0 Lacs
Pune, Maharashtra, India
On-site
Job Requisition ID # 25WD85491 Position Overview We are looking for an experienced Principal Software Engineer to join our platform team focusing on AI/ML Platform (AMP). This team builds and maintains central components to fast track the development of new ML/AI models such as model development studio, feature store, model serving and model observability. The ideal candidate would have a background in ML Ops, Data engineering and DevOps with the experience of building high scale deployment architectures and observability. As an important contributor to our engineering team, you will help shape the future of our AI/ML capabilities, delivering solutions that inspire value for our organization. You will report directly to an Engineering Manager, and you will be based in Pune. Responsibilities System design: You will design, implement and manage software systems for the AI/ML Platform and orchestrate the full ML development lifecycle for the partner teams Mentoring: Spreading your knowledge, sharing best practices and doing design reviews to step up the expertise at the team level Multi-cloud architecture: Define components which leverages strengths from multiple cloud platforms (e.g., AWS, Azure) to optimize performance, cost, and scalability AI/ML observability: You will build systems for monitoring performance of AI/ML models and find insights on the underlying data such as drift detection, data fairness/bias and anomalies ML Solution Deployment: You will develop tools for building and deploying ML artefacts in production environments and facilitating a smooth transition from development to deployment Big Data Management: Automate and orchestrate tasks related to managing big data transformation and processing and build large-scale data stores for ML artifacts Scalable Services: Design and implement low-latency, scalable prediction, and inference services to support the diverse needs of our users Cross-Functional Collaboration: Collaborate across diverse teams, including machine learning researchers, developers, product managers, software architects, and operations, fostering a collaborative and cohesive work environment End-to-end ownership: You will take the end-to-end ownership of the components and work with other engineers in the team including design, architecture, implementation, rollout and onboarding support to partner teams, production on-call support, testing/verification, investigations etc Minimum Qualifications Educational Background: Bachelor’s degree in Computer Science or equivalent practical experience Experience: Over 8 years of experience in software development and engineering, delivering production systems and services Prior experience of working with MLOps team at the intersection of the expertise across ML model deployments, DevOps and data engineering Hands-on skills: Ability to fluently translate the design into high quality code in golang, python, Java Knowledge of DevOps practices, containerization, orchestration tools such as CI/CD, Terraform, Docker, Kubernetes, Gitops Demonstrate knowledge of distributed data processing frameworks, orchestrators, and data lake architectures using technologies such as Spark, Airflow, iceberg/ parquet formats Prior collaborations with Data science teams to deploy their models, setting up ML observability for inference level monitoring Exposure for building RAG based applications by collaborating with other product teams, Data scientists/AI engineers Demonstrate creative problem-solving skills with the ability to break down problems into manageable components Knowledge of Amazon AWS and/or Azure cloud for solutioning large scale application deployments Excellent communication and collaboration skills, fostering teamwork and effective information exchange Preferred Qualifications Experience in integrating with third party vendors Experience in latency optimization with the ability to diagnose, tune, and enhance the efficiency of serving systems Familiarity with tools and frameworks for monitoring and managing the performance of AI/ML models in production (e.g., MLflow, Kubeflow, TensorBoard) Familiarity with distributed model training/inference pipelines using (KubeRay or equivalent) Exposure to leveraging GPU computing for AI/ML workloads, including experience with CUDA, OpenCL, or other GPU programming tools, to significantly enhance model training and inference performance Exposure to ML libraries such as PyTorch, TensorFlow, XGBoost, Pandas, and ScikitLearn Learn More About Autodesk Welcome to Autodesk! Amazing things are created every day with our software – from the greenest buildings and cleanest cars to the smartest factories and biggest hit movies. We help innovators turn their ideas into reality, transforming not only how things are made, but what can be made. We take great pride in our culture here at Autodesk – our Culture Code is at the core of everything we do. Our values and ways of working help our people thrive and realize their potential, which leads to even better outcomes for our customers. When you’re an Autodesker, you can be your whole, authentic self and do meaningful work that helps build a better future for all. Ready to shape the world and your future? Join us! Salary transparency Salary is one part of Autodesk’s competitive compensation package. Offers are based on the candidate’s experience and geographic location. In addition to base salaries, we also have a significant emphasis on discretionary annual cash bonuses, commissions for sales roles, stock or long-term incentive cash grants, and a comprehensive benefits package. Diversity & Belonging We take pride in cultivating a culture of belonging and an equitable workplace where everyone can thrive. Learn more here: https://www.autodesk.com/company/diversity-and-belonging Are you an existing contractor or consultant with Autodesk? Please search for open jobs and apply internally (not on this external site). Show more Show less
Posted 3 weeks ago
Upload Resume
Drag or click to upload
Your data is secure with us, protected by advanced encryption.
The mlflow job market in India is rapidly growing as companies across various industries are increasingly adopting machine learning and data science technologies. mlflow, an open-source platform for the machine learning lifecycle, is in high demand in the Indian job market. Job seekers with expertise in mlflow have a plethora of opportunities to explore and build a rewarding career in this field.
These cities are known for their thriving tech industries and have a high demand for mlflow professionals.
The average salary range for mlflow professionals in India varies based on experience: - Entry-level: INR 6-8 lakhs per annum - Mid-level: INR 10-15 lakhs per annum - Experienced: INR 18-25 lakhs per annum
Salaries may vary based on factors such as location, company size, and specific job requirements.
A typical career path in mlflow may include roles such as: 1. Junior Machine Learning Engineer 2. Machine Learning Engineer 3. Senior Machine Learning Engineer 4. Tech Lead 5. Machine Learning Manager
With experience and expertise, professionals can progress to higher roles and take on more challenging projects in the field of machine learning.
In addition to mlflow, professionals in this field are often expected to have skills in: - Python programming - Data visualization - Statistical modeling - Deep learning frameworks (e.g., TensorFlow, PyTorch) - Cloud computing platforms (e.g., AWS, Azure)
Having a strong foundation in these related skills can further enhance a candidate's profile and career prospects.
As you explore opportunities in the mlflow job market in India, remember to continuously upskill, stay updated with the latest trends in machine learning, and showcase your expertise confidently during interviews. With dedication and perseverance, you can build a successful career in this dynamic and rapidly evolving field. Good luck!
Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.
We have sent an OTP to your contact. Please enter it below to verify.
Accenture
36723 Jobs | Dublin
Wipro
11788 Jobs | Bengaluru
EY
8277 Jobs | London
IBM
6362 Jobs | Armonk
Amazon
6322 Jobs | Seattle,WA
Oracle
5543 Jobs | Redwood City
Capgemini
5131 Jobs | Paris,France
Uplers
4724 Jobs | Ahmedabad
Infosys
4329 Jobs | Bangalore,Karnataka
Accenture in India
4290 Jobs | Dublin 2