Jobs
Interviews

298 Argocd Jobs

Setup a job Alert
JobPe aggregates results for easy application access, but you actually apply on the job portal directly.

7.0 - 12.0 years

12 - 24 Lacs

gurugram

Work from Office

Kubernetes expert with GitOps (ArgoCD, FluxCD), CI/CD (Jenkins, GitLab), IaC (Terraform, Crossplane), observability (Prometheus, Grafana, ELK), service mesh (Istio, Linkerd), security (RBAC, OPA, Kyverno), and scripting (Bash, Python, Go, Ansible).

Posted 9 hours ago

Apply

5.0 - 10.0 years

7 - 17 Lacs

mumbai

Work from Office

An understanding of product development methodologies and microservices architecture. Hands-on experience with at least two major cloud providers (AWS, GCP, Azure). Multi-cloud experience is a strong advantage. Expertise in designing, implementing, and managing cloud architectures focusing on scalability, security, and resilience. Understanding and experience with cloud fundamentals like Networking, IAM, Compute, and Managed Services like DB, Storage, GKE/EKS, and KMS. Hands-on experience with cloud architecture design & setup. An in-depth understanding of Infrastructure as Code tools like Terraform, HELM is a must. Practical experience in deploying, maintaining, and scaling applications on Kubernetes clusters using Helm Charts or Kustomize Hands-on experience with any CI/CD tools like Gitlab CI, Jenkins, Github Actions. GitOps tools like ArgoCD, FluxCD is a must. Experience with Monitoring and Logging tools like Prometheus, Grafana and Elastic Stack. Experience working with PaaS is a plus Experience deploying on-prem data centre. Experience with k3s OSS / OpenShift / Rancher Kubernetes Cluster is a plus What are we looking for Learn, Architect & Build Skills & Technologies as highlighted above Product-Oriented Delivery Design, Build, and Operate Cloud Architecture & DevOps Pipeline Build on Open Source Technologies Collaboration with teams across 5 products GitOps Philosophy DevSecOps Mindset - Highly Secure Platform

Posted 12 hours ago

Apply

15.0 - 17.0 years

0 Lacs

noida, uttar pradesh, india

Remote

Req ID: 340251 NTT DATA strives to hire exceptional, innovative and passionate individuals who want to grow with us. If you want to be part of an inclusive, adaptable, and forward-thinking organization, apply now. We are currently seeking a Industry Consulting Manager to join our team in Noida, Uttar Pradesh (IN-UP), India (IN). Job Description: Technical Architect Observability & SRE Frameworks Position Title: Technical Architect Observability & Site Reliability Engineering (SRE) Location: Noida, India Experience: 15+ years (including 5+ years in observability/SRE architecture) Employment Type: Full-time Role Overview We are looking for a highly experienced Technical Architect to lead the design, strategy, and implementation of Observability and SRE frameworks for enterprise-scale, microservices-based applications. The ideal candidate will bring deep technical knowledge of both Splunk Observability Stack and Open Source tools (like OpenTelemetry, Prometheus, Grafana, Jaeger), and be capable of defining and executing architecture strategies for complex distributed systems. This role requires hands-on ability to create architecture blueprints , lead technical teams, and work directly with stakeholders and platform owners to embed observability and reliability practices across the SDLC. Key Responsibilities Architecture & Blueprinting Design and deliver end-to-end observability architecture (Metrics, Logs, Traces, Events) for cloud-native and hybrid environments. Create technical architecture diagrams, data flow maps, and integration blueprints using tools like Lucidchart, Draw.io, or Visio. Lead the definition of SLIs, SLOs, and Error Budgets aligned with business KPIs and DORA metrics. Toolchain Strategy & Implementation Architect telemetry pipelines using OpenTelemetry Collector and Splunk Observability Cloud (SignalFx, APM, RUM, Log Observer). Define tool adoption strategy and integration roadmap for OSS tools (Prometheus, Loki, Grafana, Jaeger) and Splunk-based stacks. Guide teams on instrumentation approaches (auto/manual) across languages like Java, Go, Python, .NET, etc. Reliability Engineering Enablement Lead adoption of SRE principles including incident management frameworks, resiliency testing, and runbook automation. Collaborate with DevOps to integrate observability into CI/CD pipelines (e.g., Jenkins, ArgoCD, GitHub Actions). Define health checks, golden signals, and SPoG (Single Pane of Glass) dashboards. Exposure to AIOps, ML-based anomaly detection, or business observability. Stakeholder Management & Governance Serve as a technical liaison between client leadership, SREs, developers, and infrastructure teams. Run workshops, assessments, and evangelize observability-first culture across teams. Provide guidance on data retention, access control, cost optimization, and compliance (especially with Splunk ingestion policies). Performance & Optimization Continuously monitor and fine-tune observability data flows to prevent alert fatigue and ensure actionability. Implement root cause analysis practices using telemetry correlation across metrics, logs, and traces. Lead efforts to build self-healing systems using automated playbooks and AIOps integrations (where applicable). Required Skills & Qualifications 15+ years in IT, with 5 years in Observability/SRE architecture roles Proven experience designing architecture for microservices, containers (Docker, Kubernetes), and distributed systems Strong hands-on expertise with: Splunk Observability Cloud (SignalFx, Log Observer, APM) OpenTelemetry (SDKs + Collector) Prometheus + Grafana Jaeger / Zipkin for distributed tracing CI/CD tools: Jenkins, GitHub Actions, ArgoCD Ability to build and present clear architecture diagrams and solution roadmaps Working knowledge of cloud environments (AWS, Azure, GCP) and container orchestration (K8s/OpenShift) Familiarity with SRE and DevOps best practices (error budgets, release engineering, chaos testing) Nice to Have Splunk certifications: Core Consultant, Observability Specialist, Admin Knowledge of ITIL and modern incident management frameworks (PagerDuty, OpsGenie) Experience in banking or regulated enterprise environments Soft Skills Strong leadership and cross-functional collaboration Ability to work in ambiguous, fast-paced environments Excellent documentation and communication skills Passion for mentoring teams and building best practices at scale Why This Role Matters The client is on a journey to mature its Observability and SRE ecosystem , and this role will be critical in: Unifying legacy and modern telemetry stacks Driving reliability-first mindset and tooling Establishing a scalable blueprint for production excellence About NTT DATA NTT DATA is a $30 billion trusted global innovator of business and technology services. We serve 75% of the Fortune Global 100 and are committed to helping clients innovate, optimize and transform for long term success. As a Global Top Employer, we have diverse experts in more than 50 countries and a robust partner ecosystem of established and start-up companies. Our services include business and technology consulting, data and artificial intelligence, industry solutions, as well as the development, implementation and management of applications, infrastructure and connectivity. We are one of the leading providers of digital and AI infrastructure in the world. NTT DATA is a part of NTT Group, which invests over $3.6 billion each year in R&D to help organizations and society move confidently and sustainably into the digital future. Visit us at us.nttdata.com Whenever possible, we hire locally to NTT DATA offices or client sites. This ensures we can provide timely and effective support tailored to each clients needs. While many positions offer remote or hybrid work options, these arrangements are subject to change based on client requirements. For employees near an NTT DATA office or client site, in-office attendance may be required for meetings or events, depending on business needs. At NTT DATA, we are committed to staying flexible and meeting the evolving needs of both our clients and employees. NTT DATA recruiters will never ask for payment or banking information and will only use @nttdata.com and @talent.nttdataservices.com email addresses. If you are requested to provide payment or disclose banking information, please submit a contact us form, https://us.nttdata.com/en/contact-us . NTT DATA endeavors to make https://us.nttdata.com accessible to any and all users. If you would like to contact us regarding the accessibility of our website or need assistance completing the application process, please contact us at https://us.nttdata.com/en/contact-us . This contact information is for accommodation requests only and cannot be used to inquire about the status of applications. NTT DATA is an equal opportunity employer. Qualified applicants will receive consideration for employment without regard to race, color, religion, sex, sexual orientation, gender identity, national origin, disability or protected veteran status. For our EEO Policy Statement, please click here . If you&aposd like more information on your EEO rights under the law, please click here . For Pay Transparency information, please click here . Show more Show less

Posted 1 day ago

Apply

5.0 - 10.0 years

0 - 1 Lacs

hyderabad

Work from Office

Golang Developer JD: 5 to 10 years of experience Strong knowledge in scripting: Golang Hands-on experience in Containers, Docker, Ansible, Kubernetes, K8s operators, Go Services and good understanding of Cloud-native principles Good Troubleshoot and resolve issues related to Kubernetes, containerized applications, and Go services. Proficient in GitOps Tools FluxCD, ArgoCD Expertise in OCI-compliant registries, including Harbor and Quay, to manage and secure our container images Install, configure, and maintain Git servers (e.g., GitLab, GitHub, Gitea) Familiar with setting up and maintaining repositories, branching strategies Implement and manage user access controls, permissions, and authentication

Posted 3 days ago

Apply

12.0 - 18.0 years

35 - 45 Lacs

hyderabad

Hybrid

About TechBlocks: TechBlocks is a global digital product engineering company with 16+ years of experience helping Fortune 500 enterprises and high-growth brands accelerate innovation, modernize technology, and drive digital transformation. From cloud solutions and data engineering to experience design and platform modernization, we help businesses solve complex challenges and unlock new growth opportunities. Job Title: Site Reliability Engineering (SRE) Manager Location : Hyderabad Employment Type: Full-Time Work Model - 3 Days from office (Hybrid) Summary: The SRE Manager at TechBlocks India will lead the reliability engineering function, ensuring infrastructure resiliency and optimal operational performance. This hybrid role blends technical leadership with team mentorship and cross-functional coordination. Experience Required: 12+ years total experience, with 3+ years in a leadership role in SRE or Cloud Operations. Technical Knowledge and Skills: Mandatory: Deep understanding of Kubernetes, GKE, Prometheus, Terraform Cloud: Advanced GCP administration CI/CD: Jenkins, Argo CD, GitHub Actions Incident Management: Full lifecycle, tools like OpsGenie Nice to Have : Knowledge of service mesh and observability stacks Strong scripting skills (Python, Bash) Big Query /Dataflow exposure for telemetry Scope: Build and lead a team of SREs Standardize practices for reliability, alerting, and response Engage with Engineering and Product leaders Roles and Responsibilities: Establish and lead the implementation of organizational reliability strategies, aligning SLAs, SLOs, and Error Budgets with business goals and customer expectations. Develop and institutionalize incident response frameworks, including escalation policies, on-call scheduling, service ownership mapping, and RCA process governance. Lead technical reviews for infrastructure reliability design, high-availability architectures, and resiliency patterns across distributed cloud services. Champion observability and monitoring culture by standardizing tooling, alert definitions, dashboard templates, and telemetry data schemas across all product teams. Drive continuous improvement through operational maturity assessments, toil elimination initiatives, and SRE OKRs aligned with product objectives. Collaborate with cloud engineering and platform teams to introduce self-healing systems, capacity-aware autoscaling, and latency-optimized service mesh patterns. Act as the principal escalation point for reliability-related concerns and ensure incident retrospectives lead to measurable improvements in uptime and MTTR. Own runbook standardization, capacity planning, failure mode analysis, and production readiness reviews for new feature launches. Mentor and develop a high-performing SRE team, fostering a proactive ownership culture, encouraging cross-functional knowledge sharing, and establishing technical career pathways. Collaborate with leadership, delivery, and customer stakeholders to define reliability goals, track performance, and demonstrate ROI on SRE investments

Posted 4 days ago

Apply

5.0 - 7.0 years

17 - 18 Lacs

mumbai

Work from Office

An understanding of product development methodologies and microservices architecture. Hands-on experience with at least two major cloud providers (AWS, GCP, Azure). Multi-cloud experience is a strong advantage. Expertise in designing, implementing, and managing cloud architectures focusing on scalability, security, and resilience. Understanding and experience with cloud fundamentals like Networking, IAM, Compute, and Managed Services like DB, Storage, GKE/EKS, and KMS. Hands-on experience with cloud architecture design & setup. An in-depth understanding of Infrastructure as Code tools like Terraform, HELM is a must. Practical experience in deploying, maintaining, and scaling applications on Kubernetes clusters using Helm Charts or Kustomize Hands-on experience with any CI/CD tools like Gitlab CI, Jenkins, Github Actions. GitOps tools like ArgoCD, FluxCD is a must. Experience with Monitoring and Logging tools like Prometheus, Grafana and Elastic Stack. Experience working with PaaS is a plus Experience deploying on-prem data centre. Experience with k3s OSS / OpenShift / Rancher Kubernetes Cluster is a plus What are we looking for Learn, Architect & Build Skills & Technologies as highlighted above Product-Oriented Delivery Design, Build, and Operate Cloud Architecture & DevOps Pipeline Build on Open Source Technologies Collaboration with teams across 5 products GitOps Philosophy DevSecOps Mindset - Highly Secure Platform

Posted 4 days ago

Apply

3.0 - 12.0 years

0 Lacs

hyderabad, telangana

On-site

As a Java Technical Architect at GlobalLogic, you will be responsible for defining and owning the end-to-end architecture for enterprise-grade Java applications. Your role will involve leading the design and development of microservices-based, cloud-native applications using Java, Spring Boot, and related frameworks. You will evaluate and select appropriate design patterns, frameworks, and architectural strategies to promote modularity, reusability, and performance. Additionally, you will conduct technical assessments, code reviews, and proof-of-concepts (PoCs) to validate architectural decisions and implementation approaches. Key Responsibilities: - Define and own the end-to-end architecture for enterprise-grade Java applications - Lead the design and development of microservices-based, cloud-native applications - Evaluate and select appropriate design patterns, frameworks, and architectural strategies - Conduct technical assessments, code reviews, and proof-of-concepts (PoCs) - Guide the modernization of legacy Java systems - Collaborate with DevOps and cloud engineering teams - Define API strategies including RESTful service design - Support capacity planning, performance tuning, and disaster recovery strategies - Mentor and coach developers and technical leads - Stay current with emerging trends in Java ecosystem and cloud computing Qualifications: - Bachelor's or Master's degree in Computer Science, Engineering, or a related technical discipline - 12+ years of software development experience, with 3-5 years in a Java architecture or technical leadership role - Deep expertise in Java/J2EE, Spring Framework, and microservices architecture - Proficient in RESTful API design, API management tools, and cloud platforms - Knowledge of relational and NoSQL databases, build tools, version control systems, and CI/CD pipelines - Experience with enterprise integration patterns, event-driven architecture, and service meshes - Strong analytical and problem-solving skills - Excellent communication and interpersonal skills GlobalLogic offers a culture of caring, learning and development opportunities, interesting and meaningful work, balance and flexibility, and a high-trust organization. GlobalLogic, a Hitachi Group Company, is a trusted digital engineering partner known for creating innovative digital products and experiences since 2000.,

Posted 4 days ago

Apply

6.0 - 8.0 years

0 Lacs

bengaluru, karnataka, india

On-site

Title: Senior DevOps Engineer Reporting to: Senior Director, Product Development Location: Bengaluru (Bangalore) Opportunity: Responsibilities: Infrastructure Development & Integration Design, implement, and manage cloud-native infrastructure (AWS, Azure, GCP) to support healthcare platforms, AI agents, and clinical applications. Build and maintain scalable CI/CD pipelines to enable rapid and reliable delivery of software, data pipelines, and AI/ML models. Design and manage Kubernetes (K8s) clusters for container orchestration, workload scaling, and high availability with integrated monitoring to ensure cluster health and performance Implement Kubernetes-native tools (Helm, Kustomize, ArgoCD) for deployment automation and environment management ensuring observability through monitoring dashboards and alerts Collaborate with Staff Engineers/Architects to align infrastructure with enterprise goals for scalability, reliability, and performance leveraging monitoring insights to inform architectural decisions. System Optimization & Reliability Implement and maintain comprehensive monitoring, logging, and alerting mechanisms (Prometheus, Grafana, ELK, Datadog, AWS cloudwatch, AWS cloud trail) to ensure real-time visibility into system performance, resource utilization, and potential incidents. Implement monitoring, logging, and alerting mechanisms (Prometheus, Grafana, ELK, Datadog) to ensure system reliability and proactive incident response. Ensure data pipeline workflows (ETL/ELT, real-time streaming, batch processing) are observable, reliable, and auditable. Support observability and monitoring of GenAI pipelines, embeddings, vector databases, and agentic AI workflows. Proactively analyze monitoring data to identify bottlenecks, predict failures, and drive continuous improvement in system reliability. Compliance & Security Support audit trails and compliance reporting through automated DevSecOps practices. Implement security controls for LLM-based applications, AI agents, and healthcare data pipelines, including prompt injection prevention, API rate limiting, and data governance. Collaboration & Agile Practices Partner closely with software engineers, data engineers, AI/ML engineers, and product managers to deliver integrated, secure, and scalable solutions. Contribute to agile development processes including sprint planning, stand-ups, and retrospectives. Mentor junior engineers and share best practices in cloud-native infrastructure, CI/CD, Kubernetes, and automation. Innovation & Technical Expertise Stay informed about emerging DevOps practices, cloud-native architectures, MLOps/LLMOps, and data engineering tools. Prototype and evaluate new frameworks and tools to enhance infrastructure for data pipelines, GenAI, and Agentic AI applications. Advocate for best practices in infrastructure design, focusing on modularity, maintainability, and scalability. Requirements Education & Experience Bachelor&aposs or Master&aposs degree in Computer Science, Engineering, or related technical discipline. 6+ years of experience in DevOps, Site Reliability Engineering, or related roles, with at least 5+ years building cloud-native infrastructure. Proven track record of managing production-grade Kubernetes clusters and cloud infrastructure in regulated environments. Experience supporting GenAI/LLM applications (e.g., OpenAI, Hugging Face, LangChain) and vector databases (e.g., Pinecone, Weaviate, FAISS). Hands-on experience supporting data pipeline products using ETL/ELT frameworks (Apache Airflow, dbt, Prefect) and streaming systems (Kafka, Spark, Flink). Experience deploying AI agents and orchestrating agent workflows in production environments. Technical Proficiency Expertise in Kubernetes (K8s) for orchestration, scaling, and managing containerized applications. Strong proficiency in containerization (Docker) and Kubernetes ecosystem tools (Helm, ArgoCD, Istio/Linkerd for service mesh). Hands-on experience with Infrastructure as Code (Terraform, CloudFormation, or Pulumi). Proficiency with CI/CD tools (Jenkins, GitHub Actions, GitLab CI, ArgoCD, Spinnaker). Familiarity with monitoring and observability tools (Prometheus, Grafana, ELK, Datadog, AWS cloud watch and AWS cloud trail), including setting up dashboards, alerts, and custom metrics for cloud-native and AI systems. Good to have: knowledge of healthcare data standards (FHIR, HL7) and secure deployment practices for AI/ML and data pipelines. Professional Skills Strong problem-solving skills with a focus on reliability, scalability, and security. Excellent collaboration and communication skills across cross-functional teams. Proactive, detail-oriented, and committed to technical excellence in a fast-paced healthcare environment. About Get Well: Now part of the SAI Group family, Get Well is redefining digital patient engagement by putting patients in control of their personalized healthcare journeys, both inside and outside the hospital. Get Well is combining high-tech AI navigation with high-touch care experiences driving patient activation, loyalty, and outcomes while reducing the cost of care. For almost 25 years, Get Well has served more than 10 million patients per year across over 1,000 hospitals and clinical partner sites, working to use longitudinal data analytics to better serve patients and clinicians. AI innovator SAI Group led by Chairman Romesh Wadhwani is the lead growth investor in Get Well. Get Well&aposs award-winning solutions were recognized again in 2024 by KLAS Research and AVIA Marketplace. Learn more at Get Well and follow-us on LinkedIn?and Twitter. Get Well is proud to be an equal opportunity employer. All qualified applicants will receive consideration for employment without regard to race, color, religion, gender, gender identity or expression, sexual orientation, national origin, genetics, disability, age or veteran status. About SAI Group: SAIGroup commits to $1 Billion capital, an advanced AI platform that currently processes 300M+ patients, and 4000+ global employee base to solve enterprise AI and high priority healthcare problems. SAIGroup - Growing companies with advanced AI; https://www.cnbc.com/2023/12/08/75-year-old-tech-mogul-betting-1-billion-of-his-fortune-on-ai-future.html Bio of our Chairman Dr. Romesh Wadhwani: Team - SAIGroup (Informal at Romesh Wadhwani - Wikipedia) TIME Magazine recently recognized Chairman Romesh Wadhwani as one of the Top 100 AI leaders in the world - Romesh and Sunil Wadhwani: The 100 Most Influential People in AI 2023 | TIME Show more Show less

Posted 5 days ago

Apply

5.0 - 7.0 years

0 Lacs

bengaluru, karnataka, india

On-site

Title: DevOps Engineer Reporting to: Senior Director, Product Development Location: Bengaluru (Bangalore) Opportunity: Responsibilities: Infrastructure Development & Integration Design, implement, and manage cloud-native infrastructure (AWS, Azure, GCP) to support healthcare platforms, AI agents, and clinical applications. Build and maintain scalable CI/CD pipelines to enable rapid and reliable delivery of software, data pipelines, and AI/ML models. Design and manage Kubernetes (K8s) clusters for container orchestration, workload scaling, and high availability with integrated monitoring to ensure cluster health and performance Implement Kubernetes-native tools (Helm, Kustomize, ArgoCD) for deployment automation and environment management ensuring observability through monitoring dashboards and alerts Collaborate with Staff Engineers/Architects to align infrastructure with enterprise goals for scalability, reliability, and performance leveraging monitoring insights to inform architectural decisions. System Optimization & Reliability Implement and maintain comprehensive monitoring, logging, and alerting mechanisms (Prometheus, Grafana, ELK, Datadog, AWS cloudwatch, AWS cloud trail) to ensure real-time visibility into system performance, resource utilization, and potential incidents. Implement monitoring, logging, and alerting mechanisms (Prometheus, Grafana, ELK, Datadog) to ensure system reliability and proactive incident response. Ensure data pipeline workflows (ETL/ELT, real-time streaming, batch processing) are observable, reliable, and auditable. Support observability and monitoring of GenAI pipelines, embeddings, vector databases, and agentic AI workflows. Proactively analyze monitoring data to identify bottlenecks, predict failures, and drive continuous improvement in system reliability. Compliance & Security Support audit trails and compliance reporting through automated DevSecOps practices. Implement security controls for LLM-based applications, AI agents, and healthcare data pipelines, including prompt injection prevention, API rate limiting, and data governance. Collaboration & Agile Practices Partner closely with software engineers, data engineers, AI/ML engineers, and product managers to deliver integrated, secure, and scalable solutions. Contribute to agile development processes including sprint planning, stand-ups, and retrospectives. Mentor junior engineers and share best practices in cloud-native infrastructure, CI/CD, Kubernetes, and automation. Innovation & Technical Expertise Stay informed about emerging DevOps practices, cloud-native architectures, MLOps/LLMOps, and data engineering tools. Prototype and evaluate new frameworks and tools to enhance infrastructure for data pipelines, GenAI, and Agentic AI applications. Advocate for best practices in infrastructure design, focusing on modularity, maintainability, and scalability. Requirements Education & Experience Bachelor&aposs or Master&aposs degree in Computer Science, Engineering, or related technical discipline. 5+ years of experience in DevOps, Site Reliability Engineering, or related roles, with at least 3+ years building cloud-native infrastructure. Proven track record of managing production-grade Kubernetes clusters and cloud infrastructure in regulated environments. Experience supporting GenAI/LLM applications (e.g., OpenAI, Hugging Face, LangChain) and vector databases (e.g., Pinecone, Weaviate, FAISS). Hands-on experience supporting data pipeline products using ETL/ELT frameworks (Apache Airflow, dbt, Prefect) and streaming systems (Kafka, Spark, Flink). Experience deploying AI agents and orchestrating agent workflows in production environments. Technical Proficiency Expertise in Kubernetes (K8s) for orchestration, scaling, and managing containerized applications. Strong proficiency in containerization (Docker) and Kubernetes ecosystem tools (Helm, ArgoCD, Istio/Linkerd for service mesh). Hands-on experience with Infrastructure as Code (Terraform, CloudFormation, or Pulumi). Proficiency with CI/CD tools (Jenkins, GitHub Actions, GitLab CI, ArgoCD, Spinnaker). Familiarity with monitoring and observability tools (Prometheus, Grafana, ELK, Datadog, AWS cloud watch and AWS cloud trail), including setting up dashboards, alerts, and custom metrics for cloud-native and AI systems. Good to have: knowledge of healthcare data standards (FHIR, HL7) and secure deployment practices for AI/ML and data pipelines. Professional Skills Strong problem-solving skills with a focus on reliability, scalability, and security. Excellent collaboration and communication skills across cross-functional teams. Proactive, detail-oriented, and committed to technical excellence in a fast-paced healthcare environment. About Get Well: Now part of the SAI Group family, Get Well is redefining digital patient engagement by putting patients in control of their personalized healthcare journeys, both inside and outside the hospital. Get Well is combining high-tech AI navigation with high-touch care experiences driving patient activation, loyalty, and outcomes while reducing the cost of care. For almost 25 years, Get Well has served more than 10 million patients per year across over 1,000 hospitals and clinical partner sites, working to use longitudinal data analytics to better serve patients and clinicians. AI innovator SAI Group led by Chairman Romesh Wadhwani is the lead growth investor in Get Well. Get Well&aposs award-winning solutions were recognized again in 2024 by KLAS Research and AVIA Marketplace. Learn more at Get Well and follow-us on LinkedIn?and Twitter. Get Well is proud to be an equal opportunity employer. All qualified applicants will receive consideration for employment without regard to race, color, religion, gender, gender identity or expression, sexual orientation, national origin, genetics, disability, age or veteran status. About SAI Group: SAIGroup commits to $1 Billion capital, an advanced AI platform that currently processes 300M+ patients, and 4000+ global employee base to solve enterprise AI and high priority healthcare problems. SAIGroup - Growing companies with advanced AI; https://www.cnbc.com/2023/12/08/75-year-old-tech-mogul-betting-1-billion-of-his-fortune-on-ai-future.html Bio of our Chairman Dr. Romesh Wadhwani: Team - SAIGroup (Informal at Romesh Wadhwani - Wikipedia) TIME Magazine recently recognized Chairman Romesh Wadhwani as one of the Top 100 AI leaders in the world - Romesh and Sunil Wadhwani: The 100 Most Influential People in AI 2023 | TIME Show more Show less

Posted 5 days ago

Apply

4.0 - 8.0 years

0 Lacs

bengaluru, karnataka, india

On-site

Airbnb was born in 2007 when two hosts welcomed three guests to their San Francisco home, and has since grown to over 5 million hosts who have welcomed over 2 billion guest arrivals in almost every country across the globe. Every day, hosts offer unique stays and experiences that make it possible for guests to connect with communities in a more authentic way. The Community You Will Join As a Systems Engineer within Airbnb&aposs BizTech Solutions Engineering team, you will be a key technical contributor responsible for designing, building, and managing critical corporate infrastructure applications. You&aposll work alongside a collaborative team that values innovation, technical excellence, and continuous learning while supporting Airbnb&aposs global workforce with scalable infrastructure solutions. The Difference You Will Make Technical Leadership: Serve as a subject matter expert on various applications and infrastructure components, influencing technical directions that drive organizational impact across multiple teams and stakeholders. Infrastructure Excellence: Design and implement highly available applications using containerization technologies (Docker, Kubernetes, GKE, AWS EKS) with service mesh experience (Istio), ensuring robust and scalable solutions. Advanced Observability: Implement and maintain sophisticated monitoring and alerting systems using Opensearch and Grafana, focusing on proactive incident prevention and performance optimization. Automation & DevOps: Build and maintain CI/CD pipelines using Jenkins and ArgoCD, while leveraging core tools like Terraform, Helm, Chef, Ansible, and Git to automate infrastructure provisioning and application management. Cross-Platform Expertise: Manage enterprise-scale Linux and Windows operating systems across multiple cloud environments (AWS, GCP, OCI, Azure), implementing security-first solutions and best practices. A Typical Day Your contributions may take a variety of forms, including: Monitor application and infrastructure performance proactively by evaluating performance metrics such as MTTA, MTTR, and improvement targets, implementing advanced observability solutions and alerting mechanisms to maintain high SLI/SLO/SLA standards and ensure operational efficiency aligns with organizational goals. Research and analyze technical and operational aspects of applications, documenting status through comprehensive runbooks and reports while identifying gaps and implementing strategic fixes for long-term improvements through gap analysis and risk assessments. Collaborate with cross-functional teams, including Systems Engineering and Operations Engineering, to discuss architectural designs, infrastructure rollouts, and organizational priorities while engaging with stakeholders to gather feedback, identify gaps in applications or services, and develop actionable solutions that support Airbnb&aposs engineering efficiency objectives. Design and implement complex technology stacks and highly available applications using containerization technologies, implementing automated deployment processes and ensuring optimal performance across distributed environments while addressing technical issues through advanced configuration management tools such as Terraform, Chef, Kubernetes, and Helm. Implement processes for infrastructure migrations and upgrades, addressing retirements of obsolete versions while ensuring compliance with security measures, including PKI and Hashicorp Vault, and participating in on-call rotations for critical services with a security-first mindset. Create comprehensive technical documentation, develop support processes and procedures, and provide operational training to facilitate knowledge transfer and support ownership handoff to peers and organizational units while mentoring peers and stakeholders on applications, infrastructure components, and automation practices. Execute tactical and strategic technical planning, demonstrating project management capabilities while effectively prioritizing and managing multiple concurrent initiatives, participating in team check-ins to ensure alignment on ongoing projects. Engage with technical communities through open source contributions, speaking engagements, or blogging, while coordinating groups of technical individuals to drive broader industry impact and promote innovation within the systems engineering discipline. Your Expertise Experience: 4-8 years in systems engineering with upper-level individual contributor experience and a BS in Computer Science or related discipline preferred. Technical Skills: Expert-level proficiency in cloud platforms (AWS, GCP, OCI, Azure) with hands-on experience in infrastructure provisioning and management Deep working knowledge of Terraform, Helm, Chef, Ansible, and Git as core workflow tools Advanced containerization experience with Docker, Kubernetes, and service mesh technologies (Istio) Proficient in general-purpose programming languages including Python or Go for automation and tooling development Experienced in implementing and maintaining CI/CD pipelines and associated tooling (Jenkins, ArgoCD) Skilled in advanced observability tools and practices, with preference for Opensearch and Grafana Leadership Competencies: Proven ability to establish subject matter expertise and influence technical directions for organizational impact Experience operating as part of distributed teams using asynchronous workflows and communications with global workforce Demonstrated capability in designing complex technology stacks and managing enterprise-scale infrastructure Strong emphasis on security-first solution mindset during ticket triage and incident response Excellent technical communication skills with ability to create clear documentation and training materials Track record of taking ownership, accountability, and responsibility for technical outcomes and team success Show more Show less

Posted 5 days ago

Apply

5.0 - 10.0 years

25 - 37 Lacs

pune

Work from Office

About Position: We are seeking a DevOps Engineer with ArgoCD experience of 5 years. Role: DevOps Engineer with ArgoCd Location: Pune Experience: 5 to 11 yrs Job Type: Full Time Employment What You'll Do: Build reusable workflows using Go, empowering developers to provision infrastructure, deploy applications, manage secrets, and operate at scale without needing to become Kubernetes or cloud experts Drive platform standardization and codification of best practices across cloud infrastructure, Kubernetes, and CI/CD Create developer friendly APIs and experiences while maintaining a high bar for reliability, observability, and performance Design, develop, and maintain Go-based platform tooling and self-service automation that simplifies infrastructure provisioning, application deployment, and service management. Write clean, testable code and workflows that integrate with our internal systems such as GitLab, ArgoCD, Port, AWS, and Kubernetes. Partner with product engineering, SREs, and cloud teams to identify high-leverage platform improvements and enable adoption across brands. Expertise You'll Bring: Experience in Linux, Git/GitHub, Docker, K8S, ArgoCD, Terraform, Shell-scripting (Bash), AWS ECS/S3/Fargate/Lambdas/RDS/Elasticache, DataDog. Bonus tech: DroneCI, Splunk. Establish templates, reusable libraries, and workflows to promote consistency and scalability. Maintain observability, alerting, and health monitoring for platform systems and workflows. Develop with an automation-first mindset eliminating toil and manual operations through codification. Benefits: Competitive salary and benefits package Culture focused on talent development with quarterly growth cycles, company-sponsored higher education and certifications Opportunity to work with cutting-edge technologies Employee engagement initiatives such as project parties, flexible work hours, and Long Service awards Annual health check-ups Insurance coverage: group term life, personal accident, and Mediclaim hospitalization for self, spouse, two children, and parents Values-Driven, People-Centric & Inclusive Work Environment: Persistent Ltd. is dedicated to fostering diversity and inclusion in the workplace. We invite applications from all qualified individuals, including those with disabilities, and regardless of gender or gender preference. We welcome diverse candidates from all backgrounds. We support hybrid work and flexible hours to fit diverse lifestyles. Our office is accessibility-friendly, with ergonomic setups and assistive technologies to support employees with physical disabilities. If you are a person with disabilities and have specific requirements, please inform us during the application process or at any time during your employment Let's unleash your full potential at Persistent - persistent.com/careers "Persistent is an Equal Opportunity Employer and prohibits discrimination and harassment of any kind."

Posted 5 days ago

Apply

3.0 - 6.0 years

8 - 15 Lacs

kochi

Hybrid

As a Network Engineer specializing in automation, youll streamline the deployment and operation of core network functions—such as mobile and packet core—across on-premises and cloud environments. Your focus will be on improving reliability, scalability, and efficiency through deployment automation, CI/CD pipelines, and infrastructure-as-code tools. This role requires strong knowledge of modern networking and automation technologies, along with the ability to troubleshoot, optimize workflows, and maintain high service availability. You’ll take initiative on technical tasks, work independently, and contribute to building secure and resilient network automation solutions. Key duties/ responsibilities Build and maintain CI/CD pipelines to deploy and test network functions on both on-prem and cloud environments. Develop infrastructure-as-code templates (Terraform, Helm) to create repeatable, reliable environments for core network workloads. Automate day-0/day-1/day-2 lifecycle management (installation, upgrades, scaling, patching) and reduce manual operations. Enhance monitoring and observability with logging, metrics, and alerting solutions. Improve operational reliability by integrating automated testing, resilience, and performance checks. Collaborate with operations teams to transition manual tasks into automated workflows. Participate in on-call rotation and contribute to post-incident analysis and continuous improvement. Skills Required: Experience in configuring, monitoring, and troubleshooting enterprise networks. Strong foundation in TCP/IP, routing protocols and should be good at troubleshooting network outages Automation & CI/CD: Practical experience with GitLab CI, ArgoCD, or similar tools; scripting in Python, Bash, or Go; Infrastructure as Code: Hands-on experience with Terraform and Helm for automated deployments. Kubernetes: Proficiency in deploying, managing, and scaling applications on Kubernetes clusters, including container orchestration, Helm chart development, and cluster resource management Linux & Networking: Strong Linux administration skills and solid networking fundamentals (routing, firewalls, troubleshooting) Monitoring & Observability: Familiarity with logging and metrics tools (Datadog, Prometheus, Nagios, Zabbix, etc.); performance analysis capabilities. Nice to have: Core Networks: basic understanding of Mobile core architecture and protocol SCTP, GTP-U/C, Diameter, HTTP/2 Experience with open-source mobile cores or automation frameworks used in telecom environments. Understanding of high-performance networking (DPDK, SR-IOV). Experience building Kubernetes operators or advanced networking controllers. Certifications like GCP , AWS, will be plus Ability to absorb large amounts of technical knowledge and apply it to troubleshooting and support tasks Strong troubleshooting and analytical mindset. Ability to work across operations and development teams. Clear communication and documentation skills.

Posted 5 days ago

Apply

7.0 - 10.0 years

2 - 7 Lacs

bengaluru, karnataka, india

On-site

The Role Infrastructure Specialists at Kyndryl are project-based subject matter experts in all things infrastructure good at providing analysis, documenting and diagraming work for hand-off, offering timely solutions, and generally figuring it out. This is a hands-on role where your feel for the interaction between a system and its environment will be invaluable to every one of your clients. There are two halves to this role: First, contributing to current projects where you analyze problems and tech issues, offer solutions, and test, modify, automate, and integrate systems. And second, long-range strategic planning of IT infrastructure and operational execution. This role isn't specific to any one platform, so you'll need a good feel for all of them. And because of this, you'll experience variety and growth at Kyndryl that you won't find anywhere else. You'll be involved early to offer solutions, help decide whether something can be done, and identify the technical and timeline risks up front. This means dealing with both client expectations and internal challenges in other words, there are plenty of opportunities to make a difference, and a lot of people will witness your contributions. In fact, a frequent sign of success for our Infrastructure Specialists is when clients come back to us and ask for the same person by name. That's the kind of impact you can have! This is a project-based role where you'll enjoy deep involvement throughout the lifespan of a project, as well as the chance to work closely with Architects, Technicians, and PMs. Whatever your current level of tech savvy or where you want your career to lead, you'll find the right opportunities and a buddy to support your growth. Boredom Trust us, that won't be an issue. Your future at Kyndryl There are lots of opportunities to gain certification and qualifications on the job, and you'll continuously grow as a Cloud Hyperscaler. Many of our Infrastructure Specialists are on a path toward becoming either an Architect or Distinguished Engineer, and there are opportunities at every skill level to grow in either of these directions. Who You Are You're good at what you do and possess the required experience to prove it. However, equally as important you have a growth mindset; keen to drive your own personal and professional development. You are customer-focused someone who prioritizes customer success in their work. And finally, you're open and borderless naturally inclusive in how you work with others. Required Technical and Professional Experience 7+ Years of experience in RedHat OpenShift Container (mandatory), 3+ years of experience in sysdig, ArgoCD, Podman, Helm, Github. Ticketing tool ServiceNow. Provide technical support to healthcare client using RedHat OpenShift container orchestration platform, with a focus on deployments in the healthcare industry. Possess working knowledge of Kubernetes, ServiceMesh, Certificates Management, Kubernetes Persistent Storage, SDS and more Troubleshoot and resolve OpenShift cluster issues, including pod failures, network issues, storage problems, and security vulnerabilities. Work with clients to identify the root cause of problems, implement solutions, and ensure ongoing stability. Stay up-to-date on the latest OpenShift features and functionality, including security best practices. Keep the clusters up to date on patching upgrades Collaborate with development and product teams to identify and resolve systemic issues. Maintain a strong understanding of HIPAA and other healthcare compliance regulations as they pertain to OpenShift deployments. Working knowledge of OnPrem and Cloud based OpenShift deployments OpenShift Integration: Maintain and troubleshoot integrations between OpenShift clusters and external systems, including logging platforms (e.g., ELK Stack), monitoring tools (e.g., Prometheus, Grafana), and Incident alerting systems (e.g., ServiceNow, PagerDuty, ServiceNow). Configure and manage OpenShift routes and services for communication with external applications. DevOps Expertise: Possess a working knowledge of DevOps principles and CI/CD pipelines. Understand and be able to troubleshoot deployments using Argo CD, Jenkins or similar tools. Preferred Technical and Professional Experience Exceptional judgment and decision-making abilities Ready to work on North America EST/PST Familiar with a variety of IT concepts, practices, and procedures Knowledge of design techniques, tools, and principals involved in production of precision technical plans, blueprints, drawings, and models Logic and reasoning to identify the strengths and weaknesses of alternative solutions, conclusions, or approaches to problems Strong analytical and problem-solving skills with high attention to detail Being You Diversity is a whole lot more than what we look like or where we come from, it's how we think and who we are. We welcome people of all cultures, backgrounds, and experiences. But we're not doing it single-handily: Our Kyndryl Inclusion Networks are only one of many ways we create a workplace where all Kyndryls can find and provide support and advice. This dedication to welcoming everyone into our company means that Kyndryl gives you and everyone next to you the ability to bring your whole self to work, individually and collectively, and support the activation of our equitable culture. That's the Kyndryl Way. What You Can Expect With state-of-the-art resources and Fortune 100 clients, every day is an opportunity to innovate, build new capabilities, new relationships, new processes, and new value. Kyndryl cares about your well-being and prides itself on offering benefits that give you choice, reflect the diversity of our employees and support you and your family through the moments that matter wherever you are in your life journey. Our employee learning programs give you access to the best learning in the industry to receive certifications, including Microsoft, Google, Amazon, Skillsoft, and many more. Through our company-wide volunteering and giving platform, you can donate, start fundraisers, volunteer, and search over 2 million non-profit organizations. At Kyndryl, we invest heavily in you, we want you to succeed so that together, we will all succeed.

Posted 5 days ago

Apply

5.0 - 7.0 years

0 Lacs

bengaluru, karnataka, india

On-site

NV Cloud Functions group delivers a powerful platform that seamlessly links workloads to GPUs, enabling the deployment, management, and serving of GPU-accelerated, containerized applications worldwide. This solution enables businesses to efficiently utilize and scale the applications seamlessly, in an increasingly fast-paced digital environment powered by NVIDIA DGX Cloud. NVIDIA DGX Cloud is an end-to-end, scalable AI platform for developers, offering scalable capacity built on the latest NVIDIA architecture and co-engineered with the world's leading cloud service providers (CSPs). For more details, check out NVIDIA DGX Cloud. NVIDIA is looking for a Senior Cloud Software Engineer to work on platform software based on open-source container runtimes and Kubernetes technologies. We expect you to have strong programming skills, devops experience, a deep understanding of designing and building software, especially related to GO and C, experience with Systems Software and Distributed systems, as well as excellent communication and planning skills. We also welcome out-of-the-box problem solvers who can provide new insights while strong at completing tasks. Expect to be constantly challenged, improving and evolving for the better. What you will be doing: You'll be working in a distributed team that explores innovative ways to make GPU and DPU accelerated applications easier to develop, deploy, and monitor. Work on enabling GPUs and DPUs as first-class resources in Kubernetes container orchestrator. This is an excellent opportunity to join the core group working on Cloud Native technologies enabling NVIDIA accelerators in the Kubernetes environment. Work with engineering teams across NVIDIA to ensure your software integrates seamlessly with NVIDIA Cloud eco-system. Automating and optimizing build, test, integration, and release processes for cloud native. Efficiently multitasking with varied responsibilities to efficiently address evolving priorities. What we need to see: Bachelor's or Master's Degree in Computer Science or equivalent program from an accredited university/college, and 5+ years of hands-on software engineering. Expert-level knowledge in a systems programming language (Go, C) and a solid understanding of data structures and algorithms. Expertise in a scripting language (Bash, Python) Strong understanding of Container Orchestration Systems (Kubernetes) and Container Technologies. Experience working with GitLab or a similar SCM Hands-on Automation experience in continuous integration frameworks like GitLab & ArgoCD. Knowledge and experience working with system internals of Unix/Unix-like kernels such as Linux. Strong background in cloud computing and distributed software design, and development. Understanding of performance, security, and reliability in complex distributed systems. Ways to stand out from the crowd: Background with pub-sub models and message queues Experience with developing Kubernetes Custom Resources and Operators, and deploying them in a cloud service provider NVIDIA is committed to fostering a diverse work environment and is proud to be an equal opportunity employer. As we highly value diversity in our current and future employees, we do not discriminate (including in our hiring and promotion practices) on the basis of race, religion, color, national origin, gender, gender expression, sexual orientation, age, marital status, veteran status, disability status or any other characteristic protected by law.

Posted 6 days ago

Apply

3.0 - 7.0 years

0 Lacs

ahmedabad, gujarat

On-site

You are seeking a skilled Cloud Database & DevOps Architect to play a pivotal role in overseeing the design, automation, and scalability of crucial systems within our organization. Fusing your profound expertise in databases, particularly PostgreSQL and MongoDB, with your DevOps proficiency, you will be instrumental in ensuring the efficiency, performance, and resilience of our systems across various containerized and cloud environments. Your responsibilities will span creating and implementing best practices for high availability, replication, backups, disaster recovery, monitoring, and CI/CD pipelines, all while actively engaging with both development and operations teams. Your core responsibilities will include architecting and managing PostgreSQL and MongoDB clusters with an emphasis on high availability, replication, and failover mechanisms. Additionally, you will be tasked with devising and executing backup and disaster recovery strategies, containerizing and orchestrating databases and services using Docker and Kubernetes, setting up and overseeing CI/CD pipelines for seamless application and database deployments, and employing infrastructure-as-code principles for consistent configurations. Monitoring and optimizing database and infrastructure performance, safeguarding security, ensuring compliance and data governance, and collaborating with development and quality assurance teams for streamlined delivery pipelines will also be integral parts of your role. To excel in this position, you must possess expert-level knowledge in PostgreSQL and MongoDB, substantial experience with containers (specifically Docker and Kubernetes), a proven track record in high availability, clustering, replication, and disaster recovery planning, and hands-on familiarity with DevOps tools like CI/CD pipelines, GitLab CI/Jenkins, and ArgoCD. Proficiency in Infrastructure as Code tools such as Terraform and Ansible, knowledge of cloud platforms like AWS, Azure, and GCP, and a solid background in monitoring and observability tools like Prometheus, Grafana, ELK, and Loki are essential. An aptitude for troubleshooting and automation is crucial for success in this role. Desirable skills that would further enhance your suitability for this position include familiarity with Redis, Kafka, and RabbitMQ, knowledge of service mesh and API gateway in Kubernetes environments, and experience in security compliance standards such as GDPR, HIPAA, and SOC2. This is a full-time role, and the ideal candidate will have a minimum of 5 years of experience with PostgreSQL, 3 years with MongoDB, and 3 years in database administration and Docker. Availability for overnight shifts is preferred for this position.,

Posted 6 days ago

Apply

4.0 - 9.0 years

7 - 17 Lacs

bengaluru

Work from Office

About this role: Wells Fargo is seeking a Software Engineering Manager to provide leadership for an application delivery team. This Manager will lead a high-performing software engineering team and will be responsible for overseeing the development of software solutions and ensuring alignment with strategic business and product objectives. This role requires a combination of technical expertise, leadership skills, and product management capabilities. In this role, you will: Manage, coach, and develop a team of individual contributor engineer roles with low to moderate complexity and less experienced managers who are responsible for building high quality capabilities with modern technology Ensure adherence to the Banking Platform Architecture, and meeting non-functional requirements with each release Engage with architects and experienced engineers to incorporate Wells Fargo Technology technical strategies, while understanding next generation domain architecture to enable application migration paths to target architecture; for example, cloud readiness, application modernization and data strategy Function as the technical representative for the product during cross-team collaborative efforts and planning Identify and recommend opportunities for driving resolution of technology roadblocks including code, build and deployment while also managing overall software development cycle and security standards Act as an escalation partner for scrum masters and the teams to make decisions and help remove impediments, obstacles, and friction while encouraging constant learning, experimentation, and continual improvement culture Build engineering skills side-by-side in the codebase, conduct peer reviews to evaluate quality and solution alignment to technical direction, and guide design, as needed Interpret, develop, and ensure security, stability, and scalability within functions of technology with low to moderate complexity, as well as identify, manage and mitigate technology and enterprise risk Collaborate and consult with the Product Managers/Product Owners to drive user satisfaction, influence technology requirements and priorities in the product roadmap, promote innovative and intelligent solutions, generate corporate value and articulate technical strategy while being a solid advocate of agile and DevOps practices Interact directly with third party vendors and technology service providers Manage allocation of people and financial resources for technology engineering including career development and performance management for engineers and managers on the team Hire, mentor and guide talent development of direct reports to build the skills required to effectively design and deliver innovative solutions for the supported product areas/products Required Qualifications: 4+ years of Software Engineering experience, or equivalent demonstrated through one or a combination of the following: work experience, training, military experience, education 2+ years of Leadership experience Desired Qualifications: 4+ years of Software Engineering experience and 2+ years of Leadership experience Strong experience in software development frameworks , cloud-native architectures, and platform engineering. Proficiency in Java , Python, and experience with Spring Boot, .NET Core, or equivalent frameworks . Deep understanding of CI/CD pipelines, Kubernetes, containerization, and infrastructure-as-code (IaC) tools . Hands-on experience with GitHub Actions, Jenkins, ArgoCD, Backstage, or similar DevTools . Strong grasp of performance optimization, observability (OpenTelemetry, Prometheus, Grafana), and developer workflow automation . Experience with developer portals, self-service platforms, and inner-source adoption . Knowledge of AI-powered developer tools, GitHub Copilot, or generative AI-based coding assistance . Exposure to multi-cloud environments (AWS, Azure, GCP) and security best practices (OWASP, SAST, DAST) . Job Expectations: Technical Leadership & Strategy: Define the vision, roadmap, and strategy for Engineering Frameworks, Utilities, and Developer Tools & Environments. Drive adoption of modern engineering frameworks and establish coding standards, best practices, and reusable components. Lead the design and implementation of scalable development tools that improve developer efficiency and experience . Oversee the migration and modernization of legacy frameworks and environments . Engineering Frameworks & Utilities: Develop and maintain core engineering frameworks for backend, frontend, and cloud-native development. Implement domain-driven design (DDD), event-driven architectures , and microservices patterns. Ensure performance, security, observability, and compliance in engineering frameworks. Drive automation in code generation, service scaffolding, and dependency management . Developer Tools & Environments: Lead the development of cloud-based IDEs, CI/CD pipelines, local dev environments, and code quality tools . Own the engineering productivity platform , including build acceleration, test automation, and debugging tools . Collaborate with security and compliance teams to integrate DevSecOps practices into developer environments. Optimize inner-loop development workflows to reduce friction and enhance developer experience. Cross-Team Collaboration & Enablement: Partner with Product, Architecture, SRE, Security, and Developer Experience teams to drive engineering excellence. Advocate for engineering best practices , tool adoption , and developer education initiatives . Work closely with engineering leadership to align platform investments with business priorities . People Leadership & Execution: Lead, mentor, and grow a team of high-performing engineers and technical leads . Foster a culture of innovation, accountability, and continuous improvement . Manage project execution, ensuring timely delivery and high-quality outcomes . Leadership & Collaboration: Proven track record of leading engineering teams and delivering platform-level solutions. Strong stakeholder management skills with the ability to drive adoption of engineering frameworks and tools . Excellent problem-solving skills and a data-driven approach to measuring developer productivity

Posted 6 days ago

Apply

20.0 - 30.0 years

20 - 30 Lacs

bengaluru

Work from Office

We're looking for an engineer to join our Infrastructure team which hosts platform and the underlying services in cloud platforms for our clients. On the Platform Team you'll work with engineers on the infrastructure team as we tackle problems of tooling, deployment, observability, performance, architecture, and lots of other fun stuff. Our work as a team is broad - you will gain exposure to things you've never worked with and have a chance to build deeper expertise in what you do know. This is a high impact engineering role and is best suited to someone who is excited to work in a growing startup and wants the opportunity to contribute to a growing infrastructure team. Responsibilities: Design and Build Scalable Backend Systems Develop high-performance, reliable backend services that power core Customer Data Platform (CDP) capabilities, ensuring scalability and long-term maintainability. Improve and Refactor Existing Services Continuously enhance system performance and code quality by refactoring legacy services, adopting modern best practices, and addressing technical debt. Maintain Operational Support and Platform Stability Contribute to the stability and resilience of the platform through active monitoring, troubleshooting, and participation in the Platform On-Call Rotation. Collaborate Across Teams and Mentor Others Work closely with cross-functional partners in product, frontend, and DevOps. Provide technical mentorship and knowledge sharing within the team. Drive Strategic Initiatives Play a key role in the Core CDP Infrastructure Consolidation project, Support CI/CD Processes Ensure smooth deployments and delivery by adhering to and enhancing Uniphores existing CI/CD pipelines and release practices. Requirements 5 – 7 years of experience in roles like DevOps Engineer, Strong bias toward the simplest solution to a problem Curious about how and why things work Experience with large, distributed workloads on public cloud platforms (AWS) Understanding of networking fundamentals (TCP/IP, DNS, NAT, Routing, Load Balancing, the OSI Model, etc) Experience with observability, APM, and monitoring tools like Datadog, Prometheus, New Relic, Honeycomb, etc. Experience with build and deployment systems like Jenkins, Travis, Ansible, Puppet, Chef, Terraform, Kubernetes, etc. Experience with at least one programming language Experience implementing CI/CD pipelines, particularly using Jenkins You have operated and contributed in an environment where Infrastructure as Code is the preferred approach Experience working on a SaaS product Experience achieving compliance standards like ISO, SOC, and similar Nice to Haves Experience with infrastructure and deployment tools such as Garden, Flux, and Ansible Familiarity with security tooling like Snyk and SonarQube Exposure to other cloud providers, such as Google Cloud Platform (GCP) or Microsoft Azure Strong understanding of Test Driven Development (TDD) principles and practices Experience working with or building Customer Data Platforms (CDP)

Posted 6 days ago

Apply

0.0 - 4.0 years

0 Lacs

haryana

On-site

As a Contract Logistics Specialist at Kuehne+Nagel, you will be responsible for managing end-to-end warehousing operations for customers. Your precise management will not only enhance team success but also contribute significantly to the day-to-day operations and overall success of warehouse and distribution centers. For instance, you will oversee tasks such as storing and delivering delicate flowers and fresh ingredients to local stores, catering to everyday lunches and family celebrations. Your role at Kuehne+Nagel truly plays a crucial part in various aspects beyond what one can initially imagine. We are currently looking for a motivated and inquisitive Junior DevOps Engineer to join our dynamic team. This position is well-suited for individuals with a strong service mindset, a solid technical background, and a keen interest in continuous learning and automation. You will collaborate closely with senior engineers to provide support in infrastructure maintenance and manage deployment pipelines effectively. Key Responsibilities: - Assist in deploying and maintaining WMS virtual machines utilizing tools like Jenkins and GitLab. - Work alongside development and operations teams to ensure seamless application delivery and optimal performance. - Monitor systems using tools such as Grafana and Oracle EM for efficient operations. - Contribute to troubleshooting and resolving infrastructure and application-related issues. - Document processes and actively participate in knowledge sharing within the team. Qualifications: - Bachelor's degree in Computer Science, Engineering, or related field (or equivalent practical experience). - Strong service orientation with a dedication to supporting internal teams and enhancing user experience. - Technical proficiency in areas like Linux, Oracle, Java, Virtualization, Cloud Platforms (e.g., AWS), Git, Jira, GitLab, Jenkins, ArgoCD, Terraform, Ansible, Docker, Kubernetes, OpenShift, Grafana, Loki, Mimir, Tempo. - Fundamental understanding of operations and system administration, monitoring and alerting practices, automation of repetitive tasks, troubleshooting, and root cause analysis. - Previous internship or project exposure in DevOps or related domains. - Familiarity with Agile methodologies. What's in it for you: Joining Kuehne+Nagel means becoming part of a global logistics leader that focuses on creating tangible impacts on business, customers, and careers. Here's what you can expect: - Global Exposure: Explore a world of international opportunities with a presence in over 100 countries. - People-Centric Culture: Join a team where your opinion is valued, and individuals genuinely care for each other. - Learning & Development: Evolve personally and professionally through top-notch training and career growth opportunities. - Innovation & Sustainability: Contribute to a forward-thinking organization driving real change in logistics and environmental sustainability. - Rewards & Recognition: Be acknowledged for your dedication, performance, and potential. - Stability with Agility: Experience the best of both worlds - the reliability of a trusted global brand combined with the entrepreneurial spirit of a startup. About Us: Logistics plays a vital role in everyday life, from the goods we use to the healthcare we rely on. At Kuehne+Nagel, your work transcends logistics; it influences both ordinary and extraordinary moments in people's lives worldwide. As a global leader with a rich history and a vision to propel the world forward, we provide a secure and supportive environment where your career can truly make a meaningful impact. Whether we are facilitating the delivery of life-saving medications, developing sustainable transportation solutions, or supporting local communities, your career journey with us will contribute to a greater purpose than you can envision.,

Posted 1 week ago

Apply

5.0 - 8.0 years

10 - 15 Lacs

mumbai

Work from Office

Key Responsibilities Strong understanding of product development methodologies and microservices architecture Hands-on experience with at least two major cloud providers (AWS, GCP, Azure); multi-cloud expertise is a plus Expertise in designing, implementing, and managing scalable, secure, and resilient cloud architectures Knowledge of cloud fundamentals Networking, IAM, Compute, Managed Services (DB, Storage, GKE/EKS, KMS) Hands-on experience in cloud architecture design & setup Proficiency with Infrastructure as Code (IaC) tools Terraform, Helm (must-have), Kustomize Experience deploying, maintaining, and scaling apps on Kubernetes clusters Hands-on with CI/CD tools (GitLab CI, Jenkins, GitHub Actions) and GitOps tools (ArgoCD, FluxCD must-have) Strong knowledge of Monitoring & Logging tools Prometheus, Grafana, Elastic Stack Experience with PaaS and on-prem data center deployments Exposure to k3s OSS / OpenShift / Rancher Kubernetes clusters is a plus What Were Looking For Ability to Learn, Architect & Build Hands-on expertise in the skills & technologies above Product-oriented delivery mindset Experience in Designing, Building, and Operating Cloud Architecture & DevOps Pipelines Strong background in Open Source Technologies Collaboration across multiple product teams (5+) Adherence to GitOps philosophy DevSecOps mindset building highly secure platforms

Posted 1 week ago

Apply

5.0 - 10.0 years

7 - 17 Lacs

mumbai

Work from Office

An understanding of product development methodologies and microservices architecture. Hands-on experience with at least two major cloud providers (AWS, GCP, Azure). Multi-cloud experience is a strong advantage. Expertise in designing, implementing, and managing cloud architectures focusing on scalability, security, and resilience. Understanding and experience with cloud fundamentals like Networking, IAM, Compute, and Managed Services like DB, Storage, GKE/EKS, and KMS. Hands-on experience with cloud architecture design & setup. An in-depth understanding of Infrastructure as Code tools like Terraform, HELM is a must. Practical experience in deploying, maintaining, and scaling applications on Kubernetes clusters using Helm Charts or Kustomize Hands-on experience with any CI/CD tools like Gitlab CI, Jenkins, Github Actions. GitOps tools like ArgoCD, FluxCD is a must. Experience with Monitoring and Logging tools like Prometheus, Grafana and Elastic Stack. Experience working with PaaS is a plus Experience deploying on-prem data centre. Experience with k3s OSS / OpenShift / Rancher Kubernetes Cluster is a plus What are we looking for Learn, Architect & Build Skills & Technologies as highlighted above Product-Oriented Delivery Design, Build, and Operate Cloud Architecture & DevOps Pipeline Build on Open Source Technologies Collaboration with teams across 5 products GitOps Philosophy DevSecOps Mindset - Highly Secure Platform

Posted 1 week ago

Apply

8.0 - 12.0 years

20 - 30 Lacs

bengaluru

Work from Office

AWS Cloud -EC2,S3,Lambda,I am EKS ,Helm chart ,ArgoCD ,Terraform ,prometheus and grafana . Should be ready for face to face interview. AWS and Kubernate certified.

Posted 1 week ago

Apply

5.0 - 9.0 years

15 - 19 Lacs

mumbai

Remote

Key Responsibilities Strong understanding of product development methodologies and microservices architecture Hands-on experience with at least two major cloud providers (AWS, GCP, Azure); multi-cloud expertise is a plus Expertise in designing, implementing, and managing scalable, secure, and resilient cloud architectures Knowledge of cloud fundamentals Networking, IAM, Compute, Managed Services (DB, Storage, GKE/EKS, KMS) Hands-on experience in cloud architecture design & setup Proficiency with Infrastructure as Code (IaC) tools Terraform, Helm (must-have), Kustomize Experience deploying, maintaining, and scaling apps on Kubernetes clusters Hands-on with CI/CD tools (GitLab CI, Jenkins, GitHub Actions) and GitOps tools (ArgoCD, FluxCD must-have) Strong knowledge of Monitoring & Logging tools Prometheus, Grafana, Elastic Stack Experience with PaaS and on-prem data center deployments Exposure to k3s OSS / OpenShift / Rancher Kubernetes clusters is a plus What Were Looking For Ability to Learn, Architect & Build Hands-on expertise in the skills & technologies above Product-oriented delivery mindset Experience in Designing, Building, and Operating Cloud Architecture & DevOps Pipelines Strong background in Open Source Technologies Collaboration across multiple product teams (5+) Adherence to GitOps philosophy DevSecOps mindset building highly secure platforms

Posted 1 week ago

Apply

13.0 - 18.0 years

0 Lacs

bengaluru, karnataka, india

On-site

Job Description Extremes Cloud Operations team is a group of talented engineers dedicated to building scalable, reliable, and secure cloud-based solutions. We are seeking a highly skilled and motivated Cloud Test Bed Engineer to manage and optimize cloud test beds for both functional testing and performance testing. In this role, you will be responsible for ensuring that our cloud environments are configured, automated, and optimized for end-to-end testingranging from validating core functionalities to assessing system performance under heavy loads. Experience: 13-18 Years Responsibilities Design, manage, and optimize cloud test bed environments in AWS, GCP, and Azure for both functional and performance testing. Collaborate with Development and QA teams to define testing requirements for functional validation (unit, integration, regression tests) and performance testing (load, stress, scalability). Provision, automate, and scale test environments using infrastructure-as-code tools like Terraform, CloudFormation, or similar solutions to support both functional and performance testing. Ensure that cloud test beds replicate production environments and accurately simulate real-world traffic and load for performance testing, as well as support stable, isolated environments for functional testing. Automate the setup and teardown of cloud test environments to efficiently handle continuous integration and continuous testing processes. Develop and implement monitoring solutions to track key metrics for both functional test validation and performance testing (e.g., test execution times, system resource utilization, response times). Troubleshoot issues in test environments (whether related to functional or performance testing) and work to resolve them quickly, ensuring minimal downtime for testing activities. Collaborate with teams to identify performance bottlenecks, scalability issues, and functional bugs and ensure timely resolution during testing phases. Continuously optimize and improve the efficiency of test bed management, ensuring that test environments are available, scalable, and cost-effective. Analyze test results for both functional and performance tests, identifying trends, issues, and opportunities for improvement in test bed infrastructure and testing strategies. Ensure seamless integration of performance and functional testing into the CI/CD pipeline to support rapid feedback cycles during the development process. Minimum Qualifications BS-level technical degree required; Computer Science or Engineering background preferred. 13+ years of experience in CloudOps or DevOps with a focus cloud application management. Hands-on experience with AWS, GCP, or Azure in building and maintaining cloud-based environments. Strong knowledge of Linux systems, networking, and cloud security fundamentals. Experience with containerized environments (Docker, Kubernetes) and their use in functional and performance testing scenarios. Proficiency with automation tools (Terraform, CloudFormation, Ansible, Helm, ArgoCD) for managing cloud environments. Experience with monitoring tools (e.g., Nagios, Grafana, Prometheus) to track and analyze test bed performance and functional test results. Experience in troubleshooting and optimizing cloud infrastructure for performance testing, including identifying bottlenecks and ensuring accurate results. Familiarity with CI/CD pipelines and integrating both functional and performance testing into these workflows. Strong collaboration and communication skills to work effectively with cross-functional teams across multiple time zones. Preferred Qualifications Experience with performance testing tools (e.g., JMeter, LoadRunner, Apache Benchmark) and functional testing tools (e.g., Selenium, TestNG). Experience with highly scalable microservices architectures and testing them in cloud environments. Experience with large-scale distributed systems and managing test beds that simulate complex, production-like scenarios for both functional and performance validation. Experience with performance tuning in cloud environments, identifying and resolving issues that may affect performance test results. Show more Show less

Posted 1 week ago

Apply

6.0 - 8.0 years

0 Lacs

ahmedabad, gujarat, india

On-site

Job Description: Technical Skill . Experience working on AWS, Kubernetes Orchestration and EKS. . Must have 6+ years of Production Env. Experience. . Exp. with infra-automation using Terraform| CloudFormation | Ansible. . Exp. working on Linux environment with at least one scripting language. . Exp. of CI/CD pipeline using Jenkins, Harness l ArgoCD . Exp. in Application Performance Monitoring tools such as Datadog, Instana, Grafana, Splunk, PagerDuty, Pingdom and Cloud Watch. . Good to know and at least hands on experience on DevOps and Cloud Operation Process and Agile Release Processes, and L2/L3 Ticketing process and experience in SRE role. . Should have experience with source control and management tools like Git. Essential Duties & Responsibilities: . Should be able to work in 16 X 5 shifts for support of infrastructure, Weekly rotational shifts Morning OR evening. . Design, implement and maintain highly available, scalable AWS infrastructure and services within a managed service environment and continual re-evaluation. . Manage the deployments in env. such as QA/Stage/Production and engaging stack-holders as required, and adhering to SLA (Service Level Agreement) . Build, Deploy and Manage Kubernetes clusters through automation like Terraform. . Institute infrastructure as code, security and process automation and automation of routine maintenance tasks . Security and Vulnerability Patching. . Create and deliver knowledge-sharing presentations and documentation. . Learning on the job and exploring new technologies with little supervision. . Participating in and leading war-room/critical outage calls and ensuring the RCA process is implemented. . Cross located Team Management to create a self-managed team, support skill development, and provide growth path, performance evaluation, and attrition management. . Customer Management for daily calls to discuss progress/updates, weekly & monthly status reporting, and lead calls with cross-functional team for creating new services monitors. . Ability to work with and influence cross-functional teams of customer / eInfochips. . Delivery Management to ensure well-defined SLAs (Service Level Agreement) and adherence of the same by self/team using delivery tracking tool - JIRA / Confluence. . Quality Management to contribute that all deliverables from eInfochips are following the quality processes of eInfochips and the customer. . Documentation of processes/status reporting / SOPs for future reference. . Knowledge of AI tools, and Savant with Amazon SageMaker Services will be an added advantage. Tools and Technology specific Exp. Level: Linux : 5+ Yrs. AWS : 5+ Yrs. Kubernetes : 3+ Yrs. Terraform : 3+ yrs. Total 5+ Year Production Operations and Maintenance Experience.

Posted 1 week ago

Apply

3.0 - 7.0 years

0 Lacs

haryana

On-site

As a Systems Designer at our company, you will be responsible for designing large scale systems with high complexity to support our high throughput applications. You will use your expertise to understand how to leverage infrastructure for solving such large scale problems effectively. Additionally, you will have the opportunity to develop tools and contribute to open source projects wherever possible. We encourage problem solving as a way of life and expect you to always go to the root cause to support the code you write in production. In terms of technology stack, we select the best combination of tech based on the requirements. You will be working with technologies such as Java, Kotlin, Scala, Spring Boot, JUnit, Resilience4j, Feign, MySQL/AuroraDB, DynamoDB, ELK, Kafka, Redis, TiDB, Docker, Kubernetes, ArgoCD, AWS, GCP, GitHub, IntelliJ, Gradle, Maven, npm/yarn, Flyway, Jenkins, snyk, Bigquery, Kibana, Spark, PlantUML, draw.io, miro.com, Slack, and Zoom. To qualify for this role, you should have at least 3 years of experience using Java in a professional environment. Alternatively, if you are a non-Java user, you should possess excellent skills in generalized programming languages such as Scala, Python, and Go. An interest and ability to learn other coding languages as needed is essential. Experience with NoSQL databases and distributed cache, strong fundamentals in data structures, algorithms, and object-oriented programming, as well as an in-depth understanding of concurrency and distributed computing are required. A degree in Computer Engineering or Computer Science, or 5+ years equivalent experience in SaaS platform development is preferred. Proficiency in English or Japanese is necessary. Preferred qualifications include experience with implementing platform components like RESTful APIs, Pub/Sub Systems, and Database Clients, familiarity with AWS services, expertise in system design and development of web systems handling massive transactions or requiring scalability, language ability in Japanese and English, and experience in microservices. This is a full-time position located at our Gurugram office in Wework. The development center requires your presence in the office to establish a strong core team.,

Posted 1 week ago

Apply
cta

Start Your Job Search Today

Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.

Job Application AI Bot

Job Application AI Bot

Apply to 20+ Portals in one click

Download Now

Download the Mobile App

Instantly access job listings, apply easily, and track applications.

Featured Companies