Get alerts for new jobs matching your selected skills, preferred locations, and experience range. Manage Job Alerts
15.0 - 17.0 years
0 Lacs
noida, uttar pradesh, india
Remote
Req ID: 340251 NTT DATA strives to hire exceptional, innovative and passionate individuals who want to grow with us. If you want to be part of an inclusive, adaptable, and forward-thinking organization, apply now. We are currently seeking a Industry Consulting Manager to join our team in Noida, Uttar Pradesh (IN-UP), India (IN). Job Description: Technical Architect Observability & SRE Frameworks Position Title: Technical Architect Observability & Site Reliability Engineering (SRE) Location: Noida, India Experience: 15+ years (including 5+ years in observability/SRE architecture) Employment Type: Full-time Role Overview We are looking for a highly experienced Technical Architect to lead the design, strategy, and implementation of Observability and SRE frameworks for enterprise-scale, microservices-based applications. The ideal candidate will bring deep technical knowledge of both Splunk Observability Stack and Open Source tools (like OpenTelemetry, Prometheus, Grafana, Jaeger), and be capable of defining and executing architecture strategies for complex distributed systems. This role requires hands-on ability to create architecture blueprints , lead technical teams, and work directly with stakeholders and platform owners to embed observability and reliability practices across the SDLC. Key Responsibilities Architecture & Blueprinting Design and deliver end-to-end observability architecture (Metrics, Logs, Traces, Events) for cloud-native and hybrid environments. Create technical architecture diagrams, data flow maps, and integration blueprints using tools like Lucidchart, Draw.io, or Visio. Lead the definition of SLIs, SLOs, and Error Budgets aligned with business KPIs and DORA metrics. Toolchain Strategy & Implementation Architect telemetry pipelines using OpenTelemetry Collector and Splunk Observability Cloud (SignalFx, APM, RUM, Log Observer). Define tool adoption strategy and integration roadmap for OSS tools (Prometheus, Loki, Grafana, Jaeger) and Splunk-based stacks. Guide teams on instrumentation approaches (auto/manual) across languages like Java, Go, Python, .NET, etc. Reliability Engineering Enablement Lead adoption of SRE principles including incident management frameworks, resiliency testing, and runbook automation. Collaborate with DevOps to integrate observability into CI/CD pipelines (e.g., Jenkins, ArgoCD, GitHub Actions). Define health checks, golden signals, and SPoG (Single Pane of Glass) dashboards. Exposure to AIOps, ML-based anomaly detection, or business observability. Stakeholder Management & Governance Serve as a technical liaison between client leadership, SREs, developers, and infrastructure teams. Run workshops, assessments, and evangelize observability-first culture across teams. Provide guidance on data retention, access control, cost optimization, and compliance (especially with Splunk ingestion policies). Performance & Optimization Continuously monitor and fine-tune observability data flows to prevent alert fatigue and ensure actionability. Implement root cause analysis practices using telemetry correlation across metrics, logs, and traces. Lead efforts to build self-healing systems using automated playbooks and AIOps integrations (where applicable). Required Skills & Qualifications 15+ years in IT, with 5 years in Observability/SRE architecture roles Proven experience designing architecture for microservices, containers (Docker, Kubernetes), and distributed systems Strong hands-on expertise with: Splunk Observability Cloud (SignalFx, Log Observer, APM) OpenTelemetry (SDKs + Collector) Prometheus + Grafana Jaeger / Zipkin for distributed tracing CI/CD tools: Jenkins, GitHub Actions, ArgoCD Ability to build and present clear architecture diagrams and solution roadmaps Working knowledge of cloud environments (AWS, Azure, GCP) and container orchestration (K8s/OpenShift) Familiarity with SRE and DevOps best practices (error budgets, release engineering, chaos testing) Nice to Have Splunk certifications: Core Consultant, Observability Specialist, Admin Knowledge of ITIL and modern incident management frameworks (PagerDuty, OpsGenie) Experience in banking or regulated enterprise environments Soft Skills Strong leadership and cross-functional collaboration Ability to work in ambiguous, fast-paced environments Excellent documentation and communication skills Passion for mentoring teams and building best practices at scale Why This Role Matters The client is on a journey to mature its Observability and SRE ecosystem , and this role will be critical in: Unifying legacy and modern telemetry stacks Driving reliability-first mindset and tooling Establishing a scalable blueprint for production excellence About NTT DATA NTT DATA is a $30 billion trusted global innovator of business and technology services. We serve 75% of the Fortune Global 100 and are committed to helping clients innovate, optimize and transform for long term success. As a Global Top Employer, we have diverse experts in more than 50 countries and a robust partner ecosystem of established and start-up companies. Our services include business and technology consulting, data and artificial intelligence, industry solutions, as well as the development, implementation and management of applications, infrastructure and connectivity. We are one of the leading providers of digital and AI infrastructure in the world. NTT DATA is a part of NTT Group, which invests over $3.6 billion each year in R&D to help organizations and society move confidently and sustainably into the digital future. Visit us at us.nttdata.com Whenever possible, we hire locally to NTT DATA offices or client sites. This ensures we can provide timely and effective support tailored to each clients needs. While many positions offer remote or hybrid work options, these arrangements are subject to change based on client requirements. For employees near an NTT DATA office or client site, in-office attendance may be required for meetings or events, depending on business needs. At NTT DATA, we are committed to staying flexible and meeting the evolving needs of both our clients and employees. NTT DATA recruiters will never ask for payment or banking information and will only use @nttdata.com and @talent.nttdataservices.com email addresses. If you are requested to provide payment or disclose banking information, please submit a contact us form, https://us.nttdata.com/en/contact-us . NTT DATA endeavors to make https://us.nttdata.com accessible to any and all users. If you would like to contact us regarding the accessibility of our website or need assistance completing the application process, please contact us at https://us.nttdata.com/en/contact-us . This contact information is for accommodation requests only and cannot be used to inquire about the status of applications. NTT DATA is an equal opportunity employer. Qualified applicants will receive consideration for employment without regard to race, color, religion, sex, sexual orientation, gender identity, national origin, disability or protected veteran status. For our EEO Policy Statement, please click here . If you&aposd like more information on your EEO rights under the law, please click here . For Pay Transparency information, please click here . Show more Show less
Posted 1 day ago
4.0 years
0 Lacs
bengaluru, karnataka, india
On-site
We’re transforming the software industry. We’re Flexera. With more than 50,000 customers across the world, we’re achieving that goal. But we know we can’t do any of that without our team. Ready to help us re-imagine the industry during a time of substantial growth and ambitious plans? Come and see why we’re consistently recognized by Gartner, Forrester and IDC as a category leader in the marketplace. Flexera delivers Technology Value Optimization solutions that enable some of the largest companies in the world to inform their IT so they can transform their IT. From on-prem to the cloud, companies can get the IT asset data needed to rightsize, reallocate spend, reduce risk and maximize ROI. Site Reliability Engineer - Cloud Cost Optimisation Engineer About Us We're a fast-growing, category-leading organization with ambitious objectives and a positive, inclusive culture. We're looking for passionate professionals who want to grow their talents and achieve great things. If that sounds like you, we want to talk to you about joining our team. The Cloud Enablement team is responsible for accelerating the delivery and improving the operation of our cloud-based software by providing and supporting tools and patterns which reduce the cognitive load on our development teams. We free up our developers to focus on solving problems for our customers rather than spending time on extraneous tasks. Drawing on the shared experience and expertise from our organization and industry; we create, support and evolve the paved path for teams to build, deploy and run secure and reliable software. What will you do? Design, build, advocate for and support the common tools and delivery platform used by Flexera developers. Improve developer experience and operational excellence. Foster collaboration and knowledge sharing across Flexera. Select and rollout supported defaults and standards for CI/CD tooling, Observability, Security and Runtime Environment. Work with teams across several continents, build relationships with our engineers by listening and understanding their needs and balancing this with the needs of our business. Research new tools and patterns and continuously measure and evolve our ways of doing things. Cloud Cost Optimization uses a combination of strategies, techniques, best practices and tools to help manage/reduce cloud costs. You have Developer/DevOps/SRE/Platform experience and a strong interest in software delivery and ongoing operation. Worked on rolling out automation, tools, technologies, patterns and guardrails across an organization or teams. Experience working in a globally distributed team. Extensive public cloud (preferably AWS) knowledge & experience. Deep knowledge of containers (Docker) orchestration (Kubernetes). Knowledge of tools and patterns around CI/CD (familiar with Travis CI, Circle CI, Buildkite or similar). Observability knowledge; Logs, Tracing, Metrics and experience in a few of Elastic Stack, XRay, Jaeger, Zipkin, Prometheus, Honeycomb or LightStep. Enterprise observability tools such as NewRelic, DataDog etc. Cloud cost optimization; Using automation to keep Cloud cost under control and within budget. Enabling individual Engineering teams with cloud cost optimization. Knowledge of operations, including incident management, immutable infrastructure as code (esp. Terraform or CloudFormation), and problem-solving. Produced robust well-tested code preferably in Golang; however, we will also consider Python, JavaScript, Ruby, Java or C# if you are happy to learn Go. Excellent communication skills, including experience in writing good documentation and running workshops. Critical Skills / Competencies Agile software delivery methodologies Experience managing cloud-based services e.g. AWS, Azure at scale Experience with DevOps Experience with docker Containers, Kubernetes, EKS, ECS Infrastructure as code e.g. Terraform, CloudFormation CI/CD pipelines using Jenkins, travisCI, teamcity, pipeline as code Automation / Configuration Management at scale e.g. Puppet, Chef, Ansible, Salt, Packer etc. Service mesh such as ishtio, Consul or similar Expertise in one or more of the following languages: Python / Go / Java / C# / C / C++ Experience with IaaS and Serverless services from a cloud provider A strong understanding in TCP/IP, DNS and experience designing networks Linux & Windows system administration experience Experience implementing fault detection, and automating fixes Experience designing scalable services Experience designing distributed, fault-tolerant systems A good understanding of SQL, No-SQL databases A solid understanding of data structures and algorithms A positive attitude and willingness to learn Strong conflict resolution competence Excellent written and verbal communication skills Detail oriented. The ideal candidate is one who naturally digs as deep as they need to understand the why Minimum Qualifications Bachelor's or higher degree in Computer Science, Information Technology, or a related field. At least 4 years of hands-on job experience managing services in a public cloud At least 1 years of experience working as a member of a centralized Cloud enablement / Platform or a similar team Bonus Skills The following list of items are not pre-requisites for the role but might give you a bit more of an idea about what you may expect to come across in your SRE role at Flexera: Python / Golang / Java / C# / C / C++ / Bash experience Big Data, Machine Learning, AI (DataBricks, Snowflake etc.) Platforms Experience with Monitoring systems such as New Relic, ELK, Prometheus, Datadog, X-ray etc. Security background SQL, NOSQL and Graph databases Relevant Certification e.g. AWS, GCP, Azure Experience of Disciplined Agile Delivery (DAD) Flexera is proud to be an equal opportunity employer. Qualified applicants will be considered for open roles regardless of age, ancestry, color, family or medical care leave, gender identity or expression, genetic information, marital status, medical condition, national origin, physical or mental disability, political affiliation, protected veteran status, race, religion, sex (including pregnancy), sexual orientation, or any other characteristic protected by local/national laws, policies and/or regulations. Flexera understands the value that results from employing a diverse, equitable, and inclusive workforce. We recognize that equity necessitates acknowledging past exclusion and that inclusion requires intentional effort. Our DEI (Diversity, Equity, and Inclusion) council is the driving force behind our commitment to championing policies and practices that foster a welcoming environment for all. We encourage candidates requiring accommodations to please let us know by emailing careers@flexera.com.
Posted 1 day ago
7.0 years
2 - 9 Lacs
noida
Remote
Req ID: 340254 NTT DATA strives to hire exceptional, innovative and passionate individuals who want to grow with us. If you want to be part of an inclusive, adaptable, and forward-thinking organization, apply now. We are currently seeking a Industry Consulting Snr. Consultant to join our team in Noida, Uttar Pradesh (IN-UP), India (IN). Job Title: Telemetry Engineer Location: Noida, India Employment Type: Full-Time Experience Level: Senior (7+ years preferred) Role Overview: We are seeking a highly skilled Telemetry Engineer to lead the design and implementation of telemetry pipelines across diverse environments including microservices, VM-based applications, cloud-native platforms, and on-premise systems. The ideal candidate will have deep expertise in OpenTelemetry architecture and implementation , and a strong background in observability, distributed systems, and performance monitoring. Key Responsibilities: Architect and implement end-to-end telemetry pipelines for applications deployed across cloud, on-prem, and hybrid environments. Lead the installation, configuration, and optimization of OpenTelemetry components including SDKs, Collector, and exporters. Collaborate with application, infrastructure, and DevOps teams to define telemetry standards and integrate observability into CI/CD workflows. Design scalable and resilient data collection strategies for metrics, logs, and traces . Develop and maintain instrumentation libraries for microservices and legacy applications. Ensure telemetry data is efficiently routed to observability platforms (e.g., Splunk, Prometheus, Grafana, Datadog). Conduct performance tuning and troubleshooting of telemetry pipelines. Provide architectural guidance and best practices for telemetry adoption across teams. Stay current with OpenTelemetry releases and contribute to internal tooling and automation. Required Skills & Qualifications: Proven experience in setting up telemetry pipelines from scratch across multiple environments. Strong hands-on expertise with OpenTelemetry (Collector, SDKs, OTLP protocol). Deep understanding of distributed tracing, metrics collection, and log aggregation . Experience with observability platforms such as Splunk, Prometheus, Grafana, Jaeger, Zipkin, Datadog , etc. Proficiency in one or more programming languages (e.g., Python, Go, Java, Node.js) for instrumentation. Familiarity with cloud platforms (AWS, Azure, GCP) and VM/on-prem infrastructure . Knowledge of container orchestration (Kubernetes), service meshes (Istio), and CI/CD pipelines. Excellent communication and documentation skills. Preferred Qualifications: Experience contributing to or working with the OpenTelemetry community . Certifications in cloud technologies or observability tools. Experience in regulated or enterprise-scale environments. About NTT DATA NTT DATA is a $30 billion trusted global innovator of business and technology services. We serve 75% of the Fortune Global 100 and are committed to helping clients innovate, optimize and transform for long term success. As a Global Top Employer, we have diverse experts in more than 50 countries and a robust partner ecosystem of established and start-up companies. Our services include business and technology consulting, data and artificial intelligence, industry solutions, as well as the development, implementation and management of applications, infrastructure and connectivity. We are one of the leading providers of digital and AI infrastructure in the world. NTT DATA is a part of NTT Group, which invests over $3.6 billion each year in R&D to help organizations and society move confidently and sustainably into the digital future. Visit us at us.nttdata.com Whenever possible, we hire locally to NTT DATA offices or client sites. This ensures we can provide timely and effective support tailored to each client's needs. While many positions offer remote or hybrid work options, these arrangements are subject to change based on client requirements. For employees near an NTT DATA office or client site, in-office attendance may be required for meetings or events, depending on business needs. At NTT DATA, we are committed to staying flexible and meeting the evolving needs of both our clients and employees. NTT DATA recruiters will never ask for payment or banking information and will only use @nttdata.com and @talent.nttdataservices.com email addresses. If you are requested to provide payment or disclose banking information, please submit a contact us form, https://us.nttdata.com/en/contact-us. NTT DATA endeavors to make https://us.nttdata.com accessible to any and all users. If you would like to contact us regarding the accessibility of our website or need assistance completing the application process, please contact us at https://us.nttdata.com/en/contact-us. This contact information is for accommodation requests only and cannot be used to inquire about the status of applications. NTT DATA is an equal opportunity employer. Qualified applicants will receive consideration for employment without regard to race, color, religion, sex, sexual orientation, gender identity, national origin, disability or protected veteran status. For our EEO Policy Statement, please click here. If you'd like more information on your EEO rights under the law, please click here. For Pay Transparency information, please click here.
Posted 4 days ago
8.0 years
0 Lacs
noida, uttar pradesh, india
On-site
Job Description Role : Manager - DevOps We at Pine Labs are looking for those who share our core belief - Every Day is Game day. We bring our best selves to work each day to realize our mission of enriching the world through the power of digital commerce and financial services. Role Purpose We are seeking a Manager DevOps who will lead and manage the organizations DevOps Infrastructure, Observability stack for applications, CI-CD Pipeline and support services. This role involves managing a team of DevOps engineers, architecting scalable infrastructure, and ensuring high availability and performance of our messaging and API management systems. This individual will oversee a team of IT professionals, ensure the seamless delivery of IT services, and implement strategies to align technology solutions with business objectives. The ideal candidate is a strategic thinker with strong technical expertise and proven leadership we entrust you with : Lead and mentor a team of DevOps Lead/Engineers in designing and maintaining scalable infrastructure. Architect and manage Kafka clusters for high-throughput, low-latency data streaming. Deploy, configure, and manage Kong API Gateway for secure and scalable API traffic Design and implement CI/CD pipelines for microservices and infrastructure. Automate infrastructure provisioning using tools like Terraform or Ansible. Monitor system performance and ensure high availability and disaster recovery. Collaborate with development, QA, and security teams to streamline deployments and enforce best practices. Ensure compliance with security standards and implement DevSecOps practices. Maintain documentation and provide training on Kafka and Kong usage and best practices. Strong understanding of observability pillars : metrics, logs, traces, and events. Hands-on experience with Prometheus for metrics collection and Grafana for dashboarding and visualization. Proficiency in centralized logging solutions like ELK Stack (Elasticsearch, Logstash, Kibana), Fluentd, or Splunk. Experience with distributed tracing tools such as Jaeger, Zipkin, or OpenTelemetry. Ability to implement instrumentation in applications for custom metrics and traceability. Skilled in setting up alerting and incident response workflows using tools like Alertmanager, PagerDuty, or Opsgenie. Familiarity with SLOs, SLIs, and SLA definitions and monitoring for service reliability. Experience with anomaly detection and root cause analysis (RCA) using observability data. Knowledge of cloud-native monitoring tools (e.g., AWS CloudWatch, Azure Monitor, GCP Operations Suite). Ability to build actionable dashboards and reports for technical and business stakeholders. Understanding of security and compliance monitoring within observability frameworks. Collaborative mindset to work with SREs, developers, and QA teams to define meaningful observability goals. Prepare and manage the IT budget, ensuring alignment with organizational priorities. Monitor expenditures and identify opportunities for cost savings without compromising quality. Well-spoken with good communication skills, as lot of stakeholder management is needed. What matters in this role : work experience : Bachelors or masters degree in computer science, Engineering, or related field. 8+ years of experience in DevOps or related roles, with at least 5 years in a leadership position. Strong hands-on experience with Apache Kafka (setup, tuning, monitoring, security). Proven experience with Kong API Gateway (plugins, routing, authentication, rate limiting). Proficiency in cloud platforms (AWS, Azure, or GCP). Kafka certification or Kong Gateway certification. Experience with service mesh technologies (e.g., Istio, Linkerd). Knowledge of event-driven architecture and microservices patterns. Experience with GitOps and Infrastructure as Code (IaC). Experience with containerization and orchestration (Docker, Kubernetes). Strong scripting skills (Bash, Python, etc.). Hands on with monitoring tools (Prometheus, Grafana, Mimir, ELK you should be comfortable with : Working from office : 5 days a week ( Sector 62, Noida) Pushing The Boundaries Have a big idea? See something that you feel we should do but havent done? We will hustle hard to make it happen. We encourage out of the box thinking, and if you bring that with you, we will make sure you get a bag that fits all the energy you bring along. What We Value In Our People You take the shot : You Decide Fast and You Deliver Right You are the CEO of what you do : you show ownership and make things happen You own tomorrow : by building solutions for the merchants and doing the right thing You sign your work like an artist : You seek to learn and take pride in the work you do (ref:hirist.tech)
Posted 4 days ago
15.0 years
0 Lacs
noida, uttar pradesh, india
Remote
Req ID: 340251 NTT DATA strives to hire exceptional, innovative and passionate individuals who want to grow with us. If you want to be part of an inclusive, adaptable, and forward-thinking organization, apply now. We are currently seeking a Industry Consulting Manager to join our team in Noida, Uttar Pradesh (IN-UP), India (IN). Job Description: Technical Architect – Observability & SRE Frameworks Position Title: Technical Architect – Observability & Site Reliability Engineering (SRE) Location: Noida, India Experience: 15+ years (including 5+ years in observability/SRE architecture) Employment Type: Full-time Role Overview We are looking for a highly experienced Technical Architect to lead the design, strategy, and implementation of Observability and SRE frameworks for enterprise-scale, microservices-based applications. The ideal candidate will bring deep technical knowledge of both Splunk Observability Stack and Open Source tools (like OpenTelemetry, Prometheus, Grafana, Jaeger), and be capable of defining and executing architecture strategies for complex distributed systems. This role requires hands-on ability to create architecture blueprints , lead technical teams, and work directly with stakeholders and platform owners to embed observability and reliability practices across the SDLC. Key Responsibilities Architecture & Blueprinting Design and deliver end-to-end observability architecture (Metrics, Logs, Traces, Events) for cloud-native and hybrid environments. Create technical architecture diagrams, data flow maps, and integration blueprints using tools like Lucidchart, Draw.io, or Visio. Lead the definition of SLIs, SLOs, and Error Budgets aligned with business KPIs and DORA metrics. Toolchain Strategy & Implementation Architect telemetry pipelines using OpenTelemetry Collector and Splunk Observability Cloud (SignalFx, APM, RUM, Log Observer). Define tool adoption strategy and integration roadmap for OSS tools (Prometheus, Loki, Grafana, Jaeger) and Splunk-based stacks. Guide teams on instrumentation approaches (auto/manual) across languages like Java, Go, Python, .NET, etc. Reliability Engineering Enablement Lead adoption of SRE principles including incident management frameworks, resiliency testing, and runbook automation. Collaborate with DevOps to integrate observability into CI/CD pipelines (e.g., Jenkins, ArgoCD, GitHub Actions). Define health checks, golden signals, and SPoG (Single Pane of Glass) dashboards. Exposure to AIOps, ML-based anomaly detection, or business observability. Stakeholder Management & Governance Serve as a technical liaison between client leadership, SREs, developers, and infrastructure teams. Run workshops, assessments, and evangelize observability-first culture across teams. Provide guidance on data retention, access control, cost optimization, and compliance (especially with Splunk ingestion policies). Performance & Optimization Continuously monitor and fine-tune observability data flows to prevent alert fatigue and ensure actionability. Implement root cause analysis practices using telemetry correlation across metrics, logs, and traces. Lead efforts to build self-healing systems using automated playbooks and AIOps integrations (where applicable). Required Skills & Qualifications 15+ years in IT, with 5 years in Observability/SRE architecture roles Proven experience designing architecture for microservices, containers (Docker, Kubernetes), and distributed systems Strong hands-on expertise with: Splunk Observability Cloud (SignalFx, Log Observer, APM) OpenTelemetry (SDKs + Collector) Prometheus + Grafana Jaeger / Zipkin for distributed tracing CI/CD tools: Jenkins, GitHub Actions, ArgoCD Ability to build and present clear architecture diagrams and solution roadmaps Working knowledge of cloud environments (AWS, Azure, GCP) and container orchestration (K8s/OpenShift) Familiarity with SRE and DevOps best practices (error budgets, release engineering, chaos testing) Nice to Have Splunk certifications: Core Consultant, Observability Specialist, Admin Knowledge of ITIL and modern incident management frameworks (PagerDuty, OpsGenie) Experience in banking or regulated enterprise environments Soft Skills Strong leadership and cross-functional collaboration Ability to work in ambiguous, fast-paced environments Excellent documentation and communication skills Passion for mentoring teams and building best practices at scale Why This Role Matters The client is on a journey to mature its Observability and SRE ecosystem , and this role will be critical in: Unifying legacy and modern telemetry stacks Driving reliability-first mindset and tooling Establishing a scalable blueprint for production excellence About NTT DATA NTT DATA is a $30 billion trusted global innovator of business and technology services. We serve 75% of the Fortune Global 100 and are committed to helping clients innovate, optimize and transform for long term success. As a Global Top Employer, we have diverse experts in more than 50 countries and a robust partner ecosystem of established and start-up companies. Our services include business and technology consulting, data and artificial intelligence, industry solutions, as well as the development, implementation and management of applications, infrastructure and connectivity. We are one of the leading providers of digital and AI infrastructure in the world. NTT DATA is a part of NTT Group, which invests over $3.6 billion each year in R&D to help organizations and society move confidently and sustainably into the digital future. Visit us at us.nttdata.com Whenever possible, we hire locally to NTT DATA offices or client sites. This ensures we can provide timely and effective support tailored to each client’s needs. While many positions offer remote or hybrid work options, these arrangements are subject to change based on client requirements. For employees near an NTT DATA office or client site, in-office attendance may be required for meetings or events, depending on business needs. At NTT DATA, we are committed to staying flexible and meeting the evolving needs of both our clients and employees. NTT DATA recruiters will never ask for payment or banking information and will only use @nttdata.com and @talent.nttdataservices.com email addresses. If you are requested to provide payment or disclose banking information, please submit a contact us form, https://us.nttdata.com/en/contact-us . NTT DATA endeavors to make https://us.nttdata.com accessible to any and all users. If you would like to contact us regarding the accessibility of our website or need assistance completing the application process, please contact us at https://us.nttdata.com/en/contact-us . This contact information is for accommodation requests only and cannot be used to inquire about the status of applications. NTT DATA is an equal opportunity employer. Qualified applicants will receive consideration for employment without regard to race, color, religion, sex, sexual orientation, gender identity, national origin, disability or protected veteran status. For our EEO Policy Statement, please click here . If you'd like more information on your EEO rights under the law, please click here . For Pay Transparency information, please click here .
Posted 5 days ago
7.0 years
0 Lacs
noida, uttar pradesh
Remote
Req ID: 340254 NTT DATA strives to hire exceptional, innovative and passionate individuals who want to grow with us. If you want to be part of an inclusive, adaptable, and forward-thinking organization, apply now. We are currently seeking a Industry Consulting Snr. Consultant to join our team in Noida, Uttar Pradesh (IN-UP), India (IN). Job Title: Telemetry Engineer Location: Noida, India Employment Type: Full-Time Experience Level: Senior (7+ years preferred) Role Overview: We are seeking a highly skilled Telemetry Engineer to lead the design and implementation of telemetry pipelines across diverse environments including microservices, VM-based applications, cloud-native platforms, and on-premise systems. The ideal candidate will have deep expertise in OpenTelemetry architecture and implementation , and a strong background in observability, distributed systems, and performance monitoring. Key Responsibilities: Architect and implement end-to-end telemetry pipelines for applications deployed across cloud, on-prem, and hybrid environments. Lead the installation, configuration, and optimization of OpenTelemetry components including SDKs, Collector, and exporters. Collaborate with application, infrastructure, and DevOps teams to define telemetry standards and integrate observability into CI/CD workflows. Design scalable and resilient data collection strategies for metrics, logs, and traces . Develop and maintain instrumentation libraries for microservices and legacy applications. Ensure telemetry data is efficiently routed to observability platforms (e.g., Splunk, Prometheus, Grafana, Datadog). Conduct performance tuning and troubleshooting of telemetry pipelines. Provide architectural guidance and best practices for telemetry adoption across teams. Stay current with OpenTelemetry releases and contribute to internal tooling and automation. Required Skills & Qualifications: Proven experience in setting up telemetry pipelines from scratch across multiple environments. Strong hands-on expertise with OpenTelemetry (Collector, SDKs, OTLP protocol). Deep understanding of distributed tracing, metrics collection, and log aggregation . Experience with observability platforms such as Splunk, Prometheus, Grafana, Jaeger, Zipkin, Datadog , etc. Proficiency in one or more programming languages (e.g., Python, Go, Java, Node.js) for instrumentation. Familiarity with cloud platforms (AWS, Azure, GCP) and VM/on-prem infrastructure . Knowledge of container orchestration (Kubernetes), service meshes (Istio), and CI/CD pipelines. Excellent communication and documentation skills. Preferred Qualifications: Experience contributing to or working with the OpenTelemetry community . Certifications in cloud technologies or observability tools. Experience in regulated or enterprise-scale environments. About NTT DATA NTT DATA is a $30 billion trusted global innovator of business and technology services. We serve 75% of the Fortune Global 100 and are committed to helping clients innovate, optimize and transform for long term success. As a Global Top Employer, we have diverse experts in more than 50 countries and a robust partner ecosystem of established and start-up companies. Our services include business and technology consulting, data and artificial intelligence, industry solutions, as well as the development, implementation and management of applications, infrastructure and connectivity. We are one of the leading providers of digital and AI infrastructure in the world. NTT DATA is a part of NTT Group, which invests over $3.6 billion each year in R&D to help organizations and society move confidently and sustainably into the digital future. Visit us at us.nttdata.com Whenever possible, we hire locally to NTT DATA offices or client sites. This ensures we can provide timely and effective support tailored to each client's needs. While many positions offer remote or hybrid work options, these arrangements are subject to change based on client requirements. For employees near an NTT DATA office or client site, in-office attendance may be required for meetings or events, depending on business needs. At NTT DATA, we are committed to staying flexible and meeting the evolving needs of both our clients and employees. NTT DATA recruiters will never ask for payment or banking information and will only use @nttdata.com and @talent.nttdataservices.com email addresses. If you are requested to provide payment or disclose banking information, please submit a contact us form, https://us.nttdata.com/en/contact-us. NTT DATA endeavors to make https://us.nttdata.com accessible to any and all users. If you would like to contact us regarding the accessibility of our website or need assistance completing the application process, please contact us at https://us.nttdata.com/en/contact-us. This contact information is for accommodation requests only and cannot be used to inquire about the status of applications. NTT DATA is an equal opportunity employer. Qualified applicants will receive consideration for employment without regard to race, color, religion, sex, sexual orientation, gender identity, national origin, disability or protected veteran status. For our EEO Policy Statement, please click here. If you'd like more information on your EEO rights under the law, please click here. For Pay Transparency information, please click here.
Posted 5 days ago
8.0 years
0 Lacs
noida, uttar pradesh, india
On-site
Job Description – Manager - DevOps We at Pine Labs are looking for those who share our core belief - “Every Day is Game day”. We bring our best selves to work each day to realize our mission of enriching the world through the power of digital commerce and financial services. Role Purpose We are seeking a Manager DevOps who will lead and manage the organizations DevOps Infrastructure, Observability stack for applications, CI-CD Pipeline and support services. This role involves managing a team of DevOps engineers, architecting scalable infrastructure, and ensuring high availability and performance of our messaging and API management systems. This individual will oversee a team of IT professionals, ensure the seamless delivery of IT services, and implement strategies to align technology solutions with business objectives. The ideal candidate is a strategic thinker with strong technical expertise and proven leadership experience. Responsibilities We Entrust You With Lead and mentor a team of DevOps Lead/Engineers in designing and maintaining scalable infrastructure. Architect and manage Kafka clusters for high-throughput, low-latency data streaming. Deploy, configure, and manage Kong API Gateway for secure and scalable API traffic management. Design and implement CI/CD pipelines for microservices and infrastructure. Automate infrastructure provisioning using tools like Terraform or Ansible. Monitor system performance and ensure high availability and disaster recovery. Collaborate with development, QA, and security teams to streamline deployments and enforce best practices. Ensure compliance with security standards and implement DevSecOps practices. Maintain documentation and provide training on Kafka and Kong usage and best practices. Strong understanding of observability pillars: metrics, logs, traces, and events. Hands-on experience with Prometheus for metrics collection and Grafana for dashboarding and visualization. Proficiency in centralized logging solutions like ELK Stack (Elasticsearch, Logstash, Kibana), Fluentd, or Splunk. Experience with distributed tracing tools such as Jaeger, Zipkin, or OpenTelemetry. Ability to implement instrumentation in applications for custom metrics and traceability. Skilled in setting up alerting and incident response workflows using tools like Alertmanager, PagerDuty, or Opsgenie. Familiarity with SLOs, SLIs, and SLA definitions and monitoring for service reliability. Experience with anomaly detection and root cause analysis (RCA) using observability data. Knowledge of cloud-native monitoring tools (e.g., AWS CloudWatch, Azure Monitor, GCP Operations Suite). Ability to build actionable dashboards and reports for technical and business stakeholders. Understanding of security and compliance monitoring within observability frameworks. Collaborative mindset to work with SREs, developers, and QA teams to define meaningful observability goals. Prepare and manage the IT budget, ensuring alignment with organizational priorities. Monitor expenditures and identify opportunities for cost savings without compromising quality. Well-spoken with good communication skills, as lot of stakeholder management is needed. What Matters In This Role Relevant work experience Bachelor’s or master’s degree in computer science, Engineering, or related field. 8+ years of experience in DevOps or related roles, with at least 5 years in a leadership position. Strong hands-on experience with Apache Kafka (setup, tuning, monitoring, security). Proven experience with Kong API Gateway (plugins, routing, authentication, rate limiting). Proficiency in cloud platforms (AWS, Azure, or GCP). Kafka certification or Kong Gateway certification. Experience with service mesh technologies (e.g., Istio, Linkerd). Knowledge of event-driven architecture and microservices patterns. Experience with GitOps and Infrastructure as Code (IaC). Experience with containerization and orchestration (Docker, Kubernetes). Strong scripting skills (Bash, Python, etc.). Hands on with monitoring tools (Prometheus, Grafana, Mimir, ELK Stack). Things You Should Be Comfortable With Working from office: 5 days a week ( Sector 62, Noida) Pushing the boundaries: Have a big idea? See something that you feel we should do but haven’t done? We will hustle hard to make it happen. We encourage out of the box thinking, and if you bring that with you, we will make sure you get a bag that fits all the energy you bring along. What We Value In Our People You take the shot: You Decide Fast and You Deliver Right You are the CEO of what you do: you show ownership and make things happen You own tomorrow: by building solutions for the merchants and doing the right thing You sign your work like an artist: You seek to learn and take pride in the work you do
Posted 6 days ago
5.0 - 10.0 years
0 Lacs
hyderabad, telangana
On-site
As a Java Enterprise Technical Architect, you will be responsible for designing and deploying scalable, high-performance microservices architecture. Your expertise in cloud computing, DevOps, security, and database optimization will be vital in ensuring the efficiency and security of enterprise applications. You will play a key role in fixing VAPT vulnerabilities, suggesting deployment architectures, and implementing clustering and scalability solutions. Your hands-on experience in coding with Java, Spring Boot, and microservices will be crucial in leading architecture decisions while ensuring best practices in software development. Your responsibilities will include designing secure and efficient deployment architectures, optimizing enterprise applications for security, and providing recommendations for cloud-native architectures on AWS, Azure, or GCP. You will also be responsible for fixing VAPT issues, implementing end-to-end security measures, and ensuring database security with encryption and access control mechanisms. Performance optimization, scalability, DevOps, and cloud enablement will be key focus areas where your expertise will be required. In addition to technical leadership and hands-on development, you will review and improve code quality, scalability, and security practices across development teams. You will mentor developers, conduct training sessions, and define architecture patterns and coding standards to ensure high-quality, scalable, and secure applications. Collaboration with stakeholders, evaluation of technologies, and staying updated with emerging trends will be essential in driving innovation and ensuring the security, performance, and reliability of system architecture. To qualify for this role, you should have 10+ years of hands-on experience in Java full-stack, Spring Boot, J2EE, and microservices, along with 5+ years of expertise in designing enterprise-grade deployment architectures. A strong security background, network design skills, and deep knowledge of application servers are required. Strong experience in database performance tuning, DevOps, cloud platforms, and containerization technologies will be necessary. Effective communication, problem-solving, and analytical skills will be essential to work closely with technical and non-technical stakeholders. A Bachelor's degree or Master's degree in computer science, Information Technology, or a related field is required to ensure a strong educational foundation for this role.,
Posted 1 week ago
5.0 - 9.0 years
0 Lacs
hyderabad, telangana
On-site
As a Senior Observability Engineer, you will play a crucial role in leading the design, development, and maintenance of observability solutions across our infrastructure, applications, and services. Your primary responsibility will be to implement cutting-edge monitoring, logging, and tracing solutions to ensure the reliability, performance, and availability of our complex, distributed systems. Collaboration with cross-functional teams, including Development, Infrastructure Engineers, DevOps, and SREs, will be essential to optimize system observability and enhance our incident response capabilities. Key Responsibilities: - Lead the Design & Implementation of observability solutions for cloud and on-premises environments, encompassing monitoring, logging, and tracing. - Drive the Development and maintenance of advanced monitoring tools such as Prometheus, Grafana, Datadog, New Relic, and AppDynamics. - Implement Distributed Tracing frameworks like OpenTelemetry, Jaeger, or Zipkin to enhance application performance diagnostics and troubleshooting. - Optimize Log Management and analysis strategies using tools like Elasticsearch, Splunk, Loki, and Fluentd for efficient log processing and insights. - Develop Advanced Alerting and anomaly detection strategies to proactively identify system issues and improve Mean Time to Recovery (MTTR). - Collaborate with Development & SRE Teams to enhance observability in CI/CD pipelines, microservices architectures, and various platform environments. - Automate Observability Tasks by leveraging scripting languages such as Python, Bash, or Golang to increase efficiency and scale observability operations. - Ensure Scalability & Efficiency of monitoring solutions to manage large-scale distributed systems and meet evolving business requirements. - Lead Incident Response by providing actionable insights through observability data for effective troubleshooting and root cause analysis. - Stay Abreast of Industry Trends in observability, Site Reliability Engineering (SRE), and monitoring practices to continuously improve processes. Required Qualifications: - 5+ years of hands-on experience in observability, SRE, DevOps, or related fields, with a proven track record in managing complex, large-scale distributed systems. - Expert-level proficiency in observability tools such as Prometheus, Grafana, Datadog, New Relic, AppDynamics, and the ability to design and implement these solutions at scale. - Advanced experience with log management platforms like Elasticsearch, Splunk, Loki, and Fluentd, optimizing log aggregation and analysis for performance insights. - Deep expertise in distributed tracing tools like OpenTelemetry, Jaeger, or Zipkin, focusing on performance optimization and root cause analysis. - Extensive experience with cloud environments (Azure, AWS, GCP) and Kubernetes for deploying and managing observability solutions in cloud-native infrastructures. - Advanced proficiency in scripting languages like Python, Bash, or Golang, and experience with Infrastructure as Code (IaC) tools such as Terraform and Ansible. - Strong understanding of system architecture, performance tuning, and troubleshooting production environments with scalability and high availability in mind. - Proven leadership experience and the ability to mentor teams, provide technical direction, and drive best practices for observability and monitoring. - Excellent problem-solving skills, emphasizing actionable insights and data-driven decision-making. - Ability to lead high-impact projects, communicate effectively with stakeholders, and influence cross-functional teams. - Strong communication and collaboration skills, working closely with engineering teams, leadership, and external partners to achieve observability and system reliability goals. Preferred Qualifications: - Experience with AI-driven observability tools and anomaly detection techniques. - Familiarity with microservices, serverless architectures, and event-driven systems. - Proven track record of handling on-call rotations and incident management workflows in high-availability environments. - Relevant certifications in observability tools, cloud platforms, or SRE best practices are advantageous.,
Posted 2 weeks ago
5.0 years
12 - 15 Lacs
jaipur, rajasthan, india
On-site
Responsibilities Lead Development: Lead and mentor a team of developers in designing, developing, and deploying highly scalable, performant, and secure microservices-based applications using NestJS, Laravel, and Next.js. System Architecture & Microservices Design: Architect and implement microservices solutions, ensuring clear separation of concerns, high scalability, and maintainability. API Design: Design and develop robust RESTful and GraphQL APIs that connect and communicate between microservices. Integration: Implement seamless integration patterns between microservices, leveraging message brokers (e.g., RabbitMQ, Kafka) and APIs. Cloud & Containerization: Build and deploy microservices in cloud environments (AWS, Azure, GCP) with containerized solutions (e.g., Docker, Kubernetes). Code Reviews & Mentorship: Conduct regular code reviews, ensuring adherence to best practices, high-quality code standards, and fostering a culture of continuous learning. Collaboration: Work cross-functionally with Product Managers, UI/UX designers, and other stakeholders to ensure successful delivery of scalable applications. Performance Optimization: Implement strategies to optimize microservices, ensuring high availability, low latency, and efficient resource usage across all components. CI/CD & DevOps: Manage and optimize continuous integration and deployment pipelines, automate deployment processes, and ensure smooth updates across microservices. Technical Documentation: Maintain clear and concise documentation for system design, APIs, microservice interactions, and deployment processes. Requirements Proven Experience: 5+ years of experience in software development with at least 2+ years of leadership experience in a senior engineering role. Backend Technologies: Deep knowledge of NestJS (Node.js framework) and Laravel (PHP framework) with a strong understanding of backend systems. Frontend Technologies: Solid experience with Next.js for building performant, server-side rendered React applications. Microservices Architecture: Hands-on experience designing and building microservices with a focus on scalability, fault tolerance, and distributed systems. API Design & Integration: Expertise in building and maintaining scalable RESTful APIs and GraphQL services that interconnect microservices. Containerization & Cloud: Proficiency with Docker, Kubernetes, and deploying services on cloud platforms (AWS, Azure, or GCP). Database Design: Experience with both relational databases (e.g., MySQL, PostgreSQL) and NoSQL databases (e.g., MongoDB) in a microservices context. Version Control: Expertise in using Git, GitHub, or GitLab for version control and collaboration. CI/CD: Experience in automating deployment pipelines using tools like Jenkins, GitLab CI, or CircleCI. Leadership & Communication: Strong leadership skills, with the ability to manage, guide, and motivate teams. Excellent written and verbal communication skills. Nice To Have Familiarity with event-driven architecture and tools like Kafka, RabbitMQ, or NATS. Experience with Serverless architecture and frameworks (e.g., AWS Lambda). Understanding of service mesh frameworks like Istio or Linkerd. Experience with distributed tracing tools (e.g., Jaeger, Zipkin). Experience in automated testing frameworks (Jest, Mocha, PHPUnit, etc.). Knowledge of modern frontend technologies like TypeScript, Tailwind CSS, and GraphQL. Skills: laravel,next.js,nestjs,php,microservices architecture,microservices,aws,aws lambda
Posted 2 weeks ago
5.0 - 9.0 years
0 Lacs
karnataka
On-site
The ideal candidate for the position of Java + Observability Engineer (Apache Storm) in Bengaluru should be a skilled professional with experience and expertise in Java programming and observability tools, particularly Apache Storm. As an Architect level engineer, you will be responsible for designing and building scalable systems, utilizing your knowledge in Java (Version 8/11 or higher), observability tools such as Prometheus, Grafana, OpenTelemetry, ELK, Jaeger, Zipkin, New Relic, containerization technologies like Docker and Kubernetes, CI/CD pipelines, and hands-on experience with Apache Storm. This is a unique opportunity for individuals who are passionate about modern observability stacks and distributed systems, and are looking to make a significant impact in a challenging role. The role is based in Bengaluru and is only open to candidates who can join immediately. If you meet the key skill requirements and are prepared to take on this architect-level position, we invite you to apply by sending your resume to sushil@saisservices.com.,
Posted 2 weeks ago
3.0 - 7.0 years
0 Lacs
noida, uttar pradesh
On-site
An extraordinarily talented group of individuals work together every day to drive TNS" success, from both professional and personal perspectives. Come join the excellence! TNS is looking for an Observability Engineer to support the design, implementation, and evolution of our observability stack. This role is critical in ensuring the reliability, performance, and scalability of our systems by providing deep visibility into infrastructure and application behavior. You will collaborate with cross-functional teams to define observability standards and drive adoption of best practices across the organization. Lead the design, implementation, and continuous improvement of the observability stack, including monitoring, logging, and tracing systems. Define and enforce observability standards and best practices across engineering teams to ensure consistent instrumentation and visibility. Build scalable monitoring solutions that provide real-time insights into system health, performance, and availability. Develop and maintain dashboards, alerts, and automated responses to proactively detect and resolve issues before they impact users. Collaborate with development, infrastructure, and SRE teams to integrate observability into CI/CD pipelines and production workflows. Conduct root cause analysis and post-incident reviews to identify observability gaps and drive improvements. Evaluate and implement tools such as Splunk, Splunk Observability Cloud, Netreo to support monitoring and alerting needs. Champion a culture of data-driven decision-making by enabling teams to access and interpret observability data effectively. Automating observability pipelines and alerting mechanisms. Qualifications 5+ years of experience in Site Reliability Engineering, DevOps, or Observability roles. 3+ years of experience in SRE/DevOps. Demonstrated success in deploying and managing monitoring tools and observability solutions at scale. Hands-on experience with monitoring and observability platforms such as Splunk, Splunk Observability Cloud (O11y), Grafana, Prometheus, Datadog. Proven ability to design and implement SLOs/SLIs, dashboards, and alerting strategies that align with business and operational goals. Familiarity with incident response, alert tuning, and postmortem analysis. Strong scripting or programming skills (e.g., Python, Go, Bash). Excellent communication and collaboration skills, with a focus on knowledge sharing and mentorship. Desired Strong understanding of distributed tracing tools like OpenTelemetry, Jaeger, or Zipkin. Experience integrating observability into CI/CD pipelines and Kubernetes environments. Contributions to open-source observability tools or frameworks. Strong understanding of distributed tracing tools like OpenTelemetry, Jaeger, or Zipkin. Strong knowledge of cloud platforms (AWS, Azure, or GCP) and container orchestration (Kubernetes). If you are passionate about technology, love personal growth and opportunity, come see what TNS is all about!,
Posted 2 weeks ago
10.0 years
0 Lacs
new delhi, delhi, india
On-site
Job description For IMMEDIATE JOINER on Part Time contract basis. Job description Experience: 10+ years. We are looking for a Java Tech Lead cum Architect to provide sound technical leadership in all aspects of our business. You will communicate with employees, stakeholders and customers to ensure our company's technologies are used appropriately. As a tech lead, you will be responsible for developing, overseeing, improving and upgrading enterprise services and software. You will also need to stay on top of the latest trends and technologies and keep an eye out for any software, services that might improve business processes. Strategic thinking and strong business acumen are essential in this role. We expect you to be well-versed in current technological trends and familiar with a variety of business concepts. If you are also an excellent communicator and public speaker, wed like to meet you. Responsibilities Develop technical aspects of the company's strategy to ensure alignment with its business goals Discover and implement new technologies that yield competitive advantage Help departments use technology profitably Supervise system infrastructure to ensure functionality and efficiency Build quality assurance and data protection processes Monitor KPIs and IT budgets to assess technological performance Use stakeholders feedback to inform necessary improvements and adjustments to technology Communicate technology strategy to partners and investors Requirements Skillfully write high-quality, well-tested code and should be comfortable with Object-Oriented programming Presence in the external tech community: can proactively share his expertise with others via speaking engagements, contributions to open source, blogs and more Comfortability with Agile methods, such as Extreme Programming (XP), Scrum and/or Kanban Resource should enjoy influencing others and always advocate for technical excellence while being open to change when needed. Resource should resilient in ambiguous situations and can approach challenges from multiple perspectives. Working knowledge of cloud technology such as AWS, Azure, Kubernetes and Docker is an added advantage. Proven experience as a TL or similar leadership role Knowledge of technological trends to build strategy Developing the Technical Architecture based on Micro services. Developing and coordinating Architecture Plans In hand coding experience of Java, Golang, Python. Java, JEE/J2EE Core Java 11, JDBC 4; Servlet 3; Web Services (SOAP, REST), JAX-WS, JAX-RS; EJB 3; Struts 1.2 & 2; Groovy and Grails; Spring (Core, Bean, Context, Security, AOP, ORM, REST, MVC, Repository, Transaction and Propagation); JPA 2, Hibernate 3, Spring Data, iBatis, GORM; Java Core Design Pattern and JEE Design Pattern. Microservices Spring Boot, Spring Cloud (Config, Netflix (Eureka, Hystrix, Zuul, Archaius), Cloud foundry, Sleuth (Zipkin, HTrace), Kafka, Zookeeper); Docker, Kubernetes, Cassandra, Redis Cluster, Elastic Search, GCP, GITHUB Deisgn and Development experience on PostgreSQL 8, MongoDB, MySQL. Sound knowledge of API design tools RAML, Swagger, Sandbox and Mulesoft Anypoint Studio. Understanding of budgets and business-planning Ability to conduct technological analyses and research Excellent communication skills Leadership and organizational abilities Strategic thinking Problem-solving aptitude Btech/BE in Computer Science, Engineering or a related field; M-tech or other relevant post graduate degree is a plus. Note : leading a team of Java developers, designing and implementing Java-based applications, providing technical guidance, ensuring project quality, and mentoring junior developers is core requirement of this position.
Posted 3 weeks ago
6.0 years
0 Lacs
bengaluru
On-site
Are you ready to power the World's connections? If you don’t think you meet all of the criteria below but are still interested in the job, please apply. Nobody checks every box - we’re looking for candidates that are particularly strong in a few areas, and have some interest and capabilities in others. About the role: Kong is building the future of API management for developers, empowering organisations to connect their services and systems with speed, security, and reliability. As a fast-growing and well-funded company, we take pride in our happy customers, motivated employees, and innovative products that shape the API landscape. As a Staff Product Manager for Kong’s observability offerings, you will lead the charge to build out and enhance Kong’s existing observability and tracing offerings. Observability is a key reason why customers choose Kong and Konnect, and the roadmap you drive will be critical to Kong’s growth and success. You will collaborate closely with a talented and focused engineering team while maintaining 360-degree ownership of the product, engaging with engineering, field teams, other product managers, and customers to deliver value. At Kong, your contributions will have an outsized impact, as our small and agile environment amplifies the influence of every team member. What you’ll do: Define and Drive Product Vision : Collaborate with cross-functional teams, including engineering, design, and field teams, to establish and execute a clear vision for Kong’s Event Gateway, ensuring alignment with customer needs and long-term production plans. Lead Execution : Partner with engineering and design leaders to scope, define, and deliver capabilities for the product’s releases that bring customer value and support the promotion of Kong Event Gateway as a major feature. Customer-Focused Planning : Develop a deep understanding of customer needs and use cases, supported by solid data and evidence. Translate this understanding into artefacts outlining target use cases, milestone plans, and timelines to guide the product roadmap. Ensure Cross-Functional Alignment : Work closely with engineering, design, and marketing teams to align on use case execution, milestone delivery, and launch plans. Foster collaboration and ensure clarity in planning and execution across all relevant stakeholders. Develop Roadmaps and Plans : Create and maintain a comprehensive external-facing roadmap and related materials that communicate the product direction and milestones. Collaborate with Product Marketing Management (PMM) to develop and execute a clear launch plan, ensuring all deliverables are approved and on track. Champion Communication and Transparency : Serve as the central point of communication, keeping internal and external stakeholders informed on progress, milestones, and expectations. Build strong relationships with field teams to ensure project updates and outcomes are effectively communicated. What you’ll bring: Experience: 6+ years in product management, working on developer infrastructure products and working with technical customers. Deep understanding of customer needs, onboarding, and product adoption. An understanding and sympathy for how engineering, in both small startups and large enterprises, are typically organised and operates. Knowledge of common tools and patterns in the Observability ecosystem, such as Prometheus, Grafana, Zipkin/OpenTracing. Understand the impact of sampling and cardinality on the performance and cost of metrics management. Effectively engaging with stakeholders, identifying core pain points, and articulating a clear product vision. Execution-Driven #LI-KS1 About Kong: Kong Inc., a leading developer of cloud API technologies, is on a mission to enable companies around the world to become “API-first” and securely accelerate AI adoption. Kong helps organizations globally — from startups to Fortune 500 enterprises — unleash developer productivity, build securely, and accelerate time to market. For more information about Kong, please visit www.konghq.com or follow us on X @thekonginc.
Posted 3 weeks ago
5.0 years
0 Lacs
chennai, tamil nadu, india
On-site
Job Title: Senior Java Spring Boot Developer Experience: 5+ Years Location: Chennai Job Description: We are seeking a passionate and experienced Java Spring Boot Developer with a strong background in designing and implementing scalable microservices using modern cloud-native technologies. The ideal candidate should have at least 5 years of experience in Java development, Spring ecosystem, and be enthusiastic about writing clean, maintainable code. Key Responsibilities: Design, develop, test, and deploy scalable microservices using Java Spring Boot . Integrate and manage MongoDB , Hibernate , and other persistence frameworks. Develop and deploy services in a cloud-native environment using Spring Cloud . Implement CI/CD pipelines and deploy microservices on OpenShift and Kubernetes . Collaborate with cross-functional teams to understand business requirements and deliver high-quality technical solutions. Monitor application performance using tools like Prometheus , Grafana , Zipkin , Kibana , FluentD , and Hystrix . Participate in code reviews, contribute to architectural decisions, and mentor junior developers. Continuously improve system performance through tuning and optimization. Stay updated with the latest industry trends, tools, and practices. Required Skills: Strong expertise in Java (8/11+) and Spring Boot. In-depth knowledge of Spring Cloud , REST APIs , and Microservices Architecture . Experience with MongoDB , Hibernate , and JPA. Proficient in Docker , Kubernetes , and OpenShift for container orchestration and deployment. Hands-on experience in performance tuning , debugging, and application monitoring . Familiarity with observability tools like Prometheus, Grafana, Zipkin, FluentD, Hystrix, Kibana. Solid understanding of DevOps principles and CI/CD pipelines. Excellent problem-solving, communication, and collaboration skills. Strong passion for clean, high-quality, maintainable code. Preferred Qualifications: Experience with other NoSQL databases. Exposure to reactive programming or event-driven systems. Certifications in cloud technologies (AWS/GCP/Azure/OpenShift) are a plus.
Posted 3 weeks ago
0 years
0 Lacs
chennai, tamil nadu, india
On-site
The Engineering Sr Lead Analyst is a senior management level position responsible for accomplishing results through the management of a team or department in an effort to drive a variety of engineering activities including the design, acquisition and deployment of hardware, software and network infrastructure in coordination with the Technology team. The overall objective of this role is to lead efforts to ensure quality standards are being met within existing and planned framework. Responsibilities: Design and Implement Open Telemetry Observability Platform. Cloud Integration: Ensure seamless integration with public cloud providers, enabling E2E business transaction from on-Prem to Public cloud. Stakeholder Collaboration, Senior Leadership Communication, Best Practices and Standards. Support all products within functional area, from design to implementation and ongoing performance according to business need Identify, establish and maintain strategic partnerships with executives from technology vendors within assigned area and coordinate to influence vendor product roadmaps to be in line with Citi strategic goals Ensure that Citi makes appropriate strategic choices in the onboarding, management and exiting of vendors from the product portfolio Ensure there is adequate capital budget to develop and implement the assigned strategic technology roadmap Provide technical leadership in the design of highly complex systems Develop and advocate for new solutions to undefined system needs, where no solution is available within the technical community Impact the engineering function by contributing to technical direction and strategic decisions Ensure products are ready and adhere to global operational standards for ongoing operation and support Mentor junior team members and allocate work Exhibit knowledge of how own specialty area contributes to the business and apply knowledge of competitors, products and services Appropriately assess risk when business decisions are made, demonstrating particular consideration for the firm's reputation and safeguarding Citigroup, its clients and assets, by driving compliance with applicable laws, rules and regulations, adhering to Policy, applying sound ethical judgment regarding personal behavior, conduct and business practices, and escalating, managing and reporting control issues with transparency. Deep understanding of Open Telemetry observability principles and practices (metrics, logs, tracing). Extensive hands-on experience with various observability tools and platforms such as: APM: Dynatrace, New Relic, AppDynamics Logging: Splunk, Elastic Stack (Elasticsearch, Logstash, Kibana), Sumo Logic, Azure Log Analytics, AWS CloudWatch Logs Metrics & Monitoring: Prometheus, Grafana, TICK stack (Telegraf, InfluxDB, Chronograf, Kapacitor), VictoriaMetrics, Zabbix, Nagios Tracing: Jaeger, Zipkin, OpenTelemetry Familiarity with cloud platforms (AWS, Azure, GCP) and their native monitoring services (e.g., CloudWatch, Azure Monitor, Google Cloud Monitoring). Experience with containerization and orchestration technologies (Docker, Kubernetes) and monitoring in these environments. Proficiency in Infrastructure-as-Code (IaC) tools like Terraform and Ansible for managing observability infrastructure. Experience with scripting and programming languages such as Python, Bash, and potentially Java or Go for automation and custom tooling. Understanding of network protocols, system administration, and incident response processes. Qualifications: 10+ of relevant experience in an Engineering role Experience working in Financial Services or a large complex and/or global environment Project Management experience Consistently demonstrates clear and concise written and verbal communication Comprehensive knowledge of design metrics, analytics tools, benchmarking activities and related reporting to identify best practices Demonstrated analytic/diagnostic skills Ability to work in a matrix environment and partner with virtual teams Ability to work independently, multi-task, and take ownership of various parts of a project or initiative Ability to work under pressure and manage to tight deadlines or unexpected changes in expectations or requirements Proven track record of operational process change and improvement Education: Bachelor’s degree/University degree or equivalent experience Master’s degree preferred ------------------------------------------------------ Job Family Group: Technology ------------------------------------------------------ Job Family: Systems & Engineering ------------------------------------------------------ Time Type: Full time ------------------------------------------------------ Most Relevant Skills Please see the requirements listed above. ------------------------------------------------------ Other Relevant Skills For complementary skills, please see above and/or contact the recruiter. ------------------------------------------------------ Citi is an equal opportunity employer, and qualified candidates will receive consideration without regard to their race, color, religion, sex, sexual orientation, gender identity, national origin, disability, status as a protected veteran, or any other characteristic protected by law. If you are a person with a disability and need a reasonable accommodation to use our search tools and/or apply for a career opportunity review Accessibility at Citi. View Citi’s EEO Policy Statement and the Know Your Rights poster.
Posted 3 weeks ago
5.0 - 10.0 years
10 - 15 Lacs
bengaluru
Work from Office
Roles & Responsibilities : • .NET Instrumentation: • Implement and maintain application instrumentation using OpenTelemetry SDK for distributed tracing, metrics, and logs across .NET applications (Core and Framework). Auto-Instrumentation: • Evaluate, integrate, and enhance auto-instrumentation capabilities using OpenTelemetry .NET auto-instrumentation libraries. Build custom instrumentation where necessary. • SDK & Framework Compatibility: • Ensure compatibility across various .NET runtimes and versions (.NET Core, .NET Framework, .NET 5/6/7/8+). Test and validate behavior across different environments. • Performance & Overhead Analysis: • Tune instrumentation code to minimize performance overhead. Contribute to best practices for high-efficiency observability implementations. • Tooling & Integration: • Integrate instrumentation with backend systems such as Jaeger, Prometheus, Grafana, OTLP collectors, and internal observability platforms. • Documentation & Developer Enablement: • Create clear documentation, sample apps, and templates to help internal teams adopt observability standards and SDKs with ease. • Collaboration: • Work with cross-functional teams including application developers, DevOps, and platform engineers to understand instrumentation needs and guide adoption. Mandatory Skills: • Minimum 3+ years software development experience with a focus on .NET/.NET Core (C#). • Deep understanding of .NET application internals, runtime behavior, diagnostics APIs, and async programming patterns. • Hands-on experience with OpenTelemetry SDK and concepts like spans, traces, metrics, and logs. • Experience with auto-instrumentation and diagnostics source/event listener patterns in .NET. • Familiarity with telemetry backends: Jaeger, Zipkin, OTEL Collector, Prometheus, or cloud-native observability stacks. • Strong understanding of application performance, observability concepts (golden signals), and modern microservices patterns. • Experience with CI/CD pipelines and containerized environments (Docker, Kubernetes) is a plus Good to Have Skills • Contributions to OpenTelemetry or other open-source instrumentation projects. • Experience developing reusable .NET libraries or SDKs. • Familiarity with other programming languages (Java, Python, Node.js) and crosslanguage trace correlation. • Exposure to distributed systems, message queues, databases, and cloud services.
Posted 3 weeks ago
8.0 years
0 Lacs
noida, uttar pradesh, india
On-site
Our Company Changing the world through digital experiences is what Adobe’s all about. We give everyone—from emerging artists to global brands—everything they need to design and deliver exceptional digital experiences! We’re passionate about empowering people to create beautiful and powerful images, videos, and apps, and transform how companies interact with customers across every screen. We’re on a mission to hire the very best and are committed to creating exceptional employee experiences where everyone is respected and has access to equal opportunity. We realize that new ideas can come from everywhere in the organization, and we know the next big idea could be yours! Our Company Changing the world through digital experiences is what Adobe’s all about. We give everyone—from emerging artists to global brands—everything they need to design and deliver exceptional digital experiences. We’re on a mission to hire the very best and are committed to building exceptional employee experiences where everyone is respected and has access to equal opportunity. We realize that new insights can come from everywhere, and we know the next big idea could be yours! At Adobe, you will be immersed in an exceptional work environment that is recognized around the world. The Opportunity Structured Content is at the heart of the engines powering the new age Experiences such as Chatbots, Voice Based Devices, Fluid and Omni-Channel content delivery. In Technical Communication group of Adobe, we are developing a new age Component Content Management System which powers the journey of Structured Content for large enterprises. Scale and efficiency are key here. Millions of documents are published regularly to multiple platforms and channels across the industry using this solution. We have a strong vision and we are looking for a highly motivated, technically driven and hands on leader to realize it for our product. About the Team: AEM Guides is a new age technology CCMS. It is used by Fortune-500 companies which publish millions of documents regularly using this product. It is a unique opportunity to work in a startup like environment within a large organization where all product functions collaborate closely with the business counterparts and with the large enterprise customers. Given the enterprise business domain and the startup nature of the team, we are growing fast and scaling the product scope and customer base at a very rapid pace. The Challenge As a Senior Computer Scientist, you will go beyond traditional coding responsibilities to lead and shape complex features within our systems. Your role will focus on: Architecting Scalable Solutions: Design and implement features that integrate seamlessly with our broader system architecture, leveraging Adobe Experience Manager capabilities to manage customer workflows, and drive actionable insights. Complex Systems Integration: Ensure that new features interact effectively with existing components, maintaining high performance and reliability. Advanced Algorithmic Design: Develop and optimize algorithms to solve complex problems, applying sophisticated design principles to enhance system functionality. Strategic Design Judgment: Make informed, high-level design decisions that align with long-term product goals and architectural standards. Product and Technology Expertise: Stay ahead of the emerging technologies, using this knowledge to drive innovation and continuously improve our offerings. We are looking for passionate and driven senior architects who can translate intricate product features into scalable, efficient solutions. Your expertise in architectural decision-making and team mentoring will be crucial to our success and innovation in the marketing technology space. Roles & Responsibilities: This is an individual contributor position. Expectations will be on the below lines: Responsible for design and architecture of new services and features Well versed in emerging industry technologies and trends, and the ability to channel that knowledge to the team and use it to influence product direction. Be responsible for all phases of engineering. From early specs, design/architecture, technology choice, development, unit-testing/integration automation, and deployment. Collaborate with architects, product management and other engineering teams to build the services and product features Build technical specifications, prototypes and presentations to communicate your ideas. Participate in resolution of production issues and develop solutions to prevent future issues from happening again Orchestrate with team to develop a product or parts of a large product. Required Skills & Expertise: 8+ years of experience in technical roles, proven experience across product life cycle. Well versed with microservices architecture, cloud-based web services architecture, design patterns and frameworks Experience in scaling and leading teams building solution with cloud technologies Should have excellent computer science fundamentals and a good understanding of design, and performance of algorithms Mastery of Java SE (Java 8+), including functional programming, streams, lambdas, and concurrency APIs. Proficiency in designing and developing RESTful APIs and GraphQL. Excellent database fundamentals and hands on with MySQL/Postgres/ MongoDB Understanding of API versioning, security (OAuth 2.0, JWT), and documentation using tools like Swagger/OpenAPI. Knowledge of AWS and/or Azure, Kubernetes, Docker, Jenkins, Splunk Knowledge of application security best practices (e.g., OWASP Top Ten) Experience with implementing encryption, secure communication protocols (TLS/SSL), and vulnerability management Strong ability to partner across organizations and divisions and presents an opportunity to define and contribute to solving some of the most difficult problems Basic understanding of UI/UX design and development is a plus. Ability to build consensus & drive decisions in ambiguous scenarios Excellent work ethics and highly motivated Excellent oral and written communication skills (interpersonal and client-facing) Ability to manage systems development scope and changes in the context of the business environment. Minimum of a Bachelor's degree or equivalent in Computer Science, Information Technology, Engineering, or related field. Architectural Skills Scalable Design Patterns: Proficiency in applying design patterns (e.g., Singleton, Factory, Strategy, and Observer) and architectural patterns like CQRS and Domain-Driven Design (DDD). Integration and Interoperability: Experience with integrating third-party services, SDKs, and APIs. Knowledge of data streaming and batch processing frameworks (e.g., Apache Flink, Apache Spark). Monitoring and Observability: Familiarity with monitoring tools like Prometheus, Grafana, and New Relic. Experience with distributed tracing tools like Jaeger or Zipkin Code Reviews & Standards: Skilled in conducting thorough code reviews and enforcing best practices. Data Management: Proficiency in handling large-scale data processing and ensuring data consistency across distributed systems. Knowledge of caching mechanisms (e.g., Redis, Memcached) for optimized performance Adobe is an equal opportunity employer. We welcome and encourage diversity in the workplace regardless of race, gender, religion, age, sexual orientation, gender identity, disability or veteran status. We ensure that individuals with disabilities are provided reasonable accommodation to participate in the job application or interview process, to perform crucial job functions, and to receive other benefits and privileges of employment. Please contact us to request accommodation. Other Info: Adobe Fast Facts: https://blogs.adobe.com/adobelife/category/apac/ Life@Adobe Blog: https://blogs.adobe.com/adobelife/category/apac/ Adobe Corporate Social Responsibility: http://www.adobe.com/corporateresponsibility/ Adobe Culture and Benefits: https://benefits.adobe.com/in Adobe Investor Relations: http://www.adobe.com/aboutadobe/invrelations Discover Check-In: https://www.adobe.com/check-in.html Adobe is proud to be an Equal Employment Opportunity employer. We do not discriminate based on gender, race or color, ethnicity or national origin, age, disability, religion, sexual orientation, gender identity or expression, veteran status, or any other applicable characteristics protected by law. Learn more. Adobe aims to make Adobe.com accessible to any and all users. If you have a disability or special need that requires accommodation to navigate our website or complete the application process, email accommodations@adobe.com or call (408) 536-3015.
Posted 3 weeks ago
10.0 years
0 Lacs
noida, uttar pradesh, india
On-site
Our Company Changing the world through digital experiences is what Adobe’s all about. We give everyone—from emerging artists to global brands—everything they need to design and deliver exceptional digital experiences! We’re passionate about empowering people to create beautiful and powerful images, videos, and apps, and transform how companies interact with customers across every screen. We’re on a mission to hire the very best and are committed to creating exceptional employee experiences where everyone is respected and has access to equal opportunity. We realize that new ideas can come from everywhere in the organization, and we know the next big idea could be yours! Adobe Document Cloud is transforming how people create, manage, and share documents. With a rapidly expanding subscription base, groundbreaking innovations, and cutting-edge features, this is an exciting time to join our team. We operate in agile, collaborative environments with significant autonomy, valuing technical expertise, creativity, adaptability, customer focus, and a passion for continuous learning. By empowering users to maximize the potential of their PDFs and documents, we enable millions of creatives, business professionals, and marketers to achieve success every day. This role requires exceptional technical, communication, coding, problem-solving, design, and collaboration skills. The ideal candidate is an accomplished professional with extensive development experience and a strong understanding of system architecture. If you are driven by curiosity, problem-solving, and a passion for technology and innovation, this position is a perfect fit. Candidates should demonstrate expertise in most or all of the qualifications listed below. Preferred Qualifications Education: BE/B.Tech/MCA/M.Tech or equivalent. Proficiency in at least one programming language such as Java or Python, with minimum 10+ years of professional experience. Hands-on experience with databases like MySQL, PostgreSQL, MongoDB, etc. Knowledge of caching mechanisms (e.g., Redis, Memcached) for optimized performance. Familiarity with monitoring tools like Prometheus, Grafana, and New Relic. Experience with distributed tracing tools like Jaeger or Zipkin Conceptualization to Execution: Ability to translate concepts into actionable plans, design, and development. Web API Proficiency: Familiarity with web API calls and their implementation. Previous experience as a full-stack engineer is a plus. Proven expertise in deploying large-scale, cloud-hosted applications, including automation of deployment processes. A strong understanding of security principles, is reflected in your work. Hands-on experience with Linux-based platforms, storage systems, networking, load balancers, and cloud environments like Azure or AWS. Proficiency in using modern technologies like AI tools (e.g., Copilot) to boost productivity. Excellent written and verbal communication skills, capable of conveying technical concepts effectively to both technical and non-technical audiences. Strong documentation practices and proficiency with tools like wikis, blogs, and other collaborative platforms. Familiarity with service delivery and project management principles. Ability to prioritize tasks effectively, working both independently and collaboratively as part of a team. Additional Preferred Experience Supporting large-scale SaaS-based applications and databases. Familiarity with agile software development methodologies, including software builds and source code control. Internal Opportunities Creativity, curiosity, and constant learning are celebrated aspects of your career growth journey. We're glad that you're pursuing a new opportunity at Adobe! Put your best foot forward: Update your Resume/CV and Workday profile – don't forget to include your uniquely 'Adobe' experiences and volunteer work. Visit the Internal Mobility page on Inside Adobe to learn more about the process and set up a job alert for roles you're interested in. Check out these tips to help you prep for interviews. If you are applying for a role outside of your current country, ensure you review the International Resources for Relocating Employees on Inside Adobe, including the impacts to your Benefits, AIP, Equity & Payroll. Once you apply for a role via Workday, the Talent Team will reach out to you within 2 weeks. If you move into the official interview process with the hiring team, make sure you inform your manager so they can champion your career growth. Why Adobe? At Adobe, you will be immersed in an exceptional work environment that is recognized around the world. You will also be surrounded by colleagues who are committed to helping each other grow through our unique Check-In approach where ongoing feedback flows freely. If you're looking to make an impact, Adobe's the place for you. Discover what our employees are saying about their career experiences on the Adobe Life blog and explore the meaningful benefits we offer. Adobe is an equal opportunity and affirmative action employer. We welcome and encourage diversity in the workplace regardless of gender, race or color, ethnicity or national origin, age, disability, religion, sexual orientation, gender identity or expression, veteran status, or any other characteristics protected by law. If you have a disability or special need that requires accommodation to navigate our internal careers site or to complete the application process, please contact accommodations@adobe.com. Adobe is proud to be an Equal Employment Opportunity employer. We do not discriminate based on gender, race or color, ethnicity or national origin, age, disability, religion, sexual orientation, gender identity or expression, veteran status, or any other applicable characteristics protected by law. Learn more. Adobe aims to make Adobe.com accessible to any and all users. If you have a disability or special need that requires accommodation to navigate our website or complete the application process, email accommodations@adobe.com or call (408) 536-3015.
Posted 3 weeks ago
6.0 years
10 - 28 Lacs
Hyderabad, Telangana, India
On-site
Company: Hashone Careers Website: Visit Website Business Type: Startup Company Type: Product Business Model: B2B Funding Stage: Pre-seed Salary Range: ₹ 10-28 Lacs PA Job Description 🚀 Job Overview We are seeking an experienced Lead DevOps Engineer with deep expertise in Kubernetes infrastructure design and implementation. This role requires an individual who can architect, build, and manage enterprise-grade Kubernetes clusters from the ground up. The position offers an exciting opportunity to lead infrastructure modernization initiatives and work with cutting-edge cloud-native technologies. Initial Setup Phase: First 2-3 months will be based in Mumbai for project initiation and stakeholder alignment, followed by relocation to our Bangalore CoffeeBeans office for ongoing operations. 🎯 Key Responsibilities Infrastructure Design & Implementation Design and architect enterprise-grade Kubernetes clusters across multi-cloud environments (AWS/Azure/GCP) Build production-ready Kubernetes infrastructure with high availability, scalability, and security best practices Implement Infrastructure as Code using Terraform, Helm charts, and GitOps methodologies Set up monitoring, logging, and observability solutions for Kubernetes workloads. Design disaster recovery and backup strategies for containerized applications Leadership & Team Management Lead a team of 3-4 DevOps engineers and provide technical mentorship Drive best practices for containerization, orchestration, and cloud-native development Collaborate with development teams to optimize application deployment strategies Conduct technical reviews and ensure code quality standards across infrastructure components Facilitate knowledge transfer and create comprehensive documentation Operational Excellence Manage CI/CD pipelines integrated with Kubernetes deployments Implement security policies including RBAC, network policies, and container security scanning Optimize cluster performance and resource utilization Automate routine operations and reduce manual intervention Ensure 99.9% uptime for production Kubernetes workloads Strategic Planning Define infrastructure roadmap aligned with business objectives Evaluate and recommend new tools and technologies for container orchestration Capacity planning and cost optimization for cloud infrastructure Risk assessment and mitigation strategies for production environments 🛠 Must-Have Technical Skills Core Kubernetes Expertise 6+ years of hands-on experience with Kubernetes in production environments Deep understanding of Kubernetes architecture, components (etcd, API server, scheduler, kubelet) Expertise in Kubernetes networking (CNI, Ingress controllers, Service mesh) Advanced knowledge of Kubernetes storage (CSI, Persistent Volumes, StorageClasses) Experience with Kubernetes operators and custom resource definitions (CRDs) Infrastructure as Code Terraform - Advanced proficiency for infrastructure provisioning Helm - Creating and managing complex Helm charts Ansible/Chef/Puppet - Configuration management experience GitOps workflows - ArgoCD, Flux, or similar tool Cloud Platforms - Multi-cloud experience with at least 2 major cloud providers: AWS: EKS, EC2, VPC, IAM, CloudFormation Azure: AKS, Virtual Networks, Azure Resource Manager GCP: GKE, Compute Engine, VPC, Deployment Manager CI/CD & DevOps Tools Jenkins, GitLab CI, Azure DevOps, or GitHub Actions Docker - Advanced containerization and optimization techniques Container registries - Docker Hub, ECR, ACR, GCR management Version control - Git workflows and branching strategies Monitoring & Observability Prometheus & Grafana - Metrics collection and visualization ELK Stack / EFK - Centralized logging solutions Jaeger/Zipkin - Distributed tracing implementation AlertManager - Intelligent alerting and incident management 💡 Good-to-Have Skills Advanced Technologies Service Mesh experience (Istio, Linkerd, Consul Connect) Serverless platforms (Knative, OpenFaaS, AWS Lambda) Database operations in Kubernetes (PostgreSQL, MongoDB operators) Machine Learning pipelines on Kubernetes (Kubeflow, MLflow) Security & Compliance Container security tools (Twistlock, Aqua Security, Falco) Policy management (Open Policy Agent, Gatekeeper) Compliance frameworks (SOC 2, PCI-DSS, GDPR) Certificate management (cert-manager, Let's Encrypt) Programming & Scripting Python/Go - For automation and tooling development Shell scripting (Bash/PowerShell) - Advanced automation YAML/JSON - Configuration management expertise 🎓 Required Qualifications Education Bachelor's degree in Computer Science, Engineering, or related technical field - Relevant certifications preferred: Certified Kubernetes Administrator (CKA) Certified Kubernetes Application Developer (CKAD) Cloud provider certifications (AWS/Azure/GCP) Experience 6-7 years of DevOps/Infrastructure engineering experience 4+ years of hands-on Kubernetes experience in production 2+ years in a lead/senior role managing infrastructure teams Experience with large-scale distributed systems and microservices architecture
Posted 4 weeks ago
10.0 years
0 Lacs
Hyderabad, Telangana, India
Remote
Are you ready to make an impact at DTCC? Do you want to work on innovative projects, collaborate with a dynamic and supportive team, and receive investment in your professional development? At DTCC, we are at the forefront of innovation in the financial markets. We are committed to helping our employees grow and succeed. We believe that you have the skills and drive to make a real impact. We foster a thriving internal community and are committed to creating a workplace that looks like the world that we serve. Pay and Benefits: Competitive compensation, including base pay and annual incentive Comprehensive health and life insurance and well-being benefits, based on location Pension / Retirement benefits Paid Time Off and Personal/Family Care, and other leaves of absence when needed to support your physical, financial, and emotional well-being. DTCC offers a flexible/hybrid model of 3 days onsite and 2 days remote (onsite Tuesdays, Wednesdays and a third day unique to each team or employee). The Impact you will have in this role: At DTCC, the Observability team is at the forefront of ensuring the health, performance, and reliability of our critical systems and applications. We empower the organization with real-time visibility into infrastructure and business applications by leveraging cutting-edge monitoring, reporting, and visualization tools. Our team collects and analyzes metrics, logs, and traces using platforms like Splunk and other telemetry solutions. This data is essential for assessing application health and availability, and for enabling rapid root cause analysis when issues arise—helping us maintain resilience in a fast-paced, high-volume trading environment. If you're passionate about observability, data-driven problem solving, and building systems that make a real-world impact, we’d love to have you on our team. Primary Responsibilities: As a member of DTCC’s Observability team, you will play a pivotal role in enhancing our monitoring and telemetry capabilities across critical infrastructure and business applications. Your responsibilities will include: Lead the migration from OpenText monitoring tools to Grafana and other open-source platforms. Design and deploy monitoring rules for infrastructure and business applications. Develop and manage alerting rules and notification workflows. Build real-time dashboards to visualize system health and performance. Configure and manage OpenTelemetry Collectors and Pipelines. Integrate observability tools with CI/CD, incident management, and cloud platforms. Deploy and manage observability agents across diverse environments. Perform upgrades and maintenance of observability platforms. Qualifications: Minimum of 10+ years of related experience. Bachelor's degree preferred or equivalent experience. Talent needed for success Proven experience designing intuitive, real-time dashboards (e.g., in Grafana) that effectively communicate system health, performance trends, and business KPIs. Expertise in defining and tuning monitoring rules, thresholds, and alerting logic to ensure accurate and actionable incident detection. Strong understanding of both application-level and operating system-level metrics, including CPU, memory, disk I/O, network, and custom business metrics. Experience with structured log ingestion, parsing, and analysis using tools like Splunk, Fluentd, or OpenTelemetry. Familiarity with implementing and analyzing synthetic transactions and real user monitoring to assess end-user experience and application responsiveness. Hands-on experience with application tracing tools and frameworks (e.g., OpenTelemetry, Jaeger, Zipkin) to diagnose performance bottlenecks and service dependencies. Proficiency in configuring and using AWS CloudWatch for collecting and visualizing cloud-native metrics, logs, and events. Understanding of containerized environments (e.g., Docker, Kubernetes) and how to monitor container health, resource usage, and orchestration metrics. Ability to write scripts or small applications in languages such as Python, Java, or Bash to automate observability tasks and data processing. Experience with automation and configuration management tools such as Ansible, Terraform, Chef, or SCCM to deploy and manage observability components at scale. Actual salary is determined based on the role, location, individual experience, skills, and other considerations. We are an equal opportunity employer and value diversity at our company. We do not discriminate on the basis of race, religion, color, national origin, sex, gender, gender expression, sexual orientation, age, marital status, veteran status, or disability status. We will ensure that individuals with disabilities are provided reasonable accommodation to participate in the job application or interview process, to perform essential job functions, and to receive other benefits and privileges of employment. Please contact us to request accommodation.
Posted 4 weeks ago
3.0 - 7.0 years
0 Lacs
pune, maharashtra
On-site
Sarvaha is seeking a skilled Observability Engineer with at least 3 years of experience to assist in the design, deployment, and scaling of monitoring and logging infrastructure on Kubernetes. As part of this role, you will be instrumental in establishing end-to-end visibility in cloud environments by managing Petabyte data scales, aiding teams in improving reliability, early anomaly detection, and promoting operational excellence. You will be responsible for configuring and overseeing observability agents on AWS, Azure & GCP, utilizing Infrastructure as Code (IaC) techniques like Terraform, Helm & GitOps for automating the deployment of the Observability stack. Additionally, you should have experience working with various language stacks such as Java, Ruby, Python, and Go, instrumenting services using OpenTelemetry, integrating telemetry pipelines, optimizing telemetry metrics storage with time-series databases like Mimir & NoSQL DBs, creating dashboards, setting up alerts, and tracking SLIs/SLOs. Your role will also involve enabling Root Cause Analysis (RCA) and incident response using observability data, as well as securing the observability pipeline. The ideal candidate will possess a BE/BTech/MTech (CS/IT or MCA) degree with a focus on Software Engineering, strong skills in interpreting logs, metrics, and traces, proficiency in tools like LGTM (Loki, Grafana, Tempo, Mimi), Jaeger, Datadog, Zipkin, InfluxDB, familiarity with log frameworks such as log4j, lograge, Zerolog, loguru, knowledge of OpenTelemetry, IaC, and security best practices, ability to document observability processes, logging standards & instrumentation guidelines, proactive issue identification and resolution using observability data, and a commitment to maintaining data quality and integrity throughout the observability pipeline. At Sarvaha, you can expect top-notch remuneration, excellent growth prospects, a supportive work environment with talented individuals, challenging software implementation and deployment tasks, and the flexibility of a hybrid work mode offering complete work-from-home options even prior to the pandemic.,
Posted 1 month ago
5.0 years
0 Lacs
Bengaluru South, Karnataka, India
On-site
You Lead the Way. We’ve Got Your Back. With the right backing, people and businesses have the power to progress in incredible ways. When you join Team Amex, you become part of a global and diverse community of colleagues with an unwavering commitment to back our customers, communities and each other. Here, you’ll learn and grow as we help you create a career journey that’s unique and meaningful to you with benefits, programs, and flexibility that support you personally and professionally. At American Express, you’ll be recognized for your contributions, leadership, and impact—every colleague has the opportunity to share in the company’s success. Together, we’ll win as a team, striving to uphold our company values and powerful backing promise to provide the world’s best customer experience every day. And we’ll do it with the utmost integrity, and in an environment where everyone is seen, heard and feels like they belong. Join Team Amex and let's lead the way together. How will you make an impact in this role? There are hundreds of opportunities to make your mark on technology and life at American Express. As an Engineer, Here’s just some of what you’ll be doing: - Function as a member of an agile team and helps drive consistent development and test practices with respect to tools, common components, and documentation. - Perform technical development for applications including architecture, design, developing prototypes, writing new code and API's and performing unit and assembly testing of developed software also as needed - Hand-on architecture working closely with solution architects and architecture governance technical team for solution development and design reviews - Typically spends 60-70% of time writing code and testing, and remainder of time managing team and stakeholders through ongoing product/platform release sprints - Perform code reviews, make design decisions and ensure constant focus on quality and clearing technical debt items in the backlog. - You will be using Java Technologies – Vertx, Spring MVC, Spring Data and Spring Boot and RESTful micro services. - Build & Maintain new RESTful APIs on Cloud Platform (eCP) using MapR DB and any NoSQL DBs. - Improve efficiency, reliability, and scalability of our data pipelines. - Work on cross-functional initiatives and collaborate with Engineers across the organizations. - Build CICD pipelines for continuous integration and delivery. - Build automation for application changes and deployment for faster time to market. - Develops deep understanding of tie-ins with other systems and platforms within the supported domains - Perform ongoing refactoring of code, quality assurance and testing, applying best practice methodologies and processes. - Identify opportunities to adopt innovative technologies & ideas in development / test area. - Provide continuous support for ongoing application availability. - Collaborate/influence within and across teams to create successes with an innovative mindset and challenge the status quo. Minimum Qualifications - Computer Science, or equivalent degree or industry experience. - Up to 5 years of software development experience in a professional environment. - Strong OOPs concepts, concurrency, exception handling etc. - Data structures and Algorithms, Strong OS fundamentals, Advanced Java concepts - Strong development experience using languages Java/ Go. Keen understanding of how to build highly performant systems with such languages. - Hands on experience in building distributed micro-service systems, experience with Service Mesh products/frameworks like such as Envoy, Linkerd, Istio, gRPC/ProtoBuf; Distributed System Management/tracing and monitoring such as Zipkin/Jaeger, ELK Stack, Prometheus - Have experience in designing and building Cloud native applications. Experience in cloud platforms like Docker, Kubernetes, OpenShift are a plus. - Experience in using NoSQL databases and distributed messaging systems such as Redis, Kafka, Couchbase, Cassandra in the context of low latency and high transaction volume systems. - Experience with Continuous Integration/Continuous Deployment tools such as Jenkins, GitHub - Experience working in DevOps teams and Scrums. - Ability to learn and deep dive into new technology and business domains, question status quo and think outside of the box. - Proficient oral and written communication skills. Preferred Qualifications Demonstrated experience of writing applications for deployment on cloud environment including OpenShift and Docker containers. Experience on RESTful API design and implementation is a plus 1-2 years of experience with Test Driven Development (TDD) / Behavior Driven Development (BDD) practices, unit testing, functional testing, system integration testing, regression testing and web service testing. - Experience with automated release management using Maven, Git, Jenkins. Experience with continuous integration and continuous delivery environment Looks proactively beyond the obvious for continuous improvement opportunities. Willingness to learn new technologies and exploit them to their optimal potential Excellent leadership and communication skills, with the ability to influence at all levels across functions, from both technical and non-technical perspectives alike. We back our colleagues and their loved ones with benefits and programs that support their holistic well-being. That means we prioritize their physical, financial, and mental health through each stage of life. Benefits include: Competitive base salaries Bonus incentives Support for financial-well-being and retirement Comprehensive medical, dental, vision, life insurance, and disability benefits (depending on location) Flexible working model with hybrid, onsite or virtual arrangements depending on role and business need Generous paid parental leave policies (depending on your location) Free access to global on-site wellness centers staffed with nurses and doctors (depending on location) Free and confidential counseling support through our Healthy Minds program Career development and training opportunities American Express is an equal opportunity employer and makes employment decisions without regard to race, color, religion, sex, sexual orientation, gender identity, national origin, veteran status, disability status, age, or any other status protected by law. Offer of employment with American Express is conditioned upon the successful completion of a background verification check, subject to applicable laws and regulations.
Posted 1 month ago
10.0 years
0 Lacs
New Delhi, Delhi, India
On-site
Job description For IMMEDIATE JOINER contract basis. Job description Experience: 10+ years. We are looking for a Java Tech Lead cum Architect to provide sound technical leadership in all aspects of our business. You will communicate with employees, stakeholders and customers to ensure our company's technologies are used appropriately. As a tech lead, you will be responsible for developing, overseeing, improving and upgrading enterprise services and software. You will also need to stay on top of the latest trends and technologies and keep an eye out for any software, services that might improve business processes. Strategic thinking and strong business acumen are essential in this role. We expect you to be well-versed in current technological trends and familiar with a variety of business concepts. If you are also an excellent communicator and public speaker, wed like to meet you. Responsibilities Develop technical aspects of the company's strategy to ensure alignment with its business goals Discover and implement new technologies that yield competitive advantage Help departments use technology profitably Supervise system infrastructure to ensure functionality and efficiency Build quality assurance and data protection processes Monitor KPIs and IT budgets to assess technological performance Use stakeholders feedback to inform necessary improvements and adjustments to technology Communicate technology strategy to partners and investors Requirements Skillfully write high-quality, well-tested code and should be comfortable with Object-Oriented programming Presence in the external tech community: can proactively share his expertise with others via speaking engagements, contributions to open source, blogs and more Comfortability with Agile methods, such as Extreme Programming (XP), Scrum and/or Kanban Resource should enjoy influencing others and always advocate for technical excellence while being open to change when needed. Resource should resilient in ambiguous situations and can approach challenges from multiple perspectives. Working knowledge of cloud technology such as AWS, Azure, Kubernetes and Docker is an added advantage. Proven experience as a TL or similar leadership role Knowledge of technological trends to build strategy Developing the Technical Architecture based on Micro services. Developing and coordinating Architecture Plans In hand coding experience of Java, Golang, Python. Java, JEE/J2EE Core Java 11, JDBC 4; Servlet 3; Web Services (SOAP, REST), JAX-WS, JAX-RS; EJB 3; Struts 1.2 & 2; Groovy and Grails; Spring (Core, Bean, Context, Security, AOP, ORM, REST, MVC, Repository, Transaction and Propagation); JPA 2, Hibernate 3, Spring Data, iBatis, GORM; Java Core Design Pattern and JEE Design Pattern. Microservices Spring Boot, Spring Cloud (Config, Netflix (Eureka, Hystrix, Zuul, Archaius), Cloud foundry, Sleuth (Zipkin, HTrace), Kafka, Zookeeper); Docker, Kubernetes, Cassandra, Redis Cluster, Elastic Search, GCP, GITHUB Deisgn and Development experience on PostgreSQL 8, MongoDB, MySQL. Sound knowledge of API design tools RAML, Swagger, Sandbox and Mulesoft Anypoint Studio. Understanding of budgets and business-planning Ability to conduct technological analyses and research Excellent communication skills Leadership and organizational abilities Strategic thinking Problem-solving aptitude Btech/BE in Computer Science, Engineering or a related field; M-tech or other relevant post graduate degree is a plus. Note : leading a team of Java developers, designing and implementing Java-based applications, providing technical guidance, ensuring project quality, and mentoring junior developers is core requirement of this position.
Posted 1 month ago
10.0 years
0 Lacs
Bangalore Urban, Karnataka, India
On-site
About the Role: As the SRE Architect for Flipkart’s Reliability & Productivity Charter, you will own the vision and strategic roadmap for our Reliability charter—defining what “resilient at scale” means for Flipkart and how we measure success. You will architect and drive key platform initiatives including: ● Centralized Observability Stack: End-to-end design of metrics, tracing, logging, and alerting pipelines to give every engineering team a single pane of glass into system health. ● Public Cloud Management: Define best practices, guardrails, and automation for Flipkart’s multi-region GCP footprint to ensure cost-effective, secure, and compliant operations. ● SRE Platform Innovations: Lead the architecture of chaos engineering (Chaos Platform), mass code migration (CodeLift with OpenRewrite), golden-image enforcement and artifact scanning (ImageScanning), and other next-generation reliability tools. In this role, you will collaborate closely with engineering, product, and operations stakeholders to translate high-level reliability objectives into concrete, scalable systems and processes that empower thousands of engineers to build, deploy, and operate Flipkart’s services with confidence. About Flipkart’s Reliability & Productivity Charter Join a dynamic SRE team focused on elevating Flipkart’s platform resilience, developer productivity, and operational excellence. We build and own the platforms and tooling that enable thousands of engineers to deliver high-quality features at scale and with confidence. Key Responsibilities ● Architect & Design ○ Define the end-to-end architecture for centralized observability (metrics, tracing, logs, alerting) and ensure scalability, security, and cost-efficiency ○ Drive the technical roadmap for platforms such as Chaos Platform, CodeLift, and Image Scanning ○ Establish best-practice patterns (golden paths) for multi-region, multi-cloud deployments aligned with BCP/DR requirements ● Platform Delivery & Governance ○ Lead cross-functional design reviews, proof-of-concepts, and production rollouts for new platform components ○ Ensure robust standards for API design, data modeling, and service-level objectives (SLOs) ○ Define and enforce policy as code (e.g., quota management, image enforcement, CI/CD pipelines) ● Technology Leadership & Mentorship ○ Coach and guide SRE Engineers and Platform Engineers on system design, reliability patterns, and performance optimizations ○ Evangelize “shift-left” practices: resilience testing, security scanning (Snyk, Artifactory integration), and automated feedback loops ○ Stay abreast of industry trends (service meshes, event stores, distributed tracing backends) and evaluate their applicability ● Performance & Capacity Planning ○ Collaborate with FinanceOps and CloudOps to optimize public cloud cost, capacity, and resource utilization ○ Define monitoring, alerting, and auto-remediation strategies to maintain healthy error budgets What We’re Looking For ● Experience & Expertise ○ 10+ years in large-scale distributed systems architecture, with at least 3 years in an SRE or platform engineering context ○ Hands-on mastery of observability stacks (Prometheus, OpenTelemetry, Jaeger/Zipkin, ELK/EFK, Grafana, Alertmanager) ○ Proven track record of designing chaos engineering frameworks and non-functional testing workflows ● Technical Skills ○ Deep knowledge of public cloud platforms (GCP preferred), container orchestration (Kubernetes), and IaC (Terraform, Helm) ○ Strong background in language-agnostic tooling (Go, Java, Python) and API-driven microservices architectures ○ Familiarity with OpenRewrite for mass code migration and vulnerability management tools (Snyk, Trivy) ● Leadership & Collaboration ○ Demonstrated ability to influence stakeholders across engineering, product, and operations teams ○ Excellent written and verbal communication—able to translate complex architectures into clear, actionable plans ○ Passion for mentoring and growing engineering talent in reliability and productivity best practice
Posted 1 month ago
Upload Resume
Drag or click to upload
Your data is secure with us, protected by advanced encryption.
Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.
We have sent an OTP to your contact. Please enter it below to verify.
Accenture
73564 Jobs | Dublin
Wipro
27625 Jobs | Bengaluru
Accenture in India
22690 Jobs | Dublin 2
EY
20638 Jobs | London
Uplers
15021 Jobs | Ahmedabad
Bajaj Finserv
14304 Jobs |
IBM
14148 Jobs | Armonk
Accenture services Pvt Ltd
13138 Jobs |
Capgemini
12942 Jobs | Paris,France
Amazon.com
12683 Jobs |