Get alerts for new jobs matching your selected skills, preferred locations, and experience range.
4 - 9 years
8 - 12 Lacs
Bengaluru
Work from Office
Knowledge of Linux &Windows systems, system architecture & configuration management tools, experienced with monitoring tools like Grafana, Prometheus, and Loki. The role involves understanding of TCP/IP, HTTP, DNS, & Load Balancing protocols. Required Candidate profile The role requires exp of 5 to 8 years and expertise with CI/CD pipelines, proficiency in automating tasks for enhanced efficiency & reliability &hands-on experience with Terraform and Kubernetes
Posted 3 months ago
3 - 6 years
5 - 8 Lacs
Kochi
Work from Office
This role is within the DevOps team responsible for build and delivery of multiple IBM Automation software products. Responsibilities include developing and modernizing Continuous Integration and Continuous Delivery systems and processes, to enable agile development and delivery of IBM software products, leveraging containerization technologies such as Docker, Kubernetes and Cloud infrastructure. Responsibilities: Automate and modernize software delivery processes, including testing and deployment Collaborate with development teams to eliminate manual effort, reduce human error and improve efficiency Required education Bachelor's Degree Preferred education Master's Degree Required technical and professional expertise 3+ Years of overall experience in software domain Experience with Docker and an understanding of containerized software fundamentals Experience in scripting automation languages like Python/Go Linux command line and bash scripting experience Experience with Red Hat OpenShift/ Kubernetes. Familiarity with DevOps and Agile principles, and the ability to apply them Experience developing and maintaining CI/CD pipelines using Jenkins and Tekton Strong technical and problem-solving skills with an attention to detail Self-motivated with a sense of ownership and willingness to bring forward ideas Preferred technical and professional experience Familiarity with Compute, Storage and Networking components from IBM Cloud, AWS, Azure Familiarity with LogDNA/Sysdig/Prometheus for cluster analysis in a Kubernetes environment
Posted 3 months ago
8 - 13 years
25 - 40 Lacs
Bengaluru
Work from Office
Specialized Open Positions in Azure - Cloud Engineer - DevOps Engineer - Solutions Architect - Azure Data Engineer - Security Engineer - API Developer - Data Hosting - IAM - OS - Integration C&B - Best in The Industry Required Candidate profile Mid Level - 5+ Years , Sr Level - 10+ years of Relevant Experience in Cloud Engineering Must have hands on Working Experience Send CV to itjobs@consultasia.in mention Specialization in Azure Cloud
Posted 3 months ago
10 - 15 years
12 - 17 Lacs
Bengaluru
Work from Office
Skill Set: Experience with Azure Cloud Services (IaaS, PaaS- App Service, SQL, Traffic Manager, RBAC, NLB, VNET, NSG, DNS, Azure Cognito services etc.) and knowledge on DevOps concepts. Strong Scripting experience - PowerShell, Azure CLI, Bash, etc. Good understanding of Cloud concepts and experience in working on Microsoft Azure (IAAS/PAAS/SAAS) and Azure DevOps platform Should have Strong experience in assessing the existing on prem application/DC based application towards its cloud readiness using widely used tools using Azure Migrate, Matilda or custom-built tools. Strong knowledge in conducting discovery and assessment of the applications, recommendations on R-lane, Identify move group, migration approach. Able to perform cost comparison between on prem and cloud applications, ability to negotiate and convince the stakeholders for cloud migration. Hands-on familiarity with IaaC (Infrastructure as a Code) constructs using Terraform HCL ( HashiCorp Configuration Language), Powershell and other DevOps tools Experience in using CI and CD tools. Jenkins/ Azure pipeline , C reation and deployment of Build and Release definitions including build artefact management using tools like Azure Artefacts, Artefactory Experience in working with Distributed Source Code Repositories like GitHub, Azure DevOps Services (TFS/Git) Exposure to implementing HA/DR solution architectures in Cloud Experience in Kubernetes, Docker, and other cloud deployment, container technologies Exposure to the monitoring tools like Prometheus, Grafana, Datadog, etc Exposure to migrating on-premise .NET based solutions/applications (ASP.Net, C#, MVC, SQL) to Azure cloud platform will be added advantage Stakeholder Interaction Stakeholder Type Stakeholder Identification Purpose of Interaction Internal Delivery teams Review of architectural process deployment in engagements, escalation management, issue resolution, delivery assurance and excellence Pre-sales team For solutioning and architectural design purpose Quality Assurance Periodic review of architecture Practice teams To understand the different solution portfolios Holmes engineering and roll out For automation purpose Talent Transformation Team, Competency Group Plan and support delivery of Technical Trainings, knowledge sharing HRBP For hiring and managing resources External Vendors/ Partners For strategic alignment and partnerships, training Industry forums Best practices, market intelligence, knowledge sharing Display Lists the competencies required to perform this role effectively: Functional Competencies/ Skill Domain/Industry Knowledge - Awareness and knowledge of broad economic, demographic, technological and global trends within own ecosystem - Master Market Intelligence - Deep specialized understanding of the ecosystem practice, overall market & competition and nuances of delivery in that domain - Master Systems Thinking - Understanding of the Wipro system (interrelatedness, interdependencies and boundaries) and perform problem solving in a complex environment - Expert Leveraging Technology - In-depth knowledge of and mastery over ecosystem technology that commands expert authority respect - Master Asset reusability - ability to re-use the assets to ensure scalability and optimization - Expert Technical knowledge - architecture design principles, framework and documentation - Master Competency Levels Foundation Knowledgeable about the competency requirements. Demonstrates (in parts) frequently with minimal support and guidance. Competent Consistently demonstrates the full range of the competency without guidance. Extends the competency to difficult and unknown situations as well. Expert Applies the competency in all situations and is serves as a guide to others as well. Master Coaches others and builds organizational capability in the competency area. Serves as a key resource for that competency and is recognised within the entire organization. Behavioral Competencies Strategic perspective Stakeholder Management Technology Acumen Innovation Managing Complexity Client centricity Execution excellence Self-Learning Domain Experience Change agility Passion for results Nurturing people Executive presence Deliver No. Performance Parameter Measure 1. Support sales team to create wins % of proposals with Quality Index >7, timely support of the proposals, identifying opportunities/ leads to sell services within/ outside account (lead generation), no. of proposals led 2. Delivery support CSAT, delivery as per cost, quality and timelines, Identify and develop resuable components, utilization %, Recommend tools for reuse, automation for improved productivity and reduced cycle times, # of reusable components 3. Capability development % trainings and certifications completed, increase in ACE certifications, thought leadership content developed (white papers, Wipro PoVs) 4. Practice Building Identifying opportunities for architects in the account and outside through known contacts, Contribute towards domain and technology accelerator kits, Mentoring junior architects 5. Team Management Team attrition %, Employee satisfaction score
Posted 3 months ago
3 - 5 years
4 - 7 Lacs
Gurgaon
Work from Office
GreensTurn is seeking a highly skilled DevOps Engineer to manage and optimize our cloud infrastructure, automate deployment pipelines, and enhance the security and performance of our web based platform. The ideal candidate will be responsible for ensuring high availability, scalability, and security of the system while working closely with developers, security teams, and product managers. Key Responsibilities: Cloud Infrastructure Management: Deploy, configure, and manage cloud services on AWS or Azure for scalable, cost-efficient infrastructure. CI/CD Implementation: Develop and maintain CI/CD pipelines for automated deployments using GitHub Actions, Jenkins, or GitLab CI/CD . Containerization & Orchestration: Deploy and manage applications using Docker, Kubernetes (EKS/AKS), and Helm . Monitoring & Performance Optimization: Implement real-time monitoring, logging, and alerting using Prometheus, Grafana, CloudWatch, or ELK Stack . Security & Compliance: Ensure best practices for IAM (Identity & Access Management), role-based access control (RBAC), encryption, firewalls, and vulnerability management . Infrastructure as Code (IaC): Automate infrastructure provisioning using Terraform, AWS CloudFormation, or Azure Bicep . Networking & Load Balancing: Set up VPC, security groups, load balancers (ALB/NLB), and CDN (CloudFront/Azure CDN) . Disaster Recovery & Backup: Implement automated backups, failover strategies, and disaster recovery plans . Database Management: Optimize database performance, backup policies, and replication for MongoDB Collaboration & Documentation: Work with development teams to integrate DevOps best practices and maintain proper documentation for infrastructure and deployment workflows. Preferred candidate profile Perks and benefits
Posted 3 months ago
7 - 12 years
10 - 18 Lacs
Pune, Bengaluru, Hyderabad
Work from Office
Total Yrs. of Experience* - 7+yrs Work Location* - Bangalore/Hyderabad/Pune MUST HAVE- Kubernetes Administration Log based Observability (ELK or Similar), and Monitoring/Obs using Metrics (Prometheus+VicMetrics or Similar) Good to have- Kafka, Clustered DB, OpenStack, GitLab, HA K8S, Ansible
Posted 3 months ago
5 - 7 years
35 - 40 Lacs
Pune, Mumbai, Gurgaon
Work from Office
Must have 5+ years of experience.Implement & maintain Kubernetes clusters, ensuring high availability and scalability. Established real-time monitoring with Grafana, Prometheus, and CloudWatch Night Shift Location-Mumbai,Gurugram,Chennai,Indore,Remote Bangalore , Delhi,kolkata
Posted 3 months ago
0 years
1 - 2 Lacs
Pune
Work from Office
We are a leading IT company dedicated to delivering innovative solutions using the latest technologies. Location: Pune (Baner), Maharashtra Job Type: Full-time Experience: Freshers Requirements for Professional Training: Candidate must be ready to work 5 days in a week(must be ready to work on weekends also) Bachelors degree in Computer Science, IT, or any field. Strong problem-solving skills and the ability to learn quickly. Excellent communication and teamwork skills. What We Offer: ( https://www.firstquadtechsolutions.com/#/services/professional-training ) job as a AWS and DevOps engineer with 12 months contract, means you pay only getting hired with offer and appointment letter in hand. Allowing candidates to experience the 15 days industry training before committing financially, till that candidate can quit without paying anything. Initially comprehensive industry training on AWS and DevOps practices with real-world ongoing company projects Tools and technologies covered are as Below : AWS cloud and its services Linus and its commands Terraform (IaaS) Git and Bitbucket Ansible Maven and SonarQube Jenkins Docker Kubernetes
Posted 3 months ago
6 - 8 years
2 - 6 Lacs
Delhi NCR, Noida
Work from Office
Sr Production support engineer Requirements: • Hands-on Linux experience - Must • Logging tool like ELK. - Must • Application servers(weblogic/jboss), web servers (nginx/apache web server), - Must • Scripting experience (Shell Scripting and Python) - Must Job Description: • Monitoring and alerting tools like AppD, Grafana, Prometheus and Zabbix. • Queries execution experience (Good if postgres/mongodb/Oracle) • Incident Management experience • Good troubleshooting skills
Posted 3 months ago
2 - 7 years
4 - 9 Lacs
Bengaluru
Work from Office
Primary Skillsets:Site Reliability Engineer(SRE) Application Support, Ansible, Splunk, Dynatrace (OR) AppDymanics Secondary Skills:Kibana, Elastic APM, Prometheus Location:Hyderabad / Chennai / Bangalore Level:SA / M About The Role :: 7+ years of software engineering experience 7+ years of application development and implementation experience 7+ years of application production support experience 2+ years of experience performing Linux/Unix/Windows Server, or security device, builds and configuration 2 + years of experience developing or supporting IT Service Management solutions (e.g. Incident Management, Change Management, Asset Management and Configuration Management) Experience with APM tools like Dynatrace, AppD, Elastic APM Experience with Splunk, Kibana, Prometheus Experience with Automation tools like Ansible Request to pre-screen the profiles with the below questioners and share profiles on high priority. List of technologies worked on List of Tools worked on SPLUNK How did you use the tool from your application standpoint. How can we create dashboards, alerts. As a platform support engineer, what are the different parameters you will want the application team to write into splunk logs. As a platform support engineer,what will you check in the logs, before giving a sign-off for PROD Deployment App Dynamics How did you use the tool from application standpoint Recent past scenarios where AppD helped in troubleshooting As a platform support engineer, what type of alerts can you set in AppD and why Grafana How did you use the tool from application standpoint What value-add did you get Recent past scneario where Grafana was used APIGee How did you integrate this in your app What value add did you get or implement As a platform support engineer,how will you troubleshoot a transaction in APIGee Have you heard of Denedo(my reply was YES) What does the tool do? Have you had any hands on? (Training has been scheduled to happen soon. A list of requirements are being worked by another SCRUM Team) How would you handle a scenario where the application is underperforming in cloud? If you should use any tools, what will they be? How would you handle a scenariowhere the application is underperforming in Virtual Machines?If you should use any tools, what will they be? How would you handle ascenariowhere the SQL Query in the application is underperforming? How would you handle an issue thats happening only in PROD and not in any other region? Did you troubleshoot issues in any other Operating Systems apart from Windows? Platform Support work is not like a 9-5 job and you will have to be available 24/7 sometimes in shifts, and weekends. Why would you want to get into this area from Application Development?
Posted 3 months ago
3 - 5 years
10 - 12 Lacs
Chennai, Delhi NCR, Bengaluru
Work from Office
Responsibilities Work with development teams to ideate software solutions Building and setting up new development tools and infrastructure Working on ways to automate and improve development and release processes Ensuring that systems are safe and secure against cybersecurity threats Deploy updates and fixes Perform root cause analysis for production errors Develop scripts to automate infrastructure provision Working with software developers and software engineers to ensure that development follows established processes and works as intended Requirements At least 2+ years of professional experience as a software developer / DevOps engineer or equivalent Professional experience with Golang Experience with test-driven development and the use of testing frameworks Strong communication skills Technologies we use GitOps GitHub, GitLab, BitBucket Language Golang CI/CD Jenkins, Circle CI, Travis CI, TeamCity, Azure DevOps Containerization Docker, Swarm, Kubernetes Provisioning Terraform CloudOps Azure, AWS, GCP Observability Prometheus, Grafana, GrayLog, ELK Location: Delhi NCR,Bangalore,Chennai,Pune,Kolkata,Ahmedabad,Mumbai,Hyderabad
Posted 3 months ago
3 - 8 years
30 - 35 Lacs
Delhi NCR, Mumbai, Bengaluru
Work from Office
Skill required- Azure Data Factory, Kubernetes, Azure DevOps Must-Have:- Working experience on Azure DevOps (4+ years) Working experience on Kubernetes - scripting, deployment Data Factory Terraform scripting Ansible Powershell Python, Cloud Formation, Good knowledge of ITIL process (good to have) Must have: Strong knowledge of Kubernetes, Istio Service mesh Linux - CLI and Basic knowledge of the OS Scripting (Bash and YAML) Containerization and Docker essentials Jenkins Pipeline creation and execution SCM Management such as GitHub and SVN Cloud Platform Knowledge Azure Monitoring tools like Grafana, Prometheus, ELK stack Certifications (Good to have): 1. Solutions architect associate 2. Certified Kubernetes Administrator (CKA) Location: Remote, Anywhere in- Delhi / NCR,Bangalore/Bengaluru ,Hyderabad/Secunderabad,Chennai,Pune,Kolkata,Ahmedabad,Mumbai
Posted 3 months ago
4 - 6 years
6 - 8 Lacs
Hyderabad
Work from Office
In-depth knowledge of core AWS services like EC2, S3, Lambda, RDS, VPC, IAM, CloudFormation, and CloudWatch. Hands-on experience with Jenkins, GitLab CI/CD, CircleCI, or similar tools. Proficiency in Terraform, CloudFormation, or Ansible. Strong scripting skills in Python, Bash, or PowerShell for automation. Experience with Docker and Kubernetes for managing containerized applications. Familiarity with monitoring tools like Prometheus, Grafana, Datadog, or AWS CloudWatch. Expertise in Git and branching strategies. About The Role :: In-depth knowledge of core AWS services like EC2, S3, Lambda, RDS, VPC, IAM, CloudFormation, and CloudWatch. Hands-on experience with Jenkins, GitLab CI/CD, CircleCI, or similar tools. Proficiency in Terraform, CloudFormation, or Ansible. Strong scripting skills in Python, Bash, or PowerShell for automation. Experience with Docker and Kubernetes for managing containerized applications. Familiarity with monitoring tools like Prometheus, Grafana, Datadog, or AWS CloudWatch. Expertise in Git and branching strategies. Primary skills: AWS,DevOps , Kubernets ,Docker Secondary skills: Any Scripting language
Posted 3 months ago
10 - 14 years
12 - 16 Lacs
Hyderabad
Work from Office
Overview Job Summary: We are seeking a highly skilled Kubernetes Subject Matter Expert (SME) to join our team. The ideal candidate will have 10+ years of industry experience, with 4+ years of expertise in Kubernetes and DevSecOps. The role requires hands-on experience with multi-cloud environments, preferably Azure and AWS. The candidate must hold Certified Kubernetes Administrator (CKA) or Certified Kubernetes Security Specialist (CKS) certifications and have a strong track record of implementing Kubernetes at scale for large production environments. Responsibilities Key Responsibilities: Design, deploy, and optimize Kubernetes-based platforms for large-scale production workloads. Implement DevSecOps best practices to enhance the security and reliability of Kubernetes clusters. Manage Kubernetes environments across multi-cloud platforms (Azure, AWS) with a focus on resilience and high availability. Provide technical leadership in architecting, scaling, and troubleshooting Kubernetes ecosystems. Develop automation strategies using Infrastructure-as-Code (IaC) tools such as Terraform, Helm, and Ansible. Work with security teams to ensure compliance with industry security standards and best practices. Define and implement observability and monitoring using tools like Prometheus, Grafana, and ELK Stack. Lead incident response and root cause analysis for Kubernetes-related production issues. Guide and mentor engineering teams on Kubernetes, service mesh, and container security. Qualifications Required Qualifications: 10+ years of industry experience in cloud infrastructure, container orchestration, and DevSecOps. 4+ years of hands-on experience with Kubernetes in production environments. Strong knowledge of Kubernetes security, RBAC, Network Policies, and admission controllers. Experience in multi-cloud environments (Azure, AWS preferred). Hands-on experience with Istio or other service meshes. Expertise in containerization technologies like Docker. Proficiency with CI/CD pipelines (GitOps, ArgoCD, Jenkins, Azure DevOps, or similar). Experience with Kubernetes storage and networking in enterprise ecosystems. Deep understanding of Kubernetes upgrade strategies, scaling, and optimization. Must have CKA or CKS certification. Preferred Qualifications: Experience with edge computing and Kubernetes at the edge. Familiarity with AI/ML workloads on Kubernetes. Hands-on experience with policy governance frameworks like OPA/Gatekeeper. Knowledge of serverless and event-driven architectures in Kubernetes. Strong scripting skills in Python, Go, or Bash.
Posted 3 months ago
8 - 12 years
25 - 30 Lacs
Bengaluru
Work from Office
What's the role? Lead the design and architecture of highly scalable HERE API Gateway Product ensuring the Product meet performance, reliability, and security requirements. The role requires you to be hands-on for a minimum of 80 percent of the day, including tasks such as coding, code reviews, low-level design, and high-level system design. Work closely with product managers, designers, and other stakeholders to understand requirements and deliver effective solutions that align with business goals. Provide guidance and mentorship to engineering teams on best practices for microservices development, including API design, service orchestration. Oversee the development, testing, and deployment of microservices, ensuring adherence to coding standards and best practices. Identify and address performance bottlenecks, scaling issues, and other technical challenges to ensure systems can handle large volumes of traffic and data. Stay current with industry trends and emerging technologies to continuously improve our microservices architecture and development practices. Lead the response to production issues, including diagnosing problems, coordinating with engineering teams, and implementing solutions Who are you? You are a Principal Software engineer who is collaborative, creative, problem solver and the following: Must have a BS/MS Degree with 8+ years of experience in software engineering with a strong focus on microservices architecture and distributed systems. At least 2 years of experience with technical leadership. Must be hands-on for a minimum of 80 percent of the day, including tasks such as coding, code reviews, low-level design, and high-level system design. Excellent verbal and written communication skills, with the ability to articulate complex technical concepts to both technical and non-technical stakeholders. Proficiency in Java, Python, containerization technologies Docker, Kubernetes, and cloud platforms AWS, Azure. Strong analytical and problem-solving skills, with the ability to diagnose and address technical challenges in a timely manner. Strong knowledge and work experience with HTTP , the REST-principles , and TCP/IP networking , good understanding of OAuth and other authentication and authorization Experience with GitLab, Grafana, Splunk , Helm , iPaas , Prometheus, Experience in scripting languages Python, Bash, Lua is a plus. Extensive experience designing and implementing microservices-based architectures, including service orchestration, API management, and inter-service communication. Proven track record of building systems that scale horizontally and handle high traffic and data volumes efficiently. Demonstrated ability to lead technical teams, mentor engineers, and drive complex projects to successful completion.
Posted 3 months ago
4 - 8 years
20 - 30 Lacs
Hyderabad
Work from Office
Expertise in AWS, Python, Groovy, Jenkins, Terraform, and CI/CD. Must have hands-on experience with Docker, Kubernetes (EKS), cloud automation, Unix/Linux & Windows admin, and monitoring tools. Strong troubleshooting skills required.
Posted 3 months ago
4 - 9 years
20 - 35 Lacs
Hyderabad
Hybrid
4-6experience in (L2 Skills) (AWS, Python, Groovy, Jenkins)Terraform, and CI/CD automation. Must have expertise in cloud services (EC2, RDS, Lambda, EKS), infrastructure as code, and system administration. Strong troubleshooting skills required.
Posted 3 months ago
8 - 13 years
25 - 30 Lacs
Bengaluru
Work from Office
As a DevOps/SRE Engineer, you will be responsible for defining and deploying Oracle cloud services with deep focus on architecture, production operations, capacity planning, performance management, deployment, and release engineering. You will work with multiple cross-functional teams helping deliver new and outstanding experiences to our stakeholders while ensuring reliability and performance. You will have an opportunity to solve resiliency and scalability problems in the distributed systems and data processing platforms. We provide lots of training. We share, help and learn from each other. We are passionate and motivated to grow ourselves and your career. Responsibilities Take ownership of the implementation and production operations of a wide array of core system platform solutions. Identify opportunities for automation, design and implement robust deployment pipelines, and develop scripts and tools to automate repetitive tasks, which reduce manual effort and minimizing the risk of human errors. Implement best practices for monitoring, logging, and alerting to proactively identify and address potential issues before they impact users. Understand and communicate the scale, capacity, security, performance attributes, and requirements of the service and technology stack. Utilize a deep understanding of service topology and their dependencies required to troubleshoot issues and define mitigations. Create and improve current service deployment infrastructure using automation and the latest cloud capabilities to improve agility, reliability, and observability. Support CI/CD tools integration/ operations/ change management, and maintenance. Professional curiosity and a desire to a develop deep understanding of services and technologies. Qualifications: BE degree in CS, EE, or equivalent. 4+ years of experience in infrastructure engineering, DevOps or SRE roles. Experience with production operations and best practices for deploying quality code in production and troubleshooting issues when they arise. Experience with operational support of containerized, microservice-based application(s) in a production-level Kubernetes environment for a highly available product or service offering. Experience on Containerization, Docker, Kubernetes Hands-on experience with Kubernetes, including deployment and management Familiarity with Helm for managing Kubernetes applications and deployments Familiarity with monitoring and logging technologies (e.g., Prometheus, Grafana, Splunk) Troubleshooting within Linux and Kubernetes environment during deployments. Deep knowledge of Networking (TCP, UDP, DNS, DHCP, IPSec) Hands on expertise on any cloud (AWS, OCI, Azure) Familiarity with security practices in web application delivery. Hands-on experience in installing, configuring, operating, and monitoring CI/CD pipeline tools Experience with Terraform and scripting languages such as Python, Go or PowerShell for automating tasks and managing infrastructure. Thorough understanding of DevOps culture and Agile Methodology. Ability to work effectively in a collaborative, cross-functional team environment. Good to have knowledge in Java to understand the development environment. Key Skills: Multitasking - ability to work on multiple tasks at once Problem-solving skills - use problem-solving skills to isolate and solve problems with programs to keep progress on track Strong communication and analytical skills. Excellent problem solving and analytical skills. Handles hard problems with a positive "can do" attitude. Team player and able to work with others all skill level
Posted 3 months ago
3 - 5 years
5 - 7 Lacs
Bengaluru
Work from Office
The IT AI Application Platform team is seeking a Senior Site Reliability Engineer (SRE) to develop, scale, and operate our AI Application Platform based on Red Hat technologies, including OpenShift AI (RHOAI) and Red Hat Enterprise Linux AI (RHEL AI). As an SRE you will contribute to running core AI services at scale by enabling customer self-service, making our monitoring system more sustainable, and eliminating toil through automation. On the IT AI Application Platform team, you will have the opportunity to influence the complex challenges of scale which are unique to Red Hat IT managed AI platform services, while using your skills in coding, operations, and large-scale distributed system design. We develop, deploy, and maintain Red Hats next-generation AI application deployment environment for custom applications and services across a range of hybrid cloud infrastructures. We are a global team operating on-premise and in the public cloud, using the latest technologies from Red Hat and beyond. What will you do: Working with live systems and coding automation Build and manage our large scale infrastructure and platform services, including public cloud, private cloud, and datacenter-based Automate cloud infrastructure through use of technologies (e.g. auto scaling, load balancing, etc.), scripting (bash, python and golang), monitoring and alerting solutions (e.g. Splunk, Splunk IM, Prometheus, Grafana, Catchpoint etc) Design, develop, and become expert in AI capabilities leveraging emerging industry standards Participate in the design and development of software like Kubernetes operators, webhooks, cli-tools Implement and maintain intelligent infrastructure and application monitoring designed to enable application engineering teams Ensure the production environment is operating in accordance with established procedures and best practices Provide escalation support for high severity and critical platform-impacting events Provide feedback around bugs and feature improvements to the various Red Hat Product Engineering teams Contribute software tests and participate in peer review to increase the quality of our codebase Help and develop peers capabilities through knowledge sharing, mentoring, and collaboration Participate in a regular on-call schedule, supporting the operation needs of our tenants Practice sustainable incident response and blameless postmortems Work within a small agile team to develop and improve SRE methodologies, support your peers, plan and self-improve What will you bring: 3+ years of experience of using cloud providers and technologies (Google, Azure, Amazon, OpenStack etc) 1+ years of experience administering a kubernetes based production environment 2+ years of experience with enterprise systems monitoring 2+ years of experience with enterprise configuration management software like Ansible by Red Hat, Puppet, or Chef 2+ years of experience programming with at least one object-oriented language; Golang, Java, or Python are preferred 2+ years of experience delivering a hosted service Demonstrated ability to quickly and accurately troubleshoot system issues Solid understanding of standard TCP/IP networking and common protocols like DNS and HTTP Demonstrated comfort with collaboration, open communication and reaching across functional boundaries Passion for understanding users needs and delivering outstanding user experiences Independent problem-solving and self-direction Works well alone and as part of a global team Experience working with Agile development methodologies Bachelor's degree in Computer Science or a related technical field involving software or systems engineering is required Hands-on experience that demonstrates your ability and interest in Site Reliability Engineering Experience programming in at least one of these languages: Python, Golang, Java, C, C++ or another object-oriented language Experience working with public clouds such as AWS, GCP, or Azure Collaboratively troubleshoot and solve problems in a team setting Experience troubleshooting an as-a-service offering (SaaS, PaaS, etc.) and some experience working with complex distributed systems Demonstrated ability to debug, optimize code and automate routine tasks Basic understanding of Unix/Linux operating systems.
Posted 3 months ago
8 - 13 years
20 - 35 Lacs
Hyderabad
Work from Office
8-13 years of exp in AWS, Python, Groovy, Jenkins. Strong in IaC (Terraform/CloudFormation), CI/CD (Bitbucket/GitHub, Jenkins), AWS services (EC2, RDS, Lambda, S3, EKS), Docker, Linux/Windows admin, monitoring (Prometheus, Grafana), &troubleshooting.
Posted 3 months ago
6 - 11 years
7 - 15 Lacs
Pune, Bengaluru
Work from Office
Core Competency, Datadog, Splunk, Dynatrace, Grafana, Prometheus, Thousand Eyes, Gremlin etc. Efficiency in creating Dashboard for Infra / APM / E2E workflows. Monitoring, logging, Alerting and Error budget ( 99.9 , 99.99, 99.999 % ) for software, Operations & Business. Define SLO, SLI, SLA with business/ operations / Engineering team Automation / Auto healing Python, Shell Scripting, Java scripts etcDeveloping custom services Monitoring Experience with logging, monitoring, and event detection on Cloud or Distributed platforms. ITIL Incident/ Change, Proficient in Problem management – Blameless postmortem, findings, applying permanent fixes, Documentation for lesson learn. Technical operations application support and stability, realiability and resiliency experience. AND DevOps, Ansible, Terraform, Dockers, AWS (Atlas), Jenkins CICD pipelines. Unix/Linux, Windows Server, Oracle, MSSQL, MongoDB. Role: Automation Engineer Automation Engineers Responsible to perform end to end Self-Healing automation solution to reduce manual effort/TOIL. Primary Skill –Ansible, Terraform, Python, DevOps, SRE, Dockers, AWS (Atlas), ECS Based internal tooling Secondary Skill –Shell Script, Linux, Monitoring tools – Datadog, Splunk, Dynatrace, Grafana, Thousand Eyes, Gremlin etc. Automation Engineer: 5 to 7 years of experience with Automation principals and tools ( Ansible etc.).should have worked with Toil identification and quality of life automation. Advanced working experience with two or more of the following: Unix/Linux, Windows Server, Oracle, MSSQL, MongoDB. Experience with Python, Java, Curl scripting or any other types of scripting. Experience with JIRA, Confluence, BitBucket, GitHUB, Jenkins, Jules, Terraform. Experience with two or more of the following observability tools: AppDynamics, Geneos, Dynatrace, ECS Based internal tooling, Datadog, Cloud watch, Big Panda, Elastic Search (ELK), Google Cloud Logging, Grafana, Prometheus, Splunk, Thousand Eyes etc.. Experience with logging, monitoring, and event detection on Cloud or Distributed platforms. Experience creating and modifying technical documentation such as environment flow, functional requirements, nonfunctional requirements. Effective production management – Incident & change Management, Production control, ITSM, Service Now, problem solving and analytical skills with ability to turn findings into strategic imperatives. Technical operations application support and stability, reliability and resiliency experience. Minimum 4-6 years of hands-on experience into SRE implementation of monitoring system- Dashboards development for application reliability using Splunk, Dynatrace, Grafana, App Dynamics, Datadog, Big panda. Experience working on Configuration as Code, Infrastructure as code, AWS(Altas) Provides technical direction regarding monitoring and logging to less experienced staff or develops highly complex original solutions. Acts as an Expert technical resource for modeling, simulation, and analysis efforts. Overall, we are looking for an Automation Engineer, who could reduce the toil issues and enhance the system towards reliability and scalability. Nature of the Job: 1. Collaborate with Production support team, identify the existing manual activities, and automate. 2. Identify toil area where it can be automated to avoid manual intervention 3. Build Monitoring system and observability platform for more Stack traces and alerts and Dashboards. 4. Ability to define SLA, SLO and SLI and implement the same for better monitoring 5. Scalability, reliability, and observability are the primary goals for reduction of MTTD and MTTR. has context menu
Posted 3 months ago
10 - 18 years
30 - 45 Lacs
Hyderabad
Work from Office
As the Principal Application Engineer in Release Engineering, your hands-on role focuses on delivering codeline management, CICD expertise, and operations services for Oracle ERP Fusion Applications. Goal is for those services to be as automated and self-service as possible for our engineering teams. The work is non-routine, dealing with complex issues that require advanced technical and business skills in specialization. Collaboration within a distributed team underscores the importance of cultural sensitivity to prevent conflicts. Your responsibilities also include interpreting management requirements, articulating them to the team in a simple format, creating project plans, and ensuring accurate completion Job Responsibilities: Collaborating across teams, mentoring junior members, and leading projects. Liaise between management and the team, addressing team challenges. Possess expertise in cloud platforms (e.g., OCI, Azure, AWS), Oracle products, and secure coding principles. Manage a fully automated cloud software delivery pipeline. Demonstrate proficiency in Continuous Integration, Continuous Delivery, and micro-services. Advocate for pipeline-as-code and implement improvements for end-to-end automation. Identify and implement enhancements to the software delivery pipeline, focusing on CI/CD and 100% automation. Proactive engagement with management is expected when facing hurdles, and guiding and mentoring the team on project planning while prioritizing their well-being are integral aspects of this role. Experience and Qualification: Proficient in CI release engineering tools (e.g., Jenkins, Maven/Gradle) and monitoring tools (Prometheus, Grafana). Experience in performance tuning, SQL tuning, and large-scale, Linux-based enterprise software. Strong scripting skills (Python, Linux shell scripting). Expertise in software development lifecycle, change management, and configuration practices. Innovate and automate Life Cycle Management Activities, reducing manual efforts. Drive team with new features and technology updates, ensuring quick bug resolution. Detailed understanding of source code management concepts and experience with modern systems (e.g., git). Troubleshoot across various layers (network, application, caching, etc.). Shape processes, platforms, and tools for efficient software and configuration changes. Provide updates considering both immediate needs and future requirements. 8+ years of experience in software engineering, build management, testing, or related fields. 6+ years of experience in systems automation (tooling, testing, build & release). Proven success in contributing to a team-oriented, creative, and analytical environment. Excellent written and oral communication skills in English. Strong troubleshooting, problem-solving, and analytical skills. BS/MS (MS preferred) in Computer Science or equivalent work experience. Proven ability to drive transformational changes in a dynamic environment. Dedicated with a proactive and go-getter demeanour.
Posted 3 months ago
4 - 8 years
15 - 20 Lacs
Pune
Hybrid
So, what’s the role all about? A Java developer is a software professional specializing in designing, developing, and maintaining applications and systems using the Java programming language. They play a critical role in building scalable, robust, and high-performing applications for a variety of industries, including finance, healthcare, technology, and e-commerce. How will you make an impact? Bachelor’s degree in computer science, Business Information Systems or related field or equivalent work experience is required. 2+ year experience in software development – Well established technical problem-solving skills. Experience in Java, springboot and microservices. Experience with Kafka, Kinesis , KDA ,Apache Flink Experience in Kubernetes operators, Grafana, Prometheus Experience with Snowflake or any DWH solution. Experience with AWS Technology including (EKS, EMR, S3, Kinesis, Lambda’s, Firehose, IAM, CloudWatch, etc) Excellent communication skills, problem-solving skills, decision-making skills Experience in Databases Experience in CI/CD, git, github Actions Jenkins based pipeline deployments. Strong experience in SQL Working knowledge of unit testing Working knowledge of user stories or use cases Working knowledge of design patterns or equivalent experience. Working knowledge of object-oriented software design. Team Player Have you got what it takes? Bachelor’s degree in computer science, Business Information Systems or related field or equivalent work experience is required. 2+ year ( SE) experience in software development – Well established technical problem-solving skills. Experience in Java, springboot and microservices. Experience with Kafka, Kinesis , KDA ,Apache Flink Experience in Kubernetes operators, Grafana, Prometheus Experience with Snowflake or any DWH solution. Experience with AWS Technology including (EKS, EMR, S3, Kinesis, Lambda’s, Firehose, IAM, CloudWatch, etc) You will have an advantage if you also have: Experience in Big data What’s in it for you? Join an ever-growing, market disrupting, global company where the teams – comprised of the best of the best – work in a fast-paced, collaborative, and creative environment! As the market leader, every day at NICE is a chance to learn and grow, and there are endless internal career opportunities across multiple roles, disciplines, domains, and locations. If you are passionate, innovative, and excited to constantly raise the bar, you may just be our next Nicer! Enjoy NICE-FLEX! At NICE, we work according to the NICE-FLEX hybrid model, which enables maximum flexibility: 2 days working from the office and 3 days of remote work, each week. Naturally, office days focus on face-to-face meetings, where teamwork and collaborative thinking generate innovation, new ideas, and a vibrant, interactive atmosphere. Requisition ID: 6083 Reporting into: Tech Manager Role Type: Individual Contributor
Posted 3 months ago
8 - 13 years
10 - 15 Lacs
Bengaluru
Work from Office
Responsibilities Architectural Design : Design and implement robust network monitoring solutions capable of providing real-time insights into network performance and health. Develop automation frameworks to streamline network management, including provisioning, configuration, and incident response. Leverage Kubernetes for deploying and managing containerized network monitoring and automation applications, ensuring scalability and reliability. AI/ML Integration : o Incorporate AI/ML technologies to enhance network monitoring, including predictive analytics for network performance, anomaly detection, and automated incident response. o Develop and implement machine learning models that can analyze network data and provide actionable insights for optimization. Technical Leadership : Lead the development and integration of innovative network monitoring and automation strategies. Establish and enforce industry standards and best practices for software development within the network domain. Development and Implementation : Ensure seamless integration of monitoring and automation tools with existing network infrastructure and third-party systems. Collaboration and Communication : Work closely with cross-functional teams, including Network Reliability Engineers (NRE) and Service Reliability Engineers (SRE) to ensure the effective use of monitoring and automation solutions. Performance Optimization : Continuously monitor, optimize, and tune network monitoring and automation systems to meet required service levels. Develop and implement automation scripts and tools to improve network incident response times. Security and Compliance : Integrate security considerations into the design of monitoring and automation solutions. Ensure compliance with industry regulations, including the development of automated auditing and reporting features. Continuous Improvement : Provide training and mentorship to team members on the latest technologies and methodologies. We are seeking a highly skilled Network Software Architect with extensive experience in network monitoring and automation. The ideal candidate will have a proven track record in designing and implementing scalable, high-performance solutions that enhance the reliability, security, and efficiency of complex network infrastructures. In this role, you will play a pivotal role in driving innovation, establishing best practices, and leading the development of cutting-edge technologies that empower our network operations. This position will also involve leveraging Kubernetes for containerized environments and incorporating AI/ML technologies to optimize network monitoring and automation processes, pushing the boundaries of what's possible in modern network management. Required education Bachelor's Degree Required technical and professional expertise Bachelor’s or Master’s degree in Computer Science, Information Technology or a related field. 8+ years of experience in network architecture, with a focus on network monitoring and automation. Strong programming skills in languages such as Python, Java, or Go Extensive experience with network monitoring tools (e.g., Nagios, Prometheus), automation frameworks (e.g., Ansible, Puppet, Chef), and Kubernetes. Deep knowledge of networking protocols such as BGP, VxLAN, SNMP, NetFlow, and streaming telemetry. Experience with AI/ML technologies, particularly in network monitoring and predictive analytics. Experience with DevOps practices and tools (e.g., CI/CD, Jenkins, Git). Familiarity with AI/ML-based network monitoring and predictive analytics. Knowledge of containerization technologies (e.g., Docker, Kubernetes). Deep knowledge of networking protocols such as BGP, VXLAN, SNMP, NetFlow, and streaming telemetry. Proven track record of designing and implementing scalable, high-performance network solutions. Strong problem-solving skills and the ability to work in a fast-paced, dynamic environment. Excellent communication skills, with the ability to convey complex technical concepts to non-technical audiences. Preferred technical and professional experience Industry certifications such as CCIE, ACE, JNCIE, or relevant cloud certifications are a plus
Posted 3 months ago
5 - 10 years
7 - 12 Lacs
Bengaluru
Work from Office
Responsibilities Design benchmark scenarios, implement and tune the workloads, conduct the workload measurements along with analyzing the performance data to guide Software development teams on performance improvements. This role is collaborative within a global team of IBM. Design benchmark and workload scenarios, keep them current. Implement benchmark and workload scenarios including setting up and configuring the various business automation software:Workflow, Decisions, Robotic Process Automation, Business Automation Management Open Edition. Automate and conduct performance measurements and data collection for benchmark and workload scenarios. Analyze the collected performance data to identify performance issues and bottlenecks. Develop and prototype proof of concept solutions to improve performance and drive innovations Level 3 customer engagement with performance troubleshoot and analysis. In a global team, work together with developers on improving the systems performance across the stack (Automation software, OpenShift, Databases). Required education Bachelor's Degree Preferred education Master's Degree Required technical and professional expertise 5+ years of active and recent Performance and scalability engineering and/or Data Engineering experience with system performance optimization and tooling for performance metrics from servers and software modules as well as pipeline design and development. Experience with Unix/Linux/Kubernetes commands related to system performance Experience with dockers and Kubernetes. Experience with performance script development including but not limited to JMeter, Loadrunner or Locust 3 + years in the following:Java/Scala, Golang and Python, General GC technologies, Kernel, LinuxOS stack, SQL, REST API. Knowledge and interest in AI and machine learning. English Fluent (verbal and written). Bachelor’s degree in computer science or equivalent. Preferred technical and professional experience Master’s degree in computer science or equivalent. Experience with Openshift Console and monitoring. Experience with performance assessment/monitoring tooling and technologies like NewRelic, Instana, Prometheus, Grafana. PromQL knowledge is a plus. Cloud, virtualization technologies knowledge. Experience with Microservices, Cloud DevOps
Posted 3 months ago
Upload Resume
Drag or click to upload
Your data is secure with us, protected by advanced encryption.
Prometheus is a popular monitoring and alerting tool used in the field of DevOps and software development. In India, the demand for professionals with expertise in Prometheus is on the rise. Job seekers looking to build a career in this field have a promising outlook in the Indian job market.
These cities are known for their vibrant tech industry and have a high demand for professionals skilled in Prometheus.
The salary range for Prometheus professionals in India varies based on experience levels. Entry-level positions can expect to earn around ₹5-8 lakhs per annum, whereas experienced professionals can earn up to ₹15-20 lakhs per annum.
A typical career path in Prometheus may include roles such as: - Junior Prometheus Engineer - Prometheus Developer - Senior Prometheus Engineer - Prometheus Architect - Prometheus Consultant
As professionals gain experience and expertise, they can progress to higher roles with increased responsibilities.
In addition to Prometheus, professionals in this field are often expected to have knowledge and experience in: - Kubernetes - Docker - Grafana - Time series databases - Linux system administration
Having a strong foundation in these related skills can enhance job prospects in the Prometheus domain.
As you explore opportunities in the Prometheus job market in India, remember to continuously upgrade your skills and stay updated with the latest trends in monitoring and alerting technologies. With dedication and preparation, you can confidently apply for roles in this dynamic field. Good luck!
Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.
We have sent an OTP to your contact. Please enter it below to verify.
Accenture
36723 Jobs | Dublin
Wipro
11788 Jobs | Bengaluru
EY
8277 Jobs | London
IBM
6362 Jobs | Armonk
Amazon
6322 Jobs | Seattle,WA
Oracle
5543 Jobs | Redwood City
Capgemini
5131 Jobs | Paris,France
Uplers
4724 Jobs | Ahmedabad
Infosys
4329 Jobs | Bangalore,Karnataka
Accenture in India
4290 Jobs | Dublin 2