Get alerts for new jobs matching your selected skills, preferred locations, and experience range.
1.0 - 6.0 years
6 - 13 Lacs
Bengaluru
Work from Office
Position Summary: We are seeking an experienced and highly skilled Lead LogicMonitor Administrator to architect, deploy, and manage scalable observability solutions across hybrid IT environments. This role demands deep expertise in LogicMonitor and a strong understanding of modern IT infrastructure and application ecosystems, including on premises, cloud-native, and hybrid environments. The ideal candidate will play a critical role in designing real-time service availability dashboards, optimizing performance visibility, and ensuring comprehensive monitoring coverage for business-critical services. Role & Responsibilities: Monitoring Architecture & Implementation Serve as the subject matter expert (SME) for LogicMonitor, overseeing design, implementation, and continuous optimization. Lead the development and deployment of monitoring solutions that integrate on premise infrastructure, public cloud (AWS, Azure, GCP), and hybrid environments. Develop and maintain monitoring templates, escalation chains, and alerting policies that align with business service SLAs. Ensure monitoring solutions adhere to industry standards and compliance requirements. Real-Time Dashboards & Visualization Design and build real-time service availability dashboards to provide actionable insights for operations and leadership teams. Leverage Logic Monitor’s APIs and data sources to develop custom visualizations, ensuring a single-pane-of-glass view for multi-layered service components. Collaborate with applications and service owners to define KPIs, thresholds, and health metrics. Proficient in interpreting monitoring data and metrics related to uptime and performance. Automation & Integration Automate onboarding/offboarding of monitored resources using LogicMonitor’s REST API, Groovy scripts, and Configuration Modules. Integrate LogicMonitor with ITSM tools (e.g., ServiceNow, Jira), collaboration platforms (e.g., Slack, Teams), and CI/CD pipelines. Enable proactive monitoring through synthetic transactions and anomaly detection capabilities. Streamline processes through automation and integrate monitoring with DevOps practices. Operations & Optimization Perform ongoing health checks, capacity planning, tools version upgrades, and tuning monitoring thresholds to reduce alert fatigue. Establish and enforce monitoring standards, best practices, and governance models across the organization. Lead incident response investigations, root cause analysis, and post-mortem reviews from a monitoring perspective. Optimize monitoring strategies for effective resource utilization and cost efficiency. Qualification Minimum Educational Qualifications: Bachelor’s degree in computer science, Information Technology, Engineering, or a related field Required Skills & Qualifications: 8+ years of total experience. 5+ years of hands-on experience with LogicMonitor, including custom DataSources, Property Sources, dashboards, and alert tuning. Proven expertise in IT infrastructure monitoring: networks, servers, storage, virtualization (VMware, Nutanix), and containerization (Kubernetes, Docker). Strong understanding of cloud platforms (AWS, Azure, GCP) and their native monitoring tools (e.g., CloudWatch, Azure Monitor). Experience in scripting and automation (e.g., Python, PowerShell, Groovy, Bash). Familiarity with observability stacks: ELK, Grafana, is a strong plus. Proficient with ITSM and incident management processes, including integrations with ServiceNow. Excellent problem-solving, communication, and documentation skills. Ability to work collaboratively in cross-functional teams and lead initiatives. Preferred Qualifications: LogicMonitor Certified Professional (LMCA and LMCP) or similar certification. Experience with APM tools (e.g., SolarWinds, AppDynamics, Dynatrace, Datadog) and log analytics platforms and logicmonitor observability Knowledge of DevOps practices and CI/CD pipelines. Exposure to regulatory/compliance monitoring (e.g., HIPAA, PCI, SOC 2). Experience with machine learning or AI-based monitoring solutions. Additional Information Intuitive is an Equal Employment Opportunity Employer. We provide equal employment opportunities to all qualified applicants and employees, and prohibit discrimination and harassment of any type, without regard to race, sex, pregnancy, sexual orientation, gender identity, national origin, color, age, religion, protected veteran or disability status, genetic information or any other status protected under federal, state, or local applicable laws. We will consider for employment qualified applicants with arrest and conviction records in accordance with fair chance laws.
Posted 20 hours ago
10.0 - 15.0 years
22 - 37 Lacs
Bengaluru
Work from Office
Who We Are At Kyndryl, we design, build, manage and modernize the mission-critical technology systems that the world depends on every day. So why work at Kyndryl? We are always moving forward – always pushing ourselves to go further in our efforts to build a more equitable, inclusive world for our employees, our customers and our communities. The Role As an ELK (Elastic, Logstash & Kibana) Data Engineer, you would be responsible for developing, implementing, and maintaining the ELK stack-based solutions for Kyndryl’s clients. This role would be responsible to develop efficient and effective, data & log ingestion, processing, indexing, and visualization for monitoring, troubleshooting, and analysis purposes. Responsibilities: Design, implement, and maintain scalable data pipelines using ELK Stack (Elasticsearch, Logstash, Kibana) and Beats for monitoring and analytics. Develop data processing workflows to handle real-time and batch data ingestion, transformation and visualization. Implement techniques like grok patterns, regular expressions, and plugins to handle complex log formats and structures. Configure and optimize Elasticsearch clusters for efficient indexing, searching, and performance tuning. Collaborate with business users to understand their data integration & visualization needs and translate them into technical solutions Create dynamic and interactive dashboards in Kibana for data visualization and insights that can enable to detect the root cause of the issue. Leverage open-source tools such as Beats and Python to integrate and process data from multiple sources. Collaborate with cross-functional teams to implement ITSM solutions integrating ELK with tools like ServiceNow and other ITSM platforms. Anomaly detection using Elastic ML and create alerts using Watcher functionality Extract data by Python programming using API Build and deploy solutions in containerized environments using Kubernetes. Monitor Elasticsearch clusters for health, performance, and resource utilization Automate routine tasks and data workflows using scripting languages such as Python or shell scripting. Provide technical expertise in troubleshooting, debugging, and resolving complex data and system issues. Create and maintain technical documentation, including system diagrams, deployment procedures, and troubleshooting guides If you're ready to embrace the power of data to transform our business and embark on an epic data adventure, then join us at Kyndryl. Together, let's redefine what's possible and unleash your potential. Your Future at Kyndryl Every position at Kyndryl offers a way forward to grow your career. We have opportunities that you won’t find anywhere else, including hands-on experience, learning opportunities, and the chance to certify in all four major platforms. Whether you want to broaden your knowledge base or narrow your scope and specialize in a specific sector, you can find your opportunity here. Who You Are You’re good at what you do and possess the required experience to prove it. However, equally as important – you have a growth mindset; keen to drive your own personal and professional development. You are customer-focused – someone who prioritizes customer success in their work. And finally, you’re open and borderless – naturally inclusive in how you work with others. Required Technical and Professional Experience: Minimum of 5 years of experience in ELK Stack and Python programming Graduate/Postgraduate in computer science, computer engineering, or equivalent with minimum of 10 years of experience in the IT industry. ELK Stack : Deep expertise in Elasticsearch, Logstash, Kibana, and Beats. Programming : Proficiency in Python for scripting and automation. ITSM Platforms : Hands-on experience with ServiceNow or similar ITSM tools. Containerization : Experience with Kubernetes and containerized applications. Operating Systems : Strong working knowledge of Windows, Linux, and AIX environments. Open-Source Tools : Familiarity with various open-source data integration and monitoring tools. Knowledge of network protocols, log management, and system performance optimization. Experience in integrating ELK solutions with enterprise IT environments. Strong analytical and problem-solving skills with attention to detail. Knowledge in MySQL or NoSQL Databases will be added advantage Fluent in English (written and spoken). Preferred Technical and Professional Experience “Elastic Certified Analyst” or “Elastic Certified Engineer” certification is preferrable Familiarity with additional monitoring tools like Prometheus, Grafana, or Splunk. Knowledge of cloud platforms (AWS, Azure, or GCP). Experience with DevOps methodologies and tools. Being You Diversity is a whole lot more than what we look like or where we come from, it’s how we think and who we are. We welcome people of all cultures, backgrounds, and experiences. But we’re not doing it single-handily: Our Kyndryl Inclusion Networks are only one of many ways we create a workplace where all Kyndryls can find and provide support and advice. This dedication to welcoming everyone into our company means that Kyndryl gives you – and everyone next to you – the ability to bring your whole self to work, individually and collectively, and support the activation of our equitable culture. That’s the Kyndryl Way. What You Can Expect With state-of-the-art resources and Fortune 100 clients, every day is an opportunity to innovate, build new capabilities, new relationships, new processes, and new value. Kyndryl cares about your well-being and prides itself on offering benefits that give you choice, reflect the diversity of our employees and support you and your family through the moments that matter – wherever you are in your life journey. Our employee learning programs give you access to the best learning in the industry to receive certifications, including Microsoft, Google, Amazon, Skillsoft, and many more. Through our company-wide volunteering and giving platform, you can donate, start fundraisers, volunteer, and search over 2 million non-profit organizations. At Kyndryl, we invest heavily in you, we want you to succeed so that together, we will all succeed. Get Referred! If you know someone that works at Kyndryl, when asked ‘How Did You Hear About Us’ during the application process, select ‘Employee Referral’ and enter your contact's Kyndryl email address.
Posted 3 days ago
4.0 - 5.0 years
20 - 25 Lacs
Bengaluru
Work from Office
Bachelor’s degree in a technical field: Computer Science, Engineering, or similar Experience working in an external customer facing technical support role Experience troubleshooting complex technical issues Excellent written and verbal communication skills in English Fundamental L2/L3 networking knowledge: network stacks, switching, routing, firewalls, etc. Experience working with Linux Ability to work in a dynamic, high pressure customer facing environment Ability to manage and prioritize numerous customer issues simultaneously. Additional Skills Considered a Plus Experience with Docker and Kubernetes Experience with ELK stack Experience working with Telecommunication Providers Experience with scripting: Ansible, Bash, Python
Posted 4 days ago
7.0 - 10.0 years
11 - 16 Lacs
Mumbai, Hyderabad, Pune
Work from Office
Key Responsibilities: Design, build, and maintain CI/CD pipelines for ML model training, validation, and deployment Automate and optimize ML workflows, including data ingestion, feature engineering, model training, and monitoring Deploy, monitor, and manage LLMs and other ML models in production (on-premises and/or cloud) Implement model versioning, reproducibility, and governance best practices Collaborate with data scientists, ML engineers, and software engineers to streamline end-to-end ML lifecycle Ensure security, compliance, and scalability of ML/LLM infrastructure Troubleshoot and resolve issues related to ML model deployment and serving Evaluate and integrate new MLOps/LLMOps tools and technologies Mentor junior engineers and contribute to best practices documentation Required Skills & Qualifications: 8+ years of experience in DevOps, with at least 3 years in MLOps/LLMOps Strong experience with cloud platforms (AWS, Azure, GCP) and container orchestration (Kubernetes, Docker) Proficient in CI/CD tools (Jenkins, GitHub Actions, GitLab CI, etc.) Hands-on experience deploying and managing different types of AI models (e.g., OpenAI, HuggingFace, custom models) to be used for developing solutions. Experience with model serving tools such as TGI, vLLM, BentoML, etc. Solid scripting and programming skills (Python, Bash, etc.) Familiarity with monitoring/logging tools (Prometheus, Grafana, ELK stack) Strong understanding of security and compliance in ML environments Preferred Skills: Knowledge of model explainability, drift detection, and model monitoring Familiarity with data engineering tools (Spark, Kafka, etc. Knowledge of data privacy, security, and compliance in AI systems. Strong communication skills to effectively collaborate with various stakeholders Critical thinking and problem-solving skills are essential Proven ability to lead and manage projects with cross-functional teams
Posted 4 days ago
7.0 - 10.0 years
8 - 13 Lacs
Mumbai, Hyderabad, Pune
Work from Office
Key Responsibilities: Design, build, and maintain CI/CD pipelines for ML model training, validation, and deployment Automate and optimize ML workflows, including data ingestion, feature engineering, model training, and monitoring Deploy, monitor, and manage LLMs and other ML models in production (on-premises and/or cloud) Implement model versioning, reproducibility, and governance best practices Collaborate with data scientists, ML engineers, and software engineers to streamline end-to-end ML lifecycle Ensure security, compliance, and scalability of ML/LLM infrastructure Troubleshoot and resolve issues related to ML model deployment and serving Evaluate and integrate new MLOps/LLMOps tools and technologies Mentor junior engineers and contribute to best practices documentation Required Skills & Qualifications: 8+ years of experience in DevOps, with at least 3 years in MLOps/LLMOps Strong experience with cloud platforms (AWS, Azure, GCP) and container orchestration (Kubernetes, Docker) Proficient in CI/CD tools (Jenkins, GitHub Actions, GitLab CI, etc.) Hands-on experience deploying and managing different types of AI models (e.g., OpenAI, HuggingFace, custom models) to be used for developing solutions. Experience with model serving tools such as TGI, vLLM, BentoML, etc. Solid scripting and programming skills (Python, Bash, etc.) Familiarity with monitoring/logging tools (Prometheus, Grafana, ELK stack) Strong understanding of security and compliance in ML environments Preferred Skills: Knowledge of model explainability, drift detection, and model monitoring Familiarity with data engineering tools (Spark, Kafka, etc. Knowledge of data privacy, security, and compliance in AI systems. Strong communication skills to effectively collaborate with various stakeholders Critical thinking and problem-solving skills are essential Proven ability to lead and manage projects with cross-functional teams
Posted 4 days ago
3.0 - 5.0 years
15 - 27 Lacs
Bengaluru
Work from Office
Job Summary We are seeking a skilled and innovative Cloud Engineer to join our team. As a Cloud Engineer, you will be responsible for developing and maintaining cloud-based solutions, with a focus on coding complex problems, automation using Golang and Python, and collaborating with the Site Reliability Engineering (SRE) team for feature deployment in production. Additionally, the ideal candidate should be proficient in utilizing AI tools like Copilot to enhance productivity in the areas of automation, documentation, and unit test writing. Responsibilities: • Develop, test, and maintain cloud-based applications and services using Go Lang and Python. • Write clean, efficient, and maintainable code to solve complex problems and improve system performance. • Collaborate with cross-functional teams to understand requirements and design scalable and secure cloud solutions. • Automate deployment, scaling, and monitoring of cloud-based applications and infrastructure. • Work closely with the SRE team to ensure smooth feature deployment in production environments. • Utilize AI tools like Copilot to enhance productivity in automation, documentation, and unit test writing. • Troubleshoot and resolve issues related to cloud infrastructure, performance, and security. • Stay up to date with emerging technologies and industry trends to continuously improve cloud-based solutions. • Participate in code reviews, knowledge sharing sessions, and contribute to the improvement of development processes. Job Requirements • Strong programming skills in Golang and Python. • Proficiency in using AI tools like Copilot to enhance productivity in automation, documentation, and unit test writing. • Solid understanding of cloud computing concepts and services (e.g., AWS, Azure, Google Cloud). • Experience with containerization technologies (e.g., Docker, Kubernetes) and infrastructure-as-code tools (e.g., Terraform, CloudFormation). • Proficient in designing and implementing RESTful APIs and microservices architectures. • Familiarity with CI/CD pipelines and tools (e.g., Jenkins, GitLab CI/CD). • Knowledge of networking concepts, security best practices, and system administration. • Excellent problem-solving skills and ability to work in a fast-paced, collaborative environment. • Strong communication and interpersonal skills to effectively collaborate with cross-functional teams. Preferred Skills: • Experience with other programming languages, such as Java, C++, or Ruby. • Knowledge of database technologies (e.g., SQL, NoSQL) and data storage solutions. • Familiarity with monitoring and logging tools (e.g., Prometheus, ELK stack). • Understanding of Agile/Scrum methodologies and DevOps principles. • Certifications in cloud technologies (e.g., AWS Certified Cloud Practitioner, Google Cloud Certified - Associate Cloud Engineer) would be a plus. If you are passionate about cloud technologies, have a strong problem-solving mindset, and enjoy working in a collaborative environment, we would love to hear from you. Join our team and contribute to building scalable, reliable, and secure cloud solutions. Please note that this job description is not exhaustive and may change based on the organization's needs. Education A Bachelor of Science Degree in Engineering or Computer Science with 2 years of experience, or a Master’s Degree; or equivalent experience is typically required.
Posted 6 days ago
4.0 - 7.0 years
5 - 9 Lacs
Noida
Work from Office
Proficiency in Go programming language (Golang). Solid understanding of RESTful API design and microservices architecture. Experience with SQL and NoSQL databases (e.g., PostgreSQL, MongoDB, Redis). Familiarity with container technologies (Docker, Kubernetes). Understanding of distributed systems and event-driven architecture. Version control with Git. Familiarity with CI/CD pipelines and cloud platforms (AWS, GCP, Azure). Experience with message brokers (Kafka, RabbitMQ). Knowledge of GraphQL. Exposure to performance tuning and profiling. Contributions to open-source projects or personal GitHub portfolio. Familiarity with monitoring tools (Prometheus, Grafana, ELK). Roles and Responsibilities Design, develop, and maintain backend services and APIs using Go (Golang). Write efficient, scalable, and reusable code. Collaborate with front-end developers, DevOps engineers, and product teams to deliver high-quality features. Optimize applications for performance and scalability. Develop unit and integration tests to ensure software quality. Implement security and data protection best practices. Troubleshoot and debug production issues. Participate in code reviews, architecture discussions, and continuous improvement processes.
Posted 6 days ago
8.0 - 12.0 years
30 - 35 Lacs
Pune, Chennai
Work from Office
Mandatory Skills SRE, DevOps, Scripting (Python/Bash/Perl), Automation Tools (Ansible/Terraform/Puppet), AWS Cloud, Docker, Kubernetes, Observability Tools (Prometheus/Grafana/ELK Stack/Splunk), CICD pipelines using GitLab Jenkins or similar tools Please share your resume to thulasidharan.b@ltimindtree.com Note: Only 0-30 days notice
Posted 6 days ago
5.0 - 7.0 years
25 - 40 Lacs
Pune
Work from Office
Our world is transforming, and PTC is leading the way.Our software brings the physical and digital worlds together, enabling companies to improve operations, create better products, and empower people in all aspects of their business. Our people make all the difference in our success. Today, we are a global team of nearly 7,000 and our main objective is to create opportunities for our team members to explore, learn, and grow – all while seeing their ideas come to life and celebrating the differences that make us who we are and the work we do possible. PTC is looking for hands-on engineer, experienced with site reliability and operations , for a leading CAD SaaS solution. As part of your job at PTC, you will: Collaborate with multiple teams, to monitor & observe their cloud-deployed services Implement automated pipelines for deployment into cloud environment Implement monitoring & observability solutions Handle incidents and changes Troubleshoot and resolve production issues Conduct post-mortems Handle security incidents Job requirements: Proven experience working in Cloud DevOps and Site Reliability Engineering Ability to develop observability solutions using DataDog, or ELK, Prometheus and Grafana Great communication skills, written and verbal Strong hands-on skills to support Security in Cloud environment Experience and knowledge in cloud architecture reviews, SaaS processes and handling security incidences Advantage – knowledge and experience with Azure Why PTC? Life at PTC is about more than working with today’s most cutting-edge technologies to transform the physical world. It’s about showing up as you are and working alongside some of today’s most talented industry leaders to transform the world around you. If you share our passion for problem-solving through innovation, you’ll likely become just as passionate about the PTC experience as we are. Are you ready to explore your next career move with us? Website: https://www.ptc.com LinkedIn: https://www.linkedin.com/company/ptcinc/ Facebook Page: https://www.facebook.com/ptc.inc/ Twitter Handle: @LifeatPTC '@PTC Instagram: ptc_inc Hashtag: #lifeatPTC Life at PTC is about more than working with today’s most cutting-edge technologies to transform the physical world. It’s about showing up as you are and working alongside some of today’s most talented industry leaders to transform the world around you. If you share our passion for problem-solving through innovation, you’ll likely become just as passionate about the PTC experience as we are. Are you ready to explore your next career move with us? We respect the privacy rights of individuals and are committed to handling Personal Information responsibly and in accordance with all applicable privacy and data protection laws. Review our Privacy Policy here ."
Posted 6 days ago
7.0 - 10.0 years
7 - 18 Lacs
Bengaluru
Work from Office
Responsibilities: * Design, implement & maintain monitoring solutions using Nagios, SolarWinds & Zabbix. * Develop Python scripts for automation & alerting integration with Grafana.
Posted 1 week ago
12.0 - 18.0 years
16 - 20 Lacs
Pune
Work from Office
Seasoned DevOps Architect to will lead the design, implementation, and maintenance of cloud-based infrastructure and DevOps team Collaborating closely with development, operations, and security teams, and ensure the seamless delivery of high-quality software solutions Qualifications: 18+ years of IT experience, with 8+ years dedicated to DevOps roles Deep knowledge of cloud platforms (AWS, Azure, GCP) Expertise in infrastructure automation tools (Terraform, Ansible, Puppet, Chef) Proficiency in containerization and orchestration (Docker, Kubernetes) Experience with CI/CD pipelines and tools (Jenkins, GitLab CI/CD, Azure DevOps) Strong knowledge of monitoring and logging tools (Prometheus, Grafana, ELK stack) Advanced scripting abilities (Bash, Python, Ruby) Solid understanding of security best practices and related tools Ability to work effectively both independently and within a team
Posted 1 week ago
2.0 - 7.0 years
3 - 7 Lacs
Ahmedabad
Work from Office
To help us build functional systems that improve customer experience we are now looking for an experienced DevOps Engineer. They will be responsible for deploying product updates, identifying production issues and implementing integrations that meet our customers' needs. If you have a solid background in software engineering and are familiar with Ruby or Python, wed love to speak with you. Responsibilities Work with development teams to ideate software solutions Building and setting up new development tools and infrastructure Working on ways to automate and improve development and release processes Ensuring that systems are safe and secure against cybersecurity threats Deploy updates and fixes Perform root cause analysis for production errors Develop scripts to automate infrastructure provision Working with software developers and software engineers to ensure that development follows established processes and works as intended Technologies we use GitOps GitHub, GitLab, BitBucket CI/CD Jenkins, Circle CI, Travis CI, TeamCity, Azure DevOps Containerization Docker, Swarm, Kubernetes Provisioning Terraform CloudOps Azure, AWS, GCP Observability Prometheus, Grafana, GrayLog, ELK Qualifications Graduate / Postgraduate in Technology sector Proven experience as a DevOps Engineer or similar role Effective communication and teamwork skills
Posted 1 week ago
3.0 - 8.0 years
1 - 4 Lacs
Chandigarh
Work from Office
Opportunity: We are seeking a highly skilled and experienced AI Infrastructure Engineer (or MLOps Engineer) to design, build, and maintain the robust and scalable AI/ML platforms that power our cutting-edge asset allocation strategies. In this critical role, you will be instrumental in enabling our AI Researchers and Quantitative Developers to efficiently develop, deploy, and monitor machine learning models in a high-performance, secure, and regulated financial environment. You will bridge the gap between research and production, ensuring our AI initiatives run smoothly and effectively. Responsibilities: Platform Design & Development: Architect, implement, and maintain the end-to-end AI/ML infrastructure, including data pipelines, feature stores, model training environments, inference serving platforms, and monitoring systems. Environment Setup & Management: Configure and optimize AI/ML development and production environments, ensuring access to necessary compute resources (CPUs, GPUs), software libraries, and data. MLOps Best Practices: Implement and advocate for MLOps best practices, including version control for models and data, automated testing, continuous integration/continuous deployment (CI/CD) pipelines for ML models, and robust model monitoring. Resource Optimization: Manage and optimize cloud computing resources (AWS, Azure, GCP, or on-premise) for cost-efficiency and performance, specifically for AI/ML workloads. Data Management: Collaborate with data engineers to ensure seamless ingestion, storage, and accessibility of high-quality financial and alternative datasets for AI/ML research and production. Tooling & Automation: Select, implement, and integrate various MLOps tools and platforms (e.g., Kubeflow, MLflow, Sagemaker, DataRobot, Vertex AI, Airflow, Jenkins, GitLab CI/CD) to streamline the ML lifecycle. Security & Compliance: Ensure that all AI/ML infrastructure and processes adhere to strict financial industry security standards, regulatory compliance, and data governance policies. Troubleshooting & Support: Provide expert support and troubleshooting for AI/ML infrastructure issues, resolving bottlenecks and ensuring system stability. Collaboration: Work closely with AI Researchers, Data Scientists, Software Engineers, and DevOps teams to translate research prototypes into scalable production systems. Documentation: Create and maintain comprehensive documentation for all AI/ML infrastructure components, processes, and best practices. Qualifications: Bachelor's or Master's degree in Computer Science, Software Engineering, Data Science, or a related quantitative field. Experience: 3+ years of experience in a dedicated MLOps, AI Infrastructure, DevOps, or Site Reliability Engineering role, preferably in the financial services industry. Proven experience in designing, building, and maintaining scalable data and AI/ML pipelines and platforms. Strong proficiency in cloud platforms (AWS, Azure, GCP) including services relevant to AI/ML (e.g., EC2, S3, Sagemaker, Lambda, Azure ML, Google AI Platform). Expertise in containerization technologies (Docker) and orchestration platforms (Kubernetes). Solid understanding of CI/CD principles and tools (Jenkins, GitLab CI/CD, CircleCI, Azure DevOps). Proficiency in scripting languages like Python (preferred), Bash, or similar. Experience with Infrastructure as Code (IaC) tools (e.g., Terraform, CloudFormation, Ansible). Familiarity with distributed computing frameworks (e.g., Spark, Dask) is a plus. Understanding of machine learning concepts and lifecycle, even if not directly developing models. Technical Skills: Deep knowledge of Linux/Unix operating systems. Strong understanding of networking, security, and database concepts. Experience with monitoring and logging tools (e.g., Prometheus, Grafana, ELK stack). Familiarity with data warehousing and data lake concepts. Preferred candidate profile Exceptional problem-solving and debugging skills. Proactive and self-driven with a strong sense of ownership. Excellent communication and interpersonal skills, able to collaborate effectively with diverse teams. Ability to prioritize and manage multiple tasks in a fast-paced environment. A keen interest in applying technology to solve complex financial problems.
Posted 1 week ago
4.0 - 8.0 years
18 - 27 Lacs
Hyderabad, Chennai, Bengaluru
Work from Office
Oversees development and implementation of data mining protocols, architectures, and models as well as data analysis methodologies, used to identify trends in large data sets. Implements new data mining and data analysis processes Required Candidate profile 2 or more years with relational or NoSQL databases (Oracle,Teradata,SQL Server,Hadoop,ELK) 3 or more years working with languages such as R,Python,Java 2 or more years skilled at data visualization
Posted 1 week ago
3.0 - 8.0 years
4 - 8 Lacs
Mumbai, Palai
Work from Office
About the Role As a result of our rapid growth, we are looking for a Java Backend Engineer to join our existing Cloud Engineering team and take the lead in the design and development of several key initiatives of our existing Miko3 product line as well as our new product development initiatives. Responsibilities Designing, developing and maintaining core system features, services and engines Collaborating with a cross functional team of the backend, Mobile application, AI, signal processing, robotics Engineers, Design, Content, and Linguistic Team to realize the requirements of conversational social robotics platform which includes investigate design approaches, prototype new technology, and evaluate technical feasibility Ensure the developed backend infrastructure is optimized for scale and responsiveness Ensure best practices in design, development, security, monitoring, logging, and DevOps adhere to the execution of the project. Introducing new ideas, products, features by keeping track of the latest developments and industry trends Operating in an Agile/Scrum environment to deliver high quality software against aggressive schedules Requirements Proficiency in distributed application development lifecycle (concepts of authentication/authorization, security, session management, load balancing, API gateway), programming techniques and tools (application of tested, proven development paradigms) Proficiency in working on Linux based Operating system. Working Knowledge of container orchestration platform like Kubernetes Proficiency in at least one server-side programming language like Java. Additional languages like Python and PHP are a plus Proficiency in at least one server-side framework like Servlets, Spring, java spark (Java). Proficient in using ORM/Data access frameworks like Hibernate,JPA with spring or other server-side frameworks. Proficiency in at least one data serialization framework: Apache Thrift, Google ProtoBuffs, Apache Avro,Google Json,JackSon etc. Proficiency in at least one of inter process communication frameworks WebSocket's, RPC, message queues, custom HTTP libraries/frameworks ( kryonet, RxJava ), etc. Proficiency in multithreaded programming and Concurrency concepts (Threads, Thread Pools, Futures, asynchronous programming). Experience defining system architectures and exploring technical feasibility tradeoffs (architecture, design patterns, reliability and scaling) Experience developing cloud software services and an understanding of design for scalability, performance and reliability Good understanding of networking and communication protocols, and proficiency in identification CPU, memory & I/O bottlenecks, solve read & write-heavy workloads. Proficiency is concepts of monolithic and microservice architectural paradigms. Proficiency in working on at least one of cloud hosting platforms like Amazon AWS, Google Cloud, Azure etc. Proficiency in at least one of database SQL, NO-SQL, Graph databases like MySQL, MongoDB, Orientdb Proficiency in at least one of testing frameworks or tools JMeter, Locusts, Taurus Proficiency in at least one RPC communication framework: Apache Thrift, GRPC is an added plus Proficiency in asynchronous libraries (RxJava), frameworks (Akka),Play,Vertx is an added plus Proficiency in functional programming ( Scala ) languages is an added plus Proficiency in working with NoSQL/graph databases is an added plus Proficient understanding of code versioning tools, such as Git is an added plus Working Knowledge of tools for server, application metrics logging and monitoring and is a plus Monit, ELK, graylog is an added plus Working Knowledge of DevOps containerization utilities like Ansible, Salt, Puppet is an added plus Working Knowledge of DevOps containerization technologies like Docker, LXD is an added plus
Posted 1 week ago
4.0 - 8.0 years
15 - 22 Lacs
Pune
Work from Office
Role & responsibilities - On-time delivery of KPI and Reports Ensure that Jobs and Processes runs on defined schedule Monitor Alerts and take action to resolve the issue Identify Improvement Areas in and define actions, track and close it Ensure Platform Availability Define Platform KPIs, SLOs along with Domain Ops Leads and Publish Report on regular basis Capacity Management - Regular monitoring of growth in volume/TPS and raise it to Dev/DevOps for Infra need Roster Management Re-run of jobs when required. Troubleshoot issues, apply patches and scripted workarounds Automation of daily task Creating Knowledge base and SOPs for the daily routine activities Available 24x7 to provide support to Ops team Provide on-call and weekend support as needed Work Experience & Skillset: hands-on experience in managing Technical Operation activities on Data Engineering platforms (Cloudera / Spark / Apache / ELK) Good Understanding Knowledge of ETL pipeline and components like Repartition, Sort, Roll up and Transform etc. Hands-on experience with ELK Stack and other monitoring tools like Kibana/Grafana Experience on Apache Airflow, Kafka, Nifi, Yarn Good Python / shell Scripting Quickly gain understanding of existing processes, Jobs, KPI, Reports. Good understanding of performance issue troubleshooting. Strong knowledge on Linux/Unix commands
Posted 1 week ago
10.0 - 14.0 years
11 - 16 Lacs
Hyderabad
Work from Office
Application Manager With 10+ years of experience in analysis, design, develop and support the IT solutions, you are responsible to Establish the support & operations for the given set of IT Solutions. The technology stack includes enterprise packages and custom built applications in Java, Angular, Spring, Spring Boot, Spring Batch,... Closely work with the Functional & Technical Teams during Design & Implementation Phases (Architects, Business Analysts, Lead Developers) to ensure the choices being made are sustainable for future use in operations Take lead in building & guiding the operations team (Developers, Business Analysts) to ensure the SLA/SLEs are met with QEE delivery Manage good handshake and collaboration with the Business/Process Owners and ensure the continuity and maintenance of the solution Take lead in getting periodic reports on usage, performance, health of the IT solutions with use of necessary reporting & dashboards. Effectively use the ITSM processes such as IT change, knowledge, incident, problem management and enterprise tooling. Establish good handshake & governance with the third-party suppliers / Vendors to ensure the agreed commitments on supporting the solution. With clear understanding on market & technology evolution, Create the Roadmap for the solutions so that they are kept reliable, stable always Manage good network with Application Managers community across theOrganization and seamlessly work on dependencies to ensure the E2E continuity of solution Requirements Bachelor's Degree in any Engineering or a related field. 10+ Years of experience in doing Analyse/Design/Develop/Maintain the Enterprise IT solutions Strong Technical understanding & Hands-on experience as Developer and Lead Developer (Preferably in Java, Angular, Spring, SQL, DB2) 5+ years of experience in Leading & Guiding the Developers/Business Analysts in the team to guarantee the SLA/SLEs Expertise in designing, implementing, and maintaining scalable and high-performance applications with use of Enterprise tooling (Jira, Servicenow, ELK, Containers...). Familiarity with CI/CD pipelines and DevOps Practises is desired. Strong communication (Verbal and Non-Verbal), inter-personal skills and good in preparing reporting & dashboards (Using tools like Servicenow, Power BI, Excel,..) Able to Define & Elaborate the roadmap for a given IT Solution (Custom Built / Package) and ensure its inline with stakeholder needs & market evolution Certification in any of the Application Management streams (Ex: ITIL 4) is preferred Having experience in managing applications both on-premise and cloud (Preferably Azure) Familiarity with Waterfall and Agile methodologies and tools (e.g., Jira, Azure DevOps) is mandatory. Working with SAFe principles in an added advantage.
Posted 1 week ago
8.0 - 12.0 years
35 - 50 Lacs
Bengaluru
Work from Office
Job Summary We are seeking a highly skilled Principal Infra Developer with 8 to 12 years of experience to join our team. The ideal candidate will have expertise in Splunk Admin SRE Grafana ELK and Dynatrace AppMon. This hybrid role requires a proactive individual who can contribute to our infrastructure development projects and ensure the reliability and performance of our systems. The position does not require travel and operates during day shifts. Responsibilities Systems Engineer Splunk or ElasticSearch Admin Job Requirements Build Deploy and Manage the Enterprise Lucene DB systems Splunk Elastic to ensure that the legacy physical Virtual systems and container infrastructure for businesscritical services are being rigorously and effectively served for high quality logging services with high availability. Support periodic Observability and infrastructure monitoring tool releases and tool upgrades Environment creation Performance tuning of large scale Prometheus systems Serve as Devops SRE for the internal observability systems in Visas various data centers across the globe including in Cloud environment Lead the evaluation selection design deployment and advancement of the portfolio of tools used to provide infrastructure and service monitoring. Ensure tools utilized can provide the critical visibility on modern architectures leveraging technologies such as cloud containers etc. Maintain upgrade and troubleshoot issues with SPLUNK clusters. Monitor and audit configurations and participate in the Change Management process to ensure that unauthorized changes do not occur. Manage patching and updates of Splunk hosts andor Splunk application software. Design develop recommend and implement Splunk dashboards and alerts in support of the Incident Response team. Ensure monitoring team increases use of automation and adopts a DevOpsSRE mentality Qualification 6plus years of enterprise system logging and monitoring tools experience with a desired 5plus years in a relevant critical infrastructure of Enterprise Splunk and Elasticsearch 5plus yrs of working experience as Splunk Administrator with Cluster Building Data Ingestion Management User Role Management Search Configuration and Optimization. Strong knowledge on opensource logging and monitoring tools. Experience with containers logging and monitoring solutions. Experience with Linux operating system management and administration Familiarity with LANWAN technologies and clear understanding of basic network concepts services Strong understanding of multitier application architectures and application runtime environments Monitoring the health and performance of the Splunk environment and troubleshooting any issues that arise. Worked in 247 on call environment. Knowledge of Python and other scripting languages and infrastructure automation technologies such as Ansible is desired Splunk Admin Certified is a plus
Posted 1 week ago
8.0 - 10.0 years
35 - 40 Lacs
Bengaluru
Work from Office
Job Responsibilities: Collaborates with Product and Engineering stakeholders to design and build platform services that meet key product and infrastructure requirements Produces both detailed designs for platform-level services Must be able to evaluate software and products against business requirements and turn business requirements into robust technical solutions fitting into corporate standards and strategy. Designs and implements microservices with thoughtfully defined APIs Should be conversant with frameworks & Architectures - Spring Boot, Spring Cloud, Spring Batch, Messaging Frameworks (like Kafka), Micro service Architecture Work with other areas of technology team to realize end to end solution and estimation for delivery proposals. Sound understanding of Java concepts, understanding of the technologies in the various architecture tiers - presentation, middleware, data access and integration to propose solution using Java /open-source technologies Design modules that are scalable, reusable, modular, secure. Clearly communicates design decisions, roadblocks and timelines to key stakeholders Adheres to all industry best practices and standards for Agile/Scrum Frameworks adopted by the Organization including but not limited to daily stand-ups, grooming, planning, retrospectives, sprint reviews, demos, and analytics via systems (JIRA) administration to directly support initiatives set by Product Management and the Organization at large Actively participate in Production stabilization and lead system software improvements along with team members. Technical Skills: Candidate Should have at least total 8+ years of experience in IT software development/design architecture. 3+ experience as an Architect in building distributed, highly available and scalable, microservice-based Cloud Native architecture Experience in one or more open-source Java frameworks such as Spring Boot, Spring Batch, Quartz, Spring Cloud, Spring Security, BPM, etc. Experience in single page web application framework like Angular. Experience with at least one type messaging system (Apache Kafka (Required), RabbitMQ) Experience with at least one RDBMS (MySQL, PostgreSQL, Oracle) Experience with at least one document-oriented DB (MongoDB, Preferably Couchbase DB) Experience with NoSQL DB like Elasticsearch Proficient in creating design documents - LLD documents with UML Good Exposure on Design Patterns, Microservices Architecture Design patterns and 12 factor application Experience working with observability/monitoring framework (Prometheus/Grafana, ELK) along with any APM tool Ability to conceptualize end-to-end system components across a wide range of technologies and translate into architectural design patterns for implementation Knowledge of security systems like Oauth 2, Keyclaok and SAML Familiarity with source code version control systems like Git/SVN Experience using, designing, and building REST/GRPC/ GraphQL/Web Service APIs Production experience with container orchestration (Docker, Kubernetes/CI/CD) and maintaining production environments Good understanding of public clouds GCP, AWS Etc. Good Exposure on API Gateways, Config servers Familiar with OWASP Experience in Telecom BSS (Business Support System) for CRM components added advantage. Immediate Joiner/30 days
Posted 1 week ago
0.0 - 3.0 years
3 - 5 Lacs
Mumbai
Work from Office
Responsibilities: Designing, developing and maintaining core system features, services and engines Collaborating with a cross functional team of the backend, Mobile application, AI, signal processing, robotics Engineers, Design, Content, and Linguistic Team to realize the requirements of conversational social robotics platform which includes investigate design approaches, prototype new technology, and evaluate technical feasibility Ensure the developed backend infrastructure is optimized for scale and responsiveness Ensure best practices in design, development, security, monitoring, logging, and DevOps adhere to the execution of the project. Introducing new ideas, products, features by keeping track of the latest developments and industry trends Operating in an Agile/Scrum environment to deliver high quality software against aggressive schedules Requirements Proficiency in distributed application development lifecycle (concepts of authentication/authorization, security, session management, load balancing, API gateway), programming techniques and tools (application of tested, proven development paradigms) Proficiency in working on Linux based Operating system. Proficiency in at least one server-side programming language like Java. Additional languages like Python and PHP are a plus Proficiency in at least one server-side framework like Servlets, Spring, java spark (Java). Proficient in using ORM/Data access frameworks like Hibernate,JPA with spring or other server-side frameworks. Proficiency in at least one data serialization framework: Apache Thrift, Google ProtoBuffs, Apache Avro,Google Json,JackSon etc. Proficiency in at least one of inter process communication frameworks WebSocket's, RPC, message queues, custom HTTP libraries/frameworks ( kryonet, RxJava ), etc. Proficiency in multithreaded programming and Concurrency concepts (Threads, Thread Pools, Futures, asynchronous programming). Experience defining system architectures and exploring technical feasibility tradeoffs (architecture, design patterns, reliability and scaling) Experience developing cloud software services and an understanding of design for scalability, performance and reliability Good understanding of networking and communication protocols, and proficiency in identification CPU, memory I/O bottlenecks, solve read write-heavy workloads. Proficiency is concepts of monolithic and microservice architectural paradigms. Proficiency in working on at least one of cloud hosting platforms like Amazon AWS, Google Cloud, Azure etc. Proficiency in at least one of database SQL, NO-SQL, Graph databases like MySQL, MongoDB, Orientdb Proficiency in at least one of testing frameworks or tools JMeter, Locusts, Taurus Proficiency in at least one RPC communication framework: Apache Thrift, GRPC is an added plus Proficiency in asynchronous libraries (RxJava), frameworks (Akka),Play,Vertx is an added plus Proficiency in functional programming ( Scala ) languages is an added plus Proficiency in working with NoSQL/graph databases is an added plus Proficient understanding of code versioning tools, such as Git is an added plus Working Knowledge of tools for server, application metrics logging and monitoring and is a plus Monit, ELK, graylog is an added plus Working Knowledge of DevOps containerization utilities like Ansible, Salt, Puppet is an added plus Working Knowledge of DevOps containerization technologies like Docker, LXD is an added plus Working Knowledge of container orchestration platform like Kubernetes is an added plus
Posted 1 week ago
4.0 - 8.0 years
7 - 12 Lacs
Bengaluru
Work from Office
Software Engineer ( MX Dashboard ) - Ruby on Rails/Python/Java/Scala/NodeJS/ JavaScript, 4+years of Experience About the Role: Our Dashboard teams build and maintain our Web applications, which manage millions of network devices from our cloud. Our customers use the Meraki Dashboard to monitor and configure critical IT infrastructure that serves tens of millions of people every day. As a Software Engineer on MX Dashboard team, you will collaborate with firmware and other Backend/SRE/Dashboard engineers to architect, design, and build a large-scale system running MX SDWAN & Security features. You will enable connections between over a million network nodes and our SDWAN & Security customers relying on our products to serve tens of millions of people. With the large footprint that we have, quality is our highest priority. MX Dashboard team is responsible for delivering a simple to use but very powerful, scalable, and groundbreaking cloud-managed service to customers. With help from product managers and firmware engineers, you will construct intuitive but powerful systems that will be used by customers via the Meraki Dashboard. What you will work on: Solve challenging architecture problems to build scalable and extendable systems. Work with firmware engineers and PM to build intuitive and powerful workflows to handle containers. Coordinate and align knowledge and opinions between firmware, SRE, and Dashboard developers. With the help of other engineers, implement sophisticated Backend & Dashboard systems to handle MX SDWAN & Security solutions. Identify and solve performance bottlenecks in our Backend architecture. Take complete ownership from conception to production release by leveraging your ability to influence, facilitate, and work collaboratively across teams. Lead, mentor, and spread best practices to other specialists on the team. You are an ideal fit if you have: 4 + years of experience writing professional production code and tests for large scale systems 3 + years of experience in Backend & Full Stack technologies Ruby on Rails/Python/Scala/Java/NodeJS/JavaScript. Can implement efficient database design and query performance in a relational database (Postgres, SQL) Experience with Container solutions (Kubernetes) Strategic and product-oriented approach with a desire to understand users Outstanding communication skills Bonus points for any of the following: Experience or interest in Security or Networking Experience in building rich web UIs with React (and Redux) Familiarity working with Observability tools like ELK, Grafana etc is a plus
Posted 1 week ago
7.0 - 12.0 years
9 - 14 Lacs
Mumbai
Work from Office
Skill Profile SRE Client Platform7+ years of relevant experience as an SRE/DevOps Engineer Have a background in either Systems Administration or Software Engineering Strong experience with major public Cloud Providers (ideally GCP but this is not a must have) Strong experience with Docker and Kubernetes. Strong experience with IaC (Terraform) Strong understanding of GitOps concepts and tools (ideally Flux) Excellent knowledge of technical architecture and modern design patterns, including micro-services, serverless functions, NoSQL, RESTful APIs, etc. Ability to set up and support CI/CD pipelines and tooling using Gitlab. Proficiency in a high-level programming language such as Python, Ruby or Go Experience with monitoring, log aggregation and alerting tooling (GCP Logging, Prometheus, Grafana). Additional SRE Data Platform:- SRE Data PlatformLinux administration skills and a deep understanding of networking and TCP/IP. Experience with the major cloud providers and Terraform. Knowledge of technical architecture and modern-day design patterns, including micro-services, serverless functions, NoSQL, RESTful APIs, etc. Demonstrable skills in a Configuration Management tool like Ansible. Experience in setting up and supporting CI/CD pipelines and tooling such as GitHub or Gitlab CI Proficiency in a high-level programming language such as Python or Go. Experience with monitoring, log aggregation, and alerting tooling (ELK, Prometheus, Grafana, etc). Experience with Docker and Kubernetes Experience with secret management tools like Hashicorp Vault is deemed a plus Proficient in applying SRE core tenets, including SLI/SLO/SLA measurement, toil elimination, and reliability modeling for optimizing system performance and resilience. Experience with cloud-native tools like Cluster API, service mesh, KEDA, OPA, Kubernetes Operators Experience with big data technologies such as NoSQL/RDBMS(PostgreSQL, Oracle, MongoDB), Redis, Spark, Rabbit, Kafka, etc. Experience in troubleshooting and monitoring large-scale distributed systems
Posted 1 week ago
4.0 - 7.0 years
6 - 9 Lacs
Bengaluru
Work from Office
About the Role: Our Dashboard teams build and maintain our Web applications, which manage millions of network devices from our cloud. Our customers use the Meraki Dashboard to monitor and configure critical IT infrastructure that serves tens of millions of people every day. As a Software Engineer on MX Dashboard team, you will collaborate with firmware and other Backend/SRE/Dashboard engineers to architect, design, and build a large-scale system running MX SDWAN & Security features. You will enable connections between over a million network nodes and our SDWAN & Security customers relying on our products to serve tens of millions of people. With the large footprint that we have, quality is our highest priority. MX Dashboard team is responsible for delivering a simple to use but very powerful, scalable, and groundbreaking cloud-managed service to customers. With help from product managers and firmware engineers, you will construct intuitive but powerful systems that will be used by customers via the Meraki Dashboard. What you will work on: Solve challenging architecture problems to build scalable and extendable systems. Work with firmware engineers and PM to build intuitive and powerful workflows to handle containers. Coordinate and align knowledge and opinions between firmware, SRE, and Dashboard developers. With the help of other engineers, implement sophisticated Backend & Dashboard systems to handle MX SDWAN & Security solutions. Identify and solve performance bottlenecks in our Backend architecture. Take complete ownership from conception to production release by leveraging your ability to influence, facilitate, and work collaboratively across teams. Lead, mentor, and spread best practices to other specialists on the team. You are an ideal fit if you have: 4 + years of experience writing professional production code and tests for large scale systems 3 + years of experience in Backend & Full Stack technologies Ruby on Rails/Python/Scala/Java/NodeJS/JavaScript. Can implement efficient database design and query performance in a relational database (Postgres, SQL) Experience with Container solutions (Kubernetes) Strategic and product-oriented approach with a desire to understand users Outstanding communication skills Bonus points for any of the following: Experience or interest in Security or Networking Experience in building rich web UIs with React (and Redux) Familiarity working with Observability tools like ELK, Grafana etc is a plus
Posted 1 week ago
3.0 - 5.0 years
5 - 7 Lacs
Kochi, Coimbatore, Thiruvananthapuram
Work from Office
Must have skills installation, configuration and management of Linux /Windows systems Good to have skills:" JIRA/Confluence Experience 3.5 - 5 years of experience is required Educational QualificationGraduation (Accurate educational details should capture) Job Summary As a L2 Cloud Operations Engineer , you will be operating an e-commerce solution built over on-prem and cloud infrastructure. You will be involved in maintaining and improving the clients business platforms and also will be responsible for the site reliability and platform stability. You will be expected to respond to incidents, support on problems, execute changes and be part of a project to improve or reengineer the platform. Roles and Responsibilities Continuous monitoring of the platforms performance and uptime Fast identification and resolution of incidents Resolution of service requests Managing the platform configuration to ensure it is optimized and up to date Improved efficiency by automating routine tasks Professional and Technical Skills You must have a strong technical aptitude and an organized, process driven work ethic. 3 .5 -5 years of relevant experience with installation, configuration and management of Linux / Windows systems. Strong working experience in managing and maintaining public clouds like AWS, Azure or GCP. Strong experience in setting up and configuring monitoring tools like Prometheus, Grafana, Zabbix etc Strong Experience with installation/configuration of Java application servers such as Jboss /WebLogic/Tomcat and also analyzing application logs, GC logs for troubleshooting performance and functional issues. Hands on experience in cloud provisioning tools like Terraform/CloudFormation will be an added advantage" Hands on experience with Docker/Kubernetes will be an added advantage Experience in ELK/Kafka/ Openshift /Python script will be an added advantage Good knowledge of SQL and" NoSQL databases like MySQL/Oracle/PostgreSQL/DynamoDB/MongoDB/Cassandra/Redis You will" have strong written and verbal communications skills and a track record for providing high customer satisfaction. Develop automation scripts as needed to enhance operational efficiencies Has prior experience in supporting Jira/Confluence or any other service management tool Prior experience working in Agile environment will be an advantage. Experience: 3.5 -5 years of experience is required Educational Qualification: Graduation (Accurate educational details should capture)
Posted 1 week ago
8.0 - 10.0 years
15 - 30 Lacs
Pune
Work from Office
Role Overview We are looking for experienced DevOps Engineers (8+ years) with a strong background in cloud infrastructure, automation, and CI/CD processes. The ideal candidate will have hands-on experience in building, deploying, and maintaining cloud solutions using Infrastructure-as-Code (IaC) best practices. The role requires expertise in containerization, cloud security, networking, and monitoring tools to optimize and scale enterprise-level applications. Key Responsibilities Design, implement, and manage cloud infrastructure solutions on AWS, Azure, or GCP. Develop and maintain Infrastructure-as-Code (IaC) using Terraform, CloudFormation, or similar tools. Implement and manage CI/CD pipelines using tools like GitHub Actions, Jenkins, GitLab CI/CD, BitBucket Pipelines, or AWS CodePipeline. Manage and orchestrate containers using Kubernetes, OpenShift, AWS EKS, AWS ECS, and Docker. Work on cloud migrations, helping organizations transition from on-premises data centers to cloud-based infrastructure. Ensure system security and compliance with industry standards such as SOC 2, PCI, HIPAA, GDPR, and HITRUST. Set up and optimize monitoring, logging, and alerting using tools like Datadog, Dynatrace, AWS CloudWatch, Prometheus, ELK, or Splunk. Automate deployment, configuration, and management of cloud-native applications using Ansible, Chef, Puppet, or similar configuration management tools. Troubleshoot complex networking, Linux/Windows server issues, and cloud-related performance bottlenecks. Collaborate with development, security, and operations teams to streamline the DevSecOps process. Must-Have Skills 3+ years of experience in DevOps, cloud infrastructure, or platform engineering. Expertise in at least one major cloud provider: AWS, Azure, or GCP. Strong experience with Kubernetes, ECS, OpenShift, and container orchestration technologies. Hands-on experience in Infrastructure-as-Code (IaC) using Terraform, AWS CloudFormation, or similar tools. Proficiency in scripting/programming languages like Python, Bash, or PowerShell for automation. Strong knowledge of CI/CD tools such as Jenkins, GitHub Actions, GitLab CI/CD, or BitBucket Pipelines. Experience with Linux operating systems (RHEL, SUSE, Ubuntu, Amazon Linux) and Windows Server administration. Expertise in networking (VPCs, Subnets, Load Balancing, Security Groups, Firewalls). Experience in log management and monitoring tools like Datadog, CloudWatch, Prometheus, ELK, Dynatrace. Strong communication skills to work with cross-functional teams and external customers. Knowledge of Cloud Security best practices, including IAM, WAF, GuardDuty, CVE scanning, vulnerability management. Good-to-Have Skills Knowledge of cloud-native security solutions (AWS Security Hub, Azure Security Center, Google Security Command Center). Experience in compliance frameworks (SOC 2, PCI, HIPAA, GDPR, HITRUST). Exposure to Windows Server administration alongside Linux environments. Familiarity with centralized logging solutions (Splunk, Fluentd, AWS OpenSearch). GitOps experience with tools like ArgoCD or Flux. Background in penetration testing, intrusion detection, and vulnerability scanning. Experience in cost optimization strategies for cloud infrastructure. Passion for mentoring teams and sharing DevOps best practices.
Posted 1 week ago
Upload Resume
Drag or click to upload
Your data is secure with us, protected by advanced encryption.
Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.
Accenture
36723 Jobs | Dublin
Wipro
11788 Jobs | Bengaluru
EY
8277 Jobs | London
IBM
6362 Jobs | Armonk
Amazon
6322 Jobs | Seattle,WA
Oracle
5543 Jobs | Redwood City
Capgemini
5131 Jobs | Paris,France
Uplers
4724 Jobs | Ahmedabad
Infosys
4329 Jobs | Bangalore,Karnataka
Accenture in India
4290 Jobs | Dublin 2