We re looking for a
Senior DevOps Engineer
to join Procore s Product & Technology Team
. Procore software solutions aim to improve the lives of everyone in construction and the people within Product & Technology are the driving force behind our innovative, top-rated global platform. We re a customer-centric group that encompasses engineering, product, product design and data, security and business systems. In this role, youll collaborate with engineering teams, product managers, and architects to design, build, and manage Procore s streaming data platform. Youll play a crucial role in ensuring the platforms reliability, scalability, and performance through automation and operational excellence.This role reports to a
Senior Engineering Manager
and is based in our Bangalore office
, with an immediate start.
What youll do:
- Design, implement, and manage highly scalable and resilient streaming data infrastructure, with a strong focus on Kafka and its ecosystem.
- Develop and maintain automation tools and infrastructure-as-code (IaC) to provision, configure, and operate streaming data services.
- Collaborate closely with software engineers to optimize streaming data pipelines and applications for performance, reliability, and cost-efficiency.
- Implement and manage monitoring, alerting, and logging solutions for the streaming data platform to proactively identify and resolve issues.
- Ensure the security, reliability, and performance of data platform services through continuous improvement and best practices.
- Actively participate in code reviews, system design discussions, and troubleshooting complex production issues.
- Participate in on-call rotations to support and maintain the stability and availability of the streaming data platform.
- Contribute across the full DevOps lifecycle, from infrastructure design and implementation to deployment, operations, and support., stay current with industry trends and best practices in DevOps, distributed systems, streaming platforms, and how AI tools can enhance engineering efficiency.
What were looking for:
- 5+ years of professional experience in a DevOps, Site Reliability Engineering (SRE), or similar infrastructure-focused role, with a strong emphasis on streaming data platforms.
- Extensive hands-on experience with Kafka, including administration, performance tuning, and troubleshooting.
- Proficiency in designing and implementing scalable infrastructure solutions using modern programming languages (e.g., Python, Java) for automation and tooling.
- Demonstrated experience leveraging AI-powered development tools (e.g., GitHub Copilot, ChatGPT for coding assistance, AI-driven observability) to enhance efficiency and productivity in engineering workflows.
- Deep understanding of cloud platforms (e.g., AWS, Azure, GCP), including experience with compute, storage, networking, and managed service offerings.
- Strong expertise in infrastructure-as-code (IaC) tools such as Terraform, Ansible, or similar.
- Experience with containerization and orchestration technologies (e.g., Docker, Kubernetes) familiarity with monitoring and logging tools (e.g., Prometheus, Grafana).
- Exceptional communication, collaboration, and problem-solving skills, with a proactive approach to technical challenges, experience collaborating effectively within global technical teams.