Jobs
Interviews

17543 Terraform Jobs - Page 39

Setup a job Alert
JobPe aggregates results for easy application access, but you actually apply on the job portal directly.

2.0 years

0 Lacs

India

On-site

JoVE is the world-leading producer and provider of science video solutions with the mission to improve scientific research and education. Millions of scientists, educators and students use JoVE for their research, teaching and learning. Our institutional clients comprise over 1,000 universities, colleges, and biopharma companies, including such leaders as Harvard, MIT, Yale, and Stanford. As a rapidly growing company, with offices in the USA, UK, Australia, and India servicing clients in over 60 countries, we are seeking talented and ambitious individuals to join our company.60 countries, we are seeking talented individuals to join our company. The Role We are looking for a Junior Site Reliability Engineer who will be part of our centralized Site Reliability Team. You will play an integral role in leading the deployment of highly scalable systems, optimization, documentation, and support of the infrastructure components of JoVE’s software products hosted on AWS. Cloud Infrastructure and Operations are critical in enabling JoVE to provide users with our technology offerings. Responsibilities: Design, build, test, and deploy cloud-native applications and microservices using IaC tools like Terraform and Crossplane. Maintain availability, latency, performance, efficiency, monitoring/observability, emergency response, capacity planning, setting and maintaining SLOs, SLIs, and Error Budgets, and creating dashboards. Plan for automation to reduce toil and increase development velocity. Perform application-specific production support, incident management, change management, problem management, RCAs, and service restoration as needed. Actively look for opportunities to improve the availability and performance of the system by applying the learnings from monitoring and observation. Collaborate with software development teams in the release management process to shape the future roadmap and establish strong operational readiness across teams. Spearhead implementation of reliability and observability tools (like Groundcover, Prometheus, Grafana, etc.) Support Infrastructure squad On-call practice and participate in 24x7 on-call rotations. Requirements: 2+ years of professional experience as a Software Engineer and Site Reliability Engineer (SRE). Extensive in-depth experience with cloud-based provisioning, monitoring, troubleshooting, and related SRE and DevOps technologies, in addition to networking knowledge. MUST have working experience with AWS infrastructure. MUST understand AWS VPC, subnets, Network ACLs, Security Groups, IAM Role, EKS. MUST have experience of using Crossplane MUST have working knowledge of GitOps, FluxCD, or ArgoCD Experience configuring Kubernetes RBAC Authorization, Ingress controller, ServiceAccount, and AWS role annotations. Basic experience with monitoring, and observability systems such as DataDog, Prometheus, Grafana, Kibana, CloudWatch. Ability to triage and resolve incidents and lead incident investigations. Experience working in a 24/7 on-call, highly transactional, or streaming production environment. Experience with Kubernetes Operators is a plus. Why Join JoVE? When working with JoVE, you can expect compensation packages competitively placed within the local market. You will make a direct impact in accelerating science research and in improving student learning in science education. Opportunity to work with global teams and in an environment that promotes innovation and collaboration. Our strong promotion from within culture draws a clear path to advance your career with us.

Posted 1 week ago

Apply

7.0 - 11.0 years

0 Lacs

noida, uttar pradesh

On-site

As a Principal Site Reliability Engineer at UKG, you play a crucial role in enhancing and supporting service delivery processes through developing software solutions. Your responsibilities include building and managing CI/CD deployment pipelines, automated testing, capacity planning, performance analysis, monitoring, alerting, chaos engineering, and auto remediation. It is essential to stay updated with current technology trends, innovate, and ensure a flawless customer experience by deploying services efficiently and consistently. Your primary duties involve engaging in the lifecycle of services, defining and implementing standards, supporting product and engineering teams, improving system performance, collaborating with professionals, guiding junior team members, and actively participating in incident response. You will also contribute to increasing operational efficiency, effectiveness, and service quality by treating operational challenges as software engineering problems. To be successful in this role, you should have an engineering degree or a related technical discipline, along with at least 7 years of hands-on experience in Engineering or Cloud. Additionally, a minimum of 5 years" experience with public cloud platforms, 3 years" experience in configuration and maintenance of applications/systems infrastructure, and knowledge of coding in higher-level languages are required. Familiarity with Cloud-based applications, Containerization Technologies, industry standards like Terraform, Ansible, and fundamentals in Computer Science, Cloud Architecture, Security, or Network Design is also essential. UKG is a dynamic organization at the forefront of workforce management and human capital management solutions. As we continue to grow and innovate, we are committed to fostering diversity and promoting inclusion in the workplace. If you require any disability accommodations during the application and interview process, please reach out to UKGCareers@ukg.com. Join us on this exciting journey towards a brighter tomorrow!,

Posted 1 week ago

Apply

4.0 - 8.0 years

0 Lacs

karnataka

On-site

As a qualified candidate for this position, you will be responsible for various key tasks related to systems development and operations. Your main responsibilities will include determining and developing user requirements for systems in production to ensure maximum usability. You will also be expected to possess excellent verbal and written communication skills to effectively collaborate with team members and stakeholders. In addition, your role will involve utilizing your expertise in Linux, scripting, CI/CD, Kubernetes, AWS, DevOps, Docker, monitoring, and Terraform to contribute to the successful implementation and maintenance of various systems and processes within the organization. Your technical proficiency in these areas will be crucial in driving innovation and efficiency across the organization. Overall, success in this position will be defined by your ability to effectively meet the requirements of the role, collaborate with colleagues, and contribute to the organization's broader goals and objectives. Your contributions will play a significant role in advancing the organization's technological capabilities and ensuring the seamless operation of critical systems.,

Posted 1 week ago

Apply

3.0 - 7.0 years

0 Lacs

karnataka

On-site

As a candidate for the role at Wind Pioneers, your main responsibility will be to ensure that our flagship software products are driven by cutting-edge science and methodologies. You will play a crucial role in implementing and refining various analyses related to wind data analysis, wind resource assessments, site design, wake modeling, and other intermediary analyses. The enhancements you make in these analyses will need to be seamlessly integrated into our codebase and thoroughly validated. This position is ideal for a detail-oriented scientist or engineer who is passionate about advancing the state-of-the-art in wind farm design and development. At Wind Pioneers, our vision is to lead the world in designing and evaluating new wind farm sites. We are dedicated to creating a set of tools, approaches, and processes that elevate the technical management of wind farm development to a significantly higher level of sophistication compared to standard industry practices. The company relies on its own software as a testing ground for innovative and advanced methodologies, providing you with a unique opportunity to be at the forefront of wind farm design and development. Your role will involve driving improvements to our software from conceptualization to commercial deployment. This requires a deep understanding of scientific and engineering principles to implement new analytical approaches within our software stack and conduct comprehensive validation studies. Key Responsibilities include two main areas: A. Creating Scientific Services: - Utilizing research findings to enhance the accuracy and efficiency of wind resource assessment processes by incorporating new technologies, methodologies, and data sources. - Generating detailed technical reports, documentation, and presentations to effectively communicate research findings, tool developments, and project outcomes. - Conducting research and development tasks, including validation studies. B. Software Engineering: - Assisting the development team in creating high-quality web applications for wind farm design. - Engaging in Data Engineering using technologies like Postgres, BigQuery, Pub/Sub, and Terraform to build event-driven systems and data lakes, particularly for geospatial data. - Leveraging Python and optionally Rust to develop and maintain performance analysis tools for designing and optimizing multi-GW scale wind farms. Candidate Requirements: - Enthusiasm for wind resources and the role of renewable energy in addressing climate change. - Bachelor's or master's degree in a scientific or engineering discipline from a reputable institution. PhD holders are also encouraged to apply. - 3-5 years of relevant experience, demonstrating independent work and initiative. - Wind industry experience is preferred but not mandatory. - Proficiency with Git and Git Flow is beneficial. - Basic knowledge of software development and Python is advantageous. - Excellent written English skills. - International experience is desirable. - Self-directed and proactive work approach. - Excitement for working in a dynamic, high-growth startup environment. - Positive attitude and passion for wind energy. Wind Pioneers Offering: - Join a focused team with a clear vision dedicated to revolutionizing wind farm project discovery and evaluation. - Utilize Wind Pioneers" advanced in-house tools to design top-tier wind farms. - Contribute to the development of Wind Pioneers" flagship tool while benefiting from using it as an end user. - Learn and collaborate closely with our Product Architect and Senior Engineer. - Enjoy a friendly and relaxed office atmosphere and team culture. - Flexible working conditions. - Competitive salary with the opportunity for a six-monthly bonus through Wind Pioneers" revenue share scheme.,

Posted 1 week ago

Apply

0 years

0 Lacs

Bengaluru, Karnataka, India

On-site

Greenlight is the leading family fintech company on a mission to help parents raise financially smart kids. We proudly serve more than 6 million parents and kids with our award-winning banking app for families. With Greenlight, parents can automate allowance, manage chores, set flexible spend controls, and invest for their family’s future. Kids and teens learn to earn, save, spend wisely, and invest. At Greenlight, we believe every child should have the opportunity to become financially healthy and happy. It’s no small task, and that’s why we leap out of bed every morning to come to work. Because creating a better, brighter future for the next generation depends on it. Greenlight is looking for a Staff Engineer, Production Operations to join our growing team! As a Staff Engineer, you will be a technical leader and individual contributor within our production operations function. You will be responsible for designing, building, and maintaining highly reliable, scalable, and performant cloud infrastructure and systems. You will play a critical role in driving technical excellence, mentoring junior engineers, and solving our most complex scalability and reliability challenges. What you will be doing: Lead the design, implementation, and evolution of Greenlight's core cloud infrastructure and SRE practices to ensure high availability, scalability, and performance Act as a technical authority for complex SRE and cloud engineering challenges, providing expert guidance and solutions Drive significant architectural improvements to enhance system reliability, resilience, and operational efficiency Develop, maintain, and optimize our cloud infrastructure using Infrastructure as Code (primarily Terraform) and automation tools Collaborate closely with development and security teams to embed SRE principles into the software development lifecycle, promoting secure and reliable coding practices Design and implement robust monitoring, logging, and alerting solutions to provide comprehensive visibility into system health Participate in and lead incident response, performing deep dive root cause analysis, and driving actionable blameless postmortems to prevent recurrence Mentor and provide technical guidance to other SRE and Cloud Engineers, contributing to their growth and the team's overall technical capabilities Research, evaluate, and advocate for new technologies and tools that can improve our operational posture and efficiency Contribute to the strategic planning and roadmap development for the SRE and Cloud Engineering functions Enhance existing services and applications to increase availability, reliability, and scalability in a microservices environment Build and improve engineering tooling, process, and standards to enable faster, more consistent, more reliable, and highly repeatable application delivery What you should bring: Technical Leadership: Lead complex technical projects and mentor engineers Communication: Articulate complex technical concepts clearly SRE Expertise: Apply SRE principles (SLIs, SLOs, error budgets) in production Distributed Systems: Understand and troubleshoot complex issues in distributed systems Monitoring & Alerting: Design and optimize monitoring, logging, and alerting systems (e.g., Datadog, Prometheus) Cloud Mastery (AWS): Expert-level knowledge of AWS services (e.g., EC2, S3, EKS) Infrastructure as Code (Terraform): Master IaC for cloud infrastructure management Containerization: Strong experience with Docker and Kubernetes in production Automation: Bias for automation and building self-healing systems Problem Solving: Exceptional analytical and problem-solving skills, proactively identifying bottlenecks Technologies we use: AWS MySQL, DynamoDB, Redis GitHub Actions for CI pipelines Kubernetes (specifically EKS) Ambassador, Helm, Argo CD, LinkerD REST, gRPC, graphQL React, Redux, Swift, Node.js, Kotlin, Java, Go, Python Datadog, Prometheus Who we are: It takes a special team to aim for a never-been-done-before mission like ours. We’re looking for people who love working together because they know it makes us stronger, people who look to others and ask, “How can I help?” and then “How can we make this even better?” If you’re ready to roll up your sleeves and help parents raise a financially smart generation, apply to join our team. Greenlight is an equal opportunity employer and will not discriminate against any employee or applicant based on age, race, color, national origin, gender, gender identity or expression, sexual orientation, religion, physical or mental disability, medical condition (including pregnancy, childbirth, or a medical condition related to pregnancy or childbirth), genetic information, marital status, veteran status, or any other characteristic protected by federal, state or local law. Greenlight is committed to an inclusive work environment and interview experience. If you require reasonable accommodations to participate in our hiring process, please reach out to your recruiter directly or email recruiting@greenlight.me.

Posted 1 week ago

Apply

0 years

0 Lacs

India

Remote

This is a remote position. About Us Simbian® is building Agentic AI platform for cybersecurity. Founded by repeat successful security founders, we have gathered an excellent cohort of employees, partners, and customers. Our mission is to solve security using AI and our core values are excellence, replication, and intellectual honesty. Our promise is to make Simbian the best workplace of your career and we believe a small group of thoughtful passionate people can make all the positive difference in the world. To fuel our fast growth, we are seeking an exceptional candidate who shares our core values of excellence (being the world's best at our craft), replication (share your best ideas with others), and intellectual honesty (tell the truth even if it's bitter). Our AI Agents automate security operations and provide our customers 10x leverage. Our customers include some of the world's largest companies. Our initial use cases include: SOC alert triage and investigation Prioritization and classification of vulnerabilities AI based threat hunting What you’ll do: Define DevOps strategy for cloud-native applications and infrastructure. Lead and mentor DevOps engineers and SREs across projects. Establish and enforce best practices for CI/CD, infrastructure automation, and monitoring. Align DevOps efforts with business goals and cloud cost optimization. Recommend right-sizing of infrastructure and implement auto-scaling strategies. Design and manage scalable, secure, and highly available cloud infrastructure (AWS, Azure, GCP). Implement Infrastructure as Code (IaC) using Terraform, CloudFormation etc Manage multi-cloud or hybrid environments, ensuring resilience and compliance. Architect and manage CI/CD pipelines for automated testing, integration, and deployment. Promote automation-first approaches in infrastructure provisioning, configuration, and release workflows. Integrate security checks (DevSecOps) into the CI/CD process. Oversee observability stack: monitoring, alerting, tracing, and logging tools (e.g., Prometheus, Grafana, ELK). Define and track SLA/SLO/SLI metrics.\ Lead incident response and postmortems for critical production issues. Implement cloud security best practices, including secrets management, IAM, encryption, and network policies. Ensure compliance with regulatory standards (SOC2) through automation and audits Oversee release cycles, rollout strategies (blue/green, canary), Manage Dev/staging/production environments for consistency and reliability. Work closely with development, QA, product, and security teams to ensure smooth delivery. Advocate for a DevOps culture: transparency, collaboration, and continuous improvement. What You’ll Get: High autonomy and visibility with leadership and founders A collaborative, transparent work culture with a bias for action Competitive salary with generous equity in a potentially large company Opportunity to work with some of the world's most talented and friendliest teammates

Posted 1 week ago

Apply

0 years

0 Lacs

India

On-site

hackajob is collaborating with OneAdvanced to connect them with exceptional tech professionals for this role. Proficient in scripting and programming languages such as Python, Go, and Bash. Strong hands-on experience with automation frameworks and tools including Terraform, Ansible, Chef, and Puppet. Familiarity with automation scripting tools for infrastructure and operations (e.g. Python, Terraform, Ansible). Experience working with AI-driven operations tools and AIOps platforms such as Moogsoft, BigPanda, Dynatrace, or custom ML-based pipelines. In-depth knowledge of CI/CD, GitOps, and event-driven systems for modern DevOps practices. Solid background in Linux systems and containerized environments like Docker and Kubernetes.

Posted 1 week ago

Apply

3.0 years

0 Lacs

Hyderabad, Telangana, India

On-site

Location: Hyderabad Experience: 3+ Years Type: Full-time About Us It is a leading AI Product Company that simulates user actions through AI — Universal Search, last-mile navigation, and contextual guidance. We partner with top product companies and have delivered results like: 70% reduction in support ticket volume 30%+ improvement in product adoption, higher conversions, and NPS Roles & Responsibilities Automate deployment and DevOps processes using Infrastructure as Code (IaC). Build, maintain, and optimize CI/CD pipelines for robust software delivery. Monitor and maintain system performance to ensure high availability and scalability. Collaborate closely with developers and product teams for seamless deployment integration. Manage cloud infrastructure (AWS or Azure), ensuring high security and compliance. Implement container orchestration using Docker and Kubernetes. Conduct Vulnerability Assessment & Penetration Testing (VAPT) and integrate InfoSec practices. Use IaC tools like Terraform, Ansible, or Pulumi (TypeScript preferred). Set up environments for development, testing, and production on secure cloud platforms. Perform periodic security audits and assessments aligned with compliance standards (SOC 2, ISO 27001). Integrate security best practices directly into CI/CD workflows. Requirements 1-3 years of experience as a DevOps Engineer or in a similar role. Experience in cloud platforms: AWS (preferred), GCP, or Azure. Strong command over Python, Bash/Shell, TypeScript, YAML, and optionally Groovy/Java. Solid knowledge of Linux systems and cloud networking. Hands-on experience with Infrastructure as Code (IaC) - Terraform, Pulumi, or similar. Deep understanding of CI/CD, containerization (Docker), and orchestration (Kubernetes). Familiarity with microservices architecture and agile methodologies. Practical knowledge in InfoSec principles, tools (e.g., WAF, SIEM), and frameworks (SOC 2, ISO 27001). Experience with security tooling and VAPT processes is highly desirable. Bonus: Experience with Pulumi using TypeScript. Why Join Us? Here, you’ll work at the forefront of AI and ML innovation, shaping the future of enterprise technology. Join a team that’s leveraging cutting-edge AI to power real-time automation, insights, and intelligent decision-making. We Offer Competitive salary High-impact commission structure Endless opportunities for learning and growth Come unlock your potential — and help us transform industries along the way.

Posted 1 week ago

Apply

2.0 years

0 Lacs

Bengaluru, Karnataka, India

On-site

Job Summary We are seeking a skilled and innovative Cloud Engineer to join our team. As a Cloud Engineer, you will be responsible for developing and maintaining cloud-based solutions, with a focus on coding complex problems, automation using Golang and Python, and collaborating with the Site Reliability Engineering (SRE) team for feature deployment in production. Additionally, the ideal candidate should be proficient in utilizing AI tools like Copilot to enhance productivity in the areas of automation, documentation, and unit test writing. Responsibilities Develop, test, and maintain cloud-based applications and services using Golang and Python. Write clean, efficient, and maintainable code to solve complex problems and improve system performance. Collaborate with cross-functional teams to understand requirements and design scalable and secure cloud solutions. Automate deployment, scaling, and monitoring of cloud-based applications and infrastructure. Work closely with the SRE team to ensure smooth feature deployment in production environments. Utilize AI tools like Copilot to enhance productivity in automation, documentation, and unit test writing. Troubleshoot and resolve issues related to cloud infrastructure, performance, and security. Stay up to date with emerging technologies and industry trends to continuously improve cloud-based solutions. Participate in code reviews, knowledge sharing sessions, and contribute to the improvement of development processes. Job Requirements Strong programming skills in Golang and Python. Proficiency in using AI tools like Copilot to enhance productivity in automation, documentation, and unit test writing. Solid understanding of cloud computing concepts and services (e.g., AWS, Azure, Google Cloud). Experience with containerization technologies (e.g., Docker, Kubernetes) and infrastructure-as-code tools (e.g., Terraform, CloudFormation). Proficient in designing and implementing RESTful APIs and microservices architectures. Familiarity with CI/CD pipelines and tools (e.g., Jenkins, GitLab CI/CD). Knowledge of networking concepts, security best practices, and system administration. Excellent problem-solving skills and ability to work in a fast-paced, collaborative environment. Strong communication and interpersonal skills to effectively collaborate with cross-functional teams. Preferred Skills Experience with other programming languages, such as Java, C++, or Ruby. Knowledge of database technologies (e.g., SQL, NoSQL) and data storage solutions. Familiarity with monitoring and logging tools (e.g., Prometheus, ELK stack). Understanding of Agile/Scrum methodologies and DevOps principles. Certifications in cloud technologies (e.g., AWS Certified Cloud Practitioner, Google Cloud Certified - Associate Cloud Engineer) would be a plus. If you are passionate about cloud technologies, have a strong problem-solving mindset, and enjoy working in a collaborative environment, we would love to hear from you. Join our team and contribute to building scalable, reliable, and secure cloud solutions. Please note that this job description is not exhaustive and may change based on the organization's needs. Education A Bachelor of Science Degree in Engineering or Computer Science with 2 years of experience, or a Master’s Degree; or equivalent experience is typically required. All internal movements within the Product Group via requisition will be lateral, offering valuable growth opportunities to extend your skills in a new area. Opportunities for a promotion will be reviewed in the normal course of business, aligned with our promotion process. At NetApp, we embrace a hybrid working environment designed to strengthen connection, collaboration, and culture for all employees. This means that most roles will have some level of in-office and/or in-person expectations, which will be shared during the recruitment process. Equal Opportunity Employer NetApp is firmly committed to Equal Employment Opportunity (EEO) and to compliance with all laws that prohibit employment discrimination based on age, race, color, gender, sexual orientation, gender identity, national origin, religion, disability or genetic information, pregnancy, and any protected classification. Why NetApp? We are all about helping customers turn challenges into business opportunity. It starts with bringing new thinking to age-old problems, like how to use data most effectively to run better - but also to innovate. We tailor our approach to the customer's unique needs with a combination of fresh thinking and proven approaches. We enable a healthy work-life balance. Our volunteer time off program is best in class, offering employees 40 hours of paid time off each year to volunteer with their favourite organizations. We provide comprehensive benefits, including health care, life and accident plans, emotional support resources for you and your family, legal services, and financial savings programs to help you plan for your future. We support professional and personal growth through educational assistance and provide access to various discounts and perks to enhance your overall quality of life. If you want to help us build knowledge and solve big problems, let's talk.

Posted 1 week ago

Apply

3.0 years

0 Lacs

Hyderabad, Telangana, India

On-site

Job Description Empowering projects with innovative technical solutions, our team excels in optimizing data and systems at scale while ensuring seamless application monitoring. We foster a culture of diversity, equity, inclusion, and respect, driving impactful decisions across multifaceted technologies As an Infrastructure Engineer III at JPMorgan Chase within the Consumer and Community Banking, you utilize strong knowledge of software, applications, and technical processes within the infrastructure engineering discipline. Apply your technical knowledge and problem-solving methodologies across multiple applications of moderate scope. Job Responsibilities Applies technical knowledge and problem-solving methodologies to projects of moderate scope, with a focus on improving the data and systems running at scale, and ensures end to end monitoring of applications Resolves most nuances and determines appropriate escalation path Executes conventional approaches to build or break down technical problems Identify risks and impact to applications from Infrastructure outages Making significant decisions for a project consisting of multiple technologies and applications Adds to team culture of diversity, equity, inclusion, and respect Required Qualifications, Capabilities, And Skills Formal training or certification on software engineering concepts and 3+ years applied experience A deep understanding of business technology drivers and their impact on architecture design, performance and monitoring, best practices experience building or supporting web environments on AWS, which includes working with services like EC2, ELB, RDS, and S3 Experience using DevOps tools in a cloud environment, such as Ansible, Artifactory, Docker, GitHub, Jenkins, Kubernetes, Maven, and Sonar Qube experience across the SDLC process – Design and/or Development and/or support Experience using monitoring solutions like CloudWatch, Prometheus, Datadog Experience of writing Infrastructure-as-Code (IaC), using tools like CloudFormation or Terraform Experience with one or more public cloud platforms like AWS, GCP, Azure Experience with one or more automation tools like Terraform, Puppet, Ansible Strong knowledge of one or more infrastructure disciplines such as hardware, networking terminology, databases, storage engineering, deployment practices, integration, automation, scaling, resilience, and performance assessments Strong knowledge of one or more scripting languages (e.g., Scripting, Python, etc.) Preferred Qualifications, Capabilities, And Skills A proactive approach to spotting problems, areas for improvement, and performance bottlenecks SRE mindset Culture/Approaches: To run better production systems by creating engineering solutions to operational problems.

Posted 1 week ago

Apply

0 years

0 Lacs

Kochi, Kerala, India

On-site

Role Description We are looking for an SDET- Test Engineer (Automation is preferred) with development skills in Bigdata, Aws technologies, Terraform, Python etc Hands-on experience with Manual Testing and any Test automation tools Working knowledge of SQL for data validation. Experience in at least one programming/scripting language: C# or Python Experience in development using any Bigdata technologies, Aws glue, Terraform, Spark etc Skills SDET, Manual Testing,Test Automation,Big Data,Aws

Posted 1 week ago

Apply

4.0 - 11.0 years

0 Lacs

Gurugram, Haryana, India

On-site

Hello, Greeting from Quess Corp!! Hope you are doing well we have job opportunity with one of our client Designation_ Data Engineer Location – Gurugram Experience – 4yrs to 11 Yrs Qualification – Graduate / PG ( IT) Skill Set – Data Engineer, Python, AWS, SQL Essential capabilities Enthusiasm for technology, keeping up with latest trends Ability to articulate complex technical issues and desired outcomes of system enhancements Proven analytical skills and evidence-based decision making Excellent problem solving, troubleshooting & documentation skills Strong written and verbal communication skills Excellent collaboration and interpersonal skills Strong delivery focus with an active approach to quality and auditability Ability to work under pressure and excel within a fast-paced environment Ability to self-manage tasks Agile software development practices Desired Experience Hands on in SQL and its Big Data variants (Hive-QL, Snowflake ANSI, Redshift SQL) Python and Spark and one or more of its API (PySpark, Spark SQL, Scala), Bash/Shell scripting Experience with Source code control - GitHub, VSTS etc. Knowledge and exposure to Big Data technologies Hadoop stack such as HDFS, Hive, Impala, Spark etc, and cloud Big Data warehouses - RedShift, Snowflake etc. Experience with UNIX command-line tools. Exposure to AWS technologies including EMR, Glue, Athena, Data Pipeline, Lambda, etc Understanding and ability to translate/physicalise Data Models (Star Schema, Data Vault 2.0 etc) Essential Experience It is expected that the role holder will most likely have the following qualifications and experience 4-11 years technical experience (within financial services industry preferred) Technical Domain experience (Subject Matter Expertise in Technology or Tools) Solid experience, knowledge and skills in Data Engineering, BI/software development such as ELT/ETL, data extraction and manipulation in Data Lake/Data Warehouse/Lake House environment. Hands on programming experience in writing Python, SQL, Unix Shell scripts, Pyspark scripts, in a complex enterprise environment Experience in configuration management using Ansible/Jenkins/GIT Hands on cloud-based solution design, configuration and development experience with Azure and AWS Hands on experience of using AWS Services - S3,EC2, EMR, SNS, SQS, Lambda functions, Redshift Hands on experience Of building Data pipelines to ingest, transform on Databricks Delta Lake platform from a range of data sources - Data bases, Flat files, Streaming etc.. Knowledge of Data Modelling techniques and practices used for a Data Warehouse/Data Mart application. Quality engineering development experience (CI/CD – Jenkins, Docker) Experience in Terraform, Kubernetes and Docker Experience with Source Control Tools – Github or BitBucket Exposure to relational Databases - Oracle or MS SQL or DB2 (SQL/PLSQL, Database design, Normalisation, Execution plan analysis, Index creation and maintenance, Stored Procedures) , PostGres/MySQL Skilled in querying data from a range of data sources that store structured and unstructured data Knowledge or understanding of Power BI (Recommended) Key Accountabilities Design, develop, test, deploy, maintain and improve software Develop flowcharts, layouts and documentation to identify requirements & solutions Write well designed & high-quality testable code Produce specifications and determine operational feasibility Integrate software components into fully functional platform Apply pro-actively & perform hands-on design and implementation of best practice CI/CD Coaching & mentoring of other Service Team members Develop/contribute to software verification plans and quality assurance procedures Document and maintain software functionality Troubleshoot, debug and upgrade existing systems, including participating in DR tests Deploy programs and evaluate customer feedback Contribute to team estimation for delivery and expectation management for scope. Comply with industry standards and regulatory requirements

Posted 1 week ago

Apply

0 years

5 - 7 Lacs

Gurugram, Haryana, India

On-site

Overview As a Site Reliability Engineer 2 (SRE 2), you play a dual role of a senior technical contributor and a team leader within the SRE team. In addition to ensuring system reliability, scalability, and performance, you will manage shift schedules, guide SRE 1 engineers, and ensure compliance with ITSM processes. Your focus will be both technical execution and operational excellence, ensuring that the team delivers high-quality, consistent support and reliability across environments. Key Responsibilities Infrastructure Reliability and Technical Leadership Ensure high availability, scalability, and performance of systems through proactive monitoring, automation, and continuous improvement. Lead efforts in improving infrastructure observability using tools like Prometheus, Alertmanager, Grafana, and other telemetry systems. Serve as an escalation point for complex technical incidents and outages, providing guidance to SRE 1 engineers. Team Oversight and Performance Management Provide technical and operational leadership to SRE 1 engineers, ensuring daily tasks are executed as per standards. Review SRE 1 work regularly to ensure adherence to best practices, SOPs, and incident response protocols. Mentor and train junior team members to enhance their technical skills and operational understanding. Conduct regular feedback sessions and contribute to performance evaluations. Shift Management and 24/7 Coverage Design, implement, and manage rotating shift schedules to ensure optimal 24/7 support coverage. Monitor shift adherence, workload distribution, and overall team health. Ensure proper handovers between shifts with complete documentation and context sharing. ITSM Process and Compliance Own and enforce ITSM processes, including Incident Management, Change Management, Problem Management, and Service Request Fulfillment. Ensure that all incidents, changes, and problems are logged, categorized, and resolved or escalated as per SLA. Continuously assess and improve ITSM processes in collaboration with internal stakeholders and audit teams. Incident and Problem Management Lead major incident investigations and coordinate response efforts across teams. Oversee root cause analysis and implementation of long-term fixes for recurring issues. Maintain detailed incident logs and postmortem reports for high-priority incidents. Change and Maintenance Oversight Review and approve change requests initiated by SRE 1 or other team members. Ensure execution of maintenance tasks adheres to predefined SOPs and does not impact system stability. Track and analyze impact of changes to continuously improve reliability metrics. Reporting and Stakeholder Communication Create and present weekly/monthly reports on SRE metrics, team performance, incident trends, and capacity planning. Collaborate with cross-functional teams, including engineering, QA, support, and product, to align operational goals. Provide updates to leadership on key incidents, system health, and team productivity. Skills:- Linux/Unix, Docker, Jenkins, grafana, Terraform, cicd, Python, Git and GitHub

Posted 1 week ago

Apply

0 years

0 Lacs

Gandhinagar, Gujarat, India

On-site

Job Title: Site Reliability Engineer Location: InfoCity, Gandhinagar, India (On-site) About AGIL f(x): AGIL f(x) is a pioneering force in the Life Sciences industry, dedicated to transforming enterprise teams with bespoke AI-powered business systems. We specialize in designing intelligent solutions that automate complex workflows and enhance decision-making by replicating human reasoning, significantly reducing manual effort, and accelerating operational speed. Our core offerings span critical areas including Quality Management Systems, Regulatory and Clinical Platforms, Safety and Compliance Tools, and Medical and Commercial Systems. We pride ourselves on being strategic AI partners, blending profound business understanding with deep technical expertise to deliver secure, scalable systems engineered for enduring success. About the Role: We are seeking a highly skilled and proactive Site Reliability Engineer (SRE) to join our growing team in Gandhinagar. This is a crucial full-time, on-site role where you will be instrumental in ensuring the continuous reliability, scalability, and performance of our cutting-edge AI-powered business systems. As an SRE at AGIL f(x), you will apply software engineering principles to operations, proactively identifying and resolving potential issues, automating infrastructure, and driving operational excellence to support our mission of empowering Life Sciences organizations. Key Responsibilities: System Reliability & Performance: Take ownership of the reliability, availability, and performance of our AI-powered business systems, ensuring they meet defined Service Level Objectives (SLOs). Monitoring & Alerting: Design, implement, and maintain robust tracking, logging, and alerting solutions to provide deep visibility into system health, performance, and user experience. This includes leveraging tools to track key metrics and set up proactive alerts. Troubleshooting & Incident Management: Lead and participate in the diagnosis, troubleshooting, and resolution of complex technical issues across our infrastructure and applications. Drive incident post-mortems (RCAs) to identify root causes and implement preventative measures and automation to reduce future occurrences. Infrastructure Management: Manage, maintain, and evolve our cloud infrastructure (specific cloud provider experience, e.g., AWS, Azure, GCP, to be added if known), ensuring scalability, security, and efficiency. This includes provisioning, configuring, and optimizing resources. Software Development for Operations: Develop, test, and deploy software solutions and automation scripts (e.g., Python, Go, Bash) to eliminate manual toil, improve operational efficiency, and enhance system resilience. This could involve building custom tools, integrating systems, or automating deployment processes. Deployment & Release Management: Collaborate with development teams to establish and optimize CI/CD pipelines, ensuring smooth, reliable, and frequent deployments of new features and bug fixes with minimal downtime. System Design & Architecture: Provide input into system design and architecture decisions, advocating for reliability, scalability, and operational maintainability from the outset. Collaboration & Communication: Work closely with development, product, and QA teams to understand system requirements, anticipate operational challenges, and foster a culture of shared ownership for system reliability. Communicate effectively with stakeholders during incidents and on long-term initiatives. Documentation: Create and maintain comprehensive documentation, runbooks, and playbooks for system configurations, operational procedures, and troubleshooting guides. Qualifications: Education: Bachelor's degree in Computer Science, Information Technology, or a related technical field. Experience: Proven experience in Site Reliability Engineering, DevOps, or a similar role focused on maintaining and improving system reliability and performance. System Administration: Strong proficiency in Linux/Unix system administration, including shell scripting. Troubleshooting Expertise: Excellent diagnostic and troubleshooting skills across distributed systems, networks, and applications. Infrastructure Management: Solid understanding of cloud computing concepts and hands-on experience with at least one primary cloud provider (e.g., AWS, Azure, GCP). Software Development: Demonstrated software development skills in at least one high-level programming language (e.g., Python, Go, Java, C#), with an emphasis on writing reliable, maintainable code for automation and operational tooling. Problem-Solving: Exceptional analytical and problem-solving abilities with meticulous attention to detail. Collaboration & Communication: Strong interpersonal skills with the ability to collaborate effectively with cross-functional teams and communicate complex technical information clearly. Proactive Mindset: A strong commitment to continuous improvement, automation, and a "fix it once" mentality. Bonus Points (Preferred Qualifications): Experience working with AI-powered systems, machine learning pipelines, or data-intensive applications. Familiarity with containerization technologies (e.g., Docker, Kubernetes). Experience with Infrastructure as Code (IaC) tools (e.g., Terraform, CloudFormation, Ansible). Knowledge of various monitoring and logging tools (e.g., Prometheus, Grafana, ELK stack, Datadog, Splunk). Experience with relational and/or NoSQL databases. Understanding of networking concepts (TCP/IP, DNS, Load Balancing, Firewalls). Why Join AGIL f(x)? Be part of a rapidly growing company at the forefront of AI innovation in the Life Sciences industry. Work on impactful projects that directly contribute to automating and optimizing critical business processes for our clients. Collaborate with a team of brilliant minds combining deep technical expertise with specialized industry knowledge. Opportunity to work with cutting-edge technologies and shape the future of AI-powered enterprise systems. A challenging yet rewarding environment that encourages continuous learning and professional growth.

Posted 1 week ago

Apply

3.0 - 6.0 years

7 - 10 Lacs

Hyderabad

Remote

Job Type: C2H (Contract to Hire) As a Data Engineer, you will work in a diverse, innovative team, responsible for designing, building, and optimizing the data infrastructure and pipelines for our new healthcare company's data platform. You'll architect and construct our core data backbone on a modern cloud stack, enabling the entire organization to turn complex data into life-saving insights. In this role, you will have the opportunity to solve challenging technical problems, mentor team members, and collaborate with innovative people to build a scalable, reliable, and world-class data ecosystem from the ground up. Core Responsibilities: (essential job duties and responsibilities) Design, develop, and maintain data replication streams and data flows to bring data from various SAP and non-SAP sources into Snowflake. Implement curated datasets on a modern data warehouse and data hub Interface directly with business and systems subject matter experts to understand analytic needs and determine logical data model requirements Work closely with data architects and senior analysts to identify common data requirements and develop shared solutions Support data integration and data modelers engineers Support and maintain data warehouse, ETL, and analytic platforms Required Skills and Experiences: Data warehouse and ETL background Advanced SQL programming capabilities Background in preparing data for analysis and reporting Familiar with data governance principles and tools Success in a highly dynamic environment with ability to shift priorities with agility Ability to go from whiteboard discussion to code Willingness to explore and implement new ideas and technologies Ability to effectively communicate with technical and non-technical audiences Ability to work independently with minimal supervision Minimum Qualifications: 4+ years experience with SQL. Snowflake strongly preferred. 3+ years experience with SAP Datasphere. 2+ years experience working directly with subject matter experts in both business and technology domains 2+ years experience with ERP data - preferably SAP S4, MS Dynamics and or BPCS 1+ year of experience with Salesforce, Workday, Concur or any other Enterprise application Nice-to-have: Experience with Machine Learning tools and processes Hands-on experience with Python Experience with Infrastructure as Code (IaC) principles and tools (e.g., Terraform, CloudFormation). Education: Bachelors in Computer Science, Information Systems, Engineering, science discipline, or similar.

Posted 1 week ago

Apply

7.0 - 12.0 years

10 - 20 Lacs

Hyderabad, Pune, Greater Noida

Work from Office

Azure DevOps Engineer Gr.Noida/Hyderabad/Pune Work Schedule : Work from Office (Hybrid) Design, implement, manage cloud infrastructure on Azure. Must have experience developing play books using Ansible. Automate using Terraform or ARM template. CI/CD pipelines - build, test, and deployment processes. Integrate Databricks notebooks & workflows into CI/CD pipelines for deployment. Version control, code quality, and testing are followed. Implement monitoring & logging solutions. infrastructure and data pipelines lifecycle.

Posted 1 week ago

Apply

8.0 - 12.0 years

30 - 45 Lacs

Coimbatore

Work from Office

We are seeking a highly skilled Lead Platform Engineer with 7+ years of experience to drive innovation at the intersection of DevOps, cloud automation, and artificial intelligence. The ideal candidate will have deep expertise in generative AI, machine learning, and AIOps, coupled with advanced knowledge of cloud infrastructure automation and modern engineering practices. This role involves leading the design, development, and implementation of transformative automation solutions using AI/ML and generative AI technologies. Responsibilities Architect automated workflows for cloud infrastructure provisioning and management using IaC tools like Terraform Build and optimize automation frameworks to enable scalable multi-cloud infrastructure deployment and management Develop and enhance service catalog components with integration into platforms such as Backstage, leveraging GenAI models for code generation Implement CI/CD pipelines to streamline code builds, testing, and deployments, ensuring continuous delivery across diverse cloud environments Write and maintain automation scripts using Python, Bash, or similar scripting languages Act as deployment orchestrator, driving smooth, automated deployments across cloud ecosystems Design and implement generative AI models such as RAG and agentic workflows using frameworks like Langchain or platforms like Bedrock, Vertex, Azure AI Build and manage vector document sources and vector databases (e.g., Amazon Kendra, Opensearch) for AI-driven applications Prepare datasets, apply feature engineering, and optimize inputs for AI/ML models to enhance training and inference outcomes Create and integrate agentic workflows using approaches like ReAct patterns or Langraph engineering with cloud GenAI platforms Evaluate model performance and select appropriate large language models (LLMs) for specific use cases while preventing model decay through prompt/flow engineering Develop MLOps pipelines to deploy RAG or agentic flows, monitoring and iterating to ensure long-term operational performance Collaborate with cross-functional teams to develop innovative cloud automation and AIOps capabilities, driving operational efficiency Requirements Bachelor's or Master's degree in Computer Science, Engineering, or a related technical field 7+ years of experience in cloud infrastructure automation, DevOps, and scripting Expertise in IaC tools such as Terraform, CloudFormation, or similar technologies Strong proficiency with Python and generative AI frameworks (RAG, agentic workflows) Proven experience working with GenAI platforms like Bedrock, Vertex AI, or Azure AI Competency in building and managing vector databases like Opensearch or Amazon Kendra Proficiency in data preparation, feature engineering, and dataset optimization for AI model development Background in designing and operating CI/CD pipelines and automating deployment workflows Knowledge of cloud automation tools, service catalogs, and integration platforms (e.g., Backstage) Nice to have Familiarity with data streaming solutions and data lake architectures for real-time AI insights Understanding of ReAct patterns and Langraph engineering for agentic workflows Skills in integrating GenAI models into existing operational platforms for enhanced automation Showcase of experience driving AIOps initiatives in large-scale environments Flexibility to adapt and utilize emerging AI/ML technologies in solving complex operational challenges

Posted 1 week ago

Apply

3.0 - 5.0 years

0 Lacs

Pune, Maharashtra, India

On-site

Description GPP Database Link (https://cummins365.sharepoint.com/sites/CS38534/) Provides application or infrastructure technical expertise, analysis and specifications for IT systems to meet business requirements in accordance with IT architecture policies and standards; translate requirements into technical specifications, create detailed solution design, coordinate construction, installation, configuration and testing of IT systems; and identifies, troubleshoots and resolves system technical issues. Key Responsibilities Assists in creating and managing the technical specifications from which programs and configurations will be applied to create the application, infrastructure or complete solution. Assists in technical analysis of potential solutions including evaluating the technical fit and viability of commercial-of-the-shelf products versus custom built solution. Delivers solution design, adhering to standards and leveraging re-use of components. Assists in creating high level and detail design of infrastructure OR applications, interfaces, conversions, extensions, reports, and workflows while meeting architecture, security, performance, scalability, and maintainability requirements. Assists technical and infrastructure groups in understanding the solution design and specifications. Works closely with IT technical service providers to ensure outsourced work packages are delivered to specifications, meeting key parameters of quality, schedule, cost, security, performance, and scalability. Participates in formal design reviews and code reviews (where applicable) to ensure the detailed design specifications have been understood and coding standards have been adhered to by the development team(s). Leverages re-usability of component designs to reduce costs and shorten time to deliver solution Ensures adherence to obtain efficiency in the build and deploy processes, leveraging automation where possible. Assists in the test strategy and execution of master test plan. Participates in the creation of standards, processes, procedures, and guidelines for the IT design and development community as well work instructions or ‘runbooks’ used for end-user support. Analyzes and revises existing systems and documentation to identify remediation or improvements in the application or infrastructure solution. Provides level 3 support for critical issues. Responsibilities Competencies: Customer focus - Building strong customer relationships and delivering customer-centric solutions. Global perspective - Taking a broad view when approaching issues, using a global lens. Manages complexity - Making sense of complex, high quantity, and sometimes contradictory information to effectively solve problems. Manages conflict - Handling conflict situations effectively, with a minimum of noise. Optimizes work processes - Knowing the most effective and efficient processes to get things done, with a focus on continuous improvement. Data Modeling - Creates, writes and tests data models, test scripts and build scripts using industry standards and tools, version control, and build and test automation to meet business, technical, security, governance and compliance requirements. Performance Tuning - Conceptualizes, analyzes and solves application, database and hardware problems using industry standards and tools, version control, and build and test automation to meet business, technical, security, governance and compliance requirements. Solution Configuration - Configures, creates and tests a solution for commercial off-the-shelf (COTS) applications using industry standards and tools, version control, and build and test automation to meet business, technical, security, governance and compliance requirements. Solution Design - Creates and defines the solution design complete with instrumentation and measurement, using industry standards and tools, version control, and build and test automation to synthesize diagrams, models and documentation in order to build a solution that meets buildability, business, technical, security, governance and compliance requirements. Solution Functional Fit Analysis - Composes and decomposes a system into its component parts using procedures, tools and work aides for the purpose of studying how well the component parts were designed, purchased and configured to interact holistically to meet business, technical, security, governance and compliance requirements. Solution Modeling - Creates, designs and formulates models, diagrams and documentation using industry standards, tools, version control, and build and test automation to meet business, technical, security, governance and compliance requirements. Solution Validation Testing - Validates a configuration item change or solution using the Function's defined best practices, including the Systems Development Life Cycle (SDLC) standards, tools and metrics, to ensure that it works as designed and meets customer requirements. Values differences - Recognizing the value that different perspectives and cultures bring to an organization. Education, Licenses, Certifications College, university, or equivalent degree in Computer Science, Information Technology, Business, or related subject, or relevant equivalent experience required. This position may require licensing for compliance with export controls or sanctions regulations. Experience Intermediate level of relevant work experience required. 3-5 years of experience. Qualifications 2-3 core responsibilities or activities unique to the role Work for EPIC team in collaboration with the enterprise product teams to assist and deliver infrastructure provisioning/automation in cloud(AWS, Azure) Provide level 1 & 2 support for all operational issues along with Self Service Tools, Validate/test infrastructure automation tools and perform DR drills on Application as per need Set up, Troubleshoot, resolve issues and maintain monitoring-related configuration in Dynatrace and Splunk 2-4 Skill, Education, or Experience Requirements needed for the role Strong hands on in AWS, Azure Cloud technologies Proficiency in Infrastructure automation and infrastructure as code concepts: Terraform, Ansible Hands on experience in atleast one of the observability tools: Splunk, Dynatrace, CloudWatch Experienced in Agile/SAFe methodology 1-2 Skill, Education, or Experience Preferences that are nice to have but not required) AWS Certified SysOps Administrator - Associate OR Microsoft Certified: Azure Administrator Associate and HashiCorp Certified: Terraform Authoring and Operations Professional DevOps and CI/CD methodologies and concepts, Exposure in IT infrastructure & architecture, system administration, networking, applications, database & security

Posted 1 week ago

Apply

7.0 years

0 Lacs

Gurugram, Haryana, India

On-site

We are looking for MLops Engineer@ Gurgaon Location Job Responsibilities: Design and implement CI/CD pipelines for machine learning workflows. Develop and maintain production-grade ML pipelines using tools like MLflow, Kubeflow, or Airflow. Automate model training, testing, deployment, and monitoring processes. Collaborate with Data Scientists to operationalize ML models, ensuring scalability and performance. Monitor deployed models for drift, degradation, and bias, and trigger retraining as needed. Maintain and improve infrastructure for model versioning, artifact tracking, and reproducibility. Integrate ML solutions with microservices/APIs using FastAPI or Flask. Work on containerized environments using Docker and Kubernetes. Implement logging, monitoring, and alerting for ML systems (e.g., Prometheus, Grafana). Champion best practices in code quality, testing, and documentation. Required Skills: 7+ years of experience in Python development and ML/AI-related engineering roles. Strong experience in ML Ops tools like MLflow, Kubeflow, Airflow, or similar. Deep understanding of Docker, Kubernetes, and container orchestration for ML workflows. Hands-on experience with cloud platforms (AWS, GCP, Azure) and infrastructure-as-code (Terraform/CDK). Familiarity with model deployment and serving frameworks (e.g., Seldon, TorchServe, TensorFlow Serving). Good understanding of DevOps practices and CI/CD pipelines (Jenkins, GitHub Actions, GitLab CI). Experience with data versioning tools (e.g., DVC) and model lifecycle management. Exposure to monitoring tools for ML and infrastructure healt Experience:7-12 Yrs Job Location :Gurgaon Interested candidates can share your CV to mangani.paramanandhan@bounteous.com, i will call you shortly. Please share your CV.

Posted 1 week ago

Apply

7.0 years

30 - 45 Lacs

Noida, Uttar Pradesh, India

On-site

We are looking for a customer-obsessed, analytical Sr. Staff Engineer to lead the development and growth of our Tax Compliance product suite . In this role, you’ll shape innovative digital solutions that simplify and automate tax filing, reconciliation, and compliance workflows for businesses of all sizes. You will join a fast-growing company where you’ll work in a dynamic and competitive market, impacting how businesses meet their statutory obligations with speed, accuracy, and confidence. As the Sr. Staff Engineer, you’ll work closely with product, DevOps, and data teams to architect reliable systems, drive engineering excellence , and ensure high availability across our platform. We’re looking for a technical leader who’s not just an expert in building scalable systems, but also passionate about mentoring engineers and shaping the future of fintech. Responsibilities Lead, mentor, and inspire a high-performing engineering team (or operate as a hands-on technical lead). Drive the design and development of scalable backend services using Python. Experience in Django, FastAPI, Task Orchestration Systems. Own and evolve our CI/CD pipelines with Jenkins, ensuring fast, safe, and reliable deployments. Architect and manage infrastructure using AWS and Terraform with a DevOps-first mindset. Collaborate cross-functionally with product managers, designers, and compliance experts to deliver features that make tax compliance seamless for our users. Set and enforce engineering best practices, code quality standards, and operational excellence. Stay up-to-date with industry trends and advocate for continuous improvement in engineering processes. Experience in fintech, tax, or compliance industries. Familiarity with containerization tools like Docker and orchestration with Kubernetes. Background in security, observability, or compliance automation. Requirements 7+ years of software engineering experience, with at least 2+ years in a leadership or principal-level role. Deep expertise in Python, including API development, performance optimization, and testing. Experience in Event-driven architecture, Kafka/RabbitMQ-like systems. Strong experience with AWS services (e.g., ECS, Lambda, S3, RDS, CloudWatch). Solid understanding of Terraform for infrastructure as code. Proficiency with Jenkins or similar CI/CD tooling. Comfortable balancing technical leadership with hands-on coding and problem-solving. Strong communication skills and a collaborative mindset. Skills:- Python, Django, FastAPI, PostgreSQL, MongoDB, Redis, Apache Kafka, RabbitMQ, AWS Simple Notification Service (SNS), AWS Simple Queuing Service (SQS), Amazon Web Services (AWS), Systems design, Apache Airflow and Celery

Posted 1 week ago

Apply

7.0 years

15 - 22 Lacs

Noida, Uttar Pradesh, India

On-site

About Us CLOUDSUFI, a Google Cloud Premier Partner, is a global leading provider of data-driven digital transformation across cloud-based enterprises. With a global presence and focus on Software & Platforms, Life sciences and Healthcare, Retail, CPG, financial services, and supply chain, CLOUDSUFI is positioned to meet customers where they are in their data monetization journey. Our Values We are a passionate and empathetic team that prioritizes human values. Our purpose is to elevate the quality of lives for our family, customers, partners and the community. Equal Opportunity Statement CLOUDSUFI is an equal opportunity employer. We celebrate diversity and are committed to creating an inclusive environment for all employees. All qualified candidates receive consideration for employment without regard to race, color, religion, gender, gender identity or expression, sexual orientation, and national origin status. We provide equal opportunities in employment, advancement, and all other areas of our workplace. Please explore more at https://www.cloudsufi.com/ What Are We Looking For We are seeking a highly skilled and experienced Senior DevOps Engineer to join our team. The ideal candidate will have extensive expertise in modern DevOps tools and practices, particularly in managing CI/CD pipelines, infrastructure as code, and cloud-native environments. This role involves designing, implementing, and maintaining robust, scalable, and efficient infrastructure and deployment pipelines to support our development and operations teams. Required Skills And Experience 7+ years of experience in DevOps, infrastructure automation, or related fields. Advanced expertise in Terraform for infrastructure as code. Solid experience with Helm for managing Kubernetes applications. Proficient with GitHub for version control, repository management, and workflows. Extensive experience with Kubernetes for container orchestration and management. In-depth understanding of Google Cloud Platform (GCP) services and architecture. Strong scripting and automation skills (e.g., Python, Bash, or equivalent). Excellent problem-solving skills and attention to detail. - Strong communication and collaboration abilities in agile development environments. Preferred Qualifications Experience with other CI/CD tools (e.g., Jenkins, GitLab CI/CD). Knowledge of additional cloud platforms (e.g., AWS, Azure). Certification in Kubernetes (CKA/CKAD) or Google Cloud (GCP Professional DevOps Engineer). Behavioral Competencies Must have worked with US/Europe based clients in onsite/offshore delivery models. Should have very good verbal and written communication, technical articulation, listening and presentation skills. Should have proven analytical and problem solving skills. Should have collaborative mindset for cross-functional team work Passion for solving complex search problems Should have demonstrated effective task prioritization, time management and internal/external stakeholder management skills. Should be a quick learner, self starter, go-getter and team player. Should have experience of working under stringent deadlines in a Matrix organization structure. Skills:- Google Cloud Platform (GCP), Kubernetes, Docker, helm and Terraform

Posted 1 week ago

Apply

5.0 - 8.0 years

15 - 30 Lacs

Hyderabad, Pune, Bengaluru

Work from Office

We are seeking a skilled Platform Engineer to join our Automation Engineering team, bringing expertise in cloud infrastructure automation, DevOps, scripting, and advanced AI/ML practices. The role focuses on integrating generative AI into automation workflows, enhancing operational efficiency, and supporting cloud-first initiatives. Responsibilities Design cloud automation workflows using Infrastructure-as-Code tools such as Terraform or CloudFormation Build scalable frameworks to manage infrastructure provisioning, deployment, and configuration across multiple cloud platforms Create service catalog components compatible with automation platforms like Backstage Integrate generative AI models to improve service catalog functionalities, including automated code generation and validation Architect CI/CD pipelines for automated build, test, and deployment processes Maintain deployment automation scripts utilizing technologies such as Python or Bash Implement generative AI models (e.g., RAG, agent-based workflows) for AIOps use cases like anomaly detection and root cause analysis Employ AI/ML tools such as LangChain, Bedrock, Vertex AI, or Azure AI for advanced generative AI solutions Develop vector databases and document sources using services like Amazon Kendra, OpenSearch, or custom solutions Engineer data pipelines to stream real-time operational insights that support AI-driven automation Build MLOps pipelines to deploy and monitor generative AI models, ensuring optimal performance and avoiding model decay Select appropriate LLM models for specific AIOps use cases and integrate them effectively into workflows Collaborate with cross-functional teams to design and refine automation and AI-driven processes Research emerging tools and technologies to enhance operational efficiency and scalability Requirements Bachelor's or Master's degree in Computer Science, Engineering, or related field 3-8 years of experience in cloud infrastructure automation, DevOps, and scripting Proficiency with Infrastructure-as-Code tools such as Terraform or CloudFormation Expertise in Python and generative AI frameworks like RAG and agent-based workflows Knowledge of cloud-based AI services, including Bedrock, Vertex AI, or Azure AI Familiarity with vector databases like Amazon Kendra, OpenSearch, or custom database solutions Competency in data engineering tasks such as feature engineering, labeling, and real-time data streaming Proven track record in creating and maintaining MLOps pipelines for AI/ML models in production environments Nice to have Background in Flow Engineering tools such as Langraph or platform-specific workflow orchestration tools Understanding of comprehensive AIOps processes to refine cloud-based automation solutions

Posted 1 week ago

Apply

5.0 - 8.0 years

15 - 30 Lacs

Chennai

Work from Office

We are seeking a skilled Platform Engineer to join our Automation Engineering team, bringing expertise in cloud infrastructure automation, DevOps, scripting, and advanced AI/ML practices. The role focuses on integrating generative AI into automation workflows, enhancing operational efficiency, and supporting cloud-first initiatives. Responsibilities Design, build, and maintain cloud automation workflows using Infrastructure-as-Code tools such as Terraform or CloudFormation Develop scalable frameworks for managing infrastructure provisioning, deployment, and configuration across multiple cloud platforms Create and integrate service catalog components with automation platforms like Backstage Leverage generative AI models to enhance service catalog capabilities, including automated code generation and validation Architect and implement CI/CD pipelines for automated build, test, and deployment processes Build and maintain deployment automation scripts using technologies such as Python or Bash Design and implement generative AI models (e.g., RAG, agent-based workflows) for AIOps use cases like anomaly detection and root cause analysis Utilize AI/ML tools such as LangChain, Bedrock, Vertex AI, or Azure AI for building advanced generative AI solutions Develop vector databases and document sources using services like Amazon Kendra, OpenSearch, or custom solutions Engineer data pipelines for streaming real-time operational insights to support AI-driven automation Create MLOps pipelines to deploy and monitor generative AI models, ensuring optimal performance and avoiding model decay Evaluate and select appropriate LLM models for specific AIOps use cases, integrating them efficiently into workflows Collaborate with cross-functional teams to design and improve automation and AI-driven processes Continuously research emerging tools and technologies to improve operational efficiency and scalability Requirements Bachelor's or Master's degree in Computer Science, Engineering, or related field 3-8 years of experience in cloud infrastructure automation, DevOps, and scripting Proficiency with Infrastructure-as-Code tools such as Terraform or CloudFormation Expertise in Python and generative AI frameworks like RAG and agent-based workflows Knowledge of cloud-based AI services, including Bedrock, Vertex AI, or Azure AI Familiarity with vector databases like Amazon Kendra, OpenSearch, or custom database solutions Competency in data engineering tasks such as feature engineering, labeling, and real-time data streaming Proven experience in creating and maintaining MLOps pipelines for AI/ML models in production environments Nice to have Familiarity with Flow Engineering tools such as Langraph or platform-specific workflow orchestration tools Understanding of end-to-end AIOps processes to enhance cloud-based automation solutions

Posted 1 week ago

Apply

5.0 - 7.0 years

20 - 30 Lacs

Udaipur, Gurugram

Work from Office

Job Description At GKM IT , were looking for a DevOps Engineer - Senior II who thrives at the intersection of strategy and execution. If you have a passion for building scalable, secure, and high-performing infrastructure, this role is your opportunity to make a direct impact. Youll play a key role in designing and optimising systems that support complex, high-load environmentswhile collaborating with top-tier engineering teams to bring stability, speed, and innovation to everything we deploy. If you're someone who enjoys solving real-world infrastructure challenges and thrives in fast-paced, production-critical setupswe’d love to work with you! Requirements Minimum 5+ years of experience in DevOps roles managing production-grade systems Implement CI/CD pipelines using Jenkins, GitHub Actions, CircleCI, or Azure DevOps Strong expertise in AWS, Terraform, Kubernetes, CI/CD, Linux, and network security Manage, monitor, and optimize distributed databases (PostgreSQL, MySQL, MongoDB, cloud-native databases) Define and manage Infrastructure as Code using Terraform, Ansible, or CloudFormation Demonstrate deep expertise in Linux internals, kernel tuning, scripting (Bash/Python), and networking Design and implement resilient, secure, and scalable infrastructure to support high-traffic applications Architect solutions for high availability, cost-efficiency, and performance optimization at enterprise scale Integrate and operate across multi-cloud or hybrid environments (AWS, Azure, on-prem) Design and maintain ETL/serverless data pipelines using Apache Airflow, AWS Glue, Lambda, and Talend Optimize data pipeline reliability, scheduling, error handling, and CI/CD integration Implement infrastructure-level security controls (CIS hardening, IAM, encryption, firewall rules) Practical experience with compliance frameworks like SOC 2, HIPAA, and internal audits Build and maintain observability systems (Prometheus, Grafana, ELK/Loki, Datadog, CloudWatch) Manage networking stacks (iptables, routing, DNS, SSL, load balancing) and Linux server security Automate provisioning, patching, config management, and multi-stage deployments Manage access control and identity integration with Microsoft Active Directory and Entra ID Provide mentorship and technical leadership to junior DevOps engineers and interns through code reviews, technical sessions, and team-wide knowledge-sharing initiatives Preferred certifications: AWS Solutions Architect, RHCE, or equivalent Experience with container orchestration tools like Kubernetes, ECS, or Docker Swarm Benefits We don’t just hire employees—we invest in people. At GKM IT, we’ve designed a benefits experience that’s thoughtful, supportive, and actually useful. Here’s what you can look forward to: Top-Tier Work Setup You’ll be equipped with a premium MacBook and all the accessories you need. Great tools make great work. Flexible Schedules & Remote Support Life isn’t 9-to-5. Enjoy flexible working hours, emergency work-from-home days, and utility support that makes remote life easier. Quarterly Performance Bonuses We don’t believe in waiting a whole year to celebrate your success. Perform well, and you’ll see it in your pay check—quarterly. Learning is Funded Here Conferences, courses, certifications—if it helps you grow, we’ve got your back. We even offer a dedicated educational allowance. Family-First Culture Your loved ones matter to us too. From birthday and anniversary vouchers (Amazon, BookMyShow) to maternity and paternity leaves—we’re here for life outside work. Celebrations & Gifting, The GKM IT Way Onboarding hampers, festive goodies (Diwali, Holi, New Year), and company anniversary surprises—it’s always celebration season here. Team Bonding Moments We love food, and we love people. Quarterly lunches, dinners, and fun company retreats help us stay connected beyond the screen. Healthcare That Has You Covered Enjoy comprehensive health insurance for you and your family—because peace of mind shouldn’t be optional. Extra Rewards for Extra Effort Weekend work doesn’t go unnoticed, and great referrals don’t go unrewarded. From incentives to bonuses—you’ll feel appreciated.

Posted 1 week ago

Apply

5.0 - 8.0 years

15 - 30 Lacs

Coimbatore

Work from Office

We are seeking a skilled Platform Engineer to join our Automation Engineering team, bringing expertise in cloud infrastructure automation, DevOps, scripting, and advanced AI/ML practices. The role focuses on integrating generative AI into automation workflows, enhancing operational efficiency, and supporting cloud-first initiatives. Responsibilities Design, build, and maintain cloud automation workflows using Infrastructure-as-Code tools such as Terraform or CloudFormation Develop scalable frameworks for managing infrastructure provisioning, deployment, and configuration across multiple cloud platforms Create and integrate service catalog components with automation platforms like Backstage Leverage generative AI models to enhance service catalog capabilities, including automated code generation and validation Architect and implement CI/CD pipelines for automated build, test, and deployment processes Build and maintain deployment automation scripts using technologies such as Python or Bash Design and implement generative AI models (e.g., RAG, agent-based workflows) for AIOps use cases like anomaly detection and root cause analysis Utilize AI/ML tools such as LangChain, Bedrock, Vertex AI, or Azure AI for building advanced generative AI solutions Develop vector databases and document sources using services like Amazon Kendra, OpenSearch, or custom solutions Engineer data pipelines for streaming real-time operational insights to support AI-driven automation Create MLOps pipelines to deploy and monitor generative AI models, ensuring optimal performance and avoiding model decay Evaluate and select appropriate LLM models for specific AIOps use cases, integrating them efficiently into workflows Collaborate with cross-functional teams to design and improve automation and AI-driven processes Continuously research emerging tools and technologies to improve operational efficiency and scalability Requirements Bachelor's or Master's degree in Computer Science, Engineering, or related field 3-8 years of experience in cloud infrastructure automation, DevOps, and scripting Proficiency with Infrastructure-as-Code tools such as Terraform or CloudFormation Expertise in Python and generative AI frameworks like RAG and agent-based workflows Knowledge of cloud-based AI services, including Bedrock, Vertex AI, or Azure AI Familiarity with vector databases like Amazon Kendra, OpenSearch, or custom database solutions Competency in data engineering tasks such as feature engineering, labeling, and real-time data streaming Proven experience in creating and maintaining MLOps pipelines for AI/ML models in production environments Nice to have Familiarity with Flow Engineering tools such as Langraph or platform-specific workflow orchestration tools Understanding of end-to-end AIOps processes to enhance cloud-based automation solutions

Posted 1 week ago

Apply
cta

Start Your Job Search Today

Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.

Job Application AI Bot

Job Application AI Bot

Apply to 20+ Portals in one click

Download Now

Download the Mobile App

Instantly access job listings, apply easily, and track applications.

Featured Companies