Get alerts for new jobs matching your selected skills, preferred locations, and experience range. Manage Job Alerts
10.0 years
0 Lacs
India
Remote
Job Description: Rubrik Backup Engineer IV Shift: 24/7 (Monthly Rotation) Remote Work From Home Job Summary The Rubrik Backup Engineer IV is a senior technical specialist responsible for the design, implementation, troubleshooting, and optimisation of enterprise backup and disaster recovery solutions using Rubrik. This individual acts as a technical escalation point and subject matter expert (SME), driving innovation and automation while collaborating with cross-functional teams to ensure data protection strategies meet business requirements. The role demands deep expertise in Rubrik architecture and integration, alongside strong capabilities in automation, scripting, VMware, and modern data protection frameworks. Career Level Summary Recognized expert with specialized depth in enterprise backup and recovery Leads large-scale initiatives and provides technical direction across teams Works independently on the most complex issues and initiatives Coaches and mentors junior engineers and cross-functional peers Key Responsibilities Serve as the highest level of technical escalation for Rubrik-related incidents and issues Architect and implement Rubrik backup solutions across hybrid, on-premises, and multi-cloud environments (AWS, Azure, GCP) Lead backup and recovery strategy design sessions for customers, including air-gapped, immutable, and ransomware-resilient architectures Integrate Rubrik with external systems (e.g., ServiceNow, Splunk, vSphere, Azure AD) using REST APIs and automation tools (Python, Ansible, Terraform) Design and maintain Rubrik SLA Domains, archival policies (cloud/tape), replication, and compliance workflows Collaborate with Engineering, Storage, Security, and Application teams to ensure backup consistency and performance Manage large-scale Rubrik clusters, capacity planning, and software upgrades Proactively identify and resolve systemic issues across infrastructure that impact backup performance or restore SLAs Document architectures, runbooks, and SOPs; contribute to technical training and playbooks Work closely with stakeholders and leadership on backup audit and regulatory compliance requirements Provide technical mentoring and guidance to junior engineers and partner teams Required Skills And Knowledge Expert knowledge of Rubrik CDM architecture, RBS, Polaris, and Rubrik APIs Advanced skills in backup for virtualized environments (VMware, Hyper-V) Strong understanding of file-level, database-level, and VM-level backup and restore operations Deep knowledge of cloud-native backups and cloud archiving using AWS S3, Azure Blob, and GCP storage Hands-on experience with integration and automation (e.g., Python, PowerShell, REST API, Terraform, Ansible) Proficiency in disaster recovery design, planning, and orchestration (DR runbooks) Familiarity with data compliance, encryption, and ransomware defense mechanisms Exposure to monitoring and reporting platforms (e.g., vROps, Splunk, Grafana) Good knowledge of enterprise infrastructure: storage systems (SAN/NAS), networking, Windows/Linux OS Excellent documentation and communication skills for both technical and executive-level audiences Experience / Education Minimum of 10 years of experience in IT infrastructure with at least 4 years of hands-on experience in Rubrik Proven track record of managing complex enterprise-scale backup environments Experience with backup and recovery for databases (MSSQL, Oracle), file servers, and virtual machines Bachelor's degree in Computer Science, Information Technology, or equivalent work experience Preferred Certifications: Rubrik Certified System Administrator (RCSA) or similar Rubrik platform certifications VMware VCP, AWS/Azure certifications are a plus About Rackspace Technology We are the multicloud solutions experts. We combine our expertise with the world’s leading technologies — across applications, data and security — to deliver end-to-end solutions. We have a proven record of advising customers based on their business challenges, designing solutions that scale, building and managing those solutions, and optimizing returns into the future. Named a best place to work, year after year according to Fortune, Forbes and Glassdoor, we attract and develop world-class talent. Join us on our mission to embrace technology, empower customers and deliver the future. More on Rackspace Technology Though we’re all different, Rackers thrive through our connection to a central goal: to be a valued member of a winning team on an inspiring mission. We bring our whole selves to work every day. And we embrace the notion that unique perspectives fuel innovation and enable us to best serve our customers and communities around the globe. We welcome you to apply today and want you to know that we are committed to offering equal employment opportunity without regard to age, color, disability, gender reassignment or identity or expression, genetic information, marital or civil partner status, pregnancy or maternity status, military or veteran status, nationality, ethnic or national origin, race, religion or belief, sexual orientation, or any legally protected characteristic. If you have a disability or special need that requires accommodation, please let us know.
Posted 3 days ago
0 years
0 Lacs
India
Remote
CryptoChakra is a leading cryptocurrency analytics and education platform committed to simplifying digital asset markets for traders, investors, and institutions. By integrating advanced predictive analytics, machine learning frameworks, and immersive learning ecosystems, we empower users to navigate market volatility with precision. Our platform leverages real-time blockchain intelligence, AI-driven forecasts, and scalable cloud infrastructure to deliver actionable insights and educational tools. As a remote-first innovator, we unite engineers, data scientists, and educators to democratize access to decentralized finance through cutting-edge technology and user-centric solutions. Position: Fresher Cloud Engineer Intern Remote | Full-Time Internship | Compensation: Paid/Unpaid based on suitability Role Summary Join CryptoChakra’s infrastructure team to build and optimize the cloud systems powering our analytics platform. This role offers hands-on experience in deploying scalable solutions for blockchain data processing, predictive modeling, and educational resource delivery, with mentorship from industry experts. Key Responsibilities Cloud Infrastructure: Assist in deploying and managing AWS/GCP services (EC2, S3, Lambda) for real-time crypto data pipelines and ML model hosting. DevOps Support: Implement CI/CD workflows (GitHub Actions, Jenkins) and containerization (Docker) for seamless software deployment. Database Management: Optimize SQL/NoSQL databases (PostgreSQL, MongoDB) for high-frequency transactional and on-chain data storage. Linux Administration: Monitor server performance, automate tasks with Bash/Python scripts, and troubleshoot system issues. Collaboration: Work with data engineers to ensure low-latency data access for analytics tools and predictive models. Qualifications Technical Skills Foundational knowledge of cloud platforms (AWS, GCP, or Azure) and DevOps principles. Familiarity with Linux environments, shell scripting, and basic system administration. Understanding of database management systems (SQL/NoSQL) and query optimization. Interest in blockchain technology, DeFi protocols, or crypto APIs (CoinGecko) is a plus. Professional Competencies Strong problem-solving skills to debug infrastructure bottlenecks. Self-motivated with adaptability to remote collaboration tools (Slack, Zoom). Curiosity to learn cloud engineering tools like Terraform, Ansible, or Kubernetes. Preferred (Not Required) Academic projects involving cloud deployments, distributed systems, or automation. Pursuing or holding a degree in Computer Science, IT, or related fields. What We Offer Skill Development: Master AWS/GCP, DevOps practices, and blockchain data infrastructure. Real-World Impact: Contribute to systems handling terabytes of crypto data daily. Flexibility: Remote work with mentorship tailored to your learning pace.
Posted 3 days ago
0.0 - 9.0 years
0 Lacs
Hyderabad, Telangana
On-site
General information Country India State Telangana City Hyderabad Job ID 45594 Department Development Experience Level MID_SENIOR_LEVEL Employment Status FULL_TIME Workplace Type On-site Description & Requirements As a Senior DevOps Engineer, you will be responsible for leading the design, development, and operationalization of cloud infrastructure and CI/CD processes. You will serve as a subject matter expert (SME) for Kubernetes, AWS infrastructure, Terraform automation, and DevSecOps practices. This role also includes mentoring DevOps engineers, contributing to architecture decisions, and partnering with cross-functional engineering teams to implement best-in-class cloud and deployment solutions. Essential Duties: Design, architect, and automate cloud infrastructure using Infrastructure as Code (IaC) tools such as Terraform and CloudFormation. Lead and optimize Kubernetes-based deployments, including Helm chart management, autoscaling, and custom controller integrations. Implement and manage CI/CD pipelines for microservices and serverless applications using Jenkins, GitLab, or similar tools. Champion DevSecOps principles, integrating security scanning (SAST/DAST) and policy enforcement into the pipeline. Collaborate with architects and application teams to build resilient and scalable infrastructure solutions across AWS services (EC2, VPC, Lambda, EKS, S3, IAM, etc.). Establish and maintain monitoring, alerting, and logging practices using tools like Prometheus, Grafana, CloudWatch, ELK, or Datadog. Drive cost optimization, environment standardization, and governance across cloud environments. Mentor junior DevOps engineers and participate in technical reviews, playbook creation, and incident postmortems. Develop self-service infrastructure provisioning tools and contribute to internal DevOps tooling. Actively participate in architecture design reviews, cloud governance, and capacity planning efforts. Basic Qualifications: 7–9 years of hands-on experience in DevOps, Cloud Infrastructure, or SRE roles. Strong expertise in AWS cloud architecture and automation using Terraform or similar IaC tools. Solid knowledge of Kubernetes, including experience managing EKS clusters, Helm, and custom resources. Deep experience in Linux administration, networking, and security hardening. Advanced experience building and maintaining CI/CD pipelines (Jenkins, GitLab CI, etc.). Proficient in scripting with Bash, Groovy, or Python. Strong understanding of containerization using Docker and orchestration strategies. Experience with monitoring and logging stacks like ELK, Prometheus, and CloudWatch. Familiarity with security, identity management, and cloud compliance frameworks. Excellent troubleshooting skills and a proactive approach to system reliability and resilience. Strong interpersonal skills and ability to work cross-functionally. Bachelor’s degree in Computer Science, Information Systems, or equivalent. Preferred Qualifications: Experience with GitOps using ArgoCD or FluxCD. Knowledge of multi-account AWS architecture, VPC peering, and Service Mesh. Exposure to DataOps, platform engineering, or large-scale data pipelines. Familiarity with Serverless Framework, API Gateway, and event-driven designs. Certifications such as AWS DevOps Engineer – Professional, CKA/CKAD, or equivalent. Experience in regulated environments (e.g., SOC2, ISO27001, GDPR, HIPAA). About Infor Infor is a global leader in business cloud software products for companies in industry specific markets. Infor builds complete industry suites in the cloud and efficiently deploys technology that puts the user experience first, leverages data science, and integrates easily into existing systems. Over 60,000 organizations worldwide rely on Infor to help overcome market disruptions and achieve business-wide digital transformation. For more information visit www.infor.com Our Values At Infor, we strive for an environment that is founded on a business philosophy called Principle Based Management™ (PBM™) and eight Guiding Principles: integrity, stewardship & compliance, transformation, principled entrepreneurship, knowledge, humility, respect, self-actualization. Increasing diversity is important to reflect our markets, customers, partners, and communities we serve in now and in the future. We have a relentless commitment to a culture based on PBM. Informed by the principles that allow a free and open society to flourish, PBM™ prepares individuals to innovate, improve, and transform while fostering a healthy, growing organization that creates long-term value for its clients and supporters and fulfillment for its employees. Infor is an Equal Opportunity Employer. We are committed to creating a diverse and inclusive work environment. Infor does not discriminate against candidates or employees because of their sex, race, gender identity, disability, age, sexual orientation, religion, national origin, veteran status, or any other protected status under the law. If you require accommodation or assistance at any time during the application or selection processes, please submit a request by following the directions located in the FAQ section at the bottom of the infor.com/about/careers webpage.
Posted 3 days ago
5.0 - 9.0 years
0 Lacs
karnataka
On-site
As a Senior Site Reliability Engineer (SRE) on the Network Assurance Data Platform team at Cisco ThousandEyes, you will be responsible for ensuring the reliability, scalability, and security of our cloud and big data platforms. Collaborating with cross-functional teams, including software development, product management, and security, you will design, build, and maintain systems operating at multi-region scale. Your efforts will directly impact the success of our machine learning (ML) and AI initiatives by guaranteeing that the underlying infrastructure is robust, efficient, and aligned with operational excellence. Your main responsibilities will include designing, building, and optimizing cloud and data infrastructure to ensure high availability, reliability, and scalability of big-data and ML/AI systems. You will implement Site Reliability Engineering principles such as monitoring, alerting, error budgets, and fault analysis. Working closely with development, product management, and security teams, you will develop secure, scalable solutions that support ML/AI workloads and enhance operational efficiency through automation. Troubleshooting complex technical issues in production environments, performing root cause analyses, and contributing to continuous improvement efforts will also be part of your role. You will help shape the team's technical strategy and roadmap, balancing immediate needs with long-term goals, while mentoring peers and fostering a culture of learning and technical excellence. Qualifications for this role include the ability to design and implement scalable and well-tested solutions with a focus on streamlining operations. Strong hands-on experience in cloud services, preferably AWS, and Infrastructure as Code skills, ideally with Terraform and Kubernetes, are required. Previous experience in AWS cost management, understanding of Prometheus and its ecosystem, and the ability to write high-quality code in Python, Go, or equivalent languages are essential. A good understanding of Unix/Linux systems, the kernel, system libraries, file systems, and client-server protocols is expected. Experience in building cloud, big data, and/or ML/AI infrastructure (e.g., EMR, Airflow, Comet ML, AWS SageMaker, Spark, etc) would be a bonus. Cisco values diversity in its employees and believes that diverse teams are better equipped to solve problems, innovate, and create a positive impact. The company encourages candidates from all backgrounds to apply, even if they do not meet every single qualification listed. Research shows that individuals from underrepresented groups may experience imposter syndrome and doubt their candidacy strength. Cisco aims to unlock the potential in all candidates and emphasizes that everyone has something valuable to offer.,
Posted 3 days ago
6.0 - 10.0 years
0 Lacs
hyderabad, telangana
On-site
As a Software Engineer III at NCR Atleos, you will be responsible for leveraging your 6-9 years of experience to contribute to our global technology initiatives. Your expertise in Java and Spring Boot Framework, specifically Java 11 or higher, will be pivotal in enhancing self-service banking experiences. You will play a key role in implementing microservices architecture, developing RESTful APIs, and utilizing API management platforms to optimize self-service availability for financial institutions and retailers worldwide. Your experience with Azure Cloud, SQL, and NoSQL databases will be essential in designing and developing high volume web-services using API protocols and data formats. A strong understanding of data structures and algorithms will enable you to create efficient and scalable solutions. Additionally, your familiarity with Linux based infrastructure, CI/CD pipelines, automated testing tools, and deployment platforms such as Terraform, Ansible, Jenkins, and AzureDevOps/GitHub Actions will be valuable assets in your role. In this position, you will have the opportunity to contribute to UI/UX frameworks like Angular and Node, drive code reviews, design reviews, and architecture discussions, and optimize Java applications for performance. Your ability to lead small teams for substantial projects, apply strong analytical and problem-solving skills to address complex technical challenges, and communicate effectively with stakeholders, project managers, and fellow developers will be key to your success at NCR Atleos. Join us at NCR Atleos to be a part of a dynamic team dedicated to creating exceptional self-service banking experiences and expanding financial access globally. Discover the benefits offered in your region, competitive base salary, and work/family programs on our careers site.,
Posted 3 days ago
3.0 - 5.0 years
0 Lacs
Mumbai, Maharashtra, India
On-site
About The Advanced Analytics Team The central Advanced Analytics team at the Abbott Established Pharma Division’s (EPD) headquarters in Basel helps define and lead the transformation towards becoming a global, data-driven company with the help of data and advanced technologies (e.g., Machine Learning, Deep Learning, Generative AI, Computer Vision). To us, Advanced Analytics is an important lever to reach our business targets, now and in the future; It helps differentiate ourselves from our competition and ensure sustainable revenue growth at optimal margins. Hence the central AA team is an integral part of the Strategy Management Office at EPD that has a very close link and regular interactions with the EPD Senior Leadership Team. Primary Job Function With the above requirements in mind, EPD is looking to fill a role of a Cloud Engineer reporting to the Head of AA Product Development. The Cloud Engineer will be responsible for developing applications leveraging AWS services. This role involves leading cloud initiatives, ensuring robust cloud infrastructure, and driving innovation in cloud technologies to support the business's advanced analytics needs. Core Job Responsibilities Support the development and maintenance of company-wide frameworks and libraries that enable faster, better, and more informed decision-making within the business, creating significant business value from data & analytics. Ensure data availability and accessibility for prioritized Advanced Analytics scope, and maintain stable, scalable, and modular data science pipelines from data exploration to deployment. Acquire, ingest, and process data from multiple sources and systems into our cloud platform (AWS), ensuring data integrity and security. Collaborate with data scientists to map data fields to hypotheses, and curate, wrangle, and prepare data for advanced analytical models. Implement and manage robust security measures to ensure compliant handling and management of data, including access strategies aligned with Information Security, Cyber Security, and Data Privacy principles. Develop and deploy smart automation tools based on cloud technologies, aligned with business priorities and needs. Oversee the timely delivery of Advanced Analytics solutions in coordination with the rest of the team and per requirements and timelines, ensuring alignment with business goals. Collaborate closely with the Data Science team and AI Engineers to understand platform needs and lead the development of solutions that support their work. Troubleshoot and resolve issues related to the AWS platform, ensuring minimal downtime and optimal performance. Define and document best practices and strategies regarding application deployment and infrastructure maintenance. Drive continuous improvement of the AWS Cloud platform by contributing and implementing new ideas and processes. Supervisory/Management Responsibilities Direct Reports: None. Indirect Reports: None. Position Accountability/Scope The Cloud Engineer is accountable for delivering targeted business impact per initiative in collaboration with key stakeholders. This role involves significant responsibility for the architecture and management of Abbott's strategic cloud platforms and AI/AA programs, enabling faster, better, and more informed decision-making within the business. Minimum Education Master in relevant field (e.g., computer science, electrical engineering) Minimum Experience/Training Required At least 3-5 years of relevant experience, with a strong track record in building solutions/applications using AWS services Proven ability to work across structured, semi-structured, and unstructured data, extracting information and identifying linkages across disparate data sets. Proficiency in multiple programming languages – Javascript, Python, Scala, PySpark or Java. Extensive knowledge and experience with various database technologies, including distributed processing frameworks, relational databases, MPP databases, and NoSQL data stores. Deep understanding of Information Security principles to ensure compliant handling and management of data. Significant experience with cloud platforms, preferably AWS and its ecosystem. Advanced knowledge of development in CICD (Continuous Integration and Continuous Delivery) environments. Strong background in data warehousing / ETL tools. Proficiency in DevOps practices and tools such as Jenkins, Terraform, etc. Proficiency in serverless architecture and services like AWS Lambda. Understanding of security best practices and implementation in cloud environments. Ability to understand business objectives and create cloud-based solutions to meet those objectives. Result-driven, analytical, and creative thinker. Proven ability to work with cross-functional teams and bridge the gap between business and data science. Fluency in English is a must; additional languages are a plus. Additional Technical Skills Experience with front-end frameworks preferably React JS. Knowledge of back-end frameworks like Django, Flask, or Node.js. Familiarity with database technologies such as RedShift, MySQL, or DynamoDB. Understanding of RESTful API design and development. Experience with version control systems like CodeCommit.
Posted 3 days ago
3.0 - 7.0 years
0 Lacs
pune, maharashtra
On-site
As a GCP Architect based in Pune, you will be responsible for utilizing your expertise in DevOps Model, Access control, Python, SQL, Terraform, and CI/CD to design and implement robust solutions on the Google Cloud Platform. Your primary focus will be on architecting efficient and secure cloud infrastructure that meets the specific requirements of the organization. Key responsibilities include designing and implementing access control mechanisms to ensure data security, developing automation scripts using Python and SQL for seamless operations, and leveraging Terraform for infrastructure as code. Additionally, you will be involved in setting up continuous integration and continuous deployment pipelines to streamline the software delivery process. The ideal candidate for this role should have a strong background in GCP architecture and a thorough understanding of DevOps principles. You should be adept at designing scalable and reliable cloud solutions while adhering to best practices in access management and automation. Your ability to work in a collaborative environment and communicate effectively with cross-functional teams will be crucial for the success of this role. If you are passionate about cloud architecture and possess the necessary skills to drive innovation in a dynamic environment, we encourage you to apply for this Contract to Hire opportunity as a GCP Architect in Pune.,
Posted 3 days ago
6.0 years
0 Lacs
Gurugram, Haryana, India
On-site
Profile : Site Reliability Engineer (SRE) Experience Required : 6+ Years Locations : Mumbai, Gurgaon, Chennai Work Arrangement : Hybrid Key Responsibilities Design and implement scalable, resilient cloud-native infrastructure across AWS/Azure/GCP platforms Own the SRE function including availability, latency, performance monitoring, emergency response, and capacity planning Collaborate with engineering and product teams to improve system reliability, speed, and performance Set up, maintain, and improve CI/CD pipelines using industry-standard tools Perform load and stress testing, analyze performance bottlenecks, and provide remediation strategies Manage incident response and conduct post-incident reviews Implement Infrastructure as Code using Terraform Monitor system performance and implement proactive measures for system optimization Mandatory Technical Skills Cloud Architecture : Hands-on experience with AWS/Azure/GCP platforms Terraform : Infrastructure as Code implementation and management Performance Testing : Proficiency with JMeter, Gatling, k6, or Locust Load Balancing : Experience with ALB, NLB, Azure Load Balancer, GCP Load Balancer CI/CD Pipelines : Jenkins, GitHub Actions, Azure DevOps, or GCP Build Additional Required Skills Cloud certifications (AWS/Azure/GCP Solution Architect preferred) SRE expertise in availability, performance monitoring, and capacity planning Monitoring tools : CloudWatch, Prometheus, Grafana Container technologies : Docker, Kubernetes, ECS/AKS/GKE Scripting & automation : Python, Bash Database operations : MySQL, PostgreSQL, NoSQL databases Strong incident management and troubleshooting capabilities Analytical problem-solving mindset (ref:hirist.tech)
Posted 3 days ago
35.0 years
0 Lacs
Noida, Uttar Pradesh, India
On-site
Job Description At Optimum Info, we are continually innovating and developing a range of software solutions empowering the Network Development and Field Operations businesses at Automotive, Power Sports and Equipment industries. Our integrated suite of comprehensive solutions provides a seamless and rich experience to our customers, helping them become more effective at their work and create an impact on the organization. Our sharp cultural focus on outstanding customer service and employee empowerment is core to our growth and success. As a growing company, we offer incredible opportunities for learning and growth with opportunity to manage high-impact business solution. Position Overview The Infrastructure Engineer will be responsible for maintaining Optimum's server and end-user infrastructure and work on initiatives to enhance the performance, reliability, and security of assets on the Amazon cloud. The position is based in Noida, India and will collaborate with infrastructure and Infosec team members based out of Optimum's other locations (Ahmedabad, India and Los Angeles, USA). Key Responsibilities AWS Infrastructure Management : Provision, configure, and monitor cloud infrastructure on AWS, ensuring high availability, performance, and security. Server Administration : Manage and maintain Windows and Linux servers, including patching, backup, and troubleshooting. Resource Optimization : Continuously review cloud resource utilization to optimize performance and reduce costs. Monitoring & Incident Response : Set up and manage monitoring tools, respond to alerts, and troubleshoot infrastructure issues. Security & Compliance : Ensure compliance with security policies, manage SSL certificates, and support access control mechanisms. Collaboration & Automation : Work with DevOps and Security teams to implement automation, infrastructure-as-code (IaC), and best practices. Office 365 Administration : Oversee O365 services, user management, and security settings. Desired Qualifications & Experience Bachelor's degree in engineering or a related field, with 35 years of experience managing cloud infrastructure. Cloud operations certification is a plus. Hands-on experience with AWS services such as EC2, S3, IAM, VPC, and CloudWatch. Strong knowledge of Windows and Linux server administration. Experience with cloud cost optimization strategies. Familiarity with Infrastructure-as-Code tools (Terraform, CloudFormation) is a plus. Strong English communication skills and proficiency in MS Office (Word, Excel, PowerPoint). Preferred Certifications AWS Certified SysOps Administrator Associate AWS Certified Solutions Architect Associate (ref:hirist.tech)
Posted 3 days ago
4.0 years
0 Lacs
Itanagar, Arunachal Pradesh, India
Remote
Job Title : Monitoring & Observability Engineer Datadog Specialist Experience : 4+ Years Location : [Specify Location or Remote] Job Type : Full-Time Job Summary We are looking for a talented Observability Engineer with hands-on experience in Datadog to enhance our infrastructure and application monitoring capabilities. The ideal candidate will have a strong understanding of performance monitoring, alerting, and observability in cloud-native Responsibilities : Design, implement, and maintain observability solutions using Datadog for applications, infrastructure, and cloud services. Set up dashboards, monitors, and alerts to proactively detect and resolve system issues. Collaborate with DevOps, SRE, and application teams to define SLOs, SLIs, and KPIs for performance monitoring. Integrate Datadog with services such as AWS, Kubernetes, CI/CD pipelines, and logging tools. Conduct performance tuning and root cause analysis of production incidents. Automate observability processes using infrastructure-as-code and scripting (e.g., Terraform, Python). Stay up-to-date with the latest features and best practices in Datadog and observability Skills : 4+ years of experience in monitoring/observability, with 2+ years hands-on experience in Datadog Strong experience with Datadog APM, infrastructure monitoring, custom metrics, and dashboards Familiarity with cloud platforms like AWS, GCP, or Azure Experience monitoring Kubernetes, containers, and microservices Good knowledge of log management, tracing, and alert tuning Proficient with scripting (Python, Shell) and IaC tools (Terraform preferred) Solid understanding of DevOps/SRE practices and incident Skills : Datadog certifications (e.g., Datadog Certified Observability Engineer) Experience integrating Datadog with CI/CD tools, ticketing systems, and chatops Familiarity with other monitoring tools (e.g., Prometheus, Grafana, New Relic, Splunk) Knowledge of performance testing tools (e.g., JMeter, k6) (ref:hirist.tech)
Posted 3 days ago
4.0 - 8.0 years
0 Lacs
Pune, Maharashtra, India
Remote
Job Title : DevOps Engineer / Cloud Engineer Location : Bangalore / Pune / Remote (Hybrid / Partially Remote) Employment Type : Full-time Experience : 4 to 8 Years Notice Period : Up to 30 days About The Role We are looking for an experienced DevOps / Cloud Engineer to architect, build, and maintain scalable, secure, and highly available cloud-native infrastructure. You will work closely with development and product teams to automate deployments, optimize reliability, and enable rapid delivery through modern CI/CD practices across AWS, Azure, or GCP environments. Key Responsibilities Design, implement, and manage cloud infrastructure on AWS, Azure, or GCP following best practices for scalability, security, and cost-efficiency. Build, maintain, and enhance CI/CD pipelines to enable automated build, test, and deployment workflows. Containerize applications using Docker and orchestrate them using Kubernetes (self-managed or managed services like EKS/AKS/GKE). Define, write, and maintain infrastructure-as-code using Terraform (or equivalent) to provision and version cloud resources. Monitor system health, performance, and availability; implement alerting, logging, and observability (metrics, tracing, centralized logs). Collaborate with developers to enable efficient release processes, including blue/green or canary deployments, rollback mechanisms, and feature flag integration. Implement security best practices at infrastructure and deployment layers (IAM, secrets management, network segmentation, vulnerability scanning). Manage environment configurations and secrets securely (e.g., HashiCorp Vault, AWS Secrets Manager, Azure Key Vault). Perform capacity planning, cost optimization, and disaster recovery planning. Troubleshoot production incidents, perform root cause analysis, and drive postmortems with actionable remediation. Automate routine operational tasks (backup, scaling, health checks, patching) and maintain runbooks. Assist in onboarding, mentoring junior engineers, and driving DevOps culture across the organization. Required Skills & Qualifications 4 to 8 years of hands-on experience in DevOps, Cloud Engineering, or Site Reliability Engineering. Strong experience with at least one major cloud provider : AWS, Azure, or GCP. Proficiency with Docker and container lifecycle management. Production experience with Kubernetes (deployment patterns, Helm charts, autoscaling, service mesh familiarity is a plus). Solid understanding and implementation of CI/CD pipelines using tools like GitHub Actions, GitLab CI, Jenkins, CircleCI, etc. Infrastructure-as-code expertise, especially Terraform (writing modules, state management, workspace strategies). Knowledge of networking, load balancing, DNS, VPN, firewalls, and cloud security configurations. Familiarity with logging/monitoring stacks (Prometheus/Grafana, ELK/EFK, Cloud-native equivalents). Scripting skills (Python, Bash, or equivalent) for automation. Experience with version control systems (Git), branching strategies, and release management. Strong problem-solving skills and ability to operate in a fast-paced, collaborative environment. Nice-to-Have Experience with service meshes (Istio, Linkerd) or API gateways. Exposure to serverless architectures (Lambda, Functions, Cloud Run). Knowledge of policy-as-code (e.g., Open Policy Agent) and compliance automation. Familiarity with GitOps paradigms (Flux, Argo CD). Experience with database operations in cloud (managed PostgreSQL, Redis, etc.). Working knowledge of observability platforms (OpenTelemetry). Certification(s) such as AWS Certified DevOps Engineer, Azure DevOps Engineer Expert, or Google Professional Cloud DevOps Engineer. Working Model Hybrid / Partially Remote: Flexibility to work remotely with periodic in-office collaboration (for Bangalore / Pune-based candidates). Core overlap hours to ensure team sync, with some flexibility in start/end times. Cross-functional collaboration with product, development, QA, and security teams. What We Offer Flexible work location Learning and certification support. Modern cloud-native tech stack. Health and wellness benefits (if applicable to your company). Inclusive culture with autonomy and ownership. (ref:hirist.tech)
Posted 3 days ago
0 years
0 Lacs
Chennai, Tamil Nadu, India
On-site
Job Description : Review and evaluate our existing cloud infrastructure, identifying areas for improvement and implementing necessary changes to enhance networking and firewalls (WAFs). Required Skills : Handson Experience in handling Linux Servers and Windows Handson Experience in VMWare and VSphere Configure and monitor Routers, Switches, Firewall including VPC Enabling, Windows Monitor and troubleshoot issues and configure everything in AWS. Experience in Terraform and creating a duplicate environment would be added advantage Should have strong exposure in Security focus as OS , Patches, Firewalls , Lan Debugging, Windows, Servers Added Advantage if implemented CCNA or Redhat Linux Job Description : Configure and deploy cyber security tools (e.g., Sentinel One) to protect our AWS infrastructure from potential threats and vulnerabilities. Implement best practice resiliency measures, such as spam filters, to ensure the stability and reliability of our cloud infrastructure. Conduct thorough reviews of our existing AWS infrastructure, identifying areas for optimization and implementing best practice configurations on existing web services. Design, deploy, and manage cloud-based solutions on AWS platform. Collaborate with development teams to optimize application performance and scalability in the AWS environment. Ensure the security, availability, and reliability of AWS resources and applications. Stay up to date with the latest trends and developments in cloud computing and AWS services. Certification in VMware administration is added advantage (ref:hirist.tech)
Posted 3 days ago
1.0 - 10.0 years
0 Lacs
pune, maharashtra
On-site
As a DevOps Tech Lead at BMC, you will play a crucial role in designing, developing, and implementing complex applications using cutting-edge technologies. Your responsibilities will include overseeing the organization's networking infrastructure across on-premises environments and multiple cloud platforms such as AWS, Google Cloud Platform, and Oracle. You will lead a team of network engineers and technicians, ensuring optimal network performance, security, and reliability. Your key responsibilities will include designing, implementing, and maintaining scalable, secure, and high-performance network infrastructure. You will develop and enforce network architecture standards, policies, and procedures, oversee network upgrades, expansions, and migrations to new technologies. Additionally, you will lead and mentor the network team, allocate resources effectively, and manage team performance. Monitoring network performance, ensuring high availability and reliability, implementing and managing network security measures, and conducting regular security assessments will be essential aspects of your role. You will also oversee the resolution of complex network issues, provide third-level support for network-related incidents, and ensure timely troubleshooting to minimize downtime. Your role will involve planning and managing network projects from conception through implementation, coordinating with other departments and stakeholders for successful project delivery, maintaining project documentation, and reporting progress to senior management. You will also manage relationships with network equipment vendors and service providers, evaluate and recommend new network technologies, and oversee procurement and maintenance of network hardware and software. To excel in this role, you should have a minimum of 10 years of experience in network engineering, with at least 1 year in a leadership role. Cisco Certified Network Professional (CCNP) or Cisco Certified Internetwork Expert (CCIE) certification is required, along with strong knowledge of network protocols, architecture, and security. Experience in configuring and managing various network devices and cloud technologies, as well as proficiency in scripting and automation tools, is crucial for success. While experience with cloud networking, SD-WAN, network virtualization technologies, ITIL processes, and other technologies are considered advantageous, our team is committed to helping you develop these skills. The role may require some on-site work and participation in an on-call rotation. At BMC, our culture is centered around our people, where authenticity and individuality are celebrated. We value diversity and encourage talents from varied backgrounds to join us in facing the world with the best ideas. If you are passionate about BMC and this opportunity resonates with you, we encourage you to apply, regardless of your qualifications. Your unique perspective and enthusiasm are highly valued in our team.,
Posted 3 days ago
6.0 years
0 Lacs
Greater Kolkata Area
Remote
Experience : 6+ Years Location : 100% Remote Availability : Immediate to 15 Days Job Type : Full-time We're looking for an experienced Senior Python Engineer with deep expertise in Python, FastAPI, Kafka, GCP, and cloud-native microservices. If you're passionate about building high-performance APIs and event-driven systems with best-in-class observability and DevOps practices, we'd love to connect with you! Key Responsibilities : & API Development : Design and implement scalable microservices using FastAPI, Pydantic, and Async I/O Build high-throughput, event-driven services leveraging Confluent Kafka DevOps & CI/CD : Implement and manage CI/CD pipelines using GitHub Actions (or similar) Deploy secure, containerized applications using Docker, Terraform, and GCP services like Cloud Run, Pub/Sub, and Eventarc Monitoring & Observability Integrate monitoring/logging tools such as New Relic, Cloud Logging, and Cloud Monitoring Ensure visibility, performance tuning, and reliability in production environments Team Leadership & Best Practices Define coding standards, enforce testing strategies (Pytest/unittest), conduct code reviews Mentor junior developers and ensure high code quality across the board Data Processing & Transformation Work with caching (Redis), data transformation (XSLT, XML/XSD), and various database systems Handle structured/unstructured data with Kafka in JSON, Avro, and Protobuf formats Required Skills Expert in Python with strong FastAPI experience Proficient in Async I/O, Pydantic, and Pytest/unittest Hands-on experience with Kafka (Confluent), Docker, and Terraform Cloud experience (preferably GCP) with services like Cloud Run, Pub/Sub, Eventarc Monitoring with New Relic or equivalent observability tools Knowledge of XML/XSD/XSLT transformations Familiarity with caching and database/storage systems (PostgreSQL, Redis, etc.) Git/GitHub and CI/CD with GitHub Actions Good To Have Experience with PyPI package creation, Tox, Ruff Exposure to GKE Autopilot, Artifact Registry, IAM, Secret Manager Understanding of Workload Identity Federation and VPC networking Bash scripting and familiarity with legacy systems like IBM AS/400 or MS SQL Server Soft Skills Strong leadership and mentoring ability Excellent collaboration and communication skills Analytical thinking and a drive for continuous improvement Benefits Competitive compensation Remote-first work culture and flexible hours Access to continuous learning resources Opportunity to work on cutting-edge cloud-native architecture A collaborative, high-performing tech team (ref:hirist.tech)
Posted 3 days ago
7.0 years
0 Lacs
Greater Kolkata Area
On-site
Who are we looking for? We are looking for 7+ years of administrator experience in MongoDB/Cassandra/Snowflake Databases. This role is focused on production support, ensuring database performance, availability, and reliability across multiple clusters. The ideal candidate will be responsible for ensuring the availability, performance, and security of our NoSQL database environment. You will provide 24/7 production support, troubleshoot issues, monitor system health, optimize performance, and collaborate with cross-functional teams to maintain a reliable and efficient Snowflake platform. Technical Skills Proven experience as a MongoDB/Cassandra/Snowflake Databases Administrator or similar role in production support environments. 7+ years of hands-on experience as a MongoDB DBA supporting production environments. Strong understanding of MongoDB architecture, including replica sets, sharding, and aggregation framework. Proficiency in writing and optimizing complex MongoDB queries and indexes. Experience with backup and recovery solutions (e.g., mongodump, mongorestore, Ops Manager). Solid knowledge of Linux/Unix systems and scripting (Shell, Python, or similar). Experience with monitoring tools like Prometheus, Grafana, DataStax OpsCenter, or similar. Understanding of distributed systems and high-availability concepts. Proficiency in troubleshooting cluster issues, performance tuning, and capacity planning. In-depth understanding of data management (e.g. permissions, recovery, security and monitoring Understanding of ETL/ELT tools and data integration patterns. Strong troubleshooting and problem-solving skills. Excellent communication and collaboration abilities. Ability to work in a 24/7 support rotation and handle urgent production issues. Strong understanding of relational database concepts. Experience with database design, modeling, and optimization is good to have Familiarity with data security is the best practice and backup : Support & Incident Management : Provide 24/7 support for MongoDB environments, including on-call rotation. Monitor system health and respond to s, incidents, and performance degradation issues. Troubleshoot and resolve production database issues in a timely manner. Database Administration Install, configure, and upgrade MongoDB clusters in on-prem or cloud environments. Perform routine maintenance including backups, restores, indexing, and data migration. Monitor and manage replica sets, sharding, and cluster Tuning & Optimization : Analyze query and indexing strategies to improve performance. Tune MongoDB server parameters and JVM settings where applicable. Monitor and optimize disk I/O, memory usage, and CPU utilization . Security & Compliance Implement and manage access control, roles, and authentication mechanisms (LDAP, x.509, SCRAM). Ensure encryption, auditing, and compliance with data governance and security & Monitoring : Create and maintain scripts for automation of routine tasks (e.g., backups, health and checks Set up and maintain monitoring tools (e.g., MongoDB Ops Manager, Prometheus/Grafana, MMS). Documentation & Collaboration Maintain documentation on architecture, configurations, procedures, and incident reports. Work closely with application and infrastructure teams to support new releases and : Experience with MongoDB Atlas and other cloud-managed MongoDB services. MongoDB certification (MongoDB Certified DBA Experience with automation tools like Ansible, Terraform, or Puppet. Understanding of DevOps practices and CI/CD integration. Familiarity with other NoSQL and RDBMS technologies is a plus. Education qualification : Any degree (ref:hirist.tech)
Posted 3 days ago
5.0 - 9.0 years
0 Lacs
chennai, tamil nadu
On-site
As a member of the Platform Observability Engineering team within Ford's Data Platforms and Engineering (DP&E) organization, you will contribute to building and maintaining a top-tier platform for monitoring and observability. This platform focuses on the four golden signalslatency, traffic, errors, and saturationproviding essential data to support operations, root cause analysis, continuous improvement, and cost optimization. You will collaborate with platform architects to help design, develop, and maintain a scalable and reliable platform, ensuring smooth integration with systems used across various teams. Your contributions will be key in improving MTTR and MTTX through increased visibility into system performance. Working with stakeholders, you will integrate observability data into their workflows, develop insightful dashboards and reports, continuously improve platform performance and reliability, optimize costs, and stay updated with industry best practices and technologies. The role focuses on building and maintaining a robust platform rather than developing individual monitoring tools, creating a centralized, reliable source of observability data that empowers data-driven decisions and accelerates incident response across the organization. Responsibilities: - Design and Build Data Pipelines: Architect, develop, and maintain scalable data pipelines and microservices supporting real-time and batch processing on GCP. - Service-Oriented Architecture (SOA) and Microservices: Design and implement SOA and microservices-based architectures for modular, flexible, and maintainable data solutions. - Full-Stack Integration: Contribute to the seamless integration of front-end and back-end components, ensuring robust data access and UI-driven data exploration. - Data Ingestion and Integration: Lead the ingestion and integration of data from various sources into the data platform, ensuring standardized and optimized data for analytics. - GCP Data Solutions: Utilize GCP services (BigQuery, Dataflow, Pub/Sub, Cloud Functions, etc.) to build and manage data platforms meeting business needs. - Data Governance and Security: Implement and manage data governance, access controls, and security best practices while leveraging GCP's native security features. - Performance Optimization: Continuously monitor and improve the performance, scalability, and efficiency of data pipelines and storage solutions. - Collaboration and Best Practices: Define best practices, design patterns, and frameworks for cloud data engineering by closely working with data architects, software engineers, and cross-functional teams. - Automation and Reliability: Automate data platform processes to enhance reliability, reduce manual intervention, and improve operational efficiency. Qualifications: - Technical Skills: Proficiency in Java, Angular, or any JavaScript technology with experience in designing and deploying cloud-based data pipelines and microservices using GCP tools like BigQuery, Dataflow, and Dataproc. - Service-Oriented Architecture and Microservices: Strong understanding of SOA, microservices, and their application within a cloud data platform context. Develop robust, scalable services using Java Spring Boot, Python, Angular, and GCP technologies. - Full-Stack Development: Knowledge of front-end and back-end technologies enabling collaboration on data access and visualization layers (e.g., React, Node.js). - Design and develop RESTful APIs for seamless integration across platform services. - Implement robust unit and functional tests to maintain high standards of test coverage and quality. - Database Management: Experience with relational (e.g., PostgreSQL, MySQL) and NoSQL databases, as well as columnar databases like BigQuery. - Data Governance and Security: Understanding of data governance frameworks and implementing RBAC, encryption, and data masking in cloud environments. - CI/CD and Automation: Familiarity with CI/CD pipelines, Infrastructure as Code (IaC) tools like Terraform, and automation frameworks. - Manage code changes with GitHub and troubleshoot and resolve application defects efficiently. - Ensure adherence to SDLC best practices, independently managing feature design, coding, testing, and production releases. - Problem-Solving: Strong analytical skills with the ability to troubleshoot complex data platform and microservices issues. Certifications (Preferred): GCP Data Engineer, GCP Professional Cloud,
Posted 3 days ago
5.0 - 9.0 years
0 Lacs
pune, maharashtra
On-site
As a Full Stack Developer, you will be responsible for utilizing your expertise in React.js, Node.js, and AWS Lambda to develop a custom enterprise platform that interacts with various SDLC tools. This platform aims to enhance tool administration, automate access provisioning and deprovisioning, manage licenses efficiently, and provide centralized dashboards for governance and monitoring purposes. With a minimum of 4-6 years of hands-on experience in Full Stack Development, you should possess a strong command over React.js for building component-based front-end architecture. Your backend skills in Node.js and proficiency in RESTful API development will be crucial for the success of this project. Additionally, your solid experience with AWS services such as Lambda, API Gateway, DynamoDB, and S3 will be highly valued. Your role will also involve integrating and automating workflows for SDLC tools like JIRA, Jenkins, GitLab, Bitbucket, GitHub, and SonarQube. A good understanding of OAuth2, SSO, and API key-based authentications is essential. Familiarity with CI/CD pipelines, microservices, and event-driven architectures will further enhance your contributions to the project. It is expected that you bring in-depth knowledge of Git and modern development practices to the table. Strong problem-solving skills and the ability to work independently are qualities that will be beneficial in this role. While not mandatory, experience with Infrastructure-as-Code tools like Terraform or CloudFormation would be advantageous. Familiarity with AWS EventBridge, Step Functions, or other serverless orchestration tools is considered a plus. Knowledge of enterprise-grade authentication methods such as LDAP, SAML, or Okta, as well as familiarity with monitoring/logging tools like CloudWatch, ELK, or DataDog, are also desirable skills. Join us in this exciting opportunity to work on a cutting-edge enterprise platform and contribute to streamlining processes and enhancing efficiency within the organization.,
Posted 3 days ago
3.0 - 7.0 years
0 Lacs
pune, maharashtra
On-site
As a Business Consulting-Azure Infra Consultant at NTT DATA, your primary responsibility will be to utilize your expertise in Azure infrastructure to support our clients" needs. You will be joining our team in Pune, Maharashtra, India, where you will have the opportunity to work with a global leader in business and technology services. The ideal candidate for this role will have hands-on experience with Azure infrastructure and be proficient in writing code, particularly in the areas of Pure Azure, Infrastructure as Code (IaC), Terraform, and GitHub. Your role will involve actively contributing to the development and implementation of solutions using these technologies to help clients innovate, optimize, and transform for long-term success. While SRE (Site Reliability Engineering) and DevOps experience is not required for this particular position, candidates with a strong background in Azure infrastructure and code development are encouraged to apply. At NTT DATA, we are committed to fostering an inclusive and forward-thinking environment where passionate individuals can thrive and grow alongside our diverse team of experts. NTT DATA is a trusted global innovator with a strong presence in over 50 countries, serving 75% of the Fortune Global 100 companies. As a Global Top Employer, we are dedicated to empowering our clients with cutting-edge solutions in data, artificial intelligence, industry-specific technologies, and digital infrastructure. By being a part of NTT Group, a major investor in research and development, we are at the forefront of driving organizations and societies towards a sustainable digital future. If you are ready to take on the challenge of being a key player in the digital and AI infrastructure space and contribute to the success of leading organizations worldwide, we encourage you to apply and become a part of our dynamic team at NTT DATA. Visit us at us.nttdata.com to learn more about our innovative solutions and global impact.,
Posted 3 days ago
3.0 - 7.0 years
0 Lacs
pune, maharashtra
On-site
The Infrastructure Engineer will be responsible for designing, building, and maintaining cloud infrastructure on Azure and on-prem cloud environments to support application deployments. You will manage and orchestrate containerized applications using Kubernetes, automate infrastructure provisioning and configuration management using Terraform and Ansible, and develop CI/CD pipelines for application deployment and infrastructure automation. Additionally, you will implement and configure monitoring solutions like Prometheus and Elastic Stack for system monitoring, logging, and alerting. Troubleshooting performance issues, optimizing system performance, ensuring system reliability, collaborating with other teams, defining and enforcing infrastructure-as-code and security policies, and continuously improving system performance, scalability, and cost-efficiency will be key aspects of your role. Creating detailed documentation of infrastructure and automation processes will also be part of your responsibilities. The ideal candidate for this position should have a Bachelor's degree in Computer Science, Engineering, or a related field, along with proven experience as an Infrastructure Engineer or in a similar role. You should possess strong experience with Azure Cloud Services, hands-on experience with Kubernetes cluster administration and Helm charts, proficiency in Terraform for infrastructure as code, experience with Ansible for configuration management, and solid knowledge of Prometheus and Elastic Stack (ELK) for monitoring and logging. Familiarity with CI/CD tools, networking concepts, security best practices, and cloud security principles is required. Strong scripting skills in Bash, Python, or similar languages, excellent troubleshooting skills, and attention to detail are also essential. Nice-to-have qualifications include certifications in Azure or Kubernetes, experience with other monitoring tools like Grafana or Datadog, knowledge of container security tools, and experience with GitOps tools. If you meet these qualifications and are looking to join a dynamic team where you can play a critical role in building, scaling, and maintaining infrastructure to ensure reliability, scalability, and security, we encourage you to apply for this position.,
Posted 3 days ago
4.0 - 8.0 years
0 Lacs
navi mumbai, maharashtra
On-site
As a DevOps Engineer at Allerin, you will be responsible for owning the cloud infrastructure by designing, provisioning, and maintaining multi-region AWS and GCP environments using Terraform and Helm. Your role will involve optimizing cost, performance, and scalability for high-traffic AI/IoT platforms. You will also be required to evolve GitLab pipelines to achieve true trunk-based, push-to-prod workflows with automated testing, security scans, and blue-green deployments. It will be essential to champion immutable builds and container security best practices in this capacity. In addition, your responsibilities will include implementing end-to-end monitoring with Prometheus, Grafana, Loki, and alerting via PagerDuty. You will lead post-incident reviews, driving Mean Time To Recovery (MTTR) down through runbook automation and chaos testing. Furthermore, you will embed shift-left security practices (SAST, DAST, IaC scanning) into pipelines and collaborate with the SOC2 team to maintain least-privilege IAM, secret rotation, and audit trails. Collaboration and a strong team culture are integral parts of this role. You will work closely with developers, QA, and product teams to ensure reliability is incorporated into every user story. Additionally, mentoring junior engineers in infrastructure-as-code, Kubernetes, and site reliability principles will be part of your responsibilities. To be successful in this role, you should have at least 4 years of experience running production workloads on AWS, GCP, or Azure. Strong knowledge of IaC (Terraform or Pulumi) and Kubernetes (EKS/GKE preferred), proven experience with GitLab or GitHub Actions CI/CD automation, a deep understanding of Linux internals, networking, and container runtime security, scripting skills in Bash, Python, or Go for automation and tooling, and hands-on experience with observability stacks (Prometheus/Grafana/ELK/Loki) are required qualifications for this position.,
Posted 3 days ago
3.0 - 7.0 years
0 Lacs
haryana
On-site
You will play a crucial role in redefining capital program oversight by integrating advanced AI systems at Fort Hill -LogiX. As a Platform Engineer, you are expected to have expertise in DevOps, Generative AI systems, and infrastructure automation to contribute to scaling and stabilizing our next-gen intelligence platform. Your primary responsibilities will involve building foundational systems supporting large language model integrations like Claude, Gemini, GPT, and Bedrock, facilitating real-time data workflows, and enabling intelligent decision-making in complex financial and operational environments. Your key responsibilities will include Infrastructure & Cloud Engineering, where you will design and maintain resilient infrastructure on AWS using services like EC2, ECS, RDS, Lambda, IAM, and VPC networking. You will manage and scale containerized applications through Docker and Kubernetes (EKS) in multi-environment production workflows. Implementing Infrastructure-as-Code (IaC) using tools like Terraform or CloudFormation will be essential, along with building and maintaining CI/CD pipelines for secure and frequent deployments. Additionally, configuring observability tooling such as CloudWatch, Grafana, Prometheus, and ELK will be necessary for system reliability and performance monitoring. In the realm of Generative AI Infrastructure, you will deploy and orchestrate integrations with AWS Bedrock (Claude, Titan), Google Cloud Gemini Enterprise, and OpenAI GPT APIs via Azure or OpenAI platform. Your tasks will also involve building secure and scalable AI inference flows with mechanisms like throttling, fallback, and caching. Supporting Retrieval-Augmented Generation (RAG) pipelines using LangChain, LangGraph, and vector databases will be part of your role, along with optimizing latency, cost, and throughput for AI-enabled services. Ensuring Security & Reliability will involve collaborating with platform and product teams to enforce secure deployments, encryption (TLS, KMS, IAM), and automate role-based access controls and service provisioning. You will be responsible for leading infrastructure incident response, root cause diagnostics, and uptime management in production environments. Automation & Operations will require you to automate testing, deployment, and rollback of AI services and platform features. Implementing event-driven and lazy approval workflows for secure and efficient platform operations will be essential, along with scaling internal tools and platforms to support rapid prototyping and stable production delivery. To qualify for this role, you should have at least 3+ years of experience in DevOps, Cloud Engineering, or Platform Engineering roles. A deep understanding of AWS, Kubernetes (EKS), Terraform, and CI/CD workflows is necessary, along with experience integrating Generative AI models/APIs like Claude, GPT, or Gemini. Strong scripting skills in Python, Bash, or similar languages are required, as well as an understanding of scalable APIs, distributed systems, and cloud-native microservices. Familiarity with LangGraph, LangChain, and vector stores like Pinecone, PGVector, Weaviate is beneficial. Experience deploying LLM applications with real-world data pipelines and observability, certification in AWS (e.g., DevOps Engineer, Solutions Architect), and prior work in platform-as-a-service or AI tool development environments will be advantageous for this role at Fort Hill LogiX.,
Posted 3 days ago
6.0 - 10.0 years
0 Lacs
kochi, kerala
On-site
We are looking for an experienced Lead DevOps Engineer to lead the strategy, design, and implementation of DevOps infrastructure across cloud and on-premises environments. As a technical leader, you will mentor DevOps engineers, collaborate with cross-functional teams, and establish best practices to ensure reliable, secure, and scalable infrastructure supporting the product lifecycle. Responsibilities include overseeing scalable and secure infrastructure design, implementing infrastructure as code (IaC) using tools like Terraform or CloudFormation, managing CI/CD pipelines, implementing monitoring solutions, leading incident response efforts, and developing automation strategies. You will also mentor junior/mid-level DevOps engineers and collaborate with other teams to guarantee smooth deployment and management of software infrastructure. Key qualifications for this role include 6+ years of hands-on DevOps experience, knowledge of cloud platforms like Azure, AWS, and GCP, experience in containerization with Docker and Kubernetes, and proficiency in CI/CD tools such as Jenkins, GitHub Actions, and TeamCity. Strong problem-solving skills, leadership abilities, and a Bachelor's degree in a relevant field are required. Preferred qualifications include relevant certifications, experience in fast-paced product environments, and knowledge of security best practices and compliance standards. Key competencies for this role include leadership and mentoring capabilities, strategic thinking, ability to manage multiple priorities, passion for innovation and automation, and clear communication and collaboration skills. Join us at Admaren, where we are revolutionizing the maritime domain with cutting-edge technology. As a Lead DevOps Engineer, you will drive infrastructure innovation, influence engineering culture, and work with a team committed to excellence. Shape the future of maritime software systems with us and lead from the front.,
Posted 3 days ago
8.0 - 12.0 years
0 Lacs
pune, maharashtra
On-site
As a Senior Python Developer, you will be responsible for utilizing your expertise in Python programming language to develop and maintain applications. Your primary skill set should include proficiency in Python, Hitachi Ops Center Automator, Oracle DB/MySQL/Postgres, Django, Terraform, Docker & Kubernetes. Additionally, knowledge of secondary skills such as DevOps, MongoDB, and RestAPI will be beneficial. You will be working onsite at the client location in Pune for 5 days a week, collaborating with the team to deliver high-quality solutions. Your role will involve coding, testing, debugging, and implementing software applications as per the project requirements. Join us in this dynamic environment where your skills will contribute to the success of our projects.,
Posted 3 days ago
5.0 - 9.0 years
0 Lacs
chennai, tamil nadu
On-site
As a Cloud Platform Engineer, you will play a crucial role in developing and maintaining Terraform modules and patterns for AWS and Azure. Your responsibilities will include creating platform landing zones, application landing zones, and deploying application infrastructure. Managing the lifecycle of these patterns will be a key aspect of your role, encompassing tasks such as releases, bug fixes, feature integrations, and updating test cases. You will be responsible for developing and releasing Terraform modules, landing zones, and patterns for both AWS and Azure platforms. Providing ongoing support for these patterns, including bug fixing and maintenance, will be essential. Additionally, you will need to integrate new features into existing patterns to enhance their functionality and ensure that updated and new patterns meet the current requirements. Updating and maintaining test cases for patterns will also be part of your responsibilities to guarantee reliability and performance. To qualify for this role, you should have at least 5 years of experience in AWS and Azure cloud migration. Proficiency in Cloud compute (such as EC2, EKS, Azure VM, AKS) and Storage (like s3, EBS, EFS, Azure Blob, Azure Managed Disks, Azure Files) is required. A strong knowledge of AWS and Azure cloud services, along with expertise in Terraform, is essential. Possessing AWS or Azure certification would be advantageous for this position. Key Qualifications: - 5+ years of AWS/Azure cloud migration experience - Proficiency in Cloud compute and Storage - Strong knowledge of AWS and Azure cloud services - Expertise in Terraform - AWS/Azure certification preferred Mandatory Skills: Cloud AWS DevOps (Minimum 5 Years of Migration Experience) Relevant Experience: 5-8 Years This is a Full-time, Permanent, or Contractual / Temporary job with a contract length of 12 months. Benefits: - Health insurance - Provident Fund Schedule: - Day shift, Monday to Friday, Morning shift Additional Information: - Performance bonus - Yearly bonus,
Posted 3 days ago
3.0 - 7.0 years
0 Lacs
pune, maharashtra
On-site
As a talented individual with a strong technical background, you will be part of a dynamic team where expertise meets innovation. Your role will involve combining architectural design with engineering skills to establish a robust and secure environment. Your primary responsibilities will include constructing and maintaining secure CI/CD pipelines that incorporate cutting-edge DevSecOps practices and AI tools such as Copilot, Claude, Gemini, and OpenAI. You will also be tasked with implementing Role-Based Access Control (RBAC), identity-aware controls, continuous monitoring, and enforcing a zero-trust security model. In this role, you will be expected to proficiently develop, test, and secure platforms while taking ownership of the full stack and the development process. The ideal candidate should have a minimum of 6 years of experience in backend development, infrastructure management, and AI pipelines, with expertise in languages like Python, Node, Go, and Rust. Additionally, you should possess at least 3 years of experience in AI/ML workflows and frameworks such as LangChain, RAG, and ReAct. Experience with tools like Kubernetes, GitHub Actions, Terraform, Grafana, as well as a strong understanding of security standards such as SBOMs, supply chain security, OWASP, and OAuth2 is essential for this role. Previous involvement in constructing large-scale production systems and contributing to open-source projects will be highly valued. We are looking for a proactive individual who excels in problem-solving, possesses strong architectural skills, and thrives in dynamic environments. If you are someone who can efficiently navigate chaos, write high-quality code, and consistently deliver tangible results, then we encourage you to apply for this exciting opportunity. Location: Pune, India (Hybrid Preferred),
Posted 3 days ago
Upload Resume
Drag or click to upload
Your data is secure with us, protected by advanced encryption.
Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.
We have sent an OTP to your contact. Please enter it below to verify.
Accenture
39817 Jobs | Dublin
Wipro
19388 Jobs | Bengaluru
Accenture in India
15458 Jobs | Dublin 2
EY
14907 Jobs | London
Uplers
11185 Jobs | Ahmedabad
Amazon
10459 Jobs | Seattle,WA
IBM
9256 Jobs | Armonk
Oracle
9226 Jobs | Redwood City
Accenture services Pvt Ltd
7971 Jobs |
Capgemini
7704 Jobs | Paris,France