Get alerts for new jobs matching your selected skills, preferred locations, and experience range. Manage Job Alerts
8.0 - 11.0 years
0 Lacs
Gurugram, Haryana, India
On-site
About Us: Airtel Payments Bank, India's first payments bank is a completely digital and paperless bank. The bank aims to take basic banking services to the doorstep of every Indian by leveraging Airtel's vast retail network in a quick and efficient manner. At Airtel Payments Bank, we’re transforming the way banking operates in the country. Our core business is banking and we’ve set out to serve each unbanked and underserved Indian. Our products and technology aim to take basic banking services to the doorstep of every Indian. We are a fun-loving, energetic and fast growing company that breathes innovation. We encourage our people to push boundaries and evolve from skilled professionals of today to risk-taking entrepreneurs of tomorrow. We hire people from every realm and offer them opportunities that encourage individual and professional growth. We are always looking for people who are thinkers & doers; people with passion, curiosity & conviction; people who are eager to break away from conventional roles and do 'jobs never done before’. Job Summary: We are looking for a Lead TechOps Engineer to join our team in managing and scaling containerized applications using Docker, Kubernetes, and OpenShift. You will be responsible for maintaining production environments, implementing automation, and ensuring platform stability and performance. Key Skills for TechOps Engineer (Docker, Kubernetes, OpenShift) 1. Containerization & Orchestration Expertise in Docker: building, managing, and debugging containers. Proficient in Kubernetes (K8s): deployments, services, ingress, Helm charts, namespaces. Experience with Red Hat OpenShift: operators, templates, routes, integrated CI/CD. 2. CI/CD and DevOps Toolchain Jenkins, GitLab CI/CD, other CI/CD pipelines. Familiarity with GitOps practices. 3. Monitoring & Logging Experience with Prometheus, Grafana, ELK stack, or similar tools. Understanding of health checks, metrics, and alerts. 4. Infrastructure as Code Hands-on with Terraform, Ansible, or Helm. Version control using Git. 5. Networking & Security K8s/OpenShift networking concepts (services, ingress, load balancers). Role-Based Access Control (RBAC), Network Policies, Secrets management. 6. Scripting & Automation Proficiency in Bash, Python, or Go for automation tasks. 7. Cloud Platforms (Optional but Valuable) Experience with AWS, GCP, or Azure Kubernetes Service (EKS, AKS, GKE). Responsibilities: Design, implement, and maintain Kubernetes/OpenShift clusters. Build and deploy containerized applications using Docker. Manage CI/CD pipelines for smooth application delivery. Monitor system performance and respond to alerts or issues. Develop infrastructure as code and automate repetitive tasks. Work with developers and QA to support and optimize application lifecycle. Requirements: 8-11 years of experience in TechOps/DevOps/SRE roles. Strong knowledge of Docker, Kubernetes, and OpenShift. Experience with CI/CD tools like Jenkins. Proficiency in scripting (Bash, Python) and automation tools (Ansible, Terraform). Familiarity with logging and monitoring tools (Prometheus, ELK, etc.). Knowledge of networking, security, and best practices in container environments. Good communication and collaboration skills. Nice to Have: Certifications (CKA, Red Hat OpenShift, etc.) Experience with public cloud providers (AWS, GCP, Azure). GitOps and service mesh (Istio, Linkerd) experience Why Join Us? Airtel Payments Bank is transforming from a digital-first bank to one of the largest Fintech company. There could not be a better time to join us and be a part of this incredible journey than now. We at Airtel payments bank don’t believe in all work and no play philosophy. For us, innovation is a way of life and we are a happy bunch of people who have built together an ecosystem that drives financial inclusion in the country by serving 300 million financially unbanked, underbanked, and underserved population of India. Some defining characteristics of life at Airtel Payments Bank are Responsibility, Agility, Collaboration and Entrepreneurial development : these also reflect in our core values that we fondly call RACE.. Show more Show less
Posted 3 months ago
2.0 - 5.0 years
0 Lacs
Gurugram, Haryana, India
On-site
About Us: Airtel Payments Bank, India's first payments bank is a completely digital and paperless bank. The bank aims to take basic banking services to the doorstep of every Indian by leveraging Airtel's vast retail network in a quick and efficient manner. At Airtel Payments Bank, we’re transforming the way banking operates in the country. Our core business is banking and we’ve set out to serve each unbanked and underserved Indian. Our products and technology aim to take basic banking services to the doorstep of every Indian. We are a fun-loving, energetic and fast growing company that breathes innovation. We encourage our people to push boundaries and evolve from skilled professionals of today to risk-taking entrepreneurs of tomorrow. We hire people from every realm and offer them opportunities that encourage individual and professional growth. We are always looking for people who are thinkers & doers; people with passion, curiosity & conviction; people who are eager to break away from conventional roles and do 'jobs never done before’. Job Summary: We are looking for a skilled TechOps Engineer to join our team in managing and scaling containerized applications using Docker, Kubernetes, and OpenShift. You will be responsible for maintaining production environments, implementing automation, and ensuring platform stability and performance. Key Skills for TechOps Engineer (Docker, Kubernetes, OpenShift) 1. Containerization & Orchestration Expertise in Docker: building, managing, and debugging containers. Proficient in Kubernetes (K8s): deployments, services, ingress, Helm charts, namespaces. Experience with Red Hat OpenShift: operators, templates, routes, integrated CI/CD. 2. CI/CD and DevOps Toolchain Jenkins, GitLab CI/CD, other CI/CD pipelines. Familiarity with GitOps practices. 3. Monitoring & Logging Experience with Prometheus, Grafana, ELK stack, or similar tools. Understanding of health checks, metrics, and alerts. 4. Infrastructure as Code Hands-on with Terraform, Ansible, or Helm. Version control using Git. 5. Networking & Security K8s/OpenShift networking concepts (services, ingress, load balancers). Role-Based Access Control (RBAC), Network Policies, Secrets management. 6. Scripting & Automation Proficiency in Bash, Python, or Go for automation tasks. 7. Cloud Platforms (Optional but Valuable) Experience with AWS, GCP, or Azure Kubernetes Service (EKS, AKS, GKE). Responsibilities: Design, implement, and maintain Kubernetes/OpenShift clusters. Build and deploy containerized applications using Docker. Manage CI/CD pipelines for smooth application delivery. Monitor system performance and respond to alerts or issues. Develop infrastructure as code and automate repetitive tasks. Work with developers and QA to support and optimize application lifecycle. Requirements: 2-5 years of experience in TechOps/DevOps/SRE roles. Strong knowledge of Docker, Kubernetes, and OpenShift. Experience with CI/CD tools like Jenkins. Proficiency in scripting (Bash, Python) and automation tools (Ansible, Terraform). Familiarity with logging and monitoring tools (Prometheus, ELK, etc.). Knowledge of networking, security, and best practices in container environments. Good communication and collaboration skills. Nice to Have: Certifications (CKA, Red Hat OpenShift, etc.) Experience with public cloud providers (AWS, GCP, Azure). GitOps and service mesh (Istio, Linkerd) experience Why Join Us? Airtel Payments Bank is transforming from a digital-first bank to one of the largest Fintech company. There could not be a better time to join us and be a part of this incredible journey than now. We at Airtel payments bank don’t believe in all work and no play philosophy. For us, innovation is a way of life and we are a happy bunch of people who have built together an ecosystem that drives financial inclusion in the country by serving 300 million financially unbanked, underbanked, and underserved population of India. Some defining characteristics of life at Airtel Payments Bank are Responsibility, Agility, Collaboration and Entrepreneurial development : these also reflect in our core values that we fondly call RACE.. Show more Show less
Posted 3 months ago
0 years
0 Lacs
India
Remote
Step into the world of AI innovation with the Experts Community of Soul AI (By Deccan AI). We are looking for India’s top 1% Platform Engineers for a unique job opportunity to work with the industry leaders. Who can be a part of the community? We are looking for Platform Engineers focusing on building scalable and high-performance AI/ML platforms. Strong background in cloud architecture, distributed systems, Kubernetes, and infrastructure automation is expected. If you have experience in this field then this is your chance to collaborate with industry leaders. What’s in it for you? Pay above market standards The role is going to be contract based with project timelines from 2 - 12 months , or freelancing. Be a part of an Elite Community of professionals who can solve complex AI challenges. Work location could be: Remote (Highly likely) Onsite on client location Deccan AI’s Office: Hyderabad or Bangalore Responsibilities: Architect and maintain scalable cloud infrastructure on AWS, GCP, or Azure using tools like Terraform and Cloud Formation. Design and implement Kubernetes clusters with Helm, Kustomize, and Service Mesh (Istio, Linkerd). Develop CI/CD pipelines using GitHub Actions, GitLab CI/CD, Jenkins, and Argo CD for automated deployments. Implement observability solutions (Prometheus, Grafana, ELK stack) for logging, monitoring, and tracing & automate infrastructure provisioning with tools like Ansible, Chef, Puppet, and optimize cloud costs and security. Required Skills: Expertise in cloud platforms (AWS, GCP, Azure) and infrastructure as code (Terraform, Pulumi) with strong knowledge of Kubernetes, Docker, CI/CD pipelines, and scripting (Bash, Python). Experience with observability tools (Prometheus, Grafana, ELK stack) and security practices (RBAC, IAM). Familiarity with networking (VPC, Load Balancers, DNS) and performance optimization. Nice to Have: Experience with Chaos Engineering (Gremlin, LitmusChaos), Canary or Blue-Green deployments. Knowledge of multi-cloud environments, FinOps, and cost optimization strategies. What are the next steps ? 1. Register on our Soul AI website. 2. Our team will review your profile. 3 . Clear all the screening round s: Clear the assessments once you are shortlisted. As soon as you qualify all the screening rounds (assessments, interviews) you will be added to our Expert Community! 4 . Profile matching and Project Allocatio n: Be patient while we align your skills and preferences with the available project. Skip the Noise. Focus on Opportunities Built for You! Show more Show less
Posted 3 months ago
7.0 years
0 Lacs
Sahibzada Ajit Singh Nagar, Punjab, India
On-site
Everything we do is powered by our customers! Featured on Deloitte's Technology Fast 500 list and G2's leaderboard, Maropost offers a connected experience that our customers anticipate, transforming marketing, merchandising, and operations with commerce tools designed to scale with fast-growing businesses. With a relentless focus on our customers’ success, we are motivated by curiosity, creativity, and collaboration to power 5,000+ global brands. Driven by a customer-first mentality, we empower businesses to achieve their goals and grow alongside us. If you're ready to make a significant impact and be part of our transformative journey, Maropost is the place for you. Become a part of Maropost today and help shape the future of commerce! Roles & Responsibilities Build and manage a REST API stack for Maropost Web Apps. Given the architecture strategy related to our big data, analytics and cloud native product vision, work on the concrete architecture design and, when necessary, prototype it Understanding of systems architecture and ability to design scalable, performance-driven solutions. Drive innovation within the engineering team, identifying opportunities to improve processes, tools, and technologies Drive the architecture and design governance for systems and products under scope, as well as code and design reviews. Technical leadership of the development team and ensuring that they follow industry-standard best practices Evaluating and improving the tools and frameworks used in software development Design, develop and architect complex web applications Integrate with ML and NLP engines. DevOps, DBMS & Scaling on Azure or GCP. Skills & Qualifications B.E./B.Tech. Hands-on experience with tech stacks—RoR and PostgreSQL 7+ years of experience with building, including designing and architecting backend applications, web apps, and analytics, preferably in commerce cloud or marketing automation domains. Experience in deploying applications at scale in production systems. Experience with platform security capabilities (TLS, SSL etc.) Excellent track record in designing highly scalable big data/event-streaming/cloud architectures and experience with having put them in production. Advanced HLD, LLD, and Design Patterns knowledge is a must. Experience of high-performance web-scale & real-time response systems Knowledge of tenant data segregation techniques, such as schema-based multi-tenancy, database-per-tenant, and hybrid approaches, for ensuring data isolation and privacy. Knowledge of networking protocols, security standards, and best practices. Experience in building and managing API endpoints for multimodal clients. In-depth knowledge and hands-on experience in architecting and optimizing large-scale database clusters, specifically MySQL and PostgreSQL, for performance, scalability, and reliability. Proficiency in microservices architecture and containerization technologies (e.g., Docker, Kubernetes). Experience with DevOps practices and tools (e.g., CI/CD pipelines, infrastructure as code). Expertise in database design, including SQL and NoSQL databases, with a specific focus on MySQL and PostgreSQL Experience in implementing advanced indexing strategies, query optimization techniques, and database tuning methodologies for optimizing the performance of MySQL and PostgreSQL databases. Enthusiasm to learn and contribute to a challenging & fun-filled startup. A knack for problem-solving and following efficient coding practices. Very strong interpersonal communication and collaboration skills Advanced HLD, LLD, and Design Patterns knowledge is a must. Hands-on Experience (Advantageous) Proficiency in infrastructure-as-code tools such as Terraform or AWS CloudFormation. Experience with containerization technologies such as Docker and container orchestration platforms like Kubernetes. Proficiency in implementing advanced replication topologies, such as master-slave replication, multi-master replication, and synchronous replication, for MySQL and PostgreSQL databases. Knowledge of database partitioning techniques, such as range partitioning, hash partitioning, and list partitioning, for optimizing storage and query performance in large-scale database clusters. Familiarity with high-availability architectures, such as active-passive and active-active configurations, for ensuring continuous availability and reliability of MySQL and PostgreSQL databases. Familiarity with microservices architecture and related tools such as Istio, Envoy, or Linkerd. Knowledge of CI/CD pipelines and related tools such as Jenkins, GitLab CI/CD, or CircleCI. Experience with monitoring and observability tools such as Prometheus, Grafana, ELK stack (Elasticsearch, Logstash, Kibana), or Splunk. Familiarity with configuration management tools like Ansible, Puppet, or Chef. Proficiency in version control systems such as Git. Knowledge of scripting languages such as Bash, PowerShell, Ruby, or Python for automation tasks. Understanding of cloud-native security practices and tools such as Google Identity and Access Management (IAM), AWS Key Management Service (KMS), or Azure Active Directory. Familiarity with network security concepts such as VPNs, firewalls, and intrusion detection/prevention systems (IDS/IPS). What’s in it for you? You will have the autonomy to take ownership of your role and contribute to the growth and success of our brand. If you are driven to make an immediate impact, achieve results, thrive in a high performing team and want to grow in a dynamic and rewarding environment – You belong to Maropost! Show more Show less
Posted 3 months ago
0 years
0 Lacs
Pune/Pimpri-Chinchwad Area
On-site
Job Description We are seeking a highly skilled Senior Reliability Engineer with strong backend software engineering skills to join our team. As a Senior Reliability Engineer , you will be responsible for designing, implementing, and maintaining our cloud infrastructure, ensuring the smooth operation of our applications and services. In addition, you will contribute to the development of our backend software systems, working closely with our engineering team to design, develop, and deploy scalable and reliable software solutions. This role will report to Senior Engineering Manager, Finance Engineering in Pune, Indi What you’ll do: Collaborate with your peers to envision, design, and develop solutions in your respective area with a bias toward reusability, toil reduction, and resiliency Surface opportunities across the broader organization for solving systemic issues Use a collaborative approach to make technical decisions that align with Procore’s architectural vision Partner with internal customers, peers, and leadership in planning, prioritization, and roadmap development Develop teammates by conducting code reviews, providing mentorship, pairing, and training opportunities Serve as a subject matter expert on tools, processes, and procedures and help guide others to create and maintain a healthy codebase Facilitate an “open source” mindset and culture both across teams internally and outside of Procore through active participation in and contributions to the greater community Design, develop, and deploy scalable and reliable backend software systems using languages such as Java, Python, or Go Work with engineering teams to design and implement microservices architecture Develop and maintain APIs using RESTful APIs, GraphQL, or gRPC Ensure high-quality code through code reviews, testing, and continuous integration Serve as a subject matter expert in a domain, including processes and software design that help guide others to create and maintain a healthy codebase What we’re looking for: Container orchestration (Kubernetes) K8s, preferably EKS. ArgoCD Terraform or similar IaC o11y (OpenTelemetry ideal) Public cloud (AWS, GCP, Azure) Cloud automation tooling (e.g., CloudFormation, Terraform, Ansible) Kafka and Kafka connectors Linux Systems Ensure compliance with security and regulatory requirements, such as HIPAA, SOX, FedRAMP Experience with the following is preferred: Continuous Integration Tooling (e.g., Circle CI, Jenkins, Travis, etc.) Continuous Deployment Tooling (e.g., ArgoCD, Spinnaker) Service Mesh / Discovery Tooling (e.g., Consul, Envoy, Istio, Linkerd) Networking (WAF, Cloudflare) Event-driven architecture (Event Sourcing, CQRS) Flink or other streaming processing technologies RDBMS and NoSQL databases Experience in working and developing APIs through REST, gRPC, or GraphQL Professional experience in Java, GoLang, Python preferred Additional Information Perks & Benefits At Procore, we invest in our employees and provide a full range of benefits and perks to help you grow and thrive. From generous paid time off and healthcare coverage to career enrichment and development programs, learn more details about what we offer and how we empower you to be your best. About Us Procore Technologies is building the software that builds the world. We provide cloud-based construction management software that helps clients more efficiently build skyscrapers, hospitals, retail centers, airports, housing complexes, and more. At Procore, we have worked hard to create and maintain a culture where you can own your work and are encouraged and given resources to try new ideas. Check us out on Glassdoor to see what others are saying about working at Procore. We are an equal-opportunity employer and welcome builders of all backgrounds. We thrive in a diverse, dynamic, and inclusive environment. We do not tolerate discrimination against candidates or employees on the basis of gender, sex, national origin, civil status, family status, sexual orientation, religion, age, disability, race, traveler community, status as a protected veteran or any other classification protected by law. If you'd like to stay in touch and be the first to hear about new roles at Procore, join our Talent Community. Alternative methods of applying for employment are available to individuals unable to submit an application through this site because of a disability. Contact our benefits team here to discuss reasonable accommodations. Show more Show less
Posted 3 months ago
3.0 years
0 Lacs
Gurgaon, Haryana, India
On-site
You Lead the Way. We’ve Got Your Back. With the right backing, people and businesses have the power to progress in incredible ways. When you join Team Amex, you become part of a global and diverse community of colleagues with an unwavering commitment to back our customers, communities and each other. Here, you’ll learn and grow as we help you create a career journey that’s unique and meaningful to you with benefits, programs, and flexibility that support you personally and professionally. At American Express, you’ll be recognized for your contributions, leadership, and impact—every colleague has the opportunity to share in the company’s success. Together, we’ll win as a team, striving to uphold our company values and powerful backing promise to provide the world’s best customer experience every day. And we’ll do it with the utmost integrity, and in an environment where everyone is seen, heard and feels like they belong. Join Team Amex and let's lead the way together. About Enterprise Architecture: Enterprise Architecture is an organization within the Chief Technology Office at American Express and it is a key enabler of the company’s technology strategy. The four pillars of Enterprise Architecture include: 1. Architecture as Code : this pillar owns and operates foundational technologies that are leveraged by engineering teams across the enterprise. 2. Architecture as Design : this pillar includes the solution and technical design for transformation programs and business critical projects which need architectural guidance and support. 3. Governance : this pillar is responsible for defining technical standards, and developing innovative tools that automate controls to ensure compliance. 4. Colleague Enablement: this pillar is focused on colleague development, recognition, training, and enterprise outreach. Responsibilities: · Design, implement, and maintain API gateway solutions using tools like Apigee, Gloo, Envoy, or AWS API Gateway. · Configure and manage API traffic policies, routing, throttling, authentication, and authorization. · Collaborate with developers and architects to ensure effective API lifecycle management (design, testing, publishing, monitoring, and retirement). · Implement security protocols such as OAuth2, JWT, mTLS, and rate limiting. · Develop and enforce API governance policies, versioning standards, and best practices. · Monitor API performance, error rates, and latency, and provide insights for improvements. · Automate deployment and configuration using CI/CD pipelines tools. · Create and maintain documentation for API gateway configurations and processes. · Troubleshoot API gateway issues and provide support for developers and partners. Qualifications Preferably a BS or MS degree in computer science, computer engineering, or other technical discipline 3+ years of experience in API gateway technologies like Apigee, Gloo, Envoy, or similar. Strong understanding of RESTful API concepts, OpenAPI/Swagger specs. Proficiency in API security mechanisms (OAuth2, API Keys, JWT, mTLS). Experience with Kubernetes and service mesh technologies (Istio, Linkerd) is a strong plus. Familiarity with CI/CD tools (e.g., Jenkins, GitHub Actions, GitLab CI). Knowledge of monitoring tools (e.g., Prometheus, Grafana) for tracking API metrics. Strong scripting or programming skills (e.g., Python, Bash, Go, or Node.js). Excellent problem-solving and communication skills. Experience in Process Management, Case Management, Work Management is a plus. Experience in Automations, Cognitive OCR, AI / ML driving cost savings is a plus. Ability to effectively interpret technical and business objectives and challenges and articulate solutions Willingness to learn new technologies and exploit them to their optimal potential Extensive experience in designing and implementing large scale platforms with high resiliency, availability, and reliability. Strong experience in applications with high throughput and performance Experience with micro services architectures and service mesh technologies is preferred. Every member of our team must be able to demonstrate the following technical, functional, leadership and business core competencies, including: · Agile Practices · Porting/Software Configuration · Programming Languages and Frameworks – Hands on experience in some or all of the following is preferred: o Java, Python, Go, React, Envoy, gRPC, ProtoBuf, JSON, CouchBase, Cassandra, Redis, Consul, Jenkins, Docker, Kubernetes, OpenShift, Drools, Elastic Stack, Kafka, Spark · Analytical Thinking We back you with benefits that support your holistic well-being so you can be and deliver your best. This means caring for you and your loved ones' physical, financial, and mental health, as well as providing the flexibility you need to thrive personally and professionally: Competitive base salaries Bonus incentives Support for financial-well-being and retirement Comprehensive medical, dental, vision, life insurance, and disability benefits (depending on location) Flexible working model with hybrid, onsite or virtual arrangements depending on role and business need Generous paid parental leave policies (depending on your location) Free access to global on-site wellness centers staffed with nurses and doctors (depending on location) Free and confidential counseling support through our Healthy Minds program Career development and training opportunities American Express is an equal opportunity employer and makes employment decisions without regard to race, color, religion, sex, sexual orientation, gender identity, national origin, veteran status, disability status, age, or any other status protected by law. Offer of employment with American Express is conditioned upon the successful completion of a background verification check, subject to applicable laws and regulations. Show more Show less
Posted 3 months ago
5.0 years
0 Lacs
Hyderabad, Telangana, India
On-site
We Are: At Synopsys, we drive the innovations that shape the way we live and connect. Our technology is central to the Era of Pervasive Intelligence, from self-driving cars to learning machines. We lead in chip design, verification, and IP integration, empowering the creation of high-performance silicon chips and software content. Join us to transform the future through continuous technological innovation. You Are: You are a forward-thinking Cloud DevOps Engineer with a passion for modernizing infrastructure and enhancing the capabilities of CI/CD pipelines, containerization strategies, and hybrid cloud deployments. You thrive in environments where you can leverage your expertise in cloud infrastructure, distributed processing workloads, and AI-driven automation. Your collaborative spirit drives you to work closely with development, data, and GenAI teams to build resilient, scalable, and intelligent DevOps solutions. You are adept at integrating cutting-edge technologies and best practices to enhance both traditional and AI-driven workloads. Your proactive approach and problem-solving skills make you an invaluable asset to any team. What You’ll Be Doing: Designing, implementing, and optimizing CI/CD pipelines for cloud and hybrid environments. Integrating AI-driven pipeline automation for self-healing deployments and predictive troubleshooting. Leveraging GitOps (ArgoCD, Flux, Tekton) for declarative infrastructure management. Implementing progressive delivery strategies (Canary, Blue-Green, Feature Flags). Containerizing applications using Docker & Kubernetes (EKS, AKS, GKE, OpenShift, or on-prem clusters). Optimizing service orchestration and networking with service meshes (Istio, Linkerd, Consul). Implementing AI-enhanced observability for containerized services using AIOps-based monitoring. Automating provisioning with Terraform, CloudFormation, Pulumi, or CDK. Supporting and optimizing distributed computing workloads, including Apache Spark, Flink, or Ray. Using GenAI-driven copilots for DevOps automation, including scripting, deployment verification, and infra recommendations. The Impact You Will Have: Enhancing the efficiency and reliability of CI/CD pipelines and deployments. Driving the adoption of AI-driven automation to reduce downtime and improve system resilience. Enabling seamless application portability across on-prem and cloud environments. Implementing advanced observability solutions to proactively detect and resolve issues. Optimizing resource allocation and job scheduling for distributed processing workloads. Contributing to the development of intelligent DevOps solutions that support both traditional and AI-driven workloads. What You’ll Need: 5+ years of experience in DevOps, Cloud Engineering, or SRE. Hands-on expertise with CI/CD pipelines (Jenkins, GitHub Actions, GitLab CI, ArgoCD, Tekton, etc.). Strong experience with Kubernetes, container orchestration, and service meshes. Proficiency in Terraform, CloudFormation, Pulumi, or Infrastructure as Code (IaC) tools. Experience working in hybrid cloud environments (AWS, Azure, GCP, on-prem). Strong scripting skills in Python, Bash, or Go. Knowledge of distributed data processing frameworks (Spark, Flink, Ray, or similar). Who You Are: You are a collaborative and innovative professional with a strong technical background and a passion for continuous learning. You excel in problem-solving and thrive in dynamic environments where you can apply your expertise to drive significant improvements. Your excellent communication skills enable you to work effectively with diverse teams, and your commitment to excellence ensures that you consistently deliver high-quality results. The Team You’ll Be A Part Of: You will join a dynamic team focused on optimizing cloud infrastructure and enhancing workloads to contribute to overall operational efficiency. This team is dedicated to driving the modernization and optimization of Infrastructure CI/CD pipelines and hybrid cloud deployments, ensuring that Synopsys remains at the forefront of technological innovation. Rewards and Benefits: We offer a comprehensive range of health, wellness, and financial benefits to cater to your needs. Our total rewards include both monetary and non-monetary offerings. Your recruiter will provide more details about the salary range and benefits during the hiring process. Show more Show less
Posted 3 months ago
8.0 years
0 Lacs
Trivandrum, Kerala, India
On-site
Role Description Roles & Responsibilities GitHub Actions & CI/CD Workflows (Primary Focus) Design, develop, and maintain scalable CI/CD pipelines using GitHub Actions. Create reusable and modular workflow templates using composite actions and reusable workflows. Manage and optimize GitHub self-hosted runners, including autoscaling and hardening. Monitor and enhance CI/CD performance with caching, parallelism, and proper dependency management. Review and analyze existing Azure DevOps pipeline templates. Migrate Azure DevOps YAML pipelines to GitHub Actions, adapting tasks to equivalent GitHub workflows. Azure Kubernetes Service (AKS) Deploy and manage containerized workloads on AKS. Implement cluster and pod-level autoscaling, ensuring performance and cost-efficiency. Ensure high availability, security, and networking configurations for AKS clusters. Automate infrastructure provisioning using Terraform or other IaC tools. Azure DevOps Design and build scalable YAML-based Azure DevOps pipelines. Maintain and support Azure Pipelines for legacy or hybrid CI/CD environments. ArgoCD & GitOps Implement and manage GitOps workflows using ArgoCD. Configure and manage ArgoCD applications to sync AKS deployments from Git repositories. Enforce secure, auditable, and automated deployment strategies via GitOps. Collaboration & Best Practices Collaborate with developers and platform engineers to integrate DevOps best practices across teams. Document workflow standards, pipeline configurations, infrastructure setup, and runbooks. Promote observability, automation, and DevSecOps principles throughout the lifecycle. Must-Have Skills 8+ years of overall IT experience, with at least 5+ years in DevOps roles. 3+ years hands-on experience with GitHub Actions (including reusable workflows, composite actions, and self-hosted runners). 2+ years of experience with AKS, including autoscaling, networking, and security. Strong proficiency in CI/CD pipeline design and automation. Experience with ArgoCD and GitOps workflows. Hands-on with Terraform, ARM, or Bicep for IaC. Working knowledge of Azure DevOps pipelines and YAML configurations. Proficient in Docker, Bash, and at least one scripting language (Python preferred). Experience in managing secure and auditable deployments in enterprise environments. Good-to-Have Skills Exposure to monitoring and observability tools (e.g., Prometheus, Grafana, ELK stack). Familiarity with Service Meshes like Istio or Linkerd. Experience with Secrets Management (e.g., HashiCorp Vault, Azure Key Vault). Understanding of RBAC, OIDC, and SSO integrations in Kubernetes environments. Knowledge of Helm and custom chart development. Certifications in Azure, Kubernetes, or DevOps practices. Skills Github Actions & CI/CD,Azure Kubernetes Service,AgroCD & GitOps,Devops Show more Show less
Posted 3 months ago
7.0 years
0 Lacs
Thiruvananthapuram, Kerala, India
Remote
About The Company Armada is an edge computing startup that provides computing infrastructure to remote areas where connectivity and cloud infrastructure is limited, as well as areas where data needs to be processed locally for real-time analytics and AI at the edge. We’re looking to bring on the most brilliant minds to help further our mission of bridging the digital divide with advanced technology infrastructure that can be rapidly deployed anywhere . About The Role We are looking for a highly experienced, collaborative, and detail-oriented Senior Engineer to join our growing Edge team. You will be responsible for the design, automation, optimization, and operation of our Kubernetes-based platform supporting our Galleon mobile data centers and Commander cloud integration. This is a critical role where you will leverage deep technical expertise in cloud infrastructure and Kubernetes while valuing mentorship, collaboration, and open communication. You will work on building and managing resilient, secure, and scalable Kubernetes environments across diverse edge locations and cloud infrastructure, ensuring the reliability of our distributed computing platform. Location. This role is office-based at our Trivandrum, Kerala office. What You'll Do (Key Responsibilities) Architect, design, deploy, configure, and manage highly available Kubernetes clusters across edge (Galleon data centers) and cloud (AWS, Azure, GCP) environments. This includes designing the cluster layout, resource allocation, and storage configurations Administer, maintain, and monitor the health, performance, and capacity of Kubernetes clusters and underlying infrastructure Implement and manage Kubernetes networking solutions (CNI plugins, Ingress controllers) and storage solutions (PV/PVC, Storage Classes, CSI drivers) Maintain and monitor containerized platform services running within the clusters and robust monitoring, logging, and alerting systems (e.g., Prometheus, Grafana, ELK stack) Drive Infrastructure-as-Code (IaC) initiatives using tools like Terraform, Ansible, Helm, and potentially Kubernetes Operators, promoting automation, repeatability, and reliability Support and troubleshoot complex issues related to the Kubernetes platform, containerized services, networking, and infrastructure Implement and enforce Kubernetes security best practices (RBAC, Network Policies, Secrets Management, Security Contexts, Image Scanning) Automate cluster operations, deployment pipelines (CI/CD integration), and infrastructure provisioning using Infrastructure as Code (IaC) tools (e.g., Terraform, Ansible) Optimize Kubernetes clusters for performance, scalability, and resource utilization, particularly in edge environments Develop and maintain comprehensive documentation for cluster architecture, configurations, operational procedures, and runbooks Work in collaboration with software engineering, DevOps, security teams, and product managers to ensure seamless integration, deployment, and secure operation of applications on Kubernetes Evaluate and integrate new technologies from the Kubernetes ecosystem Contribute to the operational excellence of the platform, including participating in on-call rotations, incident management, and building self-healing capabilities Required Qualifications Bachelor's degree in computer science, Engineering, Information Technology, a related technical field, or equivalent practical experience At least 7+ years of professional experience in infrastructure engineering, systems administration, or software development, with a strong focus (4+ years preferred) on building and maintaining production Kubernetes environments At least 3+ years of professional experience using and administering Linux operating systems Deep understanding of Kubernetes architecture, core components, operational best practices, and lifecycle management Strong experience with containerization technologies (Docker) Hands-on experience managing Kubernetes on at least one major cloud provider (AWS, Azure, GCP) Strong understanding and proven experience with Infrastructure as Code (IaC) solutions, particularly Terraform and/or Ansible Proficiency in scripting languages (e.g., Python, Bash) for automation Experience configuring and managing monitoring/logging tools (e.g., Prometheus, Grafana, ELK Stack) Solid understanding of Linux operating system, networking fundamentals (TCP/IP, DNS, Load Balancing, Firewalls, VPNs) and container networking (CNI) Strong understanding of Kubernetes security concepts and implementation (RBAC, Network Policies, Secrets) Ability to work independently and collaborate effectively with others to debug and solve problems Preferred Experience And Skills Experience with Red Hat OpenShift Container Platform (version 4+ is a plus) Experience deploying and maintaining CI/CD solutions for DevSecOps, such as GitLab CI or Jenkins Strong development experience using Docker, docker-compose, and/or Kubernetes Experience developing Ansible playbooks for process automation Kubernetes certifications (CKA, CKS) Experience with Kubernetes operators and Custom Resource Definitions (CRDs) Experience with service mesh technologies like Istio or Linkerd Experience managing Kubernetes in edge computing or resource-constrained environments Compensation & Benefits For India-based candidates: We offer a competitive base salary along with equity options, providing an opportunity to share in the success and growth of Armada. You're a Great Fit if You're A go-getter with a growth mindset. You're intellectually curious, have strong business acumen, and actively seek opportunities to build relevant skills and knowledge A detail-oriented problem-solver. You can independently gather information, solve problems efficiently, and deliver results with a "get-it-done" attitude Thrive in a fast-paced environment. You're energized by an entrepreneurial spirit, capable of working quickly, and excited to contribute to a growing company A collaborative team player. You focus on business success and are motivated by team accomplishment vs personal agenda Highly organized and results-driven. Strong prioritization skills and a dedicated work ethic are essential for you Equal Opportunity Statement At Armada, we are committed to fostering a work environment where everyone is given equal opportunities to thrive. As an equal opportunity employer, we strictly prohibit discrimination or harassment based on race, color, gender, religion, sexual orientation, national origin, disability, genetic information, pregnancy, or any other characteristic protected by law. This policy applies to all employment decisions, including hiring, promotions, and compensation. Our hiring is guided by qualifications, merit, and the business needs at the time. Show more Show less
Posted 3 months ago
6.0 years
0 Lacs
Hyderabad, Telangana, India
On-site
Position Summary: As "Infra Admin DevOps", you should be a professional skilled in both infrastructure maintenance and DevOps practices, which includes automation, CI/CD, and cloud technologies. Should focus on ensuring a reliable and scalable infrastructure while also streamlining software development and deployment processes. Key Responsibilities and Skills: Infrastructure Management: Building, maintaining, and managing the physical and virtual infrastructure (servers, networks, storage). DevOps Practices: Implementing automation tools, scripting, and DevOps methodologies to streamline software development and deployment. Cloud Computing: Working with cloud platforms (AWS, Azure, GCP) for provisioning, scaling, and managing resources. CI/CD: Setting up and managing Continuous Integration and Continuous Deployment pipelines. Automation: Automating tasks like server provisioning, configuration management, and deployment. Infrastructure as Code (IaC): Using tools to define and manage infrastructure as code. Scripting and Programming: Proficiency in scripting languages like Python, Bash, or Groovy. Monitoring and Logging: Setting up and maintaining infrastructure monitoring and logging systems. Troubleshooting: Identifying and resolving issues within the infrastructure and software deployment process. Collaboration: Working closely with developers, testers, and other DevOps team members. Design, implement, and maintain containerized applications using Kubernetes, ensuring scalability, high availability, and fault tolerance. Develop, deploy, and manage cloud infrastructure on AWS and Azure using infrastructure-as-code tools such as Terraform or CloudFormation, and configuration management tools like Ansible or Chef. Implement and manage CI/CD pipelines using tools such as Jenkins, GitLab CI, or CircleCI, ensuring automated build, test, and deployment processes. Monitor and analyze application and system performance, identifying potential bottlenecks and implementing solutions to optimize resource utilization and system efficiency. Collaborate with cross-functional teams to identify, troubleshoot, and resolve production issues, ensuring minimal downtime and optimal application performance. Develop and maintain documentation for system architecture, design, and operational procedures. Continuously evaluate and adopt new tools, technologies, and best practices to improve the efficiency and reliability of our infrastructure and development process. Participate in on-call rotation to ensure 24/7 system availability and provide support for production incidents. In essence, an Infra Admin DevOps professional ensures that the underlying infrastructure is robust and scalable while also enabling the rapid and reliable delivery of software, making them a vital part of any modern software development team. Skills & Education: Bachelor's degree in computer science, Engineering, or a related field. 6+ years of experience in a DevOps or similar role, with a strong focus on Kubernetes, AWS, and Azure. Strong knowledge of scripting languages such as Python, Bash, or PowerShell. Proficiency in implementing and managing CI/CD pipelines using tools such as Jenkins, GitLab CI, or CircleCI. Experience with infrastructure-as-code (IAC) and configuration management tools, such as Terraform, CloudFormation, Ansible, or Chef. Strong understanding of networking, security, and monitoring concepts in cloud environments. Excellent problem-solving, analytical, and communication skills. Ability to work effectively in a fast-paced, collaborative environment. Desirable: Relevant certifications, such as AWS Certified DevOps Engineer, Azure DevOps Engineer Expert, or Certified Kubernetes Administrator (CKA). Experience with other cloud platforms, such as Google Cloud Platform (GCP) or IBM Cloud. Hands-on experience with infrastructure-as-code (IAC) and configuration management tools, such as Terraform, CloudFormation, Ansible, or Chef. Working knowledge with container security and best practices. Experience with service mesh technologies, such as Istio or Linkerd. Requirements Bachelor's degree in computer science, Engineering, or a related field. 6+ years of experience in a DevOps or similar role, with a strong focus on Kubernetes, AWS, and Azure. Strong knowledge of scripting languages such as Python, Bash, or PowerShell. Proficiency in implementing and managing CI/CD pipelines using tools such as Jenkins, GitLab CI, or CircleCI. Experience with infrastructure-as-code (IAC) and configuration management tools, such as Terraform, CloudFormation, Ansible, or Chef. Strong understanding of networking, security, and monitoring concepts in cloud environments. Excellent problem-solving, analytical, and communication skills. Ability to work effectively in a fast-paced, collaborative environment. Show more Show less
Posted 3 months ago
0 years
0 Lacs
Surat, Gujarat, India
Remote
Experience : 4.00 + years Salary : USD 2962 / month (based on experience) Expected Notice Period : 15 Days Shift : (GMT+05:30) Asia/Kolkata (IST) Opportunity Type : Remote Placement Type : Full Time Contract for 3 Months(40 hrs a week/160 hrs a month) (*Note: This is a requirement for one of Uplers' client - Hybrid AI) What do you need for this opportunity? Must have skills required: air-gapped Kubernetes, compliance enforcement, Hybrid cloud, Service Mesh, Zero Trust, Datacentre/ bare metal infrastructure, GPU / CPU optimization, Large Language Models (LLMs), OpenShift, VMware, Kubernetes Hybrid AI is Looking for: Our platform, designed by esteemed professors in AI/ML, industry CIOs, and brought to market by experienced software industry executives, addresses the critical challenges CIOs face in navigating the rapidly evolving AI & IT landscape. Join us at HybridAI and be part of a team that is shaping the future of AI infrastructure with innovation, expertise, and a commitment to excellence. Location : Remote Experience : 4-5 years Type : 3 months contract Role Overview: We are looking for a hands-on Infrastructure Engineer with deep expertise in VMware and OpenShift Kubernetes, combined with a proven track record of GPU and CPU optimization for high-performance workloads. The ideal candidate should have experience in deploying large language models (LLMs) on GPU-accelerated infrastructure, managing GPU allocation and tuning, and implementing Operators in OpenShift environments. The candidate should have a strong understanding of infrastructure security, with practical knowledge of ISO 27001 compliance, and a passion for working in fast-paced startup environments. The right candidate is both a builder and optimizer—comfortable getting deep into systems while aligning performance, security, and compliance goals.This role will report into the Direactor of AI and Software Engineering. Key Responsibilities: Create and manage development (Dev), UAT, and production (Prod) environments on bare metal and Red Hat Linux-based servers. Harden Linux servers for security compliance, ensuring systems pass VAPT (Vulnerability Assessment & Penetration Testing). Develop CI/CD pipelines from GitHub to Linux-based VMs running OpenShift Kubernetes clusters. Ensure high availability, observability, and proactive alerting for the HybridAI SaaS platform. Automate deployment of the HybridAI InfraMetrics Collector in customer on- prem environments. Work with VMware vCenter and Kubernetes Cluster APIs to manage infrastructure resources and automate deployments and provide guidance on VM Optimizations. Enable build cycles with expertise on virtualization, container orchestration, and hybrid infrastructure. Deploy LLMs on GPU infrastructure ensuring optimal resource allocation and scaling for AI-driven applications. Monitor infrastructure performance and implement proactive scaling solutions. Collaborate with Head of Software Engineering to enforce API security, access control, and compliance policies. Implement secure and compliant infrastructure aligned with ISO 27001 standards. Experience: 4+ years of hands-on DevOps and infrastructure engineering experience managing enterprise-grade datacenter environments. Strong experience with Red Hat Linux and bare metal infrastructure management. Expertise in Linux security hardening (firewall configuration, SELinux, system patching). Deep knowledge of OpenShift Kubernetes (OCP) and container orchestration. Hands-on experience in CPU/GPU profiling, resource allocation, and performance tuning Experience with infrastructure as code (Terraform, Ansible) Proficiency in CI/CD pipelines (GitHub Actions, GitLab CI, Jenkins, ArgoCD) for OpenShift & Linux-based deployments. Hands-on experience with VMware stack (ESXi, vCenter, vMotion) Cloud and on-prem experience, with exposure to AWS, GCP, Azure, and private cloud platforms. Scripting and automation expertise (Bash, Python, Powershell). Strong security background, including API security, authentication (OAuth, JWT, mTLS), and compliance with CIS benchmarks. Experience with any of observability and monitoring tools, including: NVIDIA DCGM, Prometheus & Grafana, ELK Stack, DataDog, Splunk, or AppDynamics Solid experience in ISO 27001 compliance, security best practices, and policy implementation Comfortable working in agile, very fast-paced startup environments with ownership of infra outcomes Nice-to-Have Skills Experience with service mesh architectures (Istio, Linkerd). Familiarity with Zero Trust security models. Exposure to air-gapped Kubernetes deployments for security-sensitive environments. Experience with automated compliance enforcement tools (OpenSCAP, Falco, Aqua Security). Knowledge of hybrid cloud networking (VPCs, VPNs, private links between on-prem and cloud). Hands-on experience with HashiCorp Vault for secrets management. Exposure to additional compliance frameworks such as SOC 2 or NIST Experience with AI/ML or HPC workloads beyond LLM applications How to apply for this opportunity? Step 1: Click On Apply! And Register or Login on our portal. Step 2: Complete the Screening Form & Upload updated Resume Step 3: Increase your chances to get shortlisted & meet the client for the Interview! About Uplers: Our goal is to make hiring reliable, simple, and fast. Our role will be to help all our talents find and apply for relevant contractual onsite opportunities and progress in their career. We will support any grievances or challenges you may face during the engagement. (Note: There are many more opportunities apart from this on the portal. Depending on the assessments you clear, you can apply for them as well). So, if you are ready for a new challenge, a great work environment, and an opportunity to take your career to the next level, don't hesitate to apply today. We are waiting for you! Show more Show less
Posted 3 months ago
0 years
0 Lacs
Ahmedabad, Gujarat, India
Remote
Experience : 4.00 + years Salary : USD 2962 / month (based on experience) Expected Notice Period : 15 Days Shift : (GMT+05:30) Asia/Kolkata (IST) Opportunity Type : Remote Placement Type : Full Time Contract for 3 Months(40 hrs a week/160 hrs a month) (*Note: This is a requirement for one of Uplers' client - Hybrid AI) What do you need for this opportunity? Must have skills required: air-gapped Kubernetes, compliance enforcement, Hybrid cloud, Service Mesh, Zero Trust, Datacentre/ bare metal infrastructure, GPU / CPU optimization, Large Language Models (LLMs), OpenShift, VMware, Kubernetes Hybrid AI is Looking for: Our platform, designed by esteemed professors in AI/ML, industry CIOs, and brought to market by experienced software industry executives, addresses the critical challenges CIOs face in navigating the rapidly evolving AI & IT landscape. Join us at HybridAI and be part of a team that is shaping the future of AI infrastructure with innovation, expertise, and a commitment to excellence. Location : Remote Experience : 4-5 years Type : 3 months contract Role Overview: We are looking for a hands-on Infrastructure Engineer with deep expertise in VMware and OpenShift Kubernetes, combined with a proven track record of GPU and CPU optimization for high-performance workloads. The ideal candidate should have experience in deploying large language models (LLMs) on GPU-accelerated infrastructure, managing GPU allocation and tuning, and implementing Operators in OpenShift environments. The candidate should have a strong understanding of infrastructure security, with practical knowledge of ISO 27001 compliance, and a passion for working in fast-paced startup environments. The right candidate is both a builder and optimizer—comfortable getting deep into systems while aligning performance, security, and compliance goals.This role will report into the Direactor of AI and Software Engineering. Key Responsibilities: Create and manage development (Dev), UAT, and production (Prod) environments on bare metal and Red Hat Linux-based servers. Harden Linux servers for security compliance, ensuring systems pass VAPT (Vulnerability Assessment & Penetration Testing). Develop CI/CD pipelines from GitHub to Linux-based VMs running OpenShift Kubernetes clusters. Ensure high availability, observability, and proactive alerting for the HybridAI SaaS platform. Automate deployment of the HybridAI InfraMetrics Collector in customer on- prem environments. Work with VMware vCenter and Kubernetes Cluster APIs to manage infrastructure resources and automate deployments and provide guidance on VM Optimizations. Enable build cycles with expertise on virtualization, container orchestration, and hybrid infrastructure. Deploy LLMs on GPU infrastructure ensuring optimal resource allocation and scaling for AI-driven applications. Monitor infrastructure performance and implement proactive scaling solutions. Collaborate with Head of Software Engineering to enforce API security, access control, and compliance policies. Implement secure and compliant infrastructure aligned with ISO 27001 standards. Experience: 4+ years of hands-on DevOps and infrastructure engineering experience managing enterprise-grade datacenter environments. Strong experience with Red Hat Linux and bare metal infrastructure management. Expertise in Linux security hardening (firewall configuration, SELinux, system patching). Deep knowledge of OpenShift Kubernetes (OCP) and container orchestration. Hands-on experience in CPU/GPU profiling, resource allocation, and performance tuning Experience with infrastructure as code (Terraform, Ansible) Proficiency in CI/CD pipelines (GitHub Actions, GitLab CI, Jenkins, ArgoCD) for OpenShift & Linux-based deployments. Hands-on experience with VMware stack (ESXi, vCenter, vMotion) Cloud and on-prem experience, with exposure to AWS, GCP, Azure, and private cloud platforms. Scripting and automation expertise (Bash, Python, Powershell). Strong security background, including API security, authentication (OAuth, JWT, mTLS), and compliance with CIS benchmarks. Experience with any of observability and monitoring tools, including: NVIDIA DCGM, Prometheus & Grafana, ELK Stack, DataDog, Splunk, or AppDynamics Solid experience in ISO 27001 compliance, security best practices, and policy implementation Comfortable working in agile, very fast-paced startup environments with ownership of infra outcomes Nice-to-Have Skills Experience with service mesh architectures (Istio, Linkerd). Familiarity with Zero Trust security models. Exposure to air-gapped Kubernetes deployments for security-sensitive environments. Experience with automated compliance enforcement tools (OpenSCAP, Falco, Aqua Security). Knowledge of hybrid cloud networking (VPCs, VPNs, private links between on-prem and cloud). Hands-on experience with HashiCorp Vault for secrets management. Exposure to additional compliance frameworks such as SOC 2 or NIST Experience with AI/ML or HPC workloads beyond LLM applications How to apply for this opportunity? Step 1: Click On Apply! And Register or Login on our portal. Step 2: Complete the Screening Form & Upload updated Resume Step 3: Increase your chances to get shortlisted & meet the client for the Interview! About Uplers: Our goal is to make hiring reliable, simple, and fast. Our role will be to help all our talents find and apply for relevant contractual onsite opportunities and progress in their career. We will support any grievances or challenges you may face during the engagement. (Note: There are many more opportunities apart from this on the portal. Depending on the assessments you clear, you can apply for them as well). So, if you are ready for a new challenge, a great work environment, and an opportunity to take your career to the next level, don't hesitate to apply today. We are waiting for you! Show more Show less
Posted 3 months ago
0 years
0 Lacs
Greater Lucknow Area
Remote
Experience : 4.00 + years Salary : USD 2962 / month (based on experience) Expected Notice Period : 15 Days Shift : (GMT+05:30) Asia/Kolkata (IST) Opportunity Type : Remote Placement Type : Full Time Contract for 3 Months(40 hrs a week/160 hrs a month) (*Note: This is a requirement for one of Uplers' client - Hybrid AI) What do you need for this opportunity? Must have skills required: air-gapped Kubernetes, compliance enforcement, Hybrid cloud, Service Mesh, Zero Trust, Datacentre/ bare metal infrastructure, GPU / CPU optimization, Large Language Models (LLMs), OpenShift, VMware, Kubernetes Hybrid AI is Looking for: Our platform, designed by esteemed professors in AI/ML, industry CIOs, and brought to market by experienced software industry executives, addresses the critical challenges CIOs face in navigating the rapidly evolving AI & IT landscape. Join us at HybridAI and be part of a team that is shaping the future of AI infrastructure with innovation, expertise, and a commitment to excellence. Location : Remote Experience : 4-5 years Type : 3 months contract Role Overview: We are looking for a hands-on Infrastructure Engineer with deep expertise in VMware and OpenShift Kubernetes, combined with a proven track record of GPU and CPU optimization for high-performance workloads. The ideal candidate should have experience in deploying large language models (LLMs) on GPU-accelerated infrastructure, managing GPU allocation and tuning, and implementing Operators in OpenShift environments. The candidate should have a strong understanding of infrastructure security, with practical knowledge of ISO 27001 compliance, and a passion for working in fast-paced startup environments. The right candidate is both a builder and optimizer—comfortable getting deep into systems while aligning performance, security, and compliance goals.This role will report into the Direactor of AI and Software Engineering. Key Responsibilities: Create and manage development (Dev), UAT, and production (Prod) environments on bare metal and Red Hat Linux-based servers. Harden Linux servers for security compliance, ensuring systems pass VAPT (Vulnerability Assessment & Penetration Testing). Develop CI/CD pipelines from GitHub to Linux-based VMs running OpenShift Kubernetes clusters. Ensure high availability, observability, and proactive alerting for the HybridAI SaaS platform. Automate deployment of the HybridAI InfraMetrics Collector in customer on- prem environments. Work with VMware vCenter and Kubernetes Cluster APIs to manage infrastructure resources and automate deployments and provide guidance on VM Optimizations. Enable build cycles with expertise on virtualization, container orchestration, and hybrid infrastructure. Deploy LLMs on GPU infrastructure ensuring optimal resource allocation and scaling for AI-driven applications. Monitor infrastructure performance and implement proactive scaling solutions. Collaborate with Head of Software Engineering to enforce API security, access control, and compliance policies. Implement secure and compliant infrastructure aligned with ISO 27001 standards. Experience: 4+ years of hands-on DevOps and infrastructure engineering experience managing enterprise-grade datacenter environments. Strong experience with Red Hat Linux and bare metal infrastructure management. Expertise in Linux security hardening (firewall configuration, SELinux, system patching). Deep knowledge of OpenShift Kubernetes (OCP) and container orchestration. Hands-on experience in CPU/GPU profiling, resource allocation, and performance tuning Experience with infrastructure as code (Terraform, Ansible) Proficiency in CI/CD pipelines (GitHub Actions, GitLab CI, Jenkins, ArgoCD) for OpenShift & Linux-based deployments. Hands-on experience with VMware stack (ESXi, vCenter, vMotion) Cloud and on-prem experience, with exposure to AWS, GCP, Azure, and private cloud platforms. Scripting and automation expertise (Bash, Python, Powershell). Strong security background, including API security, authentication (OAuth, JWT, mTLS), and compliance with CIS benchmarks. Experience with any of observability and monitoring tools, including: NVIDIA DCGM, Prometheus & Grafana, ELK Stack, DataDog, Splunk, or AppDynamics Solid experience in ISO 27001 compliance, security best practices, and policy implementation Comfortable working in agile, very fast-paced startup environments with ownership of infra outcomes Nice-to-Have Skills Experience with service mesh architectures (Istio, Linkerd). Familiarity with Zero Trust security models. Exposure to air-gapped Kubernetes deployments for security-sensitive environments. Experience with automated compliance enforcement tools (OpenSCAP, Falco, Aqua Security). Knowledge of hybrid cloud networking (VPCs, VPNs, private links between on-prem and cloud). Hands-on experience with HashiCorp Vault for secrets management. Exposure to additional compliance frameworks such as SOC 2 or NIST Experience with AI/ML or HPC workloads beyond LLM applications How to apply for this opportunity? Step 1: Click On Apply! And Register or Login on our portal. Step 2: Complete the Screening Form & Upload updated Resume Step 3: Increase your chances to get shortlisted & meet the client for the Interview! About Uplers: Our goal is to make hiring reliable, simple, and fast. Our role will be to help all our talents find and apply for relevant contractual onsite opportunities and progress in their career. We will support any grievances or challenges you may face during the engagement. (Note: There are many more opportunities apart from this on the portal. Depending on the assessments you clear, you can apply for them as well). So, if you are ready for a new challenge, a great work environment, and an opportunity to take your career to the next level, don't hesitate to apply today. We are waiting for you! Show more Show less
Posted 3 months ago
0 years
0 Lacs
Nagpur, Maharashtra, India
Remote
Experience : 4.00 + years Salary : USD 2962 / month (based on experience) Expected Notice Period : 15 Days Shift : (GMT+05:30) Asia/Kolkata (IST) Opportunity Type : Remote Placement Type : Full Time Contract for 3 Months(40 hrs a week/160 hrs a month) (*Note: This is a requirement for one of Uplers' client - Hybrid AI) What do you need for this opportunity? Must have skills required: air-gapped Kubernetes, compliance enforcement, Hybrid cloud, Service Mesh, Zero Trust, Datacentre/ bare metal infrastructure, GPU / CPU optimization, Large Language Models (LLMs), OpenShift, VMware, Kubernetes Hybrid AI is Looking for: Our platform, designed by esteemed professors in AI/ML, industry CIOs, and brought to market by experienced software industry executives, addresses the critical challenges CIOs face in navigating the rapidly evolving AI & IT landscape. Join us at HybridAI and be part of a team that is shaping the future of AI infrastructure with innovation, expertise, and a commitment to excellence. Location : Remote Experience : 4-5 years Type : 3 months contract Role Overview: We are looking for a hands-on Infrastructure Engineer with deep expertise in VMware and OpenShift Kubernetes, combined with a proven track record of GPU and CPU optimization for high-performance workloads. The ideal candidate should have experience in deploying large language models (LLMs) on GPU-accelerated infrastructure, managing GPU allocation and tuning, and implementing Operators in OpenShift environments. The candidate should have a strong understanding of infrastructure security, with practical knowledge of ISO 27001 compliance, and a passion for working in fast-paced startup environments. The right candidate is both a builder and optimizer—comfortable getting deep into systems while aligning performance, security, and compliance goals.This role will report into the Direactor of AI and Software Engineering. Key Responsibilities: Create and manage development (Dev), UAT, and production (Prod) environments on bare metal and Red Hat Linux-based servers. Harden Linux servers for security compliance, ensuring systems pass VAPT (Vulnerability Assessment & Penetration Testing). Develop CI/CD pipelines from GitHub to Linux-based VMs running OpenShift Kubernetes clusters. Ensure high availability, observability, and proactive alerting for the HybridAI SaaS platform. Automate deployment of the HybridAI InfraMetrics Collector in customer on- prem environments. Work with VMware vCenter and Kubernetes Cluster APIs to manage infrastructure resources and automate deployments and provide guidance on VM Optimizations. Enable build cycles with expertise on virtualization, container orchestration, and hybrid infrastructure. Deploy LLMs on GPU infrastructure ensuring optimal resource allocation and scaling for AI-driven applications. Monitor infrastructure performance and implement proactive scaling solutions. Collaborate with Head of Software Engineering to enforce API security, access control, and compliance policies. Implement secure and compliant infrastructure aligned with ISO 27001 standards. Experience: 4+ years of hands-on DevOps and infrastructure engineering experience managing enterprise-grade datacenter environments. Strong experience with Red Hat Linux and bare metal infrastructure management. Expertise in Linux security hardening (firewall configuration, SELinux, system patching). Deep knowledge of OpenShift Kubernetes (OCP) and container orchestration. Hands-on experience in CPU/GPU profiling, resource allocation, and performance tuning Experience with infrastructure as code (Terraform, Ansible) Proficiency in CI/CD pipelines (GitHub Actions, GitLab CI, Jenkins, ArgoCD) for OpenShift & Linux-based deployments. Hands-on experience with VMware stack (ESXi, vCenter, vMotion) Cloud and on-prem experience, with exposure to AWS, GCP, Azure, and private cloud platforms. Scripting and automation expertise (Bash, Python, Powershell). Strong security background, including API security, authentication (OAuth, JWT, mTLS), and compliance with CIS benchmarks. Experience with any of observability and monitoring tools, including: NVIDIA DCGM, Prometheus & Grafana, ELK Stack, DataDog, Splunk, or AppDynamics Solid experience in ISO 27001 compliance, security best practices, and policy implementation Comfortable working in agile, very fast-paced startup environments with ownership of infra outcomes Nice-to-Have Skills Experience with service mesh architectures (Istio, Linkerd). Familiarity with Zero Trust security models. Exposure to air-gapped Kubernetes deployments for security-sensitive environments. Experience with automated compliance enforcement tools (OpenSCAP, Falco, Aqua Security). Knowledge of hybrid cloud networking (VPCs, VPNs, private links between on-prem and cloud). Hands-on experience with HashiCorp Vault for secrets management. Exposure to additional compliance frameworks such as SOC 2 or NIST Experience with AI/ML or HPC workloads beyond LLM applications How to apply for this opportunity? Step 1: Click On Apply! And Register or Login on our portal. Step 2: Complete the Screening Form & Upload updated Resume Step 3: Increase your chances to get shortlisted & meet the client for the Interview! About Uplers: Our goal is to make hiring reliable, simple, and fast. Our role will be to help all our talents find and apply for relevant contractual onsite opportunities and progress in their career. We will support any grievances or challenges you may face during the engagement. (Note: There are many more opportunities apart from this on the portal. Depending on the assessments you clear, you can apply for them as well). So, if you are ready for a new challenge, a great work environment, and an opportunity to take your career to the next level, don't hesitate to apply today. We are waiting for you! Show more Show less
Posted 3 months ago
0 years
0 Lacs
Nashik, Maharashtra, India
Remote
Experience : 4.00 + years Salary : USD 2962 / month (based on experience) Expected Notice Period : 15 Days Shift : (GMT+05:30) Asia/Kolkata (IST) Opportunity Type : Remote Placement Type : Full Time Contract for 3 Months(40 hrs a week/160 hrs a month) (*Note: This is a requirement for one of Uplers' client - Hybrid AI) What do you need for this opportunity? Must have skills required: air-gapped Kubernetes, compliance enforcement, Hybrid cloud, Service Mesh, Zero Trust, Datacentre/ bare metal infrastructure, GPU / CPU optimization, Large Language Models (LLMs), OpenShift, VMware, Kubernetes Hybrid AI is Looking for: Our platform, designed by esteemed professors in AI/ML, industry CIOs, and brought to market by experienced software industry executives, addresses the critical challenges CIOs face in navigating the rapidly evolving AI & IT landscape. Join us at HybridAI and be part of a team that is shaping the future of AI infrastructure with innovation, expertise, and a commitment to excellence. Location : Remote Experience : 4-5 years Type : 3 months contract Role Overview: We are looking for a hands-on Infrastructure Engineer with deep expertise in VMware and OpenShift Kubernetes, combined with a proven track record of GPU and CPU optimization for high-performance workloads. The ideal candidate should have experience in deploying large language models (LLMs) on GPU-accelerated infrastructure, managing GPU allocation and tuning, and implementing Operators in OpenShift environments. The candidate should have a strong understanding of infrastructure security, with practical knowledge of ISO 27001 compliance, and a passion for working in fast-paced startup environments. The right candidate is both a builder and optimizer—comfortable getting deep into systems while aligning performance, security, and compliance goals.This role will report into the Direactor of AI and Software Engineering. Key Responsibilities: Create and manage development (Dev), UAT, and production (Prod) environments on bare metal and Red Hat Linux-based servers. Harden Linux servers for security compliance, ensuring systems pass VAPT (Vulnerability Assessment & Penetration Testing). Develop CI/CD pipelines from GitHub to Linux-based VMs running OpenShift Kubernetes clusters. Ensure high availability, observability, and proactive alerting for the HybridAI SaaS platform. Automate deployment of the HybridAI InfraMetrics Collector in customer on- prem environments. Work with VMware vCenter and Kubernetes Cluster APIs to manage infrastructure resources and automate deployments and provide guidance on VM Optimizations. Enable build cycles with expertise on virtualization, container orchestration, and hybrid infrastructure. Deploy LLMs on GPU infrastructure ensuring optimal resource allocation and scaling for AI-driven applications. Monitor infrastructure performance and implement proactive scaling solutions. Collaborate with Head of Software Engineering to enforce API security, access control, and compliance policies. Implement secure and compliant infrastructure aligned with ISO 27001 standards. Experience: 4+ years of hands-on DevOps and infrastructure engineering experience managing enterprise-grade datacenter environments. Strong experience with Red Hat Linux and bare metal infrastructure management. Expertise in Linux security hardening (firewall configuration, SELinux, system patching). Deep knowledge of OpenShift Kubernetes (OCP) and container orchestration. Hands-on experience in CPU/GPU profiling, resource allocation, and performance tuning Experience with infrastructure as code (Terraform, Ansible) Proficiency in CI/CD pipelines (GitHub Actions, GitLab CI, Jenkins, ArgoCD) for OpenShift & Linux-based deployments. Hands-on experience with VMware stack (ESXi, vCenter, vMotion) Cloud and on-prem experience, with exposure to AWS, GCP, Azure, and private cloud platforms. Scripting and automation expertise (Bash, Python, Powershell). Strong security background, including API security, authentication (OAuth, JWT, mTLS), and compliance with CIS benchmarks. Experience with any of observability and monitoring tools, including: NVIDIA DCGM, Prometheus & Grafana, ELK Stack, DataDog, Splunk, or AppDynamics Solid experience in ISO 27001 compliance, security best practices, and policy implementation Comfortable working in agile, very fast-paced startup environments with ownership of infra outcomes Nice-to-Have Skills Experience with service mesh architectures (Istio, Linkerd). Familiarity with Zero Trust security models. Exposure to air-gapped Kubernetes deployments for security-sensitive environments. Experience with automated compliance enforcement tools (OpenSCAP, Falco, Aqua Security). Knowledge of hybrid cloud networking (VPCs, VPNs, private links between on-prem and cloud). Hands-on experience with HashiCorp Vault for secrets management. Exposure to additional compliance frameworks such as SOC 2 or NIST Experience with AI/ML or HPC workloads beyond LLM applications How to apply for this opportunity? Step 1: Click On Apply! And Register or Login on our portal. Step 2: Complete the Screening Form & Upload updated Resume Step 3: Increase your chances to get shortlisted & meet the client for the Interview! About Uplers: Our goal is to make hiring reliable, simple, and fast. Our role will be to help all our talents find and apply for relevant contractual onsite opportunities and progress in their career. We will support any grievances or challenges you may face during the engagement. (Note: There are many more opportunities apart from this on the portal. Depending on the assessments you clear, you can apply for them as well). So, if you are ready for a new challenge, a great work environment, and an opportunity to take your career to the next level, don't hesitate to apply today. We are waiting for you! Show more Show less
Posted 3 months ago
7.0 years
0 Lacs
Thiruvananthapuram, Kerala
Remote
Thiruvananthapuram Office, AEDGE AICC India Pvt Ltd About the Company Armada is an edge computing startup that provides computing infrastructure to remote areas where connectivity and cloud infrastructure is limited, as well as areas where data needs to be processed locally for real-time analytics and AI at the edge. We’re looking to bring on the most brilliant minds to help further our mission of bridging the digital divide with advanced technology infrastructure that can be rapidly deployed anywhere . About the role We are looking for a highly experienced, collaborative, and detail-oriented Senior Engineer to join our growing Edge team. You will be responsible for the design, automation, optimization, and operation of our Kubernetes-based platform supporting our Galleon mobile data centers and Commander cloud integration. This is a critical role where you will leverage deep technical expertise in cloud infrastructure and Kubernetes while valuing mentorship, collaboration, and open communication. You will work on building and managing resilient, secure, and scalable Kubernetes environments across diverse edge locations and cloud infrastructure, ensuring the reliability of our distributed computing platform. Location. This role is office-based at our Trivandrum, Kerala office. What You'll Do (Key Responsibilities) Architect, design, deploy, configure, and manage highly available Kubernetes clusters across edge (Galleon data centers) and cloud (AWS, Azure, GCP) environments. This includes designing the cluster layout, resource allocation, and storage configurations Administer, maintain, and monitor the health, performance, and capacity of Kubernetes clusters and underlying infrastructure Implement and manage Kubernetes networking solutions (CNI plugins, Ingress controllers) and storage solutions (PV/PVC, Storage Classes, CSI drivers) Maintain and monitor containerized platform services running within the clusters and robust monitoring, logging, and alerting systems (e.g., Prometheus, Grafana, ELK stack) Drive Infrastructure-as-Code (IaC) initiatives using tools like Terraform, Ansible, Helm, and potentially Kubernetes Operators, promoting automation, repeatability, and reliability Support and troubleshoot complex issues related to the Kubernetes platform, containerized services, networking, and infrastructure Implement and enforce Kubernetes security best practices (RBAC, Network Policies, Secrets Management, Security Contexts, Image Scanning) Automate cluster operations, deployment pipelines (CI/CD integration), and infrastructure provisioning using Infrastructure as Code (IaC) tools (e.g., Terraform, Ansible) Optimize Kubernetes clusters for performance, scalability, and resource utilization, particularly in edge environments Develop and maintain comprehensive documentation for cluster architecture, configurations, operational procedures, and runbooks Work in collaboration with software engineering, DevOps, security teams, and product managers to ensure seamless integration, deployment, and secure operation of applications on Kubernetes Evaluate and integrate new technologies from the Kubernetes ecosystem Contribute to the operational excellence of the platform, including participating in on-call rotations, incident management, and building self-healing capabilities Required Qualifications Bachelor's degree in computer science, Engineering, Information Technology, a related technical field, or equivalent practical experience At least 7+ years of professional experience in infrastructure engineering, systems administration, or software development, with a strong focus (4+ years preferred) on building and maintaining production Kubernetes environments At least 3+ years of professional experience using and administering Linux operating systems Deep understanding of Kubernetes architecture, core components, operational best practices, and lifecycle management Strong experience with containerization technologies (Docker) Hands-on experience managing Kubernetes on at least one major cloud provider (AWS, Azure, GCP) Strong understanding and proven experience with Infrastructure as Code (IaC) solutions, particularly Terraform and/or Ansible Proficiency in scripting languages (e.g., Python, Bash) for automation Experience configuring and managing monitoring/logging tools (e.g., Prometheus, Grafana, ELK Stack) Solid understanding of Linux operating system, networking fundamentals (TCP/IP, DNS, Load Balancing, Firewalls, VPNs) and container networking (CNI) Strong understanding of Kubernetes security concepts and implementation (RBAC, Network Policies, Secrets) Ability to work independently and collaborate effectively with others to debug and solve problems Preferred Experience and Skills Experience with Red Hat OpenShift Container Platform (version 4+ is a plus) Experience deploying and maintaining CI/CD solutions for DevSecOps, such as GitLab CI or Jenkins Strong development experience using Docker, docker-compose, and/or Kubernetes Experience developing Ansible playbooks for process automation Kubernetes certifications (CKA, CKS) Experience with Kubernetes operators and Custom Resource Definitions (CRDs) Experience with service mesh technologies like Istio or Linkerd Experience managing Kubernetes in edge computing or resource-constrained environments Compensation & Benefits For India-based candidates: We offer a competitive base salary along with equity options, providing an opportunity to share in the success and growth of Armada. #LI-JV1 #LI-Onsite You're a Great Fit if You're A go-getter with a growth mindset. You're intellectually curious, have strong business acumen, and actively seek opportunities to build relevant skills and knowledge A detail-oriented problem-solver. You can independently gather information, solve problems efficiently, and deliver results with a "get-it-done" attitude Thrive in a fast-paced environment. You're energized by an entrepreneurial spirit, capable of working quickly, and excited to contribute to a growing company A collaborative team player. You focus on business success and are motivated by team accomplishment vs personal agenda Highly organized and results-driven. Strong prioritization skills and a dedicated work ethic are essential for you Equal Opportunity Statement At Armada, we are committed to fostering a work environment where everyone is given equal opportunities to thrive. As an equal opportunity employer, we strictly prohibit discrimination or harassment based on race, color, gender, religion, sexual orientation, national origin, disability, genetic information, pregnancy, or any other characteristic protected by law. This policy applies to all employment decisions, including hiring, promotions, and compensation. Our hiring is guided by qualifications, merit, and the business needs at the time.
Posted 3 months ago
0 years
0 Lacs
Gurugram, Haryana, India
On-site
Role: Senior Site Reliability Engineer (SSE III) Responsibilities: Define and enforce SLOs, SLIs, and error budgets across microservices Architect an observability stack (metrics, logs, traces) and drive operational insights Automate toil and manual ops with robust tooling and runbooks Own incident response lifecycle: detection, triage, RCA, and postmortems Collaborate with product teams to build fault-tolerant systems Champion performance tuning, capacity planning, and scalability testing Optimise costs while maintaining the reliability of cloud infrastructure Must-have skills 6+ years in SRE/Infrastructure/Backend related roles using Cloud Native Technologies 2+ years in SRE-specific capacity Strong experience with monitoring/observability tools (Datadog, Prometheus, Grafana, ELK etc.) Experience with infrastructure-as-code (Terraform/Ansible) Proficiency in Kubernetes, service mesh (Istio/Linkerd), and container orchestration Deep understanding of distributed systems, networking, and failure domains Expertise in automation with Python, Bash, or Go Proficient in incident management, SLAs/SLOs, and system tuning Hands-on experience with GCP(preferred)/AWS/Azure and cloud cost optimisation Participation in on-call rotations and running large-scale production systems Nice to have skills: Familiarity with chaos engineering practices and tools (Gremlin, Litmus) Background in performance testing and load simulation (Gatling, Locust, k6, JMeter) Why us? You will be working with a lean team of passionate and talented individuals. We know that working with like-minded people is important. We are on a mission to supercharge brick-and-mortar retail stores in the era of e-commerce. Our customers give us confidence in our journey, and you will have a huge impact with your wor.k You will be free to experiment and can choose to do things differently. Lastly, we deeply care about a culture of being a solver. Come, be one with us! Equal opportunity employer Grey Orange Inc. is an equal employment opportunity employer. The company’s policy is not to discriminate against any applicant or employee based on race, color, religion, national origin, gender, age, sexual orientation, gender identity or expression, veteran status, marital status, mental or physical disability, and genetic information, or any other basis protected by applicable law. Grey Orange also prohibits harassment of applicants or employees based on any of these protected categories. Show more Show less
Posted 3 months ago
8 years
0 Lacs
Trivandrum, Kerala, India
On-site
Dear All, We’re Looking for Sr. Devops Tech Lead! Exciting opportunities await #ImmediateJoiners at our client office in Bangalore, Trivandrum, and Kochi. If you’re ready to take the next big step in your career, we’d love to hear from you! 📩 Send your resume or refer a friend to resumes@betterbeans.in and be a part of something better. Experience - 8+ YearsLocation – BLR/TVM/KOCNotice period – 30 Days (Immediate preferred)Salary- As per the industry standard 8+ years of experience in DevOps, hybrid cloud, and on-prem infrastructure automation.Expertise in Azure, or GCP is must, as well as on-premise infrastructure management.Strong experience with Kubernetes, Docker, Terraform, Helm .Proficiency in Terraform, Ansible, or CloudFormation for infrastructure automation.Hands-on experience with CI/CD tools like Jenkins, GitLab CI, or similar.Experience in monitoring, logging, and performance tuning of cloud and on-prem applications.Strong knowledge of Linux, Windows Server, networking, and security best practices.Hands-on experience with SRE principles (SLIs, SLOs, error budgets).Proficiency in scripting with Python, Bash, or Powershell.Strong problem-solving skills and ability to work in hybrid infrastructure environments.Experience leading a DevOps team or cross-functional projects.Strong understanding of networking, security, and server management.Excellent problem-solving skills and the ability to troubleshoot complex issues.Strong communication and teamwork abilities. Good To Have:Certifications in Microsoft Azure or Google Cloud.Knowledge of Service Mesh (Istio, Linkerd) and API gateways.Exposure to AI/ML pipelines and MLOps. Qualification: Bachelor’s degree in Computer Science, Information Technology, or a related field.
Posted 4 months ago
5 - 8 years
0 Lacs
Gurugram, Haryana, India
On-site
Role: Senior Site Reliability Engineer (SSE III) Responsibilities: Define and enforce SLOs, SLIs, and error budgets across microservicesArchitect an observability stack (metrics, logs, traces) and drive operational insightsAutomate toil and manual ops with robust tooling and runbooksOwn incident response lifecycle: detection, triage, RCA, and postmortemsCollaborate with product teams to build fault-tolerant systemsChampion performance tuning, capacity planning, and scalability testingOptimise costs while maintaining the reliability of cloud infrastructure Must-have skills 6+ years in SRE/Infrastructure/Backend related roles using Cloud Native Technologies2+ years in SRE-specific capacityStrong experience with monitoring/observability tools (Datadog, Prometheus, Grafana, ELK etc.)Experience with infrastructure-as-code (Terraform/Ansible)Proficiency in Kubernetes, service mesh (Istio/Linkerd), and container orchestrationDeep understanding of distributed systems, networking, and failure domainsExpertise in automation with Python, Bash, or GoProficient in incident management, SLAs/SLOs, and system tuningHands-on experience with GCP(preferred)/AWS/Azure and cloud cost optimisationParticipation in on-call rotations and running large-scale production systems Nice to have skills: Familiarity with chaos engineering practices and tools (Gremlin, Litmus)Background in performance testing and load simulation (Gatling, Locust, k6, JMeter) Why us? You will be working with a lean team of passionate and talented individuals. We know that working with like-minded people is important. We are on a mission to supercharge brick-and-mortar retail stores in the era of e-commerce. Our customers give us confidence in our journey, and you will have a huge impact with your wor.k You will be free to experiment and can choose to do things differently.Lastly, we deeply care about a culture of being a solver. Come, be one with us! Equal opportunity employer Grey Orange Inc. is an equal employment opportunity employer. The company’s policy is not to discriminate against any applicant or employee based on race, color, religion, national origin, gender, age, sexual orientation, gender identity or expression, veteran status, marital status, mental or physical disability, and genetic information, or any other basis protected by applicable law. Grey Orange also prohibits harassment of applicants or employees based on any of these protected categories.
Posted 4 months ago
5 years
0 Lacs
Ahmedabad, Gujarat, India
On-site
We are seeking a highly skilled Senior DevSecOps/DevOps Engineer with extensive experience in cloud infrastructure, automation, and security best practices. The ideal candidate must have 5+ years of overall experience, with at least 3+ years of direct, hands-on Kubernetes management experience. The candidate must have strong expertise in building, managing, and optimizing Jenkins pipelines for CI/CD workflows, with a focus on incorporating DevSecOps practices into the pipeline.Key Responsibilities: Design, deploy, and maintain Kubernetes clusters in cloud and/or on-premises environments.Build and maintain Jenkins pipelines for CI/CD, ensuring secure, automated, and efficient delivery processes.Integrate security checks (static code analysis, image scanning, etc.) directly into Jenkins pipelines.Manage Infrastructure as Code (IaC) using Terraform, Helm, and similar tools.Develop, maintain, and secure containerized applications using Docker and Kubernetes best practices.Implement monitoring, logging, and alerting using Prometheus, Grafana, and the ELK/EFK stack.Implement Kubernetes security practices including RBAC, network policies, and secrets management.Lead incident response efforts, root cause analysis, and system hardening initiatives.Collaborate with developers and security teams to embed security early in the development lifecycle (Shift-Left Security).Research, recommend, and implement best practices for DevSecOps and Kubernetes operations.Required Skills and Qualifications: 5+ years of experience in DevOps, Site Reliability Engineering, or Platform Engineering roles.3+ years of hands-on Kubernetes experience, including cluster provisioning, scaling, and troubleshooting.Strong expertise in creating, optimizing, and managing Jenkins pipelines for end-to-end CI/CD.Experience in containerization and orchestration: Docker and Kubernetes.Solid experience with Terraform Helm, and other IaC tools.Experience securing Kubernetes clusters, containers, and cloud-native applications.Scripting proficiency (Bash, Python, or Golang preferred).Knowledge of service meshes (Istio, Linkerd) and Kubernetes ingress management.Hands-on experience with security scanning tools (e.g., Trivy, Anchore, Aqua, SonarQube) integrated into Jenkins.Strong understanding of IAM, RBAC, and secret management systems like Vault or AWS Secrets Manager.
Posted 4 months ago
3.0 years
0 Lacs
Bengaluru, Karnataka, India
On-site
Our people work differently depending on their jobs and needs. From hybrid working to flexible hours, we have plenty of options that help our people to thrive. This role is based in India and as such all normal working days must be carried out in India. Job Description Join us as a Software Engineer This is an opportunity for a driven Software Engineer to take on an exciting new career challenge Day-to-day, you'll build a wide network of stakeholders of varying levels of seniority It’s a chance to hone your existing technical skills and advance your career We're offering this role as associate level What you'll do In your new role, you’ll engineer and maintain innovative, customer centric, high performance, secure and robust solutions. We are seeking a highly skilled and motivated AWS Cloud Engineer with deep expertise in Amazon EKS, Kubernetes, Docker, and Helm chart development. The ideal candidate will be responsible for designing, implementing, and maintaining scalable, secure, and resilient containerized applications in the cloud. You’ll also be: Design, deploy, and manage Kubernetes clusters using Amazon EKS. Develop and maintain Helm charts for deploying containerized applications. Build and manage Docker images and registries for microservices. Automate infrastructure provisioning using Infrastructure as Code (IaC) tools (e.g., Terraform, CloudFormation). Monitor and troubleshoot Kubernetes workloads and cluster health. Support CI/CD pipelines for containerized applications. Collaborate with development and DevOps teams to ensure seamless application delivery. Ensure security best practices are followed in container orchestration and cloud environments. Optimize performance and cost of cloud infrastructure. The skills you'll need You’ll need a background in software engineering, software design, architecture, and an understanding of how your area of expertise supports our customers. You'll need experience in Java full stack including Microservices, ReactJS, AWS, Spring, SpringBoot, SpringBatch, Pl/SQL, Oracle, PostgreSQL, Junit, Mockito, Cloud, REST API, API Gateway, Kafka and API development. You’ll also need: 3+ years of hands-on experience with AWS services, especially EKS, EC2, IAM, VPC, and CloudWatch. Strong expertise in Kubernetes architecture, networking, and resource management. Proficiency in Docker and container lifecycle management. Experience in writing and maintaining Helm charts for complex applications. Familiarity with CI/CD tools such as Jenkins, GitLab CI, or GitHub Actions. Solid understanding of Linux systems, shell scripting, and networking concepts. Experience with monitoring tools like Prometheus, Grafana, or Datadog. Knowledge of security practices in cloud and container environments. Preferred Qualifications: AWS Certified Solutions Architect or AWS Certified DevOps Engineer. Experience with service mesh technologies (e.g., Istio, Linkerd). Familiarity with GitOps practices and tools like ArgoCD or Flux. Experience with logging and observability tools (e.g., ELK stack, Fluentd). Show more Show less
Posted Date not available
7.0 - 12.0 years
15 - 25 Lacs
bengaluru
Work from Office
Position: Senior Developer / Cloud Engineer Service Mesh Work Location : Whitefield, Bangalore Work Mode: Hybrid Exp: 10+ years Strong understanding of Kubernetes architecture and networking. Hands-on experience with Service Mesh framework - Istio . Experience with mTLS, service discovery, sidecar injection, and ingress/egress gateways. Familiarity with Helm, Kustomize, or other Kubernetes deployment tools. Knowledge of CI/CD integration and GitOps practices. Experience with observability tools: Prometheus, Grafana, Jaeger, Kiali, Fluentd, ELK stack, etc. Proficiency in scripting (Bash, Python, or Go preferred). Solid understanding of cloud platforms (AWS, Azure, or GCP). Knowledge of security principles for microservices and zero-trust architecture. We are looking for a skilled engineer to lead the implementation and optimization of a Service Mesh architecture across our microservices ecosystem. The ideal candidate will have hands-on experience with Istio, Linkerd, Consul, or AWS App Mesh, and will be responsible for defining architecture, enabling observability, enforcing policies, and improving security and traffic management. Key Responsibilities: Design, deploy, and manage a Service Mesh (e.g., Istio, Linkerd) to support secure, scalable, and observable microservices communication. Define traffic routing policies, such as canary releases, blue-green deployments, and retries/failovers. Implement mutual TLS (mTLS), authorization, and access policies for service-to-service communication. Integrate the service mesh with observability stacks (Prometheus, Grafana, Jaeger, etc.) f or metrics, tracing, and logging. Work closely with development, SRE, and security teams to align mesh configuration with application requirements and security standards. Provide performance tuning, troubleshooting, and incident response related to mesh components. Document best practices and provide knowledge transfer to internal teams.
Posted Date not available
Upload Resume
Drag or click to upload
Your data is secure with us, protected by advanced encryption.
Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.
We have sent an OTP to your contact. Please enter it below to verify.
Accenture
73564 Jobs | Dublin
Wipro
27625 Jobs | Bengaluru
Accenture in India
22690 Jobs | Dublin 2
EY
20638 Jobs | London
Uplers
15021 Jobs | Ahmedabad
Bajaj Finserv
14304 Jobs |
IBM
14148 Jobs | Armonk
Accenture services Pvt Ltd
13138 Jobs |
Capgemini
12942 Jobs | Paris,France
Amazon.com
12683 Jobs |