Jobs
Interviews

1633 Grafana Jobs - Page 24

Setup a job Alert
JobPe aggregates results for easy application access, but you actually apply on the job portal directly.

4.0 - 8.0 years

0 Lacs

coimbatore, tamil nadu

On-site

As a Site Reliability Engineer (SRE) at our Coimbatore location, you will be responsible for ensuring the reliability and performance of our cloud and on-prem environments. Your key responsibilities will include driving root cause analysis to prevent incident recurrence, managing capacity planning and performance tuning, and participating in the on-call rotation for timely support and issue resolution. You will also be involved in designing, implementing, and maintaining CI/CD pipelines using tools such as Jenkins and GitHub, automating infrastructure deployment and monitoring following Infrastructure as Code (IaC) principles, and enhancing automation for operational tasks and incident response. In addition, you will implement and manage enterprise monitoring solutions like Splunk, Dynatrace, Prometheus, and Grafana, build real-time dashboards and alerts to proactively identify system anomalies, and continuously improve observability, logging, and tracing across all environments. Furthermore, you will work with AWS, Azure, and PCF environments, managing cloud-native services and infrastructure, designing and optimizing cloud architecture for reliability and cost-efficiency, and collaborating with cloud security and networking teams to ensure secure and compliant infrastructure. Your collaboration with product and development teams will ensure alignment with business objectives.,

Posted 2 weeks ago

Apply

10.0 - 14.0 years

0 Lacs

delhi

On-site

You are a highly experienced DevOps Architect and Level 4 DevOps Subject Matter Expert (SME) with over 10 years of relevant experience in the field of DevOps. Your expertise lies in building scalable, secure, and fully automated infrastructure environments, with a focus on delivering robust DevOps solutions, establishing architecture best practices, and driving automation across development and operations teams. Your role involves deep hands-on expertise in Continuous Integration And Continuous Delivery (CI/CD) tools like Jenkins, Azure DevOps, Helm, GIOPS, and ArgoCD to implement reliable, automated software delivery pipelines. You possess advanced Infrastructure as Code (IaC) experience using tools such as Terraform, Ansible, SaltStack, ARM Templates, and Google Cloud Deployment Manager for scalable and consistent infrastructure provisioning. You are an expert in container platforms, particularly Kubernetes and Docker, for orchestrated, secure, and highly available deployments. Your proficiency extends to Kubernetes operations, including production-grade cluster management, autoscaling, Helm chart development, RBAC configuration, ingress controllers, and network policy enforcement. Furthermore, you have extensive cloud experience across ROS, Azure, and GCP, with deep knowledge of core services, networking, storage, identity, and security implementations. Your scripting and automation capabilities using Bash, Python, or Go enable you to develop robust automation tools and system-level integrations. In addition, you have comprehensive monitoring and observability expertise with Prometheus, Grafana, and the ELK stack for end-to-end visibility, alerting, and performance analysis. You are skilled in designing and implementing secure, scalable, and resilient DevOps architectures aligned with industry best practices for both cloud-native and hybrid environments. Your experience also includes artifact management using JFrog Artifactory or Nexus, performing root cause analysis, developing self-healing scripts, and ensuring high system availability and minimal disruption. You are familiar with DevSecOps and compliance frameworks, mentoring engineering teams in DevOps adoption, tooling, automation strategies, and architectural decision-making. As a recognized DevOps expert and L4 SME, you continuously evaluate and recommend emerging tools, frameworks, and practices to enhance deployment speed, pipeline efficiency, and platform reliability. Your strong communication skills allow you to present and explain architectural strategies and system design decisions to both technical and non-technical stakeholders with clarity and confidence.,

Posted 2 weeks ago

Apply

4.0 - 8.0 years

0 Lacs

delhi

On-site

As a DevOps Engineer specializing in App Infrastructure & Scaling, you will be a crucial member of our technology team. Your primary responsibility will involve designing, implementing, and maintaining scalable and secure cloud infrastructure that supports our mobile and web applications. Your contributions will be essential in ensuring system reliability, performance optimization, and cost efficiency across different environments. Your key responsibilities will include designing and managing cloud infrastructure on Google Cloud Platform (GCP), implementing horizontal scaling, load balancers, auto-scaling groups, and performance monitoring systems. You will also be responsible for developing and managing CI/CD pipelines using tools like GitHub Actions, Jenkins, or GitLab CI. Setting up real-time monitoring, crash alerting, logging systems, and health dashboards using industry-leading tools will be part of your daily tasks. You will collaborate closely with Flutter and PHP (Laravel) teams to address performance bottlenecks and reduce system loads. Additionally, you will conduct infrastructure security audits, recommend best practices to prevent downtime and security breaches, and monitor and optimize cloud usage and billing for a cost-effective and scalable architecture. To be successful in this role, you should have at least 3-5 years of hands-on experience in a DevOps or Cloud Infrastructure role, preferably with GCP. Strong proficiency in Docker, Kubernetes, NGINX, and load balancing strategies is essential. Experience with CI/CD pipelines and tools like GitHub Actions, Jenkins, or GitLab CI, as well as familiarity with monitoring tools like Grafana, Prometheus, NewRelic, or Datadog, is required. Deep understanding of API architecture, PHP/Laravel backends, Firebase, and modern mobile app infrastructure is also necessary. Preferred qualifications include Google Cloud Professional certification or equivalent and experience in optimizing systems for high-concurrency, low-latency environments. Familiarity with Infrastructure as Code (IaC) tools such as Terraform or Ansible is a plus. In summary, as a DevOps Engineer specializing in App Infrastructure & Scaling, you will play a critical role in ensuring the scalability, reliability, and security of our cloud infrastructure that powers our applications. Your expertise will contribute to the overall performance and cost efficiency of our systems.,

Posted 2 weeks ago

Apply

3.0 - 7.0 years

0 Lacs

haryana

On-site

You will be responsible for implementing and managing CI/CD pipelines, container orchestration, and cloud services to enhance our software development lifecycle. Collaborate with development and operations teams to streamline processes and improve deployment efficiency. Implement and manage CI/CD tools such as GitLab CI, Jenkins, or CircleCI. Utilize Docker and Kubernetes (k8s) for containerization and orchestration of applications. Write and maintain scripts in at least one scripting language (e.g., Python, Bash) to automate tasks. Manage and deploy applications using cloud services (e.g. AWS, Azure, GCP) and their respective management tools. Understand and apply network protocols, IP networking, load balancing, and firewalling concepts. Implement infrastructure as code (IaC) practices to automate infrastructure provisioning and management. Utilize logging and monitoring tools (e.g., ELK stack, OpenSearch, Prometheus, Grafana) to ensure system reliability and performance. Familiarize with GitOps practices using tools like Flux or ArgoCD for continuous delivery. Work with Helm and Flyte for managing Kubernetes applications and workflows. Bachelors or masters degree in computer science, or a related field. Proven experience in a DevOps engineering role. Strong background in software development and system administration. Experience with CI/CD tools and practices. Proficiency in Docker and Kubernetes. Familiarity with cloud services and their management tools. Understanding of networking concepts and protocols. Experience with infrastructure as code (IaC) practices. Familiarity with logging and monitoring tools. Knowledge of GitOps practices and tools. Experience with Helm and Flyte is a plus. Preferred Qualifications: Experience with cloud-native architectures and microservices. Knowledge of security best practices in DevOps and cloud environments. Understanding database management and optimization (e.g., SQL, NoSQL). Familiarity with Agile methodologies and practices. Experience with performance tuning and optimization of applications. Knowledge of backup and disaster recovery strategies. Familiarity with emerging DevOps tools and technologies.,

Posted 2 weeks ago

Apply

3.0 - 7.0 years

0 Lacs

haryana

On-site

As a Kafka Administrator at our Merchant Ecommerce platform, located in Noida Sector 62, you will be responsible for managing, maintaining, and optimizing our distributed, multi-cluster Kafka infrastructure in an on-premise environment. Your role will require a deep understanding of Kafka internals, Zookeeper administration, performance tuning, and operational excellence in high-throughput, low-latency production systems. Additionally, experience with API gateway operations (specifically Kong) and observability tooling would be advantageous. Your key responsibilities will include managing multiple Kafka clusters with high-availability Zookeeper setups, conducting end-to-end operational support, capacity planning, implementation of backup and disaster recovery processes, enforcing security configurations, optimizing Kafka producer and consumer performance, planning and executing upgrades and patching, integrating with monitoring platforms, defining log retention and archival policies, monitoring Kafka metrics and logs, collaborating on security and compliance measures, and supporting regular vulnerability assessments. You will be expected to have at least 3+ years of hands-on Kafka administration experience in production environments, a strong understanding of Kafka internals and Zookeeper management, experience with performance tuning and troubleshooting, familiarity with security mechanisms like TLS/mTLS, ACLs, and SASL, proficiency with monitoring and logging tools, and scripting skills for operational automation. Experience with API gateways, Kubernetes-based environments, compliance standards, security hardening practices, and IaC tools would be a plus. In return, we offer you a mission-critical role in managing large-scale real-time data infrastructure, a flexible work environment, opportunities for growth, a supportive team, and access to modern observability and automation tools.,

Posted 2 weeks ago

Apply

5.0 - 9.0 years

0 Lacs

ahmedabad, gujarat

On-site

As a DevOps Engineer, you will be responsible for defining and implementing DevOps strategies that are in line with the business objectives. You will lead cross-functional teams to enhance collaboration among development, QA, and operations departments. Your role will involve designing, implementing, and managing Continuous Integration/Continuous Deployment (CI/CD) pipelines to automate build, test, and deployment processes, thus expediting release cycles. Furthermore, you will be tasked with implementing and overseeing Infrastructure as Code using tools such as Terraform, CloudFormation, Ansible, etc. Managing cloud platforms like AWS, Azure, or Google Cloud will also be a part of your responsibilities. It will be crucial for you to monitor and address security risks within CI/CD pipelines and infrastructure. Setting up observability tools like Prometheus, Grafana, Splunk, Datadog, etc., and implementing proactive alerting and incident response processes will be essential. In this role, you will take the lead in incident response and root cause analysis (RCA) activities. You will also play a key role in documenting DevOps processes, best practices, and system architectures. Additionally, you will be involved in evaluating and incorporating new DevOps tools and technologies. A significant aspect of your role will involve fostering a culture of continuous learning and sharing knowledge among team members.,

Posted 2 weeks ago

Apply

5.0 - 9.0 years

0 Lacs

thiruvananthapuram, kerala

On-site

Techvantage.ai is a next-generation technology and product engineering company at the forefront of innovation in Generative AI, Agentic AI, and autonomous intelligent systems. We build intelligent, scalable, and future-ready digital platforms that drive the next wave of AI-powered transformation. We are seeking a highly skilled and experienced Senior Node.js Developer with 5+ years of hands-on experience in backend development. As part of our engineering team, you will be responsible for architecting and building scalable APIs, services, and infrastructure that power high-performance AI-driven applications. You'll collaborate with front-end developers, DevOps, and data teams to ensure fast, secure, and efficient back-end functionality that meets the needs of modern AI-first products. What we are looking for in an ideal candidate: - Design, build, and maintain scalable server-side applications and APIs using Node.js and related frameworks. - Implement RESTful and GraphQL APIs for data-driven and real-time applications. - Collaborate with front-end, DevOps, and data teams to build seamless end-to-end solutions. - Optimize application performance, scalability, and security. - Write clean, maintainable, and well-documented code. - Integrate with third-party services and internal microservices. - Apply best practices in code quality, testing (unit/integration), and continuous integration/deployment. - Troubleshoot production issues and implement monitoring and alerting solutions. Requirements: - 5+ years of professional experience in backend development using Node.js. - Proficiency in JavaScript (ES6+) and strong experience with Express.js, NestJS, or similar frameworks. - Experience with SQL and NoSQL databases (e.g., PostgreSQL, MySQL, MongoDB). - Strong understanding of API security, authentication (OAuth2, JWT), and rate limiting. - Experience building scalable microservices and working with message queues (e.g., RabbitMQ, Kafka). - Familiarity with containerized applications using Docker and orchestration via Kubernetes. - Proficient in using Git, CI/CD pipelines, and version control best practices. - Solid understanding of performance tuning, caching, and system design. Preferred Qualifications: - Experience in cloud platforms like AWS, GCP, or Azure. - Exposure to building backends for AI/ML platforms, data pipelines, or analytics dashboards. - Familiarity with GraphQL, WebSockets, or real-time communication. - Knowledge of infrastructure-as-code tools like Terraform is a plus. - Experience with monitoring tools like Prometheus, Grafana, or New Relic. What We Offer: - The chance to work on cutting-edge products leveraging AI and intelligent automation. - A high-growth, innovation-driven environment with global exposure. - Access to modern development tools and cloud-native technologies. - Attractive compensation no constraints for the right candidate.,

Posted 2 weeks ago

Apply

10.0 - 14.0 years

0 Lacs

karnataka

On-site

As a Senior Software DevOps Engineer, you will be responsible for leading the design, implementation, and evolution of telemetry pipelines and DevOps automation to enable next-generation observability for distributed systems. Your main focus will be on leveraging a deep understanding of Open Telemetry architecture along with strong DevOps practices to construct a reliable, high-performance, and self-service observability platform that spans hybrid cloud environments such as AWS and Azure. Your primary goal will be to provide engineering teams with actionable insights through rich metrics, logs, and traces while promoting automation and innovation at all levels. In your role, you will be involved in the following key activities: Observability Strategy & Implementation: - Design and manage scalable observability solutions using OpenTelemetry (OTel), including deploying OTel Collectors for ingesting and exporting telemetry data, guiding teams on instrumentation best practices, building telemetry pipelines for data routing, and utilizing processors and extensions for advanced enrichment and routing. DevOps Automation & Platform Reliability: - Take ownership of the CI/CD experience using GitLab Pipelines, integrate infrastructure automation with Terraform, Docker, and scripting in Bash and Python, and develop resilient and reusable infrastructure-as-code modules across AWS and Azure ecosystems. Cloud-Native Enablement: - Create observability blueprints for cloud-native applications on AWS and Azure, optimize cost and performance of telemetry pipelines, and ensure SLA/SLO adherence for observability services. Monitoring, Dashboards, and Alerting: - Build and maintain role-based dashboards in tools like Grafana and New Relic for real-time visibility into service health and business KPIs, implement alerting best practices, and integrate with incident management systems. Innovation & Technical Leadership: - Drive cross-team observability initiatives to reduce MTTR and enhance engineering velocity, lead innovation projects such as self-service observability onboarding and AI-assisted root cause detection, and mentor engineering teams on telemetry standards and operational excellence. Qualifications and Skills: - 10+ years of experience in DevOps, Site Reliability Engineering, or Observability roles - Deep expertise with OpenTelemetry, GitLab CI/CD, Terraform, Docker, and scripting languages (Python, Bash, Go) - Hands-on experience with AWS and Azure services, cloud automation, and cost optimization - Proficiency with observability backends such as Grafana, New Relic, Prometheus, and Loki - Strong passion for building automated, resilient, and scalable telemetry pipelines - Excellent documentation and communication skills to drive adoption and influence engineering culture Nice to Have: - Certifications in AWS, Azure, or Terraform - Experience with OpenTelemetry SDKs in Go, Java, or Node.js - Familiarity with SLO management, error budgets, and observability-as-code approaches - Exposure to event streaming technologies (Kafka, RabbitMQ), Elasticsearch, Vault, and Consul,

Posted 2 weeks ago

Apply

2.0 - 6.0 years

0 Lacs

pune, maharashtra

On-site

You will be joining as a talented SDE1 - DevOps Engineer with the exciting opportunity to contribute towards building a top-notch DevOps infrastructure that can scale to accommodate the next 100M users. As an ideal candidate, you will be expected to tackle a variety of challenges with enthusiasm and take full ownership of your responsibilities. Your main responsibilities will include running a highly available Cloud-based software product on AWS, designing and implementing new systems in close collaboration with the Software Development team, setting up and maintaining CI/CD systems, and automating the deployment of software. You will also be tasked with continuously enhancing the security posture and operational efficiency of the Amber platform, as well as optimizing the operational costs. To excel in this role, you should possess 2-3 years of experience in a DevOps / SRE role, with a minimum of 2 years. You must have hands-on experience with AWS services such as ECS, EKS, RDS, Elasticache, and CloudFront, as well as familiarity with Google Cloud Platform. Proficiency in Infrastructure as Code tools like Terraform, CI/CD tools like Jenkins and GitHub Actions, and scripting languages such as Python and Bash is essential. Additionally, you should have a strong grasp of SCM in GitHub, networking concepts, and experience with observability and monitoring tools like Grafana, Loki, Prometheus, and ELK. Prior exposure to On-Call Rotation and mentoring junior DevOps Engineers would be advantageous. While not mandatory, knowledge of NodeJS and Ruby, including their platforms and workflows, would be considered a plus for this role.,

Posted 2 weeks ago

Apply

2.0 - 6.0 years

0 Lacs

chennai, tamil nadu

On-site

The role at Hitachi Energy India Development Centre (IDC) in Chennai offers you the opportunity to be part of a dedicated team of over 500 R&D engineers, specialists, and experts focused on creating innovative digital solutions, new products, and cutting-edge technology. As a part of the IDC team, you will collaborate with R&D and Research centers across more than 15 locations globally, contributing to the advancement of the world's energy system towards sustainability, flexibility, and security. Your primary responsibilities in this role include staying on track to meet project milestones and deadlines, actively suggesting and implementing process improvements, collaborating with a diverse team across different time zones, and enhancing processes related to continuous integration, deployment, testing, and release management. You will play a crucial role in developing, maintaining, and supporting azure infrastructure and system software components, providing guidance on azure tech components, ensuring application performance, uptime, and scalability, and leading CI/CD processes design and implementation. To excel in this position, you should possess at least 3 years of experience in azure DevOps, CI/CD, configuration management, and test automation, along with expertise in Azure PaaS, Azure Active Directory, Kubernetes, and application insight. Additionally, you should have hands-on experience with infrastructure as code automation, database management, system monitoring, security practices, containerization, and Linux system administration. Proficiency in at least one programming language, strong communication skills in English, and a commitment to Hitachi Energy's core values of safety and integrity are essential for success in this role. If you are a qualified individual with a disability and require accommodations during the job application process, you can request reasonable accommodations through our website. Please provide specific details about your needs to receive the necessary support. This opportunity is tailored for individuals seeking accessibility assistance, and inquiries for other purposes may not receive a response.,

Posted 2 weeks ago

Apply

10.0 - 14.0 years

0 Lacs

andhra pradesh

On-site

You are seeking a highly skilled Technical Architect with expertise in Java Spring Boot, React.js, IoT system architecture, and a strong foundation in DevOps practices. As the ideal candidate, you will play a pivotal role in designing scalable, secure, and high-performance IoT solutions, leading full-stack teams, and collaborating across product, infrastructure, and data teams. Your key responsibilities will include designing and implementing scalable and secure IoT platform architecture, defining data flow and event processing pipelines, architecting micro services-based solutions, and integrating them with React-based front-ends. You will also be responsible for defining CI/CD pipelines, managing containerization & orchestration, driving infrastructure automation, ensuring platform monitoring and observability, and enabling auto-scaling and zero-downtime deployments. In addition, you will need to collaborate with product managers and business stakeholders to translate requirements into technical specs, mentor and lead a team of developers and engineers, conduct code and architecture reviews, set goals and targets, and provide coaching and professional development to team members. Your role will also involve conducting unit testing, identifying risks, using coding standards and best practices to ensure quality, and maintaining a long-term outlook on the product roadmap and its enabling technologies. To be successful in this role, you must have hands-on IoT project experience, experience in designing and deploying multi-tenant SaaS platforms, strong knowledge of security best practices in IoT and cloud, excellent problem-solving, communication, and team leadership skills. It would be beneficial if you have experience with Edge Computing frameworks, AI/ML model integration into IoT pipelines, exposure to industrial protocols, experience with digital twin concepts, and certifications in relevant technologies. Ideally, you should have a Bachelor's or Master's degree in Computer Science, Engineering, or a related field. By joining us, you will have the opportunity to lead architecture for cutting-edge industrial IoT platforms, work with a passionate team in a fast-paced and innovative environment, and gain exposure to cross-disciplinary challenges in IoT, AI, and cloud-native technologies.,

Posted 2 weeks ago

Apply

6.0 - 10.0 years

0 Lacs

ahmedabad, gujarat

On-site

You will be responsible for leading a team of DevOps engineers in Ahmedabad. Your main duties will include managing and mentoring the team, overseeing the deployment and maintenance of various applications such as Odoo, Magento, and Node.js. You will also be in charge of designing and managing CI/CD pipelines using tools like Jenkins and GitLab CI, handling environment-specific configurations, and containerizing applications using Docker. In addition, you will need to implement and maintain Infrastructure as Code using tools like Terraform and Ansible, monitor application health and infrastructure, and ensure systems are secure, resilient, and compliant with industry standards. Collaboration with development, QA, and IT support teams is essential for seamless delivery, and troubleshooting performance, deployment, or scaling issues across tech stacks will also be part of your responsibilities. To be successful in this role, you should have at least 6 years of experience in DevOps/Cloud/System Engineering roles, with a minimum of 2 years managing or leading DevOps teams. Hands-on experience with Odoo, Magento, Node.js, and AWS/Azure/GCP infrastructure is required. Strong scripting skills in Bash, Python, PHP, or Node CLI, as well as a deep understanding of Linux system administration and networking fundamentals, are essential. Experience with Git, SSH, reverse proxies, and load balancers is also necessary, along with good communication skills and client management exposure. Preferred certifications that would be highly valued for this role include AWS Certified DevOps Engineer Professional, Azure DevOps Engineer Expert, and Google Cloud Professional DevOps Engineer. Bonus skills that are nice to have include experience with multi-region failover, HA clusters, MySQL/PostgreSQL optimization, GitOps, ArgoCD, Helm, VAPT 2.0, WCAG compliance, and infrastructure security best practices.,

Posted 2 weeks ago

Apply

5.0 - 9.0 years

0 Lacs

karnataka

On-site

You will be responsible for managing AWS Cloud Infrastructure services, Kubernetes, JIRA, and MySQL Database with a basic understanding of Datadog. Your technical skills should include proficiency in AWS Cloud Infrastructure, Grafana, Datadog, Azure AD, and moderate knowledge of Kubernetes. It would be beneficial to have experience with ITSM Tools such as Servicenow and JIRA, Linux operating system, and basic MySQL Database skills. Training can be provided for MQTT Protocol as needed. A typical day in this role will involve monitoring the health status of applications using Datadog or similar tools, reviewing incident/service tickets in Servicenow, conducting ticket analysis and root cause analysis, and collaborating with the L4 team for technical details. You will be required to submit incident reports, update Standard Operating Procedures (SOPs), coordinate with the Development and DevOps teams for troubleshooting, support platform version upgrades, perform builds for deployment (both frontend and backend), and manage backup and recovery processes according to guidelines. This position is based in various locations across India including Bangalore, Hyderabad, Mumbai, Chennai, Pune, Noida, Gurgaon, and Kolkata. The ideal candidate should have 5 to 7 years of experience in AWS Cloud Infrastructure and related technologies. This is a Contract-to-Hire position with a hybrid work mode, and only immediate joiners will be considered.,

Posted 2 weeks ago

Apply

5.0 - 9.0 years

0 Lacs

thane, maharashtra

On-site

As a Senior Software Engineer / Technical Architect for the B2C Platform at AjnaLens in Thane, Maharashtra, India, you will be responsible for architecting and building scalable backend services using Java (Spring Boot) and Python (FastAPI/Django). With at least 5 years of experience in backend development, you will possess a strong foundation in scalable system design and cloud-native infrastructure. Your role will involve leading key decisions on infrastructure, real-time data pipelines, LLM-integrated systems, and platform observability. You will collaborate closely with product, design, and infrastructure teams to ship user-first features at scale. Additionally, you will demonstrate a product mindset, technical depth, and the ability to deliver fast in a collaborative, purpose-driven environment. The ideal candidate will have deep hands-on experience with Java, Python, and frameworks like Spring Boot, Django, or FastAPI. You should also showcase proven expertise in building scalable microservices, streaming pipelines, and event-driven systems. Proficiency in tools like Redis, PostgreSQL, PGVector, and cloud platforms such as AWS/GCP is essential. Exposure to LLM features like RAG, vector search, GPT workflows, or chatbot architectures will be advantageous. In this role, you will be expected to possess solid system design capabilities, lead architecture reviews, and mentor junior developers. Full-stack awareness, including familiarity with front-end tools like ReactJS for internal dashboards and modern DevOps practices, will be beneficial. Experience with containerization and observability tools like Prometheus, Grafana, and OTEL is desired. Bonus points will be awarded for experience with Agentic systems, real-time infrastructure, or AI-driven product features. Prior involvement in building high-performance consumer products from scratch and a passion for conscious innovation and impactful technology will be highly valued. As part of the team at AjnaLens, you will create a robust, scalable backend for the next-gen B2C platform. You will develop real-time, intelligent user experiences that combine performance with personalization, in alignment with AjnaLens" vision of building mindfully at scale. Join us in co-creating the future of conscious technology with Lenskart as our strategic investor.,

Posted 2 weeks ago

Apply

11.0 - 15.0 years

0 Lacs

hyderabad, telangana

On-site

As an AI Azure Architect, your primary responsibility will be to develop the technical vision for AI systems that cater to the existing and future business requirements. This involves architecting end-to-end AI applications, ensuring seamless integration with legacy systems, enterprise data platforms, and microservices. Collaborating closely with business analysts and domain experts, you will translate business objectives into technical requirements and AI-driven solutions. Additionally, you will partner with product management to design agile project roadmaps aligning technical strategies with market needs. Coordinating with data engineering teams is essential to ensure smooth data flows, quality, and governance across different data sources. Your role will also involve leading the design of reference architectures, roadmaps, and best practices for AI applications. Evaluating emerging technologies and methodologies to recommend suitable innovations for integration into the organizational strategy is a crucial aspect of your responsibilities. You will be required to identify and define system components such as data ingestion pipelines, model training environments, CI/CD frameworks, and monitoring systems. Leveraging containerization (Docker, Kubernetes) and cloud services will streamline the deployment and scaling of AI systems. Implementation of robust versioning, rollback, and monitoring mechanisms to ensure system stability, reliability, and performance will be part of your duties. Moreover, you will oversee the planning, execution, and delivery of AI and ML applications, ensuring they are completed within budget and timeline constraints. Managing project goals, allocating resources, and mitigating risks will fall under your project management responsibilities. You will be responsible for overseeing the complete lifecycle of AI application developmentfrom conceptualization and design to development, testing, deployment, and post-production optimization. Emphasizing security best practices during each development phase, focusing on data privacy, user security, and risk mitigation, is crucial. In addition to technical skills, the ideal candidate for this role should possess key behavioral attributes such as the ability to mentor junior developers, take ownership of project deliverables, and contribute towards risk mitigation. Understanding business objectives and functions to support data needs is also essential. Mandatory technical skills for this position include a strong background in working with agents using langgraph, autogen, and CrewAI. Proficiency in Python, along with knowledge of machine learning libraries like TensorFlow, PyTorch, and Keras, is required. Experience with cloud computing platforms (AWS, Azure, Google Cloud Platform), containerization tools (Docker), orchestration frameworks (Kubernetes), and DevOps tools (Jenkins, GitLab CI/CD) is essential. Proficiency in SQL and NoSQL databases, designing distributed systems, RESTful APIs, GraphQL integrations, and event-driven architectures are also necessary. Preferred technical skills include experience with monitoring and logging tools, cutting-edge libraries like Hugging Face Transformers, and large-scale deployment of ML projects. Training and fine-tuning of Large Language Models (LLMs) is an added advantage. Educational qualifications for this role include a Bachelor's/Master's degree in Computer Science, along with certifications in Cloud technologies (AWS, Azure, GCP) and TOGAF certification. The ideal candidate should have 11 to 14 years of relevant work experience in this field.,

Posted 2 weeks ago

Apply

3.0 - 7.0 years

0 Lacs

chennai, tamil nadu

On-site

You are seeking a hands-on backend expert to elevate your FastAPI-based platform to the next level by developing production-grade model-inference services, agentic AI workflows, and seamless integration with third-party LLMs and NLP tooling. In this role, you will be responsible for various key areas: 1. Core Backend Enhancements: - Building APIs - Strengthening security with OAuth2/JWT, rate-limiting, SecretManager, and enhancing observability through structured logging and tracing - Adding CI/CD, test automation, health checks, and SLO dashboards 2. Awesome UI Interfaces: - Developing UI interfaces using React.js/Next.js, Redact/Context, and various CSS frameworks like Tailwind, MUI, Custom-CSS, and Shadcn 3. LLM & Agentic Services: - Designing micro/mini-services to host and route to platforms such as OpenAI, Anthropic, local HF models, embeddings & RAG pipelines - Implementing autonomous/recursive agents that orchestrate multi-step chains for Tools, Memory, and Planning 4. Model-Inference Infrastructure: - Setting up GPU/CPU inference servers behind an API gateway - Optimizing throughput with techniques like batching, streaming, quantization, and caching using tools like Redis and pgvector 5. NLP & Data Services: - Managing the NLP stack with Transformers for classification, extraction, and embedding generation - Building data pipelines to combine aggregated business metrics with model telemetry for analytics You will be working with a tech stack that includes Python, FastAPI, Starlette, Pydantic, Async SQLAlchemy, Postgres, Docker, Kubernetes, AWS/GCP, Redis, RabbitMQ, Celery, Prometheus, Grafana, OpenTelemetry, and more. Experience in building production Python REST APIs, SQL schema design in Postgres, async patterns & concurrency, UI application development, RAG, LLM/embedding workflows, cloud container orchestration, and CI/CD pipelines is essential for this role. Additionally, experience with streaming protocols, NGINX Ingress, SaaS security hardening, data privacy, event-sourced data models, and other related technologies would be advantageous. This role offers the opportunity to work on evolving products, tackle real challenges, and lead the scaling of AI services while working closely with the founder to shape the future of the platform. If you are looking for meaningful ownership and the chance to solve forward-looking problems, this role could be the right fit for you.,

Posted 2 weeks ago

Apply

2.0 - 6.0 years

0 Lacs

kerala

On-site

As a Data Analyst at our company based in Trivandrum, you will be responsible for analyzing large volumes of log data from NGINX server logs to identify user behavior patterns, anomalies, and security events. Your role will involve interpreting various fields such as IP addresses, geolocation data, user agents, request paths, status codes, and request times to derive meaningful insights. Collaboration with AI engineers is crucial as you will work together to propose relevant features based on log behavior and traffic patterns. Your responsibilities will include validating engineered features, conducting exploratory data analysis, and ensuring the quality and alignment of feature logic with real-world HTTP behavior and use cases. Furthermore, you will be involved in developing data visualizations to represent time-series trends, geo-distributions, and traffic behavior. Your collaboration with the frontend/dashboard team will be essential in defining and testing visual requirements and anomaly indicators for real-time dashboards. In addition to your analytical tasks, you will also be responsible for identifying and addressing gaps, inconsistencies, and errors in raw logs to ensure data quality. Creating documentation that explains observed behavioral patterns, feature assumptions, and traffic insights for knowledge sharing within the ML and security team will also be part of your role. The minimum qualifications for this position include a Bachelor's degree in Computer Science, Information Systems, Data Analytics, Cybersecurity, or a related field, along with at least 2 years of experience in data analysis or analytics roles. Proficiency in SQL, Elasticsearch queries, Python for data analysis, and experience working with web server logs or structured event data are required. Strong analytical thinking skills are essential to break down complex log behavior into patterns and outliers. It would be beneficial if you have familiarity with web security concepts, experience with log analytics platforms, an understanding of feature engineering concepts in ML pipelines, or experience working on anomaly detection or security analytics systems. This is a full-time position with benefits such as health insurance and Provident Fund, with a day shift schedule from Monday to Friday. If you possess the necessary qualifications and experience, we look forward to receiving your application.,

Posted 2 weeks ago

Apply

1.0 - 5.0 years

0 Lacs

karnataka

On-site

At Goldman Sachs, our Engineers don't just make things - we make things possible. We change the world by connecting people and capital with ideas, solving the most challenging and pressing engineering problems for our clients. Join our engineering teams that build massively scalable software and systems, architect low latency infrastructure solutions, proactively guard against cyber threats, and leverage machine learning alongside financial engineering to continuously turn data into action. Create new businesses, transform finance, and explore a world of opportunity at the speed of markets. Engineering, which is comprised of our Technology Division and global strategists groups, is at the critical center of our business. Our dynamic environment requires innovative strategic thinking and immediate, real solutions. If you want to push the limit of digital possibilities, start here. Goldman Sachs Engineers are innovators and problem-solvers, building solutions in risk management, big data, mobile, and more. We look for creative collaborators who evolve, adapt to change, and thrive in a fast-paced global environment. Data plays a critical role in every facet of the Goldman Sachs business. The Data Engineering group is at the core of that offering, focusing on providing the platform, processes, and governance for enabling the availability of clean, organized, and impactful data to scale, streamline, and empower our core businesses. As a Site Reliability Engineer (SRE) on the Data Engineering team, you will be responsible for observability, cost, and capacity with operational accountability for some of Goldman Sachs's largest data platforms. We engage in the full lifecycle of platforms from design to demise with an adapted SRE strategy to the lifecycle. We are looking for individuals with a background as a developer who can express themselves in code. You should have a focus on Reliability, Observability, Capacity Management, DevOps, and SDLC (Software Development Lifecycle). As a self-leader comfortable with problem statements, you should structure them into data-driven deliverables. You will drive strategy with skin in the game, participate in the team's activities, drive Postmortems, and have an attitude that the problem stops with you. **How You Will Fulfil Your Potential** - Drive adoption of cloud technology for data processing and warehousing - Drive SRE strategy for some of GS's largest platforms including Lakehouse and Data Lake - Engage with data consumers and producers to match reliability and cost requirements - Drive strategy with data **Relevant Technologies**: Snowflake, AWS, Grafana, PromQL, Python, Java, Open Telemetry, Gitlab **Basic Qualifications** - A Bachelor's or Master's degree in a computational field (Computer Science, Applied Mathematics, Engineering, or in a related quantitative discipline) - 1-4+ years of relevant work experience in a team-focused environment - 1-2 years hands-on developer experience at some point in career - Understanding and experience of DevOps and SRE principles and automation, managing technical and operational risk - Experience with cloud infrastructure (AWS, Azure, or GCP) - Proven experience in driving strategy with data - Deep understanding of multi-dimensionality of data, data curation, and data quality - In-depth knowledge of relational and columnar SQL databases, including database design - Expertise in data warehousing concepts - Excellent communication skills - Independent thinker, willing to engage, challenge, or learn - Ability to stay commercially focused and to always push for quantifiable commercial impact - Strong work ethic, a sense of ownership and urgency - Strong analytical and problem-solving skills - Ability to build trusted partnerships with key contacts and users across business and engineering teams **Preferred Qualifications** - Understanding of Data Lake / Lakehouse technologies incl. Apache Iceberg - Experience with cloud databases (e.g., Snowflake, Big Query) - Understanding concepts of data modeling - Working knowledge of open-source tools such as AWS lambda, Prometheus - Experience coding in Java or Python,

Posted 2 weeks ago

Apply

4.0 - 8.0 years

0 Lacs

pune, maharashtra

On-site

As a PLM Infrastructure Engineer, you will be responsible for operating and managing enterprise-grade 3DEXPERIENCE platforms utilized in engineering and manufacturing within the Dassault Systmes ecosystem. Your role will encompass platform operations, infrastructure automation, and release management, ensuring the seamless functioning of multiple 3DX environments across on-premises and AWS setups. Your primary duties will include maintaining and managing various 3DEXPERIENCE environments such as production, integration, development, and AWS sandboxes. You will conduct 3DX upgrades, set up new environments, and deploy software updates while administering and troubleshooting Linux, Oracle DB, and Windows Server components. Additionally, you will be tasked with packaging and deploying CATIA clients, managing license infrastructure, and automating processes using tools like Ansible, GitHub for CI/CD, and monitoring systems with Nagios and Grafana. Implementing security policies, backup procedures, and system-level controls will also fall under your purview, as well as supporting interfaces with AD, IAM, and People & Organization systems. Your role will require hands-on experience with 3DEXPERIENCE platform, CATIA V5, and PowerBY, along with strong infrastructure knowledge in Linux, Oracle DB, and Windows Server environments. Skills in Ansible, GitHub CI/CD, and basic cloud platforms like AWS or Azure are essential, as is a deep understanding of PLM ecosystem dynamics and platform lifecycle management. Experience in operating on-premises and hybrid 3DX environments, familiarity with Agile tools such as Jira and ServiceNow, and knowledge of UFT and Selenium for component testing are considered advantageous. Additionally, expertise in Infrastructure as Code (IaC) using tools like Terraform is a plus. Joining this role will offer you the opportunity to work within a mission-critical environment that is integral to global engineering teams, providing real platform ownership and the chance to contribute significantly to enterprise engineering platforms.,

Posted 2 weeks ago

Apply

1.0 - 5.0 years

0 Lacs

kochi, kerala

On-site

As a Java Backend Developer in our team specializing in the IoT domain, your role will involve designing, developing, and deploying scalable microservices utilizing Spring Boot, SQL databases, and AWS services. You will play a pivotal role in guiding the backend development team, implementing DevOps best practices, and optimizing cloud infrastructure to ensure high-performance and secure services. Your key responsibilities will include architecting and implementing high-performance backend services using Java (Spring Boot), developing RESTful APIs and event-driven microservices with a focus on scalability and reliability, designing and optimizing SQL databases (PostgreSQL, MySQL), and deploying applications on AWS utilizing services like ECS, Lambda, RDS, S3, and API Gateway. In addition, you will be tasked with implementing CI/CD pipelines using tools such as GitHub Actions, Jenkins, or similar, monitoring and optimizing backend performance, ensuring best practices for security, authentication, and authorization using OAuth, JWT, and IAM roles, and collaborating with the team to maintain high standards of efficiency and quality. The ideal candidate will possess expertise in Java (Spring Boot, Spring Cloud, Spring Security), microservices architecture, API development, SQL (PostgreSQL, MySQL), ORM (JPA, Hibernate), DevOps tools (Docker, Kubernetes, Terraform, CI/CD, GitHub Actions, Jenkins), AWS cloud services (EC2, Lambda, ECS, RDS, S3, IAM, API Gateway, CloudWatch), messaging systems (Kafka, RabbitMQ, SQS, MQTT), testing frameworks (JUnit, Mockito, Integration Testing), and logging & monitoring tools (ELK Stack, Prometheus, Grafana). Preferred skills that would be beneficial for this role include experience in the IoT domain, previous work experience in startups, familiarity with event-driven architecture using Apache Kafka, knowledge of Infrastructure as Code (IaC) with Terraform, and exposure to serverless architectures. In return, we offer a competitive salary with performance-based incentives, the opportunity to lead and mentor a high-performing tech team, hands-on experience with cutting-edge cloud and microservices technologies, and a collaborative, fast-paced work environment where your skills and expertise will be valued and further developed. If you have experience in any IoT domain and are enthusiastic about contributing to a dynamic team focused on innovation and excellence, we invite you to apply for this full-time, on-site/hybrid Java Backend Developer position in Kochi.,

Posted 2 weeks ago

Apply

4.0 - 8.0 years

0 Lacs

ahmedabad, gujarat

On-site

As an Experienced Systems Administrator, you will have a strong foundation in Linux, infrastructure management, and incident response. You will be skilled in monitoring, troubleshooting, and maintaining reliable systems across virtualized and cloud-based environments. Your main responsibilities will include collaborating with the operations team to manage escalations and oversee incident management. You will also be expected to implement strategies and solutions to enhance daily operations, focusing on system stability, security, and scalability. You will drive real-time monitoring of system performance and capacity, addressing alerts promptly to optimize systems. Leading troubleshooting efforts, you will coordinate responses to network and system issues. Your role will involve conducting and overseeing server, application, and network equipment setup and maintenance. Additionally, you will ensure effective outage notification and escalation for prompt resolution. Furthermore, mentoring and training team members on technical skills and troubleshooting methods will be a key part of your responsibilities. You will also be responsible for maintaining up-to-date documentation of processes and procedures in the WIKI. Key Skills: - Minimum 4 years of experience in Linux system administration. - Proficiency in datacenter technologies and cloud platforms such as AWS/GCP. - Experience in application deployment using tools like Git and StackStorm. - Strong troubleshooting skills across networks and systems, including familiarity with network protocols (TCP/IP, UDP, ICMP) and tools like TCPdump. - Advanced diagnostic skills in network performance and system capacity monitoring. - Proficiency in Linux command-line operations. - Analytical skills with the ability to interpret and act on data effectively. - Ability to prioritize and escalate issues efficiently. - Adaptability to shift work and capacity for multitasking in high-pressure scenarios. - Excellent leadership, communication, and interpersonal skills. - Bachelor's degree in Computer Science, Engineering (BE/B.Tech), MCA, or M.Sc. Desired Skills: - Basic experience with Configuration Management tools like Ansible, SaltStack, or StackStorm. - Basic experience with CI/CD tools like Jenkins. - Experience with monitoring tools such as Nagios, Sensu, Zabbix. - Basic experience with Log Analytics tools like Splunk, Elasticsearch, Sumo Logic, Prometheus, or Grafana. - Knowledge of Virtualization technologies like VMware, KVM. - Strong fundamentals in Linux, troubleshooting, and networking. - Knowledge of Containerization technologies like Kubernetes, Rancher. - Experience with Cloud Providers such as AWS or GCP. - Advanced knowledge of Networking concepts including BGP, F5 Load Balancer, and switching protocols. - Relevant certifications like RHCSA, CCNA, or equivalent. (hirist.tech),

Posted 2 weeks ago

Apply

10.0 - 14.0 years

0 Lacs

karnataka

On-site

As a member of the Data Center Network Services team at Cisco IT, you will be responsible for supporting network services for Cisco Engineering and business functions globally. Your primary mission will be to build a future-ready network that is adaptable and agile using Cisco's networking solutions. The networks you will be working on are deployed, monitored, and managed using a DevOps approach to facilitate rapid application changes. By investing in cutting-edge technologies, we ensure the delivery of services in a fast and reliable manner. The team culture fosters collaboration, creativity, and fun, encouraging team members to think innovatively and explore new ideas. In this environment, you will play a crucial role in designing, developing, testing, and deploying Data Center network capabilities. Your work will involve engaging with fellow engineers from different disciplines and internal clients to create innovative and high-quality solutions that enhance our clients" experience. **Minimum Requirements:** - Bachelor of Engineering or Technology with a minimum of 10 years of experience in designing, deploying, operating, and managing scalable DC network infrastructure using Nexus OS - Proficiency in technologies such as Routing, Switching, Nexus, VPC, VDC, VLAN, VXLAN, and BGP - Experience in incident, problem, and organizational change management - Familiarity with DevOps principles and comfortable with Agile practices **Preferred Qualifications:** - CCNP or CCIE/DE certification - Experience with SONiC NoS including basic configuration, network problem-solving, QoS monitoring and fix (especially for RoCEv2), BGP routing - Desirable experience with L3 Fabrics, Nvidia and Linux networking, Python, Prometheus, Splunk, Grafana, and Cisco Firepower firewalls (FTD/FMC) **Nice to have Qualifications:** - Experience with Nexus Dashboard Fabric Controller for network building and troubleshooting - Experience with VXLan-based networks and problem-solving In conclusion, at Cisco, we are at the forefront of revolutionizing how data and infrastructure connect and protect organizations in the AI era and beyond. With a history of 40 years of innovation, we create solutions that enable humans and technology to work together seamlessly across physical and digital realms. Our solutions empower customers with unrivaled security, visibility, and insights across their entire digital footprint. By leveraging our technology and global network, we continuously experiment and innovate to build impactful solutions. As part of the Cisco team, you will have limitless opportunities to grow and contribute on a global scale, collaborating with appreciation to achieve significant milestones. Cisco's impact is omnipresent, and it all starts with you.,

Posted 2 weeks ago

Apply

4.0 - 8.0 years

0 Lacs

chennai, tamil nadu

On-site

The role available is within the Services - Client & Services Assurance group, which is comprised of various core streams, with direct oversight falling under the BCC - Orion stream. As an App Support Senior Analyst, you are expected to possess strong techno-functional skills. Your primary responsibilities include investigating alerts to pinpoint root causes through in-depth analysis of application flows, ensuring clear communication with stakeholders regarding system and business service availability, and managing Major Incidents to coordinate responses and escalations to maintain the flow of TTS Services to client-facing teams. Additionally, you will be tasked with identifying and eliminating operational inefficiencies from day-to-day support tasks, leading initiatives to enhance service efficiency and stability, and acting as a bridge between Business and Technology teams to expedite incident resolutions and provide a business-oriented perspective on remediation strategies. To excel in this role, you should hold a Technology academic degree with 4-6 years of relevant work experience. Proficiency in database management tools such as Oracle, Mongo DB, and Sybase, familiarity with application monitoring platforms like Grafana, ITRS Geneos, and AppDynamics, and a basic understanding of webserver technologies (WebLogic or WebSphere) are essential. Moreover, knowledge of Middleware technologies such as MQ, Kafka, and experience with ITIL tools like ServiceNow (ITIL Foundation Certification preferred) will be beneficial. Strong troubleshooting and problem-solving abilities coupled with effective verbal and written communication skills are crucial for success in this position. A foundational grasp of Observability, Site Reliability Engineering (SRE), and Open Telemetry Principles will also be advantageous. This is a full-time role within the Technology job family group, specifically categorized under Applications Support. If you require any accommodations due to a disability to utilize our search tools or apply for this position, please refer to the Accessibility at Citi guidelines. For further details on Citis EEO Policy Statement and your rights, please review the respective documents.,

Posted 2 weeks ago

Apply

15.0 - 19.0 years

0 Lacs

pune, maharashtra

On-site

You will be joining the Data Solutions Technology team, whose goal is to provide a competitive advantage to the business by delivering high-quality, innovative, and cost-effective reference data technology solutions. The team aims to meet the needs of the business, clients, regulators, and stakeholders effectively. As the Apps Support Group Manager, you will lead a team of professionals and be responsible for managing complex and critical disciplinary areas. Your role will require a comprehensive understanding of various functions within the team and how they interact to achieve objectives. Strong commercial awareness is essential, and you will be accountable for delivering services to one or more businesses or geographic regions. Your responsibilities will include improving the service level provided to end-users, guiding development teams on application stability, managing vendor relationships, and implementing frameworks for managing capacity and latency. You will also drive cost reductions and efficiencies across supported portfolios, participate in business review meetings, and ensure adherence to support process standards. To qualify for this role, you should have at least 15 years of experience in Application Production Support, with 5+ years in a strategic role. Deep expertise in Big Data platforms such as Hadoop, Spark, Kafka, and Hive is required, along with a proven track record in driving stability, resiliency, and automation initiatives. A post-graduation in a relevant field is preferred, and you should have senior-level experience in Apps Support with people management skills. Excellent communication skills, stakeholder management experience, project management skills, and capacity planning/forecasting exposure are valuable assets for this role. You should also be able to share information effectively with support team members and other technology teams, plan and organize workloads efficiently, and demonstrate clear and concise communication skills. A Bachelors/University degree is required, with a Masters degree being preferred for this position. Your role will involve full-time employment, and the most relevant skills needed include Application Production Support, Big Data Platforms, Grafana, and ITRS Geneos. If you require a reasonable accommodation due to a disability to use our search tools or apply for a career opportunity, please review the Accessibility at Citi.,

Posted 2 weeks ago

Apply

1.0 - 5.0 years

0 Lacs

pune, maharashtra

On-site

You will be responsible for providing a high level of customer satisfaction by delivering technical back-office support for Wolters Kluwer's cloud-hosted financial management software applications. Your main tasks will involve analyzing and resolving complex problems, collaborating with consultants and clients, working with internal development teams, and providing solutions to reported issues. Your responsibilities will include offering technical support to our international customers and consultants on cloud environments, analyzing technical questions and problems, troubleshooting server and network issues, and collaborating with various teams to enhance product functionality. Additionally, you will be expected to develop and maintain expertise in specific areas of product functionality, conduct acceptance tests, and enhance automation utilities to improve integration and flexibility for customers. To be considered for this role, you should hold a Bachelor's or Master's degree in a computer science-related field, have knowledge or experience with Azure or other cloud platforms, and be proficient in programming languages such as Java, JavaScript, and SQL. Experience with operating systems like Windows (Linux is a plus), application servers like Wildfly or Jboss, and databases such as SQL (MS SQL Server, Oracle) is also required. You should be available for stand-by shifts and possess strong problem-solving, analytical, organizational, and time management skills. Nice to have qualifications include awareness of Application Monitoring Tools (e.g. Grafana), knowledge of Docker and Kubernetes, experience with cloud performance tuning and security testing, and proficiency in other languages like German or French. Applicants may be required to visit a Wolters Kluwer office as part of the recruitment process. If you meet the requirements and are looking to join a dynamic team to support customers and enhance product functionality, this role may be a great fit for you.,

Posted 2 weeks ago

Apply
cta

Start Your Job Search Today

Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.

Job Application AI Bot

Job Application AI Bot

Apply to 20+ Portals in one click

Download Now

Download the Mobile App

Instantly access job listings, apply easily, and track applications.

Featured Companies