Get alerts for new jobs matching your selected skills, preferred locations, and experience range.
4.0 years
0 Lacs
Noida, Uttar Pradesh, India
On-site
POSITION: Sr. Devops Engineer Job Type: Work From Office (5 days) Location: Sector 16A, Film City, Noida / Mumbai Relevant Experience: Minimum 4+ year Salary: Competitive Education- B.Tech About the Company: Devnagri is a AI company dedicated to personalizing business communication and making it hyper-local to attract non-English speakers. We address the significant gap in internet content availability for most of the world’s population who do not speak English. For more detail - Visit www.devnagri.com We seek a highly skilled and experienced Senior DevOps Engineer to join our dynamic team. As a key member of our technology department, you will play a crucial role in designing and implementing scalable, efficient and robust infrastructure solutions with a strong focus on DevOps automation and best practices. Roles And Responsibilities Design, plan, and implement scalable, reliable, secure, and robust infrastructure architectures Manage and optimize cloud-based infrastructure components - Architect and implement containerization technologies, such as Docker, Kubernetes Implement the CI/CD pipelines to automate the build, test, and deployment processes Design and implement effective monitoring and logging solutions for applications and infrastructure. Establish metrics and alerts for proactive issue identification and resolution Work closely with cross-functional teams to troubleshoot and resolve issues. Implement and enforce security best practices across infrastructure components Establish and enforce configuration standards across various environments. Implement and manage infrastructure using Infrastructure as Code principles Leverage tools like Terraform for provisioning and managing resources. Stay abreast of industry trends and emerging technologies. Evaluate and recommend new tools and technologies to enhance infrastructure and operations Must Have Skills Cloud ( AWS & GCP ), Redis, MongoDB, MySQL, Docker, bash scripting, Jenkins, Prometheus, Grafana, ELK Stack, Apache, Linux Good To Have Skills Kubernetes, Collaboration and Communication, Problem Solving, IAM, WAF, SAST/DAST Interview Process Screening Round then Shortlisting >> 3 technical round >> 1 Managerial round >> HR Closure with your short success story into Devops and Tech Cheers For more details, visit our website- https://www.devnagri.com Note for approver Skills:- DevOps, Linux/Unix, Apache, Amazon Web Services (AWS), Google Cloud Platform (GCP), prometheus, grafana, MongoDB, MySQL and CI/CD Show more Show less
Posted 2 days ago
18.0 years
0 Lacs
Pune, Maharashtra, India
On-site
Role: Enterprise Architect Grade: VP Location: Pune / Mumbai/Chennai Experience: 18+ Years Organization: Intellect Design Arena Ltd. www.intellectdesign.com About the Role: We are looking for a senior Enterprise Architect with strong leadership and deep technical expertise to define and evolve the architecture strategy for iGTB , our award-winning transaction banking platform. The ideal candidate will have extensive experience architecting large-scale, cloud-native enterprise applications within the BFSI domain , and will be responsible for driving innovation, ensuring engineering excellence, and aligning architecture with evolving business needs. Mandatory Skills: Cloud-native architecture Microservices-based systems PostgreSQL, Apache Kafka, ActiveMQ Spring Boot / Spring Cloud, Angular Strong exposure to BFSI domain Key Responsibilities: Architectural Strategy & Governance: Define and maintain enterprise architecture standards and principles across iGTB product suites. Set up governance structures to ensure compliance across product lines. Technology Leadership: Stay updated on emerging technologies; assess and recommend adoption to improve scalability, security, and performance. Tooling & Automation: Evaluate and implement tools to improve developer productivity, code quality, and application reliability—including automation across testing, deployment, and monitoring. Architecture Evangelism: Drive adoption of architecture guidelines and tools across engineering teams through mentorship, training, and collaboration. Solution Oversight: Participate in the design of individual modules to ensure technical robustness and adherence to enterprise standards. Performance & Security: Oversee performance benchmarking and security assessments. Engage with third-party labs for certification as needed. Customer Engagement: Represent architecture in pre-sales, CXO-level interactions, and post-production engagements to demonstrate the product's technical superiority. Troubleshooting & Continuous Improvement: Support teams in resolving complex technical issues. Capture learnings and feed them back into architectural best practices. Automation Vision: Lead the end-to-end automation charter for iGTB—across code quality, CI/CD, testing, monitoring, and release management. Profile Requirements: 18+ years of experience in enterprise and solution architecture roles, preferably within BFSI or fintech Proven experience with mission-critical, scalable, and secure systems Strong communication and stakeholder management skills, including CXO interactions Demonstrated leadership in architecting complex enterprise products and managing teams of architects Ability to blend technical depth with business context to drive decisions Passion for innovation, engineering excellence, and architectural rigor Show more Show less
Posted 2 days ago
15.0 - 20.0 years
17 - 22 Lacs
Bengaluru
Work from Office
Project Role : Data Engineer Project Role Description : Design, develop and maintain data solutions for data generation, collection, and processing. Create data pipelines, ensure data quality, and implement ETL (extract, transform and load) processes to migrate and deploy data across systems. Must have skills : Databricks Unified Data Analytics Platform Good to have skills : NAMinimum 7.5 year(s) of experience is required Educational Qualification : 15 years full time education Summary :As a Data Engineer, you will design, develop, and maintain data solutions that facilitate data generation, collection, and processing. Your typical day will involve creating data pipelines, ensuring data quality, and implementing ETL processes to migrate and deploy data across various systems. You will collaborate with cross-functional teams to understand their data needs and provide effective solutions, ensuring that the data infrastructure is robust and scalable to meet the demands of the organization. Roles & Responsibilities:- Expected to be an SME.- Collaborate and manage the team to perform.- Responsible for team decisions.- Engage with multiple teams and contribute on key decisions.- Provide solutions to problems for their immediate team and across multiple teams.- Mentor junior team members to enhance their skills and knowledge in data engineering.- Continuously evaluate and improve data processes to enhance efficiency and effectiveness. Professional & Technical Skills: - Must To Have Skills: Proficiency in Databricks Unified Data Analytics Platform.- Experience with data pipeline orchestration tools such as Apache Airflow or similar.- Strong understanding of ETL processes and data warehousing concepts.- Familiarity with cloud platforms like AWS, Azure, or Google Cloud.- Knowledge of programming languages such as Python or Scala for data manipulation. Additional Information:- The candidate should have minimum 7.5 years of experience in Databricks Unified Data Analytics Platform.- This position is based at our Bengaluru office.- A 15 years full time education is required. Qualification 15 years full time education
Posted 2 days ago
15.0 - 20.0 years
17 - 22 Lacs
Bengaluru
Work from Office
Project Role : Data Engineer Project Role Description : Design, develop and maintain data solutions for data generation, collection, and processing. Create data pipelines, ensure data quality, and implement ETL (extract, transform and load) processes to migrate and deploy data across systems. Must have skills : Databricks Unified Data Analytics Platform Good to have skills : NAMinimum 7.5 year(s) of experience is required Educational Qualification : 15 years full time education Summary :As a Data Engineer, you will design, develop, and maintain data solutions that facilitate data generation, collection, and processing. Your typical day will involve creating data pipelines, ensuring data quality, and implementing ETL processes to migrate and deploy data across various systems. You will collaborate with cross-functional teams to understand their data needs and provide effective solutions, ensuring that the data infrastructure is robust and scalable to meet the demands of the organization. Roles & Responsibilities:- Expected to be an SME.- Collaborate and manage the team to perform.- Responsible for team decisions.- Engage with multiple teams and contribute on key decisions.- Provide solutions to problems for their immediate team and across multiple teams.- Mentor junior team members to enhance their skills and knowledge in data engineering.- Continuously evaluate and improve data processes to enhance efficiency and effectiveness. Professional & Technical Skills: - Must To Have Skills: Proficiency in Databricks Unified Data Analytics Platform.- Experience with data pipeline orchestration tools such as Apache Airflow or similar.- Strong understanding of ETL processes and data warehousing concepts.- Familiarity with cloud platforms like AWS, Azure, or Google Cloud.- Knowledge of programming languages such as Python or Scala for data manipulation. Additional Information:- The candidate should have minimum 7.5 years of experience in Databricks Unified Data Analytics Platform.- This position is based at our Bengaluru office.- A 15 years full time education is required. Qualification 15 years full time education
Posted 2 days ago
15.0 - 20.0 years
17 - 22 Lacs
Chennai
Work from Office
Project Role : Data Engineer Project Role Description : Design, develop and maintain data solutions for data generation, collection, and processing. Create data pipelines, ensure data quality, and implement ETL (extract, transform and load) processes to migrate and deploy data across systems. Must have skills : PySpark Good to have skills : NAMinimum 5 year(s) of experience is required Educational Qualification : 15 years full time education Summary :As a Data Engineer, you will design, develop, and maintain data solutions that facilitate data generation, collection, and processing. Your typical day will involve creating data pipelines, ensuring data quality, and implementing ETL processes to migrate and deploy data across various systems. You will collaborate with cross-functional teams to understand their data needs and provide effective solutions, ensuring that the data infrastructure is robust and scalable to meet the demands of the organization. Roles & Responsibilities:- Expected to be an SME.- Collaborate and manage the team to perform.- Responsible for team decisions.- Engage with multiple teams and contribute on key decisions.- Provide solutions to problems for their immediate team and across multiple teams.- Mentor junior team members to enhance their skills and knowledge in data engineering.- Continuously evaluate and improve data processes to enhance efficiency and effectiveness. Professional & Technical Skills: - Must To Have Skills: Proficiency in PySpark.- Good To Have Skills: Experience with Apache Kafka.- Strong understanding of data warehousing concepts and architecture.- Familiarity with cloud platforms such as AWS or Azure.- Experience in SQL and NoSQL databases for data storage and retrieval. Additional Information:- The candidate should have minimum 5 years of experience in PySpark.- This position is based in Chennai.- A 15 years full time education is required. Qualification 15 years full time education
Posted 2 days ago
15.0 - 20.0 years
17 - 22 Lacs
Chennai
Work from Office
Project Role : Data Engineer Project Role Description : Design, develop and maintain data solutions for data generation, collection, and processing. Create data pipelines, ensure data quality, and implement ETL (extract, transform and load) processes to migrate and deploy data across systems. Must have skills : PySpark Good to have skills : NAMinimum 5 year(s) of experience is required Educational Qualification : 15 years full time education Summary :As a Data Engineer, you will design, develop, and maintain data solutions that facilitate data generation, collection, and processing. Your typical day will involve creating data pipelines, ensuring data quality, and implementing ETL processes to migrate and deploy data across various systems. You will collaborate with cross-functional teams to understand their data needs and provide effective solutions, ensuring that the data infrastructure is robust and scalable to meet the demands of the organization. Roles & Responsibilities:- Expected to be an SME.- Collaborate and manage the team to perform.- Responsible for team decisions.- Engage with multiple teams and contribute on key decisions.- Provide solutions to problems for their immediate team and across multiple teams.- Mentor junior team members to enhance their skills and knowledge in data engineering.- Continuously evaluate and improve data processes to enhance efficiency and effectiveness. Professional & Technical Skills: - Must To Have Skills: Proficiency in PySpark.- Good To Have Skills: Experience with Apache Kafka.- Strong understanding of data warehousing concepts and architecture.- Familiarity with cloud platforms such as AWS or Azure.- Experience in SQL and NoSQL databases for data storage and retrieval. Additional Information:- The candidate should have minimum 5 years of experience in PySpark.- This position is based in Chennai.- A 15 years full time education is required. Qualification 15 years full time education
Posted 2 days ago
15.0 - 20.0 years
17 - 22 Lacs
Pune
Work from Office
Project Role : Data Engineer Project Role Description : Design, develop and maintain data solutions for data generation, collection, and processing. Create data pipelines, ensure data quality, and implement ETL (extract, transform and load) processes to migrate and deploy data across systems. Must have skills : Databricks Unified Data Analytics Platform Good to have skills : NAMinimum 7.5 year(s) of experience is required Educational Qualification : 15 years full time education Summary :As a Data Engineer, you will design, develop, and maintain data solutions that facilitate data generation, collection, and processing. Your typical day will involve creating data pipelines, ensuring data quality, and implementing ETL processes to migrate and deploy data across various systems. You will collaborate with cross-functional teams to understand their data needs and provide effective solutions, ensuring that the data infrastructure is robust and scalable to meet the demands of the organization. Roles & Responsibilities:- Expected to be an SME.- Collaborate and manage the team to perform.- Responsible for team decisions.- Engage with multiple teams and contribute on key decisions.- Provide solutions to problems for their immediate team and across multiple teams.- Mentor junior team members to enhance their skills and knowledge in data engineering.- Continuously evaluate and improve data processes to enhance efficiency and effectiveness. Professional & Technical Skills: - Must To Have Skills: Proficiency in Databricks Unified Data Analytics Platform.- Experience with data pipeline orchestration tools such as Apache Airflow or similar.- Strong understanding of ETL processes and data warehousing concepts.- Familiarity with cloud platforms like AWS, Azure, or Google Cloud.- Knowledge of programming languages such as Python or Scala for data manipulation. Additional Information:- The candidate should have minimum 7.5 years of experience in Databricks Unified Data Analytics Platform.- This position is based at our Pune office.- A 15 years full time education is required. Qualification 15 years full time education
Posted 2 days ago
15.0 - 20.0 years
17 - 22 Lacs
Pune
Work from Office
Project Role : Data Engineer Project Role Description : Design, develop and maintain data solutions for data generation, collection, and processing. Create data pipelines, ensure data quality, and implement ETL (extract, transform and load) processes to migrate and deploy data across systems. Must have skills : Databricks Unified Data Analytics Platform Good to have skills : NAMinimum 7.5 year(s) of experience is required Educational Qualification : 15 years full time education Summary :As a Data Engineer, you will design, develop, and maintain data solutions that facilitate data generation, collection, and processing. Your typical day will involve creating data pipelines, ensuring data quality, and implementing ETL processes to migrate and deploy data across various systems. You will collaborate with cross-functional teams to understand their data needs and provide effective solutions, ensuring that the data infrastructure is robust and scalable to meet the demands of the organization. Roles & Responsibilities:- Expected to be an SME.- Collaborate and manage the team to perform.- Responsible for team decisions.- Engage with multiple teams and contribute on key decisions.- Provide solutions to problems for their immediate team and across multiple teams.- Mentor junior team members to enhance their skills and knowledge in data engineering.- Continuously evaluate and improve data processes to enhance efficiency and effectiveness. Professional & Technical Skills: - Must To Have Skills: Proficiency in Databricks Unified Data Analytics Platform.- Experience with data pipeline orchestration tools such as Apache Airflow or similar.- Strong understanding of ETL processes and data warehousing concepts.- Familiarity with cloud platforms like AWS, Azure, or Google Cloud.- Knowledge of programming languages such as Python or Scala for data manipulation. Additional Information:- The candidate should have minimum 7.5 years of experience in Databricks Unified Data Analytics Platform.- This position is based at our Pune office.- A 15 years full time education is required. Qualification 15 years full time education
Posted 2 days ago
15.0 - 20.0 years
17 - 22 Lacs
Pune
Work from Office
Project Role : Data Engineer Project Role Description : Design, develop and maintain data solutions for data generation, collection, and processing. Create data pipelines, ensure data quality, and implement ETL (extract, transform and load) processes to migrate and deploy data across systems. Must have skills : PySpark Good to have skills : NAMinimum 5 year(s) of experience is required Educational Qualification : 15 years full time education Summary :As a Data Engineer, you will design, develop, and maintain data solutions that facilitate data generation, collection, and processing. Your typical day will involve creating data pipelines, ensuring data quality, and implementing ETL processes to effectively migrate and deploy data across various systems. You will collaborate with cross-functional teams to understand their data needs and provide innovative solutions to enhance data accessibility and usability. Roles & Responsibilities:- Expected to be an SME.- Collaborate and manage the team to perform.- Responsible for team decisions.- Engage with multiple teams and contribute on key decisions.- Provide solutions to problems for their immediate team and across multiple teams.- Mentor junior team members to enhance their skills and knowledge in data engineering.- Continuously evaluate and improve data processes to ensure efficiency and effectiveness. Professional & Technical Skills: - Must To Have Skills: Proficiency in PySpark.- Good To Have Skills: Experience with Apache Kafka, Apache Airflow, and cloud platforms such as AWS or Azure.- Strong understanding of data modeling and database design principles.- Experience with SQL and NoSQL databases for data storage and retrieval.- Familiarity with data warehousing concepts and tools. Additional Information:- The candidate should have minimum 5 years of experience in PySpark.- This position is based in Pune.- A 15 years full time education is required. Qualification 15 years full time education
Posted 2 days ago
15.0 - 20.0 years
17 - 22 Lacs
Hyderabad
Work from Office
Project Role : Data Engineer Project Role Description : Design, develop and maintain data solutions for data generation, collection, and processing. Create data pipelines, ensure data quality, and implement ETL (extract, transform and load) processes to migrate and deploy data across systems. Must have skills : Databricks Unified Data Analytics Platform Good to have skills : NAMinimum 7.5 year(s) of experience is required Educational Qualification : 15 years full time education Summary :As a Data Engineer, you will design, develop, and maintain data solutions that facilitate data generation, collection, and processing. Your typical day will involve creating data pipelines, ensuring data quality, and implementing ETL processes to migrate and deploy data across various systems. You will collaborate with cross-functional teams to understand their data needs and provide effective solutions, ensuring that the data infrastructure is robust and scalable to meet the demands of the organization. Roles & Responsibilities:- Expected to be an SME.- Collaborate and manage the team to perform.- Responsible for team decisions.- Engage with multiple teams and contribute on key decisions.- Provide solutions to problems for their immediate team and across multiple teams.- Mentor junior team members to enhance their skills and knowledge in data engineering.- Continuously evaluate and improve data processes to enhance efficiency and effectiveness. Professional & Technical Skills: - Must To Have Skills: Proficiency in Databricks Unified Data Analytics Platform.- Experience with data pipeline orchestration tools such as Apache Airflow or similar.- Strong understanding of ETL processes and data warehousing concepts.- Familiarity with cloud platforms like AWS, Azure, or Google Cloud.- Knowledge of programming languages such as Python or Scala for data manipulation. Additional Information:- The candidate should have minimum 7.5 years of experience in Databricks Unified Data Analytics Platform.- This position is based at our Hyderabad office.- A 15 years full time education is required. Qualification 15 years full time education
Posted 2 days ago
15.0 - 20.0 years
17 - 22 Lacs
Hyderabad
Work from Office
Project Role : Data Engineer Project Role Description : Design, develop and maintain data solutions for data generation, collection, and processing. Create data pipelines, ensure data quality, and implement ETL (extract, transform and load) processes to migrate and deploy data across systems. Must have skills : Databricks Unified Data Analytics Platform Good to have skills : NAMinimum 7.5 year(s) of experience is required Educational Qualification : 15 years full time education Summary :As a Data Engineer, you will design, develop, and maintain data solutions that facilitate data generation, collection, and processing. Your typical day will involve creating data pipelines, ensuring data quality, and implementing ETL processes to migrate and deploy data across various systems. You will collaborate with cross-functional teams to understand their data needs and provide effective solutions, ensuring that the data infrastructure is robust and scalable to meet the demands of the organization. Roles & Responsibilities:- Expected to be an SME.- Collaborate and manage the team to perform.- Responsible for team decisions.- Engage with multiple teams and contribute on key decisions.- Provide solutions to problems for their immediate team and across multiple teams.- Mentor junior team members to enhance their skills and knowledge in data engineering.- Continuously evaluate and improve data processes to enhance efficiency and effectiveness. Professional & Technical Skills: - Must To Have Skills: Proficiency in Databricks Unified Data Analytics Platform.- Experience with data pipeline orchestration tools such as Apache Airflow or similar.- Strong understanding of ETL processes and data warehousing concepts.- Familiarity with cloud platforms like AWS, Azure, or Google Cloud.- Knowledge of programming languages such as Python or Scala for data manipulation. Additional Information:- The candidate should have minimum 7.5 years of experience in Databricks Unified Data Analytics Platform.- This position is based at our Hyderabad office.- A 15 years full time education is required. Qualification 15 years full time education
Posted 2 days ago
4.0 - 6.0 years
6 - 10 Lacs
Chennai
Work from Office
Description Ciklum is looking for a Full-Stack Java Engineer to join our team full-time in India. We are a custom product engineering company that supports both multinational organizations and scaling startups to solve their most complex business challenges. With a global team of over 4, 000 highly skilled developers, consultants, analysts and product owners, we engineer technology that redefines industries and shapes the way people live. About the role: As a Full-Stack Java Engineer, become a part of a high-performing engineering team focused on delivering robust, scalable applications in the financial technology domain. This role offers the opportunity to work across backend services and frontend interfaces, contributing to mission-critical systems in a dynamic Agile environment. Responsibilities Design, develop, and maintain microservices using Java (11+), Spring Boot, REST APIs, and SOAP Web Services Build and enhance user-facing components using JSF and other Java-based UI frameworks Integrate applications with databases (primarily Oracle) and support deployment pipelines using Maven, Jenkins, and Tomcat Ensure code quality through unit testing, code reviews, and best development practices Collaborate with cross-functional teams (product, QA, DevOps) to deliver end-to-end features Participate in Agile ceremonies including backlog refinement, sprint planning, and retrospectives Assist in troubleshooting and resolving issues across development and production environments Requirements 4-6 years of experience in full-stack or backend-focused Java development Backend Expertise: Java (11+), Spring Boot, REST APIs, SOAP Web Services, Oracle, Maven, Jenkins, Tomcat Frontend Proficiency: Experience with JSF or other component-based Java UI frameworks Familiarity with microservices architecture and CI/CD workflows Strong analytical and debugging skills Solid understanding of SDLC, version control (Git), and Agile methodologies Excellent communication and collaboration abilities Desirable Experience with Spring WebFlux, Kafka, MongoDB, Drools, Apache Freemarker, or Netty Exposure to cloud platforms such as AWS, Azure, or PCF (Pivotal Cloud Foundry) Hands-on experience working in a SAFe Agile setup or large-scale Agile programs Whats in it for you Care: your mental and physical health is our priority. We ensure comprehensive company-paid medical insurance, as well as financial and legal consultation Tailored education path: boost your skills and knowledge with our regular internal events (meetups, conferences, workshops), Udemy licence, language courses and company-paid certifications Growth environment: share your experience and level up your expertise with a community of skilled professionals, locally and globally Flexibility: hybrid work mode at Chennai or Pune Opportunities: we value our specialists and always find the best options for them. Our Resourcing Team helps change a project if needed to help you grow, excel professionally and fulfil your potential Global impact: work on large-scale projects that redefine industries with international and fast-growing clients Welcoming environment: feel empowered with a friendly team, open-door policy, informal atmosphere within the company and regular team-building events
Posted 2 days ago
6.0 - 11.0 years
11 - 16 Lacs
Noida
Work from Office
Assist in the upgrade of Java from version 8 to 17 and Spring Boot from version 1. 5 to 3. 2. - Develop, test, and maintain high-quality code using Java and Spring Boot. - Write unit tests using JUnit and Mockito. - Rich experience with Apache Maven as build tool - Strong experience with Java 17. - Hands-on experience with Spring Boot 3. 2. - Experience in JUnit and Mockito for unit testing. - Familiarity with RESTful APIs and microservices architecture. Responsibilities - Participate in code reviews and contribute to the development of best practices. - Collaborate with senior developers and other team members to deliver software solutions. - Troubleshoot and debug applications. - Good understanding of software development best practices. - Strong analytical and problem-solving skills. - Good communication and teamwork skills. ** Good to have **: - Experience with circuit breaker patterns and caching techniques. - Experience with Docker. - Knowledge of CI/CD pipelines. - Familiarity with Agile methodologies. Mandatory Competencies Java - Core JAVA Fundamental Technical Skills - Spring Framework/Hibernate/Junit etc. Fundamental Technical Skills - OOPS/Design Beh - Communication Beh - Communication and collaboration Database - SQL Others - Micro services Java Others - Spring Boot At Iris Software, we offer world-class benefits designed to support the financial, health and well-being needs of our associates to help achieve harmony between their professional and personal growth. From comprehensive health insurance and competitive salaries to flexible work arrangements and ongoing learning opportunities, were committed to providing a supportive and rewarding work environment. Join us and experience the difference of working at a company that values its employees success and happiness.
Posted 2 days ago
6.0 - 8.0 years
5 - 9 Lacs
Hyderabad
Work from Office
Job : Data Engineer Consultant Jobs in Hyderabad (J49138)- Job in Hyderabad Data Engineer Consultant (Job Code : J49138) Job Summary 6 - 8 Years Data Engineer Consultant BE-Comp/IT, BE-Other, BTech-Comp/IT, BTech-Other, MCA Stream of Study: Computer Science/IT IT-Software/Software Services IT Software - Client Server Key Skills: Job Post Date: Sunday, June 15, 2025 Company Description Our client is a global knowledge practice that provides consulting, technology, engineering, management and innovation services to leading businesses, governments, non-governmental organizations and not-for-profits. We focus on gaining, refining and sharing expertise in the energy and utility sector, then provide strategic advice and implement outcome-driven solutions. Working with customers across the utility value chain, we deliver sustainable and lasting improvements to their efficiency and performance, adding value to their bottom line. Demand for power, gas and water is consistently growing as the population of the planet expands. Our goal is to support large consumers of energy and water, and improve the sustainability of resources by increasing efficiency and optimizing existing operations. We also develop commercially successful ways to use renewable resources which deliver transformative advantages for our customers. As demand grows so does opportunity, something Enzen has seized on since its inception in 2006. The business has grown and developed across the globe, with a physical presence in the UK, India, Australia, USA, Spain, Turkey, Middle-East, Africa and Kazakhstan. As we push into the second decade of the company, we`re super-charging our growth by delivering exceptional value and results to our customers. For people with the right mindset, the opportunity to develop and grow in the organization has never been greater. Our expanding solutions, services and geographies mean we`re always on the lookout for individuals who can drive positive change and are hungry for the success and rewards that go with it. Job Description "Strong programming and scripting skills in SQL and Python. Experience with data pipeline tools (e. g. , Apache Airflow, Azure Data Factory, AWS Glue). Hands-on with cloud-based data platforms such as Azure, AWS. Familiarity with data modeling and warehousing concepts (e. g. , star schema, snowflake). "
Posted 2 days ago
3.0 - 7.0 years
9 - 14 Lacs
Pune
Work from Office
Some careers shine brighter than others. If you re looking for a career that will help you stand out, join HSBC and fulfil your potential. Whether you want a career that could take you to the top, or simply take you in an exciting new direction, HSBC offers opportunities, support and rewards that will take you further. HSBC is one of the largest banking and financial services organisations in the world, with operations in 64 countries and territories. We aim to be where the growth is, enabling businesses to thrive and economies to prosper, and, ultimately, helping people to fulfil their hopes and realise their ambitions. We are currently seeking an experienced professional to join our team in the role of Senior Consultant Specialist In this role will be responsible for the below: The Senior Data Engineer will be responsible for designing, building, and managing the data infrastructure and data pipeline processes for the bank. This role involves leading a team of data engineers, working closely with data scientists, analysts, and IT professionals to ensure data is accessible, reliable, and secure. The ideal candidate will have a strong background in data engineering, excellent leadership skills, and a thorough understanding of the banking industrys data requirements. Leadership and Team Management . Lead, mentor, and develop a team of data engineers. Establish best practices for data engineering and ensure team adherence. Coordinate with other IT teams, business units, and stakeholders. Data Pipelines Integration and Management: Design and implement scalable data architectures to support the banks data needs. Develop and maintain ETL (Extract, Transform, Load) processes. Ensure the data infrastructure is reliable, scalable, and secure. Oversee the integration of diverse data sources into a cohesive data platform. Ensure data quality, data governance, and compliance with regulatory requirements. Develop and enforce data security policies and procedures. Monitor and optimize data pipeline performance. Troubleshoot and resolve data-related issues promptly. Implement monitoring and alerting systems for data processes Requirements To be successful in this role, you should meet the following requirements: (Must have Requirements) 8+ years of experience in data engineering or related field Strong experience with database technologies (SQL, NoSQL), data warehousing solutions, and big data technologies (Hadoop, Spark) Proficiency in programming languages such as Python, Java, or Scala. Experience with cloud platforms (AWS, Azure, Google Cloud) and their data services. Deep understanding of ETL processes and data pipeline orchestration tools (Airflow, Apache NiFi). Knowledge of data modeling, data warehousing concepts, and data integration techniques. Strong problem-solving skills and ability to work under pressure. Excellent communication and interpersonal skills. Experience in the banking or financial services industry. Familiarity with regulatory requirements related to data security and privacy in the banking sector. Certifications in cloud platforms (AWS Certified Data Analytics, Google Professional Data Engineer, etc. ). Experience with machine learning and data science frameworks Location : Pune and Bangalore
Posted 2 days ago
5.0 - 10.0 years
7 Lacs
Hyderabad
Work from Office
Project Role : Application Developer Project Role Description : Design, build and configure applications to meet business process and application requirements. Must have skills : Apache JMeter Good to have skills : NAMinimum 5 year(s) of experience is required Educational Qualification : 15 years full time education Summary :As an Application Developer, you will be involved in designing, building, and configuring applications to meet business process and application requirements. Your typical day will revolve around creating innovative solutions to address various business needs and ensuring seamless application functionality. Roles & Responsibilities:- Expected to be an SME- Collaborate and manage the team to perform- Responsible for team decisions- Engage with multiple teams and contribute on key decisions- Provide solutions to problems for their immediate team and across multiple teams- Lead the application development process- Conduct code reviews and ensure coding standards are met- Implement best practices for application design and development Professional & Technical Skills: - Must To Have Skills: Proficiency in Apache JMeter- Strong understanding of performance testing methodologies- Experience in load testing and stress testing- Knowledge of scripting languages for test automation- Familiarity with performance monitoring tools Additional Information:- The candidate should have a minimum of 5 years of experience in Apache JMeter- This position is based at our Hyderabad office- A 15 years full-time education is required Qualification 15 years full time education
Posted 2 days ago
15.0 - 20.0 years
17 - 22 Lacs
Gurugram
Work from Office
Project Role : Application Lead Project Role Description : Lead the effort to design, build and configure applications, acting as the primary point of contact. Must have skills : Spring Boot Good to have skills : Apache SparkMinimum 5 year(s) of experience is required Educational Qualification : 15 years full time education Summary :As an Application Lead, you will lead the effort to design, build, and configure applications, acting as the primary point of contact. Your typical day will involve collaborating with various teams to ensure project milestones are met, facilitating discussions to address challenges, and guiding your team through the development process. You will also engage in strategic planning to align application development with business objectives, ensuring that the solutions provided are effective and efficient. Your role will require you to stay updated with industry trends and best practices to enhance the overall performance of the applications being developed. Roles & Responsibilities:- Expected to be an SME.- Collaborate and manage the team to perform.- Responsible for team decisions.- Engage with multiple teams and contribute on key decisions.- Provide solutions to problems for their immediate team and across multiple teams.- Facilitate knowledge sharing sessions to enhance team capabilities.- Monitor project progress and ensure adherence to timelines and quality standards. Professional & Technical Skills: - Must To Have Skills: Proficiency in Spring Boot.- Good To Have Skills: Experience with Apache Spark.- Strong understanding of microservices architecture and RESTful APIs.- Experience with cloud platforms such as AWS or Azure.- Familiarity with containerization technologies like Docker and Kubernetes. The candidate should have minimum 5 years of experience in Spring Boot. Additional Information:- This position is based at our Gurugram office.- A 15 years full time education is required. Qualification 15 years full time education
Posted 2 days ago
6.0 years
0 Lacs
Gurugram, Haryana, India
Remote
Job Title: Senior Java Developer (Remote) Experience: 6 to 8 Years Location: Remote Employment Type: Full-Time Notice period: Immediate Joiner Job Summary: We are looking for a highly skilled and experienced Senior Java Developer to join our distributed team. The ideal candidate should have a strong background in developing scalable enterprise-grade applications using Java and related technologies, with exposure to full-stack development, system integration, and performance optimization. Key Responsibilities: Design and develop high-performance, scalable, and reusable Java-based applications. Build RESTful APIs with a strong understanding of RESTful architecture. Implement enterprise integration patterns using Apache Camel or Spring Integration. Ensure application security in compliance with OWASP guidelines. Write and maintain unit, integration, and BDD tests using JUnit, Cucumber, Selenium. Conduct performance and load testing; optimize through memory and thread dump analysis. Collaborate with product owners, QA teams, and other developers in Agile/Scrum environments. Participate in code reviews, architecture discussions, and mentoring junior developers. Technical Skills & Experience Required: Core Backend: Strong proficiency in Java (8 or higher) Proficient in Spring Boot , Spring Security , Spring MVC , Spring Data Solid experience with REST API design, implementation, and testing using Postman , SoapUI Unit Testing , Integration Testing , BDD Testing Web Services and Integration: Experience with XML , Web Services (RESTful and SOAP) , Apache CXF Knowledge of Enterprise Integration Patterns Exposure to Apache Camel or Spring Integration Frontend & Full Stack: Familiarity with HTML5 , CSS3 Experience with TypeScript , JavaScript , jQuery , Node.js Working knowledge of Webpack and Gulp Database & Data Streaming: Strong in RDBMS and Database Design (e.g., Oracle , PL/SQL ) Exposure to MongoDB and NoSQL Understanding of Kafka architecture , Kafka as a data streaming platform Performance & Security: Experience in Performance Analysis and Application Tuning Understanding of Security aspects and OWASP guidelines Experience with Memory & Thread Dump Analysis Cloud & DevOps: Working knowledge of Kubernetes Familiarity with Elastic solutions at the enterprise level Experience in Identity and Access Management tools like ForgeRock About IGT Solutions: IGT Solutions is a next-gen customer experience (CX) company, defining and delivering transformative experiences for the global and most innovative brands using digital technologies. With the combination of Digital and Human Intelligence, IGT becomes the preferred partner for managing end-to-end CX journeys across Travel and High Growth Tech industries. We have a global delivery footprint, spread across 30 delivery centers in China, Colombia, Egypt, India, Indonesia, Malaysia, Philippines, Romania, South Africa, Spain, UAE, the US, and Vietnam, with 25000+ CX and Technology experts from 35+ nationalities. IGT's Digital team collaborates closely with our customers business & technology teams to take solutions faster to market while sustaining quality while focusing on business value and improving overall end-Customer Experience. Our offerings include industry solutions as well as Digital services. We work with leading global enterprise customers to improve synergies between business & technology by enabling rapid business value realization leveraging Digital Technologies. These include lifecycle transformation & rapid development / technology solution delivery services delivered leveraging traditional as well as Digital Technologies, deep functional understanding and software engineering expertise. IGT is ISO 27001:2013, CMMI SVC Level 5 and ISAE-3402 compliant for IT, and COPC® Certified v6.0, ISO 27001:2013 and PCI DSS 3.2 certified for BPO processes. The organization follows Six Sigma rigor for process improvements. It is our policy to provide equal employment opportunities to all individuals based on job-related qualifications and ability to perform a job, without regard to age, gender, gender identity, sexual orientation, race, color, religion, creed, national origin, disability, genetic information, veteran status, citizenship or marital status, and to maintain a non-discriminatory environment free from intimidation, harassment or bias based upon these grounds. Show more Show less
Posted 2 days ago
4.0 - 9.0 years
6 - 11 Lacs
Kochi
Work from Office
As Data Engineer, you will develop, maintain, evaluate and test big data solutions. You will be involved in the development of data solutions using Spark Framework with Python or Scala on Hadoop and AWS Cloud Data Platform Responsibilities: Experienced in building data pipelines to Ingest, process, and transform data from files, streams and databases. Process the data with Spark, Python, PySpark, Scala, and Hive, Hbase or other NoSQL databases on Cloud Data Platforms (AWS) or HDFS Experienced in develop efficient software code for multiple use cases leveraging Spark Framework / using Python or Scala and Big Data technologies for various use cases built on the platform Experience in developing streaming pipelines Experience to work with Hadoop / AWS eco system components to implement scalable solutions to meet the ever-increasing data volumes, using big data/cloud technologies Apache Spark, Kafka, any Cloud computing etc Required education Bachelor's Degree Preferred education Master's Degree Required technical and professional expertise Minimum 4+ years of experience in Big Data technologies with extensive data engineering experience in Spark / Python or Scala ; Minimum 3 years of experience on Cloud Data Platforms on AWS; Experience in AWS EMR / AWS Glue / Data Bricks, AWS RedShift, DynamoDB Good to excellent SQL skills Exposure to streaming solutions and message brokers like Kafka technologies Preferred technical and professional experience Certification in AWS and Data Bricks or Cloudera Spark Certified developers
Posted 2 days ago
15.0 - 20.0 years
17 - 22 Lacs
Mumbai
Work from Office
Project Role : Application Developer Project Role Description : Design, build and configure applications to meet business process and application requirements. Must have skills : Apache Spark Good to have skills : Java Enterprise EditionMinimum 5 year(s) of experience is required Educational Qualification : 15 years full time education Summary :As an Application Developer, you will design, build, and configure applications to meet business process and application requirements. A typical day involves collaborating with team members to understand project needs, developing innovative solutions, and ensuring that applications are aligned with business objectives. You will engage in problem-solving activities, participate in team meetings, and contribute to the overall success of the projects you are involved in, ensuring that the applications you develop are efficient and effective in meeting user needs. Roles & Responsibilities:- Expected to be an SME.- Collaborate and manage the team to perform.- Responsible for team decisions.- Engage with multiple teams and contribute on key decisions.- Provide solutions to problems for their immediate team and across multiple teams.- Mentor junior team members to enhance their skills and knowledge.- Continuously evaluate and improve application performance and user experience. Professional & Technical Skills: - Must To Have Skills: Proficiency in Apache Spark.- Good To Have Skills: Experience with Java Enterprise Edition.- Strong understanding of distributed computing principles.- Experience with data processing frameworks and tools.- Familiarity with cloud platforms and services. Additional Information:- The candidate should have minimum 5 years of experience in Apache Spark.- This position is based in Mumbai.- A 15 years full time education is required. Qualification 15 years full time education
Posted 2 days ago
5.0 - 10.0 years
7 - 12 Lacs
Kochi
Work from Office
Develop user-friendly web applications using Java and React.js while ensuring high performance. Design, develop, test, and deploy robust and scalable applications. Building and consuming RESTful APIs. Collaborate with the design and development teams to translate UI/UX design wireframes into functional components. Optimize applications for maximum speed and scalability. Stay up-to-date with the latest Java and React.js trends, techniques, and best practices. Participate in code reviews to maintain code quality and ensure alignment with coding standards. Identify and address performance bottlenecks and other issues as they arise. Help us shape the future of Event Driven technologies, including contributing to Apache Kafka, Strimzi, Apache Flink, Vert.x and other relevant open-source projects. Collaborate within a dynamic team environment to comprehend and dissect intricate requirements for event processing solutions. Translate architectural blueprints into actualized code, employing your technical expertise to implement innovative and effective solutions. Conduct comprehensive testing of the developed solutions, ensuring their reliability, efficiency, and seamless integration. Provide ongoing support for the implemented applications, responding promptly to customer inquiries, resolving issues, and optimizing performance. Serve as a subject matter expert, sharing insights and best practices related to product development, fostering knowledge sharing within the team. Continuously monitor the evolving landscape of event-driven technologies, remaining updated on the latest trends and advancements. Collaborate closely with cross-functional teams, including product managers, designers, and developers, to ensure a holistic and harmonious product development process. Take ownership of technical challenges and lead your team to ensure successful delivery, using your problem-solving skills to overcome obstacles. Mentor and guide junior developers, nurturing their growth and development by providing guidance, knowledge transfer, and hands-on training. Engage in agile practices, contributing to backlog grooming, sprint planning, stand-ups, and retrospectives to facilitate effective project management and iteration. Foster a culture of innovation and collaboration, contributing to brainstorming sessions and offering creative ideas to push the boundaries of event processing solutions. Maintain documentation for the developed solutions, ensuring comprehensive and up-to-date records for future reference and knowledge sharing. Involve in building and orchestrating containerized services Required education Bachelor's Degree Preferred education Bachelor's Degree Required technical and professional expertise Proven 5+ years of experience as aFull stack developer (Java and React.js) with a strong portfolio of previous projects. Proficiency in Java, JavaScript, HTML, CSS, and related web technologies. Familiarity with RESTfulAPIs and their integration into applications. Knowledge of modern CICD pipelines and tools like Jenkinsand Travis. Strong understanding of version control systems, particularly Git. Good communication skills and the ability to articulate technical concepts to both technical and non-technical team members. Familiarity with containerizationand orchestrationtechnologies like Docker and Kubernetes for deploying event processing applications. Proficiency in troubleshootingand debugging. Exceptional problem-solving and analytical abilities, with a knack for addressing technical challenges. Ability to work collaboratively in an agile and fast-paced development environment. Leadership skills to guide and mentorjunior developers, fostering their growth and skill development. Strong organizational and time management skills to manage multiple tasks and priorities effectively. Adaptability to stay current with evolving event-driven technologies and industry trends. Customer-focused mindset, with a dedication to delivering solutions that meet or exceed customer expectations. Creative thinking and innovation mindset to drive continuous improvement and explore new possibilities. Collaborative and team-oriented approach to work, valuing open communication and diverse perspectives. Preferred technical and professional ex
Posted 2 days ago
10.0 - 13.0 years
25 - 40 Lacs
Gurugram
Work from Office
We're Nagarro. We are a Digital Product Engineering company that is scaling in a big way! We build products, services, and experiences that inspire, excite, and delight. We work at scale across all devices and digital mediums, and our people exist everywhere in the world (18000+ experts across 38 countries, to be exact). Our work culture is dynamic and non-hierarchical. We're looking for great new colleagues. That's where you come in! REQUIREMENTS: Total Experience 10+years. Strong working experience in Big Data Engineering. Expertise in AWS Glue (including Crawlers and Data Catalog). Hands-on experience in Python and PySpark for data engineering tasks. Strong working experience with Terraform and/or CloudFormation. Strong experience with Snowflake including data loading, transformations, and querying. Expertise in CI/CD pipelines, preferably using GitHub Actions. Strong working knowledge of AWS services like S3, SNS, Secret Manager, Athena, and Lambda. Familiarity with JIRA and GitHub for Agile project management and version control. Excellent problem-solving skills and ability to resolve complex functional issues independently. Strong documentation skills, including creation of configuration guides, test scripts, and user manuals. RESPONSIBILITIES: Writing and reviewing great quality code. Understanding the client's business use cases and technical requirements and be able to convert them into technical design which elegantly meets the requirements Mapping decisions with requirements and be able to translate the same to developers Identifying different solutions and being able to narrow down the best option that meets the clients' requirements Defining guidelines and benchmarks for NFR considerations during project implementation Writing and reviewing design document explaining overall architecture, framework, and high-level design of the application for the developers Reviewing architecture and design on various aspects like extensibility, scalability, security, design patterns, user experience, NFRs, etc., and ensure that all relevant best practices are followed Developing and designing the overall solution for defined functional and non-functional requirements; and defining technologies, patterns, and frameworks to materialize it. Understanding and relating technology integration scenarios and applying these learnings in projects
Posted 2 days ago
4.0 years
0 Lacs
Noida, Uttar Pradesh, India
On-site
We are seeking a talented Full Stack Developer experienced in Java, Kotlin, Spring Boot, Angular, and Apache Kafka to join our dynamic engineering team. The ideal candidate will design, develop, and maintain end-to-end web applications and real-time data processing solutions, leveraging modern frameworks and event-driven architectures. Location : Noida/Pune/Bangalore/Hyderabad/Chennai Timings : 2pm to 11pm Experience : 4-6 Years Key Responsibilities Design, develop, and maintain scalable web applications using Java, Kotlin, Spring Boot, and Angular. Build and integrate RESTful APIs and microservices to connect frontend and backend components. Develop and maintain real-time data pipelines and event-driven features using Apache Kafka. Collaborate with cross-functional teams (UI/UX, QA, DevOps, Product) to define, design, and deliver new features. Write clean, efficient, and well-documented code following industry best practices and coding standards. Participate in code reviews, provide constructive feedback, and ensure code quality and consistency. Troubleshoot and resolve application issues, bugs, and performance bottlenecks in a timely manner. Optimize applications for maximum speed, scalability, and security. Stay updated with the latest industry trends, tools, and technologies, and proactively suggest improvements. Participate in Agile/Scrum ceremonies and contribute to continuous integration and delivery pipelines. Required Qualifications Experience with cloud-based technologies and deployment (Azure, GCP). Familiarity with containerization (Docker, Kubernetes) and microservices architecture. Proven experience as a Full Stack Developer with hands-on expertise in Java, Kotlin, Spring Boot, and Angular (Angular 2+). Strong understanding of object-oriented and functional programming principles. Experience designing and implementing RESTful APIs and integrating them with frontend applications. Proficiency in building event-driven and streaming applications using Apache Kafka. Experience with database systems (SQL/NoSQL), ORM frameworks (e.g., Hibernate, JPA), and SQL. Familiarity with version control systems (Git) and CI/CD pipelines. Good understanding of HTML5, CSS3, JavaScript, and TypeScript. Experience with Agile development methodologies and working collaboratively in a team environment. Excellent problem-solving, analytical, and communication skills. Show more Show less
Posted 2 days ago
14.0 years
0 Lacs
Kolkata, West Bengal, India
On-site
General Information Job Role: Lead DevOps Engineer Functional Area: DevOps Job Location: Pan India Job Shift: General Indian/ UK Shift Education: B.Sc./ B.Tech/ B.E / MTech in Any Specialization Employment Type Full Time, Permanent About Unified Infotech Embark on a transformative journey with Unified Infotech, a beacon of innovation and excellence in the tech consulting and software development landscape for over 14 years. We are dedicated to designing custom, forward-thinking web, mobile, and software solutions for a diverse clientele, from burgeoning MSMEs to towering Enterprises. Our mission is to engineer products that not only solve complex challenges but also set new benchmarks in the digital realm. At Unified, a job is not simply a job. It is a pursuit of excellence, to build and create, to understand and consult, to imagine and be creative, to reformulate UX, to invent and redefine, to code for performance, to collaborate and communicate. Role Description We are seeking a highly skilled and motivated DevOps Lead with expertise in both AWS and Azure cloud platforms to join our dynamic team. The successful candidate will collaborate with solution architects, developers, project managers, customer technical teams, and internal stakeholders to drive results. Your primary focus will be ensuring seamless customer access to applications in the cloud, managing customer workload migrations, implementing robust backup policies, overseeing hybrid cloud deployments, and building solutions for service assurance with a strong emphasis on leveraging Azure's unique capabilities. Desired Experience Define architecture, design, implement, program manage, and lead technology teams in delivering complex technical solutions for our clients across both AWS and Azure platforms. Span across DevOps, Continuous Integration (CI), and Continuous Delivery (CD) areas, providing demonstrable implementation experience in shaping value-add consulting solutions. Deploy, automate, and maintain cloud infrastructure with leading public cloud vendors such as Amazon Web Services (AWS) and Microsoft Azure, with a keen focus on integrating Azure-specific services and tools. Set up backups, replications, archiving, and implement disaster recovery measures leveraging Azure's resilience and geo-redundancy features. Utilize Azure DevOps services for better collaboration, reporting, and increasing automation in the CI/CD pipelines. Job Requirements Detail-oriented with a holistic perspective on system architecture, including at least 1 year of hands-on experience with Azure cloud services. Strong shell scripting and Linux administration skills, with a deep understanding of Linux and virtualization. Expertise in server technologies like Apache, Nginx, and Node, including optimization experience. Knowledge of database technologies such as MySQL, Redis, and MongoDB, with proficiency in management, replication, and disaster recovery. Proven experience in medium to large-scale public cloud deployments on AWS and Azure, including the migration of complex, multi-tier applications to these platforms. In-depth working knowledge of AWS and Azure, showcasing the ability to leverage Azure-specific features such as Azure Active Directory, Azure Kubernetes Service (AKS), Azure Functions, and Azure Logic Apps. Familiarity with CI/CD, automation, and monitoring processes for production-level infrastructure, including the use of Azure Monitor and Azure Automation and third party . Practical experience in setting up full-stack monitoring solutions using Prometheus, Grafana, and Loki, including long-term storage, custom dashboard creation, alerting, and integration with Kubernetes clusters. Worked extensively with Azure Front Door , including custom routing, WAF policies, SSL/TLS certificate integration, and performance optimization for global traffic. Experienced in multi Ingress Controller architecture setup and management, including namespace-specific ingress deployments Hands-on experience in setting up, configuring, and managing Azure API Management (APIM) Deep understanding of system performance and the ability to analyze root causes using tools available in Azure. Experience with Azure-specific management and governance tools, such as Azure Policy, Azure Blueprints, and Azure Resource Manager (ARM) templates. Proficiency in CI/CD automation using tools like Jenkins, Travis CI, Circle CI, or Azure DevOps. Knowledge of security infrastructure and vulnerabilities, including Azure's security tools like Azure Security Center and Azure Sentinel. Capability to analyze costs for the entire infrastructure, including cost management and optimization in Azure environments. Hands-on experience with configuration management tools like Ansible, Puppet, Chef, or similar, with an emphasis on their integration in Azure environments. Experience with container orchestration tools such as Kubernetes, Docker Swarm, and Docker containers, with a preference for those proficient in Azure Kubernetes Service (AKS). Total Exp : 6+ Years Exp in Cloud : AWS 3+, Azure 1+ Years NP : Immediate to 30 days preferred. Show more Show less
Posted 2 days ago
0 years
0 Lacs
Bengaluru, Karnataka, India
On-site
Provide support for data production systems in Nielsen Technology International Media (Television and Radio audio measurement) playing a critical role in ensuring the reliability, scalability, and security. Configure, implement, and deploy audience measurement solutions. Provide expert-level support, leading infrastructure automation initiatives, driving continuous improvement across our DevOps practices and supporting Agile processes. Core Technologies: Linux, Airflow, Bash, CI/CD, AWS services (EC2, S3, RDS, EKS, VPC), PostgreSQL, Python, Kubernetes. Responsibilities: Architect, manage, and optimize scalable and secure cloud infrastructure (AWS) using Infrastructure as Code (Terraform, CloudFormation, Ansible). Implement and maintain robust CI/CD pipelines to streamline software deployment and infrastructure changes. Identify and implement cost-optimization strategies for cloud resources. Ensure the smooth operation of production systems across 30+ countries, providing expert-level troubleshooting and incident response. Manage cloud-related migration changes and updates, supporting the secure implementation of changes/fixes. Participate in 24/7 on-call rotation for emergency support. Key Skills: Proficiency in Linux OS, particularly Fedora and Debian-based distributions (AlmaLinux, Amazon Linux, Ubuntu). Strong proficiency in scripting languages (Bash, Python) and SQL. Knowledge of scripting languages (Bash, SQL). Versed in leveraging Automations / DevOps principles with understanding of CI/CD concepts Working knowledge of infrastructure-as-code tools like Terraform, CloudFormation, and Ansible. Solid experience with AWS core services (EC2, EKS, S3, RDS, VPC, IAM, Security Groups). Hands-on experience with Docker, Kubernetes for containerized workloads. Solid understanding of DevOps practices, including monitoring, security, and high-availability design. Hands-on experience with Apache Airflow for workflow automation and scheduling. Strong troubleshooting skills, with experience in resolving issues and handling incidents in production environments. Foundational understanding of modern networking principles and cloud network architectures. Show more Show less
Posted 2 days ago
Upload Resume
Drag or click to upload
Your data is secure with us, protected by advanced encryption.
Apache is a widely used software foundation that offers a range of open-source software solutions. In India, the demand for professionals with expertise in Apache tools and technologies is on the rise. Job seekers looking to pursue a career in Apache-related roles have a plethora of opportunities in various industries. Let's delve into the Apache job market in India to gain a better understanding of the landscape.
These cities are known for their thriving IT sectors and see a high demand for Apache professionals across different organizations.
The salary range for Apache professionals in India varies based on experience and skill level. - Entry-level: INR 3-5 lakhs per annum - Mid-level: INR 6-10 lakhs per annum - Experienced: INR 12-20 lakhs per annum
In the Apache job market in India, a typical career path may progress as follows: 1. Junior Developer 2. Developer 3. Senior Developer 4. Tech Lead 5. Architect
Besides expertise in Apache tools and technologies, professionals in this field are often expected to have skills in: - Linux - Networking - Database Management - Cloud Computing
As you embark on your journey to explore Apache jobs in India, it is essential to stay updated on the latest trends and technologies in the field. By honing your skills and preparing thoroughly for interviews, you can position yourself as a competitive candidate in the Apache job market. Stay motivated, keep learning, and pursue your dream career with confidence!
Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.
We have sent an OTP to your contact. Please enter it below to verify.