Get alerts for new jobs matching your selected skills, preferred locations, and experience range. Manage Job Alerts
10.0 - 15.0 years
10 - 15 Lacs
Pune, Maharashtra, India
On-site
The Role We are looking for an experienced Data Engineer to design and develop advanced data migration pipelines from traditional OLTP databases (e.g., Oracle) to modern big data platforms such as Cloudera and Databricks. The ideal candidate will possess expertise in technologies such as Python, Java, Spark, and NiFi, along with a proven track record in managing data pipelines for tasks including initial snapshot loading, building Change Data Capture (CDC) pipelines, exception management, reconciliation, data security, and retention. This role also demands proficiency in data modeling, cataloging, taxonomy creation, and ensuring robust data provenance and lineage to support governance and compliance requirements. Key Responsibilities Design, develop, and optimize data migration pipelines from OLTP databases like Oracle to big data platforms, including Cloudera CDP/CDH and Databricks. Build scalable ETL workflows using tools like Python, Scala, Apache Spark, and Apache NiFi to support initial snapshots, CDC, exception handling, and reconciliation processes. Implement data security measures, such as encryption, access controls, and compliance with data retention policies, across all migration pipelines. Develop and maintain data models, taxonomy structures, and cataloging systems to ensure logical organization and easy accessibility of data. Establish data lineage and provenance to ensure traceability and compliance with governance frameworks. Collaborate with cross-functional teams to understand data migration requirements, ensuring high-quality and timely delivery of solutions. Monitor and troubleshoot data pipelines to ensure performance, scalability, and reliability. Stay updated on emerging technologies in data engineering and big data ecosystems, proposing improvements to existing systems and processes. Required Skills And Qualifications 10+ years of experience in data engineering, with at least 2 years in a leadership or technical lead role. Proficiency in OLTP databases, particularly Oracle, and data egress techniques. Strong programming skills in Python, Scala and Java. Expertise in Apache Spark, Flink, Kafka and data integration tools like Apache NiFi. Hands-on experience with Cloudera Data Platform CDP/CDH, Apache Ozone Familiarity with cloud-based big data ecosystems such as AWS Databrick, S3, Glue etc Familiarity with patterns such as Medallion, data layers, datalake, datawarehouse, experience in building scalable ETL pipeline, optimizing data workflows, leveraging platforms to integrate transform, and store large datasets. Knowledge of data security best practices, including encryption, data masking, and role-based access control. Exceptional problem-solving and analytical abilities Strong communication and leadership skills, with the ability to navigate ambiguity and collaborate effectively across diverse teams.Optional Awareness on regulatory compliance requirements for data handling and privacy Education: Bachelor's or Master's degree in Computer Science
Posted 3 weeks ago
5.0 - 6.0 years
10 - 20 Lacs
Hyderabad, Chennai, Bengaluru
Hybrid
Name- Teradata Developer Rounds of discussion- 2 (1st Virtual and 2nd Face to face from Wipro Premises) Location- Bangalore/Chennai / Hyderabad only Exp- 4- 7 yrs JD: Experience in Teradata 16.x , SQL/BTEQ scripting, TPT, Informatica 10.x UNIX Shell scripting, Autosys, Cloudera CDP, Hadoop (Hive, Impala, PySpark, Scala) and Data analysis. Strong Data sourcing, data modelling and provisioning skills as related to support large scale AML monitoring and surveillance data assets. Database Executed database intensive development, data migrations and conversions. Knowledge and experience working with SQL performance tuning. A person who can handle all at once - Design/Build, Hands-on development, Co-ordination of development in distributed environment Very good at engaging Business stakeholders and apt at Business engagement and requirement discovery Good experience in SDLC, Agile, Continuous Integration Continuous Delivery, and change management (CICD ) - Jira, Bitbucket,Jenkins, Artifactory, Ansible Good knowledge of Unix and development of shell scripts is required Effective communication and listening skills Software development in Agile environment. Education: BS/MS in Computer Science, Engineering, or any quantitative discipline Role & responsibilities Preferred candidate profile
Posted 1 month ago
15.0 - 20.0 years
22 - 32 Lacs
pune
Hybrid
OverView We are seeking a senior solution Architect to lead the design, integration, and delivery of a large-scale network traffic and data flow system. This role is accountable for ensuring architectural integrity, zero-error tolerance, and robust fallback mechanisms across the entire solution lifecycle. The architect will oversee subscriber data capture, DPI, DR generation, Kafka integration, DWH ingestion. and secure API-based retrieval, ensuring compliance and security regulations. Roles & responsibilities Own the end-to-end architecture spanning subscriber traffic capture, DPI, DR generation, Kafka streaming, and data lake ingestion. Design and document system architecture, data flow diagrams, and integration blueprints across DPI and traffic classification systems, nProbe, Kafka. Spark, and Cloudera CDP Implement fallback and error-handling mechanisms to ensure zero data loss and high availability across all layers. Lead cross-functional collaboration with network engineers, Kafka developers. data platform teams, and security stakeholders. Ensure data govemance, encryption, and compliance using tools like Apache Ranger, Atlas, SDX, and HashiCorp Vault. Oversee API design and exposure for customer access, including advanced search, session correlation, and audit logging. Drive SIT/UAT planning, performance benchmarking, and production rollout readiness. Provide technical leadership across multiple vendors and internal teams, ensuring alignment with Business requirements and regulatory standards Preferred candidate profile Proven experience in telecom-grade architecture involving DPI, IPFIX/NefFlow, and subscriber metadata enrichment. Deep knowledge of Apache Kafka, Spark Structured Streaming, and Cloudera CDP (HDFS, Hive, Iceberg, Ranger). Experience integrating QPtobe with Kafka and downstream analyfics platforms. Strong understanding of QoE metrics, A/B party correlation, and application traffic classification. Expertise in RESTful API design, schema management (Avro/JSON), and secure data access protocols. Familiarity with network interfaces (Gn/Gi, Radius, DNS) and traffic filtering strategies. Experience implementing fallback mechanisms, error queues, and disaster recovery strategies. Excellent communication, documentation, and stakeholder management skills. Cloudera Certified Architect / Kafka Developer / AWS or GCP Solution Architect. Security certifications (e.g., CISSP, CISM) will be advantageous
Posted Date not available
Upload Resume
Drag or click to upload
Your data is secure with us, protected by advanced encryption.
Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.
We have sent an OTP to your contact. Please enter it below to verify.
Accenture
54024 Jobs | Dublin
Wipro
24262 Jobs | Bengaluru
Accenture in India
18733 Jobs | Dublin 2
EY
17079 Jobs | London
Uplers
12548 Jobs | Ahmedabad
IBM
11704 Jobs | Armonk
Amazon
11059 Jobs | Seattle,WA
Bajaj Finserv
10656 Jobs |
Accenture services Pvt Ltd
10587 Jobs |
Oracle
10506 Jobs | Redwood City