· 5 to 10 years of experience in designing and building data pipelines using Apache Spark, Databricks or equivalent bigdata frameworks. · Handson expertise with streaming and messaging systems such as Apache Kafka (publish subscribe architecture), Confluent Cloud, RabbitMQ or Azure Event Hub. Experience creating producers, consumers and topics and integrating them into downstream processing. · Deep understanding of relational databases and CDC. Proficiency in SQL Server, Oracle, or other RDBMSs; experience capturing change events using Debezium or native CDC tools and transforming them for downstream consumption. · Implement CDC and deduplication logic. Capture change events from source databases using Debezium, built-in CDC features of SQL Server/ Oracle or other connectors. Apply watermarking and drop duplicate strategies based on primary keys and event timestamps. · Proficiency in programming languages such as Python, Scala or Java and solid knowledge of SQL for data manipulation and transformation. · Cloud platform expertise. Experience with Azure or AWS services for data storage, compute, and orchestration (e.g., ADLS, S3, Azure Data Factory, AWS Glue, Airflow, DBX, DLT). · Data modelling and warehousing. Knowledge of data Lakehouse architectures, Delta Lake, partitioning strategies, and performance optimization. · Version control and DevOps. Familiarity with Git and CI/CD pipelines; ability to automate deployment and manage infrastructure as code. Strong problem solving and communication skills. Ability to work with cross functional teams and articulate complex technical concepts to nontechnical stakeholders. Work Location: Chennai - Tharamani / Pune - Kharadi Work Mode: Hybrid - 3 Days WFO Job Types: Full-time, Permanent Pay: ₹1,000,000.00 - ₹3,200,000.00 per year Benefits: Health insurance Provident Fund Work Location: In person