Posted:22 hours ago|
Platform:
On-site
Full Time
Design and architect scalable, high-performance, and secure data platforms (Data Warehouses, Data Lakes, Lakehouse) to support analytics, AI/ML, and reporting needs.
Leverage Big Data technologies such as PySpark, Hadoop, Kafka, and Databricks for large-scale data processing and transformation.
Develop and maintain conceptual, logical, and physical data models aligned with business requirements and governance standards.
Architect data integration pipelines to enable seamless data flow across diverse systems, including legacy and modern cloud platforms (Databricks, Snowflake, AWS, Azure).
Lead modernization and migration initiatives from on-premises databases (e.g., Informix, Oracle) to cloud-native platforms (AWS, Azure, Google Cloud).
Continuously optimize data pipelines and storage for performance, scalability, cost efficiency, and SLA adherence.
Partner with business development teams to provide technical leadership during pre-sales engagements, including client workshops and architecture discussions.
Collaborate with data engineers, data scientists, business analysts, and IT teams to ensure architecture alignment with business goals.
Define and enforce data governance frameworks, quality standards, and security policies to ensure compliance with organizational and regulatory requirements.
Evaluate emerging technologies and tools to define long-term data strategy and ensure future scalability.
Key Skills
· Data modeling (conceptual, logical, physical) for high-performance platforms
· Cloud platforms: AWS, Azure, GCP; Databricks, Snowflake
· Big Data architecture and distributed processing (PySpark, Hadoop, Kafka)
· Relational (SQL Server, Oracle, MySQL) and NoSQL (DynamoDB, CosmosDB)
· ETL/ELT pipeline design, automation, and system integration
· Cloud migration: legacy to cloud transitions
· Leveraging cloud platform tools for data engineering, pipeline automation, and analytics (e.g., AWS
Glue, Azure Data Factory, GCP Dataflow, Databricks, Snowflake, Talend)
· Data security, compliance, and governance
· Agile and cross-functional team collaboration
· Lakehouse architecture: combining data lake flexibility with warehouse performance
· AI/ML and Data Science platform integration for analytics, predictive modelling, and insights
Programming & scripting: SQL, Python & PySpark
Datacrew.ai
Upload Resume
Drag or click to upload
Your data is secure with us, protected by advanced encryption.
Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.
We have sent an OTP to your contact. Please enter it below to verify.
Practice Python coding challenges to boost your skills
Start Practicing Python Nowchennai, tamil nadu, india
Salary: Not disclosed
patna, bihar, india
Salary: Not disclosed
hyderābād
2.685 - 7.23 Lacs P.A.
bengaluru
5.7 - 10.5 Lacs P.A.
noida
7.2 - 9.38 Lacs P.A.
mumbai, maharashtra, india
Salary: Not disclosed
bengaluru, karnataka
Salary: Not disclosed
gurugram, haryana, india
Salary: Not disclosed
gurgaon, haryana, india
Salary: Not disclosed
pune, delhi / ncr, mumbai (all areas)
20.0 - 25.0 Lacs P.A.