Get alerts for new jobs matching your selected skills, preferred locations, and experience range. Manage Job Alerts
6.0 years
0 Lacs
hyderabad, telangana, india
On-site
Job Overview: We are looking for a skilled and motivated Data Engineer with strong experience in Python programming and Google Cloud Platform (GCP) to join our data engineering team. The ideal candidate will be responsible for designing, developing, and maintaining robust and scalable ETL (Extract, Transform, Load) data pipelines. The role involves working with various GCP services, implementing data ingestion and transformation logic, and ensuring data quality and consistency across systems. Key Responsibilities: Design, develop, test, and maintain scalable ETL data pipelines using Python. Work extensively on Google Cloud Platform (GCP) services such as: Dataflow for real-time and batch dat...
Posted 3 weeks ago
3.0 - 7.0 years
0 Lacs
hyderabad, telangana
On-site
As a Data Engineer at our company, you will be responsible for designing, developing, and maintaining scalable and efficient data pipelines using GCP services such as Dataflow, Cloud Composer (Airflow), and Pub/Sub. Your role will involve designing and implementing robust data models optimized for analytical and operational workloads within GCP data warehousing solutions like BigQuery. You will also be tasked with developing and implementing ETL processes to ingest, cleanse, transform, and load data from various sources into our data warehouse and other data stores on GCP. Furthermore, you will play a key role in building and managing data warehousing solutions on GCP, ensuring data integrit...
Posted 3 weeks ago
5.0 - 9.0 years
0 Lacs
ahmedabad, gujarat
On-site
As a Consultant Delivery (Data Engineer) at Worldline, you will be an integral part of the Data Management team, contributing to a significant Move to Cloud (M2C) project. Your primary focus will be migrating our data infrastructure to the cloud and enhancing our data pipelines for improved performance and scalability. You will have the opportunity to work on a critical initiative that plays a key role in the organization's digital transformation. To excel in this role, you should hold a Bachelor's or Master's degree in Computer Science, Engineering, or a related field, along with a minimum of 5 years of experience as a Data Engineer. Your expertise should include a strong emphasis on cloud-...
Posted 3 weeks ago
5.0 - 10.0 years
30 - 35 Lacs
noida, hyderabad, chennai
Hybrid
Design lakehouse architectures with GCP Stack Lead batch/streaming pipeline design Guide migration&platform integration Lead workshops,requirement gathering,HLD,LLD Deliver source-to-target mapping,lineage, catalog Provide governance, best practices
Posted 3 weeks ago
0 years
0 Lacs
gurgaon, haryana, india
On-site
Introduction A career in IBM Consulting is rooted by long-term relationships and close collaboration with clients across the globe. You'll work with visionaries across multiple industries to improve the hybrid cloud and AI journey for the most innovative and valuable companies in the world. Your ability to accelerate impact and make meaningful change for your clients is enabled by our strategic partner ecosystem and our robust technology platforms across the IBM portfolio Your Role And Responsibilities As an entry level Application Developer at IBM, you'll work with clients to co-create solutions to major real-world challenges by using best practice technologies, tools, techniques, and produ...
Posted 3 weeks ago
6.0 - 10.0 years
30 - 35 Lacs
bengaluru
Work from Office
We are seeking an experienced PySpark Developer / Data Engineer to design, develop, and optimize big data processing pipelines using Apache Spark and Python (PySpark). The ideal candidate should have expertise in distributed computing, ETL workflows, data lake architectures, and cloud-based big data solutions. Key Responsibilities: Develop and optimize ETL/ELT data pipelines using PySpark on distributed computing platforms (Hadoop, Databricks, EMR, HDInsight). Work with structured and unstructured data to perform data transformation, cleansing, and aggregation. Implement data lake and data warehouse solutions on AWS (S3, Glue, Redshift), Azure (ADLS, Synapse), or GCP (BigQuery, Dataflow). Op...
Posted 3 weeks ago
8.0 - 13.0 years
12 - 15 Lacs
hyderabad
Work from Office
We are seeking a Technical Architect specializing in Healthcare Data Analytics with expertise in Google Cloud Platform (GCP). The role involves designing and implementing data solutions tailored to healthcare analytics requirements. The ideal candidate will have experience in GCP tools like BigQuery, Dataflow, Dataprep, and Healthcare APIs, and should stay up to date with GCP updates. Knowledge of healthcare data standards, compliance requirements (e.g., HIPAA), and healthcare interoperability is essential. The role requires experience in microservices, containerization (Docker, Kubernetes), and programming languages like Python and Spark. The candidate will lead the implementation of data a...
Posted 3 weeks ago
9.0 - 12.0 years
12 - 20 Lacs
hyderabad, gachibowli
Work from Office
- Data Processing: BigQuery, Apache Spark, Hadoop, Dataflow - BI Tools: Tableau, Power BI, Looker - Languages: Python, SQL, Java, Scala - ETL Tools: Apache Nifi, Talend, Informatica, Dataform - Cloud: GCP (BigQuery, Dataflow, Pub/Sub, Cloud Storage) - Data Modeling: Kimball, Star Schema, Snowflake Schema - Version Control: Git, GitLab.
Posted 3 weeks ago
9.0 - 12.0 years
12 - 20 Lacs
hyderabad
Work from Office
- Data Processing: BigQuery, Apache Spark, Hadoop, Dataflow - BI Tools: Tableau, Power BI, Looker - Languages: Python, SQL, Java, Scala - ETL Tools: Apache Nifi, Talend, Informatica, Dataform - Cloud: GCP (BigQuery, Dataflow, Pub/Sub, Cloud Storage) - Data Modeling: Kimball, Star Schema, Snowflake Schema - Version Control: Git, GitLab.
Posted 3 weeks ago
9.0 - 12.0 years
12 - 20 Lacs
hyderabad, hitech city
Work from Office
- Data Processing: BigQuery, Apache Spark, Hadoop, Dataflow - BI Tools: Tableau, Power BI, Looker - Languages: Python, SQL, Java, Scala - ETL Tools: Apache Nifi, Talend, Informatica, Dataform - Cloud: GCP (BigQuery, Dataflow, Pub/Sub, Cloud Storage) - Data Modeling: Kimball, Star Schema, Snowflake Schema - Version Control: Git, GitLab.
Posted 3 weeks ago
9.0 - 12.0 years
12 - 20 Lacs
navi mumbai
Work from Office
- Data Processing: BigQuery, Apache Spark, Hadoop, Dataflow - BI Tools: Tableau, Power BI, Looker - Languages: Python, SQL, Java, Scala - ETL Tools: Apache Nifi, Talend, Informatica, Dataform - Cloud: GCP (BigQuery, Dataflow, Pub/Sub, Cloud Storage) - Data Modeling: Kimball, Star Schema, Snowflake Schema - Version Control: Git, GitLab.
Posted 3 weeks ago
9.0 - 12.0 years
12 - 20 Lacs
mumbai suburban
Work from Office
- Data Processing: BigQuery, Apache Spark, Hadoop, Dataflow - BI Tools: Tableau, Power BI, Looker - Languages: Python, SQL, Java, Scala - ETL Tools: Apache Nifi, Talend, Informatica, Dataform - Cloud: GCP (BigQuery, Dataflow, Pub/Sub, Cloud Storage) - Data Modeling: Kimball, Star Schema, Snowflake Schema - Version Control: Git, GitLab.
Posted 3 weeks ago
9.0 - 12.0 years
12 - 20 Lacs
mumbai
Work from Office
- Data Processing: BigQuery, Apache Spark, Hadoop, Dataflow - BI Tools: Tableau, Power BI, Looker - Languages: Python, SQL, Java, Scala - ETL Tools: Apache Nifi, Talend, Informatica, Dataform - Cloud: GCP (BigQuery, Dataflow, Pub/Sub, Cloud Storage) - Data Modeling: Kimball, Star Schema, Snowflake Schema - Version Control: Git, GitLab.
Posted 3 weeks ago
6.0 - 9.0 years
8 - 11 Lacs
hyderabad
Work from Office
- Data Processing: BigQuery, Apache Spark, Hadoop, Dataflow - BI Tools: Tableau, Power BI, Looker - Languages: Python, SQL, Java, Scala - ETL Tools: Apache Nifi, Talend, Informatica, Dataform - Cloud: GCP (BigQuery, Dataflow, Pub/Sub, Cloud Storage) - Data Modeling: Kimball, Star Schema, Snowflake Schema - Version Control: Git, GitLab
Posted 3 weeks ago
6.0 - 9.0 years
8 - 11 Lacs
navi mumbai
Work from Office
- Data Processing: BigQuery, Apache Spark, Hadoop, Dataflow - BI Tools: Tableau, Power BI, Looker - Languages: Python, SQL, Java, Scala - ETL Tools: Apache Nifi, Talend, Informatica, Dataform - Cloud: GCP (BigQuery, Dataflow, Pub/Sub, Cloud Storage) - Data Modeling: Kimball, Star Schema, Snowflake Schema - Version Control: Git, GitLab
Posted 3 weeks ago
6.0 - 9.0 years
8 - 11 Lacs
hyderabad, gachibowli
Work from Office
- Data Processing: BigQuery, Apache Spark, Hadoop, Dataflow - BI Tools: Tableau, Power BI, Looker - Languages: Python, SQL, Java, Scala - ETL Tools: Apache Nifi, Talend, Informatica, Dataform - Cloud: GCP (BigQuery, Dataflow, Pub/Sub, Cloud Storage) - Data Modeling: Kimball, Star Schema, Snowflake Schema - Version Control: Git, GitLab
Posted 3 weeks ago
6.0 - 9.0 years
8 - 11 Lacs
mumbai
Work from Office
- Data Processing: BigQuery, Apache Spark, Hadoop, Dataflow - BI Tools: Tableau, Power BI, Looker - Languages: Python, SQL, Java, Scala - ETL Tools: Apache Nifi, Talend, Informatica, Dataform - Cloud: GCP (BigQuery, Dataflow, Pub/Sub, Cloud Storage) - Data Modeling: Kimball, Star Schema, Snowflake Schema - Version Control: Git, GitLab
Posted 3 weeks ago
6.0 - 9.0 years
8 - 11 Lacs
hyderabad, hitech city
Work from Office
- Data Processing: BigQuery, Apache Spark, Hadoop, Dataflow - BI Tools: Tableau, Power BI, Looker - Languages: Python, SQL, Java, Scala - ETL Tools: Apache Nifi, Talend, Informatica, Dataform - Cloud: GCP (BigQuery, Dataflow, Pub/Sub, Cloud Storage) - Data Modeling: Kimball, Star Schema, Snowflake Schema - Version Control: Git, GitLab
Posted 3 weeks ago
6.0 - 9.0 years
8 - 11 Lacs
mumbai suburban
Work from Office
- Data Processing: BigQuery, Apache Spark, Hadoop, Dataflow - BI Tools: Tableau, Power BI, Looker - Languages: Python, SQL, Java, Scala - ETL Tools: Apache Nifi, Talend, Informatica, Dataform - Cloud: GCP (BigQuery, Dataflow, Pub/Sub, Cloud Storage) - Data Modeling: Kimball, Star Schema, Snowflake Schema - Version Control: Git, GitLab
Posted 3 weeks ago
3.0 years
4 - 7 Lacs
thiruvananthapuram
On-site
Trivandrum India Technology Full time 8/26/2025 J00170006 Equifax is where you can power your possible. If you want to achieve your true potential, chart new paths, develop new skills, collaborate with bright minds, and make a meaningful impact, we want to hear from you. Equifax is seeking creative, high-energy and driven software engineers with hands-on development skills to work on a variety of meaningful projects. Our software engineering positions provide you the opportunity to join a team of talented engineers working with leading-edge technology. You are ideal for this position if you are a forward-thinking, committed, and enthusiastic software engineer who is passionate about technolo...
Posted 3 weeks ago
3.0 years
0 Lacs
trivandrum, kerala, india
On-site
Equifax is seeking creative, high-energy and driven software engineers with hands-on development skills to work on a variety of meaningful projects. Our software engineering positions provide you the opportunity to join a team of talented engineers working with leading-edge technology. You are ideal for this position if you are a forward-thinking, committed, and enthusiastic software engineer who is passionate about technology. What You’ll Do Develop applications with a focus on operational excellence, security, and scalability, including microservices and dataflow jobs. Apply modern software development practices, such as serverless computing, microservices architecture, and CI/CD. Write, d...
Posted 3 weeks ago
5.0 years
0 Lacs
india
On-site
Company Description ThreatXIntel is a startup cyber security company dedicated to providing customizable and affordable cyber security solutions for businesses and organizations. Our proactive approach involves continuous monitoring and testing to identify vulnerabilities before they can be exploited. We offer services such as cloud security, web and mobile security testing, and DevSecOps to help clients protect their digital assets. Role Description We are seeking experienced Data Engineers (freelance/contract) to support a large-scale enterprise migration of data warehouse assets and ETL pipelines from Teradata to Google Cloud Platform (GCP) . The role is hands-on and requires strong exper...
Posted 3 weeks ago
8.0 - 15.0 years
0 Lacs
chennai, tamil nadu
On-site
As a Data Architecture professional with 8-15 years of experience, your primary responsibility will be to design and implement data-centric solutions on Google Cloud Platform (GCP). You will utilize various GCP tools such as Big Query, Google Cloud Storage, Cloud SQL, Memory Store, Dataflow, Dataproc, Artifact Registry, Cloud Build, Cloud Run, Vertex AI, Pub/Sub, and GCP APIs to create efficient and scalable solutions. Your role will involve building ETL pipelines to ingest data from diverse sources into our system and developing data processing pipelines using programming languages like Java and Python for data extraction, transformation, and loading (ETL). You will be responsible for creat...
Posted 3 weeks ago
6.0 - 15.0 years
0 Lacs
pune, maharashtra
On-site
The role of Cloud Architecture and Engineering at Deutsche Bank as a Director based in Pune, India involves leading the global cloud infrastructure architecture for Corporate Bank Technology. With a focus on domains like Cash Management, Securities Services, Trade Finance, and Trust & Agency Services, you will be responsible for creating domain-level architecture roadmaps, designing solutions for TAS business, and providing technical leadership to development teams across multiple TAS Tribes and Corporate Bank Domains. Key Responsibilities: - Leading the global cloud infrastructure architecture across Corporate Bank domains - Creating domain level architecture roadmaps to ensure long-term bu...
Posted 3 weeks ago
10.0 years
0 Lacs
chennai, tamil nadu, india
On-site
Key Responsibilities Design, develop, and optimize robust data models (conceptual, logical, and physical) to support analytics and business reporting needs. Leverage Advanced SQL skills for efficient data querying, transformation, performance tuning, and optimization. Build and maintain data pipelines on Google Cloud Platform (GCP) leveraging native services and tools. Architect and implement scalable Data Warehousing (DWH) solutions for storing and analyzing large volumes of structured and semi-structured data. Integrate data from ERP systems such as SAP and other enterprise platforms into centralized repositories. Perform comprehensive data discovery, profiling, and validation to ensure da...
Posted 3 weeks ago
Upload Resume
Drag or click to upload
Your data is secure with us, protected by advanced encryption.
Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.
We have sent an OTP to your contact. Please enter it below to verify.
Accenture
79616 Jobs | Dublin
Wipro
28651 Jobs | Bengaluru
Accenture in India
23930 Jobs | Dublin 2
EY
21734 Jobs | London
Uplers
15782 Jobs | Ahmedabad
Bajaj Finserv
14649 Jobs |
IBM
14601 Jobs | Armonk
Amazon.com
13827 Jobs |
Accenture services Pvt Ltd
13693 Jobs |
Kotak Life Insurance
13606 Jobs | Jaipur