Title Google BigQuery

3 - 7 years

0 Lacs

Posted:5 days ago| Platform: Shine logo

Apply

Work Mode

On-site

Job Type

Full Time

Job Description

Role Overview: You will be responsible for having a good understanding of Hadoop eco system & Yarn architecture. Your hands-on experience in Spark with Python will be crucial in loading and manipulating large data sets using Spark & Hive into Hadoop GCP. Your knowledge on debugging and troubleshooting Hadoop jobs will be essential. In addition, you will need to have good communication and client interfacing skills. You will prepare implementation plans as per the need and build the in-scope applications in Big Data technologies. You will be responsible for all technical deliveries of the project, and you should have a good understanding of Agile & DevOps methodology. Prior experience with US customers is nice to have, and you should have worked in an offshore delivery model. You should be strong in Java/J2EE, Unix & SQL. Key Responsibilities: - Have a good understanding of Hadoop eco system & Yarn architecture - Hands-on experience in Spark with Python - Hands-on with loading and manipulating large data sets using Spark & Hive into Hadoop GCP - Knowledge on debugging and troubleshooting Hadoop jobs - Good communication and client interfacing skills - Prepare implementation plan as per the need and build the in-scope applications in Big Data technologies - Responsible for all technical deliveries of the project - Good understanding of Agile & DevOps methodology - Prior experience with US customer is nice to have - Worked in offshore delivery model - Strong in Java/J2EE, Unix & SQL Qualifications Required: - Strong experience in PySpark, Hadoop, Hive, Java - GCP experience in migrating On-Prem application to Cloud or good understanding of GCP services related to Application Migration - Knowledge in Python or willingness to learn Python (needed for GCP migration) - Knowledge in SQL - Good hands-on experience in Hadoop ecosystem - Good hands-on experience in Google Cloud Platform - Dataproc Composer Airflow - Nice to have experience in Spark with Python (Note: No additional details of the company were provided in the job description.) Role Overview: You will be responsible for having a good understanding of Hadoop eco system & Yarn architecture. Your hands-on experience in Spark with Python will be crucial in loading and manipulating large data sets using Spark & Hive into Hadoop GCP. Your knowledge on debugging and troubleshooting Hadoop jobs will be essential. In addition, you will need to have good communication and client interfacing skills. You will prepare implementation plans as per the need and build the in-scope applications in Big Data technologies. You will be responsible for all technical deliveries of the project, and you should have a good understanding of Agile & DevOps methodology. Prior experience with US customers is nice to have, and you should have worked in an offshore delivery model. You should be strong in Java/J2EE, Unix & SQL. Key Responsibilities: - Have a good understanding of Hadoop eco system & Yarn architecture - Hands-on experience in Spark with Python - Hands-on with loading and manipulating large data sets using Spark & Hive into Hadoop GCP - Knowledge on debugging and troubleshooting Hadoop jobs - Good communication and client interfacing skills - Prepare implementation plan as per the need and build the in-scope applications in Big Data technologies - Responsible for all technical deliveries of the project - Good understanding of Agile & DevOps methodology - Prior experience with US customer is nice to have - Worked in offshore delivery model - Strong in Java/J2EE, Unix & SQL Qualifications Required: - Strong experience in PySpark, Hadoop, Hive, Java - GCP experience in migrating On-Prem application to Cloud or good understanding of GCP services related to Application Migration - Knowledge in Python or willingness to learn Python (needed for GCP migration) - Knowledge in SQL - Good hands-on experience in Hadoop ecosystem - Good hands-on experience in Google Cloud Platform - Dataproc Composer Airflow - Nice to have experience in Spark with Python (Note: No additional details of the company were provided in the job description.)

Mock Interview

Practice Video Interview with JobPe AI

Start Java Interview
cta

Start Your Job Search Today

Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.

Job Application AI Bot

Job Application AI Bot

Apply to 20+ Portals in one click

Download Now

Download the Mobile App

Instantly access job listings, apply easily, and track applications.

coding practice

Enhance Your Java Skills

Practice Java coding challenges to boost your skills

Start Practicing Java Now
Cognizant logo
Cognizant

IT Services and IT Consulting

Teaneck New Jersey

RecommendedJobs for You