Sr. Data Engineer

0 years

0 Lacs

Posted:1 day ago| Platform: Linkedin logo

Apply

Work Mode

On-site

Job Type

Full Time

Job Description

Exigo Tech is a Sydney-based Technology Solutions Provider that is focused on providing solutions on three major verticals; Infrastructure, Cloud, and Application to businesses across Australia. We help companies reach operational efficiencies by empowering them with technology solutions that drive their business processes.


Exigo is looking for Full-time Sr. Data Engineer


We are ISO 27001:2022 certified organization

Visit our website: https://exigotech.co/ for more details….

LinkedIn: https://www.linkedin.com/company/exigotech/


Click Here to know more : LIFE AT EXIGO TECH


Roles and Responsibilities

  • Install, configure, and manage Apache Spark (open-source) clusters on Ubuntu, including Spark master/worker nodes and Spark environment files.
  • Configure and manage Spark UI and Spark History Server for monitoring jobs, analyzing DAGs, stages, tasks, and troubleshooting performance.
  • Develop, optimize, and deploy PySpark ETL/ELT pipelines using DataFrame API, UDFs, window functions, caching, partitioning, and broadcasting.
  • Deploy PySpark jobs using spark-submit in client/cluster mode with proper logging and error handling.
  • Install, configure, and manage Apache Airflow including UI, scheduler, webserver, connections, and variables.
  • Create, schedule, and monitor Airflow DAGs for PySpark jobs using SparkSubmitOperator, BashOperator, or PythonOperator.
  • Configure and manage cron jobs for scheduling data processing tasks where needed.
  • Install, configure, and optimize Trino (PrestoSQL) coordinator and worker nodes; configure catalogs such as S3, MySQL, or PostgreSQL.
  • Maintain Linux/Ubuntu servers including services, logs, environment variables, memory usage, and port conflict resolution.
  • Design and implement scalable data architectures using Azure Data Services including ADF, Synapse, ADLS, Azure SQL, and Databricks.
  • Develop, manage, and automate ETL/ELT pipelines using Azure Data Factory (Pipelines, Mapping Dataflows, Dataflows).
  • Monitor, troubleshoot, and optimize data pipelines across Spark, Airflow, Trino, and Azure platforms.
  • Work with structured, semi-structured, and unstructured data across multiple data sources and formats.
  • Implement data analytics, transformation, backup, and recovery solutions.
  • Perform data migration, upgrade, and modernization using Azure and database tools.
  • Implement CI/CD pipelines for data solutions using Azure DevOps and Git.
  • Ensure data quality, governance, lineage, metadata management, and security compliance across cloud and big data environments.
  • Design and optimize data models using star and snowflake schemas; build data warehouses, Delta Lake, and Lakehouse systems.
  • Develop and rebuild reports/dashboards using Power BI, Tableau, or similar tools.
  • Collaborate with internal teams, clients, and business users to gather requirements and deliver high-quality data solutions.
  • Provide documentation, runbooks, and operational guidance.


Technical Skills:

  1. Apache Spark (Open Source) & PySpark - Must
  • Apache Spark installation & cluster configuration (Ubuntu/Linux)
  • Spark master/worker setup (standalone & cluster mode)
  • Spark UI & History Server configuration and debugging
  • PySpark development (ETL pipelines, UDFs, window functions, DataFrame API)
  • Performance tuning (partitioning, caching, shuffles)
  • spark-submit deployment with monitoring and logging

2. Apache Airflow & Job Orchestration - Must

  • Airflow installation & configuration (UI, scheduler, webserver)
  • Creating and scheduling DAGs (SparkSubmitOperator, BashOperator, PythonOperator)
  • Retry logic, triggers, alerting, and log management
  • Cron job scheduling & process automation

3. Trino (PrestoSQL) - Must

  • Trino coordinator & worker node setup
  • Catalog configuration (S3, RDBMS sources)
  • Distributed SQL troubleshooting & performance optimization

4. Azure Data Services (nice to have)

  • Azure Data Factory
  • Azure Synapse Analytics
  • Azure SQL / Cosmos DB
  • Azure Data Lake Storage (Gen2)
  • Azure Databricks (Delta, Notebooks, Jobs)
  • Azure Event Hubs / Stream Analytics

5. Microsoft Fabric ( nice to have)

  • Lakehouse
  • Warehouse
  • Dataflows
  • Notebooks
  • Pipelines

6. Programming & Querying

  • Python
  • PySpark
  • SQL
  • Scala

7. Data Modeling & Warehousing

  • Star schema modeling
  • Snowflake schema modeling
  • Fact/dimension modeling
  • Data warehouse & Lakehouse design
  • Delta Lake / Lakehouse architectures

8. DevOps & CI/CD

  • Git / GitHub / Azure Repos
  • Azure DevOps pipelines (CI/CD)
  • Automated deployment for Spark, Airflow, ADF, Databricks, Fabric

9. BI Tools (Nice to have)

  • Power BI
  • Tableau
  • Report building, datasets, DAX

10. Linux/Ubuntu Server Knowledge

  • Shell scripting
  • Service management
  • Logs & environment variables


Soft Skills:

  • Excellent problem solving and communication skills
  • Able to work well in a team setting
  • Excellent organizational and time management skills
  • Taking end-to-end ownership
  • Production support & timely delivery
  • Self-driven, flexible and innovative
  • Microsoft Certified: Azure Data Engineer Associate (DP-203 / DP -300)
  • Knowledge of DevOps and CI/CD pipelines in Azure


Education:

  • BSc/BA in Computer Science, Engineering or a related field


Work Location:

Mock Interview

Practice Video Interview with JobPe AI

Start PySpark Interview
cta

Start Your Job Search Today

Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.

Job Application AI Bot

Job Application AI Bot

Apply to 20+ Portals in one click

Download Now

Download the Mobile App

Instantly access job listings, apply easily, and track applications.

coding practice

Enhance Your Skills

Practice coding challenges to boost your skills

Start Practicing Now

RecommendedJobs for You

vadodara, gujarat, india

pune, maharashtra, india

bengaluru, karnataka, india