Freelance / Remote Big Data Engineer

3 years

0 Lacs

Posted:3 weeks ago| Platform: Linkedin logo

Apply

Work Mode

Remote

Job Type

Part Time

Job Description

About HYI.AI

HYI.AI is a Virtual Assistance and GenAI platform built for startups, entrepreneurs, and tech innovators. We specialize in offering virtual talent solutions, GenAI tools, and custom AI/ML deployments to help founders and businesses scale smarter and faster. We’re on a mission to power the next wave of digital startups globally.


Role Overview

Big Data Engineer


Key Responsibilities

  • Design, develop, and manage scalable

    data pipelines

    using big data technologies
  • Build data integration workflows for

    structured and unstructured

    data sources
  • Implement ETL/ELT processes for real-time and batch processing
  • Optimize data processing jobs for

    performance, scalability, and fault tolerance

  • Ensure

    data quality, security, and governance

    throughout the data lifecycle
  • Collaborate with analytics and engineering teams to support data product development
  • Monitor and maintain large-scale distributed systems and troubleshoot performance issues
  • Maintain documentation for data architecture, pipeline configurations, and operational procedures


Required Skills & Qualifications

  • 3+ years of hands-on experience in

    Big Data engineering

  • Proficiency in technologies such as

    Hadoop

    ,

    Spark

    ,

    Hive

    ,

    Kafka

    ,

    Flink

    , or

    Presto

  • Strong programming/scripting skills in

    Python

    ,

    Java

    , or

    Scala

  • Experience with

    cloud-based data platforms

    (AWS EMR, GCP BigQuery, Azure Data Lake)
  • Familiarity with

    data warehousing

    and

    distributed computing systems

  • Solid understanding of

    SQL

    , data modeling, and query optimization
  • Experience with

    workflow orchestration tools

    like

    Airflow

    ,

    Luigi

    , or

    Dagster

  • Knowledge of

    version control (Git)

    and CI/CD pipelines for data deployments


Preferred Skills

  • Experience with

    real-time streaming

    data pipelines using Kafka, Flink, or Kinesis
  • Familiarity with

    NoSQL databases

    (Cassandra, HBase, MongoDB)
  • Exposure to

    data governance

    ,

    compliance

    , and

    metadata management

  • Experience in

    containerized environments

    (Docker, Kubernetes)
  • Understanding of

    machine learning pipelines

    and integration with data science workflows


What You’ll Get

  • Work on high-volume data infrastructure projects with global companies
  • Flexible remote work and performance-based culture
  • Opportunity to architect data solutions for cutting-edge applications
  • Access to the HYI.AI network for collaboration, growth, and career advancement




Mock Interview

Practice Video Interview with JobPe AI

Start Job-Specific Interview
cta

Start Your Job Search Today

Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.

Job Application AI Bot

Job Application AI Bot

Apply to 20+ Portals in one click

Download Now

Download the Mobile App

Instantly access job listings, apply easily, and track applications.

coding practice

Enhance Your Skills

Practice coding challenges to boost your skills

Start Practicing Now

RecommendedJobs for You