Data Engineer – Financial Infrastructure & Analytics

2 years

0 Lacs

Posted:1 week ago| Platform: Linkedin logo

Apply

Work Mode

On-site

Job Type

Full Time

Job Description

About the Role

Quantitative Data Engineer

quant engineering, data systems, and research enablement

Responsibilities

  • Architect and maintain scalable ETL pipelines

    for ingesting and transforming terabytes of structured, semi-structured, and unstructured market and alternative data.
  • Design time-series optimized data stores

    and

    streaming frameworks

    to support low-latency data access for both backtesting and live trading.
  • Develop ingestion frameworks

    integrating vendor feeds (Bloomberg, Refinitiv, Polygon, Quandl, etc.), exchange data, and internal execution systems.
  • Collaborate with quantitative researchers and ML teams

    to ensure data accuracy, feature availability, and schema evolution aligned with modeling needs.
  • Implement data quality checks, validation pipelines, and version control mechanisms

    for all datasets.
  • Monitor and optimize distributed compute environments

    (Spark, Flink, Ray, or Dask) for performance and cost efficiency.
  • Automate workflows

    using orchestration tools (Airflow, Prefect, Dagster) for reliability and reproducibility.
  • Establish best practices

    for metadata management, lineage tracking, and documentation.
  • Contribute to internal libraries and SDKs

    for seamless data access by trading and research applications.

In Trading Firms, Data Engineers Typically:

  • Build

    real-time data streaming systems

    to capture market ticks, order books, and execution signals.
  • Manage

    versioned historical data lakes

    for backtesting and model training.
  • Handle

    multi-venue data normalization

    (different exchanges and instruments).
  • Integrate

    alternative datasets

    (satellite imagery, news sentiment, ESG, supply-chain data).
  • Work closely with

    quant researchers

    to convert raw data into

    research-ready features

    .
  • Optimize pipelines for

    ultra-low latency

    where milliseconds can impact P&L.
  • Implement

    data observability frameworks

    to ensure uptime and quality.
  • Collaborate with

    DevOps and infra engineers

    to scale storage, caching, and compute.

Tech Stack

  • Languages:

    Python, SQL, Scala, Go, Rust (optional for HFT pipelines)
  • Data Processing:

    Apache Spark, Flink, Ray, Dask, Pandas, Polars
  • Workflow Orchestration:

    Apache Airflow, Prefect, Dagster
  • Databases & Storage:

    PostgreSQL, ClickHouse, DuckDB, ElasticSearch, Redis
  • Data Lakes:

    Delta Lake, Iceberg, Hudi, Parquet
  • Streaming:

    Kafka, Redpanda, Pulsar
  • Cloud & Infra:

    AWS (S3, EMR, Lambda), GCP, Azure, Kubernetes
  • Version Control & Lineage:

    DVC, MLflow, Feast, Great Expectations
  • Visualization / Monitoring:

    Grafana, Prometheus, Superset, DataDog
  • Tools for Finance:

    kdb+/q (for tick data), InfluxDB, QuestDB

What You Will Gain

  • End-to-end ownership

    of core data infrastructure in a high-impact, mission-critical domain.
  • Deep exposure to

    quantitative research workflows

    ,

    market microstructure

    , and

    real-time trading systems

    .
  • Collaboration with elite quantitative researchers, traders, and ML scientists.

  • Hands-on experience with

    cutting-edge distributed systems

    and

    time-series data technologies

    .
  • A culture that emphasizes

    technical excellence, autonomy, and experimentation.

Qualifications

  • Bachelor’s or Master’s in

    Computer Science, Data Engineering, or related field.

  • 2+ years

    of experience building and maintaining

    production-grade data pipelines

    .
  • Proficiency in

    Python

    ,

    SQL

    , and frameworks like

    Airflow

    ,

    Spark

    , or

    Flink

    .
  • Familiarity with

    cloud storage and compute (S3, GCS, EMR, Dataproc)

    and

    versioned data lakes (Delta, Iceberg)

    .
  • Experience with

    financial datasets

    ,

    tick-level data

    , or

    high-frequency time series

    is a strong plus.
  • Strong understanding of

    data modeling, schema design, and performance optimization

    .
  • Excellent communication skills with an ability to support

    multidisciplinary teams

    .

Mock Interview

Practice Video Interview with JobPe AI

Start DevOps Interview
cta

Start Your Job Search Today

Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.

Job Application AI Bot

Job Application AI Bot

Apply to 20+ Portals in one click

Download Now

Download the Mobile App

Instantly access job listings, apply easily, and track applications.

coding practice

Enhance Your Python Skills

Practice Python coding challenges to boost your skills

Start Practicing Python Now
Meril logo
Meril

Medical Devices

Ahmedabad

RecommendedJobs for You