Work from Office
Full Time
Were looking for a skilled Node.js Developer with a strong foundation in data engineering to join our engineering team. Youll be responsible for building scalable backend systems using modern Node.js frameworks and tools, while also designing and maintaining robust data pipelines and integrations. Primary Responsibilities: Build and maintain performant APIs and backend services using Node.js and frameworks like Express.js, NestJS, or Fastify. Develop and manage ETL/ELT pipelines, data models, schemas, and data transformation logic for analytics and operational use. Ensure data quality, integrity, and consistency through validation, monitoring, and logging. Work with database technologies (MySQL, PostgreSQL, MongoDB, Redis) to store and manage application and analytical data. Implement integrations with third-party APIs and internal microservices. Use ORMs like Sequelize, TypeORM, or Prisma for data modeling and interaction. Write unit, integration, and E2E tests using frameworks such as Jest, Mocha, or Supertest. Collaborate with frontend, DevOps, and data engineering teams to ship end-to-end features. Monitor and optimize system performance, logging (e.g., Winston, Pino), and error handling. Contribute to CI/CD workflows and infrastructure automation using tools like PM2, Docker and Jenkins. Required Skills: 3+ years of experience in backend development using Node.js. Hands-on experience with Express.js, NestJS, or other Node.js frameworks. Understanding of data modelling, partitioning, indexing, and query optimization. Experience in building and maintaining data pipelines, preferably using custom Node.js scripts. Familiarity with stream processing and messaging systems (e.g., Kafka, RabbitMQ, or Redis Streams). Solid understanding of SQL and NoSQL data stores and schema design. Strong knowledge of JavaScript and preferably TypeScript. Familiarity with cloud platforms (AWS/GCP/Azure) and services like S3, Lambda, or Cloud Functions. Experience with containerized environments (Docker) and CI/CD. Experience with data warehouses (e.g., BigQuery, Snowflake, Redshift). Nice To Have: Cloud Certification in AWS or GCP. Experience with distributed processing tools (eg. Spark, Trino/Presto) Experience with Data Transformation tool (ex. DBT, SQLMesh) and Data Orchestration (ex. Apache Airflow, Kestra etc) Familiarity with Serverless architectures and tools like Vercel/Netlify for deployment
Affle
Upload Resume
Drag or click to upload
Your data is secure with us, protected by advanced encryption.
My Connections Affle
Gurugram
8.0 - 12.0 Lacs P.A.
Hyderabad, Pune, Bengaluru
35.0 - 60.0 Lacs P.A.
Chennai
3.6 - 7.2 Lacs P.A.
Bengaluru
50.0 - 55.0 Lacs P.A.
Hyderabad
0.5 - 0.5 Lacs P.A.
Experience: Not specified
1.0 - 5.0 Lacs P.A.
Hyderabad, Pune
0.5 - 2.5 Lacs P.A.
Mumbai, Hyderabad, Chennai, Nellore, Bengaluru
Salary: Not disclosed
Noida, Hyderabad, Bengaluru
15.0 - 20.0 Lacs P.A.
15.0 - 27.5 Lacs P.A.