Junior Semantic Modeler & SQL Data Engineer
Overview
Junior Semantic Modeler & SQL Data Engineer
What You’ll Do
Data Modeling & Engineering
- Design and implement
semantic data models
governed by Domain-Driven Design (DDD) principles to support discoverability, self-service analytics, and scalable data product delivery. - Build data transformation workflows using
SQL-based stream processing with Apache Flink
for real-time analytics and low-latency data products. - Develop
SQL-based batch and micro-batch data pipelines using Apache Spark
to process largescale - datasets efficiently.
- Engineer robust, performant pipelines using
BigQuery, Snowflake, and Databricks SQL
—leveraging cloud-native best practices for storage, compute, and data engineering automation. - Apply
Data Mesh
and Lakehouse architectural patterns to drive domain-oriented ownership and high data quality. - Develop and maintain user-defined metadata layers and semantic catalogs for governed querying and consistent reporting.
- Optimize schema design, indexing, and data storage for both analytical and transactional performance across streaming and batch environments.
Agentic AI, Knowledge Graphs & Automation
- Rapidly upskill on RAG-based knowledge graph creation and fine-tuning, applying the latest AI models to accelerate semantic discovery.
- Rely on
AI-assisted development tools
(e.g., GitHub Copilot, Claude Code, Cursor, Aider) for code generation, documentation, and optimization—minimizing hand-coding while maximizing productivity. - Automate ETL/ELT workflows, implement robust data validation, and establish CI/CD practices for data pipelines, including regression and anomaly testing.
- Integrate and optimize
stream, batch, and micro-batch workflows
to ensure seamless, end-to-end data processing and analytics delivery.
Data Product & BI Delivery
- Design, version, and deliver
data products
with clear contracts and lineage, supporting reusable and testable analytics building blocks. - Develop advanced
BI reports and dashboards
(using tools like Tableau, Power BI, Looker, Superset) that surface key data-driven insights and metrics. - Ensure data integrity, quality, and consistency throughout the lifecycle via automated tests, validation frameworks, and detailed documentation.
- Monitor, profile, and proactively improve pipeline reliability, performance, and scalability—across real-time, batch, and micro-batch environments.
Leadership, Collaboration & Mentoring
- Partner with data scientists, business analysts, and product teams to co-create high-impact, user-centric solutions.
- Mentor junior engineers on advanced SQL, semantic modeling, AI-augmented workflows, and modern data processing paradigms (streaming, batch, micro-batch).
- Communicate complex technical concepts clearly across technical and non-technical audiences.
What Sets You Apart
Qualifications
- Bachelor’s in Computer Science, Software Engineering, or a related field from a top-tier engineering college
- 2+ years in data engineering or data product roles, with deep experience in SQL, semantic modeling, and pipeline design.
- Hands-on expertise in RDBMS, cloud data platforms (
AWS, Azure, GCP
), and cloud data warehouses (BigQuery, Snowflake, Databricks
). Proven experience in SQL-based stream processing (Apache Flink), batch/micro-batch processing (Apache Spark), and cloud-native SQL data engineering.
- Proficiency in Python (preferred) or Java/Scala for data transformation and automation.
- Familiarity with big data (Spark, Kafka, Hadoop) and NoSQL (MongoDB, Cassandra, Redis) is a plus.
- Proven experience using AI code assistant tools for productivity and workflow acceleration.
- Strong foundation in data architecture, version control, data governance, and metadata management.
Core Competencies
- Integrity, sound judgment, and a relentless commitment to data quality and innovation.
- Strategic, adaptive thinker with the ability to translate complex business needs into actionable data solutions.
- Excellent communicator, influencer, and collaborator across diverse technical and business teams.
- Self-starter with a sense of urgency, ownership, and drive to deliver high-impact results.
Skills & Mindset
- Advanced SQL mastery: Joins, CTEs, window functions, analytical functions, UDFs, stored procs, query optimization.
- Experience with
SQL-based stream processing (Flink), batch/micro-batch (Spark), and cloud SQL data platforms (BigQuery, Snowflake, Databricks) is a Plus
. - Semantic modeling: Ability to abstract, organize, and represent data for business meaning and usability.
- Metrics & dashboard development: Track record building high-value data-driven insights and reporting products.
- Critical thinking, analytical rigor, and attention to detail.
- Continuous learner—excited to work at the cutting edge of data and AI.
Join us to architect intelligent, agentic data solutions that power smarter decisions—across streaming, batch, and cloud data environments.
Location & Travel:
- Coimbatore/Chennai/Bangalore
About Steerwise:
Steerwise is a value-focused intellectual products & experiential services firm, providing business technology strategy, creative solution design, & software product delivery for organizations across industries, sectors, & geographies. We bring strategy advisory, solution design, software product development, and system operation competencies to actualize transformation & optimization endeavors. We work with clients to build capabilities that empower organizations to create digital customer experience, harmonize and monetize data assets, and exploit modern technologies for gaining competitive advantage by leveraging our Steerwise Vae™ intellectual capital. Our Business Value First™ culture makes us a valuable partner for our clients and an inspiring place to work for our employees.