Jobs
Interviews

threeS Data Solutions

1 Job openings at threeS Data Solutions
Data Engineer/ ETL Developer (Senior) coimbatore, tamil nadu, india 6 years None Not disclosed On-site Contractual

Company Description threeS Data, a cutting-edge technology startup based in Coimbatore, India, specializes in Data Architecture, Management, Governance, Analytics, Intelligence, Business Intelligence, Automation, and Machine Learning. Founded in 2024, we focus on delivering simple, smart, and significant solutions that meet our clients' desired outcomes. Our engagements are partnerships, dedicated to understanding the complexities of day-to-day operations and offering practical, honest approaches to deliver exceptional results. Role Description This is a contract role based in Coimbatore, ideal for professionals who can independently deliver high-quality ETL solutions in a cloud-native, fast-paced environment. The position is hybrid, based in Coimbatore, with some work-from-home flexibility. Day-to-day tasks include designing, developing, and maintaining data pipelines, performing data modeling, implementing ETL processes, and managing data warehousing solutions. We are looking for candidates (6+years of experience) expertise in Apache Airflow , Redshift , and SQ based data pipelines, with upcoming transitions to Snowflake . Key Responsibilities: ETL Design and Development: · Design and develop scalable and modular ETL pipelines using Apache Airflow , with orchestration and monitoring capabilities. · Translate business requirements into robust data transformation pipelines across cloud-data platforms. · Develop reusable ETL components to support a configuration-driven architecture. Data Integration and Transformation: · Integrate data from multiple sources: Redshift, flat files, APIs, Excel, and relational databases. · Implement transformation logic such as cleansing, standardization, enrichment, and de-duplication. · Manage incremental and full loads, along with SCD handling strategies. SQL and Database Development: · Write performant SQL queries for data staging and transformation within Redshift and Snowflake. · Utilize joins, window functions, and aggregations effectively. · Ensure indexing and query tuning for high-performance workloads. Performance Tuning: · Optimize data pipelines and orchestrations for large-scale data volumes. · Tune SQL queries and monitor execution plans. · Implement best practices in distributed data processing and cloud-native optimizations. Error Handling and Logging: · Implement robust error handling and logging in Airflow DAGs. · Enable retry logic, alerting mechanisms, and failure notifications. Testing and Quality Assurance: · Conduct unit and integration testing of ETL jobs. · Validate data outputs against business rules and source systems. · Support QA during UAT cycles and help resolve data defects. Deployment and Scheduling: · Deploy pipelines using Git-based CI/CD practices. · Schedule and monitor DAGs using Apache Airflow and integrated tools. · Troubleshoot failures and ensure data pipeline reliability. Documentation and Maintenance: · Document data flows, DAG configurations, transformation logic, and operational procedures. · Maintain change logs and update job dependency charts. Collaboration and Communication: · Work closely with data architects, analysts, and BI teams to define and fulfill data needs. · Participate in stand-ups, sprint planning, and post-deployment reviews. Compliance and Best Practices: · Ensure ETL processes adhere to data security, governance, and privacy regulations (HIPAA, GDPR, etc.). · Follow naming conventions, version control standards, and deployment protocols. Qualifications o 6+ years of hands-on experience in ETL development. o Proven experience with Apache Airflow , Amazon Redshift , and strong SQL. o Strong understanding of data warehousing concepts and cloud-based data ecosystems. o Familiarity with handling flat files, APIs, and external sources o Experience with job orchestration, error handling, and scalable transformation patterns. o Ability to work independently and meet deadlines. Preferred Skills: § Exposure to Snowflake or plans to migrate to Snowflake platforms. § Experience in healthcare , life sciences , or regulated environments is a plus. § Familiarity with Azure Data Factory , Power BI , or other cloud BI tools. § Knowledge of Git, Azure DevOps, or other version control and CI/CD platforms.