About ShyftLabs
ShyftLabs is a fast-growing data product company founded in early 2020, working primarily with Fortune 500 clients. We design and deliver cutting-edge digital and data-driven solutions that help businesses accelerate growth, improve decision-making, and create measurable value through innovation.
Position Overview
We’re looking for an experienced Data Engineer who’s passionate about building scalable, high-performance data solutions. In this role, you’ll collaborate with cross-functional teams-including Data Engineers, Analysts, and Product Managers-to design, implement, and maintain robust data pipelines and systems that power our clients’ most critical business decisions.
-
Design, develop, and maintain data pipelines and ETL/ELT processes using Python.
-
Build and optimize scalable, high-performance data applications.
-
Collaborate with cross-functional teams to define requirements and deliver reliable solutions.
-
Develop and manage real-time streaming pipelines using Pub/Sub or Apache Beam.
-
Participate in code reviews, architecture discussions, and continuous improvement initiatives.
-
Monitor, troubleshoot, and optimize production data systems for reliability and performance.
-
5+ years of professional experience in software or data engineering using Python.
-
Strong understanding of software engineering best practices (testing, version control, CI/CD).
-
Proven experience building and optimizing ETL/ELT pipelines and data workflows.
-
Proficiency in SQL and database concepts.
-
Experience with data processing frameworks (e.g., Pandas).
-
Understanding of software design patterns and scalable architecture principles.
-
Experience with cloud platforms (GCP preferred).
-
Knowledge of CI/CD pipelines and Infrastructure as Code tools.
-
Familiarity with containerization (Docker, Kubernetes).
-
Bachelor’s degree in Computer Science, Engineering, or a related field (or equivalent experience).
-
Excellent problem-solving, analytical, and communication skills.
-
Experience with GCP services such as Cloud Run and Dataflow.
-
Experience with stream processing technologies (e.g., Pub/Sub).
-
Familiarity with workflow orchestration tools (e.g., Airflow).
-
Exposure to data visualization tools or libraries.
-
Knowledge of GitLab CI/CD and Terraform.
-
Experience with Snowflake, BigQuery, or Databricks.
-
GCP Data Engineer Certification is a plus.
We are proud to offer a competitive salary alongside a strong insurance package. We pride ourselves on the growth of our employees, offering extensive learning and development resources.