About Costco Wholesale
Costco Wholesale is a multi-billion-dollar global retailer with warehouse club operations in eleven countries. They provide a wide selection of quality merchandise, plus the convenience of specialty departments and exclusive member services, all designed to make shopping a pleasurable experience for their members.
About Costco Wholesale India
At Costco Wholesale India, we foster a collaborative space, working to support Costco Wholesale in developing innovative solutions that improve members’ experiences and make employees’ jobs easier. Our employees play a key role in driving and delivering innovation to establish IT as a core competitive advantage for Costco Wholesale.
Position Title: Staff, Data Engineer
Roles & Responsibilities:
- Shape and drive enterprise-wide data architecture strategy: Define and evolve the long-term technical vision for scalable, resilient data infrastructure across multiple business units and domains.
- Lead large-scale, cross-functional initiatives: Architect and guide the implementation of data platforms and pipelines that enable analytics, AI/ML, and BI at an organizational scale.
- Pioneer advanced and forward-looking solutions: Introduce novel approaches in real-time processing, hybrid/multi-cloud, and AI/ML integration to transform how data is processed and leveraged across the enterprise.
- Mentor and develop senior technical leaders: Influence Principal Engineers, Engineering Managers, and other Staff Engineers; create a culture of deep technical excellence and innovation.
- Establish cross-org technical standards: Define and enforce best practices for data modeling, pipeline architecture, governance, and compliance at scale.
- Solve the most complex, ambiguous challenges: Tackle systemic issues in data scalability, interoperability, and performance that impact multiple teams or the enterprise as a whole.
- Serve as a strategic advisor to executive leadership: Provide technical insights to senior executives on data strategy, emerging technologies, and long-term investments.
- Represent the organization as a thought leader: Speak at industry events/conferences, publish thought leadership, contribute to open source and standards bodies, and lead partnerships with external research or academic institutions.
Technical Skills:
- 15+ years of experience
- Mastery of data architecture and distributed systems at enterprise scale: Deep experience in GCP .
- Advanced programming and infrastructure capabilities: Expertise in writing database queries, Python, or Java, along with infrastructure-as-code tools like Terraform or Cloud Deployment Manager.
- Leadership in streaming and big data systems: Authority in tools such as BigQuery, Dataflow, Dataproc, Pub/sub for both batch and streaming workloads.
- Enterprise-grade governance and compliance expertise: Design and implement standards for data quality, lineage, security, privacy (e.g., GDPR, HIPAA), and auditability across the organization.
- Strategic integration with AI/ML ecosystems: Architect platforms that serve advanced analytics and AI workloads (Vertex AI, TFX, MLflow).
- Exceptional ability to influence across all levels: Communicate technical vision to engineers, influence strategic direction with executives, and drive alignment across diverse stakeholders.
- Recognized industry leader: Demonstrated track record through conference presentations, publications, open-source contributions, or standards development.
Must Have Skills:
- Deep expertise in data architecture, distributed systems, and GCP.
- Python or Java, infrastructure-as-code (e.g. Terraform)
- Big data tools: BigQuery(Expert level. Having experience on performance tuning and UDFs), Dataflow, Dataproc, Pub/Sub (batch + streaming)
- Data governance, privacy, and compliance (e.g. GDPR, HIPAA)
- Data modeling and architecture - level expert, have experience on hybrid architectures
- SQL Skills level - Expert
- Deep understanding of BigQuery, have experience on partitioning, clustering and performance optimizations
- Experience on Cloud function, Composer and Cloud run, dataflow flex templates - should be able to write
- Understanding of full concepts of cloud architecture.