Job
Description
About the Job :
The Data & AI team is a highly focused effort to lead digital-first execution and transformation at Red Hat leveraging data & AI strategically for our customers, partners, and associates.What will you do?
Evolve existing source data pipelines to an ELT model of data ingestionCleanly separate source-aligned data products from aggregate data productsRichly decorate your data products with metadata to support knowledge transfer, adoption, and the application of Machine LearningTag and classify your data assets to ensure they are used responsibly throughout the organization, applying masking or restricting access where appropriateApply software engineering best practices to your code release process to support CI/CD and a high velocity collaboration model based on InnerSourceRegister and maintain the catalog entries for your data products to support discoverability and reuseEnsure your data products easily join with other business data products using common identifiers and keysDevelop automated and resilient processes that support the end-end delivery of business valuePublish and deliver on a data product SLO for your data customersResponsibly share your data products with other internal consumers while balancing the core needs of security and complianceContribute feedback and recommendations to the Data Platform team in order to remove friction and increase scale for all usersWrite custom adapters to integrate internal data sources into the centralized Warehouse environmentWhat will you bring?
Bachelor's degree in Computer Science, Computer Engineering, or related field.4+ years of software development experience with a focus on data applications & systemsExceptional software and data engineering skills that lead to elegant and maintainable data productsExpert level proficiency in using SQL for data transformationProficiency in at least one general purpose programming language, eg. Python, Go, Java, Rust, etc.Strong opinions and perspectives that you kindly debate, defend, or change to ensure that the entire team moves as oneSets and resets the bar on all things quality, from code through to data, and everything in betweenDeep empathy for your users of your data products, leading to a constant focus on removing friction, increasing adoption, and delivering business resultsPrunes and prioritizes work in order to maximize your contributions and impactBias for action and leading by examplePast experience in building enterprise data products that have a high level of governance and compliance requirementsAs a Software Engineer, you will design and build our Model Context Protocol, a low-latency service that standardizes and serves real-time data to power our AI models. You will engineer resilient pipelines from feature stores and databases to accelerate model deployment and enhance prediction accuracy.The following are considered as a plus:
Familiarity with open source or inner source development and processesFamiliarity of data mesh architectural principlesExperience with Snowflake, Fivetran, dbt, Airflow / AstronomerPersonal qualities and communication
Communication skills and experience in interacting with cross functional business and engineering teamsCapability in undertaking business needs analysis in direct consultationMotivated with a passion for quality, learning and contributing to collective goalsExcellent communication, presentation, and writing skills