Data DevOps Engineer

3 - 7 years

0 Lacs

Posted:1 week ago| Platform: Shine logo

Apply

Work Mode

On-site

Job Type

Full Time

Job Description

As a DevOps Engineer specializing in data, you will be dedicated to implementing and managing our cloud-based data infrastructure utilizing AWS and Snowflake. Your primary responsibility will involve collaborating with data engineers, data scientists, and various stakeholders to design, deploy, and maintain a robust data ecosystem that supports our analytics and business intelligence initiatives. Your expertise in modern data technology stacks, MLOps methodologies, automation, and information security will play a crucial role in improving our data pipelines and ensuring data integrity and availability. You should possess a Bachelor's degree in Computer Science, Engineering, or have at least 3 years of experience in a DevOps engineering role or a similar engineering position. A strong command of AWS services (e.g., EC2, S3, Lambda, RDS) and cloud infrastructure best practices is essential. Proficiency in Snowflake, including data modeling, performance tuning, and query optimization, is required. Experience with modern data technologies and tools (e.g., Apache Airflow, dbt, ETL processes) is also expected. Familiarity with MLOps frameworks and methodologies such as MLflow, Kubeflow, or SageMaker, as well as knowledge of containerization and orchestration tools like Docker and Kubernetes, will be beneficial. Proficiency in scripting languages such as Python, Ruby, PHP, and Perl, along with automation frameworks, is necessary. Additionally, a strong understanding of Git and GitHub workflows, databases, SQL, CI/CD tools and practices (e.g., Jenkins, GitLab CI), and information security principles is crucial. Excellent problem-solving skills, a collaborative team spirit, and strong communication skills, both verbal and written, are highly valued. Preferred qualifications include experience with data governance and compliance frameworks, familiarity with data visualization tools (e.g., Tableau, Looker), and knowledge of machine learning frameworks and concepts. Possessing relevant security certifications (e.g., CISSP, CISM, AWS Certified Security) is considered a plus. Your key responsibilities will include infrastructure management, data pipeline deployment, Snowflake administration, MLOps implementation, information security integration, CI/CD implementation, support and troubleshooting, tool development, automation and visualization, system maintenance, monitoring and performance tuning, collaboration with stakeholders, and documentation of data architecture and security protocols. unifyCX is an emerging Global Business Process Outsourcing company with a strong presence in multiple countries. We provide personalized contact centers, business processing, and technology outsourcing solutions to clients worldwide, leveraging advanced AI technologies to enhance customer experiences and drive operational efficiency. We are committed to innovation and diversity, welcoming individuals from all backgrounds to join us in supporting our international clientele.,

Mock Interview

Practice Video Interview with JobPe AI

Start Python Interview
cta

Start Your Job Search Today

Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.

Job Application AI Bot

Job Application AI Bot

Apply to 20+ Portals in one click

Download Now

Download the Mobile App

Instantly access job listings, apply easily, and track applications.

coding practice

Enhance Your Python Skills

Practice Python coding challenges to boost your skills

Start Practicing Python Now

RecommendedJobs for You

Gurgaon / Gurugram, Haryana, India

Bengaluru / Bangalore, Karnataka, India