Get alerts for new jobs matching your selected skills, preferred locations, and experience range. Manage Job Alerts
0 years
4 - 9 Lacs
Bengaluru
Remote
Your opportunity Do you love the transformative impact data can have on a business? Are you motivated to push for results and overcome all obstacles? Then we have a role for you. What you'll do Lead the building of scalable, fault tolerant pipelines with built in data quality checks that transform, load and curate data from various internal and external systems Provide leadership to cross-functional initiatives and projects. Influence architecture design and decisions. Build cross-functional relationships with Data Scientists, Product Managers and Software Engineers to understand data needs and deliver on those needs. Improve engineering processes and cross-team collaboration. Ruthlessly prioritize work to align with company priorities. Provide thought leadership to grow and evolve DE function and implementation of SDLC best practices in building internal-facing data products by staying up-to-date with industry trends, emerging technologies, and best practices in data engineering This role requires Experience in BI and Data Warehousing. Strong experience with dbt, Airflow and snowflake Experience with Apache Iceberg tables Experience and knowledge of building data-lakes in AWS (i.e. Spark/Glue, Athena), including data modeling, data quality best practices, and self-service tooling. Experience mentoring data professionals from junior to senior levels Demonstrated success leading cross functional initiatives Passionate about data quality, code quality, SLAs and continuous improvement Deep understanding of data system architecture Deep understanding of ETL/ELT patterns Development experience in at least one object-oriented language (Python,R,Scala, etc.). Comfortable with SQL and related tooling Bonus points if you have Experience with Observability Please note that visa sponsorship is not available for this position. Fostering a diverse, welcoming and inclusive environment is important to us. We work hard to make everyone feel comfortable bringing their best, most authentic selves to work every day. We celebrate our talented Relics’ different backgrounds and abilities, and recognize the different paths they took to reach us – including nontraditional ones. Their experiences and perspectives inspire us to make our products and company the best they can be. We’re looking for people who feel connected to our mission and values, not just candidates who check off all the boxes. If you require a reasonable accommodation to complete any part of the application or recruiting process, please reach out to resume@newrelic.com. We believe in empowering all Relics to achieve professional and business success through a flexible workforce model. This model allows us to work in a variety of workplaces that best support our success, including fully office-based, fully remote, or hybrid. Our hiring process In compliance with applicable law, all persons hired will be required to verify identity and eligibility to work and to complete employment eligibility verification. Note: Our stewardship of the data of thousands of customers’ means that a criminal background check is required to join New Relic. We will consider qualified applicants with arrest and conviction records based on individual circumstances and in accordance with applicable law including, but not limited to, the San Francisco Fair Chance Ordinance. Headhunters and recruitment agencies may not submit resumes/CVs through this website or directly to managers. New Relic does not accept unsolicited headhunter and agency resumes, and will not pay fees to any third-party agency or company that does not have a signed agreement with New Relic. Candidates are evaluated based on qualifications, regardless of race, religion, ethnicity, national origin, sex, sexual orientation, gender expression or identity, age, disability, neurodiversity, veteran or marital status, political viewpoint, or other legally protected characteristics. Review our Applicant Privacy Notice at https://newrelic.com/termsandconditions/applicant-privacy-policy
Posted 1 week ago
9.0 years
0 Lacs
Bengaluru, Karnataka, India
On-site
Who We Are Wayfair is moving the world so that anyone can live in a home they love – a journey enabled by more than 3,000 Wayfair engineers and a data-centric culture. Wayfair’s Advertising business is rapidly expanding, adding hundreds of millions of dollars in profits to Wayfair. We are building Sponsored Products, Display & Video Ad offerings that cater to a variety of Advertiser goals while showing highly relevant and engaging Ads to millions of customers. We are evolving our Ads Platform to empower advertisers across all sophistication levels to grow their business on Wayfair at a strong, positive ROI and are leveraging state of the art Machine Learning techniques. The Advertising Optimization & Automation Science team is central to this effort. We leverage machine learning and generative AI to streamline campaign workflows, delivering impactful recommendations on budget allocation, target Return on Ad Spend (tROAS), and SKU selection. Additionally, we are developing intelligent systems for creative optimization and exploring agentic frameworks to further simplify and enhance advertiser interactions. We are looking for an experienced Senior Machine Learning Scientist to join the Advertising Optimization & Automation Science team. In this role, you will be responsible for building intelligent, ML-powered systems that drive personalized recommendations and campaign automation within Wayfair’s advertising platform. You will work closely with other scientists, as well as members of our internal Product and Engineering teams, to apply your ML expertise to define and deliver 0-to-1 capabilities that unlock substantial commercial value and directly enhance advertiser outcomes. What You’ll do Design and build intelligent budget, tROAS, and SKU recommendations, and simulation-driven decisioning that extends beyond the current advertising platform capabilities. Lead the next phase of GenAI-powered creative optimization and automation to drive significant incremental ad revenue and improve supplier outcomes. Raise technical standards across the team by promoting best practices in ML system design and development. Partner cross-functionally with Product, Engineering, and Sales to deliver scalable ML solutions that improve supplier campaign performance. Ensure systems are designed for reuse, extensibility, and long-term impact across multiple advertising workflows. Research and apply best practices in advertising science, GenAI applications in creative personalization, and auction modeling. Keep Wayfair at the forefront of innovation in supplier marketing optimization. Collaborate with Engineering teams (AdTech, ML Platform, Campaign Management) to build and scale the infrastructure needed for automated, intelligent advertising decisioning. We Are a Match Because You Have : Bachelor's or Master’s degree in Computer Science, Mathematics, Statistics, or related field. 9+ years of experience in building large scale machine learning algorithms. 4+ years of experience working in an architect or technical leadership position. Strong theoretical understanding of statistical models such as regression, clustering and ML algorithms such as decision trees, neural networks, transformers and NLP techniques. Proficiency in programming languages such as Python and relevant ML libraries (e.g., TensorFlow, PyTorch) to develop production-grade products. Strategic thinker with a customer-centric mindset and a desire for creative problem solving, looking to make a big impact in a growing organization. Demonstrated success influencing senior level stakeholders on strategic direction based on recommendations backed by in-depth analysis; Excellent written and verbal communication. Ability to partner cross-functionally to own and shape technical roadmaps Intellectual curiosity and a desire to always be learning! Nice to have Experience with GCP, Airflow, and containerization (Docker). Experience building scalable data processing pipelines with big data tools such as Hadoop, Hive, SQL, Spark, etc. Familiarity with Generative AI and agentic workflows. Experience in Bayesian Learning, Multi-armed Bandits, or Reinforcement Learning. About Wayfair Inc. Wayfair is one of the world’s largest online destinations for the home. Through our commitment to industry-leading technology and creative problem-solving, we are confident that Wayfair will be home to the most rewarding work of your career. If you’re looking for rapid growth, constant learning, and dynamic challenges, then you’ll find that amazing career opportunities are knocking. No matter who you are, Wayfair is a place you can call home. We’re a community of innovators, risk-takers, and trailblazers who celebrate our differences, and know that our unique perspectives make us stronger, smarter, and well-positioned for success. We value and rely on the collective voices of our employees, customers, community, and suppliers to help guide us as we build a better Wayfair – and world – for all. Every voice, every perspective matters. That’s why we’re proud to be an equal opportunity employer. We do not discriminate on the basis of race, color, ethnicity, ancestry, religion, sex, national origin, sexual orientation, age, citizenship status, marital status, disability, gender identity, gender expression, veteran status, genetic information, or any other legally protected characteristic. We are interested in retaining your data for a period of 12 months to consider you for suitable positions within Wayfair. Your personal data is processed in accordance with our Candidate Privacy Notice (which can found here: https://www.wayfair.com/careers/privacy). If you have any questions regarding our processing of your personal data, please contact us at dataprotectionofficer@wayfair.com. If you would rather not have us retain your data please contact us anytime at dataprotectionofficer@wayfair.com.
Posted 1 week ago
5.0 years
0 Lacs
Chennai, Tamil Nadu, India
On-site
Overview: TekWissen is a global workforce management provider throughout India and many other countries in the world. The below clientis a global company with shared ideals and a deep sense of family. From our earliest days as a pioneer of modern transportation, we have sought to make the world a better place – one that benefits lives, communities and the planet Job Title: Software Engineer Senior Location: Chennai Work Type: Hybrid Position Description: As part of the client's DP&E Platform Observability team, you'll help build a top-tier monitoring platform focused on latency, traffic, errors, and saturation. You'll design, develop, and maintain a scalable, reliable platform, improving MTTR/MTTX, creating dashboards, and optimizing costs. Experience with large systems, monitoring tools (Prometheus, Grafana, etc.), and cloud platforms (AWS, Azure, GCP) is ideal. The focus is a centralized observability source for data-driven decisions and faster incident response. Skills Required: Spring Boot, Angular, Cloud Computing Skills Preferred: Google Cloud Platform - Biq Query, Data Flow, Dataproc, Data Fusion, TERRAFORM, Tekton,Cloud SQL, AIRFLOW, POSTGRES, Airflow PySpark, Python, API Experience Required: 5+ years of overall experience with proficiency in Java, angular or any javascript technology with experience in designing and deploying cloud-based data pipelines and microservices using GCP tools like BigQuery, Dataflow, and Dataproc. Ability to leverage best in-class data platform technologies (Apache Beam, Kafka,...) to deliver platform features, and design & orchestrate platform services to deliver data platform capabilities. Service-Oriented Architecture and Microservices: Strong understanding of SOA, microservices, and their application within a cloud data platform context. Develop robust, scalable services using Java Spring Boot, Python, Angular, and GCP technologies. Full-Stack Development: Knowledge of front-end and back-end technologies, enabling collaboration on data access and visualization layers (e.g., React, Node.js). Design and develop RESTful APIs for seamless integration across platform services. Implement robust unit and functional tests to maintain high standards of test coverage and quality. Database Management: Experience with relational (e.g., PostgreSQL, MySQL) and NoSQL databases, as well as columnar databases like BigQuery. Data Governance and Security: Understanding of data governance frameworks and implementing RBAC, encryption, and data masking in cloud environments. CI/CD and Automation: Familiarity with CI/CD pipelines, Infrastructure as Code (IaC) tools like Terraform, and automation frameworks. Manage code changes with GitHub and troubleshoot and resolve application defects efficiently. Ensure adherence to SDLC best practices, independently managing feature design, coding, testing, and production releases. Problem-Solving: Strong analytical skills with the ability to troubleshoot complex data platform and microservices issues. Experience Preferred: GCP Data Engineer, GCP Professional Cloud Education Required: Bachelor's Degree TekWissen® Group is an equal opportunity employer supporting workforce diversity.
Posted 1 week ago
4.0 years
7 - 11 Lacs
India
On-site
Experience: 4+ years Department: Data & Analytics We’re seeking a hands-on Senior Data Analyst & Scientist to drive insights, build predictive models, and develop scalable data pipelines. This hybrid role blends data science and analytics engineering to support decision-making across teams. Responsibilities: - Analyze data to identify trends and business insights - Build and deploy ML models (forecasting, segmentation, recommendations) - Design/manage ETL/ELT pipelines - Develop dashboards and reports (Tableau, Power BI, etc.) - Collaborate with product, engineering, and business teams Requirements: - 4+ years of experience in data science, analytics, or data engineering - Strong in SQL and Python (pandas, NumPy); experience with ML tools like scikit-learn - Familiar with Airflow, DBT, or similar orchestration tools - Experience with cloud platforms like BigQuery, Snowflake, or Redshift - Proficient in BI tools such as Tableau, Power BI, or Looker - Excellent communication and problem-solving skills Nice to Have: - ML model deployment experience - Cloud experience (AWS, GCP, or Azure) - Background in SaaS or product analytics Job Type: Full-time Pay: ₹60,000.00 - ₹95,000.00 per month Benefits: Flexible schedule Paid sick time Paid time off Language: English (Required) Work Location: In person
Posted 1 week ago
0 years
0 Lacs
Chennai, Tamil Nadu, India
On-site
Job Description Responsibilities: Infrastructure Provisioning and Management: Set up and configure cloud projects/Data Integration Tools and accounts using Google Cloud Platform (GCP). Configuring Cloud and Data Integration Solutions: Configure Data integration tools like Astronomer/GKE/App Integration/Airflow and Composer Data Pipeline Orchestration: Set up and configure data pipeline orchestration tools such as Astronomer, Composer, and Airflow. Manage and schedule workflows using Directed Acyclic Graphs (DAGs) to ensure efficient data processing. Monitoring and Incident Response: Maintain the availability of systems and services in production by setting service-level objectives (SLOs), service-level agreements (SLAs), and service-level indicators (SLIs). Set up monitoring and alerting systems for cloud services, data pipelines, and AI solutions. Respond to incidents, perform root cause analysis, and collaborate with teams to improve incident response practices. Security and Compliance: Ensure access and identity management in Google Cloud and data pipeline projects. Implement security measures to protect cloud-based data, services, and AI/ML workloads. Comply with global privacy laws and organizational policies. GenAI Solution Development & Delivery: Lead the development, testing, and deployment of GenAI solutions. Collaboration and Communication: Collaborate with cross-functional teams to design and implement innovative security solutions that align with industry best practices and regulatory requirements. Communicate effectively with management and users about the state and performance of cloud services, data pipelines, and AI solutions. Continuous Improvement: Automate infrastructure provisioning with tools like Terraform and Tekton. Enhance visibility and monitoring capabilities for cloud services, data pipelines, and GenAI solutions. Implement improvements based on feedback and incident reviews OnCall Support: Able to support Oncall during weekends Responsibilities Responsibilities: Infrastructure Provisioning and Management: Set up and configure cloud projects/Data Integration Tools and accounts using Google Cloud Platform (GCP). Configuring Cloud and Data Integration Solutions: Configure Data integration tools like Astronomer/GKE/App Integration/Airflow and Composer Data Pipeline Orchestration: Set up and configure data pipeline orchestration tools such as Astronomer, Composer, and Airflow. Manage and schedule workflows using Directed Acyclic Graphs (DAGs) to ensure efficient data processing. Monitoring and Incident Response: Maintain the availability of systems and services in production by setting service-level objectives (SLOs), service-level agreements (SLAs), and service-level indicators (SLIs). Set up monitoring and alerting systems for cloud services, data pipelines, and AI solutions. Respond to incidents, perform root cause analysis, and collaborate with teams to improve incident response practices. Security and Compliance: Ensure access and identity management in Google Cloud and data pipeline projects. Implement security measures to protect cloud-based data, services, and AI/ML workloads. Comply with global privacy laws and organizational policies. GenAI Solution Development & Delivery: Lead the development, testing, and deployment of GenAI solutions. Collaboration and Communication: Collaborate with cross-functional teams to design and implement innovative security solutions that align with industry best practices and regulatory requirements. Communicate effectively with management and users about the state and performance of cloud services, data pipelines, and AI solutions. Continuous Improvement: Automate infrastructure provisioning with tools like Terraform and Tekton. Enhance visibility and monitoring capabilities for cloud services, data pipelines, and GenAI solutions. Implement improvements based on feedback and incident reviews OnCall Support: Able to support Oncall during weekends Qualifications QUALIFICATIONS Education: Bachelor's degree in Computer Science, Information Technology, or a related field. Experience: Proven experience in cloud engineering, data pipeline orchestration, and AI solution development, specifically with Google Cloud Platform (GCP) and tools like Astronomer, Composer, Airflow. Programming Language : Python, Terraform, Google cloud, Tekton, Gen AI and LLM Models. Skills: Proficiency in GCP services and data pipeline orchestration tools. Strong understanding of IAM roles, policies, and security best practices. Experience with infrastructure as code tools like Terraform. Excellent problem-solving and troubleshooting skills. Strong communication and collaboration skills.
Posted 1 week ago
6.0 years
14 - 24 Lacs
India
On-site
6+ years of experience as a Data Engineer. Strong proficiency in SQL. Hands-on experience with modern cloud data warehousing solutions (Snowflake, Big Query, Redshift) Expertise in ETL/ELT processes, batch, and streaming data processing. Proven ability to troubleshoot data issues and propose effective solutions. Knowledge of AWS services (S3, DMS, Glue, Athena). Familiarity with DBT for data transformation and modeling. Must be fluent in English communication. Desired Experience 3 years of Experience with additional AWS services (EC2, ECS, EKS, VPC, IAM). Knowledge of Infrastructure as Code (IaC) tools like Terraform and Terragrunt. Proficiency in Python for data engineering tasks. Experience with orchestration tools like Dagster, Airflow, or AWS Step Functions. Familiarity with pub-sub, queuing, and streaming frameworks (AWS Kinesis, Kafka, SQS, SNS). Experience with CI/CD pipelines and automation for data processes. Skills: python,glue,automation,cloud data warehousing (snowflake, big query, redshift),pub-sub frameworks (aws kinesis, kafka, sqs, sns),aws (s3, dms, glue, athena),etl/elt processes,orchestration tools (dagster, airflow, aws step functions),batch processing,ci/cd pipelines,data,ci,streaming data processing,aws,infrastructure as code (iac),etl,sql,dbt,pyspark,lambda,terragrunt,aws services (ec2, ecs, eks, vpc, iam),cd,terraform,s3
Posted 1 week ago
0 years
0 Lacs
Bengaluru, Karnataka, India
On-site
Hi All, We are hiring for Data Engineers, kindly refer the below skillsets: Mandatory Skills: GCP Cloud (especially BigQuery and DataProc) Big Data technologies Hadoop Hive Python / PySpark Airflow and DAG orchestration Preferred Skills: Experience with visualization tools such as Tableau or Power BI Familiarity with Jethro is a plus
Posted 1 week ago
2.0 - 3.0 years
0 Lacs
Bengaluru, Karnataka, India
On-site
Job Title: Senior Data Analyst - Project Management Location: Bengaluru, Karnataka, India Experience : 2-3 Years About the Company & Role : We are one of India’s premier integrated political consulting firms specializing in building data-driven 360-degree election campaigns. We help our clients with strategic advice and implementation which brings together data-backed insights and in-depth ground intelligence into a holistic electoral campaign. We are passionate about our democracy and the politics that shape the world around us. We draw on some of the sharpest minds from distinguished institutions and diverse professional backgrounds to help us achieve our goal. The team brings in 7 years of experience in building electoral strategies that spark conversations, effect change, and help shape electoral and legislative ecosystems in our country. Job Summary: We are seeking a highly motivated and skilled Data Analyst to join our dynamic Project Management Office (PMO). This critical role involves developing, maintaining, and enhancing insightful PMO dashboards while also designing, implementing, and managing automated data pipelines. The ideal candidate will possess a strong blend of data analysis, visualization, and technical automation skills to ensure the PMO has timely, accurate data for tracking project performance, identifying trends, and making data-driven decisions. Key Responsibilities: PMO Dashboard Development & Management: Design, build, and maintain interactive dashboards using BI tools (e.g., Looker Studio, Tableau) to visualize key project metrics, resource allocation, timelines, risks, and overall PMO performance KPIs. Collaborate with PMO leadership and project managers to gather reporting requirements and translate them into effective data models and visualizations. Ensure data accuracy, consistency, and reliability within dashboards and reports. Perform data analysis to identify trends, potential issues, and areas for process improvement within project execution. Generate regular performance reports and support ad-hoc data requests from stakeholders. Data Management: Design, develop, implement, and maintain robust, automated data pipelines for Extract, Transform, Load (ETL/ELT) processes. Automate data collection from various sources including project management software, spreadsheets, databases, and APIs (e.g., Slack API). Load and process data efficiently into our data warehouse environment (e.g., Google BigQuery). Write and optimize SQL queries for data manipulation, transformation, and aggregation. Implement data quality checks, error handling, and monitoring for automated pipelines. Troubleshoot and resolve issues related to data extraction, transformation, loading, and pipeline failures. Document data sources, data models, pipeline architecture, and automation workflows. Required Qualifications & Skills: Bachelor's degree in Computer Science, Data Science, Statistics, Information Systems, Engineering, or a related quantitative field. Proven experience (approx. 2-3 years) in data analysis, business intelligence, data engineering, or a similar role. Strong proficiency in SQL for complex querying, data manipulation, and performance tuning. Hands-on experience building and maintaining dashboards using Tableau. Demonstrable experience in designing and automating data pipelines using scripting languages (Python preferred) and/or ETL/ELT tools. Solid understanding of data warehousing concepts, ETL/ELT principles, and data modeling. Excellent analytical, problem-solving, and critical thinking skills. Strong attention to detail and commitment to data accuracy. Good communication and collaboration skills, with the ability to interact with technical and non-technical stakeholders. Ability to work independently and manage priorities effectively. Preferred Qualifications & Skills: Experience working directly within a Project Management Office (PMO) or supporting project management functions. Familiarity with project management tools (e.g., Jira, Asana, MS Project) and concepts (Agile, Waterfall). Experience with cloud platforms, particularly Google Cloud Platform (GCP) and BigQuery. Experience with workflow orchestration tools (e.g., Airflow, Cloud Composer, Cloud Functions). Experience integrating data via APIs from various business systems. Basic understanding of data governance and data quality management practices. If you are a driven professional seeking a high-impact challenge and interested in joining a team of like-minded, motivated individuals who think strategically, act decisively, and get things done, email us at openings@varaheanalytics.com
Posted 1 week ago
5.0 years
0 Lacs
Hyderabad, Telangana, India
On-site
Job Title: Senior Python Developer – AI/ML Document Automation Location: Hyderabad Work Mode: Hybrid Experience: 5+ Years Job Summary: We are looking for a highly skilled Senior Python Developer with deep expertise in AI/ML and document automation . The ideal candidate will lead the design and development of intelligent systems for extracting and processing structured and unstructured data from documents such as invoices, receipts, contracts, and PDFs. This role involves both hands-on coding and architectural contributions to scalable automation platforms. Roles and Responsibilities: Design and develop modular Python applications for document parsing and intelligent automation. Build and optimize ML/NLP pipelines for tasks like Named Entity Recognition (NER), classification, and layout-aware data extraction. Integrate rule-based and AI-driven techniques (e.g., regex, spaCy, PyMuPDF, Tesseract) to handle diverse document formats. Develop and deploy models via REST APIs using FastAPI or Flask, and containerize with Docker. Collaborate with cross-functional teams to define automation goals and data strategies. Conduct code reviews, mentor junior developers, and uphold best coding practices. Monitor model performance and implement feedback mechanisms for continuous improvement. Maintain thorough documentation of workflows, metrics, and architectural decisions. Mandatory Skills: Expert in Python (OOP, asynchronous programming, modular design). Strong foundation in machine learning algorithms and natural language processing techniques. Hands-on experience with Scikit-learn, TensorFlow, PyTorch, and Hugging Face Transformers. Proficient in developing REST APIs using FastAPI or Flask. Experience in PDF/text extraction using PyMuPDF, Tesseract, or similar tools. Skilled in regex-based extraction and rule-based NER. Familiar with Git, Docker, and any major cloud platform (AWS, GCP, or Azure). Exposure to MLOps tools such as MLflow, Airflow, or LangChain.
Posted 1 week ago
0 years
0 Lacs
Hyderabad, Telangana, India
On-site
Position Overview: We are seeking a talented Data Engineer with expertise in Apache Spark, Python / Java and distributed systems. The ideal candidate will be skilled in creating and managing data pipelines using AWS. Key Responsibilities: Design, develop, and implement data pipelines for ingesting, transforming, and loading data at scale. Utilise Apache Spark for data processing and analysis. Utilise AWS services (S3, Redshift, EMR, Glue) to build and manage efficient data pipelines. Optimise data pipelines for performance and scalability, considering factors like partitioning, bucketing, and caching. Write efficient and maintainable Python code. Implement and manage distributed systems for data processing. Collaborate with cross-functional teams to understand data requirements and deliver optimal solutions. Ensure data quality and integrity throughout the data lifecycle. Qualifications: Proven experience with Apache Spark and Python / Java. Strong knowledge of distributed systems. Proficiency in creating data pipelines with AWS. Excellent problem-solving and analytical skills. Ability to work independently and as part of a team. Bachelor's degree in Computer Science, Information Technology, or a related field (or equivalent experience). Proven experience in designing and developing data pipelines using Apache Spark and Python. Experience with distributed systems concepts (Hadoop, YARN) is a plus. In-depth knowledge of AWS cloud services for data engineering (S3, Redshift, EMR, Glue). Familiarity with data warehousing concepts (data modeling, ETL) is preferred. Strong programming skills in Python (Pandas, NumPy, Scikit-learn are a plus). Experience with data pipeline orchestration tools (Airflow, Luigi) is a plus. Excellent problem-solving and analytical skills. Strong communication and collaboration skills. Preferred Qualifications: Experience with additional AWS services (e.g., AWS Glue, AWS Lambda, Amazon Redshift). Familiarity with data warehousing and ETL processes. Knowledge of data governance and best practices. Have a good understanding of the oops concept. Hands-on experience with SQL database design Experience with Python, SQL, and data visualization/exploration tools
Posted 1 week ago
0.0 - 1.0 years
8 - 14 Lacs
Hyderabad, Telangana
On-site
Job Title: Senior Python Developer – AI/ML Document Automation Location: Hyderabad Work Mode: Hybrid Experience: 5+ Years Job Summary: We are looking for a highly skilled Senior Python Developer with deep expertise in AI/ML and document automation . The ideal candidate will lead the design and development of intelligent systems for extracting and processing structured and unstructured data from documents such as invoices, receipts, contracts, and PDFs. This role involves both hands-on coding and architectural contributions to scalable automation platforms. Roles and Responsibilities: Design and develop modular Python applications for document parsing and intelligent automation. Build and optimize ML/NLP pipelines for tasks like Named Entity Recognition (NER), classification, and layout-aware data extraction. Integrate rule-based and AI-driven techniques (e.g., regex, spaCy, PyMuPDF, Tesseract) to handle diverse document formats. Develop and deploy models via REST APIs using FastAPI or Flask, and containerize with Docker. Collaborate with cross-functional teams to define automation goals and data strategies. Conduct code reviews, mentor junior developers, and uphold best coding practices. Monitor model performance and implement feedback mechanisms for continuous improvement. Maintain thorough documentation of workflows, metrics, and architectural decisions. Mandatory Skills: Expert in Python (OOP, asynchronous programming, modular design). Strong foundation in machine learning algorithms and natural language processing techniques. Hands-on experience with Scikit-learn, TensorFlow, PyTorch, and Hugging Face Transformers. Proficient in developing REST APIs using FastAPI or Flask. Experience in PDF/text extraction using PyMuPDF, Tesseract, or similar tools. Skilled in regex-based extraction and rule-based NER. Familiar with Git, Docker, and any major cloud platform (AWS, GCP, or Azure). Exposure to MLOps tools such as MLflow, Airflow, or LangChain. Job Type: Full-time Pay: ₹800,000.00 - ₹1,400,000.00 per year Benefits: Provident Fund Schedule: Day shift Monday to Friday Application Question(s): Are you an immediate Joiner? Experience: Python : 2 years (Required) AI/ML: 2 years (Required) NLP: 1 year (Required) Location: Hyderabad, Telangana (Required) Work Location: In person
Posted 1 week ago
7.0 years
0 Lacs
Bengaluru, Karnataka, India
On-site
Bangalore/Gurugram/Hyderabad YOE - 7+ years We are seeking a talented Data Engineer with strong expertise in Databricks, specifically in Unity Catalog, PySpark, and SQL, to join our data team. You’ll play a key role in building secure, scalable data pipelines and implementing robust data governance strategies using Unity Catalog. Key Responsibilities: Design and implement ETL/ELT pipelines using Databricks and PySpark. Work with Unity Catalog to manage data governance, access controls, lineage, and auditing across data assets. Develop high-performance SQL queries and optimize Spark jobs. Collaborate with data scientists, analysts, and business stakeholders to understand data needs. Ensure data quality and compliance across all stages of the data lifecycle. Implement best practices for data security and lineage within the Databricks ecosystem. Participate in CI/CD, version control, and testing practices for data pipelines Required Skills: Proven experience with Databricks and Unity Catalog (data permissions, lineage, audits). Strong hands-on skills with PySpark and Spark SQL. Solid experience writing and optimizing complex SQL queries. Familiarity with Delta Lake, data lakehouse architecture, and data partitioning. Experience with cloud platforms like Azure or AWS. Understanding of data governance, RBAC, and data security standards. Preferred Qualifications: Databricks Certified Data Engineer Associate or Professional. Experience with tools like Airflow, Git, Azure Data Factory, or dbt. Exposure to streaming data and real-time processing. Knowledge of DevOps practices for data engineering.
Posted 1 week ago
4.0 - 9.0 years
4 - 8 Lacs
Pune
Work from Office
Experience: 4+ Years. Expertise in Python Language is MUST. SQL (should be able to write complex SQL Queries) is MUST Hands on experience in Apache Flink Streaming Or Spark Streaming MUST Hands On expertise in Apache Kafka experience is MUST Data Lake Development experience. Orchestration (Apache Airflow is preferred). Spark and Hive: Optimization of Spark/PySpark and Hive apps Trino/(AWS Athena) (Good to have) Snowflake (good to have). Data Quality (good to have). File Storage (S3 is good to have) Our Offering:- Global cutting-edge IT projects that shape the future of digital and have a positive impact on environment. Wellbeing programs & work-life balance - integration and passion sharing events. Attractive Salary and Company Initiative Benefits Courses and conferences. Attractive Salary. Hybrid work culture.
Posted 1 week ago
10.0 years
0 Lacs
Gurugram, Haryana, India
On-site
✅ Job Title: Data Engineer – Apache Spark, Scala, GCP & Azure 📍 Location: Gurugram (Hybrid – 3 days/week in office) 🕒 Experience: 5–10 Years 🧑💻 Type: Full-time 📩 Apply: Share your resume with the details listed below to vijay.s@xebia.com 🕐 Availability: Immediate joiners or max 2 weeks' notice period only 🚀 About the Role Xebia is looking for a skilled Data Engineer to join our fast-paced team in Gurugram. You will work on building and optimizing scalable data pipelines, processing large datasets using Apache Spark and Scala , and deploying on cloud platforms like GCP and Azure . If you're passionate about clean architecture, high-quality data flow, and performance tuning, this is the opportunity for you. 🔧 Key Responsibilities Design and develop robust ETL pipelines using Apache Spark Write clean and efficient data processing code in Scala Handle large-scale data movement, transformation, and storage Build solutions on Google Cloud Platform (GCP) and Microsoft Azure Collaborate with teams to define data strategies and ensure data quality Optimize jobs for performance and cost on distributed systems Document technical designs and ETL flows clearly for the team ✅ Must-Have Skills Apache Spark Scala ETL design & development Cloud platforms: GCP & Azure Strong understanding of Data Engineering best practices Solid communication and collaboration skills 🌟 Good-to-Have Skills Apache tools (Kafka, Beam, Airflow, etc.) Knowledge of data lake and data warehouse concepts CI/CD for data pipelines Exposure to modern data monitoring and observability tools 💼 Why Xebia? At Xebia, you’ll be part of a forward-thinking, tech-savvy team working on high-impact, global data projects. We prioritize clean code, scalable solutions, and continuous learning. Join us to build real-time, cloud-native data platforms that power business intelligence across industries. 📤 To Apply Please share your updated resume and include the following details in your email to vijay.s@xebia.com : Full Name: Total Experience: Current CTC: Expected CTC: Current Location: Preferred Xebia Location: Gurugram Notice Period / Last Working Day (if serving): Primary Skills: LinkedIn Profile URL: Note: Only candidates who can join immediately or within 2 weeks will be considered. Build intelligent, scalable data solutions with Xebia – let’s shape the future of data together. 📊🚀
Posted 1 week ago
0 years
0 Lacs
Gurugram, Haryana, India
On-site
Hi All, We are hiring for Data Engineers, kindly refer the below skillsets: Mandatory Skills: GCP Cloud (especially BigQuery and DataProc) Big Data technologies Hadoop Hive Python / PySpark Airflow and DAG orchestration Preferred Skills: Experience with visualization tools such as Tableau or Power BI Familiarity with Jethro is a plus
Posted 1 week ago
10.0 - 14.0 years
0 Lacs
Hyderabad, Telangana, India
On-site
Role - Cloud Architect – Analytics & Data Products We’re looking for a Cloud Architect / Lead to design, build, and manage scalable AWS infrastructure that powers our analytics and data product initiatives. This role focuses on automating infrastructure provisioning , application/API hosting , and enabling data and GenAI workloads through a modern, secure cloud environment. Key Responsibilities Design and provision AWS infrastructure using Terraform or AWS CloudFormation to support evolving data product needs. Develop and manage CI/CD pipelines using Jenkins , AWS Code Pipeline , Code Build , or GitHub Actions . Deploy and host internal tools, APIs, and applications using ECS , EKS , Lambda , API Gateway , and ELB . Provision and support analytics and data platforms using S3 , Glue , Redshift , Athena , Lake Formation , and orchestration tools like Step Functions or Apache Airflow (MWAA) . Implement cloud security, networking, and compliance using IAM , VPC , KMS , CloudWatch , CloudTrail , and AWS Config . Collaborate with data engineers, ML engineers, and analytics teams to align infrastructure with application and data product requirements. Support GenAI infrastructure, including Amazon Bedrock , Sage Maker , or integrations with APIs like Open AI . Requirements 10-14 years of experience in cloud engineering, DevOps, or cloud architecture roles. Strong hands-on expertise with the AWS ecosystem and tools listed above. Proficiency in scripting (e.g., Python , Bash ) and infrastructure automation. Experience deploying containerized workloads using Docker , ECS , EKS , or Fargate . Familiarity with data engineering and GenAI workflows is a plus. AWS certifications (e.g., Solutions Architect , DevOps Engineer ) are preferred.
Posted 1 week ago
6.0 - 8.0 years
0 Lacs
Indore, Madhya Pradesh, India
On-site
Qualification 6-8 years of good hands on exposure with Big Data technologies – pySpark (Data frame and SparkSQL), Hadoop, and Hive Good hands on experience of python and Bash Scripts Good understanding of SQL and data warehouse concepts Strong analytical, problem-solving, data analysis and research skills Demonstrable ability to think outside of the box and not be dependent on readily available tools Excellent communication, presentation and interpersonal skills are a must Hands-on experience with using Cloud Platform provided Big Data technologies (i.e. IAM, Glue, EMR, RedShift, S3, Kinesis) Orchestration with Airflow and Any job scheduler experience Experience in migrating workload from on-premise to cloud and cloud to cloud migrations Good to have: Role Develop efficient ETL pipelines as per business requirements, following the development standards and best practices. Perform integration testing of different created pipeline in AWS env. Provide estimates for development, testing & deployments on different env. Participate in code peer reviews to ensure our applications comply with best practices. Create cost effective AWS pipeline with required AWS services i.e S3,IAM, Glue, EMR, Redshift etc. Experience 6 to 8 years Job Reference Number 13024
Posted 1 week ago
3.0 - 5.0 years
0 - 0 Lacs
chennai
On-site
About Hexr Factory: We are always exploring the possibilities to bridge the connection between the physical and digital world. We design and build Metaverse & Digital twin technologies for the future of industry and entertainment. Experience: 3-5 years Title: Data Engineers You are a successful candidate if you have 3+ years of experience in data engineering, preferably with real-time systems. Proficient with Python, SQL, and distributed data systems (Kinesis, Spark, Flink, etc.). Strong understanding of event-driven architectures, data lakes, and message serialization. Experience with sensor data processing, telemetry ingestion, or mobility data is a plus. Familiarity with Docker, CI/CD, Kubernetes, and cloud-native architectures. Familiarity with building data pipelines & its workflows (eg: Airflow). Preferred Qualifications: Exposure to smart city platforms, V2X ecosystems or other timeseries paradigms. Experience integrating data from Camera and other sensors. If interested, Please share your resume to jobs@hexrfactory.com Work location: Chennai, Tamil Nadu Contact: 9884099499 Web: www.hexrfactory.com
Posted 1 week ago
6.0 years
0 Lacs
Hyderabad, Telangana, India
On-site
Job Summary We are looking for a Data Engineer with strong experience in cloud platforms (AWS & Azure) , Scala programming , and a solid understanding of data architecture and governance frameworks . You will play a key role in building, optimizing, and maintaining scalable data pipelines and systems while ensuring data quality, security, and compliance across the organization. Key Responsibilities Data Engineering & Development Design and develop reliable, scalable ETL/ELT data pipelines using Scala , SQL , and orchestration tools. Integrate and process structured, semi-structured, and unstructured data from various sources (APIs, databases, flat files, etc.). Develop solutions on AWS (e.g., S3, Glue, Redshift, EMR) and Azure (e.g., Data Factory, Synapse, Blob Storage). Cloud & Infrastructure Build cloud-native data solutions that align with enterprise architecture standards. Leverage IaC tools (Terraform, CloudFormation, ARM templates) to deploy and manage infrastructure. Monitor performance, cost, and security posture of data environments in both AWS and Azure. Data Architecture & Governance Collaborate with data architects to define and implement logical and physical data models. Apply data governance principles including data cataloging , lineage tracking , data privacy , and compliance (e.g., GDPR) . Support the enforcement of data policies and data quality standards across data domains. Collaboration & Communication Work cross-functionally with data analysts, scientists, architects, and business stakeholders to support data needs. Participate in Agile ceremonies and contribute to sprint planning and reviews. Maintain clear documentation of pipelines, data models, and data flows. Required Qualifications Bachelor's degree in Computer Science, Engineering, or a related field. 3–6 years of experience in data engineering or data platform development. Hands-on experience with AWS and Azure data services. Proficient in Scala for data processing (e.g., Spark, Kafka Streams). Strong SQL skills and familiarity with distributed systems. Experience with orchestration tools (e.g., Apache Airflow, Azure Data Factory).
Posted 1 week ago
3.0 - 5.0 years
0 Lacs
Bengaluru, Karnataka, India
On-site
About The Role Data Analysts are the drivers of how data is leveraged in solving business problems within their area. They are able to use their experience to consult with stakeholders in problem-definition, setting success metrics and shaping the way forward through data insights and effective communication with their audience. We are looking for experienced data analysts who would be able to deep-dive into data to generate insights, run root cause analysis autonomously and manage business stakeholders largely independently (seeking for help in complex scenarios) based on their experience by prioritizing business impact and efficiently adapting to business needs. About The Team The successful candidate will be a key member of the Payments Accounting Data Analytics Team. They will be responsible for generating data-driven analysis, reporting, root cause analysis, and data reconciliations to support stakeholders, and help maintain the complex data ecosystem. B.Responsible Works independently on data collection and preparation. Uses their past experience and seeks for help in complex scenarios to translate business problems into data driven insights. Leverages available cloud big data platforms to run root cause analysis, data reconciliations and shares the insights with the business team. Maintains and drives key reports, metrics and workflows running within their scope Is able to communicate results and outcomes clearly to stakeholders based on their knowledge and experience. Actively participates in business and/or analytics team activities and suggests ways of achieving objectives (standup, planning meeting, retrospectives) Networks and proactively connects with craft peers beyond the team scope Has strong understanding of the big data ecosystems Collaborates and is open to giving and receiving feedback with peers and direct stakeholders. Is flexible in adopting and proposing new approaches and expanding their technical competencies when a more efficient way presents itself Expected to get significant deep knowledge about the operational, tactical and strategic workings of the department. Has a main focus on business and technical opportunities. B.Skilled Educational background in Quantitative field - Preferably a Master's degree 3-5 years of experience in data analytics, Insight generation and data visualization Should have executed big data analytics projects in Industry setting Advanced knowledge of SQL, ideally with experience in Snowflake Good knowledge with Python/Py-Spark Experience of working with ETL and Data Modelling tools like Airflow, Dagster and DBT Knowledge and experience using data analysis and visualization tools (e.g: tableau, data studio, powerbi, mixpanel, etc) Familiarity with Cloud data platforms like AWS and GIT version control is a plus Familiarity with financial metrics is a big plus Strong communication and stakeholder management skills Able to understand details while keeping an eye on the bigger picture Pre-Employment Screening If your application is successful, your personal data may be used for a pre-employment screening check by a third party as permitted by applicable law. Depending on the vacancy and applicable law, a pre-employment screening may include employment history, education and other information (such as media information) that may be necessary for determining your qualifications and suitability for the position.
Posted 1 week ago
0 years
0 Lacs
Pune, Maharashtra, India
On-site
Role: Snowflake Developer Location: Pune (Kalyani Nagar) Exp: 6+ Key Responsibilities: •Design, develop, and maintain scalable Snowflake data warehouse solutions. •Write and optimize complex SQL queries for data extraction, transformation, and reporting. •Develop and manage Snowflake stored procedures using SQL and JavaScript. •Implement and manage data integration between Snowflake and external systems (e.g., using ETL tools, APIs, or Snowpipe). •Create and maintain data models and ensure data quality and consistency across environments. •Collaborate with data engineers, analysts, and business stakeholders to understand requirements and deliver reliable solutions. •Monitor performance, diagnose issues, and implement performance tuning best practices. •Implement access controls and security policies aligned with enterprise standards. Required Skills & Qualifications: •Strong hands-on experience with Snowflake platform and architecture. •Should know python libraries. •Advanced proficiency in SQL, including writing and optimizing complex queries. •Experience with stored procedures, user-defined functions (UDFs), and task scheduling in Snowflake. •Familiarity with data integration tools (e.g., Informatica, Talend, Apache Airflow, DBT, Fivetran, or custom Python scripts). •Experience with data modeling (star/snowflake schemas) and data warehouse design. •Knowledge of cloud platforms (AWS, Azure, or GCP) and how Snowflake integrates with them. •Experience working with large datasets and performance tuning of data loads/queries. •Strong problem-solving and communication skills. Please share the resume to hema@synapsetechservice.com
Posted 1 week ago
0 years
0 Lacs
Gurugram, Haryana, India
On-site
Backend & MLOps Engineer – Integration, API, and Infrastructure Expert 1. Role Objective: Responsible for building robust backend infrastructure, managing ML operations, and creating scalable APIs for AI applications. Must excel in deploying and maintaining AI products in production environments with high availability and security standards. The engineer will be expected to build secure, scalable backend systems that integrate AI models into services (REST, gRPC), manage data pipelines, enable model versioning, and deploy containerized applications in secure (air-gapped) Naval infrastructure. 2. Key Responsibilities: 2.1. Create RESTful and/or gRPC APIs for model services. 2.2. Containerize AI applications and maintain Kubernetes-compatible Docker images. 2.3. Develop CI/CD pipelines for model training and deployment. 2.4. Integrate models as microservices using TorchServe, Triton, or FastAPI. 2.5. Implement observability (metrics, logs, alerts) for deployed AI pipelines. 2.6. Build secured data ingestion and processing workflows (ETL/ELT). 2.7. Optimize deployments for CPU/GPU performance, power efficiency, and memory usage 3. Educational Qualifications Essential Requirements: 3.1. B.Tech/ M.Tech in Computer Science, Information Technology, or Software Engineering. 3.2. Strong foundation in distributed systems, databases, and cloud computing. 3.3. Minimum 70% marks or 7.5 CGPA in relevant disciplines. Professional Certifications: 3.4. AWS Solutions Architect/DevOps Engineer Professional 3.5. Google Cloud Professional ML Engineer or DevOps Engineer 3.6. Azure AI Engineer or DevOps Engineer Expert. 3.7. Kubernetes Administrator (CKA) or Developer (CKAD). 3.8. Docker Certified Associate Core Skills & Tools 4. Backend Development: 4.1. Languages: Python, FastAPI, Flask, Go, Java, Node.js, Rust (for performance-critical components) 4.2. Web Frameworks: FastAPI, Django, Flask, Spring Boot, Express.js. 4.3. API Development: RESTful APIs, GraphQL, gRPC, WebSocket connections. 4.4. Authentication & Security: OAuth 2.0, JWT, API rate limiting, encryption protocols. 5. MLOps & Model Management: 5.1. ML Platforms: MLflow, Kubeflow, Apache Airflow, Prefect 5.2. Model Serving: TensorFlow Serving, TorchServe, ONNX Runtime, NVIDIA Triton, BentoML 5.3. Experiment Tracking: Weights & Biases, Neptune, ClearML 5.4. Feature Stores: Feast, Tecton, Amazon SageMaker Feature Store 5.5. Model Monitoring: Evidently AI, Arize, Fiddler, custom monitoring solutions 6. Infrastructure & DevOps: 6.1. Containerization: Docker, Podman, container optimization. 6.2. Orchestration: Kubernetes, Docker Swarm, OpenShift. 6.3. Cloud Platforms: AWS, Google Cloud, Azure (multi-cloud expertise preferred). 6.4. Infrastructure as Code: Terraform, CloudFormation, Pulumi, Ansible. 6.5. CI/CD: Jenkins, GitLab CI, GitHub Actions, ArgoCD. 6.6. DevOps & Infra: Docker, Kubernetes, NGINX, GitHub Actions, Jenkins. 7. Database & Storage: 7.1. Relational: PostgreSQL, MySQL, Oracle (for enterprise applications) 7.2. NoSQL: MongoDB, Cassandra, Redis, Elasticsearch 7.3. Vector Databases: Pinecone, Weaviate, Chroma, Milvus 7.4. Data Lakes: Apache Spark, Hadoop, Delta Lake, Apache Iceberg 7.5. Object Storage: AWS S3, Google Cloud Storage, MinIO 7.6. Backend: Python (FastAPI, Flask), Node.js (optional) 7.7. DevOps & Infra: Docker, Kubernetes, NGINX, GitHub Actions, Jenkins 8. Secure Deployment: 8.1. Military-grade security protocols and compliance 8.2. Air-gapped deployment capabilities 8.3. Encrypted data transmission and storage 8.4. Role-based access control (RBAC) & IDAM integration 8.5. Audit logging and compliance reporting 9. Edge Computing: 9.1. Deployment on naval vessels with air gapped connectivity. 9.2. Optimization of applications for resource-constrained environment. 10. High Availability Systems: 10.1. Mission-critical system design with 99.9% uptime. 10.2. Disaster recovery and backup strategies. 10.3. Load balancing and auto-scaling. 10.4. Failover mechanisms for critical operations. 11. Cross-Compatibility Requirements: 11.1. Define and expose APIs in a documented, frontend-consumable format (Swagger/OpenAPI). 11.2. Develop model loaders for AI Engineer's ONNX/ serialized models. 11.3. Provide UI developers with test environments, mock data, and endpoints. 11.4. Support frontend debugging, edge deployment bundling, and user role enforcement. 12. Experience Requirements 12.1. Production experience with cloud platforms and containerization. 12.2. Experience building and maintaining APIs serving millions of requests. 12.3. Knowledge of database optimization and performance tuning. 12.4. Experience with monitoring and alerting systems. 12.5. Architected and deployed large-scale distributed systems. 12.6. Led infrastructure migration or modernization projects. 12.7. Experience with multi-region deployments and disaster recovery. 12.8. Track record of optimizing system performance and cost
Posted 1 week ago
6.0 years
0 Lacs
Bangalore Urban, Karnataka, India
On-site
Take ownership of pipeline stability and performance across our GCP-based stack (BigQuery, GCS, Dataprep/Dataflow) Lead the enhancement of our existing ETL workflows to support better modularity, reusability, and error handling Help introduce lightweight governance practices —including column-level validation, source tracking, and transformation transparency Support development of a semantic layer (e.g., KPI definitions, normalized metric naming) to reduce rework and support downstream users Work with analysts and dashboard developers to structure data outputs for intuitive use and parameterization Collaborate with team leadership to prioritize improvements based on impact and feasibility Support platform readiness for automated reporting , predictive modeling , and AI-enhanced analysis Contribute to team culture through clear documentation, mentoring, and code review Participate in hiring, onboarding, and evolving our internal standards What We’re Looking For Must-Have: 4–6+ years of experience in data engineering, preferably in a fast-paced agency or multi-client environment Solid command of Google Cloud Platform , especially BigQuery, GCS, and Cloud Dataprep (Alteryx) or Dataflow Strong SQL and Python skills with a focus on transformation and data reliability Experience building and maintaining ETL pipelines in production Familiarity with metadata-driven development , version control, and task orchestration (Airflow or equivalent) Proven ability to balance individual execution with team collaboration Clear communicator, able to translate technical trade-offs to non-technical stakeholders Nice-to-Have: Experience applying basic data governance principles (e.g., lineage tracking, validation frameworks, naming conventions) Exposure to building or maintaining a semantic layer (via dbt, LookML, etc.) Familiarity with AI/ML workflows or tooling for automated insight generation Understanding of marketing or media datasets Experience developing custom marketing attribution models Experience mentoring junior team members or participating in code/process standardization
Posted 1 week ago
2.0 years
0 Lacs
India
Remote
Job description L1 Support – Data Engineering (Remote, South India) Location: Permanently based in South India (any city) – non-negotiable Work Mode: Remote | 6 days/week | 24x7x365 support (rotational shifts) Salary Range - Between INR 2.5 to 3 Lacs Per Annum Experience: 2 years Language: English proficiency mandatory ; Hindi is a plus About the Role We're looking for an experienced and motivated L1 Support Engineer – Data Engineering to join our growing team. If you have solid exposure to AWS , SQL , and Python scripting , and you're ready to thrive in a 24x7 support environment—this role is for you! What You’ll Do Monitor and support AWS services (S3, EC2, CloudWatch, IAM) Handle SQL-based issue resolution and data analysis Run and maintain Python scripts ; Shell scripting is a plus Support ETL pipelines and data workflows Monitor Apache Airflow DAGs and resolve basic issues Collaborate with cross-functional and multicultural teams What We’re Looking For B.Tech or MCA preferred , but candidates with a Bachelor’s degree in any field and the right skillset are welcome to apply. 2 years of Data Engineering Support or similar experience Strong skills in AWS , SQL , Python , and ETL processes Familiarity with data warehousing (Amazon Redshift or similar) Ability to work rotational shifts in a 6-day, 24x7 environment Excellent communication and problem-solving skills English fluency is required ; Hindi is an advantage
Posted 1 week ago
80.0 years
0 Lacs
Bengaluru, Karnataka, India
On-site
Job Title Associate Data Engineer (Internship Program to Full-time Employee) Job Description For more than 80 years, Kaplan has been a trailblazer in education and professional advancement. We are a global company at the intersection of education and technology, focused on collaboration, innovation, and creativity to deliver a best in class educational experience and make Kaplan a great place to work. Our offices in India opened in Bengaluru in 2018. Since then, our team has fueled growth and innovation across the organization, impacting students worldwide. We are eager to grow and expand with skilled professionals like you who use their talent to build solutions, enable effective learning, and improve students’ lives. The future of education is here and we are eager to work alongside those who want to make a positive impact and inspire change in the world around them. The Associate Data Engineer at Kaplan North America (KNA) within the Analytics division will work with world class psychometricians, data scientists and business analysts to forever change the face of education. This role is a hands-on technical expert who will help implement an Enterprise Data Warehouse powered by AWS RA3 as a key feature of our Lake House architecture. The perfect candidate possesses strong technical knowledge in data engineering, data observability, Infrastructure automation, data ops methodology, systems architecture, and development. You should be expert at designing, implementing, and operating stable, scalable, low cost solutions to flow data from production systems into the data warehouse and into end-user facing applications. You should be able to work with business customers in a fast-paced environment understanding the business requirements and implementing data & reporting solutions. Above all you should be passionate about working with big data and someone who loves to bring datasets together to answer business questions and drive change Responsibilities You design, implement, and deploy data solutions. You solve difficult problems generating positive feedback. Build different types of data warehousing layers based on specific use cases Lead the design, implementation, and successful delivery of large-scale, critical, or difficult data solutions involving a significant amount of work Build scalable data infrastructure and understand distributed systems concepts from a data storage and compute perspective Utilize expertise in SQL and have a strong understanding of ETL and data modeling Ensure the accuracy and availability of data to customers and understand how technical decisions can impact their business’s analytics and reporting Be proficient in at least one scripting/programming language to handle large volume data processing. 30-day notification period preferred Requirements In-depth knowledge of the AWS stack (RA3, Redshift, Lambda, Glue, SnS). Experience in data modeling, ETL development and data warehousing. Effective troubleshooting and problem-solving skills Strong customer focus, ownership, urgency and drive. Excellent verbal and written communication skills and the ability to work well in a team Preferred Qualification Proficiency with Airflow, Tableau & SSRS Location Bangalore, KA, India Additional Locations Employee Type Employee Job Functional Area Systems Administration/Engineering Business Unit 00091 Kaplan Higher ED At Kaplan, we recognize the importance of attracting and retaining top talent to drive our success in a competitive market. Our salary structure and compensation philosophy reflect the value we place on the experience, education, and skills that our employees bring to the organization, taking into consideration labor market trends and total rewards. All positions with Kaplan are paid at least $15 per hour or $31,200 per year for full-time positions. Additionally, certain positions are bonus or commission-eligible. And we have a comprehensive benefits package, learn more about our benefits here. Diversity & Inclusion Statement Kaplan is committed to cultivating an inclusive workplace that values diversity, promotes equity, and integrates inclusivity into all aspects of our operations. We are an equal opportunity employer and all qualified applicants will receive consideration for employment regardless of age, race, creed, color, national origin, ancestry, marital status, sexual orientation, gender identity or expression, disability, veteran status, nationality, or sex. We believe that diversity strengthens our organization, fuels innovation, and improves our ability to serve our students, customers, and communities. Learn more about our culture here. Kaplan considers qualified applicants for employment even if applicants have an arrest or conviction in their background check records. Kaplan complies with related background check regulations, including but not limited to, the Los Angeles County Fair Chance Ordinance for Employers and the California Fair Chance Act. There are various positions where certain convictions may disqualify applicants, such as those positions requiring interaction with minors, financial records, or other sensitive and/or confidential information. Kaplan is a drug-free workplace and complies with applicable laws.
Posted 1 week ago
Upload Resume
Drag or click to upload
Your data is secure with us, protected by advanced encryption.
Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.
We have sent an OTP to your contact. Please enter it below to verify.
Accenture
39817 Jobs | Dublin
Wipro
19388 Jobs | Bengaluru
Accenture in India
15458 Jobs | Dublin 2
EY
14907 Jobs | London
Uplers
11185 Jobs | Ahmedabad
Amazon
10459 Jobs | Seattle,WA
IBM
9256 Jobs | Armonk
Oracle
9226 Jobs | Redwood City
Accenture services Pvt Ltd
7971 Jobs |
Capgemini
7704 Jobs | Paris,France