Assistant Manager - SA2 (AI Hub - GDC)

6 - 8 years

0 Lacs

Posted:2 weeks ago| Platform: Foundit logo

Apply

Work Mode

On-site

Job Type

Full Time

Job Description

Roles & responsibilities
  • Design, develop, and implement data pipelines for scalable solutions using various technologies Databricks or Microsoft Fabric
  • Conduct Code reviews and recommend best coding practices
  • Provide effort estimates to implement solutions, development effort for the proposed solutions
  • Develop and maintain comprehensive documentation for all proposed solutions. models, including detailed technical specifications, results of testing and evaluation, and datasets used for developing the solutions
  • Lead Architect and design efforts for Product development and application development for relevant use cases providing guidance and support to team members and clients
  • Implement best practices of Data Engineering and Architectural solution design, development, Testing and documentation
  • Participate in team meetings, brainstorming sessions, and project planning activities
  • Stay up-to-date with the latest advancements in Data Engineering area, to drive innovation and maintain a competitive edge
  • Stay hands-on with the design, development and validation of systems and models deployed
  • Collaborate with audit professionals to understand business, regulatory and risk requirements and key alignment considerations for Audit
  • Drive efforts in the Data Engineering and Architecture practice area
Mandatory technical & functional skills
  • Databricks/Fabric/Spark Notebooks
  • SQL/NoSQL databases, Redis,
  • Full-Stack Development - React, Angular
  • Data Management: Design, implement, and manage AI-driven data solutions on the Microsoft Azure cloud platform, ensuring scalability and performance.
  • Data Integration: Develop and maintain data pipelines for AI applications, ensuring efficient data extraction, transformation, and loading (ETL) processes using Azure Data Factory.
  • Big Data Processing: Utilize big data technologies such as Azure Databricks and Apache Spark to handle, analyze, and process large datasets for machine learning and AI applications.
  • machine learning frameworks -TensorFlow, PyTorch
  • Backend frameworks - FastAPI, Django


Preferred Technical & Functional Skills

  • Experience working with frameworks LangChain/LlamaIndex /LlamaPrase/LlamaCloud/Semantic Kernel etc.
  • Develop real-time data ingestion and stream-analytic solutions leveraging technologies such as Kafka, Apache Spark (SQL, Scala, Java), Python and Hadoop Platform and any Cloud Data Platform.?
  • Certifications: Relevant certifications such as Microsoft Certified: AI 102, DP 700, DP 900 or AWS certiications
Key behavioral attributes/requirements
  • Strong analytical, problem-solving, and critical-thinking skills
  • Excellent collaboration skills, with the ability to work effectively in a team-oriented environment
  • Excellent written and verbal communication skills, with the ability to present complex technical concepts to non-technical audiences
  • Willingness to learn new technologies and work on them


#KGS

RESPONSIBILITIES

Roles & responsibilities
  • Design, develop, and implement data pipelines for scalable solutions using various technologies Databricks or Microsoft Fabric
  • Conduct Code reviews and recommend best coding practices
  • Provide effort estimates to implement solutions, development effort for the proposed solutions
  • Develop and maintain comprehensive documentation for all proposed solutions. models, including detailed technical specifications, results of testing and evaluation, and datasets used for developing the solutions
  • Lead Architect and design efforts for Product development and application development for relevant use cases providing guidance and support to team members and clients
  • Implement best practices of Data Engineering and Architectural solution design, development, Testing and documentation
  • Participate in team meetings, brainstorming sessions, and project planning activities
  • Stay up-to-date with the latest advancements in Data Engineering area, to drive innovation and maintain a competitive edge
  • Stay hands-on with the design, development and validation of systems and models deployed
  • Collaborate with audit professionals to understand business, regulatory and risk requirements and key alignment considerations for Audit
  • Drive efforts in the Data Engineering and Architecture practice area
Mandatory technical & functional skills
  • Databricks/Fabric/Spark Notebooks
  • SQL/NoSQL databases, Redis,
  • Full-Stack Development - React, Angular
  • Data Management: Design, implement, and manage AI-driven data solutions on the Microsoft Azure cloud platform, ensuring scalability and performance.
  • Data Integration: Develop and maintain data pipelines for AI applications, ensuring efficient data extraction, transformation, and loading (ETL) processes using Azure Data Factory.
  • Big Data Processing: Utilize big data technologies such as Azure Databricks and Apache Spark to handle, analyze, and process large datasets for machine learning and AI applications.
  • machine learning frameworks -TensorFlow, PyTorch
  • Backend frameworks - FastAPI, Django


Preferred Technical & Functional Skills

  • Experience working with frameworks LangChain/LlamaIndex /LlamaPrase/LlamaCloud/Semantic Kernel etc.
  • Develop real-time data ingestion and stream-analytic solutions leveraging technologies such as Kafka, Apache Spark (SQL, Scala, Java), Python and Hadoop Platform and any Cloud Data Platform.?
  • Certifications: Relevant certifications such as Microsoft Certified: AI 102, DP 700, DP 900 or AWS certiications
Key behavioral attributes/requirements
  • Strong analytical, problem-solving, and critical-thinking skills
  • Excellent collaboration skills, with the ability to work effectively in a team-oriented environment
  • Excellent written and verbal communication skills, with the ability to present complex technical concepts to non-technical audiences
  • Willingness to learn new technologies and work on them


#KGS

QUALIFICATIONS

This role is for you if you have the below

Educational Qualifications

  • Minimum qualification required: BTech in Computer Science/ MTech/ MCA - Fulltime education.

Work Experience

  • 6 to 8 years of experience in design, develop data centric applications using various tools and technologies eg: Databases,reporting,ETL/ELT,NoSQL etc.
  • 4+ years of experience in designing, architecting solutions using Microsoft Data technologies like ADF/SYNAPSE
  • Relevant Data Professional certifications AWS, GCP or Azure

Mock Interview

Practice Video Interview with JobPe AI

Start Job-Specific Interview
cta

Start Your Job Search Today

Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.

Job Application AI Bot

Job Application AI Bot

Apply to 20+ Portals in one click

Download Now

Download the Mobile App

Instantly access job listings, apply easily, and track applications.

coding practice

Enhance Your Skills

Practice coding challenges to boost your skills

Start Practicing Now
KPMG India logo
KPMG India

Professional Services

Pune

RecommendedJobs for You