Jobs
Interviews
7 Job openings at Hakkda
About Hakkda

Hakkda is a technology company specializing in innovative solutions that leverage artificial intelligence to empower businesses and enhance productivity.

Senior Consultant, Data Scientist

Kolkata

5 - 10 years

INR 7.0 - 12.0 Lacs P.A.

Work from Office

Full Time

ABOUT HAKKODA Hakkoda, an IBM Company, is a modern data consultancy that empowers data driven organizations to realize the full value of the Snowflake Data Cloud. We provide consulting and managed services in data architecture, data engineering, analytics and data science. We are renowned for bringing our clients deep expertise, being easy to work with, and being an amazing place to work! We are looking for curious and creative individuals who want to be part of a fast-paced, dynamic environment, where everyone s input and efforts are valued. We hire outstanding individuals and give them the opportunity to thrive in a collaborative atmosphere that values learning, growth, and hard work. Our team is distributed across North America, Latin America, India and Europe. If you have the desire to be a part of an exciting, challenging, and rapidly-growing Snowflake consulting services company, and if you are passionate about making a difference in this world, we would love to talk to you!. We are seeking an exceptional and highly motivated Lead Data Scientist with a PhD in Data Science, Computer Science, Applied Mathematics, Statistics, or a closely related quantitative field, to spearhead the design, development, and deployment of an automotive OEM s next-generation Intelligent Forecast Application. This pivotal role will leverage cutting-edge machine learning, deep learning, and statistical modeling techniques to build a robust, scalable, and accurate forecasting system crucial for strategic decision-decision-making across the automotive value chain, including demand planning, production scheduling, inventory optimization, predictive maintenance, and new product introduction. The successful candidate will be a recognized expert in advanced forecasting methodologies, possess a strong foundation in data engineering and MLOps principles, and demonstrate a proven ability to translate complex research into tangible, production-ready applications within a dynamic industrial environment. This role demands not only deep technical expertise but also a visionary approach to leveraging data and AI to drive significant business impact for a leading automotive OEM. Role Description: Strategic Leadership & Application Design: Lead the end-to-end design and architecture of the Intelligent Forecast Application, defining its capabilities, modularity, and integration points with existing enterprise systems (e.g., ERP, SCM, CRM). Develop a strategic roadmap for forecasting capabilities, identifying opportunities for innovation and the adoption of emerging AI/ML techniques (e.g., generative AI for scenario planning, reinforcement learning for dynamic optimization). Translate complex business requirements and automotive industry challenges into well-defined data science problems and technical specifications. Advanced Model Development & Research: Design, develop, and validate highly accurate and robust forecasting models using a variety of advanced techniques, including: Time Series Analysis: ARIMA, SARIMA, Prophet, Exponential Smoothing, State-space models. Machine Learning: Gradient Boosting (XGBoost, LightGBM), Random Forests, Support Vector Machines. Deep Learning: LSTMs, GRUs, Transformers, and other neural network architectures for complex sequential data. Probabilistic Forecasting: Quantile regression, Bayesian methods to capture uncertainty. Hierarchical & Grouped Forecasting: Managing forecasts across multiple product hierarchies, regions, and dealerships. Incorporate diverse data sources, including historical sales, market trends, economic indicators, competitor data, internal operational data (e.g., production schedules, supply chain disruptions), external events, and unstructured data. Conduct extensive exploratory data analysis (EDA) to identify patterns, anomalies, and key features influencing automotive forecasts. Stay abreast of the latest academic researchand industry advancements in forecasting, machine learning, and AI, actively evaluating and advocating for their practical application within the OEM. Application Development & Deployment (MLOps): Architect and implement scalable data pipelines for ingestion, cleaning, transformation, and feature engineering of large, complex automotive datasets. Develop robust and efficient code for model training, inference, and deployment within a production environment. Implement MLOps best practices for model versioning, monitoring, retraining, and performance management to ensure the continuous accuracy and reliability of the forecasting application. Collaborate closely with Data Engineering, Software Development, and IT Operations teams to ensure seamless integration, deployment, and maintenance of the application. Performance Evaluation & Optimization: Define and implement rigorous evaluation metrics for forecasting accuracy (e.g., MAE, RMSE, MAPE, sMAPE, wMAPE, Pinball Loss) and business impact. Perform A/B testing and comparative analyses of different models and approaches to continuously improve forecasting performance. Identify and mitigate sources of bias and uncertainty in forecasting models. Collaboration & Mentorship: Work cross-functionally with various business units (e.g., Sales, Marketing, Supply Chain, Manufacturing, Finance, Product Development) to understand their forecasting needs and integrate solutions. Communicate complex technical concepts and model insights clearly and concisely to both technical and non-technical stakeholders. Provide technical leadership and mentorship to junior data scientists and engineers, fostering a culture of innovation and continuous learning. Potentially contribute to intellectual property (patents) and present findings at internal and external conferences. Qualifications Education : PhD in Data Science, Computer Science, Statistics, Applied Mathematics, Operations Research, or a closely related quantitative field. Experience : 5+ years of hands-on experience in a Data Scientist or Machine Learning Engineer role, with a significant focus on developing and deploying advanced forecasting solutions in a production environment. Demonstrated experience designing and developing intelligent applications, not just isolated models. Experience in the automotive industry or a similar complex manufacturing/supply chain environment is highly desirable. Technical Skills: Expert proficiency in Python (Numpy, Pandas, Scikit-learn, Statsmodels) and/or R. Strong proficiency in SQL. Machine Learning/Deep Learning Frameworks: Extensive experience with TensorFlow, PyTorch, Keras, or similar deep learning libraries. Forecasting Specific Libraries: Proficiency with forecasting libraries like Prophet, Statsmodels, or specialized time series packages. Data Warehousing & Big Data Technologies: Experience with distributed computing frameworks (e.g., Apache Spark, Hadoop) and data storage solutions (e.g., Snowflake, Databricks, S3, ADLS). Cloud Platforms: Hands-on experience with at least one major cloud provider (Azure, AWS, GCP) for data science and ML deployments. MLOps: Understanding and practical experience with MLOps tools and practices (e.g., MLflow, Kubeflow, Docker, Kubernetes, CI/CD pipelines). Data Visualization: Proficiency with tools like Tableau, Power BI, or similar for creating compelling data stories and dashboards. Analytical Prowess: Deep understanding of statistical inference, experimental design, causal inference, and the mathematical foundations of machine learning algorithms. Problem Solving: Proven ability to analyze complex, ambiguous problems, break them down into manageable components, and devise innovative solutions. Preferred Qualifications: Publications in top-tier conferences or journals related to forecasting, time series analysis, or applied machine learning. Experience with real-time forecasting systems or streaming data analytics. Familiarity with specific automotive data types (e.g., telematics, vehicle sensor data, dealership data, market sentiment). Experience with distributed version control systems (e.g., Git). Knowledge of agile development methodologies. Soft Skills Exceptional Communication: Ability to articulate complex technical concepts and insights to a diverse audience, including senior management and non-technical stakeholders. Collaboration: Strong interpersonal skills and a proven ability to work effectively within cross-functional teams. Intellectual Curiosity & Proactiveness: A passion for continuous learning, staying ahead of industry trends, and proactively identifying opportunities for improvement. Strategic Thinking: Ability to see the big picture and align technical solutions with overall business objectives. Mentorship: Desire and ability to guide and develop less experienced team members. Resilience & Adaptability: Thrive in a fast-paced, evolving environment with complex challenges. Benefits: - Health Insurance - Paid leave - Technical training and certifications - Robust learning and development opportunities - Incentive - Toastmasters - Food Program - Fitness Program - Referral Bonus Program Hakkoda is committed to fostering diversity, equity, and inclusion within our teams. A diverse workforce enhances our ability to serve clients and enriches our culture. We encourage candidates of all races, genders, sexual orientations, abilities, and experiences to apply, creating a workplace where everyone can succeed and thrive. Ready to take your career to the next level? Apply today and join a team that s shaping the future!! Hakkoda is an IBM subsidiary which has been acquired by IBM and will be integrated in the IBM organization. Hakkoda will be the hiring entity. By Proceeding with this application, you understand that Hakkoda will share your personal information with other IBM subsidiaries involved in your recruitment process, wherever these are located. More information on how IBM protects your personal information, including the safeguards in case of cross-border data transfer, are available here.

Consultant, Data Scientist

Kolkata

2 - 5 years

INR 4.0 - 7.0 Lacs P.A.

Work from Office

Full Time

ABOUT HAKKODA Hakkoda, an IBM Company, is a modern data consultancy that empowers data driven organizations to realize the full value of the Snowflake Data Cloud. We provide consulting and managed services in data architecture, data engineering, analytics and data science. We are renowned for bringing our clients deep expertise, being easy to work with, and being an amazing place to work! We are looking for curious and creative individuals who want to be part of a fast-paced, dynamic environment, where everyone s input and efforts are valued. We hire outstanding individuals and give them the opportunity to thrive in a collaborative atmosphere that values learning, growth, and hard work. Our team is distributed across North America, Latin America, India and Europe. If you have the desire to be a part of an exciting, challenging, and rapidly-growing Snowflake consulting services company, and if you are passionate about making a difference in this world, we would love to talk to you!. We are seeking a skilled Data Scientist with 2 to 5 years of experience, specializing in Machine Learning, PySpark, and Databricks, with a proven track record in long-range demand and sales forecasting. This role is crucial for the development and implementation of an automotive OEM s next-generation Intelligent Forecast Application. The position will involve building, optimizing, and deploying large-scale machine learning models for complex, long-term forecasting challenges using distributed computing frameworks, specifically PySpark on the Databricks platform. The work will directly support strategic decision-making across the automotive value chain, including areas like long-term demand planning, production scheduling, and inventory optimization.The ideal candidate will have hands-on experience developing and deploying ML models for forecasting, particularly long-range predictions, in a production environment using PySpark and Databricks. This role requires strong technical skills in machine learning, big data processing, and time series forecasting, combined with the ability to work effectively within a technical team to deliver robust and scalable long-range forecasting solutions. Role Description: Machine Learning Model Development & Implementation for Long-Range Forecasting: Design, develop, and implement scalable and accurate machine learning models specifically for long-range demand and sales forecasting challenges. Apply advanced time series analysis techniques and integrate them with machine learning models leveraging PySpark for data processing and model training on large datasets within the Databricks environment. Implement probabilistic forecasting methods using PySpark to capture uncertainty in long-range predictions. Develop robust solutions for hierarchical and grouped long-range forecasting on distributed data. Data Processing and Feature Engineering with PySpark: Build and optimize large-scale data pipelines for ingesting, cleaning, transforming, and engineering features relevant to long-range forecasting from diverse, complex automotive datasets using PySpark on Databricks. Deployment and MLOps on Databricks: Develop and implement robust code for model training, inference, and deployment of long-range forecasting models directly within the Databricks platform. Apply MLOps principles compatible with Databricks workflows for model versioning, monitoring, retraining, and managing the lifecycle of long-range ML forecasting models in production. Collaborate with Data Engineering and IT Operations to ensure seamless deployment and operational efficiency of the forecasting application on Databricks. Performance Evaluation & Optimization: Evaluate long-range forecasting model performance using relevant metrics (e.g., MAE, RMSE, MAPE, considering metrics suitable for longer horizons) and optimize models and data processing pipelines for improved accuracy and efficiency within the PySpark/Databricks ecosystem. Technical Collaboration: Work effectively as part of a technical team, collaborating with other data scientists, data engineers, and software developers to integrate ML long-range forecasting solutions into the broader forecasting application built on Databricks. Communicate technical details and forecasting results effectively within the technical team. Qualifications Education: Bachelors or Masters degree in Data Science, Computer Science, Statistics, Applied Mathematics, or a closely related quantitative field. Experience:2 to 5 years of hands-on experience in a Data Scientist or Machine Learning Engineer role. Proven experience developing and deploying machine learning models in a production environment. Demonstrated experience in long-range demand and sales forecasting. Significant hands-on experience with PySpark for large-scale data processing and machine learning. Extensive practical experience working with the Databricks platform, including notebooks, jobs, and ML capabilities. Expert proficiency in PySpark . Expert proficiency in the Databricks platform . Strong proficiency in Python and SQL. Experience with machine learning libraries compatible with PySpark (e.g., MLlib, or integrating other libraries). Experience with advanced time series forecasting techniques and their implementation. Experience with distributed computing concepts and optimization techniques relevant to PySpark. Hands-on experience with a major cloud provider (Azure, AWS, or GCP) in the context of using Databricks. Familiarity with MLOps concepts and tools used in a Databricks environment. Experience with data visualization tools. Analytical skills with a deep understanding of machine learning algorithms and their application to forecasting. Ability to troubleshoot and solve complex technical problems related to big data and machine learning workflows. Preferred Qualifications Experience with specific long-range forecasting methodologies and libraries used in a distributed environment. Experience with real-time or streaming data processing using PySpark for near-term forecasting components that might complement long-range models. Familiarity with automotive data types relevant to long-range forecasting (e.g., economic indicators affecting car sales, long-term market trends). Experience with distributed version control systems (e.g., Git). Knowledge of agile development methodologies. Soft Skills: Collaboration: Ability to work effectively as part of a technical team. Communication: Clear and concise communication of technical details and forecasting results. Problem-Solving: Ability to tackle complex technical challenges and find efficient solutions. Learning Agility: Eagerness to learn and adapt to new technologies and methodologies within the PySpark/Databricks ecosystem and advancements in long-range forecasting. Ability to understand business needs related to long-term planning. Benefits: - Health Insurance - Paid leave - Technical training and certifications - Robust learning and development opportunities - Incentive - Toastmasters - Food Program - Fitness Program - Referral Bonus Program Hakkoda is committed to fostering diversity, equity, and inclusion within our teams. A diverse workforce enhances our ability to serve clients and enriches our culture. We encourage candidates of all races, genders, sexual orientations, abilities, and experiences to apply, creating a workplace where everyone can succeed and thrive. Ready to take your career to the next level? Apply today and join a team that s shaping the future!! Hakkoda is an IBM subsidiary which has been acquired by IBM and will be integrated in the IBM organization. Hakkoda will be the hiring entity. By Proceeding with this application, you understand that Hakkoda will share your personal information with other IBM subsidiaries involved in your recruitment process, wherever these are located. More information on how IBM protects your personal information, including the safeguards in case of cross-border data transfer, are available here.

AWS Solution Architect

Jaipur

7 - 12 years

INR 13.0 - 17.0 Lacs P.A.

Work from Office

Full Time

ABOUT HAKKODA Hakkoda, an IBM Company, is a modern data consultancy that empowers data driven organizations to realize the full value of the Snowflake Data Cloud. We provide consulting and managed services in data architecture, data engineering, analytics and data science. We are renowned for bringing our clients deep expertise, being easy to work with, and being an amazing place to work! We are looking for curious and creative individuals who want to be part of a fast-paced, dynamic environment, where everyone s input and efforts are valued. We hire outstanding individuals and give them the opportunity to thrive in a collaborative atmosphere that values learning, growth, and hard work. Our team is distributed across North America, Latin America, India and Europe. If you have the desire to be a part of an exciting, challenging, and rapidly-growing Snowflake consulting services company, and if you are passionate about making a difference in this world, we would love to talk to you!. As an AWS Managed Services Architect, you will play a pivotal role in architecting and optimizing the infrastructure and operations of a complex Data Lake environment for BOT clients. You ll leverage your strong expertise with AWS services to design, implement, and maintain scalable and secure data solutions while driving best practices. You will work collaboratively with delivery teams across the U.S., Costa Rica, Portugal, and other regions, ensuring a robust and seamless Data Lake architecture. In addition, you llproactively engage with clients to support their evolving needs, oversee critical AWS infrastructure, and guide teams toward innovative and efficient solutions. This role demands a hands-on approach, including designing solutions, troubleshooting,optimizing performance, and maintaining operational excellence. Role Description AWS Data Lake Architecture: Design, build, and support scalable, high-performance architectures for complex AWS Data Lake solutions. AWS Services Expertise: Deploy and manage cloud-native solutions using a wide range of AWS services, including but not limited to- Amazon EMR (Elastic MapReduce): Optimize and maintain EMR clusters for large-scale big data processing. AWS Batch: Design and implement efficient workflows for batch processing workloads. Amazon SageMaker: Enable data science teams with scalable infrastructure for model training and deployment. AWS Glue: Develop ETL/ELT pipelines using Glue to ensure efficient data ingestion and transformation. AWS Lambda: Build serverless functions to automate processes and handle event-driven workloads. IAM Policies: Define and enforce fine-grained access controls to secure cloud resources and maintain governance. AWS IoT & Timestream: Design scalable solutions for collecting, storing, and analyzing time-series data. Amazon DynamoDB: Build and optimize high-performance NoSQL database solutions. Data Governance & Security: Implement best practices to ensure data privacy, compliance, and governance across the data architecture. Performance Optimization: Monitor, analyze, and tune AWS resources for performance efficiency and cost optimization. Develop and manage Infrastructure as Code (IaC) using AWS CloudFormation, Terraform, or equivalent tools to automate infrastructure deployment. Client Collaboration: Work closely with stakeholders to understand business objectives and ensure solutions align with client needs. Team Leadership & Mentorship: Provide technical guidance to delivery teams through design reviews, troubleshooting, and strategic planning. Continuous Innovation: Stay current with AWS service updates, industry trends, and emerging technologies to enhance solution delivery. Documentation & Knowledge Sharing: Create and maintain architecture diagrams, SOPs, and internal/external documentation to support ongoing operations and collaboration. Qualifications 7+ years of hands-on experience in cloud architecture and infrastructure (preferably AWS). 3+ years of experience specifically in architecting and managing Data Lake or big datadata solutions on AWS. Bachelor s Degree in Computer Science, Information Systems, or a related field (preferred) AWS Certifications such as Solutions Architect Professional or Big Data Specialty. Experience with Snowflake, Matillion, or Fivetran in hybrid cloud environments. Familiarity with Azure or GCP cloud platforms. Understanding of machine learning pipelines and workflows. Technical Skills: Expertise in AWS services such as EMR, Batch, SageMaker, Glue, Lambda,IAM, IoT TimeStream, DynamoDB, and more. Strong programming skills in Python for scripting and automation. Proficiency in SQL and performance tuning for data pipelines and queries. Experience with IaC tools like Terraform or CloudFormation. Knowledge of big data frameworks such as Apache Spark, Hadoop, or similar. Data Governance & Security: Proven ability to design and implement secure solutions, with strong knowledge of IAM policies and compliance standards. Problem-Solving: Analytical and problem-solving mindset to resolve complex technical challenges. Collaboration: Exceptional communication skills to engage with technical and non-technicalstakeholders. Ability to lead cross-functional teams and provide mentorship. Benefits: - Health Insurance - Paid leave - Technical training and certifications - Robust learning and development opportunities - Incentive - Toastmasters - Food Program - Fitness Program - Referral Bonus Program Hakkoda is committed to fostering diversity, equity, and inclusion within our teams. A diverse workforce enhances our ability to serve clients and enriches our culture. We encourage candidates of all races, genders, sexual orientations, abilities, and experiences to apply, creating a workplace where everyone can succeed and thrive. Ready to take your career to the next level? Apply today and join a team that s shaping the future!! Hakkoda is an IBM subsidiary which has been acquired by IBM and will be integrated in the IBM organization. Hakkoda will be the hiring entity. By Proceeding with this application, you understand that Hakkoda will share your personal information with other IBM subsidiaries involved in your recruitment process, wherever these are located. More information on how IBM protects your personal information, including the safeguards in case of cross-border data transfer, are available here.

Sr. Consultant - Architecture

Jaipur

6 - 10 years

INR 12.0 - 16.0 Lacs P.A.

Work from Office

Full Time

ABOUT HAKKODA Hakkoda, an IBM Company, is a modern data consultancy that empowers data driven organizations to realize the full value of the Snowflake Data Cloud. We provide consulting and managed services in data architecture, data engineering, analytics and data science. We are renowned for bringing our clients deep expertise, being easy to work with, and being an amazing place to work! We are looking for curious and creative individuals who want to be part of a fast-paced, dynamic environment, where everyone s input and efforts are valued. We hire outstanding individuals and give them the opportunity to thrive in a collaborative atmosphere that values learning, growth, and hard work. Our team is distributed across North America, Latin America, India and Europe. If you have the desire to be a part of an exciting, challenging, and rapidly-growing Snowflake consulting services company, and if you are passionate about making a difference in this world, we would love to talk to you!. We are looking for people experienced with data architecture, design and development of database mapping and migration processes. This person will have direct experience optimizing new and current databases, data pipelines and implementing advanced capabilities while ensuring data integrity and security. Ideal candidates will have strong communication skills and the ability to guide clients and project team members. Acting as a key point of contact for direction and expertise. Key Responsibilities Design, develop, and optimize database architectures and data pipelines. Ensure data integrity and security across all databases and data pipelines. Lead and guide clients and project team members, acting as a key point of contact for direction and expertise. Collaborate with cross-functional teams to understand business requirements and translate them into technical solutions. Manage and support large-scale technology programs, ensuring they meet business objectives and compliance requirements. Develop and implement migration, dev/ops, and ETL/ELT ingestion pipelines using tools such as DataStage, Informatica, and Matillion. Utilize project management skills to work effectively within Scrum and Agile Development methods. Create and leverage metrics to develop actionable and measurable insights, influencing business decisions. Qualifications 7+ years of proven work experience in data warehousing, business intelligence (BI), and analytics. 3+ years of experience as a Data Architect. 3+ years of experience working on Cloud platforms (AWS, Azure, GCP). Bachelors Degree (BA/BS) in Computer Science, Information Systems, Mathematics, MIS, or a related field. Strong understanding of migration processes, dev/ops, and ETL/ELT ingestion pipelines. Proficient in tools such as DataStage, Informatica, and Matillion. Excellent project management skills and experience with Scrum and Agile Development methods. Ability to develop actionable and measurable insights and create metrics to influence business decisions. Previous consulting experience managing and supporting large-scale technology programs. Nice to Have 6-12 months of experience working with Snowflake. Understanding of Snowflake design patterns and migration architectures. Knowledge of Snowflake roles, user security, and capabilities like Snowpipe. Proficiency in SQL scripting. Cloud experience on AWS (Azure and GCP are also beneficial) Python scripting skills. Benefits: - Health Insurance - Paid leave - Technical training and certifications - Robust learning and development opportunities - Incentive - Toastmasters - Food Program - Fitness Program - Referral Bonus Program Hakkoda is committed to fostering diversity, equity, and inclusion within our teams. A diverse workforce enhances our ability to serve clients and enriches our culture. We encourage candidates of all races, genders, sexual orientations, abilities, and experiences to apply, creating a workplace where everyone can succeed and thrive. Ready to take your career to the next level? Apply today and join a team that s shaping the future!! Hakkoda is an IBM subsidiary which has been acquired by IBM and will be integrated in the IBM organization. Hakkoda will be the hiring entity. By Proceeding with this application, you understand that Hakkoda will share your personal information with other IBM subsidiaries involved in your recruitment process, wherever these are located. More information on how IBM protects your personal information, including the safeguards in case of cross-border data transfer, are available here.

Power BI - Data Analyst

Jaipur

3 - 6 years

INR 5.0 - 9.0 Lacs P.A.

Work from Office

Full Time

ABOUT HAKKODA Hakkoda, an IBM Company, is a modern data consultancy that empowers data driven organizations to realize the full value of the Snowflake Data Cloud. We provide consulting and managed services in data architecture, data engineering, analytics and data science. We are renowned for bringing our clients deep expertise, being easy to work with, and being an amazing place to work! We are looking for curious and creative individuals who want to be part of a fast-paced, dynamic environment, where everyone s input and efforts are valued. We hire outstanding individuals and give them the opportunity to thrive in a collaborative atmosphere that values learning, growth, and hard work. Our team is distributed across North America, Latin America, India and Europe. If you have the desire to be a part of an exciting, challenging, and rapidly-growing Snowflake consulting services company, and if you are passionate about making a difference in this world, we would love to talk to you!. We are looking for a skilled and motivated Data Analyst / Data Engineer to join our growing data team in Jaipur. The ideal candidate should have hands-on experience with SQL, Python, Power BI , and familiarity with Snowflake is a strong advantage. You will play a key role in building data pipelines, delivering analytical insights, and enabling data-driven decision-making across the organization. Role Description: Develop and manage robust data pipelines and workflows for data integration, transformation, and loading. Design, build, and maintain interactive Power BI dashboards and reports based on business needs. Optimize existing Power BI reports for performance, usability, and scalability . Write and optimize complex SQL queries for data analysis and reporting. Use Python for data manipulation, automation, and advanced analytics where applicable. Collaborate with business stakeholders to understand requirements and deliver actionable insights . Ensure high data quality, integrity, and governance across all reporting and analytics layers. Work closely with data engineers, analysts, and business teams to deliver scalable data solutions . Leverage cloud data platforms like Snowflake for data warehousing and analytics (good to have). Qualifications 3-6 years of professional experience in data analysis or data engineering. Bachelor s degree in computer science , Engineering, Data Science, Information Technology , or a related field. Strong proficiency in SQL with the ability to write complex queries and perform data modeling. Hands-on experience with Power BI for data visualization and business intelligence reporting. Programming knowledge in Python for data processing and analysis. Good understanding of ETL/ELT , data warehousing concepts, and cloud-based data ecosystems. Excellent problem-solving skills , attention to detail, and analytical thinking. Strong communication and interpersonal skills to work effectively with cross-functional teams . Preferred / Good to Have Experience working with large datasets and cloud platforms like Snowflake, Redshift, or BigQuery. Familiarity with workflow orchestration tools (e.g., Airflow) and version control systems (e.g., Git). Power BI Certification (e.g., PL-300: Microsoft Power BI Data Analyst). Exposure to Agile methodologies and end-to-end BI project life cycles. Benefits: - Health Insurance - Paid leave - Technical training and certifications - Robust learning and development opportunities - Incentive - Toastmasters - Food Program - Fitness Program - Referral Bonus Program Hakkoda is committed to fostering diversity, equity, and inclusion within our teams. A diverse workforce enhances our ability to serve clients and enriches our culture. We encourage candidates of all races, genders, sexual orientations, abilities, and experiences to apply, creating a workplace where everyone can succeed and thrive. Ready to take your career to the next level? Apply today and join a team that s shaping the future!! Hakkoda is an IBM subsidiary which has been acquired by IBM and will be integrated in the IBM organization. Hakkoda will be the hiring entity. By Proceeding with this application, you understand that Hakkoda will share your personal information with other IBM subsidiaries involved in your recruitment process, wherever these are located. More information on how IBM protects your personal information, including the safeguards in case of cross-border data transfer, are available here.

AWS Admin

Jaipur

10 - 15 years

INR 6.0 - 10.0 Lacs P.A.

Work from Office

Full Time

ABOUT HAKKODA Hakkoda, an IBM Company, is a modern data consultancy that empowers data driven organizations to realize the full value of the Snowflake Data Cloud. We provide consulting and managed services in data architecture, data engineering, analytics and data science. We are renowned for bringing our clients deep expertise, being easy to work with, and being an amazing place to work! We are looking for curious and creative individuals who want to be part of a fast-paced, dynamic environment, where everyone s input and efforts are valued. We hire outstanding individuals and give them the opportunity to thrive in a collaborative atmosphere that values learning, growth, and hard work. Our team is distributed across North America, Latin America, India and Europe. If you have the desire to be a part of an exciting, challenging, and rapidly-growing Snowflake consulting services company, and if you are passionate about making a difference in this world, we would love to talk to you!. We are seeking a highly skilled and experienced AWS Administrator to join a long-term project (12+ months), fully allocated and 100% hands-on. This role will backfill a senior AWS Admin with 10-15 years of experience and requires deep technical capability across AWS infrastructure service. This is not a team leadership role the ideal candidate will operate independently, take full ownership of AWS administration tasks, and contribute directly to maintaining and optimizing cloud operations. Role & Responsibilities AWS Infrastructure Management: Provision, configure, and maintain AWS services such as EC2, S3, IAM, VPC, Lambda, RDS, CloudWatch, CloudTrail, and more. Monitoring & Incident Response: Set up monitoring, logging, and alerting solutions. Respond to and resolve infrastructure issues proactively. Security & IAM: Manage IAM roles, policies, and user access with a strong focus on security best practices and compliance requirements. Automation & Scripting: Automate routine tasks using scripting (Bash, Python) and AWS CLI/SDK. Infrastructure as Code (IaC): Use tools like Terraform or CloudFormation to manage and automate infrastructure deployments and changes. Cost Optimization: Monitor resource usage and implement cost-control strategies to optimize AWS spending. Backup & Disaster Recovery: Manage backup strategies and ensure systems are resilient and recoverable. Documentation: Maintain detailed and up-to-date documentation of AWS environments, standard operating procedures, and runbooks. Skils & Qualifications 10+ years of hands-on AWS administration experience. Strong understanding of AWS core services (EC2, S3, IAM, VPC, Lambda, RDS, etc.). Experience with scripting (Python, Bash, or PowerShell) and automation tooling. Proven expertise in using Terraform or CloudFormation . Deep knowledge of IAM policy creation and security best practices. Experience with monitoring tools such as CloudWatch, Prometheus, or third-party APM tools. Familiarity with CI/CD pipelines and DevOps principles. Strong troubleshooting skills with the ability to resolve complex infrastructure issues independently. Excellent communication skills with the ability to work effectively with remote teams. Comfortable working during US Eastern Time zone hours. Preferred Qualifications: AWS Certifications (e.g., SysOps Administrator Associate , Solutions Architect Associate/Professional ). Experience in hybrid environments or with other cloud platforms (Azure, GCP). Familiarity with Snowflake, GitLab, or similar DevOps tooling. Benefits: - Health Insurance - Paid leave - Technical training and certifications - Robust learning and development opportunities - Incentive - Toastmasters - Food Program - Fitness Program - Referral Bonus Program Hakkoda is committed to fostering diversity, equity, and inclusion within our teams. A diverse workforce enhances our ability to serve clients and enriches our culture. We encourage candidates of all races, genders, sexual orientations, abilities, and experiences to apply, creating a workplace where everyone can succeed and thrive. Ready to take your career to the next level? Apply today and join a team that s shaping the future!! Hakkoda is an IBM subsidiary which has been acquired by IBM and will be integrated in the IBM organization. Hakkoda will be the hiring entity. By Proceeding with this application, you understand that Hakkoda will share your personal information with other IBM subsidiaries involved in your recruitment process, wherever these are located. More information on how IBM protects your personal information, including the safeguards in case of cross-border data transfer, are available here.

Consultant II - Data Engineer, Python

Jaipur

5 - 10 years

INR 6.0 - 10.0 Lacs P.A.

Work from Office

Full Time

ABOUT HAKKODA Hakkoda, an IBM Company, is a modern data consultancy that empowers data driven organizations to realize the full value of the Snowflake Data Cloud. We provide consulting and managed services in data architecture, data engineering, analytics and data science. We are renowned for bringing our clients deep expertise, being easy to work with, and being an amazing place to work! We are looking for curious and creative individuals who want to be part of a fast-paced, dynamic environment, where everyone s input and efforts are valued. We hire outstanding individuals and give them the opportunity to thrive in a collaborative atmosphere that values learning, growth, and hard work. Our team is distributed across North America, Latin America, India and Europe. If you have the desire to be a part of an exciting, challenging, and rapidly-growing Snowflake consulting services company, and if you are passionate about making a difference in this world, we would love to talk to you!. We are seeking a skilled and collaborative Sr. Data/Python Engineer with experience in the development of production Python-based applications (Such as Django, Flask, FastAPI on AWS) to support our data platform initiatives and application development. This role will initially focus on building and optimizing Streamlit application development frameworks, CI/CD Pipelines, ensuring code reliability through automated testing with Pytest , and enabling team members to deliver updates via CI/CD pipelines . Once the deployment framework is implemented, the Sr Engineer will own and drive data transformation pipelines in dbt and implement a data quality framework. Key Responsibilities: Lead application testing and productionalization of applications built on top of Snowflake - This includes implementation and execution of unit testing and integration testing - Automated test suites include use of Pytest and Streamlit App Tests to ensure code quality, data accuracy, and system reliability. Development and Integration of CI/CD pipelines (e.g., GitHub Actions, Azure DevOps, or GitLab CI) for consistent deployments across dev, staging, and production environments. Development and testing of AWS-based pipelines - AWS Glue, Airflow (MWAA), S3. Design, develop, and optimize data models and transformation pipelines in Snowflake using SQL and Python. Build Streamlit-based applications to enable internal stakeholders to explore and interact with data and models. Collaborate with team members and application developers to align requirements and ensure secure, scalable solutions. Monitor data pipelines and application performance, optimizing for speed, cost, and user experience. Create end-user technical documentation and contribute to knowledge sharing across engineering and analytics teams. Work in CST hours and collaborate with onshore and offshore teams. Qualifications, Skills Experience 5+ years of experience in Data Engineering or Python based application development on AWS (Flask, Django, FastAPI, Streamlit) - Experience building data data-intensive applications on python as well as data pipelines on AWS in a must. Bachelor s degree in computer science, Information Systems, Data Engineering, or a related field (or equivalent experience). Proficient in SQL and Python for data manipulation and automation tasks. Experience with developing and productionalizing applications built on Python based Frameworks such as FastAPI, Django, Flask. Experience with application frameworks such as Streamlit, Angular, React etc for rapid data app deployment. Solid understanding of software testing principles and experience using Pytest or similar Python frameworks. Experience configuring and maintaining CI/CD pipelines for automated testing and deployment. Familiarity with version control systems such as Gitlab . Knowledge of data governance, security best practices, and role-based access control (RBAC) in Snowflake. Preferred Qualifications: Experience with dbt (data build tool) for transformation modeling. Knowledge of Snowflake s advanced features (e.g., masking policies, external functions, Snowpark). Exposure to cloud platforms (e.g., AWS, Azure, GCP). Strong communication and documentation skills. Benefits: - Health Insurance - Paid leave - Technical training and certifications - Robust learning and development opportunities - Incentive - Toastmasters - Food Program - Fitness Program - Referral Bonus Program Hakkoda is committed to fostering diversity, equity, and inclusion within our teams. A diverse workforce enhances our ability to serve clients and enriches our culture. We encourage candidates of all races, genders, sexual orientations, abilities, and experiences to apply, creating a workplace where everyone can succeed and thrive. Ready to take your career to the next levelApply today and join a team that s shaping the future!! Hakkoda is an IBM subsidiary which has been acquired by IBM and will be integrated in the IBM organization. Hakkoda will be the hiring entity. By Proceeding with this application, you understand that Hakkoda will share your personal information with other IBM subsidiaries involved in your recruitment process, wherever these are located. More information on how IBM protects your personal information, including the safeguards in case of cross-border data transfer, are available here.

Hakkda logo

Hakkda

|

Technology / AI Solutions

San Francisco

50-200 Employees

7 Jobs

cta

Start Your Job Search Today

Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.

Job Application AI Bot

Job Application AI Bot

Apply to 20+ Portals in one click

Download Now

Download the Mobile App

Instantly access job listings, apply easily, and track applications.

Job Titles Overview