Jobs
Interviews

170 Impala Jobs - Page 3

Setup a job Alert
JobPe aggregates results for easy application access, but you actually apply on the job portal directly.

5.0 - 10.0 years

10 - 15 Lacs

Bengaluru

Work from Office

Overall Responsibilities: Data Pipeline Development: Design, develop, and maintain highly scalable and optimized ETL pipelines using PySpark on the Cloudera Data Platform, ensuring data integrity and accuracy. Data Ingestion: Implement and manage data ingestion processes from a variety of sources (e.g., relational databases, APIs, file systems) to the data lake or data warehouse on CDP. Data Transformation and Processing: Use PySpark to process, cleanse, and transform large datasets into meaningful formats that support analytical needs and business requirements. Performance Optimization: Conduct performance tuning of PySpark code and Cloudera components, optimizing resource utilization and reducing runtime of ETL processes. Data Quality and Validation: Implement data quality checks, monitoring, and validation routines to ensure data accuracy and reliability throughout the pipeline. Automation and Orchestration: Automate data workflows using tools like Apache Oozie, Airflow, or similar orchestration tools within the Cloudera ecosystem. Monitoring and Maintenance: Monitor pipeline performance, troubleshoot issues, and perform routine maintenance on the Cloudera Data Platform and associated data processes. Collaboration: Work closely with other data engineers, analysts, product managers, and other stakeholders to understand data requirements and support various data-driven initiatives. Documentation: Maintain thorough documentation of data engineering processes, code, and pipeline configurations. Software Requirements: Advanced proficiency in PySpark, including working with RDDs, DataFrames, and optimization techniques. Strong experience with Cloudera Data Platform (CDP) components, including Cloudera Manager, Hive, Impala, HDFS, and HBase. Knowledge of data warehousing concepts, ETL best practices, and experience with SQL-based tools (e.g., Hive, Impala). Familiarity with Hadoop, Kafka, and other distributed computing tools. Experience with Apache Oozie, Airflow, or similar orchestration frameworks. Strong scripting skills in Linux. Category-wise Technical Skills: PySpark: Advanced proficiency in PySpark, including working with RDDs, DataFrames, and optimization techniques. Cloudera Data Platform: Strong experience with Cloudera Data Platform (CDP) components, including Cloudera Manager, Hive, Impala, HDFS, and HBase. Data Warehousing: Knowledge of data warehousing concepts, ETL best practices, and experience with SQL-based tools (e.g., Hive, Impala). Big Data Technologies: Familiarity with Hadoop, Kafka, and other distributed computing tools. Orchestration and Scheduling: Experience with Apache Oozie, Airflow, or similar orchestration frameworks. Scripting and Automation: Strong scripting skills in Linux. Experience: 5-12 years of experience as a Data Engineer, with a strong focus on PySpark and the Cloudera Data Platform. Proven track record of implementing data engineering best practices. Experience in data ingestion, transformation, and optimization on the Cloudera Data Platform. Day-to-Day Activities: Design, develop, and maintain ETL pipelines using PySpark on CDP. Implement and manage data ingestion processes from various sources. Process, cleanse, and transform large datasets using PySpark. Conduct performance tuning and optimization of ETL processes. Implement data quality checks and validation routines. Automate data workflows using orchestration tools. Monitor pipeline performance and troubleshoot issues. Collaborate with team members to understand data requirements. Maintain documentation of data engineering processes and configurations. Qualifications: Bachelors or Masters degree in Computer Science, Data Engineering, Information Systems, or a related field. Relevant certifications in PySpark and Cloudera technologies are a plus. Soft Skills: Strong analytical and problem-solving skills. Excellent verbal and written communication abilities. Ability to work independently and collaboratively in a team environment. Attention to detail and commitment to data quality. S YNECHRONS DIVERSITY & INCLUSION STATEMENT Diversity & Inclusion are fundamental to our culture, and Synechron is proud to be an equal opportunity workplace and is an affirmative action employer. Our Diversity, Equity, and Inclusion (DEI) initiative Same Difference is committed to fostering an inclusive culture promoting equality, diversity and an environment that is respectful to all. We strongly believe that a diverse workforce helps build stronger, successful businesses as a global company. We encourage applicants from across diverse backgrounds, race, ethnicities, religion, age, marital status, gender, sexual orientations, or disabilities to apply. We empower our global workforce by offering flexible workplace arrangements, mentoring, internal mobility, learning and development programs, and more. All employment decisions at Synechron are based on business needs, job requirements and individual qualifications, without regard to the applicants gender, gender identity, sexual orientation, race, ethnicity, disabled or veteran status, or any other characteristic protected by law . Candidate Application Notice

Posted 2 weeks ago

Apply

5.0 - 9.0 years

0 Lacs

maharashtra

On-site

The role of a Data Engineer is crucial for ensuring the smooth operation of the Data Platform in Azure / AWS Databricks. As a Data Engineer, you will be responsible for the continuous development, enhancement, support, and maintenance of data availability, data quality, performance enhancement, and stability of the system. Your primary responsibilities will include designing and implementing data ingestion pipelines from various sources using Azure Databricks, ensuring the efficient and smooth running of data pipelines, and adhering to security, regulatory, and audit control guidelines. You will also be tasked with driving optimization, continuous improvement, and efficiency in data processes. To excel in this role, it is essential to have a minimum of 5 years of experience in the data analytics field, hands-on experience with Azure/AWS Databricks, proficiency in building and optimizing data pipelines, architectures, and data sets, and excellent skills in Scala or Python, PySpark, and SQL. Additionally, you should be capable of troubleshooting and optimizing complex queries on the Spark platform, possess knowledge of structured and unstructured data design/modelling, data access, and data storage techniques, and expertise in designing and deploying data applications on cloud solutions such as Azure or AWS. Moreover, practical experience in performance tuning and optimizing code running in Databricks environment, demonstrated analytical and problem-solving skills, particularly in a big data environment, are essential for success in this role. In terms of technical/professional skills, proficiency in Azure/AWS Databricks, Python/Scala/Spark/PySpark, HIVE/HBase/Impala/Parquet, Sqoop, Kafka, Flume, SQL, RDBMS, Airflow, Jenkins/Bamboo, Github/Bitbucket, and Nexus will be advantageous for executing the responsibilities effectively.,

Posted 2 weeks ago

Apply

2.0 - 6.0 years

0 Lacs

haryana

On-site

As a Consultant in Performance Analytics at Mastercard, you will be part of the Advisors & Consulting Services group, which specializes in Strategy & Transformation, Performance Analytics, Business Experimentation, Marketing, and Program Management. Your role will involve translating data into insights by utilizing both Mastercard and customer data to create, implement, and scale analytical solutions for clients. You will employ qualitative and quantitative analytical techniques along with enterprise applications to synthesize analyses into clear recommendations and impactful narratives. Your responsibilities will include providing creative input on projects across various industries, contributing to the development of analytics strategies for regional and global clients, collaborating with the Mastercard team to understand client needs, and developing relationships with client analysts and managers. Additionally, you will collaborate with senior project delivery consultants, identify key findings, prepare presentations, deliver recommendations to clients, and lead internal and client meetings. To qualify for this role, you should have an undergraduate degree with experience in data and analytics, proficiency in data analytics software such as Python, R, SQL, SAS, and advanced skills in Word, Excel, and PowerPoint. You must be able to analyze large datasets, synthesize key findings, manage clients or internal stakeholders, and communicate effectively in English and the local office language. Preferred qualifications include additional experience in database structures, data visualization tools, working with the Hadoop framework, and relevant industry expertise. As part of your role, you will be expected to abide by Mastercard's security policies, ensure the confidentiality and integrity of accessed information, report any suspected security violations, and complete mandatory security trainings. This position offers opportunities for professional growth and development through mentorship from performance analytics leaders. If you are passionate about leveraging data to drive business insights and solutions, and possess the required qualifications and skills, we encourage you to explore the available positions in Performance Analytics at Mastercard and apply to join our dynamic team.,

Posted 2 weeks ago

Apply

8.0 - 13.0 years

5 - 9 Lacs

Mumbai

Work from Office

Project Role : Application Developer Project Role Description : Design

Posted 2 weeks ago

Apply

3.0 - 7.0 years

0 Lacs

pune, maharashtra

On-site

The Applications Development Programmer Analyst position is an intermediate level role where you will be responsible for contributing to the establishment and implementation of new or updated application systems and programs in collaboration with the Technology team. Your main objective will be to participate in applications systems analysis and programming activities. Your responsibilities will include utilizing your knowledge of applications development procedures and concepts, as well as basic understanding of other technical areas, to identify and define necessary system enhancements. You will be expected to identify and analyze issues, provide recommendations, and implement solutions. Additionally, you will apply your knowledge of business processes, system processes, and industry standards to solve complex problems. Your role will involve analyzing information, making evaluative judgments, recommending solutions and improvements, conducting testing and debugging, utilizing script tools, and writing basic code based on design specifications. You will also need to assess the applicability of similar experiences and evaluate options under circumstances not covered by procedures. Developing a working knowledge of various technical aspects such as Citigroup's information systems, client-server application development, network operations, database administration, systems administration, data center operations, and PC-based applications will be essential. It is crucial that you appropriately assess risk when making business decisions, with a focus on safeguarding Citigroup, its clients, and assets by ensuring compliance with laws, rules, and regulations, adhering to policies, applying ethical judgment, and escalating control issues when necessary. Qualifications for this role include having 2-5 years of relevant experience, proficiency in programming/debugging for business applications, familiarity with industry practices and standards, comprehensive knowledge of a specific business area for application development, working knowledge of program languages, and demonstrating clear and concise written and verbal communication consistently. Education requirement for this position is a Bachelor's degree or equivalent experience. This job description offers a detailed overview of the job responsibilities and qualifications required. Other duties related to the role may be assigned as necessary. Skillsets required for this role include a minimum of 3+ years of hands-on experience in Data engineering stream, good knowledge of technologies such as Hadoop, Spark, Hive, Impala, Performance Tuning, Java programming language, SQL, Oracle, and any certification like Java/Big Data would be beneficial. Citi is an equal opportunity and affirmative action employer, and invites all qualified interested applicants to apply for career opportunities. If you require reasonable accommodation due to a disability to use search tools or apply for a career opportunity, review Accessibility at Citi.,

Posted 3 weeks ago

Apply

5.0 - 9.0 years

0 Lacs

indore, madhya pradesh

On-site

As a Senior Data Scientist with 5+ years of experience, you will be responsible for designing and implementing models, mining data for insights, and interpreting complex data structures to drive business decision-making. Your expertise in machine learning, including areas such as NLP, Machine vision, and Time series, will be essential in this role. You will be expected to have strong skills in Model Tuning, Model Validation, Supervised and Unsupervised Learning, and hands-on experience with model development, data preparation, training, and inference-ready deployment of models. Your proficiency in descriptive and inferential statistics, hypothesis testing, and data analysis will help in developing code for reproducible analysis of data. Experience with AWS services like Sagemaker, Lambda, Glue, Step functions, and EC2 is necessary, along with knowledge of Databricks, Anaconda distribution, and similar data science code development and deployment IDEs. Your familiarity with ML algorithms related to time-series, natural language processing, optimization, object detection, topic modeling, clustering, and regression analysis will be highly valued. You should have expertise in Hive/Impala, Spark, Python, Pandas, Keras, SKLearn, StatsModels, Tensorflow, and PyTorch. End-to-end model deployment and production experience of at least 1 year is required, along with a good understanding of Model Deployment in Azure ML platform, Anaconda Enterprise, or AWS Sagemaker. Basic knowledge of deep learning algorithms such as MaskedCNN, YOLO, and familiarity with Visualization and analytics/Reporting Tools like Power BI, Tableau, and Alteryx will be considered advantageous for this role.,

Posted 3 weeks ago

Apply

8.0 - 12.0 years

0 Lacs

karnataka

On-site

You are a strategic thinker passionate about driving solutions in Data Analytics. You have found the right team. As an Analytics Solutions Vice President in our Finance team, you will define, refine, and deliver our firm's goals. If you're a skilled data professional passionate about transforming raw data into actionable insights and eager to learn and implement new technologies, you've found the right team. Join us in the Finance Data & Insights Team, an agile product team focused on developing, producing, and transforming financial data and reporting across CCB. Your role will involve creating data visualizations and intelligence solutions for top leaders to achieve strategic goals. You'll identify opportunities to eliminate manual processes and use automation tools like Alteryx, Tableau, and ThoughtSpot to develop automated solutions. Additionally, you'll extract, analyze, and summarize data for ad hoc requests and contribute to modernizing our data environment to a cloud platform. Job responsibilities: - Lead Data & Analytics requirements gathering sessions with varying levels of leadership and complete detailed project planning using JIRA to record planned project execution steps. - Understand databases, ETL processes, and translate logic into requirements for the Technology team. - Develop and enhance Alteryx workflows by collecting data from disparate sources and summarizing it as defined in requirements gathering with stakeholders, following best practices to source data from authoritative sources. - Develop data visualization solutions using Tableau and/or ThoughtSpot to provide intuitive insights to key stakeholders. - Conduct thorough control testing of each component of the intelligence solution, providing evidence that all data and visualizations offer accurate insights and evidence in the control process. - Seek to understand stakeholder use cases to anticipate their requirements, questions, and objections. - Become a subject matter expert in these responsibilities and support team members in becoming more proficient. Required qualifications, capabilities, and skills: - Bachelor's degree in MIS or Computer Science, Mathematics, Engineering, Statistics, or other quantitative or financial subject areas - People management experience of at least 3 years is required - Experience with business intelligence analytic and data wrangling tools such as Alteryx, SAS, or Python - Experience with relational databases optimizing SQL to pull and summarize large datasets, report creation and ad-hoc analyses, Databricks, Cloud solutions - Experience in reporting development and testing, and ability to interpret unstructured data and draw objective inferences given known limitations of the data - Demonstrated ability to think beyond raw data and to understand the underlying business context and sense business opportunities hidden in data - Strong written and oral communication skills; ability to communicate effectively with all levels of management and partners from a variety of business functions - Experience with ThoughtSpot or similar tools empowering stakeholders to better understand their data - Highly motivated, self-directed, curious to learn new technologies Preferred qualifications, capabilities, and skills: - Experience with ThoughtSpot / Python major advantage - Experience with AI/ML or LLM added advantage but not a must-have. Minimum 8 years experience developing advanced data visualization and presentations preferably with Tableau - Experience with Hive, Spark SQL, Impala, or other big-data query tools. AWS, Databricks, Snowflake, or other Cloud Data Warehouse experience - Minimum of 8 years experience working with data analytics projects, preferably related to financial services domain,

Posted 3 weeks ago

Apply

3.0 - 7.0 years

0 Lacs

chennai, tamil nadu

On-site

As an experienced professional with 3-5 years of experience, you will be responsible for working with a range of technical skills including Azure Data Factory, Talend/SSIS, MSSQL, Azure, and MySQL. Your primary focus will be on Azure Data Factory, where you will utilize your expertise to handle complex data analysis tasks effectively. In this role, you will demonstrate advanced knowledge in Azure SQL DB & Synapse Analytics, Power BI, SSIS, SSRS, T-SQL, and Logic Apps. It is essential that you possess a solid understanding of Azure Data Lake and Azure Services such as Analysis Service, SQL Databases, Azure DevOps, and CI/CD processes. Furthermore, your responsibilities will include mastering data management, data warehousing, and business intelligence architecture. You will be required to apply your experience in data modeling and database design, ensuring compliance with SQL Server best practices. Effective communication is key in this role, as you will engage with stakeholders at various levels. You will contribute to the preparation of design documents, unit test plans, and code review reports. Experience in an Agile environment, specifically with Scrum, Lean, or Kanban methodologies, will be advantageous. Additionally, familiarity with Big Data technologies such as the Spark Framework, NoSQL databases, Azure Data Bricks, and the Hadoop Ecosystem (Hive, Impala, HDFS) will be beneficial for this position.,

Posted 3 weeks ago

Apply

6.0 - 11.0 years

15 - 19 Lacs

Hyderabad, Pune, Bengaluru

Work from Office

Project description During the 2008 financial crisis, many big banks failed or faced issues due to liquidity issues. Lack of liquidity can kill any financial institution over the night. That's why it's so critical to constantly monitor liquidity risks and properly maintain collaterals. We are looking for a number of talented developers, who would like to join our team in Pune, which is building liquidity risk and collateral management platform for one of the biggest investment banks over the globe. The platform is a set of front-end tools and back-end engines. Our platform helps the bank to increase efficiency and scalability, reduce operational risk and eliminate the majority of manual interventions in processing margin calls. Responsibilities The candidate will work on development of new functionality for Liqudity Risk platform closely with other teams over the globe. Skills Must have BigData experience (6 years+); Java/python J2EE, Spark, Hive; SQL Databases; UNIX Shell; Strong Experience in Apache Hadoop, Spark, Hive, Impala, Yarn, Talend, Hue; Big Data Reporting, Querying and analysis. Nice to have Spark Calculators based on business logic/rules Basic performance tuning and troubleshooting knowledge Experience with all aspects of the SDLC Experience with complex deployment infrastructures Knowledge in software architecture, design and testing Data flow automation (Apache NiFi, Airflow etc) Understanding of difference between OOP and Functional design approach Understanding of an event driven architecture Spring, Maven, GIT, uDeploy;

Posted 3 weeks ago

Apply

7.0 - 12.0 years

30 - 37 Lacs

Pune

Work from Office

: Job Title Production SR Engineer AXIOM, AVP LocationPune, India Role Description The Sr Engineer designs and implements technical solutions and configures applications in different environments in response to business problems. With the partial/ full ownership of Production platforms, Sr Engineer is required to ensure environment stability, expeditious and timely resolution of Production issues, ensuring minimal downtimes and continuity of services. Further, the SR Engineer investigates, proposes, and implements various solutions, standardizing where possible, to ensures stability and reliability of the application platforms. What well offer you , 100% reimbursement under childcare assistance benefit (gender neutral) Sponsorship for Industry relevant certifications and education Accident and Term life Insurance Your key responsibilities Production Support Supports the production platform, carrying out monitoring and remediation of issues and specific maintenance tasks as per the run book procedures. Owns the logs that track manual tasks, remediation and system improvement Second line of defence for Incident Management gathering data for investigation and identification of root cause through to resolution Defines and creates Run Book entries and knowledge articles based on incidents experienced in the production environment Application Analysis Carries out deep technical analysis of the production to identify performance, stability and resilience enhancements. Assists in design and creation of NFRs through identification of further requirements. Incident Management Prioritizes and diagnoses incidents according to agreed procedures. Investigates causes of incidents and seeks resolution. Facilitates recovery, following resolution of incidents. Documents and closes resolved incidents according to agreed procedures. Change Management Assesses, analyses, develops, documents and implements changes based on requests for change. These could be technical and/or functional solutions as provided by Development and/or other Production Engineers ensuring segregation of duties Configure the application in response to business and/or technical problems and in accordance with provided requirements and agreed design principles Problem management Investigates problems in systems, processes and services. Ensures PIR reporting and Root Cause Analysis Assists with the implementation of agreed remedies and preventative measures. Attends Program Meetings to prioritize owned stability and performance enhancements as identified in production Release and deployment : Uses the tools and techniques for specific areas of release and deployment activities. Administers the recording of activities, logging of results and documents technical activity undertaken. May carry out early life support activities such as providing support advice to initial users. Production Readiness and Testing Serves as the DevOps Lead, working closely with the Change teams in driving and embedding the Production readiness controls as part of the Software delivery lifecycle. Builds keep list of NFRs and ensures adherence through design forums Reviews requirements and specifications and defines test conditions Analyses and reports on test activities and reports risk and issues associated with own work. Your skills and experience Sound understanding of strong change management processes and controls in large organizations. Strong experience of envisioning and driving full stack automation in medium to large sized groups. Strong AXIOM Tool Knowledge and Experience and Debugging skills Excellent problem solving skills in a distributed, multi-technology ecosystem Strong technical expertise in all of the following technologiesORACLE databases, SQL, Unix Good hands on experience with databases Oracle18c, Oracle19c preferred specially working on SQL/PL SQL is essential. Strong Experience working on Cloud Technology i.e. Google Cloud (Preferred) Strong scripting experience in Java, Python and Shell Solid understanding of messaging middleware like Solace, TIBCO or MQ using JMS Solid understanding of monitoring systems like ITRS Geneos Strong Knowledge of Oracle Management, SQL scripts, performance mgmt. Knowledge of SQL and relational databases ideally both Hive/Impala/SparkSQL and a traditional RDMS, such as Oracle Strong knowledge of configuration management tools like Ansible (preferred), Chef and Puppet Strong Knowledge of Oracle Management, SQL scripts, performance mgmt. Strong knowledge of configuration management tools like Ansible (preferred), Chef and Puppet Strong understanding of Unix, Linux and Windows Understanding of Agile and Safe methodologies (preferred) How well support you . . . . About us and our teams Please visit our company website for further information: https://www.db.com/company/company.htm We strive for a culture in which we are empowered to excel together every day. This includes acting responsibly, thinking commercially, taking initiative and working collaboratively. Together we share and celebrate the successes of our people. Together we are Deutsche Bank Group. We welcome applications from all people and promote a positive, fair and inclusive work environment.

Posted 3 weeks ago

Apply

8.0 - 12.0 years

0 Lacs

chennai, tamil nadu

On-site

You will be responsible for managing one or more applications to achieve established goals and handle personnel duties for your team, including hiring and training. Your role involves designing and developing real-time and batch data transformation processes using a variety of technologies such as Hadoop, Spark Stream, Spark SQL, Python, and Hive. You will also design and develop programs to enhance functionalities in the next-generation Big Data platform and ensure data redistribution is authorized. As a Big Data Developer with 8-10 years of relevant experience, you must possess strong skills in Java/J2EE, Hadoop, Scala, Hive, Impala, Kafka, and Elastic to address data concerns and implement data remediation requirements. Your role will require you to have a good understanding of design patterns and the ability to provide solutions to complex design issues, as well as identify and resolve code issues. You will be hands-on in managing application development using Spark (Scala, Python, or Java), SQL, and the Linux-based Hadoop ecosystem (HDFS, Impala, Hive, HBase, etc.). Your experience as a senior-level professional in an Applications Development role and your proven Solution Delivery skills will be essential in this position. Additionally, you should have a basic knowledge of finance industry practices and standards. Excellent analytical and process-based skills are required, including expertise in process flow diagrams, business modeling, and functional design. Being dynamic, flexible, and maintaining a high energy level is crucial as you will be working in a demanding and rapidly changing environment. Your educational background should include a Bachelor's degree/University degree or equivalent experience.,

Posted 3 weeks ago

Apply

3.0 - 7.0 years

0 Lacs

karnataka

On-site

As a Big Data Developer, you will be responsible for leveraging your strong experience in big data technologies and associated tools such as Hadoop, Unix, HDFS, Hive, and Impala. Your proficiency in using Spark/Scala and experience with data Import/Export using Sqoop or similar tools will be crucial in this role. Additionally, you will be expected to have experience with tools like Airflow, Jenkins, or similar automation tools. Your excellent knowledge of SQL Server and database structures will be essential for writing and optimizing T-SQL queries and stored procedures. Experience working with Jira, Confluence, and GitLab will also be beneficial in this role. Your organizational skills and ability to handle multiple activities with changing priorities simultaneously will be highly valued. As part of the delivery team, your primary responsibilities will include ensuring effective design, development, validation, and support activities to meet client satisfaction in the technology domain. You will gather requirements, understand client needs, and translate them into system requirements. Additionally, you will play a key role in estimating work requirements and providing project estimations to Technology Leads and Project Managers. You will be a key contributor to building efficient programs/systems and collaborating with other Big Data developers to ensure consistency in data solutions. Your ability to partner with the business community, perform technology research, and evaluate new technologies will be crucial in enhancing the overall capability of the analytics technology stack. Key Responsibilities: - Code, test, and document new or modified data systems to create robust and scalable applications for data analytics. - Work with other Big Data developers to ensure consistency in data solutions. - Partner with the business community to understand requirements, determine training needs, and deliver user training sessions. - Perform technology and product research to define requirements, resolve issues, and enhance the analytics technology stack. - Evaluate and provide feedback on future technologies and new releases/upgrades. Job Specific Knowledge: - Support Big Data and batch/real-time analytical solutions using transformational technologies. - Work on multiple projects as a technical team member or drive user requirement analysis, design, development, testing, and automation tools. Professional Attributes: - Good communication skills. - Team player willing to collaborate throughout all phases of development, testing, and deployment. - Ability to solve problems and meet deadlines with minimal supervision. If you believe you have the skills and experience to contribute effectively to our clients" digital transformation journey, we welcome you to join our team.,

Posted 3 weeks ago

Apply

2.0 - 6.0 years

0 Lacs

karnataka

On-site

As a PySpark Data Engineer, you must have a minimum of 2 years of experience in PySpark. Strong programming skills in Python, PySpark, and Scala are preferred. It is essential to have experience in designing and implementing CI/CD, Build Management, and Development strategies. Additionally, familiarity with SQL and SQL Analytical functions is required, along with participation in key business, architectural, and technical decisions. There is an opportunity for training in AWS cloud technology. In the role of a Python Developer, a minimum of 2 years of experience in Python/PySpark is necessary. Strong programming skills in Python, PySpark, and Scala are preferred. Experience in designing and implementing CI/CD, Build Management, and Development strategies is essential. Familiarity with SQL and SQL Analytical functions and participation in key business, architectural, and technical decisions are also required. There is a potential for training in AWS cloud technology. As a Senior Software Engineer at Capgemini, you should have over 3 years of experience in Scala with a strong project track record. Hands-on experience in Scala/Spark development and SQL writing skills on RDBMS (DB2) databases are crucial. Experience in working with different file formats like JSON, Parquet, AVRO, ORC, and XML is preferred. Previous involvement in a HDFS platform development project is necessary. Proficiency in data analysis, data profiling, and data lineage, along with strong oral and written communication skills, is required. Experience in Agile projects is a plus. For the position of Data Modeler, expertise in data structures, algorithms, calculus, linear algebra, machine learning, and modeling is essential. Knowledge of data warehousing concepts such as Star schema, snowflake, or data vault for data mart or data warehousing is required. Proficiency in using data modeling software like Erwin, ER studio, MySQL Workbench to produce logical and physical data models is necessary. Hands-on knowledge and experience with tools like PL/SQL, PySpark, Hive, Impala, and other scripting tools are preferred. Experience with Software Development Lifecycle using the Agile methodology is essential. Strong communication and stakeholder management skills are crucial for this role. In this role, you will design, develop, and optimize PL/SQL procedures, functions, triggers, and packages. You will also write efficient SQL queries, joins, and subqueries for data retrieval and manipulation. Additionally, you will develop and maintain database objects such as tables, views, indexes, and sequences. Optimizing query performance and troubleshooting database issues to improve efficiency are key responsibilities. Collaboration with application developers, business analysts, and system architects to understand database requirements is essential. Ensuring data integrity, consistency, and security within Oracle databases is also a crucial aspect of the role. Developing ETL processes and scripts for data migration and integration are part of the responsibilities. Documenting database structures, stored procedures, and coding best practices is required. Staying up-to-date with Oracle database technologies, best practices, and industry trends is essential for success in this role.,

Posted 3 weeks ago

Apply

8.0 - 13.0 years

10 - 14 Lacs

Bengaluru

Work from Office

Educational Bachelor of Engineering Service Line Strategic Technology Group Responsibilities Power Programmer is an important initiative within Global Delivery to develop a team of Full Stack Developers who will be working on complex engineering projects, platforms and marketplaces for our clients using emerging technologies., They will be ahead of the technology curve and will be constantly enabled and trained to be Polyglots., They are Go-Getters with a drive to solve end customer challenges and will spend most of their time in designing and coding, End to End contribution to technology oriented development projects., Providing solutions with minimum system requirements and in Agile Mode., Collaborate with Power Programmers., Open Source community and Tech User group., Custom Development of new Platforms & Solutions ,Opportunities., Work on Large Scale Digital Platforms and marketplaces., Work on Complex Engineering Projects using cloud native architecture ., Work with innovative Fortune 500 companies in cutting edge technologies., Co create and develop New Products and Platforms for our clients., Contribute to Open Source and continuously upskill in latest technology areas., Incubating tech user group Technical and Professional : Bigdata Spark, scala, hive, kafka Preferred Skills: Technology-Big Data-Hbase Technology-Big Data-Sqoop Technology-Java-Apache-Scala Technology-Functional Programming-Scala Technology-Big Data - Data Processing-Map Reduce Technology-Big Data - Data Processing-Spark

Posted 3 weeks ago

Apply

3.0 - 5.0 years

5 - 7 Lacs

Mumbai, New Delhi, Bengaluru

Work from Office

We are seeking a skilled Big Data Developer with 3+ years of experience to develop, maintain, and optimize large-scale data pipelines using frameworks like Spark, PySpark, and Airflow. The role involves working with SQL, Impala, Hive, and PL/SQL for advanced data transformations and analytics, designing scalable data storage systems, and integrating structured and unstructured data using tools like Sqoop. The ideal candidate will collaborate with cross-functional teams to implement data warehousing strategies and leverage BI tools for insights. Proficiency in Python programming, workflow orchestration with Airflow, and Unix/Linux environments is essential. Locations : Mumbai, Delhi / NCR, Bengaluru , Kolkata, Chennai, Hyderabad, Ahmedabad, Pune, Remote

Posted 3 weeks ago

Apply

4.0 - 9.0 years

10 - 12 Lacs

Bengaluru, Doddakannell, Karnataka

Work from Office

We are seeking a highly skilled Data Engineer with expertise in ETL techniques, programming, and big data technologies. The candidate will play a critical role in designing, developing, and maintaining robust data pipelines, ensuring data accuracy, consistency, and accessibility. This role involves collaboration with cross-functional teams to enrich and maintain a central data repository for advanced analytics and machine learning. The ideal candidate should have experience with cloud-based data platforms, data modeling, and data governance processes. Location - Bengaluru,Doddakannell, Karnataka, Sarjapur Road

Posted 3 weeks ago

Apply

3.0 - 5.0 years

5 - 9 Lacs

New Delhi, Ahmedabad, Bengaluru

Work from Office

We are seeking a skilled Big Data Developer with 3+ years of experience to develop, maintain, and optimize large-scale data pipelines using frameworks like Spark, PySpark, and Airflow. The role involves working with SQL, Impala, Hive, and PL/SQL for advanced data transformations and analytics, designing scalable data storage systems, and integrating structured and unstructured data using tools like Sqoop. The ideal candidate will collaborate with cross-functional teams to implement data warehousing strategies and leverage BI tools for insights. Proficiency in Python programming, workflow orchestration with Airflow, and Unix/Linux environments is essential. Location: Remote- Delhi / NCR,Bangalore/Bengaluru, Hyderabad/Secunderabad,Chennai, Pune,Kolkata,Ahmedabad,Mumbai

Posted 3 weeks ago

Apply

5.0 - 10.0 years

7 - 12 Lacs

Mumbai

Work from Office

As a Big Data Engineer, you will develop, maintain, evaluate, and test big data solutions. You will be involved in data engineering activities like creating pipelines/workflows for Source to Target and implementing solutions that tackle the clients needs. Your primary responsibilities include: Design, build, optimize and support new and existing data models and ETL processes based on our clients business requirements. Build, deploy and manage data infrastructure that can adequately handle the needs of a rapidly growing data driven organization. Coordinate data access and security to enable data scientists and analysts to easily access to data whenever they need too Required education Bachelor's Degree Preferred education Master's Degree Required technical and professional expertise Must have 5+ years exp in Big Data -Hadoop Spark -Scala, Python HBase, Hive Good to have Aws -S3, Athena, Dynamo DB, Lambda, Jenkins GIT Developed Python and pyspark programs for data analysis. Good working experience with python to develop Custom Framework for generating of rules (just like rules engine). Developed Python code to gather the data from HBase and designs the solution to implement using Pyspark. Apache Spark Data Frames/RDD's were used to apply business transformations and utilized Hive Context objects to perform read/write operations Preferred technical and professional experience Understanding of Devops. Experience in building scalable end-to-end data ingestion and processing solutions Experience with object-oriented and/or functional programming languages, such as Python, Java and Scala

Posted 4 weeks ago

Apply

9.0 - 12.0 years

6 - 10 Lacs

Bengaluru

Work from Office

Educational Bachelor of Engineering Service Line Data & Analytics Unit Responsibilities Build robust, performing and high scalable, flexible data pipelines with a focus on time to market with quality.Responsibilities: Act as an active team member to ensure high code quality (unit testing, regression tests) delivered in time and within budget. Document the delivered code/solution Participate to the implementation of the releases following the change & release management processes Provide support to the operation team in case of major incidents for which engineering knowledge is required. Participate to effort estimations. Provide solutions (bug fixes) for problem mgt. Additional Responsibilities: Good knowledge on software configuration management systems Strong business acumen, strategy and cross-industry thought leadership Awareness of latest technologies and Industry trends Logical thinking and problem solving skills along with an ability to collaborate Two or three industry domain knowledge Understanding of the financial processes for various types of projects and the various pricing models available Client Interfacing skills Knowledge of SDLC and agile methodologies Project and Team management Technical and Professional : You have experience with most of these technologiesHDFS, Ozone, Hive, Impala, Spark, Atlas, Ranger. Knowledge of GraphQL, Venafi (Certificate Mgt) and Collibra (Data Governance) is an asset. Experience in a telecommunication environment and real-time technologies with focus on high availability and high-volume processing is an advantage:o Kafkao Flinko Spark Streaming You master programming languages as Java & Python/PySpark as well as SQL and are proficient in UNIX scripting Data formats like JSON, Parquet, XML and REST API have no secrets for you You have experience with CI/CD (GitLab/GitHub, Jenkins, Ansible, Nexus) for automated build & test. Knowledge of the Azure DevOps toolset is an asset.As project is preparing a “move to Azure”, the above will change slightly in the course of 2025. However, most of our current technological landscape remains a solid foundation for a role as EDH Data Engineer Preferred Skills: Technology-Analytics - Packages-Python - Big Data Technology-Big Data - Data Processing-Spark

Posted 1 month ago

Apply

2.0 - 7.0 years

5 - 9 Lacs

Pune

Work from Office

Educational Bachelor of Engineering Service Line Data & Analytics Unit Responsibilities A day in the life of an Infoscion As part of the Infosys delivery team, your primary role would be to interface with the client for quality assurance, issue resolution and ensuring high customer satisfaction. You will understand requirements, create and review designs, validate the architecture and ensure high levels of service offerings to clients in the technology domain. You will participate in project estimation, provide inputs for solution delivery, conduct technical risk planning, perform code reviews and unit test plan reviews. You will lead and guide your teams towards developing optimized high quality code deliverables, continual knowledge management and adherence to the organizational guidelines and processes. You would be a key contributor to building efficient programs/ systems and if you think you fit right in to help our clients navigate their next in their digital transformation journey, this is the place for you!If you think you fit right in to help our clients navigate their next in their digital transformation journey, this is the place for you! Additional Responsibilities: Knowledge of more than one technology Basics of Architecture and Design fundamentals Knowledge of Testing tools Knowledge of agile methodologies Understanding of Project life cycle activities on development and maintenance projects Understanding of one or more Estimation methodologies, Knowledge of Quality processes Basics of business domain to understand the business requirements Analytical abilities, Strong Technical Skills, Good communication skills Good understanding of the technology and domain Ability to demonstrate a sound understanding of software quality assurance principles, SOLID design principles and modelling methods Awareness of latest technologies and trends Excellent problem solving, analytical and debugging skills Technical and Professional : Primary skillsHadoop, Hive, HDFS Preferred Skills: Technology-Big Data - Hadoop-Hadoop

Posted 1 month ago

Apply

5.0 - 9.0 years

5 - 9 Lacs

Bengaluru

Work from Office

Educational Bachelor of Engineering,BCA,BSc,MCA,MTech,MSc Service Line Data & Analytics Unit Responsibilities "1. 5-8 yrs exp in Azure (Hands on experience in Azure Data bricks and Azure Data Factory)2. Good knowledge in SQL, PySpark.3. Should have knowledge in Medallion architecture pattern4. Knowledge on Integration Runtime5. Knowledge on different ways of scheduling jobs via ADF (Event/Schedule etc)6. Should have knowledge of AAS, Cubes.7. To create, manage and optimize the Cube processing.8. Good Communication Skills.9. Experience in leading a team" Additional Responsibilities: Good knowledge on software configuration management systems Strong business acumen, strategy and cross-industry thought leadership Awareness of latest technologies and Industry trends Logical thinking and problem solving skills along with an ability to collaborate Two or three industry domain knowledge Understanding of the financial processes for various types of projects and the various pricing models available Client Interfacing skills Knowledge of SDLC and agile methodologies Project and Team management Preferred Skills: Technology-Big Data - Data Processing-Spark

Posted 1 month ago

Apply

5.0 - 10.0 years

15 - 30 Lacs

Bengaluru

Work from Office

Description: To work on Data Analytical to triage and investigate data quality and data pipeline exceptions and reporting issues. Requirements: This role will support Data Operations and Reporting related projects but also will be helping with other projects as well if needed. In this role, you will leverage your strong analytical skills to triage and investigate data quality and data pipeline exceptions and reporting issues. The ideal candidate should be able to work independently and actively engage other functional teams as needed. This role requires researching transactions and events using large amounts of data. Technical Experience/Qualifications: • At least 5 years of experience in software development • At least 5 years of SQL experience in any RDBMS • Minimum 5 years of experience in Python • Strong analytical and problem-solving skill • Strong communication skill • Strong experience with data modeling • Strong experience in data analysis and reporting. • Experience with version control tools such as GitHub etc. • Experience with shell scripting and Linux • Knowledge of agile and scrum methodologies • Preferred experience in Hive SQL or related technologies such as Big Query etc. • Preferred experience in Big data technologies like Hadoop, AWS/GCP, S3, HIVE, Impala, HDFS, Spark, MapReduce • Preferred experience in reporting tools such as Looker or Tableau etc. • Preferred experience in finance and accounting but not required Job Responsibilities: Responsibilities: • Develop SQL queries as per technical requirements • Investigate and fix day to day data related issues • Develop test plan and execute test script • Data validation and analysis • Develop new reports/dashboard as per technical requirements • Modify existing reports/dashboards for bug fixes and enhancements • Develop new ETL scripts and modify existing in case of bug fixes and enhancements • Monitoring of ETL processes and fix issues in case of failure • Monitor scheduled jobs and fix issues in case of failure • Monitor data quality alerts and act on it What We Offer: Exciting Projects: We focus on industries like High-Tech, communication, media, healthcare, retail and telecom. Our customer list is full of fantastic global brands and leaders who love what we build for them. Collaborative Environment: You Can expand your skills by collaborating with a diverse team of highly talented people in an open, laidback environment — or even abroad in one of our global centers or client facilities! Work-Life Balance: GlobalLogic prioritizes work-life balance, which is why we offer flexible work schedules, opportunities to work from home, and paid time off and holidays. Professional Development: Our dedicated Learning & Development team regularly organizes Communication skills training(GL Vantage, Toast Master),Stress Management program, professional certifications, and technical and soft skill trainings. Excellent Benefits: We provide our employees with competitive salaries, family medical insurance, Group Term Life Insurance, Group Personal Accident Insurance , NPS(National Pension Scheme ), Periodic health awareness program, extended maternity leave, annual performance bonuses, and referral bonuses. Fun Perks: We want you to love where you work, which is why we host sports events, cultural activities, offer food on subsidies rates, Corporate parties. Our vibrant offices also include dedicated GL Zones, rooftop decks and GL Club where you can drink coffee or tea with your colleagues over a game of table and offer discounts for popular stores and restaurants!

Posted 1 month ago

Apply

1.0 - 3.0 years

15 - 30 Lacs

Bengaluru

Work from Office

About the Role Does digging deep for data and turning it into useful, impactful insights get you excited? Then you could be our next SDE II Data-Real Time Streaming. In this role, youll oversee your entire teams work, ensuring that each individual is working towards achieving their personal goals and Meeshos organisational goals. Moreover, youll keep an eye on all engineering projects and ensure the team is not straying from the right track. Youll also be tasked with directing programming activities, evaluating system performance, and designing new programs and features for smooth functioning. What you will do Build a platform for ingesting and processing multi terabytes of data daily Curate, build and transform raw data into scalable information Create prototypes and proofs-of-concept for iterative development Reduce technical debt with quality coding Keep a closer look at various projects and monitor the progress Carry on smooth collaborations with the sales team and engineering teams Provide management mentorship which sets the tone for holistic growth Ensure everyone is on the same page and taking ownership of the project What you will need Bachelors/Masters degree in Computer Science At least 1 to 3 years of professional experience Exceptional coding skills using Java, Scala, Python Working knowledge of Redis, MySQL and messaging systems like Kafka Knowledge of RxJava, Java Springboot, Microservices architecture Hands-on experience with the distributed systems architecture dealing with high throughput. Experience in building streaming and real-time solutions using Apache Flink/Spark Streaming/Samza. Familiarity with software engineering best practices across all stages of software development Expertise in Data system internalsStrong problem-solving and analytical skills Familiarity with Big Data systems (Spark/EMR, Hive/Impala, Delta Lake, Presto, Airflow, Data Lineage) is an advantage Familiarity with data modeling, end-to-end data pipelining, OLAP data cubes and BI tools is a plus Experience as a contributor/committer to the Big data stack is a plus Having been a contributor/committer to the big data stack Data modeling experience and end-to-end data pipelining experience is a plus Brownie points for knowledge of OLAP data cubes and BI tools

Posted 1 month ago

Apply

2.0 - 6.0 years

6 - 10 Lacs

Nagpur

Work from Office

Primine Software Private Limited is looking for BigData Engineer to join our dynamic team and embark on a rewarding career journey Develop and maintain big data solutions. Collaborate with data teams and stakeholders. Conduct data analysis and processing. Ensure compliance with big data standards and best practices. Prepare and maintain big data documentation. Stay updated with big data trends and technologies.

Posted 1 month ago

Apply

7.0 - 10.0 years

6 - 7 Lacs

Navi Mumbai, SBI Belapur

Work from Office

ISA Non captive RTH-Y Note: 1.This position requires the candidate to work from the office starting from day one clinet office. 2.Ensure that you perform basic validation and gauge the interest level of the candidate before uploading their profile to our system. 3.Candidate Band will be count as per their relevant experience. We will not entertain lesser experience profile for higher band. 4. Candidate full BGV is required before onboarding the candidate. 5. If required will regularize the candidate after 6months. Hence 6 months NOC is required from the DOJ. Mode of Interview: Face to Face (Mandatory). **JOB DESCRIPTION** Total Years of Experience : 7-10 Years Relevant Years of Experience : 7-10 Years Mandatory Skills : Cloudera DBA Detailed JD : Key Responsibilities: Provision and manage Cloudera clusters (CDP Private Cloud Base) Monitor cluster health, performance, and resource utilization Implement security (Kerberos, Ranger, TLS), HA, and backupstrategies Handle patching, upgrades, and incident response Collaborate with engineering and data teams to support workloads Skills Required: Strong hands-on with Cloudera Manager, Ambari, HDFS, Hive, Impala, Spark Linux administration and scripting skills (Shell, Python) Experience with Kerberos, Ranger, and audit/compliance setups Exposure to Cloudera Support and ticketing processes

Posted 1 month ago

Apply
cta

Start Your Job Search Today

Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.

Job Application AI Bot

Job Application AI Bot

Apply to 20+ Portals in one click

Download Now

Download the Mobile App

Instantly access job listings, apply easily, and track applications.

Featured Companies