Jobs
Interviews

8325 Pyspark Jobs - Page 7

Setup a job Alert
JobPe aggregates results for easy application access, but you actually apply on the job portal directly.

5.0 years

0 Lacs

Pune/Pimpri-Chinchwad Area

On-site

Company Description NielsenIQ is a consumer intelligence company that delivers the Full View™, the world’s most complete and clear understanding of consumer buying behavior that reveals new pathways to growth. Since 1923, NIQ has moved measurement forward for industries and economies across the globe. We are putting the brightest and most dedicated minds together to accelerate progress. Our diversity brings out the best in each other so we can leave a lasting legacy on the work that we do and the people that we do it with. NielsenIQ offers a range of products and services that leverage Machine Learning and Artificial Intelligence to provide insights into consumer behavior and market trends. This position opens the opportunity to apply the latest state of the art in AI/ML and data science to global and key strategic projects. Job Description NielsenIQ’s Innovation Team is growing our AI capabilities and is now looking to hire an AI/ML Data Scientist in India (Pune). for the Core Models team, a multidisciplinary team of researchers working on different areas of AI such as recommender systems, extreme classifiers, Large Language Models (LLMs), among others. As part of this team, you will stay up to date with the latest research in AI (with special focus on NLP, but also on Computer Vision and other AI related fields), implement current state-of-the-art algorithms in real-world and large-scale challenging problems as well as proposing novel ideas. Your main focus will be creating high-quality datasets for training and fine-tuning Custom Models for the company, LLMs and Recommender Systems, and training them to analyze the impact of the different versions of the data on model’s performance. The selected candidate will be responsible for designing and implementing scalable data pipelines and strategies to support all stages of the R&D process, e.g., fine-tuning or alignment through reinforcement learning. The results of the word will be critical to ensure the robustness, safety, and alignment of our AI models. You will also have the opportunity to produce scientific content such as patents or conference/journal papers. Job Responsibilities: Investigate, develop, and apply data pipelines with minimal technical supervision, always ensuring a combination of simplicity, scalability, reproducibility and maintainability within the ML solutions and source code. Train Deep Learning models (Transformer models) and analyze the impact of different versions of the data. Perform feasibility studies and analyze data to determine the most appropriate solution. Drive innovation and proactively contribute to our work on custom Large Language Models. Be able to communicate results to tech and non-tech audiences. To work as a member of a team, encouraging team building, motivation and cultivating effective team relations. Qualifications Required Education, Skills and Experience: Master's degree in computer science or an equivalent numerate discipline. At least 5+ years’ experience with evidence in a related field. Strong background in computer science, linear algebra, probability. Solid experience in Machine Learning and Deep Learning (special focus on Transformers). Proven experience in Natural Language Processing and Large Language Models. Proven experience building scalable data pipelines and ETLs. Able to understand scientific papers and develop ideas into executable code. Proven track record of innovation in creating novel algorithms and publishing the results in AI conferences/journals. Languages and technologies: Python, SQL, PySpark, Databricks, Pandas/Polars, PyArrow, PyTorch, Huggingface, git. Proactive attitude, constructive, intellectual curiosity, and persistence to find answers to questions. A proficient level of interpersonal and communication skills (English level B2 minimum). Keen to work as part of a diverse team of international colleagues and in a global inclusive culture. Additional Information: Preferred Education, Skills and Experience: PhD in science (NLP/Data Science is preferred) and specialized courses in one of the above-mentioned fields. Experience working with large real-world datasets and scalable ML solutions. Previous experience in e-commerce, retail and/or FMCG/Consumer Insight business. Agile methodologies development (SCRUM or Scale Agile). Additional Information Our Benefits Flexible working environment Volunteer time off LinkedIn Learning Employee-Assistance-Program (EAP) About NIQ NIQ is the world’s leading consumer intelligence company, delivering the most complete understanding of consumer buying behavior and revealing new pathways to growth. In 2023, NIQ combined with GfK, bringing together the two industry leaders with unparalleled global reach. With a holistic retail read and the most comprehensive consumer insights—delivered with advanced analytics through state-of-the-art platforms—NIQ delivers the Full View™. NIQ is an Advent International portfolio company with operations in 100+ markets, covering more than 90% of the world’s population. For more information, visit NIQ.com Want to keep up with our latest updates? Follow us on: LinkedIn | Instagram | Twitter | Facebook Our commitment to Diversity, Equity, and Inclusion NIQ is committed to reflecting the diversity of the clients, communities, and markets we measure within our own workforce. We exist to count everyone and are on a mission to systematically embed inclusion and diversity into all aspects of our workforce, measurement, and products. We enthusiastically invite candidates who share that mission to join us. We are proud to be an Equal Opportunity/Affirmative Action-Employer, making decisions without regard to race, color, religion, gender, gender identity or expression, sexual orientation, national origin, genetics, disability status, age, marital status, protected veteran status or any other protected class. Our global non-discrimination policy covers these protected classes in every market in which we do business worldwide. Learn more about how we are driving diversity and inclusion in everything we do by visiting the NIQ News Center: https://nielseniq.com/global/en/news-center/diversity-inclusion

Posted 2 days ago

Apply

8.0 - 12.0 years

0 Lacs

pune, maharashtra

On-site

As a Senior Infrastructure Architect at our organization, you will play a crucial role in our digital transformation journey. You will have the opportunity to be involved in cybersecurity, architecture, and data protection for our global organization. Collaborating with our team, you will provide expertise in designing and validating systems, infrastructure, technologies, and data protection. Your responsibilities will include participating in technical and business discussions to shape future architecture direction, analyzing data to develop architectural requirements, and contributing to infrastructure architecture governance. Additionally, you will be involved in designing and deploying infrastructure solutions that meet standardization, security, compliance, and quality requirements for various businesses. Your role will also entail researching emerging technologies and trends to support project development and operational activities, as well as coaching and mentoring team members. To excel in this role, you should hold a Bachelor's Degree with a minimum of 8 years of professional experience. You should possess experience in Azure infrastructure services, automating deployments, working in DevOps, and utilizing Databricks. Proficiency in database technologies, ETL tools, SQL queries optimization, and computing/network/storage design is essential. Furthermore, you should demonstrate an understanding of technical and business discussions, architecture standards, and requirements gathering. At our organization, we value diversity and recognize that individuals have unique working preferences. Therefore, we offer flexible working patterns, including remote work options and adaptable schedules to accommodate personal commitments. We believe in investing in our employees" well-being, fostering development, and cultivating leadership at all levels to create a supportive and inclusive work environment. Join our team at Baker Hughes, an energy technology company dedicated to innovating solutions for energy and industrial clients worldwide. With a legacy of over a century and a presence in more than 120 countries, we are committed to advancing energy technologies for a safer, cleaner, and more efficient future. If you are passionate about driving innovation and progress, we invite you to be part of our team and contribute to shaping the future of energy.,

Posted 2 days ago

Apply

8.0 - 12.0 years

0 Lacs

noida, uttar pradesh

On-site

As a Senior Data Engineering Architect at Iris Software, you will play a crucial role in leading enterprise-level data engineering projects on public cloud platforms like AWS, Azure, or GCP. Your responsibilities will include engaging with client managers to understand their business needs, conceptualizing solution options, and finalizing strategies with stakeholders. You will also be involved in team building, delivering Proof of Concepts (PoCs), and enhancing competencies within the organization. Your role will focus on building competencies in Data & Analytics, including Data Engineering, Analytics, Data Science, AI/ML, and Data Governance. Staying updated with the latest tools, best practices, and trends in the Data and Analytics field will be essential to drive innovation and excellence in your work. To excel in this position, you should hold a Bachelor's or Master's degree in a Software discipline and have extensive experience in Data architecture and implementing large-scale Data Lake/Data Warehousing solutions. Your background in Data Engineering should demonstrate leadership in solutioning, architecture, and successful project delivery. Strong communication skills in English, both written and verbal, are essential for effective collaboration with clients and team members. Proficiency in tools such as AWS Glue, Redshift, Azure Data Lake, Databricks, Snowflake, and databases, along with programming skills in Spark, Spark SQL, PySpark, and Python, are mandatory competencies for this role. Joining Iris Software offers a range of perks and benefits designed to support your financial, health, and overall well-being. From comprehensive health insurance and competitive salaries to flexible work arrangements and continuous learning opportunities, we are dedicated to providing a supportive and rewarding work environment where your success and happiness are valued. If you are inspired to grow your career in Data Engineering and thrive in a culture that values talent and personal growth, Iris Software is the place for you. Be part of a dynamic team where you can be valued, inspired, and encouraged to be your best professional and personal self.,

Posted 2 days ago

Apply

5.0 - 9.0 years

0 Lacs

karnataka

On-site

We are looking for a highly skilled and experienced Senior Python & ML Engineer with expertise in PySpark, machine learning, and large language models (LLMs). You will play a key role in designing, developing, and implementing scalable data pipelines, machine learning models, and LLM-powered applications. In this role, you will need to have a solid understanding of Python's ecosystem, distributed computing using PySpark, and practical experience in AI optimization. Your responsibilities will include designing and maintaining robust data pipelines with PySpark, optimizing PySpark jobs for efficiency on large datasets, and ensuring data integrity throughout the pipeline. You will also be involved in developing, training, and deploying machine learning models using key ML libraries such as scikit-learn, TensorFlow, and PyTorch. Tasks will include feature engineering, model selection, hyperparameter tuning, and integrating ML models into production systems for scalability and reliability. Additionally, you will research, experiment with, and integrate state-of-the-art Large Language Models (LLMs) into applications. This will involve developing solutions that leverage LLMs for tasks like natural language understanding, text generation, summarization, and question answering. You will also fine-tune pre-trained LLMs for specific business needs and datasets, and explore techniques for prompt engineering, RAG (Retrieval Augmented Generation), and LLM evaluation. Collaboration is key in this role, as you will work closely with data scientists, engineers, and product managers to understand requirements and translate them into technical solutions. You will mentor junior team members, contribute to best practices for code quality, testing, and deployment, and stay updated on the latest advancements in Python, PySpark, ML, and LLMs. Furthermore, you will be responsible for deploying, monitoring, and maintaining models and applications in production environments using MLOps principles. Troubleshooting and resolving issues related to data pipelines, ML models, and LLM applications will also be part of your responsibilities. To be successful in this role, you should have a Bachelor's or Master's degree in Computer Science, Data Science, Engineering, or a related field. Strong proficiency in Python programming, PySpark, machine learning, and LLMs is essential. Experience with cloud platforms like AWS, Azure, or GCP is preferred, along with strong problem-solving, analytical, communication, and teamwork skills. Nice-to-have skills include familiarity with R and Shiny, streaming data technologies, containerization technologies, MLOps tools, graph databases, and contributions to open-source projects.,

Posted 2 days ago

Apply

4.0 - 8.0 years

0 Lacs

karnataka

On-site

ZS is a place where passion changes lives. As a management consulting and technology firm focused on improving life and how we live it, our most valuable asset is our people. Here you'll work side-by-side with a powerful collective of thinkers and experts shaping life-changing solutions for patients, caregivers, and consumers, worldwide. ZSers drive impact by bringing a client-first mentality to each and every engagement. We partner collaboratively with our clients to develop custom solutions and technology products that create value and deliver company results across critical areas of their business. Bring your curiosity for learning, bold ideas, courage, and passion to drive life-changing impact to ZS. At ZS, we honor the visible and invisible elements of our identities, personal experiences, and belief systemsthe ones that comprise us as individuals, shape who we are, and make us unique. We believe your personal interests, identities, and desire to learn are part of your success here. Learn more about our diversity, equity, and inclusion efforts and the networks ZS supports to assist our ZSers in cultivating community spaces, obtaining the resources they need to thrive, and sharing the messages they are passionate about. ZS's Beyond Healthcare Analytics (BHCA) Team is shaping one of the key growth vector areas for ZS, Beyond Healthcare engagement, comprising clients from industries like Quick service restaurants, Technology, Food & Beverage, Hospitality, Travel, Insurance, Consumer Products Goods & other such industries across North America, Europe & South East Asia region. The BHCA India team currently has a presence across New Delhi, Pune, and Bengaluru offices and is continuously expanding further at a great pace. The BHCA India team works with colleagues across clients and geographies to create and deliver real-world pragmatic solutions leveraging AI SaaS products & platforms, Generative AI applications, and other Advanced analytics solutions at scale. What You'll Do: - Build, Refine and Use ML Engineering platforms and components. - Scaling machine learning algorithms to work on massive datasets and strict SLAs. - Build and orchestrate model pipelines including feature engineering, inferencing, and continuous model training. - Implement ML Ops including model KPI measurements, tracking, model drift & model feedback loop. - Collaborate with client-facing teams to understand business context at a high level and contribute to technical requirement gathering. - Implement basic features aligning with technical requirements. - Write production-ready code that is easily testable, understood by other developers, and accounts for edge cases and errors. - Ensure the highest quality of deliverables by following architecture/design guidelines, coding best practices, periodic design/code reviews. - Write unit tests as well as higher-level tests to handle expected edge cases and errors gracefully, as well as happy paths. - Use bug tracking, code review, version control, and other tools to organize and deliver work. - Participate in scrum calls and agile ceremonies, and effectively communicate work progress, issues, and dependencies. - Consistently contribute to researching & evaluating the latest architecture patterns/technologies through rapid learning, conducting proof-of-concepts, and creating prototype solutions. What You'll Bring: - A master's or bachelor's degree in Computer Science or related field from a top university. - 4+ years hands-on experience in ML development. - Good understanding of the fundamentals of machine learning. - Strong programming expertise in Python, PySpark/Scala. - Expertise in crafting ML Models for high performance and scalability. - Experience in implementing feature engineering, inferencing pipelines, and real-time model predictions. - Experience in ML Ops to measure and track model performance, experience working with MLFlow. - Experience with Spark or other distributed computing frameworks. - Experience in ML platforms like Sage maker, Kubeflow. - Experience with pipeline orchestration tools such as Airflow. - Experience in deploying models to cloud services like AWS, Azure, GCP, Azure ML. - Expertise in SQL, SQL DB's. - Knowledgeable of core CS concepts such as common data structures and algorithms. - Collaborate well with teams with different backgrounds/expertise/functions. Perks & Benefits: ZS offers a comprehensive total rewards package including health and well-being, financial planning, annual leave, personal growth, and professional development. Our robust skills development programs, multiple career progression options, internal mobility paths, and collaborative culture empower you to thrive as an individual and global team member. We are committed to giving our employees a flexible and connected way of working. A flexible and connected ZS allows us to combine work from home and on-site presence at clients/ZS offices for the majority of our week. The magic of ZS culture and innovation thrives in both planned and spontaneous face-to-face connections. Travel: Travel is a requirement at ZS for client-facing ZSers; business needs of your project and client are the priority. While some projects may be local, all client-facing ZSers should be prepared to travel as needed. Travel provides opportunities to strengthen client relationships, gain diverse experiences, and enhance professional growth by working in different environments and cultures. Considering applying At ZS, we're building a diverse and inclusive company where people bring their passions to inspire life-changing impact and deliver better outcomes for all. We are most interested in finding the best candidate for the job and recognize the value that candidates with all backgrounds, including non-traditional ones, bring. If you are interested in joining us, we encourage you to apply even if you don't meet 100% of the requirements listed above. ZS is an equal opportunity employer and is committed to providing equal employment and advancement opportunities without regard to any class protected by applicable law. To Complete Your Application: Candidates must possess or be able to obtain work authorization for their intended country of employment. An online application, including a full set of transcripts (official or unofficial), is required to be considered. NO AGENCY CALLS, PLEASE.,

Posted 2 days ago

Apply

4.0 - 8.0 years

0 Lacs

karnataka

On-site

As a valued member of Infosys Consulting, you will play a crucial role in supporting large Oil & Gas/Utilities prospects by showcasing Infosys" unique value proposition through practical use cases across the value chain. Your responsibilities will include gathering, identifying, and documenting business requirements, as well as creating functional specifications for new systems and processes. Utilizing your expertise in assessing current processes, conducting gap analyses, and designing future processes, you will recommend changes and drive continuous improvement using methodologies such as Six Sigma and Lean. In your role, you will be involved in Technology Project Management, which includes overseeing technology vendors and client stakeholders. You will also manage large projects and programs in a multi-vendor, globally distributed team environment, leveraging Agile principles and DevOps capabilities. Collaboration with the IT Project Management Office will be essential as you support the implementation of client-specific digital solutions, from business case development to IT strategy and tool/software selection. Your expertise in designing and implementing scalable data pipelines, ETL/ELT workflows, and optimized data models across cloud data warehouses and lakes will enable reliable access to high-quality data for business insights and strategic decision-making. You will also be responsible for building and maintaining dashboards, reports, and visualizations using tools like Power BI and Tableau, while conducting deep-dive analyses to evaluate business performance and identify opportunities. Collaboration with business stakeholders to translate strategic objectives into data-driven solutions, defining KPIs, and enabling self-service analytics will be a key aspect of your role. Additionally, you will work closely with client IT teams and business stakeholders to uncover opportunities and derive actionable insights. Participation in internal firm-building activities and supporting sales efforts for new and existing clients through proposal creation and sales presentation facilitation will also be part of your responsibilities. To qualify for this position, you should have at least 3-5 years of experience in data engineering, ideally within the Oil & Gas or Utilities sector. Strong communication skills, both written and verbal, are essential, along with a proven track record in business analysis, product design, or project management. A Bachelor's degree or Full-time MBA/PGDM from Tier 1/Tier 2 B-Schools in India or a foreign equivalent is required. Preferred qualifications include knowledge of digital technologies and agile development practices, as well as the ability to work effectively in a cross-cultural team environment. Strong teamwork, communication skills, and the ability to interact with mid-level managers of client organizations are highly valued. This position is preferred to be located in Electronic City, Bengaluru, but other locations such as Hyderabad, Chennai, Pune, Gurgaon, and Chandigarh are also considered based on business needs. Please note that the job may require extended periods of computer work and communication via telephone, email, or face-to-face interactions.,

Posted 2 days ago

Apply

5.0 - 9.0 years

0 Lacs

kolkata, west bengal

On-site

Genpact is a global professional services and solutions firm that is dedicated to delivering outcomes that shape the future. With a team of over 125,000 professionals across more than 30 countries, we are motivated by curiosity, entrepreneurial agility, and the desire to create lasting value for our clients. Our purpose is the relentless pursuit of a world that works better for people, and we serve and transform leading enterprises, including the Fortune Global 500, leveraging our deep business and industry knowledge, digital operations services, and expertise in data, technology, and AI. We are currently looking for a Principal Consultant - Snowflake Sr. Data Engineer (Snowflake + Python/Pyspark) to join our team! As a Snowflake Sr. Data Engineer, you will be responsible for providing technical direction and leading a group of developers to address a common goal. You should have experience in the IT industry and be proficient in building productionized data ingestion and processing data pipelines in Snowflake. Additionally, you should be well-versed in data warehousing concepts and have expertise in Snowflake features and integration with other data processing tools. Experience with Python programming and Pyspark for data analysis is essential for this role. Key Responsibilities: - Work on requirement gathering, analysis, designing, development, and deployment - Write SQL queries against Snowflake and develop scripts for Extract, Load, and Transform data - Understand Data Warehouse concepts and Snowflake Architecture - Hands-on experience with Snowflake utilities such as SnowSQL, SnowPipe, tables, Tasks, Streams, and more - Experience with Snowflake AWS data services or Azure data services - Proficiency in Python programming language and knowledge of packages like pandas, NumPy, etc. - Design and develop efficient ETL jobs using Python and Pyspark - Use Python and Pyspark for data cleaning, pre-processing, and transformation tasks - Implement CDC or SCD type 2 and build data ingestion pipelines - Work with workflow management tools for scheduling and managing ETL jobs Qualifications: - B.E./ Masters in Computer Science, Information Technology, or Computer Engineering - Relevant years of experience as a Snowflake Sr. Data Engineer - Skills in Snowflake, Python/Pyspark, AWS/Azure, ETL concepts, Airflow, or any orchestration tools, Data Warehousing concepts If you are passionate about leveraging your skills to drive innovative solutions and create value in a dynamic environment, we encourage you to apply for this exciting opportunity. Join us in shaping the future and making a difference!,

Posted 2 days ago

Apply

4.0 - 8.0 years

0 Lacs

karnataka

On-site

As an Assistant Manager - Analytics, you will be responsible for overseeing data-driven projects, designing intricate data solutions, offering valuable insights to stakeholders, and contributing to the advancement of our Ads product and business metrics. Your role will involve delving into deep insights on Ads core product, conducting large-scale experimentation on Adtech innovation, and forecasting demand-supply to drive growth within our Ads product and a multifaceted Ads Entertainment business. The Central Analytics team, situated within various business & product teams in a matrix structure, serves as a thought partner by providing comprehensive data insights to steer strategic decisions. This team acts as a strategic enabler for JioHotstar Ads business and product functions. By analyzing consumer experience, consumer supply, advertisers demand, and our Ad serving capabilities, we aim to achieve goals (KPIs) across Ads product, Advertisers objectives, and Entertainment business planning. We implement experiments, leverage GenAI for innovative problem-solving, and construct analytical frameworks to guide key decisions. Reporting to the Manager - Product Analytics, your key responsibilities will include applying analytics knowledge in problem-solving, generating and delivering quality data insights through reports, dashboards, and structured documentation using tools like Power BI and Tableau, developing a profound understanding of the data platform and technology stack, utilizing statistical techniques for validation, effectively communicating complex data concepts to diverse audiences, partnering with stakeholders for identifying opportunities and supporting strategic decisions, managing projects end-to-end, contributing data-driven insights in experiments, and fostering a culture of innovation, collaboration, and continuous improvement. To excel in this role, you should demonstrate expertise in predictive analysis with proficiency in R, SQL, Python, and Pyspark. Familiarity with big data platforms and tools like Hadoop, Spark, and Hive is preferred. Experience in dashboard building and data visualization using tools such as Tableau and Power BI is advantageous. You should possess advanced technical skills and the ability to collect, organize, and disseminate information accurately. A background in digital analytics, clickstream data, passion for the entertainment industry, and understanding of online video streaming platforms are desirable. Experience in Adtech and OTT platforms is also preferred. Ideally, you should hold a Bachelor's or Master's degree in Engineering, Mathematics, Operational Research, Statistics, Physics, or a related technical discipline, coupled with 4-6 years of experience in Business/Product Analytics, preferably from consumer technology companies. JioStar is an equal opportunity employer that values diversity and aims to create an inclusive workplace free from discrimination.,

Posted 2 days ago

Apply

3.0 - 7.0 years

0 Lacs

pune, maharashtra

On-site

As a Data Analytics Lead at Cummins Inc., you will be responsible for facilitating data, compliance, and environment governance processes for the assigned domain. Your role includes leading analytics projects to provide insights for the business, integrating data analysis findings into governance solutions, and ingesting key data into the data lake while ensuring the creation and maintenance of relevant metadata and data profiles. You will coach team members, business teams, and stakeholders to find necessary and relevant data, contribute to communities of practice promoting responsible analytics use, and develop the capability of peers and team members within the Analytics Ecosystem. Additionally, you will mentor and review the work of less experienced team members, integrate data from various source systems to build models for business use, and cleanse data to ensure accuracy and reduce redundancy. Your responsibilities will also involve leading the preparation of communications to leaders and stakeholders, designing and implementing data/statistical models, collaborating with stakeholders on analytics initiatives, and automating complex workflows and processes using tools like Power Automate and Power Apps. You will manage version control and collaboration using GITLAB, utilize SharePoint for project management and data collaboration, and provide regular updates on work progress via JIRA/Meets to stakeholders. Qualifications: - College, university, or equivalent degree in a relevant technical discipline, or relevant equivalent experience required. - This position may require licensing for compliance with export controls or sanctions regulations. Competencies: - Balancing stakeholders - Collaborating effectively - Communicating clearly and effectively - Customer focus - Managing ambiguity - Organizational savvy - Data Analytics - Data Mining - Data Modeling - Data Communication and Visualization - Data Literacy - Data Profiling - Data Quality - Project Management - Valuing differences Technical Skills: - Advanced Python - Databricks, Pyspark - Advanced SQL, ETL tools - Power Automate - Power Apps - SharePoint - GITLAB - Power BI - Jira - Mendix - Statistics Soft Skills: - Strong problem-solving and analytical abilities - Excellent communication and stakeholder management skills - Proven ability to lead a team - Strategic thinking - Advanced project management Experience: - Intermediate level of relevant work experience required - This is a Hybrid role Join Cummins Inc. and be part of a dynamic team where you can utilize your technical and soft skills to make a significant impact in the field of data analytics.,

Posted 2 days ago

Apply

5.0 - 10.0 years

0 Lacs

karnataka

On-site

Looking for a DBT Developer with 5 to 10 years of experience We invite applications for the role of Lead Consultant, DBT Data Engineer! As a DBT Data Engineer, you will be responsible for providing technical direction and leading a group of one or more developers to achieve a common goal. Your responsibilities will include designing, developing, and automating ETL processes using DBT and AWS. You will be tasked with building robust data pipelines to transfer data from various sources to data warehouses or data lakes. Collaborating with cross-functional teams is crucial to ensure data accuracy, completeness, and consistency. Data cleansing, validation, and transformation are essential to maintain data quality and integrity. Optimizing database and query performance will be part of your responsibilities to ensure efficient data processing. Working closely with data analysts and data scientists, you will provide clean and reliable data for analysis and modeling. Your role will involve writing SQL queries against Snowflake, developing scripts for Extract, Load, and Transform operations. Hands-on experience with Snowflake utilities such as SnowSQL, SnowPipe, Tasks, Streams, Time travel, Cloning, Optimizer, Metadata Manager, data sharing, stored procedures, and UDFs is required. Proficiency with Snowflake cloud data warehouse and AWS S3 bucket or Azure blob storage container for data integration is necessary. Additionally, you should have solid experience in Python/Pyspark integration with Snowflake and cloud services like AWS/Azure. A sound understanding of ETL tools and data integration techniques is vital for this role. You will collaborate with business stakeholders to grasp data requirements and develop ETL solutions accordingly. Strong programming skills in languages like Python, Java, and/or Scala are expected. Experience with big data technologies such as Kafka and cloud computing platforms like AWS is advantageous. Familiarity with database technologies such as SQL, NoSQL, and/or Graph databases is beneficial. Your experience in requirement gathering, analysis, designing, development, and deployment will be valuable. Building data ingestion pipelines and deploying using CI/CD tools like Azure boards, Github, and writing automated test cases are desirable skills. Client-facing project experience and knowledge of Snowflake Best Practices will be beneficial in this role. If you are a skilled DBT Data Engineer with a passion for data management and analytics, we encourage you to apply for this exciting opportunity!,

Posted 2 days ago

Apply

2.0 years

0 Lacs

India

On-site

At H1, we believe access to the best healthcare information is a basic human right. Our mission is to provide a platform that can optimally inform every doctor interaction globally. This promotes health equity and builds needed trust in healthcare systems. To accomplish this our teams harness the power of data and AI-technology to unlock groundbreaking medical insights and convert those insights into action that result in optimal patient outcomes and accelerates an equitable and inclusive drug development lifecycle. Visit h1.co to learn more about us. As a Software Engineer on the search Engineering team you will support and develop the search infrastructure of the company. This involves working with TB’s of data, indexing, ranking and retrieval of medical data to support the search in backend infra. What You'll Do At H1 The Search Engineering team is responsible for developing and maintaining the company's core search infrastructure. Our objective is to enable fast, accurate, and scalable search across terabytes of medical data. This involves building systems for efficient data ingestion, indexing, ranking, and retrieval that power key product features and user experiences. As a Software Engineer on the Search Engineering team, your day typically includes: Working with our search infrastructure – writing and maintaining code that ingests large-scale data in Elasticsearch. Designing and implementing high-performance APIs that serve search use cases with low latency. Building and maintaining end-to-end features using Node.js and GraphQL, ensuring scalability and maintainability. Collaborating with cross-functional teams – including product managers and data engineers to align on technical direction and deliver impactful features to our users. Take ownership of the search codebase–proactively debug, troubleshoot, and resolve issues quickly to ensure stability and performance. Consistently produce simple, elegant designs and write high-quality, maintainable code that can be easily understood and reused by teammates. Demonstrate a strong focus on performance optimization, ensuring systems are fast, efficient, and scalable. Communicate effectively and collaborate across teams in a fast-paced, dynamic environment. Stay up to date with the latest advancements in AI and search technologies, identifying opportunities to integrate cutting-edge capabilities into our platforms. About You You bring strong hands-on technical skills and experience in building robust backend APIs. You thrive on solving complex challenges with innovative, scalable solutions and take pride in maintaining high code quality through thorough testing.You are able to align your work with broader organizational goals and actively contribute to strategic initiatives. You proactively identify risks and propose solutions early in the project lifecycle to avoid downstream issues.You are curious, eager to learn, and excited to grow in a collaborative, high-performing engineering team environment. Requirements 1–2 years of professional experience. Strong programming skills in TypeScript, Node.js, and Python (Mandatory) Practical experience with Docker and Kubernetes Good to have: Big Data technologies (e.g., Scala, Hadoop, PySpark), Golang, GraphQL, Elasticsearch, and LLMs Not meeting all the requirements but still feel like you’d be a great fit? Tell us how you can contribute to our team in a cover letter! H1 OFFERS Full suite of health insurance options, in addition to generous paid time off Pre-planned company-wide wellness holidays Retirement options Health & charitable donation stipends Impactful Business Resource Groups Flexible work hours & the opportunity to work from anywhere The opportunity to work with leading biotech and life sciences companies in an innovative industry with a mission to improve healthcare around the globe

Posted 2 days ago

Apply

200.0 years

0 Lacs

Bengaluru, Karnataka, India

On-site

Data Science (GenAI & Prompt engineering) – Bangalore Business Analytics Analyst 2 About CITI Citi's mission is to serve as a trusted partner to our clients by responsibly providing financial services that enable growth and economic progress. We have 200 years of experience helping our clients meet the world's toughest challenges and embrace its greatest opportunities. Analytics and Information Management (AIM) Citi AIM was established in 2003, and is located across multiple cities in India – Bengaluru, Chennai, Pune and Mumbai. It is a global community that objectively connects and analyzes information, to create actionable intelligence for our business leaders. It identifies fact-based opportunities for revenue growth in partnership with the businesses. The function balances customer needs, business strategy, and profit objectives using best in class and relevant analytic methodologies. What do we do? The North America Consumer Bank – Data Science and Modeling team analyzes millions of prospects and billions of customer level transactions using big data tools and machine learning, AI techniques to unlock opportunities for our clients in meeting their financial needs and create economic value for the bank. The team extracts relevant insights, identifies business opportunities, converts business problems into modeling framework, uses big data tools, latest deep learning and machine learning algorithms to build predictive models, implements solutions and designs go-to-market strategies for a huge variety of business problems. Role Description The role will be Business Analytics Analyst 2 in the Data Science and Modeling of North America Consumer Bank team The role will report to the AVP / VP leading the team What do we offer: The Next Gen Analytics (NGA) team is a part of the Analytics & Information Management (AIM) unit. The NGA modeling team will focus on the following areas of work: Role Expectations: Client Obsession – Create client centric analytic solution to business problems. Individual should be able to have a holistic view of multiple businesses and develop analytic solutions accordingly. Analytic Project Execution – Own and deliver multiple and complex analytic projects. This would require an understanding of business context, conversion of business problems in modeling, and implementing such solutions to create economic value. Domain expert – Individuals are expected to be domain expert in their sub field, as well as have a holistic view of other business lines to create better solutions. Key fields of focus are new customer acquisition, existing customer management, customer retention, product development, pricing and payment optimization and digital journey. Modeling and Tech Savvy – Always up to date with the latest use cases of modeling community, machine learning and deep learning algorithms and share knowledge within the team. Statistical mind set – Proficiency in basic statistics, hypothesis testing, segmentation and predictive modeling. Communication skills – Ability to translate and articulate technical thoughts and ideas to a larger audience including influencing skills with peers and senior management. Strong project management skills. Ability to coach and mentor juniors. Contribute to organizational initiatives in wide ranging areas including competency development, training, organizational building activities etc. Role Responsibilities: Work with large and complex datasets using a variety of tools (Python, PySpark, SQL, Hive, etc.) and frameworks to build Deep learning/generative AI solutions for various business requirements. Primary focus areas include model training/fine-tuning, model validation, model deployment, and model governance related to multiple portfolios. Design, fine-tune and implement LLMs/GenAI applications using techniques like prompt engineering, Retrieval Augmented Generation (RAG) and model fine-tuning Responsible for documenting data requirements, data collection/processing/cleaning, and exploratory data analysis, including utilizing deep learning /generative AI algorithms and, data visualization techniques. Incumbents in this role may often be referred to as Data Scientists. Specialization in marketing, risk, digital, and AML fields possible, applying Deep learning & generative AI models to innovate in these domains. Collaborate with team members and business partners to build model-driven solutions using cutting-edge Generative AI models (e.g., Large Language Models) and also at times, ML/traditional methods (XGBoost, Linear, Logistic, Segmentation, etc.) Work with model governance & fair lending teams to ensure compliance of models in accordance with Citi standards. Appropriately assess risk when business decisions are made, demonstrating particular consideration for the firm's reputation and safeguarding Citigroup, its clients and assets, by driving compliance with applicable laws, rules, and regulations, adhering to Policy, applying sound ethical judgment regarding personal behavior, conduct and business practices, and escalating, managing and reporting control issues with transparency. What do we look for: If you are a bright and talented individual looking for a career in AI and Machine Learning with a focus on Generative AI , Citi has amazing opportunities for you. Bachelor’s Degree with atleast 3 years of experience in data analytics, or Master’s Degree with 2 years of experience in data analytics, or PhD. Technical Skills Hands-on experience in PySpark/Python/R programing along with strong experience in SQL. 2-4 years of experience working on deep learning, and generative AI applications Experience working on Transformers/ LLMs (OpenAI, Claude, Gemini etc.,), Prompt engineering, RAG based architectures and relevant tools/frameworks such as TensorFlow, PyTorch, Hugging Face Transformers, LangChain, LlamaIndex etc., Solid understanding of deep learning, transformers/language models. Familiarity with vector databases and fine-tuning techniques Experience working with large and multiple datasets, data warehouses and ability to pull data using relevant programs and coding. Strong background in Statistical Analysis. Capability to validate/maintain deployed models in production Self-motivated and able to implement innovative solutions at fast pace Experience in Credit Cards and Retail Banking is preferred Competencies Strong communication skills Multiple stake holder management Strong analytical and problem solving skills Excellent written and oral communication skills Strong team player Control orientated and Risk awareness Working experience in a quantitative field Willing to learn and can-do attitude Ability to build partnerships with cross-function leaders Education: Bachelor's / master’s degree in economics / Statistics / Mathematics / Information Technology / Computer Applications / Engineering etc. from a premier institute Other Details Employment: Full Time Industry: Credit Cards, Retail Banking, Financial Services, Banking ------------------------------------------------------ Job Family Group: Decision Management ------------------------------------------------------ Job Family: Specialized Analytics (Data Science/Computational Statistics) ------------------------------------------------------ Time Type: ------------------------------------------------------ Most Relevant Skills Please see the requirements listed above. ------------------------------------------------------ Other Relevant Skills For complementary skills, please see above and/or contact the recruiter. ------------------------------------------------------ Citi is an equal opportunity employer, and qualified candidates will receive consideration without regard to their race, color, religion, sex, sexual orientation, gender identity, national origin, disability, status as a protected veteran, or any other characteristic protected by law. If you are a person with a disability and need a reasonable accommodation to use our search tools and/or apply for a career opportunity review Accessibility at Citi. View Citi’s EEO Policy Statement and the Know Your Rights poster.

Posted 2 days ago

Apply

6.0 - 8.0 years

0 Lacs

Hyderabad, Telangana, India

On-site

Roles and responsibilities: Design and implement data pipelines for supply chain data (e.g., inventory, shipping, procurement). Develop and maintain data warehouses and data lakes.  Ensure data quality, integrity, and security. Collaborate with supply chain stakeholders to identify analytics requirements. Develop data models and algorithms for predictive analytics (e.g., demand forecasting, supply chain optimization). Implement data visualization tools (e.g., Tableau, Power BI). Integrate data from various sources (e.g., ERP, PLMs, other data sources). Develop APIs for data exchange.  Work with cross-functional teams (e.g., supply chain, logistics, IT). Communicate technical concepts to non-technical stakeholders. Experience with machine learning algorithms & concepts Knowledge of data governance and compliance. Strong problem-solving and analytical skills. Excellent communication and collaboration skills. Ability to work in a fast-paced environment. Technical Skills: Bachelor's degree in Computer Science, Information Technology, or related field. 6-8 years of experience in data engineering. Proficiency in: Programming languages - Python, Java, SQL, Spark SQL. Data technologies - Hadoop, PySpark, NoSQL databases. Data visualization tools - Qliksense, Tableau, Power BI Cloud platforms - Azure Data Factory, Azure Databricks, AWS

Posted 2 days ago

Apply

2.0 - 10.0 years

0 Lacs

India

Remote

Pay Range: ₹400-500/hour Location: Remote (India) Mode: One-to-One Sessions Only (No batch teaching) We are hiring a Part-Time PySpark, Databricks Tutor who can deliver personalized, one-on-one online sessions to college and university-level students . The ideal candidate should have hands-on experience in big data technologies , particularly PySpark and Databricks , and should be comfortable teaching tools and techniques commonly used in the computer science and data engineering fields . Key Responsibilities: Deliver engaging one-to-one remote tutoring sessions focused on PySpark, Apache Spark, Databricks , and related tools. Teach practical use cases, project implementation techniques, and hands-on coding for real-world applications. Adapt teaching style based on individual student levels – beginners to advanced. Provide support with assignments, project work, and interview preparation. Ensure clarity in communication and foster an interactive learning environment. Required Skills & Qualifications: Experience: 2 to 10 years in the field of big data, data engineering, or related roles using PySpark and Databricks. Education: Bachelor’s or Master’s degree in Computer Science, Data Science, or relevant field. Strong English communication skills – both verbal and written. Familiarity with Spark SQL, Delta Lake, notebooks, and data pipelines. Ability to teach technical concepts with simplicity and clarity. Job Requirements: Freshers with strong knowledge and teaching ability may also apply. Must have a personal laptop and stable Wi-Fi connection . Must be serious and committed to long-term part-time work. Candidates who have applied before should not reapply . 💡 Note: This is a remote, part-time opportunity , and sessions will be conducted one-to-one , not in batch format. This role is ideal for professionals, freelancers, or educators passionate about sharing knowledge. 📩 Apply now only if you agree with the pay rate (₹400-500/hr) and meet the listed criteria. Let’s inspire the next generation of data engineers!

Posted 2 days ago

Apply

3.0 years

0 Lacs

Chennai, Tamil Nadu, India

On-site

Job Description Data Engineer, Chennai We’re seeking a highly motivated Data Engineer to join our agile, cross-functional team and drive end-to-end data pipeline development in a cloud-native, big data ecosystem. You’ll leverage ETL/ELT best practices and data lakehouse paradigms to deliver scalable solutions. Proficiency in SQL, Python, Spark, and modern data orchestration tools (e.g. Airflow) is essential, along with experience in CI/CD, DevOps, and containerized environments like Docker and Kubernetes. This is your opportunity to make an impact in a fast-paced, data-driven culture. Responsibilities Responsible for data pipeline development and maintenance Contribute to development, maintenance, testing strategy, design discussions, and operations of the team Participate in all aspects of agile software development including design, implementation, and deployment Responsible for the end-to-end lifecycle of new product features / components Ensuring application performance, uptime, and scale, maintaining high standards of code quality and thoughtful application design Work with a small, cross-functional team on products and features to drive growth Learning new tools, languages, workflows, and philosophies to grow Research and suggest new technologies for boosting the product Have an impact on product development by making important technical decisions, influencing the system architecture, development practices and more Qualifications Excellent team player with strong communication skills B.Sc. in Computer Sciences or similar 3-5 years of experience in Data Pipeline development 3-5 years of experience in PySpark / Databricks 3-5 years of experience in Python / Airflow Knowledge of OOP and design patterns Knowledge of server-side technologies such as Java, Spring Experience with Docker containers, Kubernetes and Cloud environments Expertise in testing methodologies (Unit-testing, TDD, mocking) Fluent with large scale SQL databases Good problem-solving and analysis abilities Requirements - Advantage Experience with Azure cloud services Experience with Agile Development methodologies Experience with Git Additional Information Our Benefits Flexible working environment Volunteer time off LinkedIn Learning Employee-Assistance-Program (EAP) About NIQ NIQ is the world’s leading consumer intelligence company, delivering the most complete understanding of consumer buying behavior and revealing new pathways to growth. In 2023, NIQ combined with GfK, bringing together the two industry leaders with unparalleled global reach. With a holistic retail read and the most comprehensive consumer insights—delivered with advanced analytics through state-of-the-art platforms—NIQ delivers the Full View™. NIQ is an Advent International portfolio company with operations in 100+ markets, covering more than 90% of the world’s population. For more information, visit NIQ.com Want to keep up with our latest updates? Follow us on: LinkedIn | Instagram | Twitter | Facebook Our commitment to Diversity, Equity, and Inclusion NIQ is committed to reflecting the diversity of the clients, communities, and markets we measure within our own workforce. We exist to count everyone and are on a mission to systematically embed inclusion and diversity into all aspects of our workforce, measurement, and products. We enthusiastically invite candidates who share that mission to join us. We are proud to be an Equal Opportunity/Affirmative Action-Employer, making decisions without regard to race, color, religion, gender, gender identity or expression, sexual orientation, national origin, genetics, disability status, age, marital status, protected veteran status or any other protected class. Our global non-discrimination policy covers these protected classes in every market in which we do business worldwide. Learn more about how we are driving diversity and inclusion in everything we do by visiting the NIQ News Center: https://nielseniq.com/global/en/news-center/diversity-inclusion

Posted 2 days ago

Apply

5.0 - 9.0 years

0 Lacs

Hyderabad, Telangana, India

On-site

Join Amgen’s Mission of Serving Patients At Amgen, if you feel like you’re part of something bigger, it’s because you are. Our shared mission—to serve patients living with serious illnesses—drives all that we do. Since 1980, we’ve helped pioneer the world of biotech in our fight against the world’s toughest diseases. With our focus on four therapeutic areas –Oncology, Inflammation, General Medicine, and Rare Disease– we reach millions of patients each year. As a member of the Amgen team, you’ll help make a lasting impact on the lives of patients as we research, manufacture, and deliver innovative medicines to help people live longer, fuller happier lives. Our award-winning culture is collaborative, innovative, and science based. If you have a passion for challenges and the opportunities that lay within them, you’ll thrive as part of the Amgen team. Join us and transform the lives of patients while transforming your career. [Data Engineer] What You Will Do Let’s do this. Let’s change the world. In this vital role you will be responsible for designing, building, maintaining, analyzing, and interpreting data to provide actionable insights that drive business decisions. This role involves working with large datasets, developing reports, supporting and executing data governance initiatives and, visualizing data to ensure data is accessible, reliable, and efficiently managed. The ideal candidate has strong technical skills, experience with big data technologies, and a deep understanding of data architecture and ETL processes Roles & Responsibilities: Ø Design, develop, and maintain data solutions for data generation, collection, and processing Ø Be a key team member that assists in design and development of the data pipeline Ø Create data pipelines and ensure data quality by implementing ETL processes to migrate and deploy data across systems Ø Contribute to the design, development, and implementation of data pipelines, ETL/ELT processes, and data integration solutions Ø Take ownership of data pipeline projects from inception to deployment, manage scope, timelines, and risks Ø Collaborate with cross-functional teams to understand data requirements and design solutions that meet business needs Ø Develop and maintain data models, data dictionaries, and other documentation to ensure data accuracy and consistency Ø Implement data security and privacy measures to protect sensitive data Ø Leverage cloud platforms (AWS preferred) to build scalable and efficient data solutions Ø Collaborate and communicate effectively with product teams Ø Collaborate with Data Architects, Business SMEs, and Data Scientists to design and develop end-to-end data pipelines to meet fast paced business needs across geographic regions Ø Identify and resolve complex data-related challenges Ø Adhere to best practices for coding, testing, and designing reusable code/component Ø Explore new tools and technologies that will help to improve ETL platform performance Ø Participate in sprint planning meetings and provide estimations on technical implementation What We Expect Of You We are all different, yet we all use our unique contributions to serve patients. Basic Qualifications: Basic Qualifications and Experience: Master's degree / Bachelor's degree and 5 to 9 years Computer Science, IT or related field experience Functional Skills: Must-Have Skills Hands on experience with big data technologies and platforms, such as Databricks, Apache Spark (PySpark, SparkSQL), workflow orchestration, performance tuning on big data processing Proficiency in data analysis tools (eg. SQL) and experience with data visualization tools Excellent problem-solving skills and the ability to work with large, complex datasets Knowledge of data protection regulations and compliance requirements (e.g., GDPR, CCPA) Good-to-Have Skills: Experience with ETL tools such as Apache Spark, and various Python packages related to data processing, machine learning model development Strong understanding of data modeling, data warehousing, and data integration concepts Knowledge of Python/R, Databricks, SageMaker, cloud data platforms Soft Skills: Excellent critical-thinking and problem-solving skills Strong communication and collaboration skills Demonstrated awareness of how to function in a team setting Demonstrated presentation skills What You Can Expect Of Us As we work to develop treatments that take care of others, we also work to care for your professional and personal growth and well-being. From our competitive benefits to our collaborative culture, we’ll support your journey every step of the way. In addition to the base salary, Amgen offers competitive and comprehensive Total Rewards Plans that are aligned with local industry standards. Apply now and make a lasting impact with the Amgen team. careers.amgen.com As an organization dedicated to improving the quality of life for people around the world, Amgen fosters an inclusive environment of diverse, ethical, committed and highly accomplished people who respect each other and live the Amgen values to continue advancing science to serve patients. Together, we compete in the fight against serious disease. Amgen is an Equal Opportunity employer and will consider all qualified applicants for employment without regard to race, color, religion, sex, sexual orientation, gender identity, national origin, protected veteran status, disability status, or any other basis protected by applicable law. We will ensure that individuals with disabilities are provided reasonable accommodation to participate in the job application or interview process, to perform essential job functions, and to receive other benefits and privileges of employment. Please contact us to request accommodation.

Posted 2 days ago

Apply

40.0 years

0 Lacs

Hyderabad, Telangana, India

On-site

About Amgen Amgen harnesses the best of biology and technology to fight the world's toughest diseases, and make people's lives easier, fuller and longer. We discover, develop, manufacture and deliver innovative medicines to help millions of patients. Amgen helped establish the biotechnology industry more than 40 years ago and remains on the cutting-edge of innovation, using technology and human genetic data to push beyond what fs known today. About The Role Role Description: We are looking for an Associate Data Engineer with deep expertise in writing data pipelines to build scalable, high-performance data solutions. The ideal candidate will be responsible for developing, optimizing and maintaining complex data pipelines, integration frameworks, and metadata-driven architectures that enable seamless access and analytics. This role prefers deep understanding of the big data processing, distributed computing, data modeling, and governance frameworks to support self-service analytics, AI-driven insights, and enterprise-wide data management. Roles & Responsibilities: Design, develop, and maintain data solutions for data generation, collection, and processing Be a key team member that assists in design and development of the data pipeline Create data pipelines and ensure data quality by implementing ETL processes to migrate and deploy data across systems Contribute to the design, development, and implementation of data pipelines, ETL/ELT processes, and data integration solutions Collaborate with cross-functional teams to understand data requirements and design solutions that meet business needs Develop and maintain data models, data dictionaries, and other documentation to ensure data accuracy and consistency Implement data security and privacy measures to protect sensitive data Leverage cloud platforms (AWS preferred) to build scalable and efficient data solutions Collaborate and communicate effectively with product teams Identify and resolve complex data-related challenges Adhere to best practices for coding, testing, and designing reusable code/component Explore new tools and technologies that will help to improve ETL platform performance Participate in sprint planning meetings and provide estimations on technical implementation What We Expect From You We are all different, yet we all use our unique contributions to serve patients. Basic Qualifications: Bachelor’s degree and 2 to 4 years of Computer Science, IT or related field experience OR Diploma and 4 to 7 years of Computer Science, IT or related field experience Preferred Qualifications: Functional Skills: Must-Have Skills : Hands on experience with big data technologies and platforms, such as Databricks, Apache Spark (PySpark, SparkSQL), AWS, Redshift, Snowflake, workflow orchestration, performance tuning on big data processing Proficiency in data analysis tools (eg. SQL) and experience with data visualization tools. Proficient in SQL for extracting, transforming, and analyzing complex datasets from relational data stores. Experience with ETL tools such as Apache Spark, and various Python packages related to data processing, machine learning model development Good-to-Have Skills: Experience with data modeling, performance tuning on relational and graph databases ( e.g. Marklogic, Allegrograph, Stardog, RDF Triplestore). Understanding of data modeling, data warehousing, and data integration concepts Knowledge of Python/R, Databricks, SageMaker, cloud data platform Experience with Software engineering best-practices, including but not limited to version control, infrastructure-as-code, CI/CD, and automated testing Professional Certifications : AWS Certified Data Engineer preferred Databricks Certificate preferred Soft Skills: Excellent critical-thinking and problem-solving skills Strong communication and collaboration skills Demonstrated awareness of how to function in a team setting As an Associate Data Engineer at Amgen, you will be involved in the development and maintenance of data infrastructure and solutions. You will collaborate with a team of data engineers to design and implement data pipelines, perform data analysis, and ensure data quality. Your strong technical skills, problem-solving abilities, and attention to detail will contribute to the effective management and utilization of data for insights and decision-making.

Posted 2 days ago

Apply

15.0 years

0 Lacs

India

On-site

Job Summary As part of the data leadership team, the Capability Lead – Databricks will be responsible for building, scaling, and delivering Databricks-based data and AI capabilities across the organization. This leadership role involves technical vision, solution architecture, team building, partnership development, and delivery excellence using Databricks Unified Analytics Platform across industries. The individual will collaborate with clients, alliance partners (Databricks, Azure, AWS), internal stakeholders, and sales teams to drive adoption of lakehouse architectures, data engineering best practices, and AI/ML modernization. Areas of Responsibility 1. Offering and Capability Development: Develop and enhance Snowflake-based data platform offerings and accelerators Define best practices, architectural standards, and reusable frameworks for Snowflake Collaborate with alliance teams to strengthen partnership with Snowflake 2. Technical Leadership: Provide architectural guidance for Snowflake solution design and implementation Lead solutioning efforts for proposals, RFIs, and RFPs involving Snowflake Conduct technical reviews and ensure adherence to design standards. Act as a technical escalation point for complex project challenges 3. Delivery Oversight: Support delivery teams with technical expertise across Snowflake projects Drive quality assurance, performance optimization, and project risk mitigation. Review project artifacts and ensure alignment with Snowflake best practices Foster a culture of continuous improvement and delivery excellence 4. Talent Development: Build and grow a high-performing Snowflake capability team. Define skill development pathways and certification goals for team members. Mentor architects, developers, and consultants on Snowflake technologies Drive community of practice initiatives to share knowledge and innovations 5. Business Development Support: Engage with sales and pre-sales teams to position Snowflake capabilities Contribute to account growth by identifying new Snowflake opportunities Participate in client presentations, workshops, and technical discussions 6. Thought Leadership and Innovation Build thought leadership through whitepapers, blogs, and webinars Stay updated with Snowflake product enhancements and industry trends This role is highly collaborative and will work extremely closely with cross functional teams to fulfill the above responsibilities. Job Requirements: 12–15 years of experience in data engineering, analytics, and AI/ML 3–5 years of strong hands-on experience with Databricks (on Azure, AWS, or GCP) Expertise in Spark (PySpark/Scala), Delta Lake, Unity Catalog, MLflow, and Databricks notebooks Experience designing and implementing Lakehouse architectures at scale Familiarity with data governance, security, and compliance frameworks (GDPR, HIPAA, etc.) Experience with real-time and batch data pipelines (Structured Streaming, Auto Loader, Kafka, etc.) Strong understanding of MLOps and AI/ML lifecycle management Certifications in Databricks (e.g., Databricks Certified Data Engineer Professional, ML Engineer Associate) are preferred Experience with hyperscaler ecosystems (Azure Data Lake, AWS S3, GCP GCS, ADF, Glue, etc.) Experience managing large, distributed teams and working with CXO-level stakeholders Strong problem-solving, analytical, and decision-making skills Excellent verbal, written, and client-facing communication

Posted 2 days ago

Apply

5.0 - 8.0 years

0 Lacs

Chennai, Tamil Nadu, India

On-site

Job Title Data Scientist - Deputy Manager Job Description Job title: Data Scientist - Deputy Manager Your role: Implements solutions to problems using data analysis, data mining, optimization tools and machine learning techniques and statistics Build data-science and technology based algorithmic solutions to address business needs Design large scale models using Regression, Linear Models Family, Time-series models. Drive the collection of new data and the refinement of existing data sources Analyze and interpret the results of analytics experiments Applies a global approach to analytical solutions-both within a business area and across the enterprise Ability to use data for Exploratory, descriptive, Inferential, Prescriptive, and Advanced Analytics Ability to share dashboards, reports, and Analytical insights from data Experience of having done visualization on large datasets – Preferred – added advantage Technical Knowledge and Skills required Experience solving analytical problems using quantitative approaches Passion for empirical research and for answering hard questions with data Ability to manipulate and analyze complex, high-volume, high-dimensionality data from varying sources Ability to apply a flexible analytic approach that allows for results at varying levels of precision Ability to communicate complex quantitative analysis in a clear, precise, and actionable manner Expert knowledge of an analysis tool such as Pyspark and Python. Experience working with large data sets, experience working with distributed computing tools a plus (Map/Reduce, Hadoop, Hive, etc.) Familiarity with relational databases and SQL You're the right fit if: (4 x bullets max) 5 - 8 years of experience with engineering or equivalent background Experience with solving analytical problems using quantitative approaches Ability to manipulate and analyze complex, high-volume, high-dimensionality data from varying sources Ability to apply a flexible analytic approach that allows for results at varying levels of precision Ability to communicate complex quantitative analysis in a clear, precise, and actionable manner Expert knowledge of an analysis tool such as R, Python Experience working with large data sets, experience working with distributed computing tools a plus (Map/Reduce, Hadoop, Hive, etc.) Familiarity with relational databases and SQL How We Work Together We believe that we are better together than apart. For our office-based teams, this means working in-person at least 3 days per week. Onsite roles require full-time presence in the company’s facilities. Field roles are most effectively done outside of the company’s main facilities, generally at the customers’ or suppliers’ locations. Indicate if this role is an office/field/onsite role. About Philips We are a health technology company. We built our entire company around the belief that every human matters, and we won't stop until everybody everywhere has access to the quality healthcare that we all deserve. Do the work of your life to help the lives of others. Learn more about our business. Discover our rich and exciting history. Learn more about our purpose. If you’re interested in this role and have many, but not all, of the experiences needed, we encourage you to apply. You may still be the right candidate for this or other opportunities at Philips. Learn more about our culture of impact with care here.

Posted 2 days ago

Apply

8.0 years

0 Lacs

Hyderabad, Telangana, India

On-site

Job Summary 8 10 years of experience as an Azure Data Engineer with expertise in Databricks and Azure Data Factory. Programming expertise in SQL, Spark and Python is mandatory 2+ years of experience with medical claims in healthcare and/or managed care is required Expertise in developing ETL/ELT pipelines for BI/ data visualization. Familiarity with normalized, dimensional, star schema and snowflake schematic models are mandatory Prior experience using version control to manage code changes

Posted 2 days ago

Apply

3.0 - 4.0 years

0 Lacs

Mumbai Metropolitan Region

On-site

Bachelor's degree in Computer Science, Engineering, Information Technology, or a related field. 3-4 years of hands-on experience in data engineering, with a strong focus on AWS cloud services. Proficiency in Python for data manipulation, scripting, and automation. Strong command of SQL for data querying, transformation, and database management. Demonstrable Experience With AWS Data Services, Including Amazon S3: Data Lake storage and management. AWS Glue: ETL service for data preparation. Amazon Redshift: Cloud data warehousing. AWS Lambda: Serverless computing for data processing. Amazon EMR: Managed Hadoop framework for big data processing (Spark/PySpark experience highly preferred). AWS Kinesis (or Kafka): Real-time data streaming. Strong analytical, problem-solving, and debugging skills. Excellent communication and collaboration abilities, with the capacity to work effectively in an agile team environment. Responsibilities Troubleshoot and resolve data-related issues and performance bottlenecks in existing pipelines. Develop and maintain data quality checks, monitoring, and alerting mechanisms to ensure data pipeline reliability. Participate in code reviews, contribute to architectural discussions, and promote best practices in data engineering.

Posted 2 days ago

Apply

2.0 years

0 Lacs

Gurugram, Haryana, India

On-site

Achieving our goals starts with supporting yours. Grow your career, access top-tier health and wellness benefits, build lasting connections with your team and our customers, and travel the world using our extensive route network. Come join us to create what’s next. Let’s define tomorrow, together. Description United's Digital Technology team designs, develops, and maintains massively scaling technology solutions brought to life with innovative architectures, data analytics, and digital solutions. Find your future at United! We’re reinventing what our industry looks like, and what an airline can be – from the planes we fly to the people who fly them. When you join us, you’re joining a global team of 100,000+ connected by a shared passion with a wide spectrum of experience and skills to lead the way forward. Achieving our ambitions starts with supporting yours. Evolve your career and find your next opportunity. Get the care you need with industry-leading health plans and best-in-class programs to support your emotional, physical, and financial wellness. Expand your horizons with travel across the world’s biggest route network. Connect outside your team through employee-led Business Resource Groups. Create what’s next with us. Let’s define tomorrow together. Job Overview And Responsibilities Data Engineering organization is responsible for driving data driven insights & innovation to support the data needs for commercial and operational projects with a digital focus. Data Engineer will be responsible to partner with various teams to define and execute data acquisition, transformation, processing and make data actionable for operational and analytics initiatives that create sustainable revenue and share growth Design, develop, and implement streaming and near-real time data pipelines that feed systems that are the operational backbone of our business Execute unit tests and validating expected results to ensure accuracy & integrity of data and applications through analysis, coding, writing clear documentation and problem resolution This role will also drive the adoption of data processing and analysis within the Hadoop environment and help cross train other members of the team Leverage strategic and analytical skills to understand and solve customer and business centric questions Coordinate and guide cross-functional projects that involve team members across all areas of the enterprise, vendors, external agencies and partners Leverage data from a variety of sources to develop data marts and insights that provide a comprehensive understanding of the business Develop and implement innovative solutions leading to automation Use of Agile methodologies to manage projects Mentor and train junior engineers This position is offered on local terms and conditions. Expatriate assignments and sponsorship for employment visas, even on a time-limited visa status, will not be awarded. Qualifications What’s needed to succeed (Minimum Qualifications): BS/BA, in computer science or related STEM field 2+ years of IT experience in software development 2+ years of development experience using Java, Python, Scala 2+ years of experience with Big Data technologies like PySpark, Hadoop, Hive, HBASE, Kafka, Nifi 2+ years of experience with relational database systems like MS SQL Server, Oracle, Teradata Creative, driven, detail-oriented individuals who enjoy tackling tough problems with data and insights Individuals who have a natural curiosity and desire to solve problems are encouraged to apply Must be legally authorized to work in India for any employer without sponsorship Must be fluent in English and Hindi (written and spoken) Successful completion of interview required to meet job qualification Reliable, punctual attendance is an essential function of the position What will help you propel from the pack (Preferred Qualifications): Masters in computer science or related STEM field Experience with cloud based systems like AWS, AZURE or Google Cloud Certified Developer / Architect on AWS Strong experience with continuous integration & delivery using Agile methodologies Data engineering experience with transportation/airline industry Strong problem-solving skills Strong knowledge in Big Data

Posted 2 days ago

Apply

8.0 years

0 Lacs

Gurugram, Haryana, India

On-site

Data Scientist || 8 Years || Gurgaon Primary skills: • Solid experience in building ML Models. • Proficient in SQL, Python, PySpark, Spark ML language. • Good understanding of cloud platforms such as AWS (preferred), Azure or GCP. • Proficient in source code controls using Github. Secondary skills: • Experience using any Auto ML products like DataRobot / H2O AI. • Provide inputs to build Artificial Intelligence (AI) roadmap for marketing based on TE&O architecture and capability delivery timelines. • Accountable for identifying, embedding, promoting, and ensuring continuous improvement within the use of new data and advanced analytics across the teams

Posted 2 days ago

Apply

5.0 years

0 Lacs

Gurugram, Haryana, India

On-site

Achieving our goals starts with supporting yours. Grow your career, access top-tier health and wellness benefits, build lasting connections with your team and our customers, and travel the world using our extensive route network. Come join us to create what’s next. Let’s define tomorrow, together. Description Description - External United's Digital Technology team designs, develops, and maintains massively scaling technology solutions brought to life with innovative architectures, data analytics, and digital solutions. Our Values : At United Airlines, we believe that inclusion propels innovation and is the foundation of all that we do. Our Shared Purpose: "Connecting people. Uniting the world." drives us to be the best airline for our employees, customers, and everyone we serve, and we can only do that with a truly diverse and inclusive workforce. Our team spans the globe and is made up of diverse individuals all working together with cutting-edge technology to build the best airline in the history of aviation. With multiple employee-run "Business Resource Group" communities and world-class benefits like health insurance, parental leave, and space available travel, United is truly a one-of-a-kind place to work that will make you feel welcome and accepted. Come join our team and help us make a positive impact on the world. Job Overview And Responsibilities This role will be responsible for collaborating with the Business and IT teams to identify the value, scope, features and delivery roadmap for data engineering products and solutions. Responsible for communicating with stakeholders across the board, including customers, business managers, and the development team to make sure the goals are clear and the vision is aligned with business objectives. Perform data analysis using SQL Data Quality Analysis, Data Profiling and Summary reports Trend Analysis and Dashboard Creation based on Visualization technique Execute the assigned projects/ analysis as per the agreed timelines and with accuracy and quality. Complete analysis as required and document results and formally present findings to management Perform ETL workflow analysis, create current/future state data flow diagrams and help the team assess the business impact of any changes or enhancements Understand the existing Python code work books and write pseudo codes Collaborate with key stakeholders to identify the business case/value and create documentation. Should have excellent communication and analytical skills. This position is offered on local terms and conditions. Expatriate assignments and sponsorship for employment visas, even on a time-limited visa status, will not be awarded. United Airlines is an equal opportunity employer. United Airlines recruits, employs, trains, compensates, and promotes regardless of race, religion, color, national origin, gender identity, sexual orientation, physical ability, age, veteran status, and other protected status as required by applicable law. Qualifications - External Required BE, BTECH or equivalent, in computer science or related STEM field 5+ years of total IT experience as either a Data Analyst/Business Data Analyst or as a Data Engineer 2+ years of experience with Big Data technologies like PySpark, Hadoop, Redshift etc. 3+ years of experience with writing SQL queries on RDBMS or Cloud based database Experience with Visualization tools such as Spotfire, PowerBI, Quicksight etc Experience in Data Analysis and Requirements Gathering Strong problem-solving skills Creative, driven, detail-oriented focus, requiring tackling of tough problems with data and insights. Natural curiosity and desire to solve problems. Must be legally authorized to work in India for any employer without sponsorship Must be fluent in English and Hindi (written and spoken) Successful completion of interview required to meet job qualification Reliable, punctual attendance is an essential function of the position Qualifications Preferred AWS Certification preferred Strong experience with continuous integration & delivery using Agile methodologies Data engineering experience with transportation/airline industry

Posted 2 days ago

Apply

4.0 years

0 Lacs

Gurugram, Haryana, India

On-site

Achieving our goals starts with supporting yours. Grow your career, access top-tier health and wellness benefits, build lasting connections with your team and our customers, and travel the world using our extensive route network. Come join us to create what’s next. Let’s define tomorrow, together. Description As an airline, safety is our most important principle. And our Corporate Safety team is responsible for making sure safety is top of mind in every action we take. From conducting flight safety investigations and educating pilots on potential safety threats to implementing medical programs and helping prevent employee injuries, our team is instrumental in running a safe and successful airline for our customers and employees. Job Overview And Responsibilities Corporate safety is integral for ensuring a safe workplace for our employees and travel experience for our customers. This role is responsible for supporting the development and implementation of a cohesive safety data strategy and supporting the Director of Safety Management Systems (SMS) in growing United’s Corporate Safety Predictive Analytics capabilities. This Senior Analyst will serve as a subject matter expert for corporate safety data analytics and predictive insight strategy and execution. This position will be responsible for supporting new efforts to deliver insightful data analysis and build new key metrics for use by the entire United Safety organization, with the goal of enabling data driven decision making and understanding for corporate safety. The Senior Analyst will be responsible for becoming the subject matter expert in several corporate safety specific data streams and leveraging this expertise to deliver insights which are actionable and allow for a predictive approach to safety risk mitigation. Develop and implement predictive/prescriptive data analytics workflows for Safety Data Management and streamlining processes Collaborate with Digital Technology and United Operational teams to analyze, predict and reduce safety risks and provide measurable solutions Partner with Digital Technology team to develop streamlined and comprehensive data analytics workstreams Support United’s Safety Management System (SMS) with predictive data analytics by designing and developing statistical models Manage and maintain the project portfolio of SMS data team Areas of focus will include, but are not limited to: Predictive and prescriptive analytics Train and validate models Creation and maintenance of standardized corporate safety performance metrics Design and implementation of new data pipelines Delivery of prescriptive analysis insights to internal stakeholders Design and maintain new and existing corporate safety data pipelines and analytical workflows Create and manage new methods for data analysis which provide prescriptive and predictive insights on corporate safety data Partner with US and India based internal partners to establish new data analysis workflows and provide analytical support to corporate and divisional work groups Collaborate with corporate and divisional safety partners to ensure standardization and consistency between all safety analytics efforts enterprise wide Provide support and ongoing subject matter expertise regarding a set of high priority corporate safety datasets and ongoing analytics efforts on those datasets Provide tracking and status update reporting on ongoing assignments, projects, and efforts to US and India based leaders This position is offered on local terms and conditions. Expatriate assignments and sponsorship for employment visas, even on a time-limited visa status, will not be awarded. This position is for United Airlines Business Services Pvt. Ltd - a wholly owned subsidiary of United Airlines Inc. Qualifications What’s needed to succeed (Minimum Qualifications): Bachelor's degree Bachelor's degree in computer science, data science, information sytems, engineering, or another quantitative field (i.e. mathematics, statistics, economics, etc.) 4+ years experience in data analytics, predictive modeling, or statistics Expert level SQL skills Experience with Microsoft SQL Server Management Studio and hands-on experience working with massive data sets Proficiency writing complex code using both traditional and modern technologies/languages (i.e. Python, HTML, Javascript, Power Automate, Spark Node, etc.) for queries, procedures, and analytic processing to create useable data insight Ability to study/understand business needs, then design a data/technology solution that connects business processes with quantifiable outcomes Strong project management and communication skills 3-4 years working with complex data (data analytics, information science, data visualization or other relevant quantitative field Must be legally authorized to work in India for any employer without sponsorship Must be fluent in English (written and spoken) Successful completion of interview required to meet job qualification Reliable, punctual attendance is an essential function of the position What will help you propel from the pack (Preferred Qualifications): Master's degree ML / AI experience Experience with PySpark, Apache, or Hadoop to deal with massive data sets

Posted 2 days ago

Apply
cta

Start Your Job Search Today

Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.

Job Application AI Bot

Job Application AI Bot

Apply to 20+ Portals in one click

Download Now

Download the Mobile App

Instantly access job listings, apply easily, and track applications.

Featured Companies