Jobs
Interviews

8325 Pyspark Jobs - Page 10

Setup a job Alert
JobPe aggregates results for easy application access, but you actually apply on the job portal directly.

0.0 years

0 Lacs

Bengaluru, Karnataka

On-site

Job Description: Application Developer Bangalore, Karnataka, India AXA XL offers risk transfer and risk management solutions to clients globally. We offer worldwide capacity, flexible underwriting solutions, a wide variety of client-focused loss prevention services and a team-based account management approach. AXA XL recognizes data and information as critical business assets, both in terms of managing risk and enabling new business opportunities. This data should not only be high quality, but also actionable – enabling AXA XL’s executive leadership team to maximize benefits and facilitate sustained advantage. What you’ll be DOING What will your essential responsibilities include? We are seeking an experienced ETL Developer to support and evolve our enterprise data integration workflows. The ideal candidate will have deep expertise in Informatica PowerCenter, strong hands-on experience with Azure Data Factory and Databricks, and a passion for building scalable, reliable ETL pipelines. This role is critical for both day-to-day operational reliability and long-term modernization of our data engineering stack in the Azure cloud. Key Responsibilities: Maintain, monitor, and troubleshoot existing Informatica PowerCenter ETL workflows to ensure operational reliability and data accuracy. Enhance and extend ETL processes to support new data sources, updated business logic, and scalability improvements. Develop and orchestrate PySpark notebooks in Azure Databricks for data transformation, cleansing, and enrichment. Configure and manage Databricks clusters for performance optimization and cost efficiency. Implement Delta Lake solutions that support ACID compliance, versioning, and time travel for reliable data lake operations. Automate data workflows using Databricks Jobs and Azure Data Factory (ADF) pipelines. Design and manage scalable ADF pipelines, including parameterized workflows and reusable integration patterns. Integrate with Azure Blob Storage and ADLS Gen2 using Spark APIs for high-performance data ingestion and output. Ensure data quality, consistency, and governance across legacy and cloud-based pipelines. Collaborate with data analysts, engineers, and business teams to deliver clean, validated data for reporting and analytics. Participate in the full Software Development Life Cycle (SDLC) from design through deployment, with an emphasis on maintainability and audit readiness. Develop maintainable and efficient ETL logic and scripts following best practices in security and performance. Troubleshoot pipeline issues across data infrastructure layers, identifying and resolving root causes to maintain reliability. Create and maintain clear documentation of technical designs, workflows, and data processing logic for long-term maintainability and knowledge sharing. Stay informed on emerging cloud and data engineering technologies to recommend improvements and drive innovation. Follow internal controls, audit protocols, and secure data handling procedures to support compliance and operational standards. Provide accurate time and effort estimates for assigned development tasks, accounting for complexity and risk. What you will BRING We’re looking for someone who has these abilities and skills: Advanced experience with Informatica PowerCenter, including mappings, workflows, session tuning, and parameterization Expertise in Azure Databricks + PySpark, including: Notebook development Cluster configuration and tuning Delta Lake (ACID, versioning, time travel) Job orchestration via Databricks Jobs or ADF Integration with Azure Blob Storage and ADLS Gen2 using Spark APIs Strong hands-on experience with Azure Data Factory: Building and managing pipelines Parameterization and dynamic datasets Notebook integration and pipeline monitoring Proficiency in SQL, PL/SQL, and scripting languages such as Python, Bash, or PowerShell Strong understanding of data warehousing, dimensional modeling, and data profiling Familiarity with Git, CI/CD pipelines, and modern DevOps practices Working knowledge of data governance, audit trails, metadata management, and compliance standards such as HIPAA and GDPR Effective problem-solving and troubleshooting skills with the ability to resolve performance bottlenecks and job failures Awareness of Azure Functions, App Services, API Management, and Application Insights Understanding of Azure Key Vault for secrets and credential management Familiarity with Spark-based big data ecosystems (e.g., Hive, Kafka) is a plus Who WE are AXA XL, the P&C and specialty risk division of AXA, is known for solving complex risks. For mid-sized companies, multinationals and even some inspirational individuals we don’t just provide re/insurance, we reinvent it. How? By combining a comprehensive and efficient capital platform, data-driven insights, leading technology, and the best talent in an agile and inclusive workspace, empowered to deliver top client service across all our lines of business property, casualty, professional, financial lines and specialty. With an innovative and flexible approach to risk solutions, we partner with those who move the world forward. Learn more at axaxl.com What we OFFER Inclusion AXA XL is committed to equal employment opportunity and will consider applicants regardless of gender, sexual orientation, age, ethnicity and origins, marital status, religion, disability, or any other protected characteristic. At AXA XL, we know that an inclusive culture and enables business growth and is critical to our success. That’s why we have made a strategic commitment to attract, develop, advance and retain the most inclusive workforce possible, and create a culture where everyone can bring their full selves to work and reach their highest potential. It’s about helping one another — and our business — to move forward and succeed. Five Business Resource Groups focused on gender, LGBTQ+, ethnicity and origins, disability and inclusion with 20 Chapters around the globe. Robust support for Flexible Working Arrangements Enhanced family-friendly leave benefits Named to the Diversity Best Practices Index Signatory to the UK Women in Finance Charter Learn more at axaxl.com/about-us/inclusion-and-diversity. AXA XL is an Equal Opportunity Employer. Total Rewards AXA XL’s Reward program is designed to take care of what matters most to you, covering the full picture of your health, wellbeing, lifestyle and financial security. It provides competitive compensation and personalized, inclusive benefits that evolve as you do. We’re committed to rewarding your contribution for the long term, so you can be your best self today and look forward to the future with confidence. Sustainability At AXA XL, Sustainability is integral to our business strategy. In an ever-changing world, AXA XL protects what matters most for our clients and communities. We know that sustainability is at the root of a more resilient future. Our 2023-26 Sustainability strategy, called “Roots of resilience”, focuses on protecting natural ecosystems, addressing climate change, and embedding sustainable practices across our operations. Our Pillars: Valuing nature: How we impact nature affects how nature impacts us. Resilient ecosystems - the foundation of a sustainable planet and society – are essential to our future. We’re committed to protecting and restoring nature – from mangrove forests to the bees in our backyard – by increasing biodiversity awareness and inspiring clients and colleagues to put nature at the heart of their plans. Addressing climate change: The effects of a changing climate are far-reaching and significant. Unpredictable weather, increasing temperatures, and rising sea levels cause both social inequalities and environmental disruption. We're building a net zero strategy, developing insurance products and services, and mobilizing to advance thought leadership and investment in societal-led solutions. Integrating ESG: All companies have a role to play in building a more resilient future. Incorporating ESG considerations into our internal processes and practices builds resilience from the roots of our business. We’re training our colleagues, engaging our external partners, and evolving our sustainability governance and reporting. AXA Hearts in Action : We have established volunteering and charitable giving programs to help colleagues support causes that matter most to them, known as AXA XL’s “Hearts in Action” programs. These include our Matching Gifts program, Volunteering Leave, and our annual volunteering day – the Global Day of Giving. For more information, please see axaxl.com/sustainability.

Posted 3 days ago

Apply

2.0 - 6.0 years

0 Lacs

karnataka

On-site

As a Data Scientist at Setu, you will have the opportunity to be a part of a team that is revolutionizing the fintech landscape. Setu believes in empowering every company to become a fintech company by providing them with cutting-edge APIs. The Data Science team at Setu is dedicated to understanding the vast population of India and creating solutions for various fintech sectors such as personal lending, collections, PFM, and BBPS. In this role, you will have the unique opportunity to delve deep into the business objectives and technical architecture of multiple companies, leading to a customer-centric approach that fosters innovation and delights customers. The learning potential in this role is immense, with the chance to explore, experiment, and build critical, scalable, and high-value use cases. At Setu, innovation is not just a goal; it's a way of life. The team is constantly pushing boundaries and introducing groundbreaking methods to drive business growth, enhance customer experiences, and streamline operational processes. From computer vision to natural language processing and Generative AI, each day presents new challenges and opportunities for breakthroughs. To excel in this role, you will need a minimum of 2 years of experience in Data Science and Machine Learning. Strong knowledge in statistics, tree-based techniques, machine learning, inference, hypothesis testing, and optimizations is essential. Proficiency in Python programming, building Data Pipelines, feature engineering, pandas, sci-kit-learn, SQL, and familiarity with TensorFlow/PyTorch are also required. Experience with deep learning techniques and understanding of DevOps/MLOps will be a bonus. Setu offers a dynamic and inclusive work environment where you will have the opportunity to work closely with the founding team who built and scaled public infrastructure such as UPI, GST, and Aadhaar. The company is dedicated to your growth and provides various benefits such as access to a fully stocked library, tickets to conferences, learning sessions, and development allowance. Additionally, Setu offers comprehensive health insurance, access to mental health counselors, and a beautiful office space designed to foster creativity and collaboration. If you are passionate about making a tangible difference in the fintech landscape, Setu offers the perfect platform to contribute to financial inclusion and improve millions of lives. Join us in our audacious mission and obsession with craftsmanship in code as we work together to build infrastructure that directly impacts the lives of individuals across India.,

Posted 3 days ago

Apply

2.0 - 5.0 years

0 Lacs

India

Remote

Location: Remote, Preferably Bangalore with occasional travel for collaboration and client meetings Engagement Type: Contract (initial 3 months with potential for extension based on project needs and fitment) About Optron: At Optron (a venture of Blue Boy Consulting LLP), we are at the forefront of leveraging cutting-edge AI to transform how enterprises interact with and derive insights from their data. We believe in building intelligent, autonomous systems that drive unprecedented efficiency and innovation for our clients. Our culture is one of continuous learning, fearless exploration, and solving complex, real-world challenges with elegant, intelligent solutions. We are a lean, agile team passionate about pushing the boundaries of what's possible with AI. Our leadership team has extensive global top-tier strategy consulting experience, coupled with deep technical acumen. This unique blend means we don't just build technology; we build solutions that truly impact global businesses, and you'll have the freedom to shape the future direction of the company and its offerings. The Opportunity: Accelerate Enterprise Transformation with Data & Process Mining Are you a bright, driven data engineer with a passion for crafting robust data solutions and a knack for quickly mastering new technologies? Do you thrive in environments where your direct impact is tangible, and your innovative ideas can genuinely shape the future of enterprise data strategy? If so, we're looking for you! We're not just seeking a data engineer; we're seeking a highly intelligent, exceptionally quick-learning problem-solver eager to delve into the intricate world of enterprise processes. This role is pivotal in building accelerators and tools that will empower our consultants to deliver best-in-class process mining and intelligent process execution solutions for our global enterprise clients. You'll bridge the gap between raw process data and actionable insights by building robust data models that automate the discovery, analysis, and optimization of complex business processes. This is not about maintaining legacy systems; it's about pioneering the next generation of data interaction and automation through intelligent data models. We are looking for a smart, foundational developer who thrives on intellectual challenge, possesses an insatiable curiosity, and is eager to dive deep into sophisticated data environments. We are looking for raw talent, a sharp mind, and the ability to rapidly acquire and apply new knowledge. If you're a problem-solver at heart, passionate about data, and want to build solutions that redefine industry standards, this is your chance to make a significant impact. What You'll Be Doing (Key Responsibilities & Goals) As a Data Engineer, you'll drive the data backbone of our process intelligence initiatives, specifically: Architecting Process Mining Data Models: Designing, developing, and optimizing highly efficient data models to capture and prepare event data for process mining analysis. This involves deep engagement with complex datasets from critical enterprise IT systems like SAP ERP, SAP S/4HANA, Salesforce , and other bespoke client applications. Databricks & PySpark Development: Leveraging your experience (2-5 years preferred) with Databricks and PySpark (with occasional SQL Spark) to create scalable, robust, and efficient data ingestion and transformation pipelines. This includes working with core Databricks features such as Delta Lake, and optimizing data processing through techniques like Z-ordering and partitioning. End-to-End Data Pipeline Ownership: Implementing core data engineering concepts such as Change Data Capture (CDC) , to build real-time data ingestion and transformation pipelines from various sources Storage Management: Working with various data storage solutions like Azure Data Lake, Unity Catalogue, and Delta Lake for efficient data storage. Cloud & DevOps Setup: Taking ownership of setting up cloud environments, establishing robust CI/CD pipelines , and managing code repositories to ensure seamless, modular, and version-controlled development. This includes leveraging Git / Databricks Repos and Databricks Workflows for streamlined development and orchestration. Data Governance & Security: Implementing and maintaining data governance, privacy and security best practices within the Databricks environment to handle sensitive enterprise data. Synthetic Data Generation: Developing sophisticated synthetic training datasets that accurately emulate the complex data structures, event logs, and behaviours found within diverse enterprise IT systems, crucial for our analytical models. Staying Updated: Keeping up-to-date with the latest Databricks features, best practices, and industry trends to continuously enhance our solutions. What We're Looking For (Required & Preferred Qualifications) We prioritize a sharp mind and a strong foundation. While specific experience is valuable, your ability to learn and adapt quickly is paramount. Educational Background: A Bachelor of Engineering (B.E.) / Bachelor of Technology (B.Tech) in Computer Science, Information Technology, or a closely related engineering discipline is preferred. Core Data Engineering Acumen: Demonstrated understanding of fundamental data engineering principles, including data warehousing, ETL/ELT methodologies, data quality, and data governance. Databricks & Spark Exposure: 2-5 years of practical experience with Databricks , with a focus on building pipelines and data solutions using PySpark. Conceptual Depth: A clear grasp of concepts like CDC, data pipeline creation, efficient data ingestion, optimization strategies, efficient cloud cost management, and modular code development. Problem-Solving & Adaptability: A proven track record of tackling complex technical challenges with innovative solutions and a genuine eagerness to quickly master new tools and paradigms. Enterprise Data Context (Preferred): While not mandatory, prior exposure to or understanding of data structures and IT workloads within large enterprise environments (e.g., SAP, Salesforce) would be advantageous. Why Join Us? Join a team where your contributions are celebrated, and your growth is prioritized: Groundbreaking Work: Be at the forefront of data innovation, building solutions that don't just optimize, but fundamentally transform how enterprises operate. Intellectual Challenge: Work on complex, unsolved problems that will stretch your abilities and foster rapid personal and professional growth. Learning-Centric Environment & 20% Time: We deeply value continuous learning. You'll receive 20% dedicated time to explore new technologies, learn new skills, or pursue personal pet projects that spark your interest and contribute to your growth. Global Exposure: Gain invaluable experience working with diverse global clients and collaborating with colleagues from various backgrounds , expanding your professional network and worldview. High Impact & Shaping the Future: Your contributions will directly influence our clients' success and, critically, you'll have the freedom to shape the future direction of the company , contributing directly to product strategy, technical roadmap, and innovative service offerings, working closely with our visionary IIM alumni leadership. Autonomy & Trust: We trust our team members to take ownership, innovate, and deliver high-quality results. Collaborative & Supportive Team: Work alongside other bright, passionate individuals who are eager to learn and build together. Competitive Compensation: We offer attractive contractor rates commensurate with your skills and potential. Ready to Redefine Enterprise Intelligence with Data? If you're a brilliant problem-solver with a strong technical foundation and a burning desire to master the art of data engineering for enterprise transformation, we encourage you to apply. This is more than a contract; it's an opportunity to build something truly revolutionary. To Apply: Click on Easy Apply, and submit your latest resume. Ensure you have at least one key relevant project mentioned in detail on the resume.

Posted 3 days ago

Apply

2.0 - 6.0 years

0 Lacs

hyderabad, telangana

On-site

You are ready to gain the skills and experience required to progress within your role and advance your career, and there is an excellent software engineering opportunity waiting for you. As a Software Engineer II at JPMorgan Chase in the Corporate Technology organization, you play a crucial role in the Data Services Team dedicated to enhancing, building, and delivering trusted market-leading Generative AI products securely, stably, and at scale. Being a part of the software engineering team, you will implement software solutions by designing, developing, and troubleshooting multiple components within technical products, applications, or systems while continuously enhancing your skills and experience. Your responsibilities include executing standard software solutions, writing secure and high-quality code in at least one programming language, designing and troubleshooting with consideration of upstream and downstream systems, applying tools within the Software Development Life Cycle for automation, and employing technical troubleshooting to solve basic complexity technical problems. Additionally, you will analyze large datasets to identify issues and contribute to decision-making for secure and stable application development, learn and apply system processes for developing secure code and systems, and contribute to a team culture of diversity, equity, inclusion, and respect. The qualifications, capabilities, and skills required for this role include formal training or certification in software engineering concepts with a minimum of 2 years of applied experience, experience with large datasets and predictive models, developing and maintaining code in a corporate environment using modern programming languages and database querying languages, proficiency in programming languages like Python, TensorFlow, PyTorch, PySpark, numpy, pandas, SQL, and familiarity with cloud services such as AWS/Azure. You should have a strong ability to analyze and derive insights from data, experience across the Software Development Life Cycle, exposure to agile methodologies, and emerging knowledge of software applications and technical processes within a technical discipline. Preferred qualifications include understanding of SDLC cycles for data platforms, major upgrade releases, patches, bug/hot fixes, and associated documentations.,

Posted 3 days ago

Apply

7.0 - 11.0 years

0 Lacs

karnataka

On-site

You will be joining our dynamic team as a highly skilled Senior Data Engineer/DE Architect with 7-10 years of experience. Your expertise in data engineering technologies, particularly SQL, Databricks, Azure services, and client interaction will be crucial for this role. Your responsibilities will include: - Hands-on experience with SQL, Databricks, pyspark, Python, Azure Cloud, and Power BI. - Designing, developing, and optimizing pyspark workloads. - Writing scalable, modular, and reusable code in SQL, python, and pyspark. - Collaborating with client stakeholders and cross-functional teams. - Gathering and analyzing requirements, translating business needs into technical solutions. - Providing regular project updates and reports on progress. - Ensuring alignment of data solutions with business requirements. - Working in US shift hours to coordinate with global teams. We expect you to have: - 8-10 years of experience in data engineering or related fields. - Proficiency in SQL, Databricks, PySpark, Python, Azure Cloud, and Power BI. - Strong written and verbal communication skills. - Proven ability to collaborate effectively with global stakeholders. - Strong problem-solving skills and attention to detail. Apply now and be part of our innovative team!,

Posted 3 days ago

Apply

12.0 - 16.0 years

0 Lacs

karnataka

On-site

You have a deep experience in developing data processing tasks using PySpark/spark such as reading data from external sources, merging data, performing data enrichment, and loading into target data destinations. Your responsibilities will include developing, programming, and maintaining applications using the Apache Spark and python open-source framework. You will work with different aspects of the Spark ecosystem, including Spark SQL, DataFrames, Datasets, and streaming. As a Spark Developer, you must have strong programming skills in Python, Java, or Scala. It is essential that you are familiar with big data processing tools and techniques, have a good understanding of distributed systems, and possess proven experience as a Spark Developer or a related role. Your problem-solving and analytical thinking skills should be excellent. Experience with building APIs for provisioning data to downstream systems is required. Working experience on any Cloud technology like AWS, Azure, Google is an added advantage. Hands-on experience with AWS S3 Filesystem operations will be beneficial for this role.,

Posted 3 days ago

Apply

3.0 - 7.0 years

0 Lacs

karnataka

On-site

As a Data Engineer, you will be responsible for designing, developing, and delivering ADF pipelines for the Accounting & Reporting Stream. Your role will involve creating and maintaining scalable data pipelines using PySpark and ETL workflows in Azure Databricks and Azure Data Factory. You will also work on data modeling and architecture to optimize data structures for analytics and business requirements. Your responsibilities will include monitoring, tuning, and troubleshooting pipeline performance for efficiency and reliability. Collaboration with business analysts and stakeholders is key to understanding data needs and delivering actionable insights. Implementing data governance practices to ensure data quality, security, and compliance with regulations is essential. You will also be required to develop and maintain documentation for data pipelines and architecture. Experience in testing and test automation is necessary for this role. Collaboration with cross-functional teams to comprehend data requirements and provide technical advice is crucial. Strong background in data engineering is required, with proficiency in SQL, Azure Databricks, Blob Storage, Azure Data Factory, and programming languages like Python or Scala. Knowledge of Logic App and Key Vault is also necessary. Strong problem-solving skills and the ability to communicate complex technical concepts to non-technical stakeholders are essential for effective communication within the team.,

Posted 3 days ago

Apply

4.0 years

0 Lacs

Gurugram, Haryana, India

On-site

Line of Service Advisory Industry/Sector Not Applicable Specialism Data, Analytics & AI Management Level Senior Manager Job Description & Summary At PwC, our people in data and analytics focus on leveraging data to drive insights and make informed business decisions. They utilise advanced analytics techniques to help clients optimise their operations and achieve their strategic goals. In business intelligence at PwC, you will focus on leveraging data and analytics to provide strategic insights and drive informed decision-making for clients. You will develop and implement innovative solutions to optimise business performance and enhance competitive advantage. Why PWC At PwC, you will be part of a vibrant community of solvers that leads with trust and creates distinctive outcomes for our clients and communities. This purpose-led and values-driven work, powered by technology in an environment that drives innovation, will enable you to make a tangible impact in the real world. We reward your contributions, support your wellbeing, and offer inclusive benefits, flexibility programmes and mentorship that will help you thrive in work and life. Together, we grow, learn, care, collaborate, and create a future of infinite experiences for each other. Learn more about us. At PwC, we believe in providing equal employment opportunities, without any discrimination on the grounds of gender, ethnic background, age, disability, marital status, sexual orientation, pregnancy, gender identity or expression, religion or other beliefs, perceived differences and status protected by law. We strive to create an environment where each one of our people can bring their true selves and contribute to their personal growth and the firm’s growth. To enable this, we have zero tolerance for any discrimination and harassment based on the above considerations Responsibilities Exp : 4-6 years Good knowledge of Data Warehousing , Data Lakehouse & Data Modelling concepts Hands-on experience in Azure DataBricks & Pyspark Design and develop robust and scalable data pipelines using PySpark and Databricks. Implement ETL processes and metadata-driven frameworks to optimize data flow and quality. Should have experience in understanding the source to target mapping document and building optimized unit tested ETL pipeline Should have experience in Data profiling and ensure data quality and integrity throughout the data lifecycle. Hands-on experience in handling large data volumes and performance tuning Should have experience working in onsite/offshore model Should have good communication and documentation skills Mandatory Skill Sets Azure/ETL Preferred Skill Sets Azure/ETL Years Of Experience Required 4-6 years Education Qualification BE/BTech/MBA/MCA Education (if blank, degree and/or field of study not specified) Degrees/Field of Study required: Bachelor of Technology, Bachelor of Engineering, Master of Business Administration Degrees/Field Of Study Preferred Certifications (if blank, certifications not specified) Required Skills Extract Transform Load (ETL), Microsoft Azure Optional Skills Accepting Feedback, Accepting Feedback, Active Listening, Analytical Thinking, Business Case Development, Business Data Analytics, Business Intelligence and Reporting Tools (BIRT), Business Intelligence Development Studio, Coaching and Feedback, Communication, Competitive Advantage, Continuous Process Improvement, Creativity, Data Analysis and Interpretation, Data Architecture, Database Management System (DBMS), Data Collection, Data Pipeline, Data Quality, Data Science, Data Visualization, Embracing Change, Emotional Regulation, Empathy, Inclusion {+ 24 more} Desired Languages (If blank, desired languages not specified) Travel Requirements Not Specified Available for Work Visa Sponsorship? No Government Clearance Required? No Job Posting End Date

Posted 3 days ago

Apply

6.0 - 10.0 years

0 Lacs

karnataka

On-site

As an Azure Databricks Professional at YASH Technologies, you will be utilizing your 6-8 years of experience to work with cutting-edge technologies in Azure services and data bricks. Your role will involve a strong understanding of medallion architecture, as well as proficiency in Python and Pyspark. YASH Technologies is a leading technology integrator that focuses on helping clients reimagine operating models, enhance competitiveness, optimize costs, foster exceptional stakeholder experiences, and drive business transformation. Our team is comprised of bright individuals who are dedicated to making real positive changes in an increasingly virtual world. Working at YASH, you will have the opportunity to shape your career in an inclusive team environment. We believe in continuous learning and development, leveraging career-oriented skilling models and technology to empower our employees to grow and adapt at a rapid pace. Our Hyperlearning workplace is based on principles such as flexible work arrangements, free spirit, emotional positivity, agile self-determination, trust, transparency, open collaboration, support for realizing business goals, and a stable employment environment with a great atmosphere and ethical corporate culture.,

Posted 3 days ago

Apply

0 years

0 Lacs

Pune, Maharashtra, India

On-site

Join us as a Senior Automation QA at Barclays, where you will be responsible for supporting the successful delivery of location strategy projects to plan, budget, agreed quality and governance standards. You'll spearhead the evolution of our digital landscape, driving innovation and excellence. You will harness cutting-edge technology to revolutionise our digital offerings, ensuring unparalleled customer experiences. To be successful as a Senior Automation QA you should have experience with: Spark SQL Python/Pyspark scripting ETL concepts Some Other Highly Valued Skills May Include AWS exposure Jupyter Notebook You may be assessed on key critical skills relevant for success in role, such as risk and controls, change and transformation, business acumen, strategic thinking and digital and technology, as well as job-specific technical skills. This role is based out of Pune. Purpose of the role To design, develop, and execute testing strategies to validate functionality, performance, and user experience, while collaborating with cross-functional teams to identify and resolve defects, and continuously improve testing processes and methodologies, to ensure software quality and reliability. Accountabilities Development and implementation of comprehensive test plans and strategies to validate software functionality and ensure compliance with established quality standards. Creation and execution automated test scripts, leveraging testing frameworks and tools to facilitate early detection of defects and quality issues. . Collaboration with cross-functional teams to analyse requirements, participate in design discussions, and contribute to the development of acceptance criteria, ensuring a thorough understanding of the software being tested. Root cause analysis for identified defects, working closely with developers to provide detailed information and support defect resolution. Collaboration with peers, participate in code reviews, and promote a culture of code quality and knowledge sharing. Stay informed of industry technology trends and innovations, and actively contribute to the organization's technology communities to foster a culture of technical excellence and growth. Assistant Vice President Expectations To advise and influence decision making, contribute to policy development and take responsibility for operational effectiveness. Collaborate closely with other functions/ business divisions. Lead a team performing complex tasks, using well developed professional knowledge and skills to deliver on work that impacts the whole business function. Set objectives and coach employees in pursuit of those objectives, appraisal of performance relative to objectives and determination of reward outcomes If the position has leadership responsibilities, People Leaders are expected to demonstrate a clear set of leadership behaviours to create an environment for colleagues to thrive and deliver to a consistently excellent standard. The four LEAD behaviours are: L – Listen and be authentic, E – Energise and inspire, A – Align across the enterprise, D – Develop others. OR for an individual contributor, they will lead collaborative assignments and guide team members through structured assignments, identify the need for the inclusion of other areas of specialisation to complete assignments. They will identify new directions for assignments and/ or projects, identifying a combination of cross functional methodologies or practices to meet required outcomes. Consult on complex issues; providing advice to People Leaders to support the resolution of escalated issues. Identify ways to mitigate risk and developing new policies/procedures in support of the control and governance agenda. Take ownership for managing risk and strengthening controls in relation to the work done. Perform work that is closely related to that of other areas, which requires understanding of how areas coordinate and contribute to the achievement of the objectives of the organisation sub-function. Collaborate with other areas of work, for business aligned support areas to keep up to speed with business activity and the business strategy. Engage in complex analysis of data from multiple sources of information, internal and external sources such as procedures and practises (in other areas, teams, companies, etc).to solve problems creatively and effectively. Communicate complex information. 'Complex' information could include sensitive information or information that is difficult to communicate because of its content or its audience. Influence or convince stakeholders to achieve outcomes. All colleagues will be expected to demonstrate the Barclays Values of Respect, Integrity, Service, Excellence and Stewardship – our moral compass, helping us do what we believe is right. They will also be expected to demonstrate the Barclays Mindset – to Empower, Challenge and Drive – the operating manual for how we behave.

Posted 3 days ago

Apply

0 years

0 Lacs

Pune, Maharashtra, India

On-site

Join us as a ETL Data Engineer at Barclays where you will spearhead the evolution of our infrastructure and deployment pipelines, driving innovation and operational excellence. You will harness cutting-edge technology to build and manage robust, scalable and secure infrastructure, ensuring seamless delivery of our digital solutions. To be successful as a ETL Data Engineer , you should have experience with: Ab Initio Unix Shell Scripting Oracle PySpark Some Other Highly Valued Skills May Include Python Teradata Java Machine Learning You may be assessed on key critical skills relevant for success in role, such as risk and controls, change and transformation, business acumen, strategic thinking and digital and technology, as well as job-specific technical skills. The role is based out of Pune. Purpose of the role To build and maintain the systems that collect, store, process, and analyse data, such as data pipelines, data warehouses and data lakes to ensure that all data is accurate, accessible, and secure. Accountabilities Build and maintenance of data architectures pipelines that enable the transfer and processing of durable, complete and consistent data. Design and implementation of data warehoused and data lakes that manage the appropriate data volumes and velocity and adhere to the required security measures. Development of processing and analysis algorithms fit for the intended data complexity and volumes. Collaboration with data scientist to build and deploy machine learning models. Analyst Expectations To perform prescribed activities in a timely manner and to a high standard consistently driving continuous improvement. Requires in-depth technical knowledge and experience in their assigned area of expertise Thorough understanding of the underlying principles and concepts within the area of expertise They lead and supervise a team, guiding and supporting professional development, allocating work requirements and coordinating team resources. If the position has leadership responsibilities, People Leaders are expected to demonstrate a clear set of leadership behaviours to create an environment for colleagues to thrive and deliver to a consistently excellent standard. The four LEAD behaviours are: L – Listen and be authentic, E – Energise and inspire, A – Align across the enterprise, D – Develop others. OR for an individual contributor, they develop technical expertise in work area, acting as an advisor where appropriate. Will have an impact on the work of related teams within the area. Partner with other functions and business areas. Takes responsibility for end results of a team’s operational processing and activities. Escalate breaches of policies / procedure appropriately. Take responsibility for embedding new policies/ procedures adopted due to risk mitigation. Advise and influence decision making within own area of expertise. Take ownership for managing risk and strengthening controls in relation to the work you own or contribute to. Deliver your work and areas of responsibility in line with relevant rules, regulation and codes of conduct. Maintain and continually build an understanding of how own sub-function integrates with function, alongside knowledge of the organisations products, services and processes within the function. Demonstrate understanding of how areas coordinate and contribute to the achievement of the objectives of the organisation sub-function. Make evaluative judgements based on the analysis of factual information, paying attention to detail. Resolve problems by identifying and selecting solutions through the application of acquired technical experience and will be guided by precedents. Guide and persuade team members and communicate complex / sensitive information. Act as contact point for stakeholders outside of the immediate function, while building a network of contacts outside team and external to the organisation. All colleagues will be expected to demonstrate the Barclays Values of Respect, Integrity, Service, Excellence and Stewardship – our moral compass, helping us do what we believe is right. They will also be expected to demonstrate the Barclays Mindset – to Empower, Challenge and Drive – the operating manual for how we behave.

Posted 3 days ago

Apply

3.0 - 7.0 years

0 Lacs

bhubaneswar

On-site

As a Pyspark Developer_VIS, your primary responsibility will be to develop high-performance Pyspark applications for large-scale data processing. You will collaborate with data engineers and analysts to integrate data pipelines and design ETL processes using Pyspark. Optimizing existing data models and workflows to enhance overall performance is also a key aspect of your role. Additionally, you will need to analyze large datasets to derive actionable insights and ensure data quality and integrity throughout the data processing lifecycle. Utilizing SQL for querying databases and validating data is essential, along with working with cloud technologies to deploy and maintain data solutions. You will participate in code reviews, maintain version control, and document all processes, workflows, and system changes clearly. Providing support in resolving production issues and assisting stakeholders, as well as mentoring junior developers on best practices in data processing, are also part of your responsibilities. Staying updated on emerging technologies and industry trends, implementing data security measures, contributing to team meetings, and offering insights for project improvements are other expectations from this role. Qualifications required for this position include a Bachelor's degree in Computer Science, Engineering, or a related field, along with 3+ years of experience in Pyspark development and data engineering. Strong proficiency in SQL and relational databases, experience with ETL tools and data processing frameworks, familiarity with Python for data manipulation and analysis, and knowledge of big data technologies such as Apache Hadoop and Spark are necessary. Experience working with cloud platforms like AWS or Azure, understanding data warehousing concepts and strategies, excellent problem-solving and analytical skills, attention to detail, commitment to quality, ability to work independently and as part of a team, excellent communication and interpersonal skills, experience with version control systems like Git, managing multiple priorities in a fast-paced environment, willingness to learn and adapt to new technologies, strong organizational skills, and meeting deadlines are also essential for this role. In summary, the ideal candidate for the Pyspark Developer_VIS position should possess a diverse skill set including cloud technologies, big data, version control, data warehousing, Pyspark, ETL, Python, Azure, Apache Hadoop, data analysis, Apache Spark, SQL, AWS, and more. ,

Posted 3 days ago

Apply

2.0 - 6.0 years

0 Lacs

karnataka

On-site

As a Microsoft Azure Data Engineer based in the APAC region, your primary role will involve designing, developing, and maintaining data solutions for APAC Advisory products. Your responsibilities will include ensuring data quality, security, scalability, and performance of data products. Collaborating with product managers, developers, analysts, and stakeholders is essential to understand business requirements and translate them into data models and architectures. You will have the opportunity to work closely with Data Architects, Engineers, and Data Scientists within the organization to support the development and maintenance of various data products. Your expertise will contribute significantly to enhancing product effectiveness and user experiences. Leveraging the organization's data assets for decision-making, analytics, and operational efficiency will be a key focus of your role. Your main responsibilities will revolve around designing, implementing, and managing data pipelines and architectures on the Azure Synapse platform. Utilizing tools such as PySpark, Synapse pipelines, and API integrations, you will be instrumental in developing robust data solutions that align with our business needs. Implementing the Medallion Architecture framework, including Bronze, Silver, and Gold layers, will be crucial for efficient data processing and storage. Key Responsibilities: - Design and implement data pipelines using PySpark Notebooks. - Manage and optimize data storage and processing with the Medallion Architecture framework. - Develop and maintain ETL processes for data ingestion, transformation, and loading. - Integrate APIs to facilitate data exchange between systems. - Create and manage table views for data visualization. - Ensure data quality, consistency, and security across all layers. - Collaborate with stakeholders to understand data requirements and deliver solutions. - Monitor and troubleshoot data pipelines to ensure reliability and performance. - Stay updated with the latest Azure technologies and best practices. Qualifications: - Bachelor's degree in computer science, Information Technology, or a related field. - Proven experience as a Data Engineer focusing on Azure technologies. - Proficiency in PySpark notebooks and Synapse pipelines with a minimum of 2 years of demonstrable experience. - Experience with Medallion Architecture and data management through different layers. - Familiarity with API integration and master data management tools like Profisee. - Ability to create and manage table views for data visualization. - Strong problem-solving skills, attention to detail, and communication skills. - Ability to thrive in a fast-paced, dynamic environment. Preferred Qualifications: - Azure certifications, such as Azure Data Engineer Associate. - Experience with other Azure services like Azure Databricks, Azure Synapse Analytics, and Azure SQL Database. - Knowledge of data governance and security best practices. - Familiarity with additional master data management tools and techniques. INCO: Cushman & Wakefield,

Posted 3 days ago

Apply

5.0 - 9.0 years

0 Lacs

hyderabad, telangana

On-site

As a Back-End Developer at our company, you will be responsible for developing an AI-driven prescriptive remediation model for SuperZoom, CBRE's data quality platform. Your primary focus will be on analyzing invalid records flagged by data quality rules and providing suggestions for corrected values based on historical patterns. It is crucial that the model you develop learns from past corrections to continuously enhance its future recommendations. The ideal candidate for this role should possess a solid background in machine learning, natural language processing (NLP), data quality, and backend development. Your key responsibilities will include developing a prescriptive remediation model to analyze and suggest corrections for bad records, implementing a feedback loop for continuous learning, building APIs and backend workflows for seamless integration, designing a data pipeline for real-time processing of flagged records, optimizing model performance for large-scale datasets, and collaborating effectively with data governance teams, data scientists, and front-end developers. Additionally, you will be expected to ensure the security, scalability, and performance of the system in handling sensitive data. To excel in this role, you should have at least 5 years of backend development experience with a focus on AI/ML-driven solutions. Proficiency in Python, including skills in Pandas, PySpark, and NumPy, is essential. Experience with machine learning libraries like Scikit-Learn, TensorFlow, or Hugging Face Transformers, along with a solid understanding of data quality, fuzzy matching, and NLP techniques for text correction, will be advantageous. Strong SQL skills and familiarity with databases such as PostgreSQL, Snowflake, or MS SQL Server are required, as well as expertise in building RESTful APIs and integrating ML models into production systems. Your problem-solving and analytical abilities will also be put to the test in handling diverse data quality issues effectively. Nice-to-have skills for this role include experience with vector databases (e.g., Pinecone, Weaviate) for similarity search, familiarity with LLMs and fine-tuning for data correction tasks, experience with Apache Airflow for workflow automation, and knowledge of reinforcement learning to enhance remediation accuracy over time. Your success in this role will be measured by the accuracy and relevance of suggestions provided for data quality issues in flagged records, improved model performance through iterative learning, seamless integration of the remediation model into SuperZoom, and on-time delivery of backend features in collaboration with the data governance team.,

Posted 3 days ago

Apply

6.0 - 14.0 years

0 Lacs

kolkata, west bengal

On-site

You have the opportunity to join our team as a Data Engineer with expertise in PySpark. You will be based in Kolkata, working in a hybrid model with 3 days in the office. With a minimum of 6 to 14 years of experience, you will play a crucial role in building and deploying Bigdata applications using PySpark. Your responsibilities will include having a minimum of 6 years of experience in building and deploying Bigdata applications using PySpark. You should also have at least 2 years of experience with AWS Cloud, focusing on data integration with Spark and AWS Glue/EMR. A deep understanding of Spark architecture and distributed systems is essential, along with good exposure to Spark job optimizations. Your expertise in handling complex large-scale Big Data environments will be key in this role. You will be expected to design, develop, test, deploy, maintain, and enhance data integration pipelines. Mandatory skills for this role include over 4 years of experience in PySpark, as well as 2+ years of experience in AWS Glue/EMR. A strong grasp of SQL is necessary, along with excellent written and verbal communication skills, and effective time management. Nice-to-have skills include any cloud skills and ETL knowledge. This role offers an exciting opportunity for a skilled Data Engineer to contribute to cutting-edge projects and make a significant impact within our team.,

Posted 3 days ago

Apply

5.0 - 9.0 years

0 Lacs

karnataka

On-site

The purpose of this role is to provide solutions and bridge the gap between technology and business know-how to deliver any client solution. You will be responsible for bridging the gap between project and support teams through techno-functional expertise. For new business implementation projects, you will drive the end-to-end process from business requirement management to integration & configuration and production deployment. It will be your responsibility to check the feasibility of the new change requirements and provide optimal solutions to the client with clear timelines. You will provide techno-functional solution support for all new business implementations while building the entire system from scratch. Your role will also involve supporting the solutioning team from architectural design, coding, testing, and implementation. In this role, you must understand the functional design as well as technical design and architecture to be implemented on the ERP system. You will customize, extend, modify, localize, or integrate the existing product through coding, testing, and production. Implementing the business processes, requirements, and the underlying ERP technology to translate them into ERP solutions will also be part of your responsibilities. Writing code as per the developmental standards to decide upon the implementation methodology will be crucial. Providing product support and maintenance to clients for a specific ERP solution and resolving day-to-day queries/technical problems that may arise are also key aspects of this role. Additionally, you will be required to create and deploy automation tools/solutions to ensure process optimization and increase efficiency. Your role will involve bridging technical and functional requirements of the project and providing solutioning/advice to the client or internal teams accordingly. Supporting on-site managers with necessary details regarding any change and providing off-site support will also be expected. Skill upgradation and competency building are essential in this role, including clearing Wipro exams and internal certifications from time to time to upgrade skills. Attending trainings and seminars to enhance knowledge in functional/technical domains and writing papers, articles, case studies, and publishing them on the intranet are also part of the responsibilities. Stakeholder Interaction involves interacting with internal stakeholders such as Lead Consultants and Onsite Project Manager/Project Teams for reporting, updates, and off-site support as per client requirements. External stakeholder interaction includes clients for solutioning and support. Competencies required for this role include Systems Thinking, Leveraging Technology, and Functional/Technical Knowledge at varying competency levels ranging from Foundation to Master. Additionally, behavioral competencies like Formulation & Prioritization, Innovation, Managing Complexity, Client Centricity, Execution Excellence, and Passion for Results are crucial for success in this role. In terms of performance parameters, your contribution to customer projects will be measured based on quality, SLA, ETA, number of tickets resolved, problems solved, number of change requests implemented, zero customer escalation, and CSAT. Automation will be evaluated based on process optimization, reduction in process/steps, and reduction in the number of tickets raised. Skill upgradation will be measured by the number of trainings & certifications completed and the number of papers/articles written in a quarter.,

Posted 3 days ago

Apply

5.0 - 9.0 years

0 Lacs

hyderabad, telangana

On-site

As a Full Stack Developer, you will be responsible for developing and maintaining both front-end and back-end components of web applications. You will utilize the .NET framework and related technologies for server-side development while leveraging React.js to build interactive and responsive user interfaces on the client-side. Your role will involve building and maintaining RESTful APIs to facilitate communication between front-end and back-end systems, as well as implementing authentication, authorization, and data validation mechanisms within APIs. In terms of Database Management, you will design, implement, and manage databases using technologies such as SQL Server or Azure SQL Database. Your responsibilities will include ensuring efficient data storage, retrieval, and manipulation to support application functionality. You will also be involved in Data Pipeline Management, where you will design, implement, and manage data pipelines using technologies such as PySpark, Python, and SQL. Building and maintaining pipelines in Databricks will be part of your tasks. Cloud Services Integration will be a key aspect of your role, requiring you to utilize Azure services for hosting, scaling, and managing web applications. You will implement cloud-based solutions for storage, caching, and data processing, as well as configure and manage Azure resources such as virtual machines, databases, and application services. In terms of DevOps and Deployment, you will implement CI/CD pipelines for automated build, test, and deployment processes using Jenkins. It will be essential to ensure robust monitoring, logging, and error handling mechanisms are in place. Documentation and Collaboration are important aspects of this role, where you will document technical designs, implementation details, and operational procedures. Collaborating with product managers, designers, and other stakeholders to understand requirements and deliver high-quality solutions will be part of your responsibilities. Continuous Learning is encouraged in this role, requiring you to stay updated with the latest technologies, tools, and best practices in web development and cloud computing. You will continuously improve your skills and knowledge through self-learning, training, and participation in technical communities. Requirements for this role include a Bachelor's Degree or equivalent experience, along with 5+ years of software engineering experience in reliable and resilient Microservice development and deployment. Strong knowledge of RESTful API, React.js, Azure, Python, PySpark, Databricks, Typescript, Node.js, relational databases like SQL Server, and No-SQL data store such as Redis and ADLS is essential. Experience with Data Engineering, Jenkins, Artifactory, and Automation testing frameworks is desirable. Prior experience with Agile, CI/CD, Docker, Kubernetes, Kafka, Terraform, or similar technologies is also beneficial. A passion for learning and disseminating new knowledge is highly valued in this role.,

Posted 3 days ago

Apply

3.0 - 7.0 years

0 Lacs

hyderabad, telangana

On-site

As a PySpark Developer with expertise in AWS and SQL, your main responsibility will be to create and enhance data pipelines within a cloud setting. Your key duties will include developing ETL workflows using PySpark, constructing and overseeing data pipelines on AWS (including S3, Glue, EMR, Lambda), crafting and fine-tuning SQL queries for data manipulation and reporting, as well as ensuring data quality, performance, and dependability. You will also be expected to collaborate closely with data engineers, analysts, and architects. To excel in this role, you must possess a high level of proficiency in PySpark and SQL, along with hands-on experience with AWS cloud services. Strong problem-solving and debugging skills are crucial, and familiarity with data lake and data warehouse concepts will be an added advantage.,

Posted 3 days ago

Apply

2.0 - 6.0 years

0 Lacs

noida, uttar pradesh

On-site

We are looking for an experienced AI/ML Architect to spearhead the design, development, and deployment of cutting-edge AI and machine learning systems. As the ideal candidate, you should possess a strong technical background in Python and data science libraries, profound expertise in AI and ML algorithms, and hands-on experience in crafting scalable AI solutions. This role demands a blend of technical acumen, leadership skills, and innovative thinking to enhance our AI capabilities. Your responsibilities will include identifying, cleaning, and summarizing complex datasets from various sources, developing Python/PySpark scripts for data processing and transformation, and applying advanced machine learning techniques like Bayesian methods and deep learning algorithms. You will design and fine-tune machine learning models, build efficient data pipelines, and leverage distributed databases and frameworks for large-scale data processing. In addition, you will lead the design and architecture of AI systems, with a focus on Retrieval-Augmented Generation (RAG) techniques and large language models. Your qualifications should encompass 5-7 years of total experience with 2-3 years in AI/ML, proficiency in Python and data science libraries, hands-on experience with PySpark scripting and AWS services, strong knowledge of Bayesian methods and time series forecasting, and expertise in machine learning algorithms and deep learning frameworks. You should also have experience in structured, unstructured, and semi-structured data, advanced knowledge of distributed databases, and familiarity with RAG systems and large language models for AI outputs. Strong collaboration, leadership, and mentorship skills are essential. Preferred qualifications include experience with Spark MLlib, SciPy, StatsModels, SAS, and R, a proven track record in developing RAG systems, and the ability to innovate and apply the latest AI techniques to real-world business challenges. Join our team at TechAhead, a global digital transformation company known for AI-first product design thinking and bespoke development solutions. With over 14 years of experience and partnerships with Fortune 500 companies, we are committed to driving digital innovation and delivering excellence. At TechAhead, you will be part of a dynamic team that values continuous learning, growth, and crafting tailored solutions for our clients. Together, let's shape the future of digital innovation worldwide!,

Posted 3 days ago

Apply

5.0 - 9.0 years

0 Lacs

karnataka

On-site

As a Principal Analyst at Citi's Analytics and Information Management (AIM) team in Bangalore, India, you will play a crucial role in creating client-centric analytical solutions for various business challenges. With a focus on client obsession and stakeholder management, you will be responsible for owning and delivering complex analytical projects. Your expertise in business context understanding, data analysis, and project management will be essential in identifying trends, patterns, and presenting high-quality solutions to senior management. Your primary responsibilities will include developing business critical dashboards, assessing and optimizing marketing programs, sizing the impact of strategic changes, and streamlining existing processes. By leveraging your skills in SQL, Python, Pyspark, Hive, and Impala, you will work with large datasets to extract insights that drive revenue growth and business decisions. Additionally, your experience in Investment Analytics, Retail Analytics, Credit Cards, and Financial Services will be valuable in delivering actionable intelligence to business leaders. To excel in this role, you should possess a master's or bachelor's degree in Engineering, Technology, or Computer Science from premier institutes, along with 5-6 years of experience in delivering analytical solutions. Your ability to articulate and solve complex business problems, along with excellent communication and interpersonal skills, will be key in collaborating with cross-functional teams and stakeholders. Moreover, your hands-on experience in Tableau and project management skills will enable you to mentor and guide junior team members effectively. If you are passionate about data, eager to tackle new challenges, and thrive in a dynamic work environment, this position offers you the opportunity to contribute to Citi's mission of enabling growth and economic progress through innovative analytics solutions. Join us in driving business success and making a positive impact on the financial services industry. Citi is an equal opportunity and affirmative action employer, offering full-time employment in the field of Investment Analytics, Retail Analytics, Credit Cards, and Financial Services. If you are ready to take your analytics career to the next level, we invite you to apply and be part of our global community at Citi.,

Posted 3 days ago

Apply

5.0 - 9.0 years

0 Lacs

thiruvananthapuram, kerala

On-site

As a Senior Machine Learning Engineer Contractor specializing in AWS ML Pipelines, your primary responsibility will be to design, develop, and deploy advanced ML pipelines within an AWS environment. You will work on cutting-edge solutions that automate entity matching for master data management, implement fraud detection systems, handle transaction matching, and integrate GenAI capabilities. The ideal candidate for this role should possess extensive hands-on experience in AWS services like SageMaker, Bedrock, Lambda, Step Functions, and S3. Moreover, you should have a strong command over CI/CD practices to ensure a robust and scalable solution. Your key responsibilities will include designing and developing end-to-end ML pipelines focusing on entity matching, fraud detection, and transaction matching. You will be integrating generative AI solutions using AWS Bedrock to enhance data processing and decision-making. Collaboration with cross-functional teams to refine business requirements and develop data-driven solutions tailored to master data management needs will also be a crucial aspect of your role. In terms of AWS ecosystem expertise, you will be required to utilize SageMaker for model training, deployment, and continuous improvement. Additionally, leveraging Lambda and Step Functions to orchestrate serverless workflows for data ingestion, preprocessing, and real-time processing will be part of your daily tasks. Managing data storage, retrieval, and scalability concerns using AWS S3 will also be within your purview. Furthermore, you will need to develop and integrate automated CI/CD pipelines to streamline model testing, deployment, and version control. Ensuring rapid iteration and robust deployment practices to maintain high availability and performance of ML solutions will be essential. Data security and compliance will be a critical aspect of your role. You will need to implement security best practices to safeguard sensitive data, ensuring compliance with organizational and regulatory requirements. Incorporating monitoring and alerting mechanisms to maintain the integrity and performance of deployed ML models will be part of your responsibilities. Collaboration and documentation will also play a significant role in your day-to-day activities. Working closely with business stakeholders, data engineers, and data scientists to ensure solutions align with evolving business needs will be crucial. You will also need to document all technical designs, workflows, and deployment processes to support ongoing maintenance and future enhancements. Providing regular progress updates and adapting to changing priorities or business requirements in a dynamic environment are expected. To qualify for this role, you should have at least 5+ years of professional experience in developing and deploying ML models and pipelines. Proven expertise in AWS services including SageMaker, Bedrock, Lambda, Step Functions, and S3 is necessary. Strong proficiency in Python and/or PySpark, demonstrated experience with CI/CD tools and methodologies, and practical experience in building solutions for entity matching, fraud detection, and transaction matching within a master data management context are also required. Familiarity with generative AI models and their application within data processing workflows will be an added advantage. Strong analytical and problem-solving skills are essential for this role. You should be able to transform complex business requirements into scalable technical solutions and possess strong data analysis capabilities with a track record of developing models that provide actionable insights. Excellent verbal and written communication skills, the ability to work independently as a contractor while effectively collaborating with remote teams, and a proven record of quickly adapting to new technologies and agile work environments are also preferred qualities for this position. A Bachelor's or Master's degree in Computer Science, Data Science, Engineering, or a related field is a plus. Experience with additional AWS services such as Kinesis, Firehose, and SQS, prior experience in a consulting or contracting role demonstrating the ability to manage deliverables under tight deadlines, and experience within industries where data security and compliance are critical will be advantageous.,

Posted 3 days ago

Apply

0.0 - 4.0 years

0 Lacs

karnataka

On-site

We are looking for a Data Engineer to join our data team. You will be responsible for managing our master data set, developing reports, and troubleshooting data issues. To excel in this role, attention to detail, experience as a data analyst, and a deep understanding of popular data analysis tools and databases are essential. Your responsibilities include: - Building, maintaining, and managing data pipelines for efficient data flow between systems. - Collaborating with stakeholders to design and manage customized data pipelines. - Testing various ETL (Extract, Transform, Load) tools for data ingestion and processing. - Assisting in scaling the data infrastructure to meet the organization's growing data demands. - Monitoring data pipeline performance and troubleshooting data issues. - Documenting pipeline architectures and workflows for future reference and scaling. - Evaluating data formats, sources, and transformation techniques. - Working closely with data scientists to ensure data availability and reliability for analytics. We require the following skill sets/experience: - Proficiency in Python, PySpark, and Big Data concepts such as Data Lakes and Data Warehouses. - Strong background in SQL. - Familiarity with cloud computing platforms like AWS, Azure, or Google Cloud. - Basic knowledge of containerization technologies like Docker. - Exposure to data orchestration tools like Apache Airflow or Luigi. Pedigree: - Bachelor's degree in Computer Science, Electrical Engineering, or IT.,

Posted 3 days ago

Apply

9.0 years

0 Lacs

Gurugram, Haryana, India

Remote

Job Description This is a remote position. Job Summary We are looking for an experienced Senior Data Engineer to lead the development of scalable AWS-native data lake pipelines with a strong focus on time series forecasting and upsert-ready architectures. This role requires end-to-end ownership of the data lifecycle, from ingestion to partitioning, versioning, and BI delivery. The ideal candidate must be highly proficient in AWS data services, PySpark, versioned storage formats like Apache Hudi/Iceberg, and must understand the nuances of data quality and observability in large-scale analytics systems. Responsibilities Design and implement data lake zoning (Raw → Clean → Modeled) using Amazon S3, AWS Glue, and Athena. Ingest structured and unstructured datasets including POS, USDA, Circana, and internal sales data. Build versioned and upsert-friendly ETL pipelines using Apache Hudi or Iceberg. Create forecast-ready datasets with lagged, rolling, and trend features for revenue and occupancy modeling. Optimize Athena datasets with partitioning, CTAS queries, and metadata tagging. Implement S3 lifecycle policies, intelligent file partitioning, and audit logging. Build reusable transformation logic using dbt-core or PySpark to support KPIs and time series outputs. Integrate robust data quality checks using custom logs, AWS CloudWatch, or other DQ tooling. Design and manage a forecast feature registry with metrics versioning and traceability. Collaborate with BI and business teams to finalize schema design and deliverables for dashboard consumption. Requirements Essential Skills: Job Deep hands-on experience with AWS Glue, Athena, S3, Step Functions, and Glue Data Catalog. Strong command over PySpark, dbt-core, CTAS query optimization, and partition strategies. Working knowledge of Apache Hudi, Iceberg, or Delta Lake for versioned ingestion. Experience in S3 metadata tagging and scalable data lake design patterns. Expertise in feature engineering and forecasting dataset preparation (lags, trends, windows). Proficiency in Git-based workflows (Bitbucket), CI/CD, and deployment automation. Strong understanding of time series KPIs, such as revenue forecasts, occupancy trends, or demand volatility. Data observability best practices including field-level logging, anomaly alerts, and classification tagging. Personal Independent, critical thinker with the ability to design for scale and evolving business logic. Strong communication and collaboration with BI, QA, and business stakeholders. High attention to detail in ensuring data accuracy, quality, and documentation. Comfortable interpreting business-level KPIs and transforming them into technical pipelines. Preferred Skills Job Experience with statistical forecasting frameworks such as Prophet, GluonTS, or related libraries. Familiarity with Superset or Streamlit for QA visualization and UAT reporting. Understanding of macroeconomic datasets (USDA, Circana) and third-party data ingestion. Personal Proactive, ownership-driven mindset with a collaborative approach. Strong communication and collaboration skills. Strong problem-solving skills with attention to detail. Have the ability to work under stringent deadlines and demanding client conditions. Strong analytical and problem-solving skills. Ability to work in fast-paced, delivery-focused environments. Strong mentoring and documentation skills for scaling the platform. Other Relevant Information Bachelor’s degree in Computer Science, Information Technology, or a related field. Minimum 9+ years of experience in data engineering & architecture. Benefits This role offers the flexibility of working remotely in India. LeewayHertz is an equal opportunity employer and does not discriminate based on race, color, religion, sex, age, disability, national origin, sexual orientation, gender identity, or any other protected status. We encourage a diverse range of applicants. check(event) ; career-website-detail-template-2 => apply(record.id,meta)" mousedown="lyte-button => check(event)" final-style="background-color:#6875E2;border-color:#6875E2;color:white;" final-class="lyte-button lyteBackgroundColorBtn lyteSuccess" lyte-rendered="">

Posted 3 days ago

Apply

10.0 years

0 Lacs

Kolkata, West Bengal, India

Remote

JOB_POSTING-3-72996-2 Job Description Role Title: AVP, Cloud Solution Architect (L11) Company Overview COMPANY OVERVIEW: Synchrony (NYSE: SYF) is a premier consumer financial services company delivering one of the industry’s most complete digitally enabled product suites. Our experience, expertise and scale encompass a broad spectrum of industries including digital, health and wellness, retail, telecommunications, home, auto, outdoors, pet and more. We have recently been ranked #5 among India’s Best Companies to Work for 2023, #21 under LinkedIn Top Companies in India list, and received Top 25 BFSI recognition from Great Place To Work India. We have been ranked Top 5 among India’s Best Workplaces in Diversity, Equity, and Inclusion, and Top 10 among India’s Best Workplaces for Women in 2022. We offer 100% Work from Home flexibility for all our Functional employees and provide some of the best-in-class Employee Benefits and Programs catering to work-life balance and overall well-being. In addition to this, we also have Regional Engagement Hubs across India and a co-working space in Bangalore. Organizational Overview Organizational Overview: This role will be part of the Data Architecture & Analytics group part of CTO organization. Data team is responsible for designing and developing scalable data pipelines for efficient data ingestion, transformation, and loading(ETL). Collaborating with cross-functional teams to integrate new data sources and ensure data quality and consistency. Building and maintaining data models to facilitate data access and analysis by Data Scientists and Analysts. Responsible for the SYF public cloud platform & services. Govern health, performance, capacity, and costs of resources and ensure adherence to service levels Build well defined processes for cloud application development and service enablement. Role Summary/Purpose The Cloud Solution Architect – will play a key role in modernizing SAS workloads by leading vendor refactoring efforts, break-fix execution, and user enablement strategies. This position requires a deep understanding of SAS, AWS analytics services (EMR Studio, S3, Redshift, Glue), and Tableau, combined with strong user engagement, training development, and change management skills. The role involves collaborating with vendors, business users, and cloud engineering teams to refactor legacy SAS code, ensure seamless execution of fixes, and develop comprehensive training materials and user job aids. Additionally, the Cloud Solution Architect will oversee user testing, validation, and sign-offs, ensuring a smooth transition to modern cloud-based solutions while enhancing adoption and minimizing disruption. This is an exciting opportunity to lead cloud migration initiatives, enhance analytics capabilities, and drive user transformation efforts within a cutting-edge cloud environment. Key Responsibilities Lead refactoring efforts to modernize and migrate SAS-based workloads to cloud-native or alternative solutions. Oversee break/fix execution by ensuring timely resolution of system issues and performance optimizations. Engage with end-users to gather requirements, address pain points, and ensure smooth adoption of cloud solutions. Develop and deliver custom training programs, including user job aids and self-service documentation. Facilitate user sign-offs and testing by coordinating validation processes and ensuring successful implementation. Drive user communication efforts related to system changes, updates, and migration timelines. Work closely with AWS teams to optimize EMR Studio, Redshift, Glue, and other AWS services for analytics and reporting. Ensure seamless integration with Tableau and other visualization tools to support business reporting needs. Implement best practices for user change management, minimizing disruption and improving adoption. Required Skills/Knowledge Bachelor’s Degree in Computer Science, Software Engineering, or a related field. Advanced degrees (Master’s or Ph.D.) can be a plus but are not always necessary if experience is significant. Experience in scripting languages (Python, SQL, or PySpark) for data transformations. Proven expertise in SAS, including experience with SAS code refactoring and optimization. Strong AWS experience, particularly with EMR Studio, S3, Redshift, Glue, and Lambda. Experience in user change management, training development, and communication strategies. Desired Skills/Knowledge Experience with AWS cloud services. Certifications in AWS or any other cloud platform. Experience with Agile project management methods and practices. Proficiency in Tableau for analytics and visualization. Hands-on experience with cloud migration projects, particularly SAS workloads. Excellent communication and stakeholder engagement skills. Familiarity with other cloud platforms like Azure or GCP is a plus. Eligibility Criteria 10+ years of experience in data analytics, cloud solutions, or enterprise architecture, with a focus on SAS migration and AWS cloud adoption. or in lieu of a degree 12+ years of experience Work Timings: 3 PM to 12 AM IST (WORK TIMINGS: This role qualifies for Enhanced Flexibility and Choice offered in Synchrony India and will require the incumbent to be available between 06:00 AM Eastern Time – 11:30 AM Eastern Time (timings are anchored to US Eastern hours and will adjust twice a year locally). This window is for meetings with India and US teams. The remaining hours will be flexible for the employee to choose. Exceptions may apply periodically due to business needs. Please discuss this with the hiring manager for more details .) For Internal Applicants Understand the criteria or mandatory skills required for the role, before applying Inform your manager and HRM before applying for any role on Workday Ensure that your professional profile is updated (fields such as education, prior experience, other skills) and it is mandatory to upload your updated resume (Word or PDF format) Must not be any corrective action plan (First Formal/Final Formal, PIP) L9+ Employees who have completed 18 months in the organization and 12 months in current role and level are only eligible. L9 + Employees can apply. Level / Grade : 11 Job Family Group Information Technology

Posted 3 days ago

Apply

4.0 - 11.0 years

0 Lacs

Gurugram, Haryana, India

On-site

Hello, Greeting from Quess Corp!! Hope you are doing well we have job opportunity with one of our client Designation_ Data Engineer Location – Gurugram Experience – 4yrs to 11 Yrs Qualification – Graduate / PG ( IT) Skill Set – Data Engineer, Python, AWS, SQL Essential capabilities Enthusiasm for technology, keeping up with latest trends Ability to articulate complex technical issues and desired outcomes of system enhancements Proven analytical skills and evidence-based decision making Excellent problem solving, troubleshooting & documentation skills Strong written and verbal communication skills Excellent collaboration and interpersonal skills Strong delivery focus with an active approach to quality and auditability Ability to work under pressure and excel within a fast-paced environment Ability to self-manage tasks Agile software development practices Desired Experience Hands on in SQL and its Big Data variants (Hive-QL, Snowflake ANSI, Redshift SQL) Python and Spark and one or more of its API (PySpark, Spark SQL, Scala), Bash/Shell scripting Experience with Source code control - GitHub, VSTS etc. Knowledge and exposure to Big Data technologies Hadoop stack such as HDFS, Hive, Impala, Spark etc, and cloud Big Data warehouses - RedShift, Snowflake etc. Experience with UNIX command-line tools. Exposure to AWS technologies including EMR, Glue, Athena, Data Pipeline, Lambda, etc Understanding and ability to translate/physicalise Data Models (Star Schema, Data Vault 2.0 etc) Essential Experience It is expected that the role holder will most likely have the following qualifications and experience 4-11 years technical experience (within financial services industry preferred) Technical Domain experience (Subject Matter Expertise in Technology or Tools) Solid experience, knowledge and skills in Data Engineering, BI/software development such as ELT/ETL, data extraction and manipulation in Data Lake/Data Warehouse/Lake House environment. Hands on programming experience in writing Python, SQL, Unix Shell scripts, Pyspark scripts, in a complex enterprise environment Experience in configuration management using Ansible/Jenkins/GIT Hands on cloud-based solution design, configuration and development experience with Azure and AWS Hands on experience of using AWS Services - S3,EC2, EMR, SNS, SQS, Lambda functions, Redshift Hands on experience Of building Data pipelines to ingest, transform on Databricks Delta Lake platform from a range of data sources - Data bases, Flat files, Streaming etc.. Knowledge of Data Modelling techniques and practices used for a Data Warehouse/Data Mart application. Quality engineering development experience (CI/CD – Jenkins, Docker) Experience in Terraform, Kubernetes and Docker Experience with Source Control Tools – Github or BitBucket Exposure to relational Databases - Oracle or MS SQL or DB2 (SQL/PLSQL, Database design, Normalisation, Execution plan analysis, Index creation and maintenance, Stored Procedures) , PostGres/MySQL Skilled in querying data from a range of data sources that store structured and unstructured data Knowledge or understanding of Power BI (Recommended) Key Accountabilities Design, develop, test, deploy, maintain and improve software Develop flowcharts, layouts and documentation to identify requirements & solutions Write well designed & high-quality testable code Produce specifications and determine operational feasibility Integrate software components into fully functional platform Apply pro-actively & perform hands-on design and implementation of best practice CI/CD Coaching & mentoring of other Service Team members Develop/contribute to software verification plans and quality assurance procedures Document and maintain software functionality Troubleshoot, debug and upgrade existing systems, including participating in DR tests Deploy programs and evaluate customer feedback Contribute to team estimation for delivery and expectation management for scope. Comply with industry standards and regulatory requirements

Posted 3 days ago

Apply
cta

Start Your Job Search Today

Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.

Job Application AI Bot

Job Application AI Bot

Apply to 20+ Portals in one click

Download Now

Download the Mobile App

Instantly access job listings, apply easily, and track applications.

Featured Companies