Jobs
Interviews

5284 Informatica Jobs

Setup a job Alert
JobPe aggregates results for easy application access, but you actually apply on the job portal directly.

10.0 years

0 Lacs

Gurugram, Haryana, India

On-site

Acuity Knowledge Partners (Acuity) is a leading provider of bespoke research, analytics and technology solutions to the financial services sector, including asset managers, corporate and investment banks, private equity and venture capital firms, hedge funds and consulting firms. Its global network of over 6,000 analysts and industry experts, combined with proprietary technology, supports more than 600 financial institutions and consulting companies to operate more efficiently and unlock their human capital, driving revenue higher and transforming operations. Acuity is headquartered in London and operates from 10 locations worldwide. The company fosters a diverse, equitable and inclusive work environment, nurturing talent, regardless of race, gender, ethnicity or sexual orientation. Acuity was established as a separate business from Moody’s Corporation in 2019, following its acquisition by Equistone Partners Europe (Equistone). In January 2023, funds advised by global private equity firm Permira acquired a majority stake in the business from Equistone, which remains invested as a minority shareholder. For more information, visit www.acuitykp.com Position Title- Associate Director (Senior Architect – Data) Department-IT Location- Gurgaon/ Bangalore Job Summary The Enterprise Data Architect will enhance the company's strategic use of data by designing, developing, and implementing data models for enterprise applications and systems at conceptual, logical, business area, and application layers. This role advocates data modeling methodologies and best practices. We seek a skilled Data Architect with deep knowledge of data architecture principles, extensive data modeling experience, and the ability to create scalable data solutions. Responsibilities include developing and maintaining enterprise data architecture, ensuring data integrity, interoperability, security, and availability, with a focus on ongoing digital transformation projects. Key Responsibilities Strategy & Planning Develop and deliver long-term strategic goals for data architecture vision and standards in conjunction with data users, department managers, clients, and other key stakeholders. Create short-term tactical solutions to achieve long-term objectives and an overall data management roadmap. Establish processes for governing the identification, collection, and use of corporate metadata; take steps to assure metadata accuracy and validity. Establish methods and procedures for tracking data quality, completeness, redundancy, and improvement. Conduct data capacity planning, life cycle, duration, usage requirements, feasibility studies, and other tasks. Create strategies and plans for data security, backup, disaster recovery, business continuity, and archiving. Ensure that data strategies and architectures are aligned with regulatory compliance. Develop a comprehensive data strategy in collaboration with different stakeholders that aligns with the transformational projects’ goals. Ensure effective data management throughout the project lifecycle. Acquisition & Deployment Ensure the success of enterprise-level application rollouts (e.g. ERP, CRM, HCM, FP&A, etc.) Liaise with vendors and service providers to select the products or services that best meet company goals Operational Management o Assess and determine governance, stewardship, and frameworks for managing data across the organization. o Develop and promote data management methodologies and standards. o Document information products from business processes and create data entities o Create entity relationship diagrams to show the digital thread across the value streams and enterprise o Create data normalization across all systems and data base to ensure there is common definition of data entities across the enterprise o Document enterprise reporting needs develop the data strategy to enable single source of truth for all reporting data o Address the regulatory compliance requirements of each country and ensure our data is secure and compliant o Select and implement the appropriate tools, software, applications, and systems to support data technology goals. o Oversee the mapping of data sources, data movement, interfaces, and analytics, with the goal of ensuring data quality. o Collaborate with project managers and business unit leaders for all projects involving enterprise data. o Address data-related problems regarding systems integration, compatibility, and multiple-platform integration. o Act as a leader and advocate of data management, including coaching, training, and career development to staff. o Develop and implement key components as needed to create testing criteria to guarantee the fidelity and performance of data architecture. o Document the data architecture and environment to maintain a current and accurate view of the larger data picture. o Identify and develop opportunities for data reuse, migration, or retirement. Data Architecture Design: Develop and maintain the enterprise data architecture, including data models, databases, data warehouses, and data lakes. Design and implement scalable, high-performance data solutions that meet business requirements. Data Governance: Establish and enforce data governance policies and procedures as agreed with stakeholders. Maintain data integrity, quality, and security within Finance, HR and other such enterprise systems. Data Migration: Oversee the data migration process from legacy systems to the new systems being put in place. Define & Manage data mappings, cleansing, transformation, and validation to ensure accuracy and completeness. Master Data Management: Devise processes to manage master data (e.g., customer, vendor, product information) to ensure consistency and accuracy across enterprise systems and business processes. Provide data management (create, update and delimit) methods to ensure master data is governed Stakeholder Collaboration: Collaborate with various stakeholders, including business users, other system vendors, and stakeholders to understand data requirements. Ensure the enterprise system meets the organization's data needs. Training and Support: Provide training and support to end-users on data entry, retrieval, and reporting within the candidate enterprise systems. Promote user adoption and proper use of data. 10 Data Quality Assurance: Implement data quality assurance measures to identify and correct data issues. Ensure the Oracle Fusion and other enterprise systems contain reliable and up-to-date information. Reporting and Analytics: Facilitate the development of reporting and analytics capabilities within the Oracle Fusion and other systems Enable data-driven decision-making through robust data analysis. Continuous Improvement: Continuously monitor and improve data processes and the Oracle Fusion and other system's data capabilities. Leverage new technologies for enhanced data management to support evolving business needs. Technology and Tools: Oracle Fusion Cloud Data modeling tools (e.g., ER/Studio, ERwin) ETL tools (e.g., Informatica, Talend, Azure Data Factory) Data Pipelines: Understanding of data pipeline tools like Apache Airflow and AWS Glue. Database management systems: Oracle Database, MySQL, SQL Server, PostgreSQL, MongoDB, Cassandra, Couchbase, Redis, Hadoop, Apache Spark, Amazon RDS, Google BigQuery, Microsoft Azure SQL Database, Neo4j, OrientDB, Memcached) Data governance tools (e.g., Collibra, Informatica Axon, Oracle EDM, Oracle MDM) Reporting and analytics tools (e.g., Oracle Analytics Cloud, Power BI, Tableau, Oracle BIP) Hyperscalers / Cloud platforms (e.g., AWS, Azure) Big Data Technologies such as Hadoop, HDFS, MapReduce, and Spark Cloud Platforms such as Amazon Web Services, including RDS, Redshift, and S3, Microsoft Azure services like Azure SQL Database and Cosmos DB and experience in Google Cloud Platform services such as BigQuery and Cloud Storage. Programming Languages: (e.g. using Java, J2EE, EJB, .NET, WebSphere, etc.) SQL: Strong SQL skills for querying and managing databases. Python: Proficiency in Python for data manipulation and analysis. Java: Knowledge of Java for building data-driven applications. Data Security and Protocols: Understanding of data security protocols and compliance standards. Key Competencies Qualifications: Education: Bachelor’s degree in computer science, Information Technology, or a related field. Master’s degree preferred. Experience: 10+ years overall and at least 7 years of experience in data architecture, data modeling, and database design. Proven experience with data warehousing, data lakes, and big data technologies. Expertise in SQL and experience with NoSQL databases. Experience with cloud platforms (e.g., AWS, Azure) and related data services. Experience with Oracle Fusion or similar ERP systems is highly desirable. Skills: Strong understanding of data governance and data security best practices. Excellent problem-solving and analytical skills. Strong communication and interpersonal skills. Ability to work effectively in a collaborative team environment. Leadership experience with a track record of mentoring and developing team members. Excellent in documentation and presentations. Good knowledge of applicable data privacy practices and laws. Certifications: Relevant certifications (e.g., Certified Data Management Professional, AWS Certified Big Data – Specialty) are a plus. Behavioral A self-starter, an excellent planner and executor and above all, a good team player Excellent communication skills and inter-personal skills are a must Must possess organizational skills, including multi-task capability, priority setting and meeting deadlines Ability to build collaborative relationships and effectively leverage networks to mobilize resources Initiative to learn business domain is highly desirable Likes dynamic and constantly evolving environment and requirements

Posted 9 hours ago

Apply

3.0 years

4 - 10 Lacs

Gurgaon

On-site

Technical Support: Provide L1, L2 and L3 support for software applications. Troubleshoot and resolve application-related issues for end-users. Collaborate with developers, IT teams, and external vendors to address issues and implement fixes. Escalate unresolved issues to higher-level support or specialized teams. Monitoring and Maintenance: Monitor scheduled Jobs and ensure their successful completion. Perform routine maintenance tasks, including system updates, backups, and configuration changes. Assist with system upgrades, patches, and migrations to ensure continuity of service. Incident Management: Log, track, and manage incidents and service requests via ticketing systems. Follow established procedures for incident escalation and resolution. Participate in root cause analysis and problem management efforts. Documentation and Reporting: Maintain and update application documentation, including configuration settings and user guides. Create and present reports on system performance, issues, and resolution times. Document changes, fixes, and enhancements in a knowledge base for future reference. Collaboration: Work with cross-functional teams (IT, development, business users) to gather requirements and improve applications. Participate in system testing and quality assurance activities. Assist in the development and implementation of new application modules or features. Key Skills and Qualifications: Technical Skills: 3+ Years of experience in Informatica B2B (DX, DT) 3+ Years of experience in Informatica Power Center and IICS 3+ Years of experience with databases (MS-SQL server), Experience with application monitoring tools. Experience in Informatica Cloud Data Governance Catalog is preferred Analytical Skills: Ability to diagnose and solve complex technical problems. Communication: Strong verbal and written communication skills, able to explain technical concepts to non-technical users. Customer Service: Ability to provide excellent customer service under pressure and manage competing priorities. Knowledge of ITIL processes (incident, problem, and change management).

Posted 9 hours ago

Apply

3.0 - 5.0 years

0 Lacs

Hyderabad, Telangana, India

On-site

Job Description: About Us At Bank of America, we are guided by a common purpose to help make financial lives better through the power of every connection. Responsible Growth is how we run our company and how we deliver for our clients, teammates, communities and shareholders every day. One of the keys to driving Responsible Growth is being a great place to work for our teammates around the world. We’re devoted to being a diverse and inclusive workplace for everyone. We hire individuals with a broad range of backgrounds and experiences and invest heavily in our teammates and their families by offering competitive benefits to support their physical, emotional, and financial well-being. Bank of America believes both in the importance of working together and offering flexibility to our employees. We use a multi-faceted approach for flexibility, depending on the various roles in our organization. Working at Bank of America will give you a great career with opportunities to learn, grow and make an impact, along with the power to make a difference. Join us! Global Business Services Global Business Services delivers Technology and Operations capabilities to Lines of Business and Staff Support Functions of Bank of America through a centrally managed, globally integrated delivery model and globally resilient operations. Global Business Services is recognized for flawless execution, sound risk management, operational resiliency, operational excellence and innovation. In India, we are present in five locations and operate as BA Continuum India Private Limited (BACI), a non-banking subsidiary of Bank of America Corporation and the operating company for India operations of Global Business Services. Job Description: This job is responsible for performing commoditized activities which may include monitoring, managing events, servicing requests, and engineering. Key responsibilities may include assisting with network, systems, applications, access requests, and production support or security engineering. Responsibilities Incident tracking and resolutions Triage incidents, take complete ownership & accountability to ensure restoral of service/issue/failure Analyze batch flow, identify issues, fix the same and ensure proper batch completion. Drive closure of incidents & problems meeting defined SLA. Contribute in building the knowledge base for the team where ever required. Proactive root cause analysis and problem management skills. Provide regular communication to all the stakeholders. Flexible to support on weekends/holidays as and when required on rotation. Strive towards process improvement and automation. Education……B.E/B.Tech/B.Sc/B.C.A/M.C.A Certifications If Any Experience Range - 3 to 5 Required Skills Strong Production support back ground with exposure to Level two production support tasks. Experience on working with Oracle/ PL SQL, Unix/Linux, Autosys, JIL script, windows/networking basics. Knowledge in Informatica/Micro strategy would be preferred. Understanding of Incident, Problem, Change management Good Analytical and reasoning skills. Excellent communication & Interpersonal and relationship development skills Overall experience of 3 to 5 years Exposure to working in Global team Desired Skills Dotnet Understanding. Work Timings : 7 30 AM to 4 30 PM IST/11 30 AM to 8 30 PM IST/12 30 PM to 9 30 PM IST (Rotational) Weekend support (Rotational) Job Location: Hyderabad/Chennai.

Posted 9 hours ago

Apply

3.0 years

0 Lacs

Bangalore Urban, Karnataka, India

On-site

We are looking for a detail-oriented QA Engineer to ensure the quality and accuracy of data migration projects. The ideal candidate will be responsible for validating data integrity, testing migration processes, and identifying discrepancies or issues. This role requires expertise in QA methodologies, strong analytical skills, and familiarity with data migration processes and tools. Key Responsibilities Data Validation and Testing Develop and execute comprehensive test plans and test cases to validate data migration processes. Ensure data integrity, accuracy, and consistency across source and target systems. Perform pre- and post-migration data checks to verify successful migration. Test Automation Design and implement automated test scripts for data validation and reconciliation. Use appropriate tools to streamline testing processes and reduce manual effort. Defect Identification and Resolution Identify, document, and report issues or discrepancies in the data migration process. Collaborate with development teams to troubleshoot and resolve data-related defects. Collaboration and Communication Work closely with data engineers, business analysts, and stakeholders to understand migration requirements and objectives. Provide regular updates on testing progress, results, and identified risks. Process Improvement Recommend and implement best practices for data migration testing and validation. Continuously improve QA processes to enhance efficiency and effectiveness. Documentation Maintain clear and detailed documentation of test plans, test cases, and test results. Ensure proper tracking and reporting of issues using defect management tools. Requirements Bachelors degree in Computer Science, Information Technology, or a related field. 3+ years of experience in quality assurance or data testing, preferably in data migration projects. Strong knowledge of SQL for querying and validating data. Familiarity with data migration tools and ETL processes (e.g., Informatica, Talend, or similar). Hands-on experience with test automation tools (e.g., Selenium, TestNG, or similar). Understanding of data governance, privacy, and security principles. Strong analytical skills with attention to detail. Excellent communication and collaboration abilities. Preferred Qualifications Experience with cloud-based data migration (e.g., AWS, Azure, GCP). Familiarity with big data frameworks and tools (e.g., Hadoop, Spark). Knowledge of Agile methodologies and tools like Jira or Confluence.

Posted 10 hours ago

Apply

8.0 years

0 Lacs

Pune, Maharashtra, India

On-site

Job Purpose: We are looking for a highly skilled and experienced Data Engineering professional to lead our data engineering team. The ideal candidate will possess a strong technical background, strong project management abilities, and excellent client handling/stakeholder management skills. This role requires a strategic thinker who can drive the design, development and implementation of data solutions that meet our clients’ needs while ensuring the highest standards of quality and efficiency. Job Responsibilities Technology Leadership – Lead guide the team independently or with little support to design, implement deliver complex cloud-based data engineering / data warehousing project assignments Managing projects in fast paced agile ecosystem and ensuring quality deliverables within stringent timelines Responsible for Risk Management, maintaining the Risk documentation and mitigations plan. Drive continuous improvement in a Lean/Agile environment, implementing DevOps delivery approaches encompassing CI/CD, build automation and deployments. Communication & Logical Thinking – Demonstrates strong analytical skills, employing a systematic and logical approach to data analysis, problem-solving, and situational assessment. Capable of effectively presenting and defending team viewpoints, while securing buy-in from both technical and client stakeholders. Handle Client Relationship – Manage client relationship and client expectations independently. Should be able to deliver results back to the Client independently. Should have excellent communication skills. Work Experience Should have expertise and 8+ years of working experience in at least two ETL tools among Matillion, DBT, Pyspark/python, Informatica, and Talend Should have expertise and working experience in at least two databases among Databricks, Redshift, Snowflake, SQL Server, Oracle Should have strong Data Warehousing, Data Integration and Data Modeling fundamentals like Star Schema, Snowflake Schema, Dimension Tables and Fact Tables. Strong experience on SQL building blocks. Creating complex SQL queries and Procedures. Experience in AWS or Azure cloud and its service offerings Aware of techniques such as: Data Modelling, Performance tuning and regression testing Willingness to learn and take ownership of tasks. Excellent written/verbal communication and problem-solving skills and Understanding and working experience on Pharma commercial data sets like IQVIA, Veeva, Symphony, Liquid Hub, Cegedim etc. would be an advantage Good experience working in pharma or life sciences domain projects Education BE/B.Tech, MCA, M.Sc., M. Tech with 60%+ Why Axtria: - Axtria is a global provider of cloud software and data analytics to the Life Sciences industry. We help Life Sciences companies transform the product commercialization journey to drive sales growth and improve healthcare outcomes for patients. We are acutely aware that our work impacts millions of patients and lead passionately to improve their lives. We will provide– (Employee Value Proposition) Offer an inclusive environment that encourages diverse perspectives and ideas Deliver challenging and unique opportunities to contribute to the success of a transforming organization Opportunity to work on technical challenges that may impact across geographies Vast opportunities for self-development: online Axtria Institute, knowledge sharing opportunities globally, learning opportunities through external certifications Sponsored Tech Talks & Hackathons Possibility to relocate to any Axtria office for short and long-term projects Benefit package: Health benefits Retirement benefits Paid time off Flexible Benefits Hybrid /FT Office Axtria is an equal-opportunity employer that values diversity and inclusiveness in the workplace. A few more links are mentioned below, you may want to go through to know more about Axtria’s journey as an Organization, its culture, products and solutions offerings. For White papers: Research Hub: https://www.axtria.com/axtria-research-hub-pharmaceutical-industry/ For Axtria product and capability related content: 5 step guides: https://www.axtria.com/axtria-5-step-guides-sales-marketing-data-management-best-practices/ For recent marketing videos, including Jassi’s public discussions: Video Wall: https://www.axtria.com/video-wall/ Infographic Points of view on industry, Therapy areas etc.: https://www.axtria.com/video-wall/

Posted 10 hours ago

Apply

6.0 years

0 Lacs

Noida, Uttar Pradesh, India

On-site

Your Responsibilities includes, but not limited to: Participate in overall architecture, Capacity planning, development, and implementation of Master Data Management solutions (MDM). Using MDM technologies and tools across an enterprise to enable the management and integration of master data. Understand the technical landscape current as well as desired future state Assess the current state architecture & understand current business processes for managing Master Data Management solutions. Assess the functional and non-functional requirements of desired future state MDM solution Prepare the to-be architecture including data ingestion, data quality rules, data model, match/merge, workflows, UI, batch integration and real-time services Extensive hands on experience in installation and configuration of core Informatica MDM Hub components such as Hub console, Hub Store, Hub Server, Cleanse/Match Server and Cleanse Adaptor. Ability to deliver full lifecycle MDM projects for clients including Data modeling, Metadata management, design and configuration of matching and merging rules, design and configuration of standardizing, cleansing and deduplication rules. Ability to fine-tune and optimize MDM Hub performance Hands-on experience with ActiveVOS, Informatica MDM Service Integration Framework (SIF) and Informatica Business Entity Services (BES) Create Design Document and data models addressing business needs for the client MDM environment - Contribute to creating reusable assets and accelerators for MDM platforms. Will also be involved in integration/transfer of data across multiple systems, streamlining data processes and providing access to MDM data across the enterprise. Make technology decisions related to the Client MDM environment & Interpret requirements and architect MDM solutions. Provide subject matter expertise on data architecture and data integration implementations across various downstream system. Coordinate with Project Managers and participate in project planning and recurring meetings Collaborate with other team members to review prototypes and develop iterative revisions Must have Skills : 6+ years of experience & should have hands on experience of working in MDM Projects and hands on experience in one or more MDM tools like Informatica or Reltio and has expertise in defining matching/ merging & survivor-ship rules Hands on experience in industry data quality tools like Informatica IDQ, IBM Data Quality. Must be proficient reading and understanding data models and experience working with data and databases. Must be comfortable with the concept of services (important for integrating with an ESB for operational MDM) Strong technical experience in the areas of Master Data Management, Meta data management, Data Quality, Data Governance, Data Integration (ETL) and Data Security Experience with (all stages of MDM SDLC) planning, designing, building, deploying and maintaining scalable, highly available, mission critical enterprise wide applications for large enterprises Should have experience in integrating MDM with Data Warehouses and Data Lakes Excellent query writing skills with Working knowledge of Oracle, SQL server, and other major databases Good knowledge of SOA/Real-time integration , Pub-Sub Model and Data Integration with Various CRM systems like Veeva, siebel Should have strong commercial knowledge of key business processes & compliance requirements within Pharma Industry across multiple master data domains like Physician & Product Should have experience working with 3rd Party Data Providers like IMS, LASH, HMS etc Expertise in engaging with business users to understand the business requirements and articulate the value proposition Why consider Axtria? Axtria is a data analytics and software technology company – we focus heavily on sales and marketing functions in the life sciences. We provide cloud-based solutions to help life science clients with digital transformation of their commercial operations. Axtria combines strong process knowledge of Pharma commercial operations, data analytics and software. Our cloud platforms SalesIQ™, MarketingIQ™ and DataMAx™ are the most advanced and built specifically for the life-science industry. A strong track record in delivering customer value by being innovative, flexible and transparent has enabled Axtria to become the fastest and now one of the largest providers of Pharma commercial operations solutions. In short eight years, the company has grown to over 950 employees. We work with over 75 life-science customers including many large and specialty companies. Axtria is looking for exceptional talent to join our rapidly growing global team. People are our biggest perk! Our transparent and collaborative culture offers a chance to work with some of the brightest minds in the industry. Axtria Institute, our in-house university, offers the best training in the industry and an opportunity to learn in a structured environment. A customized career progression plan ensures every associate is setup for success and able to do meaningful work in a fun environment. We want our legacy to be the leaders we produce for the industry. Will you be next?

Posted 10 hours ago

Apply

3.0 years

0 Lacs

Gurugram, Haryana, India

On-site

Project Role : Application Developer Project Role Description : Design, build and configure applications to meet business process and application requirements. Must have skills : Data Warehouse ETL Testing Good to have skills : NA Minimum 3 Year(s) Of Experience Is Required Educational Qualification : 15 years full time education Summary: As an Application Developer, you will design, build, and configure applications to meet business process and application requirements. Your typical day will involve collaborating with team members to develop innovative solutions and ensure seamless application functionality. Roles & Responsibilities: - Expected to perform independently and become an SME. - Required active participation/contribution in team discussions. - Contribute in providing solutions to work-related problems. - Develop and implement ETL test cases to ensure data accuracy. - Conduct data validation and reconciliation processes. - Collaborate with cross-functional teams to troubleshoot and resolve data issues. - Create and maintain test documentation for ETL processes. - Identify opportunities for process improvement and optimization. Professional & Technical Skills: - Must To Have Skills: Proficiency in Data Warehouse ETL Testing. - Strong understanding of SQL and database concepts. - Experience with ETL tools such as Informatica or Talend. - Knowledge of data warehousing concepts and methodologies. - Hands-on experience in data quality assurance and testing. Additional Information: - The candidate should have a minimum of 3 years of experience in Data Warehouse ETL Testing. - This position is based at our Gurugram office. - A 15 years full-time education is required.

Posted 11 hours ago

Apply

4.0 years

0 Lacs

Hyderabad, Telangana, India

On-site

Overview We are seeking a Platform Architect with expertise in Informatica PowerCenter and Informatica Intelligent Cloud Services (IICS) to design, implement, and optimize enterprise-level data integration platforms. The ideal candidate will have a strong background in ETL/ELT architecture, cloud data integration, and platform modernization, ensuring scalability, security, and performance across on-prem and cloud environments. Responsibilities Platform Engineering & Administration Oversee installation, configuration, and optimization of PowerCenter and IICS environments. Manage platform scalability, performance tuning, and troubleshooting. Implement data governance, security, and compliance (e.g., role-based access, encryption, data lineage tracking). Optimize connectivity and integrations with various sources (databases, APIs, cloud storage, SaaS apps). Cloud & Modernization Initiatives Architect and implement IICS-based data pipelines for real-time and batch processing. Migrate existing PowerCenter workflows to IICS, leveraging serverless and cloud-native features. Ensure seamless integration with cloud platforms (AWS, Azure, GCP) and modern data lakes/warehouses (Snowflake, Redshift, BigQuery). Qualifications 4 years of experience in data integration and ETL/ELT architecture. Expert-level knowledge of Informatica PowerCenter and IICS (Cloud Data Integration, API & Application Integration, Data Quality). Hands-on experience with cloud platforms (AWS, Azure, GCP) and modern data platforms (Snowflake, Databricks, Redshift, BigQuery). Strong SQL, database tuning, and performance optimization skills. Deep understanding of data governance, security, and compliance best practices. Experience in automation, DevOps (CI/CD), and Infrastructure-as-Code (IaC) tools for data platforms. Excellent communication, leadership, and stakeholder management skills. Preferred Qualifications Informatica certifications (IICS, PowerCenter, Data Governance). Proficient to Power Center to IDMC Conversions Understanding on real-time streaming (Kafka, Spark Streaming). Knowledge of API-based integration and event-driven architectures. Familiarity with Machine Learning and AI-driven data processing.

Posted 11 hours ago

Apply

5.0 - 15.0 years

0 Lacs

India

On-site

**********************************4 months contract opportunity********************************** Job Summary: We are seeking a skilled and detail-oriented Snowflake Developer to design, develop, and maintain scalable data solutions using the Snowflake platform. The ideal candidate will have experience in data warehousing, ETL/ELT processes, and cloud-based data architecture. Key Responsibilities: Design and implement data pipelines using Snowflake, SQL, and ETL tools. Develop and optimize complex SQL queries for data extraction and transformation. Create and manage Snowflake objects such as databases, schemas, tables, views, and stored procedures. Integrate Snowflake with various data sources and third-party tools. Monitor and troubleshoot performance issues in Snowflake environments. Collaborate with data engineers, analysts, and business stakeholders to understand data requirements. Ensure data quality, security, and governance standards are met. Automate data workflows and implement best practices for data management. Required Skills and Qualifications: Proficiency in Snowflake SQL and Snowflake architecture. Experience with ETL/ELT tools (e.g., Informatica, Talend, dbt, Matillion). Strong knowledge of cloud platforms (AWS, Azure, or GCP). Familiarity with data modeling and data warehousing concepts. Experience with Python, Java, or Shell scripting is a plus. Understanding of data security, role-based access control, and data sharing in Snowflake. Excellent problem-solving and communication skills. Preferred Qualifications: Snowflake certification (e.g., SnowPro Core). Experience with CI/CD pipelines and DevOps practices. Knowledge of BI tools like Tableau, Power BI, or Looker. Req 5-15 years of experience is preferred. Experience with Agile based development Problem solving skills: Proficiency in writing performant SQL Queries/Scripts to generate business insights and drive better organizational decision making.

Posted 13 hours ago

Apply

7.5 years

0 Lacs

Navi Mumbai, Maharashtra, India

On-site

Project Role : Application Lead Project Role Description : Lead the effort to design, build and configure applications, acting as the primary point of contact. Must have skills : Informatica MDM Good to have skills : NA Minimum 7.5 Year(s) Of Experience Is Required Educational Qualification : 15 years full time education Summary: As an Application Lead, you will lead the effort to design, build, and configure applications, acting as the primary point of contact. Your typical day will involve collaborating with various teams to ensure project milestones are met, addressing any challenges that arise, and providing guidance to team members to foster a productive work environment. You will also engage in strategic discussions to align project goals with organizational objectives, ensuring that the applications developed meet the needs of stakeholders effectively. Roles & Responsibilities: - Expected to be an SME. - Collaborate and manage the team to perform. - Responsible for team decisions. - Engage with multiple teams and contribute on key decisions. - Provide solutions to problems for their immediate team and across multiple teams. - Facilitate knowledge sharing sessions to enhance team capabilities. - Monitor project progress and implement necessary adjustments to meet deadlines. Professional & Technical Skills: - Must To Have Skills: Proficiency in Informatica MDM. - Strong understanding of data integration and data quality processes. - Experience with data modeling and metadata management. - Familiarity with ETL processes and data warehousing concepts. - Ability to troubleshoot and resolve technical issues efficiently. Additional Information: - The candidate should have minimum 7.5 years of experience in Informatica MDM. - This position is based in Mumbai. - A 15 years full time education is required.

Posted 19 hours ago

Apply

10.0 - 14.0 years

0 Lacs

haryana

On-site

As a Technical Consultant/Technical Architect with expertise in Fund Accounting, Oracle, and Informatica, you will collaborate with Delivery Managers, System/Business Analysts, and other subject matter experts to comprehend project requirements and design effective solutions. You will play a key role in estimating efforts for new projects and proposals, as well as producing/reviewing technical specifications and unit test cases for ongoing interfaces development. Your responsibilities will include developing and implementing standards, procedures, and best practices for data maintenance, reconciliation, and exception management. You will be expected to demonstrate technical leadership, produce design/technical specifications, propose solutions, and estimate project timelines. Additionally, your role will involve guiding and mentoring junior team members in developing solutions on the GFDR platform. Requirements: - Possess 10-12 years of experience in technical leadership within data warehousing and Business Intelligence domains - Proficiency in Oracle SQL/PLSQL and stored procedures - Familiarity with Source Control Tools (Clear Case preferred) - Sound understanding of Data Warehouse, Datamart, and ODS concepts - Experience in UNIX and PERL scripting - Proficiency in standard ETL tools like Informatica Power Centre - Demonstrated technical leadership in Eagle, Oracle, Unix Scripting, Perl, and scheduling tools such as Autosys/Control - Experience with job scheduling tools (Control-M preferred) - Strong knowledge of data modeling, data normalization, and performance optimization techniques - Ability to guide/mentor juniors in solution building and troubleshooting - Exposure to fund accounting concepts/systems and master data management is desirable - Familiarity with data distribution and access concepts and ability to translate conceptual models into physical ones - Excellent interpersonal and communication skills - Capability to collaborate effectively with cross-functional teams - Willingness to work as part of a team engaged in both development and production support activities Industry: IT/Computers-Software Role: Technical Architect Key Skills: Oracle, PL/SQL, Informatica, Autosys/Control, Fund Accounting, Eagle Education: B.E/B.Tech If you meet the qualifications and are excited about taking on a challenging and rewarding role as a Technical Consultant/Technical Architect with a focus on Fund Accounting, Oracle, and Informatica, we encourage you to reach out to us at jobs@augustainfotech.com.,

Posted 21 hours ago

Apply

10.0 - 14.0 years

0 Lacs

pune, maharashtra

On-site

You should have over 10 years of experience in data architecture, data engineering, or related roles. Your expertise should include designing and implementing enterprise-level data solutions with a hands-on technical approach. You should have a proven track record of managing client relationships and leading technical teams. In terms of technical skills, you must be well-versed in data modeling, data warehousing, and database design, including both relational and NoSQL databases. You should have a strong proficiency in data engineering, which includes experience with ETL tools, data integration frameworks, and big data technologies. Hands-on experience with Google Cloud data platform and modern data processing frameworks is crucial. Moreover, familiarity with scripting and programming languages like Python and SQL for hands-on development and troubleshooting is essential. Experience with data governance frameworks & solutions such as Informatica, Collibra, Purview, etc., will be a plus. Soft skills required for this role include exceptional client management and communication skills to confidently interact with both technical and non-technical stakeholders. You should possess proven team management and leadership abilities, including mentoring, coaching, and project management. Strong analytical and problem-solving skills with a proactive, detail-oriented approach are necessary. The ability to work collaboratively in a fast-paced, dynamic environment while successfully driving multiple projects to completion is important. Preferred certifications for this position include Professional Cloud Architect (GCP), Data Architect, Certified Data Management Professional (CDMP), or similar credentials.,

Posted 21 hours ago

Apply

5.0 years

0 Lacs

Pune, Maharashtra, India

On-site

Job Description Process Manager - GCP Data Engineer Mumbai/Pune | Full-time (FT) | Technology Services Shift Timings - EMEA(1pm-9pm)| Management Level - PM| Travel Requirements - NA The ideal candidate must possess in-depth functional knowledge of the process area and apply it to operational scenarios to provide effective solutions. The role enables to identify discrepancies and propose optimal solutions by using a logical, systematic, and sequential methodology. It is vital to be open-minded towards inputs and views from team members and to effectively lead, control, and motivate groups towards company objects. Additionally, candidate must be self-directed, proactive, and seize every opportunity to meet internal and external customer needs and achieve customer satisfaction by effectively auditing processes, implementing best practices and process improvements, and utilizing the frameworks and tools available. Goals and thoughts must be clearly and concisely articulated and conveyed, verbally and in writing, to clients, colleagues, subordinates, and supervisors. Process Manager Roles And Responsibilities Participate in Stakeholder interviews, workshops, and design reviews to define data models, pipelines, and workflows. Analyse business problems and propose data-driven solutions that meet stakeholder objectives. Experience on working on premise as well as cloud platform (AWS/GCP/Azure) Should have extensive experience in GCP with a strong focus on Big Query, and will be responsible for designing, developing, and maintaining robust data solutions to support analytics and business intelligence needs. (GCP is preferable over AWS & Azure) Design and implement robust data models to efficiently store, organize, and access data for diverse use cases. Design and build robust data pipelines (Informatica / Fivertan / Matillion / Talend) for ingesting, transforming, and integrating data from diverse sources. Implement data processing pipelines using various technologies, including cloud platforms, big data tools, and streaming frameworks (Optional). Develop and implement data quality checks and monitoring systems to ensure data accuracy and consistency. Technical And Functional Skills Bachelor’s Degree with 5+ years of experience with relevant 3+ years hands-on of experience in GCP with BigQuery. Good knowledge of any 1 of the databases scripting platform (Oracle preferable) Work would involve analysis, development of code/pipelines at modular level, reviewing peers code and performing unit testing and owning push to prod activities. With 5+ of work experience and worked as Individual contributor for 5+ years Direct interaction and deep diving with VPs of deployment Should work with cross functional team/ stakeholders Participate in Backlog grooming and prioritizing tasks Worked on Scrum Methodology. GCP certification desired. About EClerx eClerx is a global leader in productized services, bringing together people, technology and domain expertise to amplify business results. Our mission is to set the benchmark for client service and success in our industry. Our vision is to be the innovation partner of choice for technology, data analytics and process management services. Since our inception in 2000, we've partnered with top companies across various industries, including financial services, telecommunications, retail, and high-tech. Our innovative solutions and domain expertise help businesses optimize operations, improve efficiency, and drive growth. With over 18,000 employees worldwide, eClerx is dedicated to delivering excellence through smart automation and data-driven insights. At eClerx, we believe in nurturing talent and providing hands-on experience. About About eClerx Technology eClerx’s Technology Group collaboratively delivers Analytics, RPA, AI, and Machine Learning digital technologies that enable our consultants to help businesses thrive in a connected world. Our consultants and specialists’ partner with our global clients and colleagues to build and implement digital solutions through a broad spectrum of activities. To know more about us, visit https://eclerx.com eClerx is an Equal Opportunity Employer. All qualified applicants will receive consideration for employment without regard to race, color, religion, sex, national origin, disability or protected veteran status, or any other legally protected basis, per applicable law

Posted 21 hours ago

Apply

4.0 - 8.0 years

0 Lacs

delhi

On-site

The ideal candidate should possess extensive expertise in SQL, data modeling, ETL/ELT pipeline development, and cloud-based data platforms like Databricks or Snowflake. You will be responsible for designing scalable data models, managing reliable data workflows, and ensuring the integrity and performance of critical financial datasets. Collaboration with engineering, analytics, product, and compliance teams is a key aspect of this role. Responsibilities: - Design, implement, and maintain logical and physical data models for transactional, analytical, and reporting systems. - Develop and oversee scalable ETL/ELT pipelines to process large volumes of financial transaction data. - Optimize SQL queries, stored procedures, and data transformations for enhanced performance. - Create and manage data orchestration workflows using tools like Airflow, Dagster, or Luigi. - Architect data lakes and warehouses utilizing platforms such as Databricks, Snowflake, BigQuery, or Redshift. - Ensure adherence to data governance, security, and compliance standards (e.g., PCI-DSS, GDPR). - Work closely with data engineers, analysts, and business stakeholders to comprehend data requirements and deliver solutions. - Conduct data profiling, validation, and quality assurance to maintain clean and consistent data. - Maintain comprehensive documentation for data models, pipelines, and architecture. Required Skills & Qualifications: - Proficiency in advanced SQL, including query tuning, indexing, and performance optimization. - Experience in developing ETL/ELT workflows with tools like Spark, dbt, Talend, or Informatica. - Familiarity with data orchestration frameworks such as Airflow, Dagster, Luigi, etc. - Hands-on experience with cloud-based data platforms like Databricks, Snowflake, or similar technologies. - Deep understanding of data warehousing principles like star/snowflake schema, slowly changing dimensions, etc. - Knowledge of cloud services (AWS, GCP, or Azure) and data security best practices. - Strong analytical and problem-solving skills in high-scale environments. Preferred Qualifications: - Exposure to real-time data pipelines like Kafka, Spark Streaming. - Knowledge of data mesh or data fabric architecture paradigms. - Certifications in Snowflake, Databricks, or relevant cloud platforms. - Familiarity with Python or Scala for data engineering tasks.,

Posted 21 hours ago

Apply

2.0 - 6.0 years

0 Lacs

hyderabad, telangana

On-site

A career within Functional and Industry Technologies services will provide you with the opportunity to build secure and new digital experiences for customers, employees, and suppliers. We focus on improving apps or developing new apps for traditional and mobile devices as well as conducting usability testing to find ways to improve our clients user experience. As part of our team, youll help clients harness technology systems in financial services focusing on areas such as insurance, sales performance management, retirement and pension, asset management, and banking & capital markets. To really stand out and make us fit for the future in a constantly changing world, each and every one of us at PwC needs to be a purpose-led and values-driven leader at every level. To help us achieve this we have the PwC Professional; our global leadership development framework. It gives us a single set of expectations across our lines, geographies and career paths, and provides transparency on the skills we need as individuals to be successful and progress in our careers, now and in the future. Responsibilities As a Senior Associate, you'll work as part of a team of problem solvers, helping to solve complex business issues from strategy to execution. PwC Professional skills and responsibilities for this management level include but are not limited to: - Use feedback and reflection to develop self awareness, personal strengths and address development areas. - Delegate to others to provide stretch opportunities, coaching them to deliver results. - Demonstrate critical thinking and the ability to bring order to unstructured problems. - Use a broad range of tools and techniques to extract insights from current industry or sector trends. - Review your work and that of others for quality, accuracy and relevance. - Know how and when to use tools available for a given situation and can explain the reasons for this choice. - Seek and embrace opportunities which give exposure to different situations, environments and perspectives. - Use straightforward communication, in a structured way, when influencing and connecting with others. - Able to read situations and modify behavior to build quality relationships. - Uphold the firm's code of ethics and business conduct. Years of Experience - 2 to 5 years of experience Education Qualification: BTech/BE/MTech/MS/MCA Preferred Skill Set/Roles and Responsibility - - Hands-on Experience in P&C Insurance on Guidewire DataHub/InfoCenter Platform. - Experience in mapping Guidewire Insurance Suite of products (PC/BC/CC/CM) to DHIC. - Works with business in identifying detailed analytical and operational reporting/extracts requirements. - Able to create Microsoft SQL / ETL / SSIS complex queries. - Participates in Sprint development, test, and integration activities. - Creates detailed source to target mappings. - Creates and validates data dictionaries - Writes and validates data translation and migration scripts. - Communicating with business to gather business requirements. - Performs GAP analysis between existing (legacy) and new (GW) data related solutions. - Working with Informatica ETL devs. - Knowledge of Cloud AWS,

Posted 22 hours ago

Apply

3.0 - 7.0 years

0 Lacs

pune, maharashtra

On-site

The IT Quality Analyst 2 role is a developing professional position where you will apply your specialty area knowledge to monitor, assess, analyze, and evaluate processes and data. Your responsibilities will include identifying policy gaps, formulating policies, interpreting data, making recommendations, and researching factual information. You will be expected to identify inconsistencies in data or results, define business issues, and formulate recommendations on policies, procedures, or practices. Additionally, you will integrate established disciplinary knowledge within your own specialty area with a basic understanding of related industry practices. As the IT Quality Analyst 2, you will play a key role in owning and driving automation in the Data and Analytics Team to achieve 90% automation in the Data and ETL space. You will implement CI/CD pipelines to automate regression suites for hands-free execution and design and develop an integrated portal to consolidate utilities and cater to user needs. Your responsibilities will also include supporting initiatives related to automation on Data & Analytics testing requirements for process and product rollout into production. You will work with the technology team to design and implement appropriate automation scripts/plans for application testing, meeting required KPI and automation effectiveness. You will ensure the timely completion of Sanity and Regression automated suites, report issues, and make recommendations for solutions. Documenting new utilities and transitioning them to testers for execution, as well as providing support for troubleshooting when required, will also be part of your role. You will monitor and review code check-ins from peers and help maintain the project repository. Additionally, you should have the ability to work independently as well as collaborate within groups on various assigned projects and effectively manage multiple priorities in a fast-paced, dynamic environment. To be successful in this role, you should have 3 to 6 years of experience in ETL testing, data validation, and data analysis. Strong SQL skills and experience working with ETL tools such as Informatica, Talend, or SSIS are also required. Experience with testing tools like JIRA and HP ALM, excellent communication skills, and the ability to collaborate with cross-functional teams are essential. Familiarity with Agile/Scrum software development methodology, the ability to work independently and as a team member, and strong analytical and problem-solving skills are key qualifications for this role. A Bachelor's/University degree or equivalent experience is required for this position.,

Posted 23 hours ago

Apply

2.0 - 6.0 years

0 Lacs

karnataka

On-site

As a Salesforce consulting generalist at PwC, your role involves providing consulting services for a variety of business applications, with a focus on Salesforce. Your responsibilities include analyzing client needs, implementing software solutions, and offering training and support to ensure effective utilization of Salesforce applications. By leveraging your broad range of consulting skills and experience, you will assist clients in optimizing operational efficiency and achieving their strategic objectives. Your approach is centered around building meaningful client relationships and continuously learning how to manage and inspire others. You navigate complex situations with ease, deepening your technical expertise and awareness of your strengths. Anticipating the needs of your teams and clients, you consistently deliver quality results. Embracing ambiguity, you ask questions and view uncertainties as opportunities for growth. To excel in this role, you need to possess a diverse set of skills, knowledge, and experiences. This includes responding effectively to different perspectives, utilizing various tools and techniques to generate ideas, employing critical thinking to solve complex problems, understanding project objectives in the broader context, and interpreting data to derive insights and recommendations. Furthermore, you must uphold professional and technical standards while adhering to the Firm's code of conduct and independence requirements. Key Qualifications: - Educational Background: BE/B.Tech/MCA/M.Sc/M.E/M.Tech - Years of Experience: 2 to 4 years in IT - Skills: Salesforce, Lightning, Apex, Visualforce, Triggers, Integration, Force.com, Implementation, and deployment Must-Have Skills: - Total IT experience of 2 to 4 years - Minimum 2+ years of SFDC experience - Minimum 1+ year of Lightning experience with the ability to write clean code - Proficiency in Salesforce Lightning and Integration - Strong experience in Force.com platform using APEX and Visualforce - Solid Implementation experience in Sales/Service/Custom cloud - Expertise in SFDC configuration and mapping business requirements - Experience in Data Migration using SFDC utilities - Implementing integration solutions between CRM, ERP, and Financial systems - Strong knowledge of RDBMS and SQL queries - Familiarity with HTML, CSS, Ajax, JavaScript, and JQuery Responsibilities: - Collaborate with a team of experienced consultants to help clients solve complex business issues - Manage multiple clients while reporting to Managers - Lead and train staff members - Develop effective relationships with clients - Contribute to enhancing technical skills within the team - Stay informed about local and national business trends - Engage in business development activities to identify opportunities - Cultivate internal relationships and strengthen the PwC brand Desired Skills (Good To Have): - Object-Oriented programming knowledge (Java, Ruby, C++) - Experience with Bootstrap, Angular JS, and Lightning components - Familiarity with marketing tools like Marketing Cloud and Exact Target - Proficiency in products like Apttus, Veeva, nCino, Adobe Flex - Expertise in data management, including data load, translation, hygiene, migration, and integration - Ability to evaluate technical processes strategically - Recommend optimal solutions to team members and customers - Commitment to code quality, modularity, and version control - Experience with custom solutions on SAP, Oracle, MS-SQL Server, or other databases - Understanding of integration platforms such as Cast Iron, Boomi, Informatica, Tibco, and Fusion - Translate customer requirements into functional Salesforce configurations - Track record of managing consulting engagements effectively - Ability to think independently and creatively, embracing technical challenges - Awareness of evolving Cloud technologies and willingness to adapt to new trends,

Posted 1 day ago

Apply

6.0 - 10.0 years

0 Lacs

hyderabad, telangana

On-site

As a Lead Data Engineer specializing in Snowflake Migration at Anblicks, you will be a key player in our Data Modernization Center of Excellence (COE). You will be at the forefront of transforming traditional data platforms by utilizing Snowflake, cloud-native tools, and intelligent automation to help enterprises unlock the power of the cloud. Your primary responsibility will be to lead the migration of legacy data warehouses such as Teradata, Netezza, Oracle, or SQL Server to Snowflake. You will re-engineer and modernize ETL pipelines using cloud-native tools and frameworks like DBT, Snowflake Tasks, Streams, and Snowpark. Additionally, you will design robust ELT pipelines on Snowflake that ensure high performance, scalability, and cost optimization, while integrating Snowflake with AWS, Azure, or GCP. In this role, you will also focus on implementing secure and compliant architectures with RBAC, masking policies, Unity Catalog, and SSO. Automation of repeatable tasks, ensuring data quality and parity between source and target systems, and mentoring junior engineers will be essential aspects of your responsibilities. Collaboration with client stakeholders, architects, and delivery teams to define migration strategies, as well as presenting solutions and roadmaps to technical and business leaders, will also be part of your role. To qualify for this position, you should have at least 6 years of experience in Data Engineering or Data Warehousing, with a minimum of 3 years of hands-on experience in Snowflake design and development. Strong expertise in migrating ETL pipelines from Talend and/or Informatica to cloud-native alternatives, proficiency in SQL, data modeling, ELT design, and pipeline performance tuning are prerequisites. Familiarity with tools like DBT Cloud, Airflow, Snowflake Tasks, or similar orchestrators, as well as a solid understanding of cloud data architecture, security frameworks, and data governance, are also required. Preferred qualifications include Snowflake certifications (SnowPro Core and/or SnowPro Advanced Architect), experience with custom migration tools, metadata-driven pipelines, or LLM-based code conversion, familiarity with domain-specific architectures in Retail, Healthcare, or Manufacturing, and prior experience in a COE or modernization-focused consulting environment. By joining Anblicks as a Lead Data Engineer, you will have the opportunity to lead enterprise-wide data modernization programs, tackle complex real-world challenges, and work alongside certified Snowflake architects, cloud engineers, and innovation teams. You will also have the chance to build reusable IP that scales across clients and industries, while experiencing accelerated career growth in the dynamic Data & AI landscape.,

Posted 1 day ago

Apply

7.5 years

0 Lacs

Navi Mumbai, Maharashtra, India

On-site

Project Role : Application Lead Project Role Description : Lead the effort to design, build and configure applications, acting as the primary point of contact. Must have skills : Databricks Unified Data Analytics Platform, Informatica Intelligent Cloud Services Good to have skills : NA Minimum 7.5 Year(s) Of Experience Is Required Educational Qualification : 15 years full time education Summary: As an Application Lead, you will lead the effort to design, build, and configure applications, acting as the primary point of contact. Your typical day will involve collaborating with various teams to ensure project milestones are met, facilitating discussions to address challenges, and guiding your team through the development process while maintaining a focus on quality and efficiency. You will also engage in strategic planning to align application development with organizational goals, ensuring that all stakeholders are informed and involved throughout the project lifecycle. Roles & Responsibilities: - Expected to be an SME. - Collaborate and manage the team to perform. - Responsible for team decisions. - Engage with multiple teams and contribute on key decisions. - Provide solutions to problems for their immediate team and across multiple teams. - Facilitate training and development opportunities for team members to enhance their skills. - Monitor project progress and implement necessary adjustments to meet deadlines. Professional & Technical Skills: - Must To Have Skills: Proficiency in Databricks Unified Data Analytics Platform, Informatica Intelligent Cloud Services. - Good To Have Skills: Experience with cloud-based data integration tools. - Strong understanding of data engineering principles and practices. - Experience with big data technologies such as Apache Spark and Hadoop. - Familiarity with data governance and data quality frameworks. Additional Information: - The candidate should have minimum 7.5 years of experience in Databricks Unified Data Analytics Platform. - This position is based in Mumbai. - A 15 years full time education is required.

Posted 1 day ago

Apply

12.0 - 18.0 years

0 Lacs

Bengaluru East, Karnataka, India

On-site

Nasdaq Technology is looking for a passionate Accounts Payable and Procurement Technical Specialist with expertise in the Workday Financials and Coupa product suite to join the Bangalore technology center in India. If Innovation and effectiveness drive you forward this is the place for you! Nasdaq is continuously revolutionizing markets and undergoing transformations while we adopt new technologies to develop innovative solutions, constantly aiming to rewrite tomorrow. As a senior technical analyst, you will play a key role in the delivery of complex technical systems of varying sizes to new and existing customers and will be part of discovering new technologies within the FinTech industry. We are looking for a technical specialist to strengthen the Enterprise Solutions team in Bangalore. You will be responsible for driving the execution of central initiatives across the Nasdaq corporate technology portfolio of Software Products and Software Services. We are looking for candidates who share those values along with a genuine desire and drive to deliver top technology solutions to today's markets. With this position we offer : Join the development organization for Nasdaq’s Enterprise Solutions team. The main objective for the development team in Bangalore is to build state of the art corporate software for all of Nasdaq’s employees. As our new member, you will work with a group of enthusiastic and experienced team members. Together with your team, you will be responsible for delivery of enhancements and supporting rearchitecture of our enterprise products to new software paradigms. Role Responsibilities: You will be responsible for cross-functional work across the globe, delivering critical solutions and services to Nasdaq’s finance processes and operations. You will be responsible for crucial design activities and internal customer interaction, building strong relations and trust with key business and technology stakeholders. In this role, you will have the opportunity to work in a strong experienced team delivering sophisticated technology solutions worldwide. You will work with and learn from subject matter experts, and have the opportunity to take on more responsibility and progress within the Enterprise Solutions team. Besides working closely with your colleagues in Bangalore, you will also work closely with Nasdaq teams in other countries. Establish and enforce standards and best development practices for the team Integrate and automate AP and procurement processes at Nasdaq Evaluate external software packages and develop recommendations for future use at Nasdaq, including ROI calculations. Proactively identify and recommend solutions for problems or issues identified in the configuration of ERP platforms and ancillary systems supporting AP, procurement, and asset management. Develop and provide executive-level architecture presentations related to the Corporate Finance suite of plaforms We expect you to have At least 12 to 18 years of experience in software implementation and configuration in the financial ERP space. Possessing knowledge of business processes and related accounting as relevant to Workday Finance modules (GL, AP, AR, etc.) – e.g., Procure-to-Pay Cycle, Record to Report , etc Must have experience in Workday’s Finance modules, Workday Extend, and Workday Studio Coupa, Navan or other relevant procurement and expense management platforms Proficiency in accounting processes, with special emphasis on Accounts Payable, Asset Capitalization and Management, and Accruals Experience in executive-level presentation skills, both written and oral Education Qualification: Bachelor/Master in computer science or other engineering fields. It would be great if you . Experience in Informatica is added advantage. Knowledge of finance organization processes, including AP, procurement, GL accounting, Asset Capitalization, and Planning & Forecasting Experience working in multinational, multi-geographic companies and organizations Does it sound like you? What happens now? As the selection process is ongoing, please submit your application in English as soon as possible. We will get back to you in 2-3 weeks. What will it be like working here? Nasdaq is a vibrant and entrepreneurial company where everyone is encouraged to take initiative, challenge status quo, and take intelligent risks. We want everyone to feel welcome and bring their authentic self to work. Every day, we are building a culture where we all feel connected, supported and empowered. We are a Hybrid-first environment that embraces work-life balance and fosters well-being. What do we offer you? Annual monetary bonus. An opportunity to become a Nasdaq shareholder Employee Stock Purchase Program Nasdaq stocks with a discount Health Insurance Program Flexible working schedule and hybrid way of work Flex day program (up to 6 paid days off a year) in addition to standard vacations and holidays Internal mentorship program – get a mentor or become one Wide selection of online learning resources, e.g., Udemy Come as You Are Nasdaq is an equal opportunity employer. We positively encourage applications from suitably qualified and eligible candidates regardless of age, color, disability, national origin, ancestry, race, religion, gender, sexual orientation, gender identity and/or expression, veteran status, genetic information, or any other status protected by applicable law. We will ensure that individuals with disabilities are provided reasonable accommodation to participate in the job application or interview process, to perform essential job functions, and to receive other benefits and privileges of employment. Please contact us to request an accommodation.

Posted 1 day ago

Apply

15.0 years

0 Lacs

India

On-site

Hiring for an US Based Product Company Position: Senior Cloud Infrastructure Engineer Experience: 15+ Years Roles & Responsibilities: Cloud experience: AWS, Azure. Kubernetes expert – Container orchestration and management Harness – Kubernetes Cost Control With Built-In Intelligence Densify – Kubernetes resource optimization Strong experience in CI/CD PROCESS AND TOOLS Strong knowledge and experience in Disaster Recovery / Business Continuity Plans Tools: Terraform, Ansible – Cloud infrastructure automation, Pulumi – IaC platform for any programming language, Jenkins – CI/CD for complex workflows, GitHub Actions – Native CI/CD built into GitHub etc. Cloud Testing Tools : Apache JMeter – Open-source load testing for web and APIs, BlazeMeter – Scalable cloud-based load testing platform, LoadRunner – Performance testing software tool etc. Cloud Provisioning Tools : AWS CloudFormation, Azure Resource Manager – Native Provisioning for Azure, Google Cloud Infrastructure Manager – Native Terraform provisioning for GCP, Cloudsfer – Cloud migration tool etc. Multi-Cloud Management Solutions: Lacework FortiCNAPP – Cloud security tool, Cloudify – Open-source, multi-cloud orchestration platform, CoreStack – Next-gen cloud business accelerator etc. Data Integration And Management Platforms: Informatica – Cloud data integration platform, AWS hosting, managing and deployment would be a big plus including archiving, backup and restore, cloud migration, DevOps. Interested can share resume at chandni@thepremierconsultants.com #aws #azure #cloudinfrastructure #devops #kubernates #terraform #cloudsecurity #dockeransible

Posted 1 day ago

Apply

3.0 years

0 Lacs

Gurgaon, Haryana, India

On-site

Technical Support: Provide L1, L2 and L3 support for software applications. Troubleshoot and resolve application-related issues for end-users. Collaborate with developers, IT teams, and external vendors to address issues and implement fixes. Escalate unresolved issues to higher-level support or specialized teams. Monitoring and Maintenance: Monitor scheduled Jobs and ensure their successful completion. Perform routine maintenance tasks, including system updates, backups, and configuration changes. Assist with system upgrades, patches, and migrations to ensure continuity of service. Incident Management: Log, track, and manage incidents and service requests via ticketing systems. Follow established procedures for incident escalation and resolution. Participate in root cause analysis and problem management efforts. Documentation and Reporting: Maintain and update application documentation, including configuration settings and user guides. Create and present reports on system performance, issues, and resolution times. Document changes, fixes, and enhancements in a knowledge base for future reference. Collaboration: Work with cross-functional teams (IT, development, business users) to gather requirements and improve applications. Participate in system testing and quality assurance activities. Assist in the development and implementation of new application modules or features. Key Skills and Qualifications: Technical Skills: 3+ Years of experience in Informatica B2B (DX, DT) 3+ Years of experience in Informatica Power Center and IICS 3+ Years of experience with databases (MS-SQL server), Experience with application monitoring tools. Experience In Informatica Cloud Data Governance Catalog Is Preferred Analytical Skills: Ability to diagnose and solve complex technical problems. Communication: Strong verbal and written communication skills, able to explain technical concepts to non-technical users. Customer Service: Ability to provide excellent customer service under pressure and manage competing priorities. Knowledge of ITIL processes (incident, problem, and change management).

Posted 1 day ago

Apply

2.0 - 6.0 years

3 - 8 Lacs

Hyderābād

On-site

Join Amgen’s Mission of Serving Patients At Amgen, if you feel like you’re part of something bigger, it’s because you are. Our shared mission—to serve patients living with serious illnesses—drives all that we do. Since 1980, we’ve helped pioneer the world of biotech in our fight against the world’s toughest diseases. With our focus on four therapeutic areas –Oncology, Inflammation, General Medicine, and Rare Disease– we reach millions of patients each year. As a member of the Amgen team, you’ll help make a lasting impact on the lives of patients as we research, manufacture, and deliver innovative medicines to help people live longer, fuller happier lives. Our award-winning culture is collaborative, innovative, and science based. If you have a passion for challenges and the opportunities that lay within them, you’ll thrive as part of the Amgen team. Join us and transform the lives of patients while transforming your career. Associate IS Engineer - Veeva Vault Promomats_Medcomms What you will do Let’s do this. Let’s change the world. In this vital role in the Veeva Vault team you will be responsible for designing, developing, and maintaining software applications and solutions in Amgen’s Vault PromoMats and Vault MedComm, that meet business needs and ensuring the availability and performance of critical systems and applications. This role involves working closely with product managers, designers, and other engineers to create high-quality, scalable software solutions and automating operations, monitoring system health, and responding to incidents to minimize downtime. Roles & Responsibilities: Possesses strong rapid prototyping skills and can quickly translate concepts into working code. Lead day to day operations and maintenance of Amgen’s Amgen’s Vault PromoMats and Vault MedComm and its hosted applications. Stay updated with the latest trends, advancements and standard processes for Veeva Vault Platform ecosystem. Design, develop, and implement applications and modules, including custom reports, SDKs, interfaces, and enhancements. Analyze and understand the functional & technical requirements of applications, solutions and systems, translate them into software architecture and design specifications. Develop and complete unit tests, integration tests, and other testing strategies to ensure the quality of the software following IS change control and GxP Validation process while exhibiting expertise in Risk Based Validation methodology. Work closely with multi-functional teams, including product management, design, and QA, to deliver high-quality software on time. Maintain detailed documentation of software designs, code, and development processes. Work on integrating with other systems and platforms to ensure seamless data flow and functionality. Stay up to date on Veeva Vault Features, new releases and standard methodologies around Veeva Platform Governance. What we expect of you Basic Qualifications and Experience: Bachelor’s degree and 2 to 6 years of Information Systems experience or related field Functional Skills: Must-Have Skills: Experience with Amgen’s Vault PromoMats and Vault MedComm, including Veeva configuration settings and custom builds. Strong knowledge of information systems and network technologies. Experience in building configured and custom solutions on Veeva Vault Platform. Experience in managing systems, implementing and validating projects in GxP regulated environments. Extensive expertise in SDLC, including requirements, design, testing, data analysis, creating and managing change controls. Proficiency in programming languages such as Python, JavaScript etc. Solid understanding of software development methodologies, including Agile and Scrum. Experience with version control systems such as Git. Good-to-Have Skills: Familiarity with relational databases (such as MySQL, SQL server, PostgreSQL etc.) Proficiency in programming languages such as Python, JavaScript or other programming languages Outstanding written and verbal communication skills, and ability to translate technical concepts for non-technical audiences. Experience with ETL Tools (Informatica, Databricks). Experience with API integrations such as MuleSoft. Solid understanding & Proficiency in writing SQL queries. Hands on experience on reporting tools such as Tableau, Spotfire & Power BI. Professional Certifications: Veeva Vault Platform Administrator or Equivalent Vault Certification (Must-Have) SAFe for Teams (Preferred) Soft Skills: Excellent analytical and fix skills. Strong verbal and written communication skills. Ability to work effectively with global, virtual teams. Team-oriented, with a focus on achieving team goals. Strong presentation and public speaking skills. Work Hours: This position requires you to work a later shift and may be assigned a second or third shift schedule. Candidates must be willing and able to work during evening or night shifts, as required. Potential Shifts (subject to change based on business requirements): Second Shift: 2:00pm – 10:00pm IST; Third Shift: 10:00 pm – 7:00 am IST. What you can expect of us As we work to develop treatments that take care of others, we also work to care for your professional and personal growth and well-being. From our competitive benefits to our collaborative culture, we’ll support your journey every step of the way. In addition to the base salary, Amgen offers competitive and comprehensive Total Rewards Plans that are aligned with local industry standards. Apply now and make a lasting impact with the Amgen team. careers.amgen.com As an organization dedicated to improving the quality of life for people around the world, Amgen fosters an inclusive environment of diverse, ethical, committed and highly accomplished people who respect each other and live the Amgen values to continue advancing science to serve patients. Together, we compete in the fight against serious disease. Amgen is an Equal Opportunity employer and will consider all qualified applicants for employment without regard to race, color, religion, sex, sexual orientation, gender identity, national origin, protected veteran status, disability status, or any other basis protected by applicable law. We will ensure that individuals with disabilities are provided reasonable accommodation to participate in the job application or interview process, to perform essential job functions, and to receive other benefits and privileges of employment. Please contact us to request accommodation.

Posted 1 day ago

Apply

0 years

0 Lacs

Telangana

On-site

We are looking for an experienced and motivated Senior Data Engineer to join our dynamic team. In this role, The role primarily focuses on MDM and associated ETL and real-time feeds monitoring and support. This engineer will be part of the global L1/L2 production support team , which is split between Chubb Engineering Centers in India and Mexico. Key responsibilities will include monitoring ETL processes, handling automated issues, and ensuring compliance with security policies. A good understanding of MDM Informatica and Data Factory is preferred The ideal candidate will have experience with Powercenter, MDM, Azure Data Factory, be able to identify and resolve data quality issues, proactively monitor production systems, performance bottlenecks, and other ETL-related problems. Responsibilities: Monitor ETL jobs including Powercenter/IICS, kafka based near real-time updates, batch processes. Troubleshoot production incidents Understands data mapping and data modeling methodologies including normal form, star, and snowflake to reduce data redundancy and improve data integrity. Maintains knowledge on current and emerging developments/trends for assigned area(s) of responsibility, assesses the impact, and collaborates with Scrum Team and Leadership to

Posted 1 day ago

Apply

4.0 years

0 Lacs

Telangana

On-site

We are looking for an experienced and motivated Senior Data Engineer to join our dynamic team. In this role, The role primarily focuses on MDM and associated ETL and real-time feeds monitoring and support. This engineer will be part of the global L1/L2 production support team , which is split between Chubb Engineering Centers in India and Mexico. Key responsibilities will include monitoring ETL processes, handling automated issues, and ensuring compliance with security policies. A good understanding of MDM Informatica and Data Factory is preferred The ideal candidate will have experience with Powercenter, MDM, Azure Data Factory, be able to identify and resolve data quality issues, proactively monitor production systems, performance bottlenecks, and other ETL-related problems. Responsibilities: Monitor ETL jobs including Powercenter/IICS, kafka based near real-time updates, batch processes. Troubleshoot production incidents Understands data mapping and data modeling methodologies including normal form, star, and snowflake to reduce data redundancy and improve data integrity. Maintains knowledge on current and emerging developments/trends for assigned area(s) of responsibility, assesses the impact, and collaborates with Scrum Team and Leadership to 4 Year/bachelor’s degree or equivalent work experience (4 years of experience in lieu of Bachelors)_ At least 5+ years of Strong understanding of ETL development concepts and tools such as ETL development solutions (e.g. Powercenter and/or IICS, Informatica MDM, Azure Data Factory, Snowflake) Experience with Data Warehousing and Business Intelligence concepts and technologies Knowledge of SQL and advanced programming languages such as Python and Java Demonstrated critical thinking skills and the ability to identify and resolve data quality issues, performance bottlenecks, and other ETL-related problems Experience with Agile methodologies and project-management skills Excellent communication and interpersonal skills 2+ years of experience in scheduling jobs using Autosys (or comparable distributed scheduler) 3+ years of experience writing Unix/Linux or Windows Scripts in tools such as PERL, Shell script, Python, etc. 3+ years of experience in creating complex technical specifications from business requirements/specifications

Posted 1 day ago

Apply

Exploring Informatica Jobs in India

The informatica job market in India is thriving with numerous opportunities for skilled professionals in this field. Companies across various industries are actively hiring informatica experts to manage and optimize their data integration and data quality processes.

Top Hiring Locations in India

  1. Bangalore
  2. Pune
  3. Hyderabad
  4. Chennai
  5. Mumbai

Average Salary Range

The average salary range for informatica professionals in India varies based on experience and expertise: - Entry-level: INR 3-5 lakhs per annum - Mid-level: INR 6-10 lakhs per annum - Experienced: INR 12-20 lakhs per annum

Career Path

A typical career progression in the informatica field may include roles such as: - Junior Developer - Informatica Developer - Senior Developer - Informatica Tech Lead - Informatica Architect

Related Skills

In addition to informatica expertise, professionals in this field are often expected to have skills in: - SQL - Data warehousing - ETL tools - Data modeling - Data analysis

Interview Questions

  • What is Informatica and why is it used? (basic)
  • Explain the difference between a connected and unconnected lookup transformation. (medium)
  • How can you improve the performance of a session in Informatica? (medium)
  • What are the various types of cache in Informatica? (medium)
  • How do you handle rejected rows in Informatica? (basic)
  • What is a reusable transformation in Informatica? (basic)
  • Explain the difference between a filter and router transformation in Informatica. (medium)
  • What is a workflow in Informatica? (basic)
  • How do you handle slowly changing dimensions in Informatica? (advanced)
  • What is a mapplet in Informatica? (medium)
  • Explain the difference between an aggregator and joiner transformation in Informatica. (medium)
  • How do you create a mapping parameter in Informatica? (basic)
  • What is a session and a workflow in Informatica? (basic)
  • What is a rank transformation in Informatica and how is it used? (medium)
  • How do you debug a mapping in Informatica? (medium)
  • Explain the difference between static and dynamic cache in Informatica. (advanced)
  • What is a sequence generator transformation in Informatica? (basic)
  • How do you handle null values in Informatica? (basic)
  • Explain the difference between a mapping and mapplet in Informatica. (basic)
  • What are the various types of transformations in Informatica? (basic)
  • How do you implement partitioning in Informatica? (medium)
  • Explain the concept of pushdown optimization in Informatica. (advanced)
  • How do you create a session in Informatica? (basic)
  • What is a source qualifier transformation in Informatica? (basic)
  • How do you handle exceptions in Informatica? (medium)

Closing Remark

As you prepare for informatica job opportunities in India, make sure to enhance your skills, stay updated with the latest trends in data integration, and approach interviews with confidence. With the right knowledge and expertise, you can excel in the informatica field and secure rewarding career opportunities. Good luck!

cta

Start Your Job Search Today

Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.

Job Application AI Bot

Job Application AI Bot

Apply to 20+ Portals in one click

Download Now

Download the Mobile App

Instantly access job listings, apply easily, and track applications.

Featured Companies