Get alerts for new jobs matching your selected skills, preferred locations, and experience range. Manage Job Alerts
5.0 years
0 Lacs
Bengaluru East, Karnataka, India
On-site
Our Company Changing the world through digital experiences is what Adobe’s all about. We give everyone—from emerging artists to global brands—everything they need to design and deliver exceptional digital experiences! We’re passionate about empowering people to create beautiful and powerful images, videos, and apps, and transform how companies interact with customers across every screen. We’re on a mission to hire the very best and are committed to creating exceptional employee experiences where everyone is respected and has access to equal opportunity. We realize that new ideas can come from everywhere in the organization, and we know the next big idea could be yours! Job Description Role Summary Digital Experience (DX) ( https://www.adobe.com/experience-cloud.html) is a USD 4B+ business serving the needs of enterprise businesses including 95%+ of fortune 500 organizations. Adobe Marketo Engage, within Adobe DX, the leading marketing automation platform, helps businesses engage customers effectively. It lets enterprises do effective engagement through various surfaces and touchpoints. We are looking for strong and passionate engineers to join our team as we scale the business by building the next gen products and contributing to our existing offerings. If you’re passionate about innovative technology, then we would be excited to talk to you! What You'll Do Collaborate with architects, product management and engineering teams to build solutions that increase the product's value. Develop technical specifications, prototypes and presentations to communicate your ideas. Proficient in emerging industry technologies and trends, and the ability to communicate that knowledge to the team and use it to influence product direction. Exceptional coding skill Write unit tests, ensuring code quality and code coverage. Ensure code is always checked in and ensure source control standards are followed. What you need to succeed 5+ years of experience in software development Expertise in Java, Spring Boot, Rest Services, MySQL or Postgres, MongoDB Good working knowledge of Azure ecosystem, Azure data factory. Good understanding of working with Cassandra, Solr, ElasticSearch, Snowflake Ambitious and not afraid to tackle unknowns, demonstrates a strong bias to action Knowledge in apache spark, scala is added advantage Strong interpersonal, analytical, problem-solving and conflict resolution skills Excellent speaking, writing, and presentation skills, as well as the ability to persuade, encourage, and empower others Bachelors/Masters in Computer Science or a related field. Adobe is proud to be an Equal Employment Opportunity employer. We do not discriminate based on gender, race or color, ethnicity or national origin, age, disability, religion, sexual orientation, gender identity or expression, veteran status, or any other applicable characteristics protected by law. Learn more. Adobe aims to make Adobe.com accessible to any and all users. If you have a disability or special need that requires accommodation to navigate our website or complete the application process, email accommodations@adobe.com or call (408) 536-3015.
Posted 4 days ago
7.0 years
0 Lacs
Hyderabad, Telangana, India
On-site
• 7-9 years of experience with data analytics, data modeling, and database design. • 3+ years of coding and scripting (Python, Java, Scala) and design experience. • 3+ years of experience with Spark framework. • 5+ Experience with ELT methodologies and tools. • 5+ years mastery in designing, developing, tuning and troubleshooting SQL. • Knowledge of Informatica Power center and Informatica IDMC. • Knowledge of distributed, column- orientated technology to create high-performant database technologies like - Vertica, Snowflake. • Strong data analysis skills for extracting insights from financial data • Proficiency in reporting tools (e.g., Power BI, Tableau). The Ideal Qualifications Technical Skills: • Domain knowledge of Investment Management operations including Security Masters, Securities Trade and Recon Operations, Reference data management, and Pricing. • Familiarity with regulatory requirements and compliance standards in the investment management industry. • Experience with IBOR’s such as Blackrock Alladin, CRD, Eagle STAR (ABOR), Eagle Pace, and Eagle DataMart. • Familiarity with investment data platforms such as GoldenSource, FINBOURNE, NeoXam, RIMES, and JPM Fusion. Soft Skills: • Strong analytical and problem-solving abilities. • Exceptional communication and interpersonal skills. • Ability to influence and motivate teams without direct authority. • Excellent time management and organizational skills, with the ability to prioritize multiple initiatives. What to Expect as Part of our Team • Regular meetings with the Corporate Technology leadership team • Focused one-on-one meetings with your manager • Access to mentorship opportunities • Access to learning content on Degreed and other informational platforms
Posted 4 days ago
5.0 - 9.0 years
0 Lacs
pune, maharashtra
On-site
You will be joining our team as a Senior Data Scientist with expertise in Artificial Intelligence (AI) and Machine Learning (ML). The ideal candidate should possess a minimum of 5-7 years of experience in data science, focusing on AI/ML applications. You are expected to have a strong background in various ML algorithms, programming languages such as Python, R, or Scala, and data processing frameworks like Apache Spark. Proficiency in data visualization tools and experience in model deployment using Docker, Kubernetes, and cloud services will be essential for this role. Your responsibilities will include end-to-end AI/ML project delivery, from data processing to model deployment. You should have a good understanding of statistics, probability, and mathematical concepts used in AI/ML. Additionally, familiarity with big data tools, natural language processing techniques, time-series analysis, and MLOps will be advantageous. As a Senior Data Scientist, you are expected to lead cross-functional project teams and manage data science projects in a production setting. Your problem-solving skills, communication skills, and curiosity to stay updated with the latest advancements in AI and ML are crucial for success in this role. You should be able to convey technical insights clearly to diverse audiences and quickly adapt to new technologies. If you are an innovative, analytical, and collaborative team player with a proven track record in AI/ML project delivery, we invite you to apply for this exciting opportunity.,
Posted 4 days ago
4.0 years
0 Lacs
Noida, Uttar Pradesh, India
On-site
Our Company Changing the world through digital experiences is what Adobe’s all about. We give everyone—from emerging artists to global brands—everything they need to design and deliver exceptional digital experiences! We’re passionate about empowering people to create beautiful and powerful images, videos, and apps, and transform how companies interact with customers across every screen. We’re on a mission to hire the very best and are committed to creating exceptional employee experiences where everyone is respected and has access to equal opportunity. We realize that new ideas can come from everywhere in the organization, and we know the next big idea could be yours! What You'll Do This is an individual contributor position. Expectations will be on the below lines: Responsible for design, architecture & implementation of new features. Be responsible for all phases of engineering. From early specs, design/architecture, technology choice, development, unit-testing/integration automation, and deployment. Collaborate with architects, product management and other engineering teams to build the technical vision, and road map for the team. Build technical specifications, prototypes and presentations to communicate your ideas. Be well versed in emerging industry technologies and trends, and have the ability to communicate that knowledge to the team and use it to influence product direction. Orchestrate with team to develop a product or parts of a large product. Requirements B.Tech / M.Tech degree in Computer Science. 4+ years of experience in front end technologies such as React, Node.js etc, along with decent experience in backend programming in Java/Scala. Should have excellent computer science fundamentals and a good understanding of design, and performance of algorithms Knowledge and experience in cloud services - Azure and/or AWS and container orchestration platform - kubernates Good understanding of distributed systems. Decent experience in Java/Scala Programming Strong understanding of RESTful APIs and GraphQL. Proficient in modern JavaScript frameworks like React, Node.js, etc Experience in writing unit, integration, and end-to-end tests. Effective oral and written communication skills, with the ability to interact with customers and cross-functional teams. Excellent work ethics and highly motivated Adobe is proud to be an Equal Employment Opportunity and affirmative action employer. We do not discriminate based on gender, race or color, ethnicity or national origin, age, disability, religion, sexual orientation, gender identity or expression, veteran status, or any other applicable characteristics protected by law. Learn more. Adobe aims to make Adobe.com accessible to any and all users. If you have a disability or special need that requires accommodation to navigate our website or complete the application process, email accommodations@adobe.com or call (408) 536-3015. Adobe values a free and open marketplace for all employees and has policies in place to ensure that we do not enter into illegal agreements with other companies to not recruit or hire each other’s employees.
Posted 4 days ago
7.0 years
0 Lacs
Noida, Uttar Pradesh, India
On-site
Our Company Changing the world through digital experiences is what Adobe’s all about. We give everyone—from emerging artists to global brands—everything they need to design and deliver exceptional digital experiences! We’re passionate about empowering people to create beautiful and powerful images, videos, and apps, and transform how companies interact with customers across every screen. We’re on a mission to hire the very best and are committed to creating exceptional employee experiences where everyone is respected and has access to equal opportunity. We realize that new ideas can come from everywhere in the organization, and we know the next big idea could be yours! Our Company Changing the world through digital experiences is what Adobe’s all about. We give everyone—from emerging artists to global brands—everything they need to design and deliver exceptional digital experiences. We’re passionate about empowering people to craft beautiful and powerful images, videos, and apps, and transform how companies interact with customers across every screen. We’re on a mission to hire the very best and are committed to building exceptional employee experiences where everyone is respected and has access to equal opportunity. We realize that new ideas can come from everywhere in the organization, and we know the next big idea could be yours! Digital Experience (DX) (https://www.adobe.com/experience-cloud.html) is a USD 3B+ business serving the needs of enterprise businesses including 95%+ of fortune 500 organizations. Adobe Journey Optimizer (AJO) within DX provides a platform for designing cross-channel customer experiences and provides an environment for visual campaign orchestration, real time interaction management and cross channel execution. It is built natively on the Adobe Experience Platform and combines a unified, real-time customer profile, an API-first open framework, centralized offer decisioning, and artificial intelligence (AI) and machine learning (ML) for personalization and optimization. Beyond the usual responsibility of designing, developing, documenting, and thoroughly testing code, Computer Scientists @ Adobe would own features of varying complexity, which may require understanding interactions with other parts of the system, moderately sophisticated algorithms and good design judgment. We are looking for strong and passionate engineers to join our team as we scale the business by building the next gen products and contributing to our existing offerings. What You'll Do This is an individual contributor position. Expectations will be on the below lines: Responsible for design and architecture of new products. Work in full DevOps mode, be responsible for all phases of engineering. From early specs, design/architecture, technology choice, development, unit-testing/integration automation, and deployment. Collaborate with architects, product management and other engineering teams to build the technical vision, and road map for the team. Build technical specifications, prototypes and presentations to communicate your ideas. Be well versed in emerging industry technologies and trends, and have the ability to communicate that knowledge to the team and use it to influence product direction. Orchestrate with team to develop a product or parts of a large product. Requirements B.Tech / M.Tech degree in Computer Science from a premier institute. 7-9.5years of relevant experience in software development. Should have excellent computer science fundamentals and a good understanding of design, and performance of algorithms Proficient in Java/Scala Programming Proficient in writing code that is reliable, maintainable, secure, and performant Knowledge of Azure services and/or AWS. Internal Opportunities We’re glad that you’re pursuing career development opportunities at Adobe. Here’s what you’ll need to do: Apply with your complete LinkedIn profile or resume/CV. Schedule a Check-in meeting with your manager to discuss this internal opportunity and your career aspirations. Check-ins should include ongoing discussions about expectations, feedback and career development. Learn more about Check-in here. Learn more about the internal career opportunities process in this FAQ. If you’re contacted for an interview, here are some tips. At Adobe, you will be immersed in an exceptional work environment that is recognized throughout the world on Best Companies lists. You will also be surrounded by colleagues who are committed to helping each other grow through our unique Check-In approach where ongoing feedback flows freely. If you’re looking to make an impact, Adobe's the place for you. Discover what our employees are saying about their career experiences on the Adobe Life blog and explore the meaningful benefits we offer. Adobe is an equal opportunity employer. We welcome and encourage diversity in the workplace regardless of gender, race or color, ethnicity or national origin, age, disability, religion, sexual orientation, gender identity or expression, or veteran status. Adobe is proud to be an Equal Employment Opportunity employer. We do not discriminate based on gender, race or color, ethnicity or national origin, age, disability, religion, sexual orientation, gender identity or expression, veteran status, or any other applicable characteristics protected by law. Learn more. Adobe aims to make Adobe.com accessible to any and all users. If you have a disability or special need that requires accommodation to navigate our website or complete the application process, email accommodations@adobe.com or call (408) 536-3015.
Posted 4 days ago
10.0 years
0 Lacs
Noida, Uttar Pradesh, India
On-site
Our Company Changing the world through digital experiences is what Adobe’s all about. We give everyone—from emerging artists to global brands—everything they need to design and deliver exceptional digital experiences! We’re passionate about empowering people to create beautiful and powerful images, videos, and apps, and transform how companies interact with customers across every screen. We’re on a mission to hire the very best and are committed to creating exceptional employee experiences where everyone is respected and has access to equal opportunity. We realize that new ideas can come from everywhere in the organization, and we know the next big idea could be yours! About Connect Adobe Connect, within Adobe DALP BU is one of the best online webinar and training delivery platform. The product has a huge customer base which has been using it for many years. The product has evolved magnificently over a period of time ensuring it stay on top of the latest tech stack. It offers opportunity to look at plethora of technologies on both client and server side. What You’ll Do: Hands-on Machine Learning Engineer who will release models in production. Develop classifiers, predictive models, and multi-variate optimization algorithms on large-scale datasets using advanced statistical modeling, machine learning, and data mining. Special focus on R&D that will be building predictive models for conversion optimization, Bidding algorithms for pacing & optimization, Reinforcement learning problems, and Forecasting. Collaborate with Product Management to bring AI-based Assistive experiences to life. Socialize what’s possible now or in the near future to inform the roadmap. Responsible for driving all aspects of ML product development: ML modeling, data/ML pipelines, quality evaluations, productization, and ML Ops. Create and instill a team culture that focuses on sound scientific processes and encourages deep engagement with our customers. Handle project scope and risks with data, analytics, and creative problem-solving. What you require: Solid foundation in machine learning, classifiers, statistical modeling and multivariate optimization techniques Experience with control systems, reinforcement learning problems, and contextual bandit algos. Experience with DNN frameworks like TensorFlow or PyTorch on large-scale data sets. TensorFlow, R, scikit, pandas Proficient in one or more: Python, Java/Scala, SQL, Hive, Spark Good to have - Git, Docker, Kubernetes GenAI, RAG pipelines a must have technology Cloud based solutions is good to have General understanding of data structures, algorithms, multi-threaded programming, and distributed computing concepts Ability to be a self-starter and work closely with other data scientists and software engineers to design, test, and build production-ready ML and optimization models and distributed algorithms running on large-scale data sets. Ideal Candidate Profile: A total of 10+ years of experience, including at least 5 years in technical roles involving Data Science, Machine Learning, or Statistics. Masters or B.Tech in Computer Science/ Statistics Comfort with ambiguity, adaptability to evolving priorities, and the ability to lead a team while working autonomously. Proven management experience with highly diverse and global teams. Demonstrated ability to influence technical and non-technical stakeholders. Proven ability to effectively manage in a high-growth, matrixed organization. Track record of delivering cloud-scale, data-driven products, and services that are widely adopted with large customer bases. An ability to think strategically, look around corners, and create a vision for the current quarter, the year, and five years down the road. A relentless pursuit of great customer experiences and continuous improvements to the product. Adobe is proud to be an Equal Employment Opportunity employer. We do not discriminate based on gender, race or color, ethnicity or national origin, age, disability, religion, sexual orientation, gender identity or expression, veteran status, or any other applicable characteristics protected by law. Learn more. Adobe aims to make Adobe.com accessible to any and all users. If you have a disability or special need that requires accommodation to navigate our website or complete the application process, email accommodations@adobe.com or call (408) 536-3015.
Posted 4 days ago
5.0 - 7.0 years
0 Lacs
India
On-site
Shift Ahead Technologies, based in Pune required a couple of Senior Engineers (5 to 7 years) experience in SCALA development , self-sufficient, that can work autonomously under customer supervision. This role shall be work from home. Should be willing to join in 7 odd days. Desirably excellent english communication and can work independently with client. Design, develop, and maintain Scala-based applications and software solutions. Write clean, efficient, and scalable code following functional programming principles and best practices Participate in architectural decisions and contribute to the design and development process of projects. Test, debug, and optimize applications to ensure high performance, security, and scalability Collaborate with cross-functional teams including developers, analysts, QA engineers, and stakeholders throughout the development cycle Collaborate with cross-functional teams including developers, analysts, QA engineers, and stakeholders throughout the development cycle Integrate Scala solutions with other platforms, frameworks (such as Akka, Play, or Spark), and APIs for data or service integration Confident candidates may apply or mail to careers@shiftahead.tech
Posted 4 days ago
3.0 - 7.0 years
0 Lacs
kolkata, west bengal
On-site
Genpact (NYSE: G) is a global professional services and solutions firm committed to delivering outcomes that shape the future. With over 125,000 employees spread across more than 30 countries, we are fueled by our innate curiosity, entrepreneurial agility, and the aspiration to create lasting value for our clients. Driven by our purpose - the relentless pursuit of a world that works better for people - we cater to and transform leading enterprises, including the Fortune Global 500, leveraging our profound business and industry expertise, digital operations services, and proficiency in data, technology, and AI. We are currently seeking applications for the position of Lead Consultant-Data Bricks Senior Engineer! As a Lead Consultant-Data Bricks Senior Engineer, your responsibilities will include working closely with Software Designers to ensure adherence to best practices, providing suggestions for enhancing code proficiency and maintainability, occasional customer interaction to analyze user needs and determine technical requirements, designing, building, and maintaining scalable and reliable data pipelines using DataBricks, developing high-quality code focusing on performance, scalability, and security, collaborating with cross-functional teams to comprehend data requirements and deliver solutions aligning with business needs, implementing data transformations and intricate algorithms within the DataBricks environment, optimizing data processing and refining data architecture to enhance system efficiency and data quality, mentoring junior engineers, and contributing to the establishment of best practices within the team. Additionally, staying updated with emerging trends and technologies in data engineering and cloud computing is imperative. Qualifications we are looking for: Minimum Qualifications: - Experience in data engineering or a related field - Strong hands-on experience with DataBricks, encompassing development of code, pipelines, and data transformations - Proficiency in at least one programming language (e.g., Python, Scala, Java) - In-depth knowledge of Apache Spark and its integration within DataBricks - Experience with cloud services (AWS, Azure, or GCP) and their data-related products - Familiarity with CI/CD practices, version control (Git), and automated testing - Exceptional problem-solving abilities with the capacity to work both independently and as part of a team - Bachelor's degree in computer science, Engineering, Mathematics, or a related technical field If you are enthusiastic about leveraging your skills and expertise as a Lead Consultant-Data Bricks Senior Engineer, join us at Genpact and be a part of shaping a better future for all. Location: India-Kolkata Schedule: Full-time Education Level: Bachelor's / Graduation / Equivalent Job Posting: Jul 30, 2024, 5:05:42 AM Unposting Date: Jan 25, 2025, 11:35:42 PM,
Posted 4 days ago
5.0 years
0 Lacs
Ahmedabad, Gujarat, India
On-site
Job Profile: Data Engineer Experience: 5+ Years Who we are: Innovatics is a place where innovation blends with analytics. We, Innovatics, take pride in knowing the notion of bleeding-edge technologies, strategic business moves, and radiant business transformation. We deliver never thought before business growth opportunities and assist businesses to accelerate their digital transformation journey. About the role: We're looking for a Data Engineer who's passionate about delivering tangible results, who has a positive attitude, and who enjoys solving problems. Requirements: Technical Skills: 3+ years of experience in a Data Engineer role, Experience with object-oriented/object function scripting languages: Python, Scala, Golang, Java, etc. Experience with Big data tools such as Spark, Hadoop/ Kafka/ Airflow/Hive Experience with Streaming data: Spark/Kinesis/Kafka/Pubsub/Event Hub Experience with GCP/Azure data factory/AWS Strong in SQL Scripting Experience with ETL tools Knowledge of Snowflake Data Warehouse Knowledge of Orchestration frameworks: Airflow/Luig Good to have knowledge of Data Quality Management frameworks Good to have knowledge of Master Data Management Self-learning abilities are a must Familiarity with upcoming new technologies is a strong plus. Should have a bachelor's degree in big data analytics, computer engineering, or a related field Personal Competency: Strong communication skills is a MUST Self-motivated, detail-oriented Strong organizational skills Ability to prioritize workloads and meet deadlines
Posted 4 days ago
2.0 - 9.0 years
0 Lacs
karnataka
On-site
We are seeking a Data Architect / Sr. Data and Pr. Data Architects to join our team. In this role, you will be involved in a combination of hands-on contribution, customer engagement, and technical team management. As a Data Architect, your responsibilities will include designing, architecting, deploying, and maintaining solutions on the MS Azure platform using various Cloud & Big Data Technologies. You will be managing the full life-cycle of Data Lake / Big Data solutions, starting from requirement gathering and analysis to platform selection, architecture design, and deployment. It will be your responsibility to implement scalable solutions on the Cloud and collaborate with a team of business domain experts, data scientists, and application developers to develop Big Data solutions. Moreover, you will be expected to explore and learn new technologies for creative problem solving and mentor a team of Data Engineers. The ideal candidate should possess strong hands-on experience in implementing Data Lake with technologies such as Data Factory (ADF), ADLS, Databricks, Azure Synapse Analytics, Event Hub & Streaming Analytics, Cosmos DB, and Purview. Additionally, experience with big data technologies like Hadoop (CDH or HDP), Spark, Airflow, NiFi, Kafka, Hive, HBase, MongoDB, Neo4J, Elastic Search, Impala, Sqoop, etc., is required. Proficiency in programming and debugging skills in Python and Scala/Java is essential, with experience in building REST services considered beneficial. Candidates should also have experience in supporting BI and Data Science teams in consuming data in a secure and governed manner, along with a good understanding of using CI/CD with Git, Jenkins / Azure DevOps. Experience in setting up cloud-computing infrastructure solutions, hands-on experience/exposure to NoSQL Databases, and Data Modelling in Hive are all highly valued. Applicants should have a minimum of 9 years of technical experience, with at least 5 years on MS Azure and 2 years on Hadoop (CDH/HDP).,
Posted 4 days ago
4.0 - 8.0 years
0 Lacs
hyderabad, telangana
On-site
As an Automation QA Engineer (Python & Scala) at our global IT leader client based in Hyderabad, you will be responsible for designing and developing automation testing frameworks using web and object-oriented technologies. With a focus on ensuring product quality, you will conduct functional and automation QA processes. Your expertise in Python or Java programming will be crucial in leveraging strong object-oriented programming skills to achieve project goals. Collaboration with team members to meet specific timelines is essential for success in this role. Ideally, you should possess 4 to 6 years of experience, with at least 2 years in automation or development being preferred. The offered salary for this position is 30 LPA, and the notice period required is immediate to 15 days. The option for relocation is available for interested candidates. Candidates with exposure to Big Data processing systems, Hadoop, or Scala will have an added advantage. Strong analytical and testing skills are essential for this role to excel in the fast-paced environment of our client's IT consulting, business process management, and digital transformation services.,
Posted 4 days ago
5.0 - 9.0 years
0 Lacs
hyderabad, telangana
On-site
You will be responsible for developing ETL processes using Informatica for data integration from various sources. This includes creating specifications, test scripts, and ensuring code coverage for all integrations. You will also support the migration of integration code from lower to higher environments, such as production. Your experience in working with XML and JSON for real-time integrations will be crucial for this role. Additionally, you should have expertise in performing full and incremental ETL using Informatica Power Center. Experience with AWS Cloud services and working with iPaaS for integration configurations is required. It is essential to have a strong background in developing ETL processes for Data Warehouse Integration and supporting integration configurations through connected apps or web services. Familiarity with reporting tools, especially MicroStrategy, is preferred. You should also have experience in production support and be willing to be on-call during selected off-shift hours. Experience with Agile framework is necessary for this position. Additionally, knowledge of Python for data extraction and manipulation, AWS terraform, New Relic setup and maintenance, GIT, Rally, and Scala would be advantageous.,
Posted 4 days ago
4.0 years
0 Lacs
Pune, Maharashtra
Remote
Job Description What you will do: The Zendesk Core Services Packaging and Consumption team is looking for a Senior Software Engineer - Backend for a project that drives successful feature adoption for Zendesk customers. The ideal candidate will have experience with analysing various data sources with good SQL skills, a good understanding of domain driven design and the willingness to explore the unknowns.  On a day-to-day basis, a strong command over one of the backend languages like Scala or Java is highly beneficial. Past experience with developing on Rails Framework will be good as well. Your responsibilities will include: Collaborating with product management, architecture, engineers (front end and back end) to design beautifully simple solutions to complicated problems. You will be relied on from concept through development, QA, staging, and production deployment. Ensuring delivery on commitments. It is your responsibility to ensure code quality, debug code, and seek guidance to unblock tasks pending. Following best practices in all our frameworks and tools. Championing best practices and proper test coverage. We ship code frequently and fast, but stability and reliability must never be compromised. Actively participate in code reviews and design discussions Partner across all areas of the SDLC, including requirements gathering, requirements analysis, building services and solutions. Work across teams and organization boundaries and timezones to standardize and integrate services, tools, and workflows What you bring to the role:  4+ years of relevant experience in at least one object oriented language like Scala or Java (Scala preferred and should have hands on experience) Experience with databases like MySQL and/or DynamoDB  Analytical mindset, good articulation skills and pragmatic approach to problem solving. Experience with CI/CD and delivery systems (Github Actions, Jenkins) Knowledge of API design, distributed systems and Kafka Experience using Datadog or other log aggregation tools Customer-first mentality when dealing with service incident management, data analysis, and root-cause analysis A hunger for learning new technologies and an eagerness to grow your knowledge and capabilities A team-first collaborative attitude that thrives in a fast-moving agile development environment Excellent written and verbal communication skills Bonus Skills: Experience with JavaScript/Typescript Experience working on SaaS-based products Experience with AWS stack (ie. Aurora), Datawarehouse technologies like Snowflake Experience with Ruby on Rails. Please note that Zendesk can only hire candidates who are physically located and plan to work from Karnataka or Maharashtra. Please refer to the location posted on the requisition for where this role is based. Hybrid: In this role, our hybrid experience is designed at the team level to give you a rich onsite experience packed with connection, collaboration, learning, and celebration - while also giving you flexibility to work remotely for part of the week. This role must attend our local office for part of the week. The specific in-office schedule is to be determined by the hiring manager. The intelligent heart of customer experience Zendesk software was built to bring a sense of calm to the chaotic world of customer service. Today we power billions of conversations with brands you know and love. Zendesk believes in offering our people a fulfilling and inclusive experience. Our hybrid way of working, enables us to purposefully come together in person, at one of our many Zendesk offices around the world, to connect, collaborate and learn whilst also giving our people the flexibility to work remotely for part of the week. Zendesk is an equal opportunity employer, and we’re proud of our ongoing efforts to foster global diversity, equity, & inclusion in the workplace. Individuals seeking employment and employees at Zendesk are considered without regard to race, color, religion, national origin, age, sex, gender, gender identity, gender expression, sexual orientation, marital status, medical condition, ancestry, disability, military or veteran status, or any other characteristic protected by applicable law. We are an AA/EEO/Veterans/Disabled employer. If you are based in the United States and would like more information about your EEO rights under the law, please click here . Zendesk endeavors to make reasonable accommodations for applicants with disabilities and disabled veterans pursuant to applicable federal and state law. If you are an individual with a disability and require a reasonable accommodation to submit this application, complete any pre-employment testing, or otherwise participate in the employee selection process, please send an e-mail to peopleandplaces@zendesk.com with your specific accommodation request.
Posted 4 days ago
5.0 - 9.0 years
0 Lacs
haryana
On-site
The role of Lead, Software Engineer at Mastercard involves playing a crucial part in the Data Unification process across different data assets to create a unified view of data from multiple sources. This position will focus on driving insights from available data sets and supporting the development of new data-driven cyber products, services, and actionable insights. The Lead, Software Engineer will collaborate with various teams such as Product Manager, Data Science, Platform Strategy, and Technology to understand data needs and requirements for delivering data solutions that bring business value. Key responsibilities of the Lead, Software Engineer include performing data ingestion, aggregation, and processing to derive relevant insights, manipulating and analyzing complex data from various sources, identifying innovative ideas and delivering proof of concepts, prototypes, and proposing new products and enhancements. Moreover, integrating and unifying new data assets to enhance customer value, analyzing transaction and product data to generate actionable recommendations for business growth, and collecting feedback from clients, development, product, and sales teams for new solutions are also part of the role. The ideal candidate for this position should have a good understanding of streaming technologies like Kafka and Spark Streaming, proficiency in programming languages such as Java, Scala, or Python, experience with Enterprise Business Intelligence Platform/Data platform, strong SQL and higher-level programming skills, knowledge of data mining and machine learning algorithms, and familiarity with data integration tools like ETL/ELT tools including Apache NiFi, Azure Data Factory, Pentaho, and Talend. Additionally, they should possess the ability to work in a fast-paced, deadline-driven environment, collaborate effectively with cross-functional teams, and articulate solution requirements for different groups within the organization. It is essential for all employees working at or on behalf of Mastercard to adhere to the organization's security policies and practices, ensure the confidentiality and integrity of accessed information, report any suspected information security violations or breaches, and complete all mandatory security trainings in accordance with Mastercard's guidelines. The Lead, Software Engineer role at Mastercard offers an exciting opportunity to contribute to the development of innovative data-driven solutions that drive business growth and enhance customer value proposition.,
Posted 4 days ago
5.0 - 9.0 years
0 Lacs
karnataka
On-site
We are seeking experienced and talented engineers to join our team. Your main responsibilities will include designing, building, and maintaining the software that drives the global logistics industry. WiseTech Global is a leading provider of software for the logistics sector, facilitating connectivity for major companies like DHL and FedEx within their supply chains. Our organization is product and engineer-focused, with a strong commitment to enhancing the functionality and quality of our software through continuous innovation. Our primary Research and Development center in Bangalore plays a pivotal role in our growth strategies and product development roadmap. As a Lead Software Engineer, you will serve as a mentor, a leader, and an expert in your field. You should be adept at effective communication with senior management while also being hands-on with the code to deliver effective solutions. The technical environment you will work in includes technologies such as C#, Java, C++, Python, Scala, Spring, Spring Boot, Apache Spark, Hadoop, Hive, Delta Lake, Kafka, Debezium, GKE (Kubernetes Engine), Composer (Airflow), DataProc, DataStreams, DataFlow, MySQL RDBMS, MongoDB NoSQL (Atlas), UIPath, Helm, Flyway, Sterling, EDI, Redis, Elastic Search, Grafana Dashboard, and Docker. Before applying, please note that WiseTech Global may engage external service providers to assess applications. By submitting your application and personal information, you agree to WiseTech Global sharing this data with external service providers who will handle it confidentially in compliance with privacy and data protection laws.,
Posted 4 days ago
3.0 - 7.0 years
0 Lacs
chennai, tamil nadu
On-site
Cloud Kinetics is seeking a candidate with expertise in Bigdata, Hadoop, Hive SQLs, Spark, and other tools within the Bigdata Eco System. As a member of our team, you will be responsible for developing code, optimizing queries for performance, setting up environments, ensuring connectivity, and deploying code into production post-testing. Strong functional and technical knowledge is essential to fulfill project requirements, particularly in the context of Banking terminologies. Additionally, you may lead small to medium-sized projects and act as the primary contact for related tasks. Proficiency in DevOps and Agile Development Framework is crucial for this role. In addition to the core requirements, familiarity with Cloud computing, particularly AWS or Azure Cloud Services, is advantageous. The ideal candidate will possess strong problem-solving skills, adaptability to ambiguity, and a quick grasp of new and complex concepts. Experience in collaborating with teams within complex organizational structures is preferred. Knowledge of BI tools like MSTR and Tableau, as well as a solid understanding of object-oriented programming and HDFS concepts, will be beneficial. As a member of the team, your responsibilities will include working as a developer in Bigdata, Hadoop, or Data Warehousing Tools, and Cloud Computing. This entails working on Hadoop, Hive SQLs, Spark, and other tools within the Bigdata Eco System. Furthermore, you will create Scala/Spark jobs for data transformation and aggregation, develop unit tests for Spark transformations and helper methods, and design data processing pipelines to streamline operations. If you are a proactive individual with a strong technical background and a passion for leveraging cutting-edge technologies to drive innovation, we encourage you to apply for this exciting opportunity at Cloud Kinetics.,
Posted 4 days ago
5.0 - 9.0 years
0 Lacs
hyderabad, telangana
On-site
As a Big Data Engineer with over 5 years of experience, you will be responsible for leading a team of engineers and demonstrating a positive attitude towards learning and implementing solutions in the domain. Your expertise will primarily lie in Spark, with proficiency in Scala/Java, Airflow Orchestration, and AWS. Your role will involve defining the system's scope and delivering effective Big Data solutions, while also collaborating with software research and development teams. In this position, you will be required to train staff on data resource management, utilizing your strong educational background with an Engineering or Master's degree in computer engineering or computer science. Your in-depth knowledge of Hadoop, Spark, and similar frameworks will be essential in driving innovative solutions. Moreover, your excellent interpersonal and communication skills will facilitate effective collaboration within the team. Your ability to solve complex networking, data, and software issues will be crucial in ensuring the successful implementation of Big Data solutions.,
Posted 4 days ago
8.0 years
0 Lacs
Greater Kolkata Area
Remote
AI / Generative AI Engineer Location: Remote ( Pan India ). Job Type: Fulltime. NOTE: "Only immediate joiners or candidates with a notice period of 15 days or less will be We are seeking a highly skilled and motivated AI/Generative AI Engineer to join our innovative team. The ideal candidate will have a strong background in designing, developing, and deploying artificial intelligence and machine learning models, with a specific focus on cutting-edge Generative AI technologies. This role requires hands-on experience with one or more major cloud platforms (Google Cloud Platform GCP, Amazon Web Services AWS) and/or modern data platforms (Databricks, Snowflake). You will be instrumental in building and scaling AI solutions that drive business value and transform user experiences. Key Responsibilities Design and Development : Design, build, train, and deploy scalable and robust AI/ML models, including traditional machine learning algorithms and advanced Generative AI models (e.g., Large Language Models LLMs, diffusion models). Develop and implement algorithms for tasks such as natural language processing (NLP), text generation, image synthesis, speech recognition, and forecasting. Work extensively with LLMs, including fine-tuning, prompt engineering, retrieval-augmented generation (RAG), and evaluating their performance. Develop and manage data pipelines for data ingestion, preprocessing, feature engineering, and model training, ensuring data quality and integrity. Platform Expertise Leverage cloud AI/ML services on GCP (e.g., Vertex AI, AutoML, BigQuery ML, Model Garden, Gemini), AWS (e.g., SageMaker, Bedrock, S3), Databricks, and/or Snowflake to build and deploy solutions. Architect and implement AI solutions ensuring scalability, reliability, security, and cost-effectiveness on the chosen platform(s). Optimize data storage, processing, and model serving components within the cloud or data platform ecosystem. MLOps And Productionization Implement MLOps best practices for model versioning, continuous integration/continuous deployment (CI/CD), monitoring, and lifecycle management. Deploy models into production environments and ensure their performance, scalability, and reliability. Monitor and optimize the performance of AI models in production, addressing issues related to accuracy, speed, and resource utilization. Collaboration And Innovation Collaborate closely with data scientists, software engineers, product managers, and business stakeholders to understand requirements, define solutions, and integrate AI capabilities into applications and workflows. Stay current with the latest advancements in AI, Generative AI, machine learning, and relevant cloud/data platform technologies. Lead and participate in the ideation and prototyping of new AI applications and systems. Ensure AI solutions adhere to ethical standards, responsible AI principles, and regulatory requirements, addressing issues like data privacy, bias, and fairness. Documentation And Communication Create and maintain comprehensive technical documentation for AI models, systems, and processes. Effectively communicate complex AI concepts and results to both technical and non-technical audiences. Required Qualifications 8+ years of experience with software development in one or more programming languages, and with data structures/algorithms/Data Architecture. 3+ years of experience with state of the art GenAI techniques (e.g., LLMs, Multi-Modal, Large Vision Models) or with GenAI-related concepts (language modeling, computer vision). 3+ years of experience with ML infrastructure (e.g., model deployment, model evaluation, optimization, data processing, debugging). Bachelor's or Master's degree in Computer Science, Artificial Intelligence, Machine Learning, Data Science, or a related technical field. Proven experience as an AI Engineer, Machine Learning Engineer, or a similar role. Strong programming skills in Python. Familiarity with other languages like Java, Scala, or R is a plus. Solid understanding of machine learning algorithms (supervised, unsupervised, reinforcement learning), deep learning concepts (e.g., CNNs, RNNs, Transformers), and statistical modeling. Hands-on experience with developing and deploying Generative AI models and techniques, including working with Large Language Models (LLMs like GPT, BERT, LLaMA, etc.). Proficiency in using common AI/ML frameworks and libraries such as TensorFlow, PyTorch, scikit-learn, Keras, Hugging Face Transformers, LangChain, etc. Demonstrable experience with at least one of the following cloud/data platforms: GCP: Experience with Vertex AI, BigQuery ML, Google Cloud Storage, and other GCP AI/ML services. AWS: Experience with SageMaker, Bedrock, S3, and other AWS AI/ML services. Databricks: Experience building and scaling AI/ML solutions on the Databricks Lakehouse Platform, including MLflow. Snowflake: Experience leveraging Snowflake for data warehousing, data engineering for AI/ML workloads, and Snowpark. Experience with data engineering, including data acquisition, cleaning, transformation, and building ETL/ELT pipelines. Knowledge of MLOps tools and practices for model deployment, monitoring, and management. Familiarity with containerization technologies like Docker and orchestration tools like Kubernetes. Strong analytical and problem-solving skills. Excellent communication and collaboration abilities. (ref:hirist.tech)
Posted 4 days ago
3.0 - 7.0 years
0 Lacs
karnataka
On-site
As a Data Engineer, you will be responsible for designing, developing, and delivering ADF pipelines for the Accounting & Reporting Stream. Your role will involve creating and maintaining scalable data pipelines using PySpark and ETL workflows in Azure Databricks and Azure Data Factory. You will also work on data modeling and architecture to optimize data structures for analytics and business requirements. Your responsibilities will include monitoring, tuning, and troubleshooting pipeline performance for efficiency and reliability. Collaboration with business analysts and stakeholders is key to understanding data needs and delivering actionable insights. Implementing data governance practices to ensure data quality, security, and compliance with regulations is essential. You will also be required to develop and maintain documentation for data pipelines and architecture. Experience in testing and test automation is necessary for this role. Collaboration with cross-functional teams to comprehend data requirements and provide technical advice is crucial. Strong background in data engineering is required, with proficiency in SQL, Azure Databricks, Blob Storage, Azure Data Factory, and programming languages like Python or Scala. Knowledge of Logic App and Key Vault is also necessary. Strong problem-solving skills and the ability to communicate complex technical concepts to non-technical stakeholders are essential for effective communication within the team.,
Posted 4 days ago
4.0 - 8.0 years
0 Lacs
pune, maharashtra
On-site
YASH Technologies is a leading technology integrator specializing in helping clients reimagine operating models, enhance competitiveness, optimize costs, foster exceptional stakeholder experiences, and drive business transformation. At YASH, you will be part of a team of innovative professionals working with cutting-edge technologies. Our purpose is anchored in bringing real positive changes in an increasingly virtual world, transcending generational gaps and future disruptions. We are currently seeking SQL Professionals for the role of Data Engineer with 4-6 years of experience. The ideal candidate must have a strong academic background. As a Data Engineer at BNY Mellon in Pune, you will be responsible for designing, developing, and maintaining scalable data pipelines and ETL processes using Apache Spark and SQL. You will collaborate with data scientists and analysts to understand data requirements, optimize and query large datasets, ensure data quality and integrity, implement data governance and security best practices, participate in code reviews, and troubleshoot data-related issues promptly. Qualifications for this role include 4-6 years of experience in data engineering, proficiency in SQL and data processing frameworks like Apache Spark, knowledge of database technologies such as SQL Server or Oracle, experience with cloud platforms like AWS, Azure, or Google Cloud, familiarity with data warehousing solutions, understanding of Python, Scala, or Java for data manipulation, excellent analytical and problem-solving skills, and good communication skills to work effectively in a team environment. Joining YASH means being empowered to shape your career in an inclusive team environment. We offer career-oriented skilling models and promote continuous learning, unlearning, and relearning at a rapid pace. Our workplace is based on four principles: flexible work arrangements, free spirit, and emotional positivity; agile self-determination, trust, transparency, and open collaboration; all support needed for the realization of business goals; and stable employment with a great atmosphere and ethical corporate culture.,
Posted 4 days ago
3.0 - 7.0 years
0 Lacs
hyderabad, telangana
On-site
As an AWS Data Engineer, you should have at least 3 years of experience in AWS Data Engineering. Your main responsibilities will include designing and building ETL pipelines and Data lakes to automate the ingestion of both structured and unstructured data. You will need to be proficient in working with AWS big data technologies such as Redshift, S3, AWS Glue, Kinesis, Athena, DMS, EMR, and Lambda for Serverless ETL processes. Knowledge in SQL and NoSQL programming languages is essential, along with experience in batch and real-time pipelines. Your role will require excellent programming and debugging skills in either Scala or Python, as well as expertise in Spark. You should have a good understanding of Data Lake formation, Apache Spark, Python, and hands-on experience in deploying models. Experience in Production migration processes is a must, and it would be advantageous to have familiarity with Power BI visualization tools and connectivity. In this position, you will be tasked with designing, building, and operationalizing large-scale enterprise data solutions and applications. You will also need to analyze, re-architect, and re-platform on-premise data warehouses to data platforms within the AWS cloud environment. Creating production data pipelines from ingestion to consumption using Python or Scala within the AWS big data architecture will be part of your routine. Additionally, you will be responsible for conducting detailed assessments of current state data platforms and developing suitable transition paths to the AWS cloud. If you possess strong data engineering skills and are looking for a challenging role in AWS Data Engineering, this opportunity may be the right fit for you.,
Posted 4 days ago
4.0 - 12.0 years
0 Lacs
hyderabad, telangana
On-site
At Medtronic, you can embark on a life-long career focused on exploration and innovation, all while advocating for healthcare access and equity for everyone. You will play a vital role in fostering a more connected and compassionate world through purpose-driven leadership. As a key member and technical leader in the field of medical device cybersecurity, you will be at the forefront of creating, deploying, and monitoring cybersecurity and information security solutions for Medtronic's medical devices and supporting IT infrastructure. Your responsibilities will include collaborating with external and internal cybersecurity researchers to identify and address vulnerabilities in Medtronic products and systems. Additionally, you will work closely with R&D teams to ensure comprehensive security risk assessments are conducted and appropriate solutions are implemented. You will also be responsible for developing project security management deliverables to comply with regulatory standards and effectively communicate cybersecurity technology to various stakeholders. Your duties may involve, but are not limited to: - Leading and executing cybersecurity-related activities related to products and devices, such as incident response, vulnerability assessments, and mitigation implementation. - Conducting product-level intrusion detection activities. - Performing product risk assessments in coordination with R&D teams and recommending specific security controls. - Participating in the development and testing of product security-related requirements and processes. - Managing security-related deliverables for regulatory bodies to ensure compliance with standards. - Evaluating and testing security risks throughout the development lifecycle. - Supporting emerging cybersecurity certification initiatives. - Maintaining and updating security documentation. - Creating and managing threat models using STRIDE. Requirements: - Bachelor's or graduate degree in computer science, computer engineering, electrical engineering, or a related field. - CISSP or similar certification, or equivalent demonstrated experience. - Experience in embedded devices vulnerability assessment, threat modeling, and risk scoring. - Formal education in cybersecurity and information assurance. - Minimum of 12 years of experience with at least 4 years in technical, cybersecurity-related roles. - Proficiency in security posture analysis, vulnerability assessment, penetration testing, and static code analysis. - Software product development experience and programming skills in languages such as C, C++, Python, Java, .NET, Go, Ruby, or Scala. - Understanding of national and international laws, regulations, and policies related to regulated medical device cybersecurity. - Familiarity with information security practices, risk management processes, cybersecurity principles, and incident response methodologies. In addition to a competitive salary, Medtronic offers a flexible benefits package that supports employees at every stage of their career and life. The company is committed to recognizing and rewarding employee contributions while providing a wide range of resources and compensation plans. Medtronic is a global leader in healthcare technology dedicated to addressing the most pressing health challenges worldwide. The company's mission of alleviating pain, restoring health, and extending life unites a diverse team of over 90,000 passionate individuals. Medtronic's commitment to diversity and innovation drives the team to engineer real solutions for real people, from the R&D lab to the factory floor and beyond.,
Posted 4 days ago
0 years
0 Lacs
Pune, Maharashtra, India
On-site
Position Overview Job Title: Full Stack Developer with Java, SQL, React, Python Location: Pune, India Corporate Title: VP Role Description Technology underpins our entire business. Our Technology, Data and Innovation (TDI) strategy is focused on strengthening engineering expertise, introducing an agile delivery model, as well as modernising the bank's IT infrastructure. We continue to invest and build a team of visionary tech talent, providing you with the training, freedom and opportunity to do pioneering work. As an [Engineer] you will develop and deliver significant components of engineering solutions to satisfy complex and diverse business goals. You will engage and partner with the business whilst working within a broader creative, collaborative and innovative team, with a strong desire to make an impact. You will be joining the dbSleuth Team within Regulatory & Cross Product IT delivering Trader and Counterparty surveillance across all business sections of Deutsche Bank. We are an engineering focused organization, striving for the highest quality architecture, design and code across our teams. You will help to build our surveillance systems, working in a fast-paced, agile environment. Our workload for new deliveries is high, using, React for UI development, Python/Spark/Scala for services, Hadoop Big Data and data science for anomaly detection using machine learning and statistical risk models. What We’ll Offer You As part of our flexible scheme, here are just some of the benefits that you’ll enjoy Best in class leave policy Gender neutral parental leaves 100% reimbursement under childcare assistance benefit (gender neutral) Sponsorship for Industry relevant certifications and education Employee Assistance Program for you and your family members Comprehensive Hospitalization Insurance for you and your dependents Accident and Term life Insurance Complementary Health screening for 35 yrs. and above Your Key Responsibilities Provide leadership within a delivery team, Modelling, Coding & testing, and collaborating to understand requirements, create stories, design solutions, implement them and help test them. Help create a culture of learning and continuous improvement within your team and be accountable for successful delivery of a regulatory critical workstream Employ a range of techniques to analyse problems and evaluate multiple solutions against engineering, business & strategic criteria Identify and resolve barriers to business deliveries implementing solutions which iteratively deliver value Design solutions using common design patterns with a range of design tools & techniques Conduct peer reviews to ensure designs are fit for purpose, extensible & re-usable Design & build solutions which are secure & controlled Your Skills And Experience Analytical thinker, team player and possess strong communication skills Enable experimentation and fast learning approaches to creating business solutions Familiar in the use of solution design tools Understand key elements of security, risk & control Track record in identifying and making improvements to the delivery process Working with very large datasets using technologies such as Python, React JS and SQL and utilizing a good understanding of UI functioning & infrastructure. Utilizing Data Modelling tools, Domain Driven design and a strong knowledge of SQL and advanced data analysis to deliver good quality code within enterprise scale development (CI/CD) Experience with development utilising SDLC tools - Git, JIRA, Artifactory, Jenkins/TeamCity, OpenShift How We’ll Support You Training and development to help you excel in your career Coaching and support from experts in your team A culture of continuous learning to aid progression A range of flexible benefits that you can tailor to suit your needs About Us And Our Teams Please visit our company website for further information: https://www.db.com/company/company.htm We strive for a culture in which we are empowered to excel together every day. This includes acting responsibly, thinking commercially, taking initiative and working collaboratively. Together we share and celebrate the successes of our people. Together we are Deutsche Bank Group. We welcome applications from all people and promote a positive, fair and inclusive work environment.
Posted 4 days ago
5.0 - 9.0 years
0 Lacs
hyderabad, telangana
On-site
We are seeking multiple Data Engineers with extensive expertise in Databricks, Kafka, Python, Spark, Scala, and more for a hybrid position based in either our Hyderabad or Chennai office. This is a full-time/permanent opportunity. If you or someone you know is interested, please reach out! The ideal candidate should have experience in real-time Customer Data/service Analytics. Requirements: - Minimum experience level of 5+ years - Proficiency in Databricks, Kafka, Python, Spark, Scala, etc. - Strong background in real-time Customer Data/service Analytics For any inquiries, please contact Alvin P. at alvinp@infoservices.com or call at +91-8977025597.,
Posted 4 days ago
4.0 - 8.0 years
0 - 0 Lacs
guwahati, assam
On-site
As a Senior Data Engineer specialized in Scala, you will be responsible for leading Spark 3.X, SCALA, Delta lake implementation, and streaming solution implementation for IOT in Spark streaming. Your expertise in Kafka is essential for this role. Any prior experience with MFG BI, DWH, and Datalake implementation will be considered a bonus. This position offers the flexibility of working from home in India and requires a total experience of 10+ years with at least 4-5 years in Scala. The role is permanent under us with an annual salary range of INR 20-25 LPA. The notice period for this position is immediate to 30 days, and the interview process consists of 2 or 3 rounds. Key Responsibilities: - Understand the factories, manufacturing process, data availability, and avenues for improvement. - Collaborate with engineering, manufacturing, and quality teams to identify problems solvable using the acquired data in the data lake platform. - Define necessary data and collaborate with connectivity engineers and users to collect the data. - Develop and maintain optimal data pipeline architecture. - Assemble large, complex datasets meeting functional and non-functional business requirements. - Identify, design, and implement process improvements, automate manual processes, and optimize data delivery for scalability. - Work on data preparation, data deep dive, and help engineering, process, and quality teams understand process/machine behavior closely using available data. - Deploy and monitor solutions. - Collaborate with data and analytics experts to enhance functionality in data systems. - Work alongside Data Architects and data modeling teams. Skills / Competencies: - Solid knowledge of the business vertical with experience in solving use cases in manufacturing or similar industries. - Ability to apply cross-industry learning to enhance manufacturing processes. - Strong problem scoping, solving, and quantification skills. - Proficient in working with unstructured datasets and building data transformation processes. - Experience with message queuing, stream processing, and scalable big data data stores. - Skilled in data mining and data wrangling techniques for analytical dataset creation. - Proficient in building and optimizing big data pipelines, architectures, and datasets. - Adaptive mindset to address data challenges and drive desired outcomes. - Experience with Spark, Delta, CDC, NiFi, Kafka, relational SQL, NoSQL databases, and query languages. - Proficiency in object-oriented languages such as Scala, Java, C++. - Knowledge of visualization tools like PowerBI, Tableau for data presentation. - Ability to analyze data, generate findings, insights through exploratory data analysis. - Strong understanding of data transformation and connection across various data types. - Proficient in numerical, analytical skills, and identifying data acquisition opportunities. - Experience in enhancing data quality, reliability, building algorithms, and prototypes. - Ability to optimize existing frameworks for better performance. If you have the requisite expertise in Scala, Spark, and data engineering, and are keen to work on cutting-edge solutions for manufacturing processes, this role offers an exciting opportunity to make a significant impact in the domain.,
Posted 4 days ago
Upload Resume
Drag or click to upload
Your data is secure with us, protected by advanced encryption.
Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.
We have sent an OTP to your contact. Please enter it below to verify.
Accenture
39581 Jobs | Dublin
Wipro
19070 Jobs | Bengaluru
Accenture in India
14409 Jobs | Dublin 2
EY
14248 Jobs | London
Uplers
10536 Jobs | Ahmedabad
Amazon
10262 Jobs | Seattle,WA
IBM
9120 Jobs | Armonk
Oracle
8925 Jobs | Redwood City
Capgemini
7500 Jobs | Paris,France
Virtusa
7132 Jobs | Southborough