Get alerts for new jobs matching your selected skills, preferred locations, and experience range. Manage Job Alerts
10.0 years
0 Lacs
Gurugram, Haryana, India
On-site
Acuity Knowledge Partners (Acuity) is a leading provider of bespoke research, analytics and technology solutions to the financial services sector, including asset managers, corporate and investment banks, private equity and venture capital firms, hedge funds and consulting firms. Its global network of over 6,000 analysts and industry experts, combined with proprietary technology, supports more than 600 financial institutions and consulting companies to operate more efficiently and unlock their human capital, driving revenue higher and transforming operations. Acuity is headquartered in London and operates from 10 locations worldwide. The company fosters a diverse, equitable and inclusive work environment, nurturing talent, regardless of race, gender, ethnicity or sexual orientation. Acuity was established as a separate business from Moody’s Corporation in 2019, following its acquisition by Equistone Partners Europe (Equistone). In January 2023, funds advised by global private equity firm Permira acquired a majority stake in the business from Equistone, which remains invested as a minority shareholder. For more information, visit www.acuitykp.com Position Title- Associate Director (Senior Architect – Data) Department-IT Location- Gurgaon/ Bangalore Job Summary The Enterprise Data Architect will enhance the company's strategic use of data by designing, developing, and implementing data models for enterprise applications and systems at conceptual, logical, business area, and application layers. This role advocates data modeling methodologies and best practices. We seek a skilled Data Architect with deep knowledge of data architecture principles, extensive data modeling experience, and the ability to create scalable data solutions. Responsibilities include developing and maintaining enterprise data architecture, ensuring data integrity, interoperability, security, and availability, with a focus on ongoing digital transformation projects. Key Responsibilities Strategy & Planning Develop and deliver long-term strategic goals for data architecture vision and standards in conjunction with data users, department managers, clients, and other key stakeholders. Create short-term tactical solutions to achieve long-term objectives and an overall data management roadmap. Establish processes for governing the identification, collection, and use of corporate metadata; take steps to assure metadata accuracy and validity. Establish methods and procedures for tracking data quality, completeness, redundancy, and improvement. Conduct data capacity planning, life cycle, duration, usage requirements, feasibility studies, and other tasks. Create strategies and plans for data security, backup, disaster recovery, business continuity, and archiving. Ensure that data strategies and architectures are aligned with regulatory compliance. Develop a comprehensive data strategy in collaboration with different stakeholders that aligns with the transformational projects’ goals. Ensure effective data management throughout the project lifecycle. Acquisition & Deployment Ensure the success of enterprise-level application rollouts (e.g. ERP, CRM, HCM, FP&A, etc.) Liaise with vendors and service providers to select the products or services that best meet company goals Operational Management o Assess and determine governance, stewardship, and frameworks for managing data across the organization. o Develop and promote data management methodologies and standards. o Document information products from business processes and create data entities o Create entity relationship diagrams to show the digital thread across the value streams and enterprise o Create data normalization across all systems and data base to ensure there is common definition of data entities across the enterprise o Document enterprise reporting needs develop the data strategy to enable single source of truth for all reporting data o Address the regulatory compliance requirements of each country and ensure our data is secure and compliant o Select and implement the appropriate tools, software, applications, and systems to support data technology goals. o Oversee the mapping of data sources, data movement, interfaces, and analytics, with the goal of ensuring data quality. o Collaborate with project managers and business unit leaders for all projects involving enterprise data. o Address data-related problems regarding systems integration, compatibility, and multiple-platform integration. o Act as a leader and advocate of data management, including coaching, training, and career development to staff. o Develop and implement key components as needed to create testing criteria to guarantee the fidelity and performance of data architecture. o Document the data architecture and environment to maintain a current and accurate view of the larger data picture. o Identify and develop opportunities for data reuse, migration, or retirement. Data Architecture Design: Develop and maintain the enterprise data architecture, including data models, databases, data warehouses, and data lakes. Design and implement scalable, high-performance data solutions that meet business requirements. Data Governance: Establish and enforce data governance policies and procedures as agreed with stakeholders. Maintain data integrity, quality, and security within Finance, HR and other such enterprise systems. Data Migration: Oversee the data migration process from legacy systems to the new systems being put in place. Define & Manage data mappings, cleansing, transformation, and validation to ensure accuracy and completeness. Master Data Management: Devise processes to manage master data (e.g., customer, vendor, product information) to ensure consistency and accuracy across enterprise systems and business processes. Provide data management (create, update and delimit) methods to ensure master data is governed Stakeholder Collaboration: Collaborate with various stakeholders, including business users, other system vendors, and stakeholders to understand data requirements. Ensure the enterprise system meets the organization's data needs. Training and Support: Provide training and support to end-users on data entry, retrieval, and reporting within the candidate enterprise systems. Promote user adoption and proper use of data. 10 Data Quality Assurance: Implement data quality assurance measures to identify and correct data issues. Ensure the Oracle Fusion and other enterprise systems contain reliable and up-to-date information. Reporting and Analytics: Facilitate the development of reporting and analytics capabilities within the Oracle Fusion and other systems Enable data-driven decision-making through robust data analysis. Continuous Improvement: Continuously monitor and improve data processes and the Oracle Fusion and other system's data capabilities. Leverage new technologies for enhanced data management to support evolving business needs. Technology and Tools: Oracle Fusion Cloud Data modeling tools (e.g., ER/Studio, ERwin) ETL tools (e.g., Informatica, Talend, Azure Data Factory) Data Pipelines: Understanding of data pipeline tools like Apache Airflow and AWS Glue. Database management systems: Oracle Database, MySQL, SQL Server, PostgreSQL, MongoDB, Cassandra, Couchbase, Redis, Hadoop, Apache Spark, Amazon RDS, Google BigQuery, Microsoft Azure SQL Database, Neo4j, OrientDB, Memcached) Data governance tools (e.g., Collibra, Informatica Axon, Oracle EDM, Oracle MDM) Reporting and analytics tools (e.g., Oracle Analytics Cloud, Power BI, Tableau, Oracle BIP) Hyperscalers / Cloud platforms (e.g., AWS, Azure) Big Data Technologies such as Hadoop, HDFS, MapReduce, and Spark Cloud Platforms such as Amazon Web Services, including RDS, Redshift, and S3, Microsoft Azure services like Azure SQL Database and Cosmos DB and experience in Google Cloud Platform services such as BigQuery and Cloud Storage. Programming Languages: (e.g. using Java, J2EE, EJB, .NET, WebSphere, etc.) SQL: Strong SQL skills for querying and managing databases. Python: Proficiency in Python for data manipulation and analysis. Java: Knowledge of Java for building data-driven applications. Data Security and Protocols: Understanding of data security protocols and compliance standards. Key Competencies Qualifications: Education: Bachelor’s degree in computer science, Information Technology, or a related field. Master’s degree preferred. Experience: 10+ years overall and at least 7 years of experience in data architecture, data modeling, and database design. Proven experience with data warehousing, data lakes, and big data technologies. Expertise in SQL and experience with NoSQL databases. Experience with cloud platforms (e.g., AWS, Azure) and related data services. Experience with Oracle Fusion or similar ERP systems is highly desirable. Skills: Strong understanding of data governance and data security best practices. Excellent problem-solving and analytical skills. Strong communication and interpersonal skills. Ability to work effectively in a collaborative team environment. Leadership experience with a track record of mentoring and developing team members. Excellent in documentation and presentations. Good knowledge of applicable data privacy practices and laws. Certifications: Relevant certifications (e.g., Certified Data Management Professional, AWS Certified Big Data – Specialty) are a plus. Behavioral A self-starter, an excellent planner and executor and above all, a good team player Excellent communication skills and inter-personal skills are a must Must possess organizational skills, including multi-task capability, priority setting and meeting deadlines Ability to build collaborative relationships and effectively leverage networks to mobilize resources Initiative to learn business domain is highly desirable Likes dynamic and constantly evolving environment and requirements
Posted 6 hours ago
9.0 years
0 Lacs
Noida, Uttar Pradesh, India
On-site
Job Description Join us with your skills and experience as an Architect to design, develop, and maintain robust test automation frameworks for our Autonomous Network solutions. This includes Orchestration/fulfillment (FlowOne, CDPA, CDFF, NoRC), Assurance/NAC, Inventory (UIV, Discovery and Reconciliation), SSO/Security product suites (NIAM), and Analytics. You will work closely with development teams, product owners, and other stakeholders to understand requirements and translate them into practical, high-impact test strategies. You will champion best practices in test automation, driving continuous improvement and innovation within the testing lifecycle. How You Will Contribute And What You Will Learn Develop and maintain comprehensive test automation frameworks and strategies aligned with Agile methodologies and CI/CD pipelines. Design and implement automated tests covering various aspects of service functionality, including performance, security, scalability, reliability, and integration. Develop coding standards, procedures, and methodologies for automated testing, collaborating with other QA leaders and architects. Create detailed test plans that specify automation architecture, positive/negative testing techniques, and reporting mechanisms. Drive end-to-end test automation, aiming for zero manual testing and integrated status reporting. Work with service development and release engineering to integrate automated tests into the CI/CD flow. Act as a subject matter expert on test automation best practices and technologies. Troubleshoot and resolve complex testing issues. Stay current with the latest testing technologies and industry trends. Key Skills And Experience If you have: Bachelor's degree in engineering/technology or equivalent with 9+ years of experience in software testing, with at least 5 years in a Test Architect or similar role for designing and implementing automated test frameworks. Practical Experience on software testing methodologies (Agile, Waterfall) and driving the full Software Testing Life Cycle (STLC). Experience with programming languages such as Java, JavaScript, Python, and scripting languages (Perl, Shell, etc.). It would be nice if you had: Good understanding of databases (Oracle, Postgres, MongoDB, MariaDB, Neo4j) and SQL. Exposure to Linux and containerization technologies (e.g., Docker, Kubernetes). Knowledge of testing in cloud environments (AWS, Azure, GCP) is a plus. About Us Come create the technology that helps the world act together Nokia is committed to innovation and technology leadership across mobile, fixed and cloud networks. Your career here will have a positive impact on people’s lives and will help us build the capabilities needed for a more productive, sustainable, and inclusive world. We challenge ourselves to create an inclusive way of working where we are open to new ideas, empowered to take risks and fearless to bring our authentic selves to work What we offer Nokia offers continuous learning opportunities, well-being programs to support you mentally and physically, opportunities to join and get supported by employee resource groups, mentoring programs and highly diverse teams with an inclusive culture where people thrive and are empowered. Nokia is committed to inclusion and is an equal opportunity employer Nokia has received the following recognitions for its commitment to inclusion & equality: One of the World’s Most Ethical Companies by Ethisphere Gender-Equality Index by Bloomberg Workplace Pride Global Benchmark At Nokia, we act inclusively and respect the uniqueness of people. Nokia’s employment decisions are made regardless of race, color, national or ethnic origin, religion, gender, sexual orientation, gender identity or expression, age, marital status, disability, protected veteran status or other characteristics protected by law. We are committed to a culture of inclusion built upon our core value of respect. Join us and be part of a company where you will feel included and empowered to succeed. About The Team As Nokia's growth engine, we create value for communication service providers and enterprise customers by leading the transition to cloud-native software and as-a-service delivery models. Our inclusive team of dreamers, doers and disruptors push the limits from impossible to possible.
Posted 6 hours ago
8.0 years
0 Lacs
Trivandrum, Kerala, India
Remote
Role-AIML Engineer Location- Remote Expereince-8 to 12 years Notice-Immediate Only Interested candidated share your resume to sunilkumar@xpetize.com Job description: Seeking a highly experienced and technically adept AI/ML Engineer to spearhead a strategic initiative focused on analyzing annual changes in IRS-published TRCs and identifying their downstream impact on codebases. Role demands deep expertise in machine learning, knowledge graph construction, and software engineering processes. The ideal candidate will have a proven track record of delivering production-grade AI solutions in complex enterprise environments. Key Responsibilities: Design and development of an AI/ML-based system to detect and analyze differences in IRS TRC publications year-over-year. Implement knowledge graphs to model relationships between TRC changes and impacted code modules. Collaborate with tax domain experts, software engineers, and DevOps teams to ensure seamless integration of the solution into existing workflows. Define and enforce engineering best practices, including CI/CD, version control, testing, and model governance. Drive the end-to-end lifecycle of the solution—from data ingestion and model training to deployment and monitoring. Ensure scalability, performance, and reliability of the deployed system in a production environment. Mentor junior engineers and contribute to a culture of technical excellence and innovation. Required Skills & Experience: 8+ years of experience in software engineering, with at least 5 years in AI/ML solution delivery. Strong understanding of tax-related data structures, especially IRS TRCs, is a plus. Expertise in building and deploying machine learning models using Python, TensorFlow/PyTorch, and ML Ops frameworks. Hands-on experience with Knowledge graph technologies (e.g., Neo4j, RDF, SPARQL, GraphQL). Deep familiarity with software architecture, microservices, and API design. Experience with NLP techniques for document comparison and semantic analysis. Proven ability to lead cross-functional teams and deliver complex projects on time. Strong communication and stakeholder management skills.
Posted 11 hours ago
8.0 years
0 Lacs
Noida, Uttar Pradesh, India
On-site
Company Description About Sopra Steria Sopra Steria, a major Tech player in Europe with 50,000 employees in nearly 30 countries, is recognised for its consulting, digital services and solutions. It helps its clients drive their digital transformation and obtain tangible and sustainable benefits. The Group provides end-to-end solutions to make large companies and organisations more competitive by combining in-depth knowledge of a wide range of business sectors and innovative technologies with a collaborative approach. Sopra Steria places people at the heart of everything it does and is committed to putting digital to work for its clients in order to build a positive future for all. In 2024, the Group generated revenues of €5.8 billion. Job Description The world is how we shape it. Should have at least 8+ years of experience in Java/Angular development, Design and Application development Proven experience as a Full Stack Developer with expertise in Spring Boot, Neo4j, Angular, and AngularJS. Should have -: A strong understanding of RESTful API design principles and experience in building and consuming APIs. Proficiency in HTML5, CSS3, and JavaScript, along with experience in frontend frameworks like Angular and AngularJS. Good experience of DB Modeling and other ORM frameworks like Hibernate, JPA etc. Solid understanding of software development lifecycle, version control systems (e.g., Git), and agile methodologies. Excellent problem-solving skills and ability to troubleshoot and debug complex issues. Strong communication and collaboration skills, with the ability to work effectively in a team environment. Good to have -: An experience with NoSQL databases, particularly Neo4j, and proficiency in Cypher query language. Familiarity with XSLT transformation, Elastic Search and orchestration technologies such as Docker and Kubernetes. Knowledge of continuous integration and continuous deployment (CI/CD) pipelines. Experience with Graph SQL, cloud platforms like AWS and other modern web technologies is a plus. Join us and be part of a dynamic team where you can contribute to cutting-edge projects and collaborate with talented professionals in a supportive and innovative environment Total Experience Expected: 08-10 years Qualifications B.Tech/MCA with at least 8+ years of experience in Java/Angular development, Design and Application development Additional Information Responsibilities : Design, develop, and maintain backend services using Spring Boot for a variety of web applications. Implement data modeling and database interactions using Neo4j, ensuring efficient data storage and retrieval. Collaborate with frontend developers to integrate backend services with Angular and AngularJS frontend applications. Work closely with UI/UX designers to ensure seamless user experiences and responsive designs. Optimize application performance and scalability through code refactoring, database optimization, and caching techniques. Write clean, maintainable, and well-documented code following best practices and coding standards. Conduct code reviews, provide constructive feedback, and mentor junior developers. Stay updated with the latest technologies and industry trends, and continuously improve skills and knowledge. At our organization, we are committed to fighting against all forms of discrimination. We foster a work environment that is inclusive and respectful of all differences. All of our positions are open to people with disabilities.
Posted 1 day ago
4.0 years
7 - 9 Lacs
Gurgaon
On-site
As the global leader in high-speed connectivity, Ciena is committed to a people-first approach. Our teams enjoy a culture focused on prioritizing a flexible work environment that empowers individual growth, well-being, and belonging. We’re a technology company that leads with our humanity—driving our business priorities alongside meaningful social, community, and societal impact. How You Will Contribute: As a Senior Software Developer within the Blue Planet team, you will play a key role in designing, developing, testing, and supporting scalable software solutions tailored for carrier-class networks and cloud environments. This role requires a strong technical foundation, attention to detail, and a collaborative mindset to deliver high-quality, modular code that is built to scale and last. You will: Work closely with cross-functional teams to design and develop high-performing software modules and features. Write and maintain backend and frontend code with strong emphasis on quality, performance, and maintainability. Support system design, documentation, and end-to-end development including unit testing and debugging. Participate in global agile development teams to deliver against project priorities and milestones. Contribute to the development of telecom inventory management solutions integrated with cloud platforms and advanced network technologies. The Must Haves: Bachelor's or Master’s degree in Computer Science, Engineering, or a related technical field. 4+ years of software development experience. Backend: Java 11+, Spring (Security, Data, MVC), SpringBoot, J2EE, Maven, JUnit. Frontend: TypeScript, JavaScript, Angular 2+, HTML, CSS, SVG, Protractor, Jasmine. Databases: Neo4j (Graph DB), PostgreSQL, TimescaleDB. Experience with SSO implementations (LDAP, SAML, OAuth2). Proficiency with Docker, Kubernetes, and cloud platforms (preferably AWS). Strong understanding of algorithms, data structures, and software design patterns. Assets: Experience with ElasticSearch, Camunda/BPMN, Drools, Kafka integration. Knowledge of RESTful APIs using Spring MVC. Knowledge in Inventory Management Systems (e.g., Cramer, Granite, Metasolv). Familiarity with tools like Node.js, Gulp, and build/test automation. Exposure to telecom/networking technologies such as DWDM/OTN, SONET, MPLS, GPON, FTTH. Understanding of OSS domains and exposure to telecom network/service topology and device modeling. Prior experience working in a global, agile development environment. #LI-FA Not ready to apply? Join our Talent Community to get relevant job alerts straight to your inbox. At Ciena, we are committed to building and fostering an environment in which our employees feel respected, valued, and heard. Ciena values the diversity of its workforce and respects its employees as individuals. We do not tolerate any form of discrimination. Ciena is an Equal Opportunity Employer, including disability and protected veteran status. If contacted in relation to a job opportunity, please advise Ciena of any accommodation measures you may require.
Posted 1 day ago
0 years
0 Lacs
Pune, Maharashtra, India
On-site
Job Reference # 322768BR Job Type Full Time Your role Are you a creative and passionate developer with a knack for building state-of-the-art Front-End solutions? Do you possess an innovative engineering mindset and enjoy leveraging cutting-edge technology to develop diverse solutions? We are seeking a hands-on Full Stack Developer to join UBS Group Technology. In this role, you will contribute to the implementation of a micro-service Front-End for our Enterprise Knowledge Graph, revolutionizing how we handle data within the bank. You will: Design minimalistic, user-friendly interfaces to solve complex problems. Develop innovative micro-sites on top of our enterprise knowledge graph. Enhance existing UI components for reuse across our micro-sites. Your team You will be part of a nimble, multi-disciplinary Data Architecture team within Group CTO, collaborating closely with specialists across various areas of Group Technology. Our team provides the foundation for data-driven management, facilitating processes from strategic and architecture planning to demand management, development, and deployment. The team is globally distributed, with members primarily based in Switzerland, the UK, and the US. Your expertise You have: proven track record in hands-on development and design of Front-End and middleware solutions. strong command of application, data, and infrastructure architecture disciplines. experience working in agile, delivery-oriented teams. Desired: proficiency in JavaScript, Svelte, and CSS. Knowledge of SPARQL and SQL is a plus. experience with graph visualization. expertise in designing and consuming RESTful and GraphQL APIs. experience building modern solutions, including data streaming and cloud-based architectures. familiarity with graph databases (e.g., GraphDB, Anzograph, Jena, Neo4J) is advantageous. experience delivering solutions for Data Analytics users, such as PowerBI. you are: willing to take full ownership of problems and code, with the ability to hit the ground running and deliver exceptional solutions. strong problem solver who anticipates issues and resolves them proactively. skilled in communicating effectively with both technical and non-technical audiences. About Us UBS is the world’s largest and the only truly global wealth manager. We operate through four business divisions: Global Wealth Management, Personal & Corporate Banking, Asset Management and the Investment Bank. Our global reach and the breadth of our expertise set us apart from our competitors. We have a presence in all major financial centers in more than 50 countries. How We Hire We may request you to complete one or more assessments during the application process. Learn more Join us At UBS, we know that it's our people, with their diverse skills, experiences and backgrounds, who drive our ongoing success. We’re dedicated to our craft and passionate about putting our people first, with new challenges, a supportive team, opportunities to grow and flexible working options when possible. Our inclusive culture brings out the best in our employees, wherever they are on their career journey. We also recognize that great work is never done alone. That’s why collaboration is at the heart of everything we do. Because together, we’re more than ourselves. We’re committed to disability inclusion and if you need reasonable accommodation/adjustments throughout our recruitment process, you can always contact us. Disclaimer / Policy Statements UBS is an Equal Opportunity Employer. We respect and seek to empower each individual and support the diverse cultures, perspectives, skills and experiences within our workforce.
Posted 1 day ago
7.0 years
0 Lacs
India
Remote
Role: Neo4j Engineer Overall IT Experience: 7+ years Relevant experience: (Graph Databases: 4+ years, Neo4j: 2+ years) Location: Remote Company Description Bluetick Consultants is a technology-driven firm that supports hiring remote developers, building technology products, and enabling end-to-end digital transformation. With previous experience in top technology companies such as Amazon, Microsoft, and Craftsvilla, we understand the needs of our clients and provide customized solutions. Our team has expertise in emerging technologies, backend and frontend development, cloud development, and mobile technologies. We prioritize staying up-to-date with the latest technological advances to create a long-term impact and grow together with our clients. Key Responsibilities • Graph Database Architecture: Design and implement Neo4j graph database schemas optimized for fund administration data relationships and AI-powered queries • Knowledge Graph Development: Build comprehensive knowledge graphs connecting entities like funds, investors, companies, transactions, legal documents, and market data • Graph-AI Integration: Integrate Neo4j with AI/ML pipelines, particularly for enhanced RAG (Retrieval-Augmented Generation) systems and semantic search capabilities • Complex Relationship Modeling: Model intricate relationships between Limited Partners, General Partners, fund structures, investment flows, and regulatory requirements • Query Optimization: Develop high-performance Cypher queries for real-time analytics, relationship discovery, and pattern recognition • Data Pipeline Integration: Build ETL processes to populate and maintain graph databases from various data sources including FundPanel.io, legal documents, and external market data using domain specific ontologies • Graph Analytics: Implement graph algorithms for fraud detection, risk assessment, relationship scoring, and investment opportunity identification • Performance Tuning: Optimize graph database performance for concurrent users and complex analytical queries • Documentation & Standards: Establish graph modelling standards, query optimization guidelines, and comprehensive technical documentation Key Use Cases You'll Enable • Semantic Search Enhancement: Create knowledge graphs that improve AI search accuracy by understanding entity relationships and context • Investment Network Analysis: Map complex relationships between investors, funds, portfolio companies, and market segments • Compliance Graph Modelling: Model regulatory relationships and fund terms to support automated auditing and compliance validation • Customer Relationship Intelligence: Build relationship graphs for customer relations monitoring and expansion opportunity identification • Predictive Modelling Support: Provide graph-based features for investment prediction and risk assessment models • Document Relationship Mapping: Connect legal documents, contracts, and agreements through entity and relationship extraction Required Qualifications • Bachelor's degree in Computer Science, Data Engineering, or related field • 7+ years of overall IT Experience • 4+ years of experience with graph databases, with 2+ years specifically in Neo4j • Strong background in data modelling, particularly for complex relationship structures • Experience with financial services data and regulatory requirements preferred • Proven experience integrating graph databases with AI/ML systems • Understanding of knowledge graph concepts and semantic technologies • Experience with high-volume, production-scale graph database implementations Technology Skills • Graph Databases: Neo4j (primary), Cypher query language, APOC procedures, Neo4j Graph Data Science library • Programming: Python, Java, or Scala for graph data processing and integration • AI Integration: Experience with graph-enhanced RAG systems, vector embeddings in graph context, GraphRAG implementations • Data Processing: ETL pipelines, data transformation, real-time data streaming (Kafka, Apache Spark) • Cloud Platforms: Neo4j Aura, Azure integration, containerized deployments • APIs: Neo4j drivers, REST APIs, GraphQL integration • Analytics: Graph algorithms (PageRank, community detection, shortest path, centrality measures) • Monitoring: Neo4j monitoring tools, performance profiling, query optimization • Integration: Elasticsearch integration, vector database connections, multi-modal data handling Specific Technical Requirements • Knowledge Graph Construction: Entity resolution, relationship extraction, ontology modelling • Cypher Expertise: Advanced Cypher queries, stored procedures, custom functions • Scalability: Clustering, sharding, horizontal scaling strategies • Security: Graph-level security, role-based access control, data encryption • Version Control: Graph schema versioning, migration strategies • Backup & Recovery: Graph database backup strategies, disaster recovery planning Industry Context Understanding • Fund Administration: Understanding of fund structures, capital calls, distributions, and investor relationships • Financial Compliance: Knowledge of regulatory requirements and audit trails in financial services • Investment Workflows: Understanding of due diligence processes, portfolio management, and investor reporting • Legal Document Structures: Familiarity with LPA documents, subscription agreements, and fund formation documents Collaboration Requirements • AI/ML Team: Work closely with GenAI engineers to optimize graph-based AI applications • Data Architecture Team: Collaborate on overall data architecture and integration strategies • Backend Developers: Integrate graph databases with application APIs and microservices • DevOps Team: Ensure proper deployment, monitoring, and maintenance of graph database infrastructure • Business Stakeholders: Translate business requirements into effective graph models and queries Performance Expectations • Query Performance: Ensure sub-second response times for standard relationship queries • Scalability: Support 100k+ users with concurrent access to graph data • Accuracy: Maintain data consistency and relationship integrity across complex fund structures • Availability: Ensure 99.9% uptime for critical graph database services • Integration Efficiency: Seamless integration with existing FundPanel.io systems and new AI services This role offers the opportunity to work at the intersection of advanced graph technology and artificial intelligence, creating innovative solutions that will transform how fund administrators understand and leverage their data relationships.
Posted 1 day ago
3.0 years
0 Lacs
Pune, Maharashtra, India
On-site
Velotio Technologies is a product engineering company working with innovative startups and enterprises. We have provided full-stack product development for 110+ startups across the globe, building products in the cloud-native, data engineering, B2B SaaS, IoT & Machine Learning space. Our team of 400+ elite software engineers solves hard technical problems while transforming customer ideas into successful products. Requirements Roles and Responsibilities Lead the creation, development, and implementation of critical system design changes, enhancements, and software projects Ensure timely execution of project deliverables Work with other engineers to ensure the system and product is consistent and aligned through all processes Improve product quality, performance, and security through substantial process improvements Follow development standards and promote best practices Individual contributor as an engineer. Requirement and Qualification: 3+ years experience in Python programming Experience with Neo4j for graph database management and querying Familiarity with Postgres and Clickhouse for database management and optimization Experience with cloud platforms including AWS, Azure, and GCP Understanding of serverless architecture for building and deploying applications Experience with SaaS (Software as a Service) /product development. Experience with containerization and orchestration technologies (e.g., Docker, Kubernetes) Exceptional problem-solving and analytical skills Excellent communication and teamwork abilities. Bonus points if you... Experience in AWS ECS, EKS Familiarity with any open-source vulnerability/secret scanning tool Benefits Our Culture: We have an autonomous and empowered work culture encouraging individuals to take ownership and grow quickly Flat hierarchy with fast decision making and a startup-oriented "get things done" culture A strong, fun & positive environment with regular celebrations of our success. We pride ourselves in creating an inclusive, diverse & authentic environment We want to hire smart, curious and ambitious folks so please reach out even if you do not have all of the requisite experience. We are looking for engineers with the potential to grow! At Velotio, we embrace diversity. Inclusion is a priority for us, and we are eager to foster an environment where everyone feels valued. We welcome applications regardless of ethnicity or cultural background, age, gender, nationality, religion, disability or sexual orientation.
Posted 1 day ago
0.0 - 3.0 years
0 Lacs
chennai, tamil nadu
On-site
Embark on a transformative journey with SwaaS, where innovation meets opportunity. Explore thrilling career prospects at the cutting edge of technology. Join our dynamic team, dedicated to shaping the future of IT. At SwaaS, we offer more than just jobs; we provide a platform for growth, collaboration, and impactful contributions. Discover a workplace where your aspirations align with limitless possibilities. Your journey towards a rewarding career in technology begins here, with SwaaS as your guide. Perks and Benefits We go beyond Salaries and provide guaranteed benefits that speak about Swaas value and culture. Our employees get common benefits and also performance-based individual benefits Performance-based benefits We promote a culture of equity. Accept the challenge, deliver the results, and get rewarded. Healthcare Our comprehensive medical insurance helps you cover your urgent medical needs. Competitive Salary We assure with pride that we are on par with the industry leaders in terms of our salary package. Employee Engagement A break is always needed out of the regular monotonous work assignments. Our employee engagement program helps our employees enhance their team bonding. Upskilling We believe in fostering a culture of Learning and harnessing the untapped potential in our employees. Everyone is encouraged and rewarded for acquiring new skills and certifications. Junior AI/ML Developer (Entry-Level) (Experience: 0-2 years) Tech Stack: Python, Node.js (Javascript), LangChain, LLama Index, OpenAI API, Perplexity.ai API, Neo4j, PostgreSQL Responsibilities: - Assist in developing AI-driven solutions using LLMs (ChatGPT, Perplexity.ai) and RAG (Retrieval-Augmented Generation). - Work on intent extraction and chatbot development, integrating APIs like OpenAI or LLama. - Support the design and testing of AI-enhanced workflows. - Implement database interactions (MySQL or PostgreSQL for structured data). - Write and optimize Python/Node.js scripts for the applications. - Debug and refine LLM-powered chatbots. Requirements: - Strong programming skills in Python (FastAPI, Flask) or Node.js. - Exposure to NLP, LLMs, AI APIs (ChatGPT, Perplexity.ai, LangChain). - Familiarity with RESTful APIs and Graph Databases (Neo4j). - Basic understanding of cloud platforms (AWS, Azure, GCP). - Passion for AI, NLP, and chatbot development. - Bonus: Knowledge of UI frameworks (React, Next.js). - Good to have - Pinecone or equivalent vector databases.,
Posted 1 day ago
5.0 - 9.0 years
0 Lacs
pune, maharashtra
On-site
You will be responsible for developing scalable web applications using Python (FastAPI), React.js, and cloud-native technologies. Specifically, you will work on building a low-code/no-code AI agent platform, designing an intuitive workflow UI, and integrating with LLMs, enterprise connectors, and role-based access controls. As a Full-Stack Developer, your responsibilities will include developing and optimizing APIs using FastAPI, integrating with LangChain, Pinecone/Weaviate vector databases, and enterprise connectors like Airbyte/Nifi for backend development. For frontend development, you will build an interactive drag-and-drop workflow UI using React.js along with supporting libraries like React Flow, D3.js, and TailwindCSS. You will also be tasked with implementing authentication mechanisms such as OAuth2, Keycloak, and role-based access controls for multi-tenant environments. Database design will involve working with PostgreSQL for structured data, MongoDB for unstructured data, and Neo4j for knowledge graphs. Your role will extend to DevOps and deployment using Docker, Kubernetes, and Terraform across various cloud platforms like Azure, AWS, and GCP. Performance optimization will be crucial as you strive to enhance API performance and frontend responsiveness for an improved user experience. Collaboration with AI and Data Engineers will be essential to ensure seamless integration of AI models. To excel in this role, you should have at least 5 years of experience in FastAPI, React.js, and cloud-native applications. A strong understanding of REST APIs, GraphQL, and WebSockets is required. Experience with JWT authentication, OAuth2, and multi-tenant security is essential. Proficiency in databases such as PostgreSQL, MongoDB, Neo4j, and Redis is expected. Knowledge of workflow automation tools like n8n, Node-RED, and Temporal.io will be beneficial. Familiarity with containerization tools (Docker, Kubernetes) and CI/CD pipelines is preferred. Any experience with Apache Kafka, WebSockets, or AI-driven chatbots would be considered a bonus.,
Posted 1 day ago
8.0 years
20 - 40 Lacs
India
On-site
Role: Senior Graph Data Engineer (Neo4j & AI Knowledge Graphs) Experience: 8+ years Type: Contract We’re hiring a Graph Data Engineer to design and implement advanced Neo4j-powered knowledge graph systems for our next-gen AI platform. You'll work at the intersection of data engineering, AI/ML, and financial services , helping build the graph infrastructure that powers semantic search, investment intelligence, and automated compliance for venture capital and private equity clients. This role is ideal for engineers who are passionate about graph data modeling , Neo4j performance , and enabling AI-enhanced analytics through structured relationships. What You'll Do Design Knowledge Graphs: Build and maintain Neo4j graph schemas modeling complex fund administration relationships — investors, funds, companies, transactions, legal docs, etc. Graph-AI Integration: Work with GenAI teams to power RAG systems, semantic search, and graph-enhanced NLP pipelines. ETL & Data Pipelines: Develop scalable ingestion pipelines from sources like FundPanel.io, legal documents, and external market feeds using Python, Spark, or Kafka. Optimize Graph Performance: Craft high-performance Cypher queries, leverage APOC procedures, and tune for real-time analytics. Graph Algorithms & Analytics: Implement algorithms for fraud detection, relationship scoring, compliance, and investment pattern analysis. Secure & Scalable Deployment: Implement clustering, backups, and role-based access on Neo4j Aura or containerized environments. Collaborate Deeply: Partner with AI/ML, DevOps, data architects, and business stakeholders to translate use cases into scalable graph solutions. What You Bring 7+ years in software/data engineering; 2+ years in Neo4j and Cypher. Strong experience in graph modeling, knowledge graphs, and ontologies. Proficiency in Python, Java, or Scala for graph integrations. Experience with graph algorithms (PageRank, community detection, etc.). Hands-on with ETL pipelines, Kafka/Spark, and real-time data ingestion. Cloud-native experience (Neo4j Aura, Azure, Docker/K8s). Familiarity with fund structures, LP/GP models, or financial/legal data a plus. Strong understanding of AI/ML pipelines, especially graph-RAG and embeddings. Use Cases You'll Help Build AI Semantic Search over fund documents and investment entities. Investment Network Analysis for GPs, LPs, and portfolio companies. Compliance Graphs modeling fund terms and regulatory checks. Document Graphs linking LPAs, contracts, and agreements. Predictive Investment Models enhanced by graph relationships. Skills: java,machine learning,spark,apache spark,neo4j aura,ai,azure,cloud-native technologies,data,ai/ml pipelines,scala,python,cypher,graphs,ai knowledge graphs,graph data modeling,apoc procedures,semantic search,etl pipelines,data engineering,neo4j,etl,cypher query,pipelines,graph schema,kafka,kafka streams,graph algorithms
Posted 2 days ago
5.0 years
0 Lacs
Gurugram, Haryana, India
On-site
Who You'll Work With You are someone who thrives in a high-performance environment, bringing a growth mindset and entrepreneurial spirit to tackle meaningful challenges that have a real impact. In return for your drive, determination, and curiosity, we’ll provide the resources, mentorship, and opportunities to help you quickly broaden your expertise, grow into a well-rounded professional, and contribute to work that truly makes a difference. When you join us, you will have: Continuous learning: Our learning and apprenticeship culture, backed by structured programs, is all about helping you grow while creating an environment where feedback is clear, actionable, and focused on your development. The real magic happens when you take the input from others to heart and embrace the fast-paced learning experience, owning your journey. A voice that matters: From day one, we value your ideas and contributions. You’ll make a tangible impact by offering innovative ideas and practical solutions. We not only encourage diverse perspectives, but they are critical in driving us toward the best possible outcomes. Global community: With colleagues across 65+ countries and over 100 different nationalities, our firm’s diversity fuels creativity and helps us come up with the best solutions. Plus, you’ll have the opportunity to learn from exceptional colleagues with diverse backgrounds and experiences. Exceptional benefits: On top of a competitive salary (based on your location, experience, and skills), we provide a comprehensive benefits package to enable holistic well-being for you and your family. Your Impact You will be having deep experiences in practicing, role modeling and coaching teams in key areas of SRE related competencies Culture and Organization: You will champion CI/CD practices, the concept of error budgets and blameless post mortems. You will continuously help to remove team boundaries (dev, ops, others). Your advanced knowledge within the SRE chapter and practice will contribute back to the community within McK and beyond. You will be a full-stack engineer with DR/BCP experience, proficient in cloud-native models and in reliability engineering advisory, and have a strong knowledge of loosely coupled API based component architecture models. You will be proficient in SCM and CI/CD tooling and practices for container workloads, a variety of related developer workflows and principles, and master multiple programming and IaC languages. You will be a competent enabler of automated zero-downtime deployments. You will bring expertise in TDD principles and practices as well as key test automation tools and frameworks. You are well versed in chaos-engineering practices and in wheels-of-misfortune exercises. You will be seasoned in outcome-centric monitoring / measurement (cloud native log management and monitoring tools, SLOs, SLIs, error budgets, toil budgets etc. You will bring expertise in stakeholder specific reporting as well. You will work with our Secure Foundations - MCS team, which is part of McKinsey’s Tech Ecosystem organization, developing new products/services and integrating them into our client work. Our company is moving fast from the traditional IT world to a Digital era embracing Agile principles. We are looking for highly skilled developers with an SRE mindset to help us with this transformation. You will work in small teams (incl. product managers, developers and operations people) in a highly collaborative way, use the latest technologies and enjoy seeing the direct impact from your work. You will combine ‘Agile’ with expertise in cloud, big data and mobile to create and maintain custom solutions, in a way consistent with SRE principles, that will help clients increase productivity and make timely decisions. This includes, but is not limited to: Development, implementation and operation of IT systems, processes supporting SaaS applications and platforms, automation of provisioning, quality controls, security auditing and maintenance, and continuous measurement and improvement of efficiency of operational activities and resources. Your Qualifications and Skills 5+ years of experience with software engineering best practice Proficiency in one or more programming languages, such as Python, JavaScript, Golang, or Ruby. Hands-on experience implementing infrastructure as code using Terraform, or similar automation tools like Ansible and CloudFormation Experience designing and building CI/CD pipelines using tools like GitHub Actions, ArgoCD, CircleCI, or Jenkins along with package management tools like Jfrog or Nexus Experience with public cloud environments, specifically AWS and either Azure or Google Cloud Platform (GCP). Expertise with container technologies and orchestration tools, including Docker, Kubernetes, Helm, and service mesh solutions such as Linkerd or Istio Experience with infrastructure and reliability testing frameworks such as Test-Kitchen, AWSpec and InSpec Experience in managing front-end and back-end workloads such as React, TypeScript, Python, Node.js, Nginx, and API management tools like Apigee and AWS API Gateway Proficiency with databases such as Neo4j, Redis, PostgreSQL, and MongoDB.Familiarity with monitoring and logging tools such as Dynatrace, Splunk, CloudWatch, and other similar platforms like ELK, Prometheus, or Grafana Expertise in networking concepts, including prior experience managing CDN+WAF configurations in Akamai, Cloudflare, AWS CloudFront, and experience with VPCs, Load Balancers, and SSH tunnels Experience with Okta, Azure AD, Ping Identity, and other OIDC/OAuth2 providers and Implementing and managing RBAC for least-privilege access Proficiency with HashiCorp Vault for managing secrets and implementing token rotation Experience with SOC 2 audits, vulnerability management, and SSL certificate management Strong skills in developing technical documentation such as architecture diagrams, runbooks, and technical documents, with experience in complex platform migrations and managing multiple workstreams
Posted 2 days ago
4.0 years
0 Lacs
Gurugram, Haryana, India
On-site
As the global leader in high-speed connectivity, Ciena is committed to a people-first approach. Our teams enjoy a culture focused on prioritizing a flexible work environment that empowers individual growth, well-being, and belonging. We’re a technology company that leads with our humanity—driving our business priorities alongside meaningful social, community, and societal impact. How You Will Contribute As a Senior Software Developer within the Blue Planet team, you will play a key role in designing, developing, testing, and supporting scalable software solutions tailored for carrier-class networks and cloud environments. This role requires a strong technical foundation, attention to detail, and a collaborative mindset to deliver high-quality, modular code that is built to scale and last. You Will Work closely with cross-functional teams to design and develop high-performing software modules and features. Write and maintain backend and frontend code with strong emphasis on quality, performance, and maintainability. Support system design, documentation, and end-to-end development including unit testing and debugging. Participate in global agile development teams to deliver against project priorities and milestones. Contribute to the development of telecom inventory management solutions integrated with cloud platforms and advanced network technologies. The Must Haves Bachelor's or Master’s degree in Computer Science, Engineering, or a related technical field. 4+ years of software development experience. Backend: Java 11+, Spring (Security, Data, MVC), SpringBoot, J2EE, Maven, JUnit. Frontend: TypeScript, JavaScript, Angular 2+, HTML, CSS, SVG, Protractor, Jasmine. Databases: Neo4j (Graph DB), PostgreSQL, TimescaleDB. Experience with SSO implementations (LDAP, SAML, OAuth2). Proficiency with Docker, Kubernetes, and cloud platforms (preferably AWS). Strong understanding of algorithms, data structures, and software design patterns. Assets Experience with ElasticSearch, Camunda/BPMN, Drools, Kafka integration. Knowledge of RESTful APIs using Spring MVC. Knowledge in Inventory Management Systems (e.g., Cramer, Granite, Metasolv). Familiarity with tools like Node.js, Gulp, and build/test automation. Exposure to telecom/networking technologies such as DWDM/OTN, SONET, MPLS, GPON, FTTH. Understanding of OSS domains and exposure to telecom network/service topology and device modeling. Prior experience working in a global, agile development environment. Not ready to apply? Join our Talent Community to get relevant job alerts straight to your inbox. At Ciena, we are committed to building and fostering an environment in which our employees feel respected, valued, and heard. Ciena values the diversity of its workforce and respects its employees as individuals. We do not tolerate any form of discrimination. Ciena is an Equal Opportunity Employer, including disability and protected veteran status. If contacted in relation to a job opportunity, please advise Ciena of any accommodation measures you may require.
Posted 2 days ago
3.0 - 4.0 years
3 - 6 Lacs
India
On-site
Job Title: Python Backend Developer (Data Layer) Location: Mohali, Punjab Company: RevClerx About RevClerx: RevClerx Pvt. Ltd., founded in 2017 and based in the Chandigarh/Mohali area (India), is a dynamic Information Technology firm providing comprehensive IT services with a strong focus on client-centric solutions. As a global provider, we cater to diverse business needs including website designing and development, digital marketing, lead generation services (including telemarketing and qualification), and appointment setting. Job Summary: We are seeking a skilled Python Backend Developer with a strong passion and proven expertise in database design and implementation. This role requires 3-4 years of backend development experience, focusing on building robust, scalable applications and APIs. The ideal candidate will not only be proficient in Python and common backend frameworks but will possess significant experience in designing, modeling, and optimizing various database solutions, including relational databases (like PostgreSQL) and, crucially, graph databases (specifically Neo4j). You will play a vital role in architecting the data layer of our applications, ensuring efficiency, scalability, and the ability to handle complex, interconnected data. Key Responsibilities: ● Design, develop, test, deploy, and maintain scalable and performant Python-based backend services and APIs. ● Lead the design and implementation of database schemas for relational (e.g., PostgreSQL) and NoSQL databases, with a strong emphasis on Graph Databases (Neo4j). ● Model complex data relationships and structures effectively, particularly leveraging graph data modeling principles where appropriate. ● Write efficient, optimized database queries (SQL, Cypher, potentially others). ● Develop and maintain data models, ensuring data integrity, consistency, and security. ● Optimize database performance through indexing strategies, query tuning, caching mechanisms, and schema adjustments. ● Collaborate closely with product managers, frontend developers, and other stakeholders to understand data requirements and translate them into effective database designs. ● Implement data migration strategies and scripts as needed. ● Integrate various databases seamlessly with Python backend services using ORMs (like SQLAlchemy, Django ORM) or native drivers. ● Write unit and integration tests, particularly focusing on data access and manipulation logic. ● Contribute to architectural decisions, especially concerning data storage, retrieval, and processing. ● Stay current with best practices in database technologies, Python development, and backend systems. Minimum Qualifications: ● Bachelor's degree in Computer Science, Engineering, Information Technology, or a related field, OR equivalent practical experience. ● 3-4 years of professional software development experience with a primary focus on Python backend development. ● Strong proficiency in Python and its standard libraries. ● Proven experience with at least one major Python web framework (e.g., Django, Flask, FastAPI). ● Demonstrable, hands-on experience designing, implementing, and managing relational databases (e.g., PostgreSQL). ● Experience with at least one NoSQL database (e.g., MongoDB, Redis, Cassandra). ● Solid understanding of data structures, algorithms, and object-oriented programming principles. ● Experience designing and consuming RESTful APIs. ● Proficiency with version control systems, particularly Git. ● Strong analytical and problem-solving skills, especially concerning data modeling and querying. ● Excellent communication and teamwork abilities. Preferred (Good-to-Have) Qualifications: ● Graph Database Expertise: ○ Significant, demonstrable experience designing and implementing solutions using Graph Databases (Neo4j strongly preferred). ○ Proficiency in graph query languages, particularly Cypher. ○ Strong understanding of graph data modeling principles, use cases (e.g., recommendation engines, fraud detection, knowledge graphs, network analysis), and trade-offs. ● Advanced Database Skills: ○ Experience with database performance tuning and monitoring tools. ○ Experience with Object-Relational Mappers (ORMs) like SQLAlchemy or Django ORM in depth. ○ Experience implementing data migration strategies for large datasets. ● Cloud Experience: Familiarity with cloud platforms (e.g., AWS, Azure, Google Cloud Platform) and their managed database services (e.g., RDS, Aurora, Neptune, DocumentDB, MemoryStore). ● Containerization & Orchestration: Experience with Docker and Kubernetes. ● Asynchronous Programming: Experience with Python's asyncio and async frameworks. ● Data Pipelines: Familiarity with ETL processes or data pipeline tools (e.g., Apache Airflow). ● Testing: Experience writing tests specifically for database interactions and data integrity. What We Offer: ● Challenging projects with opportunities to work on cutting-edge technologies especially in the field of AI. ● Competitive salary and comprehensive benefits package. ● Opportunities for professional development and learning (e.g., conferences, courses, certifications). ● A collaborative, innovative, and supportive work environment. How to Apply: Interested candidates are invited to submit their resume and a cover letter outlining their relevant experience, specifically highlighting their database design expertise (including relational, NoSQL, and especially Graph DB/Neo4j experience) Job Types: Full-time, Permanent Pay: ₹30,000.00 - ₹55,373.94 per month Benefits: Food provided Health insurance Schedule: Day shift Monday to Friday
Posted 2 days ago
10.0 years
0 Lacs
India
On-site
🧠 Co-Founder & CTO at Biolligence 📍 Location: Hybrid (UK/US/India preferred) 💼 Type: Co-Founder Role | Equity-based | Strategic Leadership About Us Biolligence is building the future of predictive safety in drug development. Our flagship platform, BioTox , uses AI agents and multi-omics data to forecast clinical toxicity at a population scale before a trial begins. We're working at the cutting edge of clinical AI, real-world evidence, and digital twins to reduce trial failures, protect patients, and accelerate new therapies to market. Backed by domain experts, advisors, and early traction from industry conversations, we’re now looking for a Co-Founder & CTO to help take this vision to the next level. The Opportunity We're seeking a technical co-founder who is both hands-on and visionary—a leader who can architect, build, and scale the BioTox platform while shaping our product roadmap and AI strategy. As CTO, you will: Lead all technology strategy, architecture, and execution Build and mentor the core engineering and data science team Drive development of AI/ML pipelines (e.g. toxicity prediction, KG agents, LLM workflows) Oversee integration of multi-omics datasets, clinical trial data, and real-world safety signals Work closely with the CEO on investor pitches, technical due diligence, and roadmap prioritization Champion best practices in data security, compliance (GxP, GDPR), and platform scalability What We’re Looking For ✅ Strong technical background with 10+ years in software engineering, AI/ML, or bioinformatics ✅ Hands-on experience with: AI/ML pipelines (LLMs, AI Agents, Transformers, etc.) Knowledge graphs (Neo4j, NetworkX, RDF) Cloud platforms (AWS/GCP/Azure), MLOps, and APIs ✅ Experience building and scaling platforms in healthcare, biotech, or life sciences ✅ Bonus: Experience with multi-omics, clinical data standards (CDISC, HL7), or FDA/EMA workflows ✅ Start-up mindset: comfortable with ambiguity, fast iteration, and building from scratch ✅ Passion for transforming drug discovery and clinical development through technology What We Offer 🎯 True co-founder role — you’ll shape the company’s mission, roadmap, and culture 📈 Significant equity stake 🤝 Work with a passionate founder (ex-Recursion, AI Scientist, PhD) and world-class advisors 💡 Build a product with real-world impact on patients, regulators, and the pharma ecosystem Let’s Talk Excited by the challenge of building AI agents that could save patients’ lives and billions in drug development? Let’s build Biolligence together. 📩 Reach out to Sujit Tangadpalliwar at hello@biolligence.com or message us directly.
Posted 2 days ago
6.0 - 9.0 years
0 Lacs
Bengaluru, Karnataka, India
On-site
Basic understanding of machine learning frameworks such as TensorFlow, PyTorch, or Hugging Face. • Knowledge of neural network architectures, particularly in areas like Transformers and basic deep learning models. • Familiarity with Python programming and essential ML libraries (NumPy, Pandas, Scikit-learn). • Exposure to NLP (Natural Language Processing) concepts and basic text processing tasks. • Some experience with cloud platforms (AWS, GCP, or Azure) for deploying simple AI models. • Understanding of basic databases and their integration with AI systems (NoSQL or SQL databases). Soft Skills: • Strong eagerness to learn and adapt to new technologies in the AI and machine learning field. • Ability to work under guidance and collaborate within a team environment. • Good problem-solving abilities and analytical thinking. • Effective communication skills to discuss technical issues and progress with the team. Education and Experience: • Bachelor’s degree in computer science, Data Science, Artificial Intelligence, or related fields. • 6-9 years of experience in machine learning, AI, or related areas (internships or academic projects are a plus). Preferred Skills (Nice-to-Have): • Exposure to basic machine learning deployment tools or practices. • Familiarity with any vector databases (ChromaDB, Pinecone, Weaviate, Milvus, FAISS) or graph databases (Neo4j, TigerGraph). • Interest in generative AI or graph-based AI solutions. • Involvement in open-source projects or personal contributions to machine learning communities. • Understanding of ethical AI principles or data privacy basics. Role Summary: As a Junior Machine Learning Developer, you will be part of a dynamic team working on cutting-edge AI and machine learning solutions. This role offers an exciting opportunity for a motivated individual to learn and grow their skills in a fast-paced, collaborative environment. You will assist senior developers in developing, testing, and deploying AI models, while gaining hands-on experience with machine learning frameworks and real-world AI applications.
Posted 2 days ago
5.0 - 9.0 years
0 Lacs
pune, maharashtra
On-site
The Engineer Intmd Analyst is an intermediate level position responsible for a variety of engineering activities including the design, acquisition, and development of hardware, software, and network infrastructure in coordination with the Technology team. The overall objective of this role is to ensure quality standards are being met within existing and planned frameworks. Responsibilities: - Provide assistance with a product or product component development within the technology domain - Conduct product evaluations with vendors and recommend product customization for integration with systems - Assist with training activities, mentor junior team members, and ensure teams" adherence to all control and compliance initiatives - Assist with application prototyping and recommend solutions around implementation - Provide third-line support to identify the root cause of issues and react to systems and application outages or networking issues - Support projects and provide project status updates to project manager or Sr. Engineer - Partner with development teams to identify engineering requirements and assist with defining application/system requirements and processes - Create installation documentation, training materials, and deliver technical training to support the organization - Appropriately assess risk when business decisions are made, demonstrating particular consideration for the firm's reputation and safeguarding Citigroup, its clients, and assets, by driving compliance with applicable laws, rules, and regulations, adhering to Policy, applying sound ethical judgment regarding personal behavior, conduct, and business practices, and escalating, managing, and reporting control issues with transparency. Qualifications: - 5-8 years of relevant experience in an Engineering role - Experience working in Financial Services or a large complex and/or global environment - Involvement in DevOps activities (SRE/LSE Auto Deployment/Self Healing) and Application Support Tech Stack: Basic - Java/python, Unix, Oracle Essential Skills: - IT experience working in one of Hbase, HDFS, Kafka, Neo4J, Akka, Spark, Storm, and GemFire - IT Support experience working in Unix, Cloud & Windows environments - Experience supporting RDBMS DB like MongoD, ORACLE, Sybase, MS SQL & DB2 - Supported Applications deployed in Websphere, Weblogic, IIS, and Tomcat - Familiar with Autosys and setup - Understanding of client-server architecture (clustered and non-clustered) - Basic Networking knowledge (Load balancers, Network Protocols) - Working knowledge of Lookup Active Directory Protocol(LDAP) and Single Sign On concepts - Service Now expertise - Experience working in Multiple Application Support Model is preferred Other Essential Attributes: - Consistently demonstrates clear and concise written and verbal communication - Comprehensive knowledge of design metrics, analytics tools, benchmarking activities, and related reporting to identify best practices - Demonstrated analytic/diagnostic skills - Ability to work in a matrix environment and partner with virtual teams - Ability to work independently, prioritize, and take ownership of various parts of a project or initiative - Ability to work under pressure and manage tight deadlines or unexpected changes in expectations or requirements - Proven track record of operational process change and improvement Education: - Bachelors degree/University degree or equivalent experience Job Family Group: - Technology Job Family: - Systems & Engineering Time Type: - Full time Most Relevant Skills: Please see the requirements listed above.,
Posted 2 days ago
2.0 - 9.0 years
0 Lacs
karnataka
On-site
We are seeking a Data Architect / Sr. Data and Pr. Data Architects to join our team. In this role, you will be involved in a combination of hands-on contribution, customer engagement, and technical team management. As a Data Architect, your responsibilities will include designing, architecting, deploying, and maintaining solutions on the MS Azure platform using various Cloud & Big Data Technologies. You will be managing the full life-cycle of Data Lake / Big Data solutions, starting from requirement gathering and analysis to platform selection, architecture design, and deployment. It will be your responsibility to implement scalable solutions on the Cloud and collaborate with a team of business domain experts, data scientists, and application developers to develop Big Data solutions. Moreover, you will be expected to explore and learn new technologies for creative problem solving and mentor a team of Data Engineers. The ideal candidate should possess strong hands-on experience in implementing Data Lake with technologies such as Data Factory (ADF), ADLS, Databricks, Azure Synapse Analytics, Event Hub & Streaming Analytics, Cosmos DB, and Purview. Additionally, experience with big data technologies like Hadoop (CDH or HDP), Spark, Airflow, NiFi, Kafka, Hive, HBase, MongoDB, Neo4J, Elastic Search, Impala, Sqoop, etc., is required. Proficiency in programming and debugging skills in Python and Scala/Java is essential, with experience in building REST services considered beneficial. Candidates should also have experience in supporting BI and Data Science teams in consuming data in a secure and governed manner, along with a good understanding of using CI/CD with Git, Jenkins / Azure DevOps. Experience in setting up cloud-computing infrastructure solutions, hands-on experience/exposure to NoSQL Databases, and Data Modelling in Hive are all highly valued. Applicants should have a minimum of 9 years of technical experience, with at least 5 years on MS Azure and 2 years on Hadoop (CDH/HDP).,
Posted 2 days ago
2.0 - 6.0 years
0 Lacs
pune, maharashtra
On-site
About Velotio Technologies: Velotio Technologies is a product engineering company that collaborates with innovative startups and enterprises. With a track record of providing full-stack product development for 110+ startups worldwide, we specialize in building products in areas such as cloud-native solutions, data engineering, B2B SaaS, IoT, and Machine Learning. Our dedicated team of 400+ elite software engineers excels in solving complex technical challenges while translating customer concepts into successful products. Velotio has joined forces with Chamberlain Group (CG) to recruit top talent who are enthusiastic about contributing to a technology-driven company focused on innovation and purpose. As a Blackstone portfolio company and a global leader in smart access solutions, CG's products are present in over 50 million residential and commercial properties globally. Through brands like LiftMaster, CG leads the market in garage door openers, gate operators, commercial door openers, and dock levelers. Their cutting-edge products and services leverage myQ technology to connect hardware to the cloud, offering users secure access to their garage, home, community, or business. By leveraging CG's industry expertise with Velotio's engineering capabilities, we aim to onboard a team of 150+ engineers to develop the next generation of smart access products. Job Description: As part of our team, you will: - Build automation tools and test automation frameworks from scratch - Collaborate closely with Software engineers to grasp application design and specifications prior to developing comprehensive test automation suites with meticulously crafted scenarios and test data - Design and implement test automation frameworks for both web and mobile applications - Configure, execute, and monitor cloud and DevOps environments related to automation and projects - Identify and formulate test strategies, plans, and test cases - Automate tests using in-house and open-source test frameworks - Lead Automation initiatives across multiple projects utilizing the latest test frameworks - Create reusable tools for test automation and explore new automation methods - Engage with Architects, Tech Leads, and Software Engineers to comprehend automation requirements - Execute tests and conduct failure analysis to support manual testing efforts and meet release timelines with high quality You will thrive in this role if you: - Are passionate about automation and have experience with Cloud Platforms such as AWS and Kubernetes - Are keen on enhancing existing QA processes while introducing new ones - Possess a strong interest in exploring new automation frameworks, technologies, and trends Desired Skills & Experience: - 4+ years of experience in building test automation for web or mobile apps with embedded or IoT components - 2+ years" experience in REST API test automation using languages like C#, JavaScript, Python, Java, or TypeScript - Familiarity with end-to-end testing tools like Selenium, Appium, Webdriver IO, Cypress, and Nightwatch - Ability to implement continuous integration and build automation systems like Jenkins, Bamboo, Octopus, etc - Knowledge of cloud-based platforms such as Perfecto, Source Lab, etc - Experience with BDD tools such as Cucumber or similar - Proficiency in relational databases like Oracle/SQL server and no-SQL databases like Couchbase/Neo4J - Hands-on experience in API testing techniques and tools - Experience in the IoT domain and testing of cloud-hosted applications and SaaS - Understanding of CI/CD tools like Jenkins, Gitlab, CircleCi - Familiarity with Source Code Management Tools (e.g., Git, TFS, RTC) - Ability to create test data based on business scenarios and develop tools and scripts for test data generation - Interest in emerging automation technologies and trends with the ability to apply concepts practically - Strong foundation in computer science, with expertise in data structures, algorithms, and software design - Effective communication skills, welcoming collaboration, and excelling in verbal and written communication - Bachelor's degree in Computer Science or equivalent experience Bonus points if you have: - Proficient in Linux system administration - Experience with Cloud-based networking, storage, and virtualization - Knowledge of microservices architecture using Kubernetes and Docker containers Benefits: Our Culture: - Autonomous and empowered work culture promoting ownership and rapid growth - Flat hierarchy with quick decision-making and a startup-oriented "get things done" mindset - Vibrant, positive environment with regular celebrations of success, fostering inclusivity, diversity, and authenticity We seek smart, inquisitive, and ambitious individuals. Even if you don't meet all requirements, reach out if you have the potential to grow and contribute to our team.,
Posted 2 days ago
8.0 - 12.0 years
0 Lacs
chennai, tamil nadu
On-site
You will be joining InnovaESI, a prominent IT partner firm dedicated to driving digital transformation by offering comprehensive data solutions globally across the USA, Canada, Mexico, India, and Singapore. The core mission at InnovaESI is to equip businesses with cutting-edge and sustainable IT solutions that optimize operations and foster expansion. As a collaborative effort between Enterprise Solutions, Inc., USA, and Innova, India, we are steadfast in our commitment to providing premium, cost-effective services to our esteemed clients. In this role, you will be based in Chennai with a hybrid work mode arrangement. With a minimum of 8 years of experience, you will be responsible for various key tasks and requirements essential for this position. It is imperative that you possess hands-on expertise in Load Runner web testing protocol, monitoring tools, and Log parsing tools for efficient performance analysis. Your proficiency in managing complex correlations and conducting initial analysis using performance monitoring tools to pinpoint potential system and resources bottlenecks will be crucial. Your responsibilities will also include designing scripts for API through Load Runner for Load and Stress Testing, as well as hands-on experience in Server monitoring and analysis. Knowledge in Workload modelling, Pacing/TT calculation, Stress, Soak, and Scalability tests will be advantageous. A strong scripting background is necessary to develop scenario-based scripts with adept handling of parameterization. A positive attitude, result-oriented mindset, and the ability to showcase accountability and ownership through various techniques and collaboration within the team are highly valued. While not mandatory, familiarity with Kafka Confluent/Server-side Monitoring and neo4j skills will be beneficial. Additional knowledge in other Performance Test Tools like Stress Stimulus is considered a plus. Proficiency in tools and technologies such as Data Base, Soap UI, JMeter, ADO Test Management Tools, Basic SQL/Oracle, PostGre, and programming languages like JAVA and C as secondary skills are preferred. Extensive API test experience is a must for this role. If you believe you meet the requirements and are excited about this opportunity, please share your resume with pragya.tiwari@innovaesi.com.,
Posted 2 days ago
0 years
0 Lacs
Chennai, Tamil Nadu, India
Remote
Job Title: AI Research Engineer Intern (Fresher) Reporting to: Lead – Research & Innovation Lab Location: remote/ Hybrid (Chennai, India) Engagement: 6-month, full-time paid internship with pre-placement-offer track 1. Why this role exists Stratsyn AI Technology Services is turbo-charging Stratsyn’s cloud-native Enterprise Intelligence & Management Suite —a modular SaaS ecosystem that fuses advanced AI, low-code automation, multimodal search, and next-generation “Virtual workforce” agents. The platform unifies strategic planning, document intelligence, workflow orchestration, and real-time analytics, empowering C-suite leaders to simulate scenarios, orchestrate execution, and convert insight into action with unmatched speed and scalability. To keep pushing that frontier, we need sharp, curious minds who can translate cutting-edge research into production-grade capabilities for this suite. This internship is our talent-funnel into future Research Engineer and Product Scientist roles. 2. What you’ll do (core responsibilities) % FocusKey Responsibility 30 %Rapid Prototyping & Experimentation – implement state-of-the-art papers (LLMs, graph learning, causal inference, agents), design ablation studies, benchmark against baselines, and iterate fast. 25 %Data Engineering for Research – build reproducible datasets, craft synthetic data when needed, automate ETL pipelines, and enforce experiment tracking (MLflow / Weights & Biases). 20 %Model Evaluation & Explainability – create evaluation harnesses (BLEU, ROUGE, MAPE, custom KPIs), visualize error landscapes, and generate executive-ready insights. 15 %Collaboration & Documentation – author tech memos, well-annotated notebooks, and contribute to internal knowledge bases; present findings in weekly research stand-ups. 10 %Innovation Scouting – scan arXiv, ACL, NeurIPS, ICML, and startup ecosystems; summarize high-impact research and propose areas for IP creation within the Suite. 3. What you will learn / outcomes to achieve Master the end-to-end research workflow: literature review → hypothesis → prototype → validation → deployment shadow. Deliver one peer-review-quality technical report and two production-grade proof-of-concepts for the Suite. Achieve a measurable impact (e.g., 8-10 % forecasting-accuracy lift or 30 % latency reduction) on a live micro-service. 4. Minimum qualifications (freshers welcome) B.E./B.Tech/M.Sc./M.Tech in CS, Data Science, Statistics, EE, or related (2024-2026 pass-out). Fluency in Python and at least one deep-learning framework (PyTorch preferred). Solid grasp of linear algebra, probability, optimization, and algorithms. Hands-on academic or personal projects in NLP, CV, time-series, or RL (GitHub links highly valued). 5. Preferred extras Publications or Kaggle/ML-competition record. Experience with distributed training (GPU clusters, Ray, Lightning) and experiment-tracking tools. Familiarity with MLOps (Docker, CI/CD, Kubernetes) or data-centric AI. Domain knowledge in supply-chain, fintech, climate, or marketing analytics. 6. Key attributes & soft skills First-principles thinker – questions assumptions, proposes novel solutions. Bias for action – prototypes in hours, not weeks; embraces agile experimentation. Storytelling ability – explains complex models in clear, executive-friendly language. Ownership mentality – treats the prototype as a product, not just a demo. 7. Tech stack you’ll touch Python | PyTorch | Hugging Face | TensorRT | LangChain | Neo4j/GraphDB | PostgreSQL | Airflow | MLflow | Weights & Biases | Docker | GitHub Actions | JAX (exploratory) 8. Internship logistics & perks Competitive monthly stipend + performance bonus. High-end workstation + GPU credits on our private cloud. Dedicated mentor and 30-60-90-day learning plan. Access to premium research portals and paid conference passes. Culture of radical candor, weekly brown-bag tech talks, and hack days. Fast-track to full-time AI Research Engineer upon successful completion. 9. Application process Apply via email: Send résumé, brief statement of purpose, and GitHub/portfolio links to HR@stratsyn.ai . Online coding assessment: algorithmic + ML fundamentals. Technical interview (2 rounds): deep dive into projects, math, and research reasoning. Culture-fit discussion: with Research Lead & CPO. Offer & onboarding – target turnaround < 3 weeks.
Posted 2 days ago
0.0 - 4.0 years
0 - 0 Lacs
Sahibzada Ajit Singh Nagar, Mohali, Punjab
On-site
Job Title: Python Backend Developer (Data Layer) Location: Mohali, Punjab Company: RevClerx About RevClerx: RevClerx Pvt. Ltd., founded in 2017 and based in the Chandigarh/Mohali area (India), is a dynamic Information Technology firm providing comprehensive IT services with a strong focus on client-centric solutions. As a global provider, we cater to diverse business needs including website designing and development, digital marketing, lead generation services (including telemarketing and qualification), and appointment setting. Job Summary: We are seeking a skilled Python Backend Developer with a strong passion and proven expertise in database design and implementation. This role requires 3-4 years of backend development experience, focusing on building robust, scalable applications and APIs. The ideal candidate will not only be proficient in Python and common backend frameworks but will possess significant experience in designing, modeling, and optimizing various database solutions, including relational databases (like PostgreSQL) and, crucially, graph databases (specifically Neo4j). You will play a vital role in architecting the data layer of our applications, ensuring efficiency, scalability, and the ability to handle complex, interconnected data. Key Responsibilities: ● Design, develop, test, deploy, and maintain scalable and performant Python-based backend services and APIs. ● Lead the design and implementation of database schemas for relational (e.g., PostgreSQL) and NoSQL databases, with a strong emphasis on Graph Databases (Neo4j). ● Model complex data relationships and structures effectively, particularly leveraging graph data modeling principles where appropriate. ● Write efficient, optimized database queries (SQL, Cypher, potentially others). ● Develop and maintain data models, ensuring data integrity, consistency, and security. ● Optimize database performance through indexing strategies, query tuning, caching mechanisms, and schema adjustments. ● Collaborate closely with product managers, frontend developers, and other stakeholders to understand data requirements and translate them into effective database designs. ● Implement data migration strategies and scripts as needed. ● Integrate various databases seamlessly with Python backend services using ORMs (like SQLAlchemy, Django ORM) or native drivers. ● Write unit and integration tests, particularly focusing on data access and manipulation logic. ● Contribute to architectural decisions, especially concerning data storage, retrieval, and processing. ● Stay current with best practices in database technologies, Python development, and backend systems. Minimum Qualifications: ● Bachelor's degree in Computer Science, Engineering, Information Technology, or a related field, OR equivalent practical experience. ● 3-4 years of professional software development experience with a primary focus on Python backend development. ● Strong proficiency in Python and its standard libraries. ● Proven experience with at least one major Python web framework (e.g., Django, Flask, FastAPI). ● Demonstrable, hands-on experience designing, implementing, and managing relational databases (e.g., PostgreSQL). ● Experience with at least one NoSQL database (e.g., MongoDB, Redis, Cassandra). ● Solid understanding of data structures, algorithms, and object-oriented programming principles. ● Experience designing and consuming RESTful APIs. ● Proficiency with version control systems, particularly Git. ● Strong analytical and problem-solving skills, especially concerning data modeling and querying. ● Excellent communication and teamwork abilities. Preferred (Good-to-Have) Qualifications: ● Graph Database Expertise: ○ Significant, demonstrable experience designing and implementing solutions using Graph Databases (Neo4j strongly preferred). ○ Proficiency in graph query languages, particularly Cypher. ○ Strong understanding of graph data modeling principles, use cases (e.g., recommendation engines, fraud detection, knowledge graphs, network analysis), and trade-offs. ● Advanced Database Skills: ○ Experience with database performance tuning and monitoring tools. ○ Experience with Object-Relational Mappers (ORMs) like SQLAlchemy or Django ORM in depth. ○ Experience implementing data migration strategies for large datasets. ● Cloud Experience: Familiarity with cloud platforms (e.g., AWS, Azure, Google Cloud Platform) and their managed database services (e.g., RDS, Aurora, Neptune, DocumentDB, MemoryStore). ● Containerization & Orchestration: Experience with Docker and Kubernetes. ● Asynchronous Programming: Experience with Python's asyncio and async frameworks. ● Data Pipelines: Familiarity with ETL processes or data pipeline tools (e.g., Apache Airflow). ● Testing: Experience writing tests specifically for database interactions and data integrity. What We Offer: ● Challenging projects with opportunities to work on cutting-edge technologies especially in the field of AI. ● Competitive salary and comprehensive benefits package. ● Opportunities for professional development and learning (e.g., conferences, courses, certifications). ● A collaborative, innovative, and supportive work environment. How to Apply: Interested candidates are invited to submit their resume and a cover letter outlining their relevant experience, specifically highlighting their database design expertise (including relational, NoSQL, and especially Graph DB/Neo4j experience) Job Types: Full-time, Permanent Pay: ₹30,000.00 - ₹55,373.94 per month Benefits: Food provided Health insurance Schedule: Day shift Monday to Friday
Posted 3 days ago
5.0 - 7.0 years
12 - 19 Lacs
Pune
Hybrid
Role & responsibilities Preferred candidate profile Required Skills : Neo4j Expertise (5+ years hands-on): Proven, in-depth experience with Neo4j, including its core concepts (nodes, relationships, properties, labels), architectural components, and deployment models (standalone, causal cluster). Mastery of the Cypher query language for complex graph traversals, pattern matching, and data manipulation. Strong understanding of Neo4j indexing strategies (schema indexes, full-text indexes) and their impact on query performance. Graph Database Solutions: Demonstrated experience in designing, implementing, and maintaining scalable graph database solutions and architectures. Familiarity with graph theory concepts, graph data modeling principles, and their application in real-world scenarios Neo4J graph analytics Dev Neo4J expertise - Proven experience with Neo4J, including its core concepts, Cypher query language and best practices Designing and implementing graph database solutions : This includes creating and maintaining graph schemas, models and architectures Familiarity with graph theory , graph data modelling and other graph database technologies Developing and optimizing Cypher queries Integrating Neo4J with BI and other systems Providing technical guidance to junior developers Creating and maintaining documentation for system architecture , design and operational processes
Posted 3 days ago
0.0 - 3.0 years
12 - 24 Lacs
Chennai, Tamil Nadu
On-site
We are looking for a forward-thinking Data Scientist with expertise in Natural Language Processing (NLP), Large Language Models (LLMs), Prompt Engineering, and Knowledge Graph construction. You will be instrumental in designing intelligent NLP pipelines involving Named Entity Recognition (NER), Relationship Extraction, and semantic knowledge representation. The ideal candidate will also have practical experience in deploying Python-based APIs for model and service integration. This is a hands-on, cross-functional role where you’ll work at the intersection of cutting-edge AI models and domain-driven knowledge extraction. Key Responsibilities: Develop and fine-tune LLM-powered NLP pipelines for tasks such as NER, coreference resolution, entity linking, and relationship extraction. Design and build Knowledge Graphs by structuring information from unstructured or semi-structured text. Apply Prompt Engineering techniques to improve LLM performance in few-shot, zero-shot, and fine-tuned scenarios. Evaluate and optimize LLMs (e.g., OpenAI GPT, Claude, LLaMA, Mistral, or Falcon) for custom domain-specific NLP tasks. Build and deploy Python APIs (using Flask/Fast API) to serve ML/NLP models and access data from graph database. Collaborate with teams to translate business problems into structured use cases for model development. Understanding custom ontologies and entity schemas for corresponding domain. Work with graph databases like Neo4j or similar DBs and query using Cypher or SPARQL. Evaluate and track performance using both standard metrics and graph-based KPIs. Required Skills & Qualifications: Strong programming experience in Python and libraries such as PyTorch, TensorFlow, spaCy, scikit-learn, Hugging Face Transformers, LangChain, and OpenAI APIs. Deep understanding of NER, relationship extraction, co-reference resolution, and semantic parsing. Practical experience in working with or integrating LLMs for NLP applications, including prompt engineering and prompt tuning. Hands-on experience with graph database design and knowledge graph generation. Proficient in Python API development (Flask/FastAPI) for serving models and utilities. Strong background in data preprocessing, text normalization, and annotation frameworks. Understanding of LLM orchestration with tools like LangChain or workflow automation. Familiarity with version control, ML lifecycle tools (e.g., MLflow), and containerization (Docker). Nice to Have: Experience using LLMs for Information Extraction, summarization, or question answering over knowledge bases. Exposure to Graph Embeddings, GNNs, or semantic web technologies (RDF, OWL). Experience with cloud-based model deployment (AWS/GCP/Azure). Understanding of retrieval-augmented generation (RAG) pipelines and vector databases (e.g., Chroma, FAISS, Pinecone). Job Type: Full-time Pay: ₹1,200,000.00 - ₹2,400,000.00 per year Ability to commute/relocate: Chennai, Tamil Nadu: Reliably commute or planning to relocate before starting work (Preferred) Education: Bachelor's (Preferred) Experience: Natural Language Processing (NLP): 3 years (Preferred) Language: English & Tamil (Preferred) Location: Chennai, Tamil Nadu (Preferred) Work Location: In person
Posted 3 days ago
3.0 - 7.0 years
0 Lacs
haryana
On-site
ZS is a place where passion changes lives. As a management consulting and technology firm focused on improving life and how we live it, our most valuable asset is our people. Here you'll work side-by-side with a powerful collective of thinkers and experts shaping life-changing solutions for patients, caregivers, and consumers worldwide. ZSers drive impact by bringing a client-first mentality to each and every engagement. We partner collaboratively with our clients to develop custom solutions and technology products that create value and deliver company results across critical areas of their business. Bring your curiosity for learning, bold ideas, courage, and passion to drive life-changing impact to ZS. Our most valuable asset is our people. At ZS, we honor the visible and invisible elements of our identities, personal experiences, and belief systems - the ones that comprise us as individuals, shape who we are, and make us unique. We believe your personal interests, identities, and desire to learn are part of your success here. Learn more about our diversity, equity, and inclusion efforts and the networks ZS supports to assist our ZSers in cultivating community spaces, obtaining the resources they need to thrive, and sharing the messages they are passionate about. **What you'll do:** We are looking for experienced Knowledge Graph developers who have the following set of technical skillsets and experience. Undertake complete ownership in accomplishing activities and assigned responsibilities across all phases of the project lifecycle to solve business problems across one or more client engagements. Apply appropriate development methodologies (e.g., agile, waterfall) and best practices (e.g., mid-development client reviews, embedded QA procedures, unit testing) to ensure successful and timely completion of assignments. Collaborate with other team members to leverage expertise and ensure seamless transitions; Exhibit flexibility in undertaking new and challenging problems and demonstrate excellent task management. Assist in creating project outputs such as business case development, solution vision and design, user requirements, prototypes, and technical architecture (if needed), test cases, and operations management. Bring transparency in driving assigned tasks to completion and report accurate status. Bring a Consulting mindset in problem-solving, innovation by leveraging technical and business knowledge/expertise and collaborate across other teams. Assist senior team members, delivery leads in project management responsibilities. Build complex solutions using Programming languages, ETL service platform, etc. **What you'll bring:** - Bachelor's or master's degree in computer science, Engineering, or a related field. - 4+ years of professional experience in Knowledge Graph development in Neo4j or AWS Neptune or Anzo knowledge graph Database. - 3+ years of experience in RDF ontologies, Data modeling & ontology development. - Strong expertise in python, pyspark, SQL. - Strong ability to identify data anomalies, design data validation rules, and perform data cleanup to ensure high-quality data. - Project management and task planning experience, ensuring smooth execution of deliverables and timelines. - Strong communication and interpersonal skills to collaborate with both technical and non-technical teams. - Experience with automation testing. - Performance Optimization: Knowledge of techniques to optimize knowledge graph operations like data inserts. - Data Modeling: Proficiency in designing effective data models within Knowledge Graph, including relationships between tables and optimizing data for reporting. - Motivation and willingness to learn new tools and technologies as per the team's requirements. **Additional Skills:** - Strong communication skills, both verbal and written, with the ability to structure thoughts logically during discussions and presentations. - Experience in pharma or life sciences data: Familiarity with pharmaceutical datasets, including product, patient, or healthcare provider data, is a plus. - Experience in manufacturing data is a plus. - Capability to simplify complex concepts into easily understandable frameworks and presentations. - Proficiency in working within a virtual global team environment, contributing to the timely delivery of multiple projects. - Travel to other offices as required to collaborate with clients and internal project teams. **Perks & Benefits:** ZS offers a comprehensive total rewards package including health and well-being, financial planning, annual leave, personal growth, and professional development. Our robust skills development programs, multiple career progression options, and internal mobility paths and collaborative culture empower you to thrive as an individual and global team member. We are committed to giving our employees a flexible and connected way of working. A flexible and connected ZS allows us to combine work from home and on-site presence at clients/ZS offices for the majority of our week. The magic of ZS culture and innovation thrives in both planned and spontaneous face-to-face connections. **Travel:** Travel is a requirement at ZS for client-facing ZSers; business needs of your project and client are the priority. While some projects may be local, all client-facing ZSers should be prepared to travel as needed. Travel provides opportunities to strengthen client relationships, gain diverse experiences, and enhance professional growth by working in different environments and cultures. **Considering applying ** At ZS, we're building a diverse and inclusive company where people bring their passions to inspire life-changing impact and deliver better outcomes for all. We are most interested in finding the best candidate for the job and recognize the value that candidates with all backgrounds, including non-traditional ones, bring. If you are interested in joining us, we encourage you to apply even if you don't meet 100% of the requirements listed above. ZS is an equal opportunity employer and is committed to providing equal employment and advancement opportunities without regard to any class protected by applicable law. **To Complete Your Application:** Candidates must possess or be able to obtain work authorization for their intended country of employment. An online application, including a full set of transcripts (official or unofficial), is required to be considered. NO AGENCY CALLS, PLEASE. Find Out More At: www.zs.com,
Posted 3 days ago
Upload Resume
Drag or click to upload
Your data is secure with us, protected by advanced encryption.
Neo4j, a popular graph database management system, is seeing a growing demand in the job market in India. Companies are looking for professionals who are skilled in working with Neo4j to manage and analyze complex relationships in their data. If you are a job seeker interested in Neo4j roles, this article will provide you with valuable insights to help you navigate the job market in India.
The average salary range for Neo4j professionals in India varies based on experience levels. - Entry-level: INR 4-6 lakhs per annum - Mid-level: INR 8-12 lakhs per annum - Experienced: INR 15-20 lakhs per annum
In the Neo4j skill area, a typical career progression may look like: - Junior Developer - Developer - Senior Developer - Tech Lead
Apart from expertise in Neo4j, professionals in this field are often expected to have or develop skills in: - Cypher Query Language - Data modeling - Database management - Java or Python programming
As you explore Neo4j job opportunities in India, it's essential to not only possess the necessary technical skills but also be prepared to showcase your expertise during interviews. Stay updated with the latest trends in Neo4j and continuously enhance your skills to stand out in the competitive job market. Prepare thoroughly, demonstrate your knowledge confidently, and land your dream Neo4j job in India. Good luck!
Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.
We have sent an OTP to your contact. Please enter it below to verify.
Accenture
39581 Jobs | Dublin
Wipro
19070 Jobs | Bengaluru
Accenture in India
14409 Jobs | Dublin 2
EY
14248 Jobs | London
Uplers
10536 Jobs | Ahmedabad
Amazon
10262 Jobs | Seattle,WA
IBM
9120 Jobs | Armonk
Oracle
8925 Jobs | Redwood City
Capgemini
7500 Jobs | Paris,France
Virtusa
7132 Jobs | Southborough