DHIRA Company Overview DHIRA is a leading company specializing in intelligent transformation, where we leverage advanced AI/ML and data-driven solutions to revolutionize business operations. Unlike traditional digital transformation, which focuses on transaction automation, our intelligent transformation encompasses both transactional automation and deep analytics for comprehensive insights. Our expertise in data engineering, data quality, and master data management ensures robust and scalable AI/ML applications. Utilizing cutting-edge technologies across AWS, Azure, GCP, and on-premises Hadoop systems, we deliver efficient and innovative data solutions. Our vision is embodied in the Akashic platform, designed to provide seamless, end-to-end analytics. At DHIRA, we are committed to excellence, driving impactful contributions to the industry. Join us to be part of a dynamic team at the forefront of intelligent transformation Role- Data Architect – Evolution of Databases, Data Modeling, and Modern Data Practices Location : Bangalore, Remote Position Overview: We are seeking a Principal Data Architect with 5+ years of experience who has a comprehensive understanding of the evolution of databases , from OLTP to OLAP, and relational systems to NoSQL, Graph, and emerging Vector Databases . This role requires deep expertise in data modeling , from traditional ER modeling to advanced dimensional, graph, and vector schemas, along with a strong grasp of the history, best practices, and future trends in data management. The ideal candidate will bring both historical context and cutting-edge expertise to architect scalable, high-performance data solutions, driving innovation while maintaining strong governance and best practices. This is a leadership role that demands a balance of technical excellence, strategic vision, and team mentorship. Key Responsibilities: 1. Data Modeling Expertise: – Design and implement Entity-Relationship Models (ER Models) for OLTP systems, ensuring normalization and consistency. – Transition ER models into OLAP environments with robust dimensional modeling, including star and snowflake schemas. – Develop hybrid data models that integrate relational, NoSQL, Graph, and Vector Database schemas. – Establish standards for schema design across diverse database systems, focusing on scalability and query performance. 2. Database Architecture Evolution: – Architect solutions across the database spectrum: • Relational databases (PostgreSQL, Oracle, MySQL) • NoSQL databases (MongoDB, Cassandra, DynamoDB) • Graph databases (Neo4j, Amazon Neptune) • Vector databases (Pinecone, Weaviate, Milvus). – Implement hybrid data architectures combining OLTP, OLAP, NoSQL, Graph, and Vector systems for diverse business needs. – Ensure compatibility and performance optimization across these systems for real-time and batch processing. 3. Data Warehousing and Analytics: – Lead the development of enterprise-scale Data Warehouses capable of supporting advanced analytics and business intelligence. – Design high-performance ETL/ELT pipelines to handle structured and unstructured data with minimal latency. – Optimize OLAP systems for petabyte-scale data storage and low-latency querying. 4. Emerging Database Technologies: – Drive adoption of Vector Databases for AI/ML applications, enabling semantic search and embedding-based queries. – Explore cutting-edge technologies in data lakes, lakehouses, and real-time processing systems. – Evaluate and integrate modern database paradigms, ensuring scalability for future business requirements. 5. Strategic Leadership: – Define the organization’s data strategy , aligning with long-term goals and emerging trends. – Collaborate with business and technical stakeholders to design systems that balance transactional and analytical workloads. – Lead efforts in data governance, ensuring compliance with security and privacy regulations. 6. Mentorship and Innovation: – Mentor junior architects and engineers, fostering a culture of learning and technical excellence. – Promote innovation by introducing best practices, emerging tools, and modern methodologies in data architecture. – Act as a thought leader in database evolution, presenting insights to internal teams and external forums. Required Skills & Qualifications: • Experience: – 6+ years of experience in data architecture, with demonstrated expertise across OLTP, OLAP, NoSQL, Graph, and Vector databases. – Proven experience designing and implementing data models across relational, NoSQL, graph, and vector systems. – A strong understanding of the evolution of databases and their impact on modern data architectures. • Technical Proficiency: – Deep expertise in ER modeling , dimensional modeling, and schema design for modern database systems. – Proficient in SQL and query optimization for relational and analytical databases. – Hands-on experience with NoSQL databases like MongoDB, Cassandra, or DynamoDB. – Strong knowledge of Graph databases (Neo4j, Amazon Neptune) and Vector databases (Pinecone, Milvus, or Weaviate). – Familiarity with modern cloud-based DW platforms (e.g., Snowflake, BigQuery, Redshift) and lakehouse solutions. • Knowledge of Data Practices: – Historical and practical understanding of data practices, from schema-on-write to schema-on-read approaches. – Experience in implementing real-time and batch processing systems for diverse workloads. – Strong grasp of data lifecycle management, governance, and security practices. • Leadership and Communication: – Ability to lead large-scale data initiatives, balancing technical depth and strategic alignment. – Excellent communication skills to articulate complex ideas to technical and non-technical audiences. – Proven ability to mentor and upskill teams, fostering a collaborative environment. Preferred Skills: • Experience integrating Vector Databases into existing architectures for AI/ML workloads. • Knowledge of real-time streaming systems (Kafka, Pulsar) and their integration with modern databases. • Certifications in data-related technologies (e.g., AWS, GCP, Snowflake, Neo4j). • Hands-on experience with BI tools (e.g., Tableau, Power BI) and AI/ML platforms.
About the Role: We’re looking for a data enthusiast who loves connecting dots — not just within datasets, but between ideas, tools, and people. This role is a blend of data analyst precision and data engineer problem-solving , perfect for someone who enjoys rolling up their sleeves, experimenting creatively, and finding elegant ways to make things work. If you’re the kind of person who says “let me try a different angle” instead of “this can’t be done,” you’ll fit right in. What you'll Do: Gather, clean, and model data from multiple sources to enable analysis and reporting. Build and maintain simple yet scalable data pipelines and transformations. Develop dashboards, reports, and metrics that actually tell stories, not just show numbers. Collaborate with business and technical teams to understand data needs and design creative solutions. Automate recurring processes and simplify data workflows wherever possible. Continuously explore better ways to represent, store, and use data efficiently. What We're Looking For: Bachelor’s degree in Engineering, Computer Science, Statistics, or related field (or equivalent self-taught skill). 0–2 years of experience working with data (internships, projects, or personal work count). Basic comfort with SQL , Python , and at least one data visualization tool ( Power BI , Tableau , or similar). Understanding of ETL concepts , data modeling , and cloud/data warehouse tools is a plus. A mindset that’s equal parts analytical , creative , and resourceful — someone who figures things out. Good communication and curiosity — you ask “why” before “how.” What you'll Love Here: A culture that values initiative over instruction. The freedom to experiment with ideas — and the expectation to back them with data. Opportunities to work across the full data lifecycle — from engineering to insights. A small, dynamic team that appreciates both logic and imagination. Bonus Points For: Experience with APIs, cloud data tools (Azure, AWS, or GCP), or basic scripting for automation. An eye for design or storytelling in data presentation. GitHub projects, hackathon participation, or personal analytics experiments.