Senior . NET Developer
We are seeking a highly skilled .NET Backend Developer with expertise in C#, SQL Server, MongoDB, MySQL, and large-scale data processing as core skill . This role focuses on efficient data ingestion, structured data integration, and high-speed processing of large datasets while ensuring optimal memory and resource utilization. The ideal candidate should have deep experience in handling structured and unstructured data, multi-threaded processing, efficient database optimization, and real-time data synchronization to support scalable and performance-driven backend architecture . Key Focus Areas - Efficient Data Ingestion & Processing: Developing scalable pipelines to process large structured/unstructured data files . - Data Integration & Alignment: Merging datasets from multiple sources with consistency . - Database Expertise & Performance Optimization: Designing high-speed relational database structures for efficient storage and retrieval. - High-Performance API Development: Developing low-latency RESTful APIs to handle large data exchanges efficiently. - Multi-Threaded Processing & Parallel Execution: Implementing concurrent data processing techniques to optimize system performance. - Caching Strategies & Load Optimization: Utilizing in-memory caching & indexing to reduce I/O overhead. - Real-Time Data Processing & Streaming: Using message queues and data streaming for optimized data distribution. Required Skills & Technologies Backend Development: C#, .NET Core, ASP.NET Core Web API Data Processing & Integration: Efficient Data Handling, Multi-Source Data Processing Database Expertise: SQL Server MongoDB ,MySQL (Schema Optimization, Indexing, Query Optimization, Partitioning, Bulk Processing) Performance Optimization: Multi-threading, Parallel Processing, High-Throughput Computing Caching & Memory Management: Redis, Memcached, IndexedDB, Database Query Caching Real-Time Data Processing: Kafka, RabbitMQ, WebSockets, SignalR File Processing & ETL Pipelines: Efficient Data Extraction, Transformation, and Storage Pipelines Logging & Monitoring: Serilog, Application Insights, ELK Stack CI/CD & Cloud Deployments: Azure DevOps, Kubernetes, Docker Key Responsibilities 1. Data Ingestion & Processing Develop scalable data pipelines to handle high-throughput structured and unstructured data ingestion. Implement multi-threaded data processing mechanisms to optimize efficiency. Optimize memory management techniques to handle large-scale data operations. 2. Data Integration & Alignment Implement high-speed algorithms to merge and integrate datasets efficiently. Ensure data consistency and accuracy across multiple sources. Optimize data buffering & streaming techniques to prevent processing bottlenecks. 3. High-Performance API Development Design and develop high-speed APIs for efficient data retrieval and updates . Implement batch processing & streaming capabilities to manage large data payloads. Optimize API response times and query execution plans . 4. Database Expertise & Optimization (SQL Server , MongoDB ,MySql ) Design efficient database schema structures to support large-scale data transactions. Implement bulk data operations, indexing, and partitioning for high-speed retrieval . Optimize stored procedures and concurrency controls to support high-frequency transactions . Use sharding and distributed database techniques for enhanced scalability. 5. Caching & Load Balancing Deploy Redis / Memcached / IndexedDB caching to improve database query performance. Implement data pre-fetching & cache invalidation strategies for real-time accuracy. Optimize load balancing techniques for efficient request distribution. 6. Real-Time Data Synchronization & Streaming Implement event-driven architectures using message queues (Kafka, RabbitMQ, etc.) . Utilize WebSockets / SignalR for real-time data synchronization . Optimize incremental updates instead of full data reloads for better resource efficiency. Preferred Additional Experience Experience handling large-scale databases and high-throughout data environments . Expertise in distributed database architectures for large-scale structured data storage. Hands-on experience with query profiling & performance tuning tools . Apply arrow_forward_ios highly skilled .NET Backend Developer with expertise in efficient data ingestion, structured data integration, and high-speed processing of large datasets while ensuring optimal memory and resource utilization. handling structured and unstructured data, multi-threaded processing, efficient database optimization, and real-time data synchronization to support Efficient Data Ingestion & Processing: Developing scalable Data Integration & Alignment: Merging Database Expertise & Performance Optimization: Designing high-speed relational database structures for efficient storage and retrieval. High-Performance API Development: Developing low-latency RESTful APIs to handle large data exchanges efficiently. Multi-Threaded Processing & Parallel Execution: Implementing concurrent data processing techniques to optimize system performance. Caching Strategies & Load Optimization: Utilizing in-memory caching & indexing to reduce I/O overhead. Real-Time Data Processing & Streaming: Using message queues and data streaming for optimized data distribution. scalable data pipelines to handle high-throughput structured and unstructured data ingestion. multi-threaded data processing mechanisms to optimize efficiency. memory management techniques to handle large-scale data operations. high-speed algorithms to merge and integrate datasets efficiently. data consistency and accuracy across multiple sources. data buffering & streaming techniques to prevent processing bottlenecks. high-speed APIs for efficient batch processing & streaming capabilities to manage large data payloads. efficient database schema structures to support large-scale data transactions. bulk data operations, indexing, and partitioning for stored procedures and concurrency controls to support sharding and distributed database techniques for enhanced scalability. Redis / Memcached / IndexedDB caching to improve database query performance. data pre-fetching & cache invalidation strategies for real-time accuracy. load balancing techniques for efficient request distribution. event-driven architectures using WebSockets / SignalR for incremental updates instead of full data reloads for better resource efficiency. Expertise in distributed database architectures for large-scale structured data storage.