Jobs
Interviews

315 Scrapy Jobs - Page 2

Setup a job Alert
JobPe aggregates results for easy application access, but you actually apply on the job portal directly.

2.0 years

0 Lacs

ahmedabad, gujarat, india

On-site

We're looking for an entrepreneurial, passionate, and driven Data Engineer to join Startup Gala Intelligence backed by Navneet Tech Venture. As we're building our technology platform from scratch, you'll have the unique opportunity to shape our technology vision, architecture, and engineering culture right from the ground up. You’ll directly contribute to foundational development and establish best practices, while eventually building and contributing to our engineering team. This role is ideal for someone eager to own the entire tech stack, who thrives on early-stage challenges, and loves building innovative, scalable solutions from day zero. What You’ll Do Web Scraping & Crawling: Build and maintain automated scrapers to extract structured and unstructured data from websites, APIs, and public datasets. Scalable Scraping Systems: Develop multi-threaded, distributed crawlers capable of handling high-volume data collection without interruptions. Data Parsing & Cleaning: Normalize scraped data, remove noise, and ensure consistency before passing to data pipelines. Anti-bot & Evasion Tactics: Implement proxy rotation, captcha solving, and request throttling techniques to handle scraping restrictions. Integration with Pipelines: Deliver clean, structured datasets into NoSQL stores and ETL pipelines for further enrichment and graph-based storage. Data Quality & Validation: Ensure data accuracy, deduplicate records, and maintain a trust scoring system for data confidence. Documentation & Maintenance: Keep scrapers updated when websites change, and document scraping logic for reproducibility. Who You Are Technical Skills: 2+ years of experience in web scraping , crawling, or data collection. Strong proficiency in Python (libraries like BeautifulSoup, Scrapy, Selenium, Playwright, Requests). Familiarity with NoSQL databases (MongoDB, DynamoDB) and data serialization formats (JSON, CSV, Parquet). Experience in handling large-scale scraping with proxy management and rate-limiting. Basic knowledge of ETL processes and integration with data pipelines. Exposure to graph databases (Neo4j) is a plus. Soft Skills: Detail-oriented, ensuring accuracy and reliability of collected data. Strong problem-solving skills, particularly in adapting scrapers to evolving web structures. Curious mindset with a drive to discover new data sources. Comfortable working in a fast-paced, early-stage startup environment. Who We Are & Our Culture Gala Intelligence , backed by Navneet Tech Ventures , is a tech-driven startup dedicated to solving one of the most pressing business challenges - fraud detection and prevention. We're building cutting-edge, real-time products designed to empower consumers and businesses to stay ahead of fraudsters, leveraging innovative technology and deep domain expertise. Our culture and values: We’re united by a single, critical mission - stopping fraud before it impacts businesses. Curiosity, innovation, and proactive action define our approach. We value transparency, collaboration, and individual ownership, creating an environment where talented people can do their best work. Problem-Driven Innovation : We're deeply committed to solving real challenges that genuinely matter for our customers. Rapid Action & Ownership : We encourage autonomy and accountability—own your projects, move quickly, and shape the future of Gala Intelligence. Collaborative Excellence : Cross-team collaboration ensures alignment, sparks innovation, and drives us forward together. Continuous Learning : Fraud evolves rapidly, and so do we. Continuous improvement, experimentation, and learning are core to our success. If you're excited by the opportunity to leverage technology in the fight against fraud, and you're ready to build something impactful from day one, we want to hear from you!

Posted 6 days ago

Apply

0 years

0 Lacs

bhopal, madhya pradesh, india

On-site

Relu Consultancy is seeking a Data Extraction Engineer with expertise in Python (Selenium). In this role, you will design, implement, and maintain robust data scraping solutions that drive our projects forward. This is your chance to contribute to cutting-edge initiatives while enjoying a work schedule tailored to your needs. Job Title: Data Extraction Engineer Location: Bhopal, MP Job Type: Full-Time CTC: 5LPA Responsibility: 1. Work on web scraping or data extraction through Selenium/Scrapy or other frameworks and related libraries. 2. Working knowledge in various DBSs, message queues & web Restful APIs. 3. Design, build, and maintain high-performance, reusable, and reliable Python code. 4. Ensure the best possible performance, quality, and responsiveness of the application. 5. Identify and correct bottlenecks and fix bugs. 6. Help maintain code quality, organization, and documentation. Qualification: 1. Experience with the Python platform, and object-oriented programming. 2. Python libraries - Pandas, NumPy, Matplotlib, Beautiful Soup, Selenium, Tabula. 3. Data Base - MySQL, SQL, Mongo DB. 4. IDE - PyCharm, Spyder, and Jupiter notebook. 5. Communication skills - Python developers need strong verbal communication skills to work with other members of the programming team and participate in a collaborative environment. 6. Analytical ability - because Python developers analyze programs to improve their functionality, these professionals have strong analytical skills and critical thinking abilities. Why Join Us: Opportunity to work with a reputable consultancy firm. Flexible schedule. Competitive hourly rate. Collaborative and supportive team environment. Opportunity for professional growth. If you're a Data Extraction Engineer seeking a Full-time role that offers the chance to work with a dynamic consultancy firm, we'd love to hear from you. Join us in contributing to our organization's success and in making a positive impact on our team and clients.

Posted 1 week ago

Apply

1.0 - 4.0 years

0 Lacs

gurugram, haryana, india

On-site

Be a part of India’s largest and most admired news network! Network18 is India's most diversified Media Company in the fast growing Media market. The Company has a strong Heritage and we possess a strong presence in Magazines, Television and Internet domains. Our brands like CNBC, Forbes and Moneycontrol are market leaders in their respective segments. The Company has over 7,000 employees across all major cities in India and has been consistently managed to stay ahead of the growth curve of the industry. Network 18 brings together employees from varied backgrounds under one roof united by the hunger to create immersive content and ideas. We take pride in our people, who we believe are the key to realizing the organization’s potential. We continually strive to enable our employees to realize their own goals, by providing opportunities to learn, share and grow. Role Overview: We are seeking a passionate and skilled Data Scientist with over a year of experience to join our dynamic team. You will be instrumental in developing and deploying machine learning models, building robust data pipelines, and translating complex data into actionable insights. This role offers the opportunity to work on cutting-edge projects involving NLP, Generative AI, data automation, and cloud technologies to drive business value. Key Responsibilities: Design, develop, and deploy machine learning models, with a strong focus on NLP (including advanced techniques and Generative AI) and other AI applications. Build, maintain, and optimize ETL pipelines for automated data ingestion, transformation, and standardization from various sources Work extensively with SQL for data extraction, manipulation, and analysis in environments like BigQuery. Develop solutions using Python and relevant data science/ML libraries (Pandas, NumPy, Hugging Face Transformers, etc.). Utilize Google Cloud Platform (GCP) services for data storage, processing, and model deployment. Create and maintain interactive dashboards and reporting tools (e.g., Power BI) to present insights to stakeholders. Apply basic Docker concepts for containerization and deployment of applications. Collaborate with cross-functional teams to understand business requirements and deliver data-driven solutions. Stay abreast of the latest advancements in AI/ML and NLP best practices. Required Qualifications & Skills: 1 to 4 years of hands-on experience as a Data Scientist or in a similar role. Solid understanding of machine learning fundamentals, algorithms, and best practices. Proficiency in Python and relevant data science libraries. Good SQL skills for complex querying and data manipulation. Demonstrable experience with Natural Language Processing (NLP) techniques, including advanced models (e.g., transformers) and familiarity with Generative AI concepts and applications. Excellent problem-solving and analytical skills. Strong communication and collaboration skills. Preferred Qualifications & Skills: Familiarity and hands-on experience with Google Cloud Platform (GCP) services, especially BigQuery, Cloud Functions, and Vertex AI. Basic understanding of Docker and containerization for deploying applications. Experience with dashboarding tools like Power BI and building web applications with Streamlit. Experience with web scraping tools and techniques (e.g., BeautifulSoup, Scrapy, Selenium). Knowledge of data warehousing concepts and schema design. Experience in designing and building ETL pipelines. Disclaimer: Please note Network18 and related group companies do not use the services of vendors or agents for recruitment. Please beware of such agents or vendors providing assistance. Network18 will not be responsible for any losses incurred. “We correspond only from our official email address”

Posted 1 week ago

Apply

2.0 - 3.0 years

0 Lacs

dwarka, delhi, india

On-site

Position: Data Mining Analyst Please Note: Only Candidates with 2-3 years of experience should apply. This is a Delhi-based position and work from office only! Work Location: Sector 23 Dwarka, Delhi We are seeking a skilled Data Mining Analyst with expertise in automating data extraction processes from web platforms. The ideal candidate will be experienced in Python, Selenium, Pandas, SQL, and APIs, with the ability to design and implement efficient and scalable data scraping systems. If you have a passion for working with data and a solid understanding of web technologies, we want to hear from you! Key Responsibilities: Design, develop, and maintain robust web scraping solutions to extract structured and unstructured data from various websites and APIs. Use tools like Python, Selenium, BeautifulSoup, Scrapy, and Pandas for data scraping and processing. Build and manage automated scripts to scrape dynamic websites, including handling JavaScript-driven content. Optimize scraping workflows to ensure data extraction is efficient, accurate, and scalable. Work with APIs to gather and integrate data, ensuring proper rate limits and authentication handling. Clean, preprocess, and store extracted data in databases (SQL) or cloud-based systems. Collaborate with data analysts and other stakeholders to provide required data for further analysis and reporting. Debug and troubleshoot issues in scraping pipelines and scripts. Ensure compliance with ethical data scraping standards, including legal considerations like website terms of use and robots.txt policies. Required Skills & Qualifications: Experience : 2-3 years of hands-on experience in web scraping and data extraction. Technical Skills : Strong proficiency in Python. Experience with web scraping frameworks and libraries like Selenium, Scrapy, BeautifulSoup, and Requests. Experience with data manipulation libraries like Pandas. Familiarity with API integration (REST, GraphQL, etc.). Proficiency in SQL for data querying, database design, and managing large datasets. Knowledge of JavaScript and front-end technologies to work with dynamic web pages. Experience with version control (Git) and collaborative development environments. Other Skills : Problem-solving skills with attention to detail. Ability to write clean, maintainable code and automate workflows. Good understanding of HTTP, HTML, CSS, and JavaScript. Familiarity with cloud services (AWS, Azure, GCP) is a plus. Python Nice to Have: Experience with cloud-based scraping tools or services (e.g., AWS Lambda, Google Cloud Functions). Familiarity with distributed scraping and data pipeline management. Experience with large-scale data collection and storage systems. Knowledge of ethical and legal issues related to web scraping. About Nuvoretail (www.nuvoretail.com) Nuvoretail Enlytical Technologies Private Limited is an e-commerce analytics and automation company. Our proprietary digital shelf analytics and automation platform called Enlytical.ai helps e-commerce brands solve the complexities in today’s e-commerce landscape by offering a unified and all- encompassing business view on the various aspects of e-commerce business. Our platform leverages insights drawn from multiple data points that help our clients win in e-commerce by gaining a competitive edge with data-driven insights for sharper decision-making. The insights cover all aspects of e-commerce such as digital product portfolio analysis, supply chain analytics, e-commerce operations automation, pricing, and competitor benchmarking, and Amazon advertising automation using our proprietary algorithms. As a leading e-commerce service provider, we offer the most comprehensive end-to-end e-commerce solutions to brands, both in India and abroad. Right from preparing a road map to writing our client’s e- commerce success story to assisting them In increasing their online sales, we do everything via our diverse e-commerce services and bespoke strategies and technology. Our services span across the brand’s e-commerce enablement including content and digital asset creation for product listing, On Platform, and Off Platform marketing services with deep expertise in Amazon Marketing Services (AMS), Amazon SEO through keyword research, e-Commerce operations across various e-commerce platforms, website development, social media marketing, and AI-enabled e-Commerce MIS Dashboards. Awards & Recognition: Thanks to the faith reposed on us by our clients, NuvoRetail has been featured as "The Most Promising Ecommerce Technology Service Providers in India 2020” by CIOReviewIndia Magazine. Our leadership is often acknowledged by leading e-commerce services, digital marketing, consulting, and other e- commerce programs around the world. We are now one of the very few companies in India that have become an Amazon Ads Advanced partner.

Posted 1 week ago

Apply

2.0 - 6.0 years

0 Lacs

ahmedabad, gujarat

On-site

As a Senior Python Developer specializing in web scraping and automation at Actowiz Solutions in Ahmedabad, you will be a key member of our dynamic team. Actowiz Solutions is a prominent provider of data extraction, web scraping, and automation solutions, enabling businesses to leverage clean, structured, and scalable data for informed decision-making. By utilizing cutting-edge technology, we strive to deliver actionable insights that drive the future of data intelligence. Your primary responsibility will be to design, develop, and optimize large-scale web scraping solutions using Scrapy, a mandatory requirement for this role. You will work with a variety of additional libraries and tools such as BeautifulSoup, Selenium, Playwright, and Requests to enhance the efficiency and effectiveness of our scraping frameworks. Implementing robust error handling, data parsing, and storage mechanisms (JSON, CSV, SQL/NoSQL databases) will be crucial in ensuring the reliability and scalability of our solutions. Collaboration with product managers, QA, and DevOps teams is essential to ensure timely project delivery. You will also be expected to research and adopt new scraping technologies that can further improve performance, scalability, and efficiency of our data extraction processes. To excel in this role, you should have at least 2 years of experience in Python development with a strong expertise in Scrapy. Proficiency in automation libraries such as Playwright or Selenium, experience with REST APIs, asynchronous programming, and concurrency are also required. Familiarity with databases (SQL/NoSQL) and cloud-based data pipelines will be advantageous, along with strong problem-solving skills and the ability to deliver within Agile methodologies. Preferred qualifications include knowledge of DevOps tools like Docker, GitHub Actions, or CI/CD pipelines. In return, Actowiz Solutions offers a competitive salary, a 5-day work week (Monday to Friday), a flexible and collaborative work environment, and ample opportunities for career growth and skill development. Join us in shaping the future of data intelligence and drive impactful decision-making with our innovative solutions.,

Posted 1 week ago

Apply

3.0 - 7.0 years

0 Lacs

haryana

On-site

You will be responsible for designing, developing, and maintaining web scraping scripts using Python. Your expertise will be essential in utilizing web scraping libraries like Beautiful Soup, Scrapy, Selenium, and other tools to extract data from various websites. It will be crucial to write reusable, testable, and efficient code to extract both structured and unstructured data effectively. Additionally, you will play a key role in developing and maintaining software documentation for the web scraping scripts you create. Collaboration with software developers, data scientists, and other stakeholders will be necessary to strategize, design, develop, and launch new web scraping projects. Troubleshooting, debugging, and optimizing web scraping scripts will also be part of your responsibilities. Staying informed about the latest industry trends and technologies in automated data collection and cleaning will be expected. Your involvement in maintaining code quality, organizing projects, participating in code reviews, and ensuring solutions align with standards will be crucial for success in this role. Creating automated test cases to validate the functionality and performance of the code will also be part of your duties. Integration of data storage solutions such as SQL/NoSQL databases, message brokers, and data streams for storing and analyzing scraped data will be another aspect of your work. A bachelor's degree in Computer Science, Software Engineering, or a related field from a reputable institution is required, along with a minimum of 3-4 years of deep coding experience in Python. Experience with Python development and web scraping techniques is essential for this position. Familiarity with web frameworks like Django and Flask, as well as technologies such as SQL, Git, and Linux, is also necessary. Strong analytical and problem-solving skills, along with effective communication and teamwork abilities, will be key attributes for success in this role. This position is located in Gurgaon, and the work model is Hybrid. If you possess the required skills and experience, we encourage you to apply for this job or refer someone who would be a good fit for this role. Stay updated with the latest industry trends and technologies to excel in this dynamic work environment.,

Posted 1 week ago

Apply

2.0 - 6.0 years

0 Lacs

karnataka

On-site

About DataWeave: Digital Commerce capabilities play a crucial role in attracting today's shoppers. The global eCommerce landscape has seen rapid growth, particularly due to the pandemic, offering consumers various options to research, compare, and purchase products. In this scenario, actionable data-driven insights have become essential for eCommerce expansion and are pivotal in optimizing investments. DataWeave emerges as a leading provider of advanced sales optimization solutions for Digital Commerce businesses, consumer brands, delivery intermediaries, and marketplaces. Through our machine learning-enabled SaaS platform, we cater to over 400 global brands and retailers, offering streamlined access to Digital Shelf Analytics and Dynamic Pricing solutions. These tools aid in formulating data-driven growth strategies and facilitate high-frequency business decisions that drive revenue growth and enhance profitability. Our actionable Digital Market Intelligence is derived from the aggregation and analysis of data collected daily on more than 100 million products available on the Web. This includes tracking competitive KPI performance in crucial business areas such as Pricing, Promotions, Content Audits, Product Availability, Merchandising, Search and Media, Sales and Market Share, as well as Ratings and Reviews. Platforms at DataWeave: The Data Platform team at DataWeave is dedicated to constructing and maintaining the core SaaS infrastructure responsible for data aggregation, processing, and distribution. DevOps forms an integral part of the platform team. Given the substantial volume of data that needs to be processed, stored, and secured across various computing requirements like Machine Learning, Report Generation, and Image processing, there is a necessity for a massive, optimized infrastructure. This is achieved by adopting a hybrid cloud approach combined with container-based, highly manageable, and scalable deployments. How we work: At DataWeave, we are passionate about tackling challenging data problems on a daily basis. We thrive on finding solutions to complex issues and excel in making sense of vast amounts of public data available on the web, operating at a serious scale. Roles & Responsibilities: We are currently looking for an experienced Node.js Engineer to join our team. The ideal candidate should possess a strong understanding of web automation and web scraping techniques and apply this knowledge to develop and maintain our web scraping and automation projects. The key responsibilities include designing and developing web scraping and automation projects using Node.js and Python, optimizing processes for performance and scalability, troubleshooting and debugging issues, collaborating with team members to enhance the architecture and performance of projects, and staying updated on new technologies and advancements in web scraping and automation. Skills & Requirements: The ideal candidate should have at least 2+ years of experience working with web scraping and automation projects, proficiency in Node.js, Python, JavaScript, and web scraping frameworks, familiarity with web automation tools, expertise in Git and other version control systems, and experience with AWS or other cloud platforms would be advantageous. Advantages: Candidates with a robust background in Node.js, Python, and web scraping libraries like Puppeteer, Cheerio, Scrapy, and BeautifulSoup, experience with web automation tools such as Selenium, TestCafe, and Cypress, knowledge of browser devtools and network protocols, understanding of web scraping best practices and techniques, strong problem-solving and debugging skills, experience with Git and other version control systems, familiarity with AWS or other cloud platforms, and excellent communication and team collaboration skills will be at an advantage. Job Location: Bangalore, India Address: InfoWeave Analytics Pvt Ltd, 75/B Windsor, 3rd Floor, Bannerghatta Rd, opp. Christ University, Hulimavu, Bengaluru Karnataka 560076 +(91) 80 618 52 222,

Posted 1 week ago

Apply

1.0 - 5.0 years

0 Lacs

hyderabad, telangana

On-site

You are a Data Analyst with 1-2 years of experience, skilled in data visualization, data analysis, and web scraping. Your primary role involves collecting, processing, and analyzing large datasets from various sources to provide meaningful insights for business decisions. Your responsibilities include end-to-end data analysis and reporting, creating interactive dashboards using tools like Power BI and Tableau, and conducting data scraping using Python tools such as BeautifulSoup and Selenium. You will also be responsible for data cleaning, transformation, and validation using Excel, SQL, or Python libraries like pandas and numpy. Collaboration with cross-functional teams to understand data requirements, ensuring data quality and security in scraping workflows, and providing ad hoc data reports are key aspects of your role. Proficiency in tools like MySQL, PostgreSQL, and Google Sheets is beneficial, along with a strong analytical mindset and experience in Agile environments. Qualifications include a Bachelor's degree in Computer Science or related field, 1-2 years of data analytics experience, and knowledge of data integrity and compliance. Preferred attributes include the ability to handle large datasets efficiently, familiarity with cloud data platforms like AWS, and experience in a day shift work environment. As a Full-time Data Analyst, you will enjoy benefits such as Provident Fund and work in person at the Hyderabad location.,

Posted 1 week ago

Apply

1.0 - 3.0 years

3 - 8 Lacs

bengaluru

Hybrid

About the Role: Grade Level (for internal use): 08 Job Title: Associate Data Engineer The Team: The Automotive Insights - Supply Chain and Technology and IMR department at S&P Global is dedicated to delivering critical intelligence and comprehensive analysis of the automotive industry's supply chain and technology. Our team provides actionable insights and data-driven solutions that empower clients to navigate the complexities of the automotive ecosystem, from manufacturing and logistics to technological innovations and market dynamics. We collaborate closely with industry stakeholders to ensure our research supports strategic decision-making and drives growth within the automotive sector. Join us to be at the forefront of transforming the automotive landscape with cutting-edge insights and expertise. Responsibilities and Impact: Develop and maintain automated data pipelines to extract, transform, and load data from diverse online sources, ensuring high data quality. Build, optimize, and document web scraping tools using Python and related libraries to support ongoing research and analytics. Implement DevOps practices for deploying, monitoring, and maintaining machine learning workflows in production environments. Collaborate with data scientists and analysts to deliver reliable, well-structured data for analytics and modeling. Perform data quality checks, troubleshoot pipeline issues, and ensure alignment with internal taxonomies and standards. Stay current with advancements in data engineering, DevOps, and web scraping technologies, contributing to team knowledge and best practices. What Were Looking For: Basic Required Qualifications: Bachelors degree in computer science, Engineering, or a related field. 1 to 3 years of hands-on experience in data engineering, including web scraping and ETL pipeline development using Python. Proficiency with Python programming and libraries such as Pandas, BeautifulSoup, Selenium, or Scrapy. Exposure to implementing and maintaining DevOps workflows, including model deployment and monitoring. Familiarity with containerization technologies (e.g., Docker) and CI/CD pipelines for data and ML workflows. Familiarity with the cloud platforms (preferably AWS). Key Soft Skills: Strong analytical and problem-solving skills, with attention to detail. Excellent communication and collaboration abilities for effective teamwork. Ability to work independently and manage multiple priorities. Curiosity and a proactive approach to learning and applying new technologies.

Posted 1 week ago

Apply

2.0 years

4 - 4 Lacs

mohali

On-site

We are looking for an enthusiastic and proactive Python Developer with core Python expertise and hands-on experience in Generative AI (GenAI) to join our development team. Experience Required: 2-3 Years Mode of Work: On-Site Only (Mohali, Punjab) Mode of Interview : Face to Face( On-Site) Contact for Queries: +91-9872993778 (Mon–Fri, 11 AM – 6 PM) Note: This number will be unavailable on weekends and public holidays. Key Responsibilities: Backend Development: Assist in the development of clean, efficient, and scalable Python applications to meet business needs. Generative AI Experience: Working knowledge and experience in building applications using GenAI technologies is mandatory. API Integration: Support the creation, management, and optimization of RESTful APIs to connect backend and frontend components. Collaboration: Work closely with frontend developers to integrate backend services into ReactJS applications, ensuring smooth data flow and functionality. Testing and Debugging: Help with debugging, troubleshooting, and optimizing applications for performance and reliability. Code Quality: Write readable, maintainable, and well-documented code while following best practices. Learning and Development: Continuously enhance your skills by learning new technologies and methodologies. Required Skills and Experience Problem Solving: Strong analytical skills with an ability to identify and resolve issues effectively. Previous working experience on LLM and AI Agents is a Plus. Teamwork: Ability to communicate clearly and collaborate well with cross-functional teams. Programming Languages: Python (Core and Advanced) , JavaScript , HTML, CSS Frameworks: Django , Flask , FastAPI , LangChain Libraries & Tools: Pandas, NumPy , Selenium, Scrapy, BeautifulSoup , Git, Postman, OpenAI API, REST APIs Databases : MySQL , PostgreSQL , SQLite Cloud & Deployment: Hands-on experience with AWS services (EC2, S3, etc.) , Building and managing cloud-based scalable applications Automation: Familiarity with Retrieval-Augmented Generation (RAG) architecture. Automation of workflows and intelligent systems using Python. Preferred Qualifications: Education: A degree in Computer Science, Software Engineering, or a related field (or equivalent practical experience). Job Types: Full-time, Permanent Pay: ₹35,000.00 - ₹40,000.00 per month Experience: Python: 2 years (Preferred) Work Location: In person

Posted 1 week ago

Apply

1.5 years

1 - 4 Lacs

ahmedabad

On-site

Senior Python Developer — Web Scraping (1.5 years industry experience) Location: Ahemdabad ( Work from Office ) Employment type: Full-time Company: Actowiz Solutions Role overview We are looking for a pragmatic, hands-on Senior Python Developer with ~1.5 years of direct experience in the web-scraping industry . You’ll design, build and maintain high-quality, reliable data extraction systems (scrapers, crawlers, parsers) and the API layers that serve that data to downstream services. The ideal candidate is comfortable working across the full stack of a scraping product — headless browsers, HTTP clients, extraction pipelines, storage, and REST APIs — and can write production-ready code, tests, and deployment pipelines. Key responsibilities Design, develop, and maintain scalable scrapers and crawlers (headless browser & HTTP-based) to extract structured data from web and app sources. Build and maintain backend APIs (Django REST / FastAPI) that serve processed data to clients and internal services. Implement robust request flows using requests, curl_cffi (or equivalent), and HTTP/2 where appropriate. Develop and operate browser-based scraping using Selenium and Playwright (including stealth variants / evasion best-practices at a high level). Parse and normalize diverse data formats (HTML, JSON, XML), using CSS selectors, XPath, regex, and DOM traversal. Implement retry, back-off, rate limiting, queueing and monitoring for scraping pipelines. Work with data stores (MongoDB, PostgreSQL, Redis, S3 or equivalent) and write efficient ingestion/ETL jobs. Write unit/integration tests, CI/CD pipelines, and containerize services (Docker). Perform code reviews, mentor junior developers, and collaborate with product, QA, and ops teams. Monitor, debug and resolve production issues, and continuously improve reliability and performance. Ensure respectful, legal and ethical data collection practices (observe robots policies, rate limits, data usage constraints as applicable). Required (must-have) technical skills Python — Strong, idiomatic Python (3.8+). Scraping tools & HTTP clients: requests (advanced), curl_cffi (or similar low-level cURL bindings). Browser automation: Selenium (WebDriver), Playwright (including experience with stealth/anti-detection approaches). APIs: Design & implement REST APIs using Django (Django REST Framework) and FastAPI . Data parsing: HTML parsing (BeautifulSoup, lxml), XPath/CSS selectors, JSON/XML handling. Asynchronous programming: asyncio, aiohttp or equivalent for high-throughput crawlers. Containers & devops basics: Docker, Git, basic CI/CD (GitHub Actions / GitLab CI / Jenkins). Datastores: MongoDB and/or PostgreSQL; familiarity with Redis for queues/caching. Debugging & monitoring: Logging best practices, error tracking, metrics, alerting. Linux & networking fundamentals: Comfort working on Linux servers, HTTP internals, headers, proxies. Code quality: Tests (unit/integration), code reviews, clear documentation. Nice-to-have Experience with curl_cffi in production scraping stacks (performance/streaming use-cases). Familiarity with anti-bot/anti-detection tradeoffs at a high level (without providing instructions to break laws) — e.g., fingerprinting awareness, browser fingerprint mitigation strategies, polite rate-limiting. Hands-on experience with cloud platforms (AWS/GCP/Azure) and managed services (S3, ECS/EKS, Lambda). Experience with queueing systems (RabbitMQ, Celery, Kafka) and orchestration. Knowledge of proxies, proxy management, and CAPTCHA solving integrations (high-level architectural experience). Prior product/enterprise experience delivering SLAs for data pipelines. Soft skills & behavioral Strong problem-solving and debugging skills — comfortable diagnosing flaky scrapers and transient production issues. Good communication: explain technical tradeoffs to non-engineers and write clear design docs. Team player: experience mentoring junior engineers and participating in code reviews. Ownership mindset: you ship, operate, and iterate on what you build. Job Types: Full-time, Permanent Pay: ₹15,000.00 - ₹40,000.00 per month Benefits: Flexible schedule Paid time off Application Question(s): Where do you stay? What is your Current Salary ? What is your Expected Salary ? Notice Period ? Are you okay to relocate to Ahemdabad? Work Location: In person Speak with the employer +91 7862049428

Posted 1 week ago

Apply

0.0 - 2.0 years

0 - 0 Lacs

mohali, punjab

On-site

We are looking for an enthusiastic and proactive Python Developer with core Python expertise and hands-on experience in Generative AI (GenAI) to join our development team. Experience Required: 2-3 Years Mode of Work: On-Site Only (Mohali, Punjab) Mode of Interview : Face to Face( On-Site) Contact for Queries: +91-9872993778 (Mon–Fri, 11 AM – 6 PM) Note: This number will be unavailable on weekends and public holidays. Key Responsibilities: Backend Development: Assist in the development of clean, efficient, and scalable Python applications to meet business needs. Generative AI Experience: Working knowledge and experience in building applications using GenAI technologies is mandatory. API Integration: Support the creation, management, and optimization of RESTful APIs to connect backend and frontend components. Collaboration: Work closely with frontend developers to integrate backend services into ReactJS applications, ensuring smooth data flow and functionality. Testing and Debugging: Help with debugging, troubleshooting, and optimizing applications for performance and reliability. Code Quality: Write readable, maintainable, and well-documented code while following best practices. Learning and Development: Continuously enhance your skills by learning new technologies and methodologies. Required Skills and Experience Problem Solving: Strong analytical skills with an ability to identify and resolve issues effectively. Previous working experience on LLM and AI Agents is a Plus. Teamwork: Ability to communicate clearly and collaborate well with cross-functional teams. Programming Languages: Python (Core and Advanced) , JavaScript , HTML, CSS Frameworks: Django , Flask , FastAPI , LangChain Libraries & Tools: Pandas, NumPy , Selenium, Scrapy, BeautifulSoup , Git, Postman, OpenAI API, REST APIs Databases : MySQL , PostgreSQL , SQLite Cloud & Deployment: Hands-on experience with AWS services (EC2, S3, etc.) , Building and managing cloud-based scalable applications Automation: Familiarity with Retrieval-Augmented Generation (RAG) architecture. Automation of workflows and intelligent systems using Python. Preferred Qualifications: Education: A degree in Computer Science, Software Engineering, or a related field (or equivalent practical experience). Job Types: Full-time, Permanent Pay: ₹35,000.00 - ₹40,000.00 per month Experience: Python: 2 years (Preferred) Work Location: In person

Posted 1 week ago

Apply

0 years

0 Lacs

india

On-site

We are looking for an experienced Python Developer with expertise in web scraping and OCR to design and optimize data extraction solutions. The role involves building scalable scripts, integrating OCR workflows, and ensuring clean, structured data for downstream use. Key Responsibilities: Develop and maintain Python scripts for web scraping from structured and unstructured sources. Implement OCR solutions to extract text/data from scanned images, PDFs, and other documents. Optimize data extraction workflows for performance and accuracy. Collaborate with teams to integrate scraped/OCR data into applications or databases. Key Skills: Strong proficiency in Python Hands-on experience with web scraping frameworks (BeautifulSoup, Scrapy, Selenium, etc.) Knowledge of OCR tools/libraries (Tesseract, OpenCV, PyMuPDF, etc.) Good understanding of APIs, data cleaning, and storage mechanisms Familiarity with databases (SQL/NoSQL)

Posted 1 week ago

Apply

1.0 - 3.0 years

0 Lacs

dwarka, delhi, india

On-site

Position: Data Mining Analyst This is a Delhi-based position and work from office only! Work Location: Sector 23 Dwarka, Delhi We are seeking a skilled Data Mining Analyst with expertise in automating data extraction processes from web platforms. The ideal candidate will be experienced in Python, Selenium, Pandas, SQL, and APIs, with the ability to design and implement efficient and scalable data scraping systems. If you have a passion for working with data and a solid understanding of web technologies, we want to hear from you! Key Responsibilities: Design, develop, and maintain robust web scraping solutions to extract structured and unstructured data from various websites and APIs. Use tools like Python, Selenium, BeautifulSoup, Scrapy, and Pandas for data scraping and processing. Build and manage automated scripts to scrape dynamic websites, including handling JavaScript-driven content. Optimize scraping workflows to ensure data extraction is efficient, accurate, and scalable. Work with APIs to gather and integrate data, ensuring proper rate limits and authentication handling. Clean, preprocess, and store extracted data in databases (SQL) or cloud-based systems. Collaborate with data analysts and other stakeholders to provide required data for further analysis and reporting. Debug and troubleshoot issues in scraping pipelines and scripts. Ensure compliance with ethical data scraping standards, including legal considerations like website terms of use and robots.txt policies. Required Skills & Qualifications: Experience : 1-3 years of hands-on experience in web scraping and data extraction. Technical Skills : Strong proficiency in Python. Experience with web scraping frameworks and libraries like Selenium, Scrapy, BeautifulSoup, and Requests. Experience with data manipulation libraries like Pandas. Familiarity with API integration (REST, GraphQL, etc.). Proficiency in SQL for data querying, database design, and managing large datasets. Knowledge of JavaScript and front-end technologies to work with dynamic web pages. Experience with version control (Git) and collaborative development environments. Other Skills : Problem-solving skills with attention to detail. Ability to write clean, maintainable code and automate workflows. Good understanding of HTTP, HTML, CSS, and JavaScript. Familiarity with cloud services (AWS, Azure, GCP) is a plus. PythonNice to Have: Experience with cloud-based scraping tools or services (e.g., AWS Lambda, Google Cloud Functions). Familiarity with distributed scraping and data pipeline management. Experience with large-scale data collection and storage systems. Knowledge of ethical and legal issues related to web scraping. About Nuvoretail (www.nuvoretail.com) Nuvoretail Enlytical Technologies Private Limited is an e-commerce analytics and automation company. Our proprietary digital shelf analytics and automation platform called Enlytical.ai helps e-commerce brands solve the complexities in today’s e-commerce landscape by offering a unified and all- encompassing business view on the various aspects of e-commerce business. Our platform leverages insights drawn from multiple data points that help our clients win in e-commerce by gaining a competitive edge with data-driven insights for sharper decision-making. The insights cover all aspects of e-commerce such as digital product portfolio analysis, supply chain analytics, e-commerce operations automation, pricing, and competitor benchmarking, and Amazon advertising automation using our proprietary algorithms. As a leading e-commerce service provider, we offer the most comprehensive end-to-end e-commerce solutions to brands, both in India and abroad. Right from preparing a road map to writing our client’s e- commerce success story to assisting them In increasing their online sales, we do everything via our diverse e-commerce services and bespoke strategies and technology. Our services span across the brand’s e-commerce enablement including content and digital asset creation for product listing, On Platform, and Off Platform marketing services with deep expertise in Amazon Marketing Services (AMS), Amazon SEO through keyword research, e-Commerce operations across various e-commerce platforms, website development, social media marketing, and AI-enabled e-Commerce MIS Dashboards. Awards & Recognition: Thanks to the faith reposed on us by our clients, NuvoRetail has been featured as "The Most Promising Ecommerce Technology Service Providers in India 2020” by CIOReviewIndia Magazine. Our leadership is often acknowledged by leading e-commerce services, digital marketing, consulting, and other e- commerce programs around the world. We are now one of the very few companies in India that have become an Amazon Ads Advanced partner.

Posted 1 week ago

Apply

2.0 years

0 Lacs

india

Remote

Position: Data Analyst Experience: 2+ Years Location: Remote Joining: Immediate Joiners Preferred About the Role: We are looking for a skilled Data Analyst with strong Python expertise and hands-on experience in handling large datasets, data cleaning, analysis, and visualization. The ideal candidate should be capable of building data pipelines, performing web scraping, and generating actionable insights through dashboards and reports. Key Responsibilities: Develop, test, and maintain efficient Python scripts for data collection and transformation. Design and implement web scraping solutions to extract structured and unstructured data from diverse sources. Perform data cleaning, preprocessing, and analysis to identify patterns, trends, and insights. Work with SQL and NoSQL databases for efficient data storage, retrieval, and manipulation. Create and maintain interactive dashboards and visualizations to present metrics and business insights to stakeholders. Collaborate with cross-functional teams to understand business needs and translate them into data-driven solutions. Optimize existing data workflows and pipelines for performance and scalability. Document code, workflows, and methodologies for clarity and maintainability. Required Skills & Qualifications: Bachelor’s degree in Computer Science, Data Science, or a related field, or equivalent practical experience. Minimum 2 years of professional experience as a Data Analyst or in a similar role. Strong proficiency in Python and key data libraries (Pandas, NumPy, Scikit-learn). Hands-on experience with web scraping tools and libraries (Scrapy, BeautifulSoup, Selenium). Solid understanding of databases with strong SQL skills ; exposure to databases like PostgreSQL, MongoDB is a plus. Experience with data visualization libraries (Matplotlib, Seaborn, Plotly) and/or BI tools (Power BI, Tableau, Looker). Strong problem-solving ability, analytical mindset, and keen attention to detail. Experience using version control systems (Git) . Preferred Qualifications: Exposure to cloud platforms such as AWS, Google Cloud Platform (GCP), or Azure. Familiarity with REST APIs for data collection and integration. Understanding of data warehousing concepts . Prior experience working in an Agile development environment .

Posted 1 week ago

Apply

5.0 years

0 Lacs

maharashtra, india

Remote

Location: Remote (Maharashtra) Experience: 2–5 Years Salary: ₹6 to ₹10 LPA Type: Full-Time About the Role We are seeking a talented and passionate AI/ML Developer to join our team in building a next-generation LLM-based intelligent chatbot system . This role encompasses backend system design, LLM integration, API development, and the creation of a scalable AI-first chatbot architecture. You’ll also contribute to scraping relevant training data, deploying fine-tuned LLMs, and building real-time conversational systems on top of modern cloud AI services. Key Responsibilities Design, develop, and maintain a scalable backend architecture for an LLM-based chatbot. Build, document, and optimize RESTful APIs and microservices to interface with the frontend, databases, and AI models. Integrate open-source LLMs (e.g., LLaMA, Mistral, Falcon, GPT-J) using frameworks like HuggingFace , Ollama , or LangChain . Implement and manage the MCP (Modular Chatbot Platform) design pattern for scalable component interactions. Perform web scraping and data automation using reliable tools to collect and structure high-quality data. Handle LLM fine-tuning , embeddings generation, and vector database integration (FAISS, ChromaDB, Pinecone). Collaborate with frontend, product, and data teams to improve chatbot accuracy, performance, and user experience. Ensure security , performance , and robustness of backend and API infrastructure. Write clean, tested, and documented code; participate in regular code reviews and design discussions. Required Skills Strong programming skills in Python with a solid foundation in Data Structures and Algorithms Hands-on with NumPy, Pandas, Scikit-learn, TensorFlow, Keras Proven backend experience using FastAPI , Flask , or Django Experience with Transformers and LLM frameworks (HuggingFace, LangChain, LlamaIndex) Solid grasp of RESTful API development , microservices, and async programming Knowledge of web scraping tools like BeautifulSoup, Scrapy, Selenium, or Playwright Familiar with LLM training , prompt engineering , and RAG pipelines Understanding of vector databases (FAISS, ChromaDB, Weaviate, Pinecone) Familiarity with MCP (Modular Chatbot Platform) design or similar scalable architectures Experience with Git , CI/CD, and Linux terminal usage Good understanding of authentication , rate limiting , and backend security practices Nice to Have Experience with containerization (Docker) and deploying ML models to the cloud (GCP, AWS, Azure) Exposure to GCP Vertex AI , AWS SageMaker , or Azure AI Services Familiar with MLOps tools like MLflow, Weights & Biases, or DVC Experience working with OpenAI , Mistral , Anthropic , or Cohere APIs Prior experience in chatbot development , NLP applications Ability to collaborate with frontend frameworks (React, Next.js, Streamlit) Education Bachelor’s or Master’s degree in Computer Science, Artificial Intelligence, Data Science , or a related field Equivalent hands-on experience in industry projects and open-source contributions is also considered What We Offer Competitive salary: ₹6 – ₹10 LPA (depending on experience and expertise) End-to-end ownership and leadership on greenfield AI projects Access to GPUs , paid API credits , and cutting-edge LLM tools Fully remote-first culture with flexible work timings An opportunity to work in a fast-paced , product-driven environment with real-world impact How to Apply Send your resume , GitHub/portfolio , and a brief note on your AI/ML experience to: hr.india@orisoninfosystem.com Subject: Application – AI/ML Developer

Posted 1 week ago

Apply

10.0 years

0 Lacs

bengaluru, karnataka, india

On-site

Experience range - 10+ years Type of job - Full time, In office Location - Bengaluru Roles and Responsibilities: Convert broad vision and concepts into a structured data science roadmap, and guide a team to successfully execute on it. Handling end-to-end client AI & analytics programs in a fluid environment. Your role will be a combination of hands-on contribution, technical team management, and client interaction. Proven ability to discover solutions hidden in large datasets and to drive business results with their data-based insights Contribute to internal product development initiatives related to data science. Drive excellent project management required to deliver complex projects, including effort/time estimation. Be proactive, with full ownership of the engagement. Build scalable client engagement level processes for faster turnaround & higher accuracy Define Technology/ Strategy and Roadmap for client accounts, and guides implementation of that strategy within projects Manage the team-members, to ensure that the project plan is being adhered to over the course of the project Build a trusted advisor relationship with the IT management at clients and internal accounts leadership. Mandated Skills A B-Tech/M-Tech/MBA from a top tier Institute preferably in a quantitative subject 15+ overall experience with 12+ years of hands-on experience in applied Machine Learning, AI and analytics Experience of scientific programming in scripting languages like Python, R, SQL, NoSQL, Spark with ML tools & Cloud Technology (AWS, Azure, GCP) Experience in Python libraries such as numpy, pandas, scikit-learn, tensor-flow, scrapy, BERT etc. Strong grasp of depth and breadth of machine learning, deep learning, data mining, and statistical concepts and experience in developing models and solutions in these areas Expertise with client engagement, understanding complexproblem statements, and offering solutions in the domains of Supply Chain, Manufacturing, CPG, Marketing etc. Desired Skills Deep understanding of ML algorithms for common use cases in both structured and unstructured data ecosystems. Comfortable with large scale data processing and distributed computing Providing required inputs to sales, and pre-sales activities A self-starter who can work well with minimal guidance Excellent written and verbal communication skills AI,ML

Posted 1 week ago

Apply

0 years

0 Lacs

jamnagar, gujarat, india

On-site

Technical Skills Proven experience working with Python and its core concepts Knowledge of and working experience with the Scrapy framework Understanding of web frameworks like Flask and Django Version Control Knowledge of probability and statistics Familiarity with ORM techniques Database integration Soft Skills Analytical thinking and problem-solving Good communication and orientation Designing and time management skills Ability to adapt and meet deadlines

Posted 1 week ago

Apply

3.0 - 7.0 years

0 Lacs

ahmedabad, gujarat

On-site

You will be part of Xwiz Analytics, a company specializing in providing innovative data-driven solutions focusing on e-commerce, social media, and market analytics. As a Python Developer with expertise in web scraping, you will play a crucial role in developing robust data extraction pipelines. Your responsibilities will include creating and optimizing scalable web scraping solutions, building scraping scripts using Python libraries like Scrapy, BeautifulSoup, Selenium, and Playwright, managing scraping bots with advanced automation capabilities, and handling anti-scraping mechanisms effectively. Moreover, you will be responsible for ensuring data accuracy and quality through data validation and parsing rules, integrating web scraping pipelines with databases and APIs for seamless data storage and retrieval, debugging and troubleshooting scraping scripts, and implementing robust error-handling mechanisms to enhance the reliability of scraping systems. Collaboration with the team to define project requirements, deliverables, and timelines will also be a key aspect of your role. To excel in this position, you must have a strong proficiency in Python, particularly in web scraping libraries, experience in parsing structured and unstructured data formats, hands-on experience with anti-bot techniques, a solid understanding of HTTP protocols, cookies, and headers management, experience with multi-threading or asynchronous programming, familiarity with databases, knowledge of API integration, cloud platforms, and strong debugging skills. Furthermore, it would be advantageous to have experience with Docker, task automation tools, machine learning techniques, and a good understanding of ethical scraping practices and compliance with legal frameworks. Key attributes that will contribute to your success in this role include strong analytical and problem-solving skills, attention to detail, the ability to thrive in a fast-paced, collaborative environment, and a proactive approach to challenges. In return, Xwiz Analytics offers you the opportunity to work on exciting projects with large-scale data challenges, a competitive salary and benefits package, a supportive environment for professional growth, and a flexible work culture that promotes innovation and learning.,

Posted 1 week ago

Apply

4.0 years

6 - 7 Lacs

mohali

On-site

We are seeking a highly skilled and passionate Python Developer to join our team. The ideal candidate should have strong expertise in backend development using Django, DRF, or FastAPI, along with extensive experience in advanced web scraping and hands-on knowledge in AI/LLM model training . Key Responsibilities: Develop and maintain RESTful APIs using Django, DRF, or FastAPI . Perform very advanced web scraping using tools like BeautifulSoup, Scrapy, and Selenium . Example platforms : Amazon, LinkedIn, Twitter, job portals, e-commerce sites, and real estate listings. Example content : Product data, job listings, user reviews, pricing data, dynamic JS-rendered content, etc. Implement and manage third-party API integrations (e.g., Stripe, Twilio, Google APIs). Contribute to AI/LLM model training using custom or open-source datasets (e.g., for classification, summarization, chatbot use cases). Write clean, scalable, and well-documented code with unit testing. Collaborate with cross-functional teams to define and deliver innovative solutions. Required Skills: Strong Python programming experience (4+ years) Frameworks: Django, DRF, FastAPI Web scraping tools: BeautifulSoup, Scrapy, Selenium Experience with scraping JS-heavy websites and handling captchas, proxies, headless browsers Hands-on experience with AI/ML models , fine-tuning LLMs, or working with tools like HuggingFace Transformers, LangChain, or OpenAI APIs Experience in integrating and working with third-party APIs Solid knowledge of REST API development and optimization Familiarity with Docker, Git, PostgreSQL/MySQL, Redis To Apply: Submit your resume at muskan@impingeonline.com Job Types: Full-time, Permanent Pay: ₹50,410.63 - ₹65,382.15 per month Benefits: Paid time off Provident Fund Work Location: In person

Posted 1 week ago

Apply

2.0 years

4 - 4 Lacs

mohali

On-site

Male applicants are preferred We are looking for an enthusiastic and proactive Python Developer with core Python expertise and hands-on experience in Generative AI (GenAI) to join our development team. Experience Required: 2-3 Years Mode of Work: On-Site Only (Mohali, Punjab) Mode of Interview : Face to Face( On-Site) Contact for Queries: +91-9872993778 (Mon–Fri, 11 AM – 6 PM) Note: This number will be unavailable on weekends and public holidays. Key Responsibilities: Backend Development: Assist in the development of clean, efficient, and scalable Python applications to meet business needs. Generative AI Experience: Working knowledge and experience in building applications using GenAI technologies is mandatory. API Integration: Support the creation, management, and optimization of RESTful APIs to connect backend and frontend components. Collaboration: Work closely with frontend developers to integrate backend services into ReactJS applications, ensuring smooth data flow and functionality. Testing and Debugging: Help with debugging, troubleshooting, and optimizing applications for performance and reliability. Code Quality: Write readable, maintainable, and well-documented code while following best practices. Learning and Development: Continuously enhance your skills by learning new technologies and methodologies. Required Skills and Experience Problem Solving: Strong analytical skills with an ability to identify and resolve issues effectively. Previous working experience on LLM and AI Agents is a Plus. Teamwork: Ability to communicate clearly and collaborate well with cross-functional teams. Programming Languages: Python (Core and Advanced) , JavaScript , HTML, CSS Frameworks: Django , Flask , FastAPI , LangChain Libraries & Tools: Pandas, NumPy , Selenium, Scrapy, BeautifulSoup , Git, Postman, OpenAI API, REST APIs Databases : MySQL , PostgreSQL , SQLite Cloud & Deployment: Hands-on experience with AWS services (EC2, S3, etc.) , Building and managing cloud-based scalable applications Automation: Familiarity with Retrieval-Augmented Generation (RAG) architecture. Automation of workflows and intelligent systems using Python. Preferred Qualifications: Education: A degree in Computer Science, Software Engineering, or a related field (or equivalent practical experience). Job Types: Full-time, Permanent Pay: ₹35,000.00 - ₹40,000.00 per month Experience: Python: 2 years (Preferred) Work Location: In person

Posted 1 week ago

Apply

0 years

1 - 1 Lacs

barnāla

On-site

Job Overview We are seeking a detail-oriented Data Entry Specialist with strong skills in online research, data cross-checking, and updating records from multiple online sources. The ideal candidate should be proficient in web scraping , Excel (including advanced formulas/macros) , and able to work quickly and accurately. Good command of spoken and written English is essential for effective communication and task execution. Key Responsibilities Collect, verify, and update data from multiple online platforms. Cross-check and validate information to ensure data accuracy. Use scraping tools or custom scripts to extract data efficiently. Maintain and organize large datasets in Excel, applying formulas, macros, and automation where necessary. Identify and correct errors or inconsistencies in data. Collaborate with team members to streamline processes and reporting. Deliver tasks on time while maintaining high speed and accuracy. Qualifications & Skills Proven experience in data entry, research, or similar roles . Strong knowledge of Excel (advanced formulas, pivot tables, VBA/macros preferred). Familiarity with web scraping techniques (e.g., Python, BeautifulSoup, Scrapy, or similar tools). Excellent attention to detail and organizational skills. Ability to work quickly without compromising quality. Good command of spoken and written English (must be able to read, understand, and communicate clearly). Self-motivated, reliable, and able to work independently. Preferred Experience with Google Sheets and online databases. Basic knowledge of APIs or automation tools (Zapier, Power Query, etc.). Prior experience in a fast-paced data-driven role. * Job Type: Full-time Pay: ₹12,500.00 - ₹16,000.00 per month Benefits: Provident Fund Language: English (Preferred) Work Location: In person

Posted 1 week ago

Apply

1.0 - 3.0 years

0 Lacs

dwarka, delhi, india

On-site

Position: Data Mining Analyst This is a Delhi-based position and work from office only! Work Location: Sector 23 Dwarka, Delhi We are seeking a skilled Data Mining Analyst with expertise in automating data extraction processes from web platforms. The ideal candidate will be experienced in Python, Selenium, Pandas, SQL, and APIs, with the ability to design and implement efficient and scalable data scraping systems. If you have a passion for working with data and a solid understanding of web technologies, we want to hear from you! Key Responsibilities: Design, develop, and maintain robust web scraping solutions to extract structured and unstructured data from various websites and APIs. Use tools like Python, Selenium, BeautifulSoup, Scrapy, and Pandas for data scraping and processing. Build and manage automated scripts to scrape dynamic websites, including handling JavaScript-driven content. Optimize scraping workflows to ensure data extraction is efficient, accurate, and scalable. Work with APIs to gather and integrate data, ensuring proper rate limits and authentication handling. Clean, preprocess, and store extracted data in databases (SQL) or cloud-based systems. Collaborate with data analysts and other stakeholders to provide required data for further analysis and reporting. Debug and troubleshoot issues in scraping pipelines and scripts. Ensure compliance with ethical data scraping standards, including legal considerations like website terms of use and robots.txt policies. Required Skills & Qualifications: Experience : 1-3 years of hands-on experience in web scraping and data extraction. Technical Skills : Strong proficiency in Python. Experience with web scraping frameworks and libraries like Selenium, Scrapy, BeautifulSoup, and Requests. Experience with data manipulation libraries like Pandas. Familiarity with API integration (REST, GraphQL, etc.). Proficiency in SQL for data querying, database design, and managing large datasets. Knowledge of JavaScript and front-end technologies to work with dynamic web pages. Experience with version control (Git) and collaborative development environments. Other Skills : Problem-solving skills with attention to detail. Ability to write clean, maintainable code and automate workflows. Good understanding of HTTP, HTML, CSS, and JavaScript. Familiarity with cloud services (AWS, Azure, GCP) is a plus. PythonNice to Have: Experience with cloud-based scraping tools or services (e.g., AWS Lambda, Google Cloud Functions). Familiarity with distributed scraping and data pipeline management. Experience with large-scale data collection and storage systems. Knowledge of ethical and legal issues related to web scraping. About Nuvoretail (www.nuvoretail.com) Nuvoretail Enlytical Technologies Private Limited is an e-commerce analytics and automation company. Our proprietary digital shelf analytics and automation platform called Enlytical.ai helps e-commerce brands solve the complexities in today’s e-commerce landscape by offering a unified and all- encompassing business view on the various aspects of e-commerce business. Our platform leverages insights drawn from multiple data points that help our clients win in e-commerce by gaining a competitive edge with data-driven insights for sharper decision-making. The insights cover all aspects of e-commerce such as digital product portfolio analysis, supply chain analytics, e-commerce operations automation, pricing, and competitor benchmarking, and Amazon advertising automation using our proprietary algorithms. As a leading e-commerce service provider, we offer the most comprehensive end-to-end e-commerce solutions to brands, both in India and abroad. Right from preparing a road map to writing our client’s e- commerce success story to assisting them In increasing their online sales, we do everything via our diverse e-commerce services and bespoke strategies and technology. Our services span across the brand’s e-commerce enablement including content and digital asset creation for product listing, On Platform, and Off Platform marketing services with deep expertise in Amazon Marketing Services (AMS), Amazon SEO through keyword research, e-Commerce operations across various e-commerce platforms, website development, social media marketing, and AI-enabled e-Commerce MIS Dashboards. Awards & Recognition: Thanks to the faith reposed on us by our clients, NuvoRetail has been featured as "The Most Promising Ecommerce Technology Service Providers in India 2020” by CIOReviewIndia Magazine. Our leadership is often acknowledged by leading e-commerce services, digital marketing, consulting, and other e- commerce programs around the world. We are now one of the very few companies in India that have become an Amazon Ads Advanced partner.

Posted 1 week ago

Apply

2.0 - 6.0 years

0 Lacs

warangal, telangana

On-site

You will be responsible for building, improving, and extending NLP capabilities. Your tasks will include selecting proper annotated datasets for supervised learning techniques, using effective text representation techniques to develop useful features, identifying and utilizing correct algorithms for specific NLP projects, developing NLP projects as per prescribed requirements, training developed NLP models, and evaluating their effectiveness. Additionally, you will conduct statistical analyses of models, adjust models where necessary, and extend machine learning frameworks and libraries for NLP projects. The ideal candidate should have 2-4 years of experience in Analytics and Machine Learning & Deep Learning, including expertise in areas such as Sentiment Analysis, Text Mining, Entity Extraction, Document Classification, Topic Modeling, Natural Language Understanding (NLU), and Natural Language Generation (NLG). You are expected to possess good experience in Web Scraping or data extraction through Selenium/Scrapy or other frameworks and related libraries like BeautifulSoup. Knowledge of advanced scraping techniques such as overcoming captcha, proxy, browser fingerprinting, and bot detection bypassing is preferred. You should have a working knowledge of various databases like MySQL, HBase, MongoDB, message queues, and Web RestFul APIs. Expertise in open-source NLP toolkits such as CoreNLP, OpenNLP, NLTK, Gensim, LingPipe, Keras, TensorFlow, Mallet, ML/math toolkits like scikit-learn, MLlib, Theano, NumPy, etc., is essential. Experience in testing and deployment of machine learning/deep learning projects on the desired platform is required. Strong logical and analytical skills along with good soft and communication skills are desired qualities. Desirable certifications for this role include Google Tensor Flow, Hadoop, Python, PySpark, SQL, and R.,

Posted 1 week ago

Apply

5.0 - 9.0 years

0 Lacs

chennai, tamil nadu

On-site

As a Senior Lead, you will be involved in Python-based automation initiatives encompassing AI/ML model testing and Framework development. Your responsibilities will include designing Frameworks, developing Proof of Concepts, Accelerators, and executing tasks effectively and independently. You will also be designing and implementing automation for Data, Web, and Mobile applications, as well as participating in Demos, Planning, and collaborating with various teams including business consultants, data scientists, engineers, and application developers. Desired Skills and Experience: - Strong knowledge and experience in Python Programming Language. - Basic understanding of AI/ML and Data science models along with related libraries such as scikit-learn, matplotlib, etc. - Ability to autonomously design, develop, and architect highly available, highly scalable accelerators and frameworks from scratch. - Hands-on experience with Pytest, Selenium, Pyspark, Numpy, Pandas. - Proficiency in SQL and database Knowledge. - Solid grasp of CI/CD pipeline tools like Azure/Jenkins. - Good understanding of RESTful interfaces and Microservice concepts. - Extensive Experience in building frameworks using Appium, Selenium, Pytest, Requests, and related Libraries. - Sound knowledge of Web Scraping utilizing Python-based tools like Scrapy, and BeautifulSoup. - Familiarity with Dockers and Kubernetes is advantageous. - Knowledge of cloud platforms AWS/Azure/GCP. - Ability to rapidly grasp and apply complex technical information to testing scenarios. - Attention to detail and adeptness in escalating and managing issues effectively. - Familiarity with Agile methodology. - Capability to manage multiple assignments simultaneously and adhere to delivery timelines. - Knowledge of Rally/JIRA is a plus. - Excellent written and verbal communication skills. Join us in shaping the world's premier AI and advanced analytics team where equal opportunities are valued. Our competitive compensation packages ensure you are rewarded based on your expertise and experience.,

Posted 1 week ago

Apply
cta

Start Your Job Search Today

Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.

Job Application AI Bot

Job Application AI Bot

Apply to 20+ Portals in one click

Download Now

Download the Mobile App

Instantly access job listings, apply easily, and track applications.

Featured Companies