Jobs
Interviews

311 Scrapy Jobs

Setup a job Alert
JobPe aggregates results for easy application access, but you actually apply on the job portal directly.

0 years

0 Lacs

gurugram, haryana, india

On-site

Responsibilities • Develop Quick Process Automations (QPAs) using one of the following platforms - VBA, MS Office pack, Java, SQL, Python. • Create appropriate project documentation and train operations team on how to use the developed tool. • Follow standard team coding structure and maintenance of already developed solutions. • Proactive to identify workarounds, fix bugs, implement changes and come up with out of the box solutions • Manage project delivery throughout entire lifecycle: analyze, plan, design, build, test, deploy, maintain • Self-driven, being able to work independently and take the project to completion with minimum supervision • Self-learner, self-starter, reach out to colleagues when you cannot find a solution • Understanding the requirement of the operations team and building the tool accordingly. Qualifications we seek in you! Minimum qualifications • BE/B- Tech, BCA • Knowledge of one or more of the following computer languages knowledge required: • Python: o Should have demonstrable, practical knowledge in the following data/file processing libraries such as Scrapy, Numpy, BeautifulSoup, Numpy, Pandas, Matplotlib, Bokeh (optional) o Should have knowledge of setting up a Python environment and all the s/w that’s needed • VBA-Use Visual Basic in Microsoft Applications, including Excel, Access, Word and PowerPoint • Advanced Excel skills (i.e. complex formulas) • Able to debug/code functions/macros using VBA • Ability to normalize complex data/define referential integrity in Access • Ability to programmatically manipulate worksheet and cell properties using VBA • Java: (would be plus if have Java knowledge) • Hands on Core java (1.7 or higher version) • Strong knowledge of Java Design pattern and their implementation. • JDBC with (Oracle/ SQL Server / My SQL) • Webservice (REST / SOAP) • Data parsing API (PDF, Excel, CSV, XML ) • Web Application development with MVC design pattern • Hand on Web framework (Struts, Spring, Hibernate etc..) • DOS/UNIX Scripting: Should have a sound knowledge of file processing in basic OS platforms and how to manipulate files/folders, how to connect to systems etc. • Database and SQL: Sound knowledge of relational databases; ability to define and create basic database, fundamentals of loading, ETL • Experience on developing automated solutions on SAP and Oracle systems. • Knowledge of backend services and integrations; • Understanding of the cloud infrastructure environment. • Ability to work in an Agile environment; • Critical thinking and analytical skills to examine bug reports and prioritize necessary tests & estimate task execution; • Must be able to proactively research solutions • Ability to work independently and also within a team. • software development experience, Preferred qualifications • Knowledge of the OM domain; • Machine Learning understanding • Excellent communication/ interpersonal skills • Working knowledge of Java/VBA would be an add on.

Posted 19 hours ago

Apply

14.0 years

0 Lacs

bengaluru, karnataka, india

On-site

Sigmoid enables business transformation using data and analytics, leveraging real-time insights to make accurate and fast business decisions, by building modern data architectures using cloud and open source. Some of the world’s largest data producers engage with Sigmoid to solve complex business problems. Sigmoid brings deep expertise in data engineering, predictive analytics, artificial intelligence, and DataOps. Sigmoid has been recognized as one of the fastest growing technology companies in North America, 2021, by Financial Times, Inc. 5000, and Deloitte Technology Fast 500. Offices: New York | Dallas | San Francisco | Lima | Bengaluru This role is for our Bengaluru office. Why Join Sigmoid? • Sigmoid provides the opportunity to push the boundaries of what is possible by seamlessly combining technical expertise and creativity to tackle intrinsically complex business problems and convert them into straight-forward data solutions. • Despite being continuously challenged, you are not alone. You will be part of a fast-paced diverse environment as a member of a high-performing team that works together to energize and inspire each other by challenging the status quo • Vibrant inclusive culture of mutual respect and fun through both work and play Roles and Responsibilities: • Convert broad vision and concepts into a structured data science roadmap, and guide a team to successfully execute on it. • Handling end-to-end client AI & analytics programs in a fluid environment. Your role will be a combination of hands-on contribution, technical team management, and client interaction. • Proven ability to discover solutions hidden in large datasets and to drive business results with their data-based insights • Contribute to internal product development initiatives related to data science. • Drive excellent project management required to deliver complex projects, including effort/time estimation. • Be proactive, with full ownership of the engagement. Build scalable client engagement level processes for faster turnaround & higher accuracy • Define Technology/ Strategy and Roadmap for client accounts, and guides implementation of that strategy within projects • Manage the team-members, to ensure that the project plan is being adhered to over the course of the project • Build a trusted advisor relationship with the IT management at clients and internal accounts leadership. Mandated Skills: • A B-Tech/M-Tech/MBA from a top tier Institute preferably in a quantitative subject • 14+ years of hands-on experience in applied Machine Learning, AI and analytics • Experience of scientific programming in scripting languages like Python, R, SQL, NoSQL, Spark with ML tools & Cloud Technology (AWS, Azure, GCP) • Experience in Python libraries such as numpy, pandas, scikit-learn, tensor-flow, scrapy, BERT etc. • Strong grasp of depth and breadth of machine learning, deep learning, data mining, and statistical concepts and experience in developing models and solutions in these areas • Expertise with client engagement, understanding complex problem statements, and offering solutions in the domains of Supply Chain, Manufacturing, CPG, Marketing etc. Desired Skills: ● Deep understanding of ML algorithms for common use cases in both structured and unstructured data ecosystems. ● Comfortable with large scale data processing and distributed computing ● Providing required inputs to sales, and pre-sales activities ● A self-starter who can work well with minimal guidance ● Excellent written and verbal communication skills

Posted 23 hours ago

Apply

5.0 years

0 Lacs

gurugram, haryana, india

On-site

We are seeking highly motivated and results-driven Business Development Executives to join our sales team. The successful candidate will be responsible for supporting overall sales operations, strengthening client relationships, and contributing to revenue growth. A core aspect of this role will be lead generation and data acquisition from digital platforms to build and maintain a qualified sales pipeline. Key Responsibilities Lead Generation: Identify, qualify, and generate prospective leads through online tools, platforms, and databases (e.g., LinkedIn, industry events, company websites, and email sources). Client facing capabilities to include ability to pitch new business and ideas to potential leads and clients. Sales Operations Support: Assist in managing the sales pipeline, ensuring systematic follow-up and accurate documentation of prospects and client interactions. CRM Management: Maintain accurate and up-to-date records of sales activities within the Customer Relationship Management (CRM) system. Market Research: Conduct research to monitor market trends, analyze competitor activity, and identify new business opportunities. Pipeline Development: Establish and maintain a consistent pipeline of qualified prospects. Tools & Reporting: Prepare presentations, reports, and sales materials using MS Office (PowerPoint, Excel) and Canva. Qualifications Experience: Minimum of 5 years of experience in Business Development or Sales. Prior experience in an agency or news media environment is mandatory . Technical Skills: Familiarity with lead generation platforms, CRM systems (Salesforce, HubSpot), and web scraping tools (Scrapy, BeautifulSoup, LinkedIn Sales Navigator). Communication: Strong verbal and written communication skills with the ability to engage effectively with clients and internal teams. Organizational Skills: Demonstrated ability to manage multiple priorities, meet deadlines, and perform effectively in a fast-paced environment. Analytical & Problem-Solving: Ability to assess challenges, identify solutions, and support informed decision-making. Education: Bachelor’s degree in Business, Marketing, Communications, or a related field is required.

Posted 1 day ago

Apply

2.0 - 5.0 years

6 - 16 Lacs

hyderabad, chennai, bengaluru

Work from Office

Python programming,data science methodologies AI systems or chatbot development Deep learning, transformer, GAN’s , VAE’s and LLM processing, prompt engineering TensorFlow, PyTorch, spaCy send cv - sairamglobal.hr@gmail.com Required Candidate profile Education 15 years full time regular study required Notice period - 0 to 45 days Location- Bangalore, Hyderabad, Chennai Good Communication skills required

Posted 2 days ago

Apply

3.0 - 7.0 years

4 - 7 Lacs

thane, navi mumbai, mumbai (all areas)

Work from Office

Key Responsibilities: Develop and maintain automated web scraping scripts using Python libraries such as Beautiful Soup, Scrapy, and Selenium. Optimize scraping pipelines for performance, scalability, and resource efficiency. Handle dynamic websites, CAPTCHA-solving, and implement IP rotation techniques for uninterrupted scraping. Process and clean raw data, ensuring accuracy and integrity in extracted datasets. Collaborate with cross-functional teams to understand data requirements and deliver actionable insights. Leverage APIs when web scraping is not feasible, managing authentication and request optimization. Document processes, pipelines, and troubleshooting steps for maintainable and reusable scraping solutions. Ensure compliance with legal and ethical web scraping practices, implementing security safeguards. Requirements: Education : Bachelors degree in Computer Science, Engineering, or a related field. Experience : 2+ years of Python development experience, with at least 1 year focused on web scraping. Technical Skills : Proficiency in Python and libraries like Beautiful Soup, Scrapy, and Selenium. Experience with regular expressions (Regex) for data parsing. Strong knowledge of HTTP protocols, cookies, headers, and user-agent rotation. Familiarity with databases (SQL and NoSQL) for storing scraped data. Hands-on experience with data manipulation libraries such as pandas and NumPy. Experience working with APIs and managing third-party integrations. Familiarity with version control systems like Git. Bonus Skills : Knowledge of containerization tools like Docker. Experience with distributed scraping solutions and task queues (e.g., Celery, RabbitMQ). Basic understanding of data visualization tools. Non-Technical Skills : Strong analytical and problem-solving skills. Excellent communication and documentation skills. Ability to work independently and collaboratively in a team environmen

Posted 3 days ago

Apply

1.0 - 3.0 years

2 - 5 Lacs

mumbai

Work from Office

We are seeking an experienced Web Scraping Engineer with deep expertise in Scrapy to develop, maintain, and optimize web crawlers. The ideal candidate will have a strong background in extracting, processing, and managing large-scale web data efficiently. Responsibilities : - Write and maintain web scraping scripts using Python Optimize custom web scraping tools and workflows. - Familiarity with Python and web scraping frameworks (Scrapy, Selenium, BeautifulSoup, Requests, Playwright). - Troubleshoot and resolve scraping challenges, including CAPTCHAs, rate limiting, and IP blocking. - Experience with proxy management (Rotating IPs, VPNs, Residential proxies, etc.) - Handle dynamic content using headless browsers and solve CAPTCHAs & IP bans. - Collaborate with senior developers to improve code quality and efficiency. Skills Required : - Python and scraping libraries such as Scrapy and BeautifulSoup - Experience with proxy management, CAPTCHA bypass techniques, and anti-bot evasion. - Ability to optimize crawlers for performance and minimize website detection risks. - Strong background in databases and data storage systems . - Prior Experience of scraping large-scale E-Commerce websites would be plus. - Willingness to learn and take feedback constructively. - Excellent communication and leadership abilities.

Posted 3 days ago

Apply

4.0 years

0 Lacs

ahmedabad, gujarat, india

On-site

Position: Lead Developer Experience: 4+ Years Location: Ahmedabad (Onsite) Employment Type: Full-Time Salary: As per industry standards Must Have Skills:- TypeScript, JavaScript, Web Scrapping & Python Position Summary: We are seeking a skilled and experienced Backend Developer with strong expertise in TypeScript, Python, and web scraping. You will be responsible for designing, developing, and maintaining scalable backend services and APIs that power our data-driven products. Your role will involve collaborating with cross-functional teams, optimizing system performance, ensuring data integrity, and contributing to the design of efficient and secure architectures. Job Responsibility: ● Design, develop, and maintain backend systems and services using Python and TypeScript. ● Develop and maintain web scraping solutions to extract, process, and manage large-scale data from multiple sources. ● Work with relational and non-relational databases, ensuring high availability, scalability, and performance. ● Implement authentication, authorization, and security best practices across services. ● Write clean, maintainable, and testable code following best practices and coding standards. ● Collaborate with frontend engineers, data engineers, and DevOps teams to deliver robust solutions and troubleshoot, debug, and upgrade existing applications. ● Stay updated with backend development trends, tools, and frameworks to continuously improve processes. Utilize core crawling experience to design efficient strategies for scraping the data from different websites and applications. ● Collaborate with technology teams, data collection teams to build end to end technology-enabled ecosystems and partner in research projects to analyze the massive data inputs. ● Responsible for the design and development of web crawlers, able to independently solve various problems encountered in the actual development process. ● Stay updated with the latest web scraping techniques, tools, and industry trends to continuously improve the scraping processes. Job Requirements: ● 4+ years of professional experience in backend development with TypeScript and Python. ● Strong understanding of TypeScript-based server-side frameworks (e.g., Node.js, NestJS, Express) and Python frameworks (e.g., FastAPI, Django, Flask). ● Experience with tools and libraries for web scraping (e.g., Scrapy, BeautifulSoup, Selenium, Puppeteer) ● Hands-on experience with Temporal for creating and orchestrating workflows ● Proven hands-on experience in web scraping, including crawling, data extraction, deduplication, and handling dynamic websites. ● Proficient in implementing proxy solutions and handling bot-detection challenges (e.g., Cloudflare). ● Experience working with Docker, containerized deployments, and cloud environments (GCP or Azure). ● Proficiency with database systems such as MongoDB and ElasticSearch. ● Hands-on experience with designing and maintaining scalable APIs. ● Knowledge of software testing practices (unit, integration, end-to-end). ● Familiarity with CI/CD pipelines and version control systems (Git). ● Strong problem-solving skills, attention to detail, and ability to work in agile environments. ● Great communication skills and ability to navigate in undirected situations. Job Exposure: ● Opportunity to apply creative methods in acquiring and filtering the North American government, agencies data from various websites, sources ● In depth industry exposure on data harvesting techniques to build, scale the robust and sustainable model, using open-source applications ● Effectively collaboration with IT team to design the tailor-made solutions basis upon clients’ requirement ● Unique opportunity to research on various agencies, vendors, products as well as technology tools to compose a solution

Posted 3 days ago

Apply

0.0 - 3.0 years

2 - 4 Lacs

bengaluru

Work from Office

We are hiring a Python Developer (0.52 yrs) for web scraping. Responsibilities: build & optimize scrapers, handle dynamic sites, proxies & CAPTCHAs, ensure data accuracy. Skills: Python, Scrapy, BeautifulSoup, Selenium, regex, debugging. Provident fund Health insurance

Posted 3 days ago

Apply

2.0 - 3.0 years

0 Lacs

dwarka, delhi, india

On-site

Position: Data Mining Analyst Please Note: Only Candidates with 2-3 years of experience should apply. This is a Delhi-based position and work from office only! Work Location: Sector 23 Dwarka, Delhi We are seeking a skilled Data Mining Analyst with expertise in automating data extraction processes from web platforms. The ideal candidate will be experienced in Python, Selenium, Pandas, SQL, and APIs, with the ability to design and implement efficient and scalable data scraping systems. If you have a passion for working with data and a solid understanding of web technologies, we want to hear from you! Key Responsibilities: Design, develop, and maintain robust web scraping solutions to extract structured and unstructured data from various websites and APIs. Use tools like Python, Selenium, BeautifulSoup, Scrapy, and Pandas for data scraping and processing. Build and manage automated scripts to scrape dynamic websites, including handling JavaScript-driven content. Optimize scraping workflows to ensure data extraction is efficient, accurate, and scalable. Work with APIs to gather and integrate data, ensuring proper rate limits and authentication handling. Clean, preprocess, and store extracted data in databases (SQL) or cloud-based systems. Collaborate with data analysts and other stakeholders to provide required data for further analysis and reporting. Debug and troubleshoot issues in scraping pipelines and scripts. Ensure compliance with ethical data scraping standards, including legal considerations like website terms of use and robots.txt policies. Required Skills & Qualifications: Experience : 2-3 years of hands-on experience in web scraping and data extraction. Technical Skills : Strong proficiency in Python. Experience with web scraping frameworks and libraries like Selenium, Scrapy, BeautifulSoup, and Requests. Experience with data manipulation libraries like Pandas. Familiarity with API integration (REST, GraphQL, etc.). Proficiency in SQL for data querying, database design, and managing large datasets. Knowledge of JavaScript and front-end technologies to work with dynamic web pages. Experience with version control (Git) and collaborative development environments. Other Skills : Problem-solving skills with attention to detail. Ability to write clean, maintainable code and automate workflows. Good understanding of HTTP, HTML, CSS, and JavaScript. Familiarity with cloud services (AWS, Azure, GCP) is a plus. Python Nice to Have: Experience with cloud-based scraping tools or services (e.g., AWS Lambda, Google Cloud Functions). Familiarity with distributed scraping and data pipeline management. Experience with large-scale data collection and storage systems. Knowledge of ethical and legal issues related to web scraping. About Nuvoretail (www.nuvoretail.com) Nuvoretail Enlytical Technologies Private Limited is an e-commerce analytics and automation company. Our proprietary digital shelf analytics and automation platform called Enlytical.ai helps e-commerce brands solve the complexities in today’s e-commerce landscape by offering a unified and all- encompassing business view on the various aspects of e-commerce business. Our platform leverages insights drawn from multiple data points that help our clients win in e-commerce by gaining a competitive edge with data-driven insights for sharper decision-making. The insights cover all aspects of e-commerce such as digital product portfolio analysis, supply chain analytics, e-commerce operations automation, pricing, and competitor benchmarking, and Amazon advertising automation using our proprietary algorithms. As a leading e-commerce service provider, we offer the most comprehensive end-to-end e-commerce solutions to brands, both in India and abroad. Right from preparing a road map to writing our client’s e- commerce success story to assisting them In increasing their online sales, we do everything via our diverse e-commerce services and bespoke strategies and technology. Our services span across the brand’s e-commerce enablement including content and digital asset creation for product listing, On Platform, and Off Platform marketing services with deep expertise in Amazon Marketing Services (AMS), Amazon SEO through keyword research, e-Commerce operations across various e-commerce platforms, website development, social media marketing, and AI-enabled e-Commerce MIS Dashboards. Awards & Recognition: Thanks to the faith reposed on us by our clients, NuvoRetail has been featured as "The Most Promising Ecommerce Technology Service Providers in India 2020” by CIOReviewIndia Magazine. Our leadership is often acknowledged by leading e-commerce services, digital marketing, consulting, and other e- commerce programs around the world. We are now one of the very few companies in India that have become an Amazon Ads Advanced partner.

Posted 4 days ago

Apply

4.0 - 8.0 years

0 Lacs

ahmedabad, gujarat

On-site

Role Overview: As a Tech Lead specializing in Python & Web Data Solutions, you will play a crucial role in leading a team of developers to design, develop, and implement scalable data scraping solutions. Your expertise in advanced scraping techniques, reverse engineering, and automation will be vital in driving the team towards success. Key Responsibilities: - Design and develop scalable data scraping solutions using tools like Scrapy and Python libraries. - Lead and mentor a team of 5+ developers, managing project timelines and deliverables. - Implement advanced blocking and captcha-solving techniques to bypass scraping restrictions. - Conduct source code reverse engineering and automate web and app interactions. - Manage proxies, IP rotation, and SSL unpinning to ensure effective scraping. - Maintain and improve API integrations and data pipelines. - Ensure code quality through effective version control, error handling, and documentation. - Collaborate with cross-functional teams for project planning and execution. - Monitor performance and provide solutions under high-pressure environments. Qualifications Required: - Minimum 4 years of experience in the IT scraping industry. - Minimum 2 years of experience leading a team of 5+ developers. - Proficiency in tools such as Scrapy, Threading, requests, and web automation. - Technical expertise in advanced Python, captcha solving, blocking handling, source reverse engineering, proxy management, IP rotation, app automation, SSL Unpin, Frida, API Management, Version Control Systems, Error Handling, SQL, MongoDB, and Pandas. - Leadership skills including basic project management, moderate documentation, team handling, pressure management, flexibility, adaptability, and high accountability. Additional Details: The preferred (good to have) qualifications include experience with Linux and knowledge of tools like Appium, Fiddler, and Burp Suite.,

Posted 4 days ago

Apply

2.0 - 6.0 years

0 Lacs

bhopal, madhya pradesh

On-site

As a Data Extraction Engineer at Relu Consultancy, your role will involve designing, implementing, and maintaining robust data scraping solutions using Python (Selenium) to drive forward cutting-edge projects. You will have the opportunity to enjoy a work schedule tailored to your needs. Key Responsibilities: - Work on web scraping or data extraction through Selenium/Scrapy or other frameworks and related libraries. - Utilize working knowledge in various DBSs, message queues & web Restful APIs. - Design, build, and maintain high-performance, reusable, and reliable Python code. - Ensure the best possible performance, quality, and responsiveness of the application. - Identify and correct bottlenecks, and fix bugs. - Assist in maintaining code quality, organization, and documentation. Qualifications: - Experience with the Python platform and object-oriented programming. - Proficiency in Python libraries such as Pandas, NumPy, Matplotlib, Beautiful Soup, Selenium, and Tabula. - Familiarity with databases like MySQL, SQL, and Mongo DB. - Proficient in IDEs like PyCharm, Spyder, and Jupiter notebook. - Strong communication skills to collaborate with team members effectively. - Analytical ability for improving program functionality. If you decide to join us, you will have the opportunity to work with a reputable consultancy firm, enjoy a flexible schedule, competitive hourly rate, and be part of a collaborative and supportive team environment. Professional growth opportunities are also available for you. If you are a Data Extraction Engineer looking for a Full-time role with a dynamic consultancy firm, we are excited to hear from you. Come and be a part of our organization's success, making a positive impact on our team and clients.,

Posted 4 days ago

Apply

2.0 - 6.0 years

0 Lacs

ahmedabad, gujarat

On-site

As an entrepreneurial, passionate, and driven Data Engineer at Startup Gala Intelligence backed by Navneet Tech Venture, you will play a crucial role in shaping the technology vision, architecture, and engineering culture of the company right from the beginning. Your contributions will be foundational in developing best practices and establishing the engineering team. **Key Responsibilities:** - **Web Scraping & Crawling:** Build and maintain automated scrapers to extract structured and unstructured data from websites, APIs, and public datasets. - **Scalable Scraping Systems:** Develop multi-threaded, distributed crawlers capable of handling high-volume data collection without interruptions. - **Data Parsing & Cleaning:** Normalize scraped data, remove noise, and ensure consistency before passing to data pipelines. - **Anti-bot & Evasion Tactics:** Implement proxy rotation, captcha solving, and request throttling techniques to handle scraping restrictions. - **Integration with Pipelines:** Deliver clean, structured datasets into NoSQL stores and ETL pipelines for further enrichment and graph-based storage. - **Data Quality & Validation:** Ensure data accuracy, deduplicate records, and maintain a trust scoring system for data confidence. - **Documentation & Maintenance:** Keep scrapers updated when websites change, and document scraping logic for reproducibility. **Qualifications Required:** - 2+ years of experience in web scraping, crawling, or data collection. - Strong proficiency in Python (libraries like BeautifulSoup, Scrapy, Selenium, Playwright, Requests). - Familiarity with NoSQL databases (MongoDB, DynamoDB) and data serialization formats (JSON, CSV, Parquet). - Experience in handling large-scale scraping with proxy management and rate-limiting. - Basic knowledge of ETL processes and integration with data pipelines. - Exposure to graph databases (Neo4j) is a plus. As part of Gala Intelligence, you will be working in a tech-driven startup dedicated to solving fraud detection and prevention challenges. The company values transparency, collaboration, and individual ownership, creating an environment where talented individuals can thrive and contribute to impactful solutions. If you are someone who enjoys early-stage challenges, thrives on owning the entire tech stack, and is passionate about building innovative, scalable solutions, we encourage you to apply. Join us in leveraging technology to combat fraud and make a meaningful impact from day one.,

Posted 4 days ago

Apply

2.0 - 6.0 years

0 Lacs

karnataka

On-site

As a Node.js Engineer at DataWeave, you will be responsible for designing and developing web scraping and automation projects using Node.js and Python. Your role will involve optimizing processes for performance and scalability, troubleshooting and debugging issues, and collaborating with team members to enhance overall architecture and performance. Stay updated with new technologies and advancements in web scraping and automation to ensure the projects are cutting-edge. **Key Responsibilities:** - Design and develop web scraping and automation projects using Node.js and Python - Optimize processes for performance and scalability - Troubleshoot and debug issues - Collaborate with team members to enhance architecture and performance - Stay updated with new technologies in web scraping and automation **Qualifications Required:** - 2+ years of experience in web scraping and automation projects - Proficiency in Node.js, Python, JavaScript, and web scraping frameworks - Familiarity with web automation tools - Experience with Git and other version control systems - Knowledge of AWS or other cloud platforms is a plus DataWeave, a leading provider of advanced sales optimization solutions for Digital Commerce businesses, offers a machine learning-enabled SaaS platform to global brands and retailers. The platform provides Digital Shelf Analytics and Dynamic Pricing solutions to develop data-driven growth strategies and make high-frequency business decisions. With actionable Digital Market Intelligence, DataWeave analyzes data on over 100 million products to track competitive KPI performance across various business areas. If you are passionate about solving challenging data problems and working at scale, DataWeave provides an opportunity to address some of the hardest data problems on the web. Join us in our mission to make sense of messy public data and contribute to the growth of eCommerce businesses globally. *Job Location:* Bangalore, India *Company Address:* InfoWeave Analytics Pvt Ltd, 75/B Windsor, 3rd Floor, Bannerghatta Rd, opp. Christ University, Hulimavu, Bengaluru, Karnataka 560076 *Contact:* +(91) 80 618 52 222,

Posted 4 days ago

Apply

2.0 - 4.0 years

2 - 4 Lacs

himatnagar

Work from Office

Responsibilities: * Design, develop & maintain web scrapers using Scrapy & Python * Collaborate with cross-functional teams on project requirements * Ensure data accuracy & compliance with industry standards Work from home

Posted 4 days ago

Apply

10.0 years

0 Lacs

india

Remote

🚀 We’re Hiring: Lead Data Engineer (SQL Server, DevOps, Python & AI) 📍 Fully Remote | Full-time We are seeking an experienced Lead Data Engineer to design, develop, and optimize scalable data platforms while driving AI-powered innovation. This role combines hands-on technical expertise with leadership responsibilities, mentoring a team of data engineers and collaborating with cross-functional teams to shape the future of data-driven IT asset management. 🔑 What You’ll Do Design and optimize SQL Server architectures with performance tuning & high availability Build and manage ETL pipelines and scalable data workflows Implement DevOps practices for database CI/CD workflows Lead Python-based data applications for processing, analytics & AI model deployment Collaborate with data scientists & engineers to deliver AI-driven solutions Ensure data governance, security & compliance standards are met Explore & recommend new data technologies for continuous innovation ✅ What We’re Looking For (Qualifications) Bachelor’s/Master’s in Computer Science, Engineering, or related field 10+ years in data engineering/database development & 2+ years in leadership Strong expertise in SQL Server (tuning, optimization, high availability) 7+ years of Python experience (Pandas, NumPy, PySpark, SQLAlchemy) Hands-on with DevOps (Azure DevOps, GitHub Actions), Docker, Kubernetes Familiar with real-time data processing (Kafka, Spark Streaming) Knowledge of data modeling, warehouses & cloud (Azure preferred) Exposure to AI/ML frameworks (TensorFlow, PyTorch, scikit-learn) & MLOps tools Experience with web scraping (Beautiful Soup, Scrapy, Selenium) Strong leadership, problem-solving, and excellent communication skills 📩 Interested? Share your resume at: 📧 seenas@theknowledgeartisans.com meghnaj@theknowledgeartisans.com

Posted 4 days ago

Apply

2.0 - 6.0 years

0 Lacs

bhopal, madhya pradesh

On-site

As a Data Extraction Engineer at Relu Consultancy, you will play a crucial role in designing, implementing, and maintaining robust data scraping solutions using Python (Selenium). Your expertise will drive forward cutting-edge projects, offering you the opportunity to work on initiatives tailored to your preferences. Responsibilities: - Work on web scraping or data extraction using Selenium/Scrapy or other related frameworks and libraries. - Utilize working knowledge in various DBSs, message queues, and web Restful APIs. - Design, build, and maintain high-performance, reusable, and reliable Python code. - Ensure optimal performance, quality, and responsiveness of applications. - Identify and rectify bottlenecks and bugs. - Assist in maintaining code quality, organization, and documentation. Qualifications: - Proficiency in Python platform and object-oriented programming. - Familiarity with Python libraries such as Pandas, NumPy, Matplotlib, Beautiful Soup, Selenium, and Tabula. - Knowledge of databases including MySQL, SQL, and MongoDB. - Experience with IDEs like PyCharm, Spyder, and Jupyter notebook. - Strong communication skills to collaborate with team members effectively. - Analytical ability for program analysis and critical thinking. Relu Consultancy offers you the following: - Opportunity to work with a reputable consultancy firm. - Flexible schedule tailored to your needs. - Competitive hourly rate. - Collaborative and supportive team environment. - Chance for professional growth and development. If you are a Data Extraction Engineer looking for a full-time role with a dynamic consultancy firm, we invite you to join us in contributing to our organization's success and making a positive impact on our team and clients.,

Posted 5 days ago

Apply

3.0 years

0 Lacs

india

Remote

Please see further details and apply here: https://apply.workable.com/polonord-adeste-srl/j/41BC0C3F26/ About Us: We are an e-commerce company dedicated to driving sales on our product listings on Amazon. As we continue to grow, we are looking for a talented Python Developer to join our team. If you have a knack for Data Integration & Automation, we’d love to hear from you! Our line of product is mainly in the health category. About the Role: We are seeking a skilled Python Developer to join our team and take ownership of building data pipelines and automation workflows. The ideal candidate will be responsible for connecting multiple data sources (APIs, web scraping, Google Sheets, Amazon Seller Central & Vendor Central, etc.), ensuring seamless integration, and setting up recurrent automations that run reliably at scale. This role requires someone who is fast, precise, and reliable, with a strong ability to write clean, maintainable code that can be easily understood and extended by other developers. Compensation: 450 - 600 € USD / month Only candidates who submit their resume in English will be eligible for consideration. Application closing date: September 25th, 2025. Responsibilities: Develop Python scripts and applications to connect diverse data sources (APIs, scraping, internal databases, Google Sheets, Amazon Seller Central & Vendor Central, etc.). Build and maintain automated pipelines for data collection, transformation, and reporting. Implement recurrent automations with proper error handling and monitoring. Deploy, manage, and scale automations in AWS environments. Collaborate with other developers and stakeholders to ensure scalability and maintainability. Write clear documentation for developed systems and automations. Continuously improve performance, reliability, and security of data integrations. Requirements: Proven experience as a Python Developer (3+ years preferred). Strong knowledge of Python libraries for APIs, web scraping (requests, BeautifulSoup, Selenium, Scrapy), and data handling (pandas, openpyxl, etc.). Hands-on experience with Amazon Seller Central and Vendor Central APIs. Experience with Google Sheets API and other third-party APIs. Proficiency with AWS services (Lambda, S3, EC2, CloudWatch, etc.) for automation and deployment. Ability to write clean, well-structured, and reusable code. Familiarity with version control (Git) and collaborative development workflows. Strong problem-solving skills and attention to detail. Excellent communication skills and ability to work in a team environment. Detail-oriented and able to manage multiple projects with tight deadlines. Excellent attention to detail, time management, and communication skills. Ability to work independently and meet tight deadlines Set-up Requirements: · Able to work in a full-time remote position. · Reside in India. · Able to work as an independent contractor. · Able to work during our core global office hours (Mon- Fri 9 am - 1 pm, 2 pm - 6 pm CET). · Quiet home office and comfortable to work from home. · Reliable, high-speed internet access (a minimum download speed of 25 Mbps and a minimum upload speed of 6 Mbps). · A newer computer is required (no older than 3 years). Additional requirements: · Very strong English proficiency. · Accountable and reliable, taking responsibility for your tasks and actions. · Self-motivated and results-oriented, taking a proactive approach to achieving goals. · Detail-oriented with strong attention to detail to produce high-quality work consistently. · Able to communicate complex tasks in writing and verbally. · Open to new challenges and responsibilities. · Actively looking to learn, improve, and grow. · Collaborate effectively with others and contribute positively to the team's success.

Posted 5 days ago

Apply

3.0 - 6.0 years

10 - 14 Lacs

coimbatore

Remote

Job Title: Web Scraping Specialist Experience: 3 - 6 Years Location: Remote (Work from Home) About the job We are seeking a highly skilled Web Scraping Specialist to join our team. The successful candidate will be responsible for designing, implementing, and maintaining web scraping processes to gather data from various online sources efficiently and accurately. As a Web Scraping Specialist, you will play a crucial role in collecting data for competitor analysis, and other business intelligence purposes. Responsibilities: Scalability/Performance: Lead and provide expertise in scraping at scale e-commerce marketplaces. Data Source Identification: Identify relevant websites and online sources from which data needs to be scraped. Collaborate with the team to understand data requirements and objectives Web Scraping Design: Develop and implement effective web scraping strategies to extract data from targeted websites. This includes selecting appropriate tools, libraries, or frameworks for the task Data Extraction: Create and maintain web scraping scripts or programs to extract the required data. Ensure the code is optimized, reliable, and can handle changes in the website's structure Data Cleansing and Validation: Cleanse and validate the collected data to eliminate errors, inconsistencies, and duplicates. Ensure data integrity and accuracy throughout the process Monitoring and Maintenance: Continuously monitor and maintain the web scraping processes. Address any issues that arise due to website changes, data format modifications, or anti-scraping mechanisms Scalability and Performance: Optimize web scraping procedures for efficiency and scalability, especially when dealing with a large volume of data or multiple data sources Compliance and Legal Considerations: Stay up-to-date with legal and ethical considerations related to web scraping, including website terms of service, copyright, and privacy regulations Documentation: Maintain detailed documentation of web scraping processes, data sources, and methodologies. Create clear and concise instructions for others to follow Collaboration: Collaborate with other teams such as data analysts, developers, and business stakeholders to understand data requirements and deliver insights effectively Security: Implement security measures to ensure the confidentiality and protection of sensitive data throughout the scraping process Requirements: Proven experience of 3+ years as a Web Scraping Specialist or similar role, with a track record of successful web scraping projects Expertise in handling dynamic content, user-agent rotation, bypass CAPTCHAs, rate limits, and utilization of proxy services Knowledge on browser fingerprinting Has leadership experience Proficiency in programming languages commonly used for web scraping, such as Python, BeautifulSoup, Scrapy, or Selenium Strong knowledge of HTML, CSS, XPath, and other web technologies relevant to web scraping and Coding Knowledge and experience in best of class data storage and retrieval of large volume of scraped data. Understanding of web scraping best practices, including handling dynamic content, user-agent rotation, and IP address management Attention to detail and the ability to handle and process large volumes of data accurately Familiarity with data cleansing techniques and data validation processes Good communication skills and the ability to collaborate effectively with cross-functional teams Knowledge of web scraping ethics, legal considerations, and compliance with website terms of service Strong problem-solving skills and the ability to adapt to changing web environments Preferred Qualifications: Bachelor's degree in Computer Science, Data Science, Information Technology, or related fields Experience with cloud-based solutions and distributed web scraping systems Familiarity with APIs and data extraction from non-public sources Knowledge of machine learning techniques for data extraction and natural language processing is desired but not mandatory Prior experience in handling large-scale data projects and working with big data frameworks Understanding of various data formats such as JSON, XML, CSV, etc Experience with version control systems like Git

Posted 5 days ago

Apply

5.0 years

0 Lacs

noida, uttar pradesh, india

On-site

Job Title: Lead Data Scientist Location: Noida Employment Type: Full-time Responsibilities: Lead and manage a team of data scientists, providing guidance, mentorship, and technical expertise Collaborate with stakeholders to understand business objectives and develop data science strategies aligned with those objectives Design, develop, and implement advanced machine learning and statistical models to solve complex business problems. Conduct exploratory data analysis, data cleansing, and feature engineering to prepare datasets for analysis. Good in fine-tuning, and optimization of LLMs like Open AI, Gemini AI, Copilot, and open-source models like LLAMA, DeepSeek, and BERT Hands-on experience in Prompt Engineering and RAG for real-world applications Develop Conversational AI, ChatBots, VoiceBots using modern NLP frameworks like SpaCy, Rasa, LangChain, Hugging Face Use vector databases like Pinecone, Vertex AI, ChromaDB, or SingleStore to manage embeddings at scale Deploy and monitor models on cloud platforms like Azure and GCP Lead client engagements, RFP responses generation, and PoC development for both internal and external stakeholders Provide technical guidance, code reviews, and performance optimization support Explore and utilize various data mining and machine learning techniques to extract valuable insights and patterns from large datasets. Develop predictive/prescriptive models, algorithms, and prototypes to support business. Communicate findings and insights to technical and non-technical stakeholders through reports, presentations, and data visualizations. Play an active role in software development processes. Stay up to date with the latest trends and advancements in data science, machine learning, and AI technologies and identify opportunities to apply them to improve business outcomes. Qualifications: A B.Tech / M.Tech / MBA from a top tier Institute preferably in a quantitative subject 5+ years of hands-on experience in applied Machine Learning, AI and analytics with a proven track record of developing AI/LLM models and algorithms A master’s or bachelor’s degree in data science, Computer Science, Statistics, or a related field Strong programming skills in scripting languages like Python, R, SQL, NoSQL, Spark with ML tools & Cloud Technology (AWS, Azure, GCP) Expertise in Python libraries such as numpy, pandas, scikit-learn, tensor-flow, scrapy, BERT etc. Strong expertise in Machine Learning, Deep Learning, Data mining, and Statistical concepts and experience in developing models and solutions in these areas Expertise in client engagement, understanding complex problem statements, and offering solutions in the domains of Supply Chain, Manufacturing, CPG, Marketing etc. Strong publication record in top-tier journals and conferences. Functional thought leader, sought after for key tech decisions. Can successfully sell ideas to an executive level decision maker. Mentors and trains the research scientist community on complex technical issues. Experience developing software in traditional programming languages (C++, Java, etc..). Excellent written and spoken communication skills About Us E-Solutions is a leading AI consulting firm delivering cutting-edge solutions across recruitment & staffing, IT services, finance, retail, education, and manufacturing. We specialize in building advanced AI/ML products including large-scale recommender systems, LLM-based chatbots, AI-powered resume engines, GenAI copilots, and enterprise-wide automation systems. As part of our growing AI leadership team, we are looking for a Principal Data Scientist who can architect, lead, and implement end-to-end AI/ML solutions using the latest in Generative AI, LLMs, Vector Databases, and NLP frameworks. Reimagine the Future of IT Consulting & Development with ML, GenAI, & Agentic AI At E-Solutions, we empower global businesses by enhancing their IT infrastructure and software capabilities. Over two decades, we’ve grown into a trusted global partner, operating in 35+ countries with a team of 5,000+ professionals. Our expertise in Managed IT Services and Software Product Engineering, combined with robust testing, cloud solutions and AI/ML/Gen AI in NLP & Computer Vision domains ensures that the enterprise thrives in a competitive digital landscape. E-Solutions has achieved significant success in implementing ML/GenAI/Agentic AI projects like AI enabled matching of professional’s profiles based on customer requirements, Agentic AI based end-to-end recruitment platform which will process & summarize customer requirements and AI agent will conduct online interviews and have efficient assessment to retrieve relevant profiles, Conversational AI Engine, Detecting Suspicious Behaviour during interviews etc. Simultaneously, the company successfully executed mergers and acquisitions, driving significant growth and expansion and embarking on a transformative digital journey. Join us to be the part of this innovative journey to build cutting-edge advance Agentic AI platform to deliver AI/ML/GenAI use cases across industries.

Posted 5 days ago

Apply

10.0 years

0 Lacs

noida, uttar pradesh, india

On-site

Job Title: Principal Data Scientist Location: Noida Employment Type: Full-time Responsibilities: Convert broad vision and concepts into a structured data science roadmap and guide a team to successfully execute on it Collaborate with AI Product managers, UI/UX, stakeholders, etc. to design, develop, and deploy AI/LLM models Good in fine-tuning, and optimization of LLMs like Open AI, Gemini AI, Copilot, and open-source models like LLAMA, DeepSeek, and BERT Hands-on experience in Prompt Engineering and RAG for real-world applications Develop Conversational AI, ChatBots, VoiceBots using modern NLP frameworks like SpaCy, Rasa, LangChain, Hugging Face Use vector databases like Pinecone, Vertex AI, ChromaDB, or SingleStore to manage embeddings at scale Deploy and monitor models on cloud platforms like Azure and GCP Lead client engagements, RFP responses generation, and PoC development for both internal and external stakeholders Provide technical guidance, code reviews, and performance optimization support Expertise in designing solutions using semantic search and vector embeddings Translate high-level business goals into measurable, AI-powered deliverables Proven ability to discover solutions hidden in large datasets and to drive business results with their data-based insights Contribute to internal product development initiatives related to data science. Drive excellent project management skills required to deliver complex projects, including effort/time estimation Be proactive, with full ownership of the engagement. Build scalable client engagement level processes for faster turnaround & higher accuracy Qualifications: A B.Tech / M.Tech / MBA from a top tier Institute preferably in a quantitative subject 10+ years of hands-on experience in applied Machine Learning, AI and analytics Experience of scientific programming in scripting languages like Python, R, SQL, NoSQL, Spark with ML tools & Cloud Technology (AWS, Azure, GCP) Expertise in Python libraries such as numpy, pandas, scikit-learn, tensor-flow, scrapy, BERT etc. Strong expertise in Machine Learning, Deep Learning, Data mining, and Statistical concepts and experience in developing models and solutions in these areas Expertise in client engagement, understanding complex problem statements, and offering solutions in the domains of Supply Chain, Manufacturing, CPG, Marketing etc. Strong publication record in top-tier journals and conferences. Functional thought leader, sought after for key tech decisions. Can successfully sell ideas to an executive level decision maker. Mentors and trains the research scientist community on complex technical issues. Experience developing software in traditional programming languages (C++, Java, etc..). Excellent written and spoken communication skills About Us E-Solutions is a leading AI consulting firm delivering cutting-edge solutions across recruitment & staffing, IT services, finance, retail, education, and manufacturing. We specialize in building advanced AI/ML products including large-scale recommender systems, LLM-based chatbots, AI-powered resume engines, GenAI copilots, and enterprise-wide automation systems. As part of our growing AI leadership team, we are looking for a Principal Data Scientist who can architect, lead, and implement end-to-end AI/ML solutions using the latest in Generative AI, LLMs, Vector Databases, and NLP frameworks. Reimagine the Future of IT Consulting & Development with ML, GenAI, & Agentic AI At E-Solutions, we empower global businesses by enhancing their IT infrastructure and software capabilities. Over two decades, we’ve grown into a trusted global partner, operating in 35+ countries with a team of 5,000+ professionals. Our expertise in Managed IT Services and Software Product Engineering, combined with robust testing, cloud solutions and AI/ML/Gen AI in NLP & Computer Vision domains ensures that the enterprise thrives in a competitive digital landscape. E-Solutions has achieved significant success in implementing ML/GenAI/Agentic AI projects like AI enabled matching of professional’s profiles based on customer requirements, Agentic AI based end-to-end recruitment platform which will process & summarize customer requirements and AI agent will conduct online interviews and have efficient assessment to retrieve relevant profiles, Conversational AI Engine, Detecting Suspicious Behaviour during interviews etc. Simultaneously, the company successfully executed mergers and acquisitions, driving significant growth and expansion and embarking on a transformative digital journey. Join us to be the part of this innovative journey to build cutting-edge advance Agentic AI platform to deliver AI/ML/GenAI use cases across industries.

Posted 5 days ago

Apply

4.0 years

0 Lacs

ahmedabad, gujarat, india

On-site

Job Title: Tech Lead - Python & Web Data Solutions Location : Ahmedabad / WFO Experience Level : Senior (4+ years) Employment Type : Full-time Job Summary We are seeking a highly skilled and experienced Lead Data Scraping Engineer to join our team. The ideal candidate will have a minimum of 4 years of hands-on experience in IT scraping, with at least 2 years leading a team of 5+ developers. This role requires deep technical knowledge in advanced scraping techniques, reverse engineering, automation, and leadership skills to drive the team towards success. Key Responsibilities Design and develop scalable data scraping solutions using tools like Scrapy and Python libraries. Lead and mentor a team of 5+ developers, managing project timelines and deliverables. Implement advanced blocking and captcha-solving techniques to bypass scraping restrictions. Conduct source code reverse engineering and automate web and app interactions. Manage proxies, IP rotation, and SSL unpinning to ensure effective scraping. Maintain and improve API integrations and data pipelines. Ensure code quality through effective version control, error handling, and documentation. Collaborate with cross-functional teams for project planning and execution. Monitor performance and provide solutions under high-pressure environments. Required Skills and Experience Data Scraping: Minimum 4 years in IT scraping industry Leadership : Minimum 2 years leading a team of 5+ developers Scraping Tools: Scrapy, Threading, requests, web automation Technical Proficiency: Advanced Python Captcha solving and blocking handling Source reverse engineering Proxy management & IP rotation App automation, SSL Unpin, Frida API Management, Version Control Systems Error Handling, SQL, MongoDB, Pandas Leadership Skills Basic project management Moderate documentation Team handling Pressure management Flexibility and adaptability High accountability Preferred (Good to Have) Experience with Linux Knowledge of Appium, Fiddler, Burp Suite

Posted 5 days ago

Apply

2.0 years

0 Lacs

ahmedabad, gujarat, india

On-site

We're looking for an entrepreneurial, passionate, and driven Data Engineer to join Startup Gala Intelligence backed by Navneet Tech Venture. As we're building our technology platform from scratch, you'll have the unique opportunity to shape our technology vision, architecture, and engineering culture right from the ground up. You’ll directly contribute to foundational development and establish best practices, while eventually building and contributing to our engineering team. This role is ideal for someone eager to own the entire tech stack, who thrives on early-stage challenges, and loves building innovative, scalable solutions from day zero. What You’ll Do Web Scraping & Crawling: Build and maintain automated scrapers to extract structured and unstructured data from websites, APIs, and public datasets. Scalable Scraping Systems: Develop multi-threaded, distributed crawlers capable of handling high-volume data collection without interruptions. Data Parsing & Cleaning: Normalize scraped data, remove noise, and ensure consistency before passing to data pipelines. Anti-bot & Evasion Tactics: Implement proxy rotation, captcha solving, and request throttling techniques to handle scraping restrictions. Integration with Pipelines: Deliver clean, structured datasets into NoSQL stores and ETL pipelines for further enrichment and graph-based storage. Data Quality & Validation: Ensure data accuracy, deduplicate records, and maintain a trust scoring system for data confidence. Documentation & Maintenance: Keep scrapers updated when websites change, and document scraping logic for reproducibility. Who You Are Technical Skills: 2+ years of experience in web scraping , crawling, or data collection. Strong proficiency in Python (libraries like BeautifulSoup, Scrapy, Selenium, Playwright, Requests). Familiarity with NoSQL databases (MongoDB, DynamoDB) and data serialization formats (JSON, CSV, Parquet). Experience in handling large-scale scraping with proxy management and rate-limiting. Basic knowledge of ETL processes and integration with data pipelines. Exposure to graph databases (Neo4j) is a plus. Soft Skills: Detail-oriented, ensuring accuracy and reliability of collected data. Strong problem-solving skills, particularly in adapting scrapers to evolving web structures. Curious mindset with a drive to discover new data sources. Comfortable working in a fast-paced, early-stage startup environment. Who We Are & Our Culture Gala Intelligence , backed by Navneet Tech Ventures , is a tech-driven startup dedicated to solving one of the most pressing business challenges - fraud detection and prevention. We're building cutting-edge, real-time products designed to empower consumers and businesses to stay ahead of fraudsters, leveraging innovative technology and deep domain expertise. Our culture and values: We’re united by a single, critical mission - stopping fraud before it impacts businesses. Curiosity, innovation, and proactive action define our approach. We value transparency, collaboration, and individual ownership, creating an environment where talented people can do their best work. Problem-Driven Innovation : We're deeply committed to solving real challenges that genuinely matter for our customers. Rapid Action & Ownership : We encourage autonomy and accountability—own your projects, move quickly, and shape the future of Gala Intelligence. Collaborative Excellence : Cross-team collaboration ensures alignment, sparks innovation, and drives us forward together. Continuous Learning : Fraud evolves rapidly, and so do we. Continuous improvement, experimentation, and learning are core to our success. If you're excited by the opportunity to leverage technology in the fight against fraud, and you're ready to build something impactful from day one, we want to hear from you!

Posted 6 days ago

Apply

0 years

0 Lacs

bhopal, madhya pradesh, india

On-site

Relu Consultancy is seeking a Data Extraction Engineer with expertise in Python (Selenium). In this role, you will design, implement, and maintain robust data scraping solutions that drive our projects forward. This is your chance to contribute to cutting-edge initiatives while enjoying a work schedule tailored to your needs. Job Title: Data Extraction Engineer Location: Bhopal, MP Job Type: Full-Time CTC: 5LPA Responsibility: 1. Work on web scraping or data extraction through Selenium/Scrapy or other frameworks and related libraries. 2. Working knowledge in various DBSs, message queues & web Restful APIs. 3. Design, build, and maintain high-performance, reusable, and reliable Python code. 4. Ensure the best possible performance, quality, and responsiveness of the application. 5. Identify and correct bottlenecks and fix bugs. 6. Help maintain code quality, organization, and documentation. Qualification: 1. Experience with the Python platform, and object-oriented programming. 2. Python libraries - Pandas, NumPy, Matplotlib, Beautiful Soup, Selenium, Tabula. 3. Data Base - MySQL, SQL, Mongo DB. 4. IDE - PyCharm, Spyder, and Jupiter notebook. 5. Communication skills - Python developers need strong verbal communication skills to work with other members of the programming team and participate in a collaborative environment. 6. Analytical ability - because Python developers analyze programs to improve their functionality, these professionals have strong analytical skills and critical thinking abilities. Why Join Us: Opportunity to work with a reputable consultancy firm. Flexible schedule. Competitive hourly rate. Collaborative and supportive team environment. Opportunity for professional growth. If you're a Data Extraction Engineer seeking a Full-time role that offers the chance to work with a dynamic consultancy firm, we'd love to hear from you. Join us in contributing to our organization's success and in making a positive impact on our team and clients.

Posted 6 days ago

Apply

1.0 - 4.0 years

0 Lacs

gurugram, haryana, india

On-site

Be a part of India’s largest and most admired news network! Network18 is India's most diversified Media Company in the fast growing Media market. The Company has a strong Heritage and we possess a strong presence in Magazines, Television and Internet domains. Our brands like CNBC, Forbes and Moneycontrol are market leaders in their respective segments. The Company has over 7,000 employees across all major cities in India and has been consistently managed to stay ahead of the growth curve of the industry. Network 18 brings together employees from varied backgrounds under one roof united by the hunger to create immersive content and ideas. We take pride in our people, who we believe are the key to realizing the organization’s potential. We continually strive to enable our employees to realize their own goals, by providing opportunities to learn, share and grow. Role Overview: We are seeking a passionate and skilled Data Scientist with over a year of experience to join our dynamic team. You will be instrumental in developing and deploying machine learning models, building robust data pipelines, and translating complex data into actionable insights. This role offers the opportunity to work on cutting-edge projects involving NLP, Generative AI, data automation, and cloud technologies to drive business value. Key Responsibilities: Design, develop, and deploy machine learning models, with a strong focus on NLP (including advanced techniques and Generative AI) and other AI applications. Build, maintain, and optimize ETL pipelines for automated data ingestion, transformation, and standardization from various sources Work extensively with SQL for data extraction, manipulation, and analysis in environments like BigQuery. Develop solutions using Python and relevant data science/ML libraries (Pandas, NumPy, Hugging Face Transformers, etc.). Utilize Google Cloud Platform (GCP) services for data storage, processing, and model deployment. Create and maintain interactive dashboards and reporting tools (e.g., Power BI) to present insights to stakeholders. Apply basic Docker concepts for containerization and deployment of applications. Collaborate with cross-functional teams to understand business requirements and deliver data-driven solutions. Stay abreast of the latest advancements in AI/ML and NLP best practices. Required Qualifications & Skills: 1 to 4 years of hands-on experience as a Data Scientist or in a similar role. Solid understanding of machine learning fundamentals, algorithms, and best practices. Proficiency in Python and relevant data science libraries. Good SQL skills for complex querying and data manipulation. Demonstrable experience with Natural Language Processing (NLP) techniques, including advanced models (e.g., transformers) and familiarity with Generative AI concepts and applications. Excellent problem-solving and analytical skills. Strong communication and collaboration skills. Preferred Qualifications & Skills: Familiarity and hands-on experience with Google Cloud Platform (GCP) services, especially BigQuery, Cloud Functions, and Vertex AI. Basic understanding of Docker and containerization for deploying applications. Experience with dashboarding tools like Power BI and building web applications with Streamlit. Experience with web scraping tools and techniques (e.g., BeautifulSoup, Scrapy, Selenium). Knowledge of data warehousing concepts and schema design. Experience in designing and building ETL pipelines. Disclaimer: Please note Network18 and related group companies do not use the services of vendors or agents for recruitment. Please beware of such agents or vendors providing assistance. Network18 will not be responsible for any losses incurred. “We correspond only from our official email address”

Posted 6 days ago

Apply

2.0 - 3.0 years

0 Lacs

dwarka, delhi, india

On-site

Position: Data Mining Analyst Please Note: Only Candidates with 2-3 years of experience should apply. This is a Delhi-based position and work from office only! Work Location: Sector 23 Dwarka, Delhi We are seeking a skilled Data Mining Analyst with expertise in automating data extraction processes from web platforms. The ideal candidate will be experienced in Python, Selenium, Pandas, SQL, and APIs, with the ability to design and implement efficient and scalable data scraping systems. If you have a passion for working with data and a solid understanding of web technologies, we want to hear from you! Key Responsibilities: Design, develop, and maintain robust web scraping solutions to extract structured and unstructured data from various websites and APIs. Use tools like Python, Selenium, BeautifulSoup, Scrapy, and Pandas for data scraping and processing. Build and manage automated scripts to scrape dynamic websites, including handling JavaScript-driven content. Optimize scraping workflows to ensure data extraction is efficient, accurate, and scalable. Work with APIs to gather and integrate data, ensuring proper rate limits and authentication handling. Clean, preprocess, and store extracted data in databases (SQL) or cloud-based systems. Collaborate with data analysts and other stakeholders to provide required data for further analysis and reporting. Debug and troubleshoot issues in scraping pipelines and scripts. Ensure compliance with ethical data scraping standards, including legal considerations like website terms of use and robots.txt policies. Required Skills & Qualifications: Experience : 2-3 years of hands-on experience in web scraping and data extraction. Technical Skills : Strong proficiency in Python. Experience with web scraping frameworks and libraries like Selenium, Scrapy, BeautifulSoup, and Requests. Experience with data manipulation libraries like Pandas. Familiarity with API integration (REST, GraphQL, etc.). Proficiency in SQL for data querying, database design, and managing large datasets. Knowledge of JavaScript and front-end technologies to work with dynamic web pages. Experience with version control (Git) and collaborative development environments. Other Skills : Problem-solving skills with attention to detail. Ability to write clean, maintainable code and automate workflows. Good understanding of HTTP, HTML, CSS, and JavaScript. Familiarity with cloud services (AWS, Azure, GCP) is a plus. Python Nice to Have: Experience with cloud-based scraping tools or services (e.g., AWS Lambda, Google Cloud Functions). Familiarity with distributed scraping and data pipeline management. Experience with large-scale data collection and storage systems. Knowledge of ethical and legal issues related to web scraping. About Nuvoretail (www.nuvoretail.com) Nuvoretail Enlytical Technologies Private Limited is an e-commerce analytics and automation company. Our proprietary digital shelf analytics and automation platform called Enlytical.ai helps e-commerce brands solve the complexities in today’s e-commerce landscape by offering a unified and all- encompassing business view on the various aspects of e-commerce business. Our platform leverages insights drawn from multiple data points that help our clients win in e-commerce by gaining a competitive edge with data-driven insights for sharper decision-making. The insights cover all aspects of e-commerce such as digital product portfolio analysis, supply chain analytics, e-commerce operations automation, pricing, and competitor benchmarking, and Amazon advertising automation using our proprietary algorithms. As a leading e-commerce service provider, we offer the most comprehensive end-to-end e-commerce solutions to brands, both in India and abroad. Right from preparing a road map to writing our client’s e- commerce success story to assisting them In increasing their online sales, we do everything via our diverse e-commerce services and bespoke strategies and technology. Our services span across the brand’s e-commerce enablement including content and digital asset creation for product listing, On Platform, and Off Platform marketing services with deep expertise in Amazon Marketing Services (AMS), Amazon SEO through keyword research, e-Commerce operations across various e-commerce platforms, website development, social media marketing, and AI-enabled e-Commerce MIS Dashboards. Awards & Recognition: Thanks to the faith reposed on us by our clients, NuvoRetail has been featured as "The Most Promising Ecommerce Technology Service Providers in India 2020” by CIOReviewIndia Magazine. Our leadership is often acknowledged by leading e-commerce services, digital marketing, consulting, and other e- commerce programs around the world. We are now one of the very few companies in India that have become an Amazon Ads Advanced partner.

Posted 6 days ago

Apply
cta

Start Your Job Search Today

Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.

Job Application AI Bot

Job Application AI Bot

Apply to 20+ Portals in one click

Download Now

Download the Mobile App

Instantly access job listings, apply easily, and track applications.

Featured Companies