Home
Jobs

83 Scrapy Jobs - Page 3

Filter
Filter Interviews
Min: 0 years
Max: 25 years
Min: ₹0
Max: ₹10000000
Setup a job Alert
JobPe aggregates results for easy application access, but you actually apply on the job portal directly.

3.0 - 7.0 years

1 - 2 Lacs

Mumbai, Thane, Navi Mumbai

Work from Office

Naukri logo

Key Responsibilities: Develop and maintain automated web scraping scripts using Python libraries such as BeautifulSoup, Scrapy, and Selenium. Optimize scraping pipelines for performance, scalability, and resource efficiency. Handle dynamic websites, CAPTCHA-solving, and implement IP rotation techniques for uninterrupted scraping. Process and clean raw data, ensuring accuracy and integrity in extracted datasets. Collaborate with cross-functional teams to understand data requirements and deliver actionable insights. Leverage APIs when web scraping is not feasible, managing authentication and request optimization. Document processes, pipelines, and troubleshooting steps for maintainable and reusable scraping solutions. Ensure compliance with legal and ethical web scraping practices, implementing security safeguards. Requirements: Education : Bachelors degree in Computer Science, Engineering, or a related field. Experience : 2+ years of Python development experience, with at least 1 year focused on web scraping. Technical Skills : Proficiency in Python and libraries like BeautifulSoup, Scrapy, and Selenium. Experience with regular expressions (Regex) for data parsing. Strong knowledge of HTTP protocols, cookies, headers, and user-agent rotation. Familiarity with databases (SQL and NoSQL) for storing scraped data. Hands-on experience with data manipulation libraries such as pandas and NumPy. Experience working with APIs and managing third-party integrations. Familiarity with version control systems like Git. Bonus Skills : Knowledge of containerization tools like Docker. Experience with distributed scraping solutions and task queues (e.g., Celery, RabbitMQ). Basic understanding of data visualization tools. Non-Technical Skills : Strong analytical and problem-solving skills. Excellent communication and documentation skills. Ability to work independently and collaboratively in a team environment. CANDIDATES AVAILABLE FOR FACE-TO-FACE INTERVIEWS ARE PREFERRED.

Posted 3 weeks ago

Apply

2.0 - 4.0 years

0 Lacs

Gurugram, Haryana, India

On-site

Linkedin logo

The D. E. Shaw group is a global investment and technology development firm with more than $65 billion in investment capital as of December 1, 2024, and offices in North America, Europe, and Asia. Since our founding in 1988, our firm has earned an international reputation for successful investing based on innovation, careful risk management, and the quality and depth of our staff. We have a significant presence in the world's capital markets, investing in a wide range of companies and financial instruments in both developed and developing economies. We are looking for resourceful and exceptional candidates for the Data Engineer role within our product development teams based out of Hyderabad. At DESIS, the Data Engineers develop Web Robots, or Web Spiders, that crawl through the web and retrieve data in the form of HTML, plain text, PDFs, Excel, and any other format that is either structured or unstructured. The job functions of the engineer also include scraping the website data into a structured format and building automated and custom reports on the downloaded data that are used as knowledge for business purposes. The team also works on automating end-to-end data pipelines. WHAT YOU'LL DO DAY-TO-DAY: As a member of the Data Engineering team, you will be responsible for various aspects of data extraction, such as understanding the data requirements of the business group, reverse-engineering the website, its technology, and the data retrieval process, re-engineering by developing web robots to automate the extraction of the data, and building monitoring systems to ensure the integrity and quality of the extracted data. You will also be responsible for managing the changes to the website's dynamics and layout to ensure clean downloads, building scraping and parsing systems to transform raw data into a structured form, and offering operations support to ensure high availability and zero data losses. Additionally, you will be involved in other tasks such as storing the extracted data in the recommended databases, building high-performing, scalable data extraction systems, and automating data pipelines. WHO WE’RE LOOKING FOR: Basic qualifications: 2-4 years of experience in website data extraction and scraping Good knowledge of relational databases, writing complex queries in SQL, and dealing with ETL operations on databases Proficiency in Python for performing operations on data Expertise in Python frameworks like Requests, UrlLib2, Selenium, Beautiful Soup, and Scrapy A good understanding of HTTP requests and responses, HTML, CSS, XML, JSON, and JavaScript Expertise with debugging tools in Chrome to reverse engineer website dynamics A good academic background and accomplishments A BCA/MCA/BS/MS degree with a good foundation and practical application of knowledge in data structures and algorithms Problem-solving and analytical skills Good debugging skills Interested candidates can apply through our website: https://www.deshawindia.com/recruit/jobs/Adv/Link/SnrMemDEFeb25 We encourage candidates with relevant experience looking to restart their careers after a break to apply for this position. Learn about Recommence, our gender-neutral return-to-work initiative. The Firm offers excellent benefits, a casual, collegial working environment, and an attractive compensation package. For further information about our recruitment process, including how applicant data will be processed, please visit https://www.deshawindia.com/careers Members of the D. E. Shaw group do not discriminate in employment matters on the basis of sex, race, colour, caste, creed, religion, pregnancy, national origin, age, military service eligibility, veteran status, sexual orientation, marital status, disability, or any other protected class. Show more Show less

Posted 3 weeks ago

Apply

2.0 - 4.0 years

0 Lacs

India

On-site

Linkedin logo

Alternative Path is seeking skilled software developers to collaborate on client projects with an asset management firm. In this role, you will collaborate with individuals across various company departments to shape and innovate new products and features for our platform, enhancing existing ones. You will have a large degree of independence and trust, but you won't be isolated; the support of the Engineering team leads, the Product team leads, and every other technology team member is behind you. This is an opportunity to join a team-first meritocracy and help grow an entrepreneurial group inside Alternative Path. You will be asked to contribute, given ownership, and will be expected to make your voice heard. Role Summary: Performing Web Scraping using various scraping techniques and then utilizing Python’s Pandas library for data cleaning and manipulation. Then ingesting the data into a Database/Warehouse, and scheduling the scrapers using Airflow or other tools Role Overview The Web Scraping Team at Alternative Path is seeking a creative and detail-oriented developer to contribute to client projects. The team develops essential applications, datasets, and alerts for various teams within the client's organization, supporting their daily investment decisions. The mission is to maintain operational excellence by delivering high-quality proprietary datasets, timely notifications, and exceptional service. We are seeking someone who is self-motivated, self-sufficient, with a passion for tinkering and a love for automation. In your role, you will: ➢ Collaborate with analysts to understand and anticipate requirements. ➢ Design, implement, and maintain Web scrapers for a wide variety of alternative datasets. ➢ Perform Data Cleaning, Exploration, Transformation etc. of scraped data. ➢ Collaborate with cross-functional teams to understand data requirements and implement efficient data processing workflows. ➢ Author QC checks to validate data availability and integrity. ➢ Maintain alerting systems and investigate time-sensitive data incidents to ensure smooth day-to-day operations. ➢ Design and implement products and tools to enhance the Web scraping Platform. Qualifications Must have ➢ Bachelor's/master’s degree in computer science or in any related field ➢ 2-4 years of software development experience ➢ Strong Python and SQL/Database skills ➢ Strong expertise in using the Pandas library (Python) is a must ➢ Experience with web technologies (HTML/JS, APIs, etc.) ➢ Proven work experience in working with large data sets for Data cleaning, Data transformation, Data manipulation, and Data replacements. ➢ Excellent verbal and written communication skills ➢ Aptitude for designing infrastructure, data products, and tools for Data Scientists Preferred ➢ Familiarity with scraping and common scraping tools (Selenium, scrapy, Fiddler, Postman, xpath) ➢ Experience containerizing workloads with Docker (Kubernetes a plus) ➢ Experience with build automation (Jenkins, Gitlab CI/CD) ➢ Experience with AWS technologies like S3, RDS, SNS, SQS, Lambda, etc. Show more Show less

Posted 3 weeks ago

Apply

0.0 - 1.0 years

0 Lacs

Mohali, Punjab

On-site

Indeed logo

Python Developer Experience: 1- 3 years Responsibilities: · Develop and maintain web applications using Python frameworks like Django and Flask. · Utilize data analysis libraries such as NumPy, pandas, and matplotlib for extracting insights from large datasets. · Implement machine learning models using TensorFlow and PyTorch for various applications. · Handle HTTP requests efficiently using libraries like Requests. · Perform web scraping tasks using Beautiful Soup and Scrapy. · Collaborate with cross-functional teams to design and implement scalable solutions. · Ensure the security and integrity of data by integrating with database systems like SQL (e.g., PostgreSQL, MySQL) or NoSQL databases (e.g., MongoDB). · Stay updated with the latest trends and technologies in the Python ecosystem. Requirements: · Bachelor's degree in Computer Science, Engineering, or related field. · Strong proficiency in Python programming language. · Experience with Python frameworks such as Django and Flask for web development. · Familiarity with data analysis libraries like NumPy, pandas, and matplotlib. · Knowledge of machine learning frameworks such as TensorFlow and PyTorch. · Experience in handling HTTP requests and web scraping. · Understanding of database systems like SQL (e.g., PostgreSQL, MySQL) or NoSQL databases (e.g., MongoDB). · Excellent problem-solving skills and attention to detail. · Ability to work independently as well as in a team environment. · Strong communication and interpersonal skills. Job Type: Full Time, Permanent, Regular 5 Days Working Location: Plot no.968, Sector-82, JLPL Industrial Area, Mohali, Punjab (140306) Job Types: Full-time, Permanent Schedule: Day shift Monday to Friday Supplemental Pay: Overtime pay Performance bonus Ability to commute/relocate: Mohali, Punjab 140306: Reliably commute or planning to relocate before starting work (Required) Education: Bachelor's (Preferred) Experience: Total work: 1 year (Required) Python: 1 year (Required) Work Location: In person

Posted 3 weeks ago

Apply

5.0 years

0 Lacs

Gurugram, Haryana

On-site

Indeed logo

Position: Data Engineer/Analyst Experience: Minimum 5 years in data engineer/analyst roles Location: Sector 20, Gurugram, Haryana, India Job Type: Full-Time Working Hours: 7:00 AM – 4:00 PM IST, Monday to Friday Key Responsibilities: Data Extraction & Automation Build or maintain scripts/APIs to automatically extract business data (name, address, phone, etc.) from Google Maps and other relevant sources. Clean and structure the scraped data for sales/marketing campaigns. Customer & Lead Data Analysis Analyse existing customer and lead datasets to identify trends, segments, and upsell/cross-sell opportunities. Create and maintain dashboards or reports to monitor KPIs and campaign effectiveness. Implement customer segmentation models to support targeted outreach strategies. Telemarketing Data Analysis Analyse cold calling and telemarketing performance data to uncover productivity gaps and suggest improvements. Help optimise call strategies by identifying the best times, industries, and segments to target. Collaboration & Strategy Work with sales and marketing teams to translate business needs into technical solutions. Recommend data-driven strategies to improve conversion, retention, and sales growth. Required Skills & Experience Proven experience as a Data Analyst, Data Engineer, or similar role. Experience with web scraping tools (e.g., Python with Beautiful Soup, Selenium, Scrapy, or Google Maps API). Strong skills in SQL, Python or R for data manipulation and analysis. Proficiency with data visualisation tools (e.g., Power BI, Tableau, Looker, or Google Data Studio). Understanding of sales and marketing workflows, particularly lead generation and telemarketing processes. Experience in working with CRM platforms and marketing data is a plus Preferred Qualifications: Bachelor’s degree in Data Science, Computer Science, Statistics, or a related field. Familiarity with privacy regulations and best practices for data handling. Previous experience in B2B sales environments is highly desirable. Team HR Vision tech shweta.visiotech03@gmail.com 8368356119 Job Type: Full-time Pay: ₹35,000.00 - ₹40,000.00 per month Benefits: Provident Fund Schedule: Morning shift Supplemental Pay: Performance bonus Work Location: In person

Posted 3 weeks ago

Apply

3.0 - 6.0 years

0 Lacs

Sadar, Uttar Pradesh, India

On-site

Linkedin logo

Profile : Python Developer Experience : 3 To 6 Years Requirement : Expertise in Python Development, AWS, Web Crawling, Databases (MYSQL, SQL SERVER), etc. Location : Work From Office (Work From Office) Working Days : 5 Prefer Immediate Joiners. Job Description Collaborate with development teams to identify and define application requirements. Develop scalable and efficient code using the Python programming language. Conduct thorough testing and debugging to ensure functionality and performance. Design and implement back-end components to enhance system performance. Evaluate and prioritize client feature requests, aligning them with business objectives. Integrate data storage solutions to improve application efficiency. Optimize existing databases to streamline operations and enhance functionality. Develop digital tools for monitoring and analyzing online traffic, providing insights for continuous improvement. Write clean, efficient, and scalable code to support application growth. Perform rigorous testing to ensure program stability and reliability. Enhance existing systems by implementing necessary improvements and updates. Coordinate with internal teams to gather user requirements and deliver tailored technical solutions. Develop and maintain web scrapers using Python libraries such as Scrapy, BeautifulSoup, Selenium, and Requests. Automate data extraction, processing, and storage pipelines to optimize workflow efficiency. Work with SQL databases (MySQL, PostgreSQL, etc.) to ensure optimal data storage and retrieval. Handle API integrations for seamless third-party data collection and automation. Implement proxy management, CAPTCHA bypass techniques, and session handling to ensure uninterrupted data extraction. Ensure compliance with legal and ethical guidelines in web scraping practices. Collaborate with data analysts and developers to seamlessly integrate extracted data into applications. Troubleshoot and resolve issues related to web crawling and automation scripts to maintain smooth operations. (ref:hirist.tech) Show more Show less

Posted 3 weeks ago

Apply

0 years

0 Lacs

Ahmedabad, Gujarat, India

On-site

Linkedin logo

Skills: Python, TensorFlow, PyTorch, Scikit-learn, NLP, Pandas, NumPy, Data Visualization, About Us We are a growing tech company based in Ahmedabad, building AI-powered enterprise applications for the BFSI sector, government bodies, and e-Auction and ProcureToPay domains. Our platforms include eAuction, FinTech Applications, and procurement automationpowered by cutting-edge AI technologies. Requirements Final year student / recent graduate (Computer Science, IT, Data Science, etc.) Solid foundation in Python programming Academic or personal project experience in any ML framework (XGBoost preferred) Familiar with or eager to learn: Vector Databases (FAISS, Weaviate) LangChain / RAG framework Scrapy or web scraping tools Pandas, Numpy, Scikit-learn, Transformers Nice to Have Exposure to OpenAI API, HuggingFace, or LLM-based projects Interest in Finanance / eAuction / Procure To Pay / e-Governance domains Git or version control understanding What We Offer Chance to work upon the Live AI projects Performance-based full-time opportunity after internship A collaborative, growth-focused work culture Show more Show less

Posted 3 weeks ago

Apply

0.0 - 1.0 years

0 Lacs

Mohali, Punjab

On-site

Indeed logo

Python Developer Experience: 1- 3 years Responsibilities: · Develop and maintain web applications using Python frameworks like Django and Flask. · Utilize data analysis libraries such as NumPy, pandas, and matplotlib for extracting insights from large datasets. · Implement machine learning models using TensorFlow and PyTorch for various applications. · Handle HTTP requests efficiently using libraries like Requests. · Perform web scraping tasks using Beautiful Soup and Scrapy. · Collaborate with cross-functional teams to design and implement scalable solutions. · Ensure the security and integrity of data by integrating with database systems like SQL (e.g., PostgreSQL, MySQL) or NoSQL databases (e.g., MongoDB). · Stay updated with the latest trends and technologies in the Python ecosystem. Requirements: · Bachelor's degree in Computer Science, Engineering, or related field. · Strong proficiency in Python programming language. · Experience with Python frameworks such as Django and Flask for web development. · Familiarity with LangChain/LlamaIndex/ML/vectoryDB technologies will be a plus, enhancing team's capabilities in data management and analysis. · Proficiency in deploying Python applications to AWS, Azure, or GCP cloud platforms. · Experience with FastAPI in API development framework, facilitating efficient and robust development of web APIs to support our software ecosystem. · Familiarity with data analysis libraries like NumPy, pandas, and matplotlib. · Knowledge of machine learning frameworks such as TensorFlow and PyTorch. · Experience in handling HTTP requests and web scraping. · Understanding of database systems like SQL (e.g., PostgreSQL, MySQL) or NoSQL databases (e.g., MongoDB). · Excellent problem-solving skills and attention to detail. · Ability to work independently as well as in a team environment. · Strong communication and interpersonal skills. Job Type: Full Time, Permanent, Regular 5 Days Working Location: Plot no.968, Sector-82, JLPL Industrial Area, Mohali, Punjab (140306) Supplement Pay: Yearly Performance Based Bonus, Overtime pays, and others. Benefits: Personality and Training Development, Paid leave, Mensuration leaves, Maternity leaves, Compensatory Off and other special leaves. Job Types: Full-time, Permanent Pay: ₹15,114.49 - ₹40,000.00 per month Schedule: Morning shift Supplemental Pay: Overtime pay Performance bonus Education: Bachelor's (Preferred) Experience: Python: 1 year (Preferred) Work Location: In person

Posted 3 weeks ago

Apply

0.0 - 1.0 years

0 Lacs

Pitampura, Delhi, Delhi

On-site

Indeed logo

Job Title: Data Analyst (Python & Web Scraping Expert) Location : Netaji Subhash Place, Pitampura, New Delhi Department : Data Analytics / Share Recovery Job Overview: We are seeking a detail-oriented and results-driven Data Analyst to join our team. The ideal candidate will have expertise in Python programming, web scraping, and data analysis, with a focus on IEPF share recovery . The role involves collecting, processing, and analyzing data from multiple online sources, providing actionable insights to support business decision-making. Key Responsibilities: Data Scraping : Use Python and web scraping techniques to gather data from financial, regulatory, and shareholding-related websites for IEPF (Investor Education and Protection Fund) share recovery. Data Cleaning & Preprocessing : Clean, process, and structure raw data for analysis. Ensure data quality and integrity by identifying and correcting errors in datasets. Data Analysis & Visualization : Analyze large datasets to extract actionable insights regarding share recovery and trends in investor shareholding. Present findings through visualizations (e.g., graphs, dashboards). Reporting : Prepare and present detailed reports on share recovery patterns, trends, and forecasts based on analysis. Present findings to the management team to help drive business decisions. Automation & Optimization : Build and maintain automated web scraping systems to regularly fetch updated shareholding data, optimizing the data pipeline for efficiency. Collaboration : Work closely with business stakeholders to understand data requirements and deliver reports or visualizations tailored to specific needs related to IEPF share recovery. Required Skills & Qualifications: Technical Skills : Strong proficiency in Python for data analysis and automation. Expertise in web scraping using libraries such as BeautifulSoup , Selenium , and Scrapy . Experience with data manipulation and analysis using Pandas , NumPy , and other relevant libraries. Familiarity with SQL for data extraction and querying relational databases. Knowledge of data visualization tools like Matplotlib , Seaborn , or Tableau for presenting insights in an easy-to-understand format. Experience : Minimum of 2-3 years of experience as a Data Analyst or in a similar role, with a focus on Python programming and web scraping. Experience working with financial or investment data, particularly in areas such as IEPF , share recovery , or investor relations . Strong problem-solving skills with the ability to analyze complex datasets and generate actionable insights. Additional Skills : Strong attention to detail and ability to work with large datasets. Ability to work in a collaborative team environment. Familiarity with cloud platforms (e.g., AWS, Google Cloud) and data storage (e.g., databases, cloud data lakes) is a plus. Education : Bachelor’s or Master’s degree in Data Science , Computer Science , Statistics , Finance , or a related field. Soft Skills : Strong communication skills, with the ability to explain technical concepts to non-technical stakeholders. Ability to prioritize tasks and manage multiple projects simultaneously. Strong organizational skills and time management. Preferred Skills: Experience working in the financial industry or understanding of regulatory frameworks (e.g., IEPF regulations and procedures). Familiarity with machine learning models and predictive analytics for forecasting share recovery trends. Ability to automate workflows and optimize existing data collection pipelines. Job Requirements: Comfortable working in a fast-paced environment. Ability to think critically and provide insights that drive strategic decisions. Must be self-motivated and capable of working independently with minimal supervision. Willingness to stay updated with the latest data analysis techniques and web scraping technologies. Job Type: Full-time Pay: ₹20,000.00 - ₹32,000.00 per month Schedule: Day shift Education: Bachelor's (Preferred) Experience: total work: 1 year (Required) Work Location: In person

Posted 3 weeks ago

Apply

0.0 - 1.0 years

0 Lacs

Jaipur, Rajasthan

On-site

Indeed logo

Experience : 1+ year Location: Jaipur Roles and Responsibilities Expert in Python, with knowledge of at least one Python web framework, such as Django, Flask, etc Develop and maintain service that extracts websites data using scrapers and APIs across multiple websites Understanding of the threading limitations of Python, and multi-process architecture Extract structured / unstructured data Writing reusable, testable, and efficient code Manage testing and bug fixes Knowledge of scraping and scraping frameworks, APIs(integration and creation) and web crawlers. Familiarity with front-end technologies (like JavaScript and HTML5). Performance Goals Sets and achieves realistic, challenging goals, honours commitments and delivers on promises Deal with clients and other employees professionally always Strong team player, able to collaborate effectively with the team to deliver software on time and within budget Ability to work independently Should have the good analytical and problem-solving capability Excellent verbal and written communication Qualifications- Good knowledge of Python, MySQL Programming experience in relational platforms like MySQL. Non-relational platforms like DynamoDB/MongoDB (no-SQL) would be added on Good knowledge of web scraping and APIs Familiarity with some scraping tools like BeautifulSoup and Selenium etc Development Experience with strong skills using the Pandas, Json, Multiprocessing, and NumPy libraries Good troubleshooting and debugging skills Strong interpersonal, verbal and written communication skills in English. Job Type: Full-time Pay: ₹7,000.00 - ₹35,000.00 per month Schedule: Day shift Ability to commute/relocate: Jaipur, Rajasthan: Reliably commute or planning to relocate before starting work (Required) Education: Bachelor's (Required) Experience: Beautiful Soup or Scrapy: 1 year (Required) APIs: 1 year (Preferred) Python: 1 year (Required) Work Location: In person

Posted 3 weeks ago

Apply

0 years

0 Lacs

Ahmedabad, Gujarat, India

On-site

Linkedin logo

About The Role Grade Level (for internal use): 10 The Team : As a member of the Data Transformation - Cognitive Engineering team you will work on building and deploying ML powered products and capabilities to power natural language understanding, data extraction, information retrieval and data sourcing solutions for S&P Global Market Intelligence and our clients. You will spearhead deployment of AI products and pipelines while leading-by-example in a highly engaging work environment. You will work in a (truly) global team and encouraged for thoughtful risk-taking and self-initiative. What’s In It For You Be a part of a global company and build solutions at enterprise scale Lead a highly skilled and technically strong team (including leadership) Contribute to solving high complexity, high impact problems Build production ready pipelines from ideation to deployment Responsibilities Design, Develop and Deploy ML powered products and pipelines Mentor a team of Senior and Junior data scientists / ML Engineers in delivering large scale projects Play a central role in all stages of the AI product development life cycle, including: Designing Machine Learning systems and model scaling strategies Research & Implement ML and Deep learning algorithms for production Run necessary ML tests and benchmarks for model validation Fine-tune, retrain and scale existing model deployments Extend existing ML library’s and write packages for reproducing components Partner with business leaders, domain experts, and end-users to gain business understanding, data understanding, and collect requirements Interpret results and present them to business leaders Manage production pipelines for enterprise scale projects Perform code reviews & optimization for your projects and team Lead and mentor by example, including project scrums Technical Requirements Proven track record as a senior / lead ML engineer Expert proficiency in Python (Numpy, Pandas, Spacy, Sklearn, Pytorch/TF2, HuggingFace etc.) Excellent exposure to large scale model deployment strategies and tools Excellent knowledge of ML & Deep Learning domain Solid exposure to Information Retrieval, Web scraping and Data Extraction at scale Exposure to the following technologies - R-Shiny/Dash/Streamlit, SQL, Docker, Airflow, Redis, Celery, Flask/Django/FastAPI, PySpark, Scrapy Experience with SOTA models related to NLP and expertise in text matching techniques, including sentence transformers, word embeddings, and similarity measures Open to learning new technologies and programming languages as required A Master’s / PhD from a recognized institute in a relevant specialization Good To Have 6-7+ years of relevant experience in ML Engineering Prior substantial experience from the Economics/Financial industry Prior work to show on Github, Kaggle, StackOverflow etc. What’s In It For You? Our Purpose Progress is not a self-starter. It requires a catalyst to be set in motion. Information, imagination, people, technology–the right combination can unlock possibility and change the world. Our world is in transition and getting more complex by the day. We push past expected observations and seek out new levels of understanding so that we can help companies, governments and individuals make an impact on tomorrow. At S&P Global we transform data into Essential Intelligence®, pinpointing risks and opening possibilities. We Accelerate Progress. Our People We're more than 35,000 strong worldwide—so we're able to understand nuances while having a broad perspective. Our team is driven by curiosity and a shared belief that Essential Intelligence can help build a more prosperous future for us all. From finding new ways to measure sustainability to analyzing energy transition across the supply chain to building workflow solutions that make it easy to tap into insight and apply it. We are changing the way people see things and empowering them to make an impact on the world we live in. We’re committed to a more equitable future and to helping our customers find new, sustainable ways of doing business. We’re constantly seeking new solutions that have progress in mind. Join us and help create the critical insights that truly make a difference. Our Values Integrity, Discovery, Partnership At S&P Global, we focus on Powering Global Markets. Throughout our history, the world's leading organizations have relied on us for the Essential Intelligence they need to make confident decisions about the road ahead. We start with a foundation of integrity in all we do, bring a spirit of discovery to our work, and collaborate in close partnership with each other and our customers to achieve shared goals. Benefits We take care of you, so you can take care of business. We care about our people. That’s why we provide everything you—and your career—need to thrive at S&P Global. Our Benefits Include Health & Wellness: Health care coverage designed for the mind and body. Flexible Downtime: Generous time off helps keep you energized for your time on. Continuous Learning: Access a wealth of resources to grow your career and learn valuable new skills. Invest in Your Future: Secure your financial future through competitive pay, retirement planning, a continuing education program with a company-matched student loan contribution, and financial wellness programs. Family Friendly Perks: It’s not just about you. S&P Global has perks for your partners and little ones, too, with some best-in class benefits for families. Beyond the Basics: From retail discounts to referral incentive awards—small perks can make a big difference. For more information on benefits by country visit: https://spgbenefits.com/benefit-summaries Global Hiring And Opportunity At S&P Global At S&P Global, we are committed to fostering a connected and engaged workplace where all individuals have access to opportunities based on their skills, experience, and contributions. Our hiring practices emphasize fairness, transparency, and merit, ensuring that we attract and retain top talent. By valuing different perspectives and promoting a culture of respect and collaboration, we drive innovation and power global markets. Equal Opportunity Employer S&P Global is an equal opportunity employer and all qualified candidates will receive consideration for employment without regard to race/ethnicity, color, religion, sex, sexual orientation, gender identity, national origin, age, disability, marital status, military veteran status, unemployment status, or any other status protected by law. Only electronic job submissions will be considered for employment. If you need an accommodation during the application process due to a disability, please send an email to: EEO.Compliance@spglobal.com and your request will be forwarded to the appropriate person. US Candidates Only: The EEO is the Law Poster http://www.dol.gov/ofccp/regs/compliance/posters/pdf/eeopost.pdf describes discrimination protections under federal law. Pay Transparency Nondiscrimination Provision - https://www.dol.gov/sites/dolgov/files/ofccp/pdf/pay-transp_%20English_formattedESQA508c.pdf IFTECH202.1 - Middle Professional Tier I (EEO Job Group) Job ID: 315679 Posted On: 2025-05-20 Location: Gurgaon, Haryana, India Show more Show less

Posted 4 weeks ago

Apply

6.0 - 7.0 years

0 Lacs

Gurugram, Haryana

On-site

Indeed logo

About the Role: Grade Level (for internal use): 10 The Team : As a member of the Data Transformation - Cognitive Engineering team you will work on building and deploying ML powered products and capabilities to power natural language understanding, data extraction, information retrieval and data sourcing solutions for S&P Global Market Intelligence and our clients. You will spearhead deployment of AI products and pipelines while leading-by-example in a highly engaging work environment. You will work in a (truly) global team and encouraged for thoughtful risk-taking and self-initiative. What’s in it for you: Be a part of a global company and build solutions at enterprise scale Lead a highly skilled and technically strong team (including leadership) Contribute to solving high complexity, high impact problems Build production ready pipelines from ideation to deployment Responsibilities: Design, Develop and Deploy ML powered products and pipelines Mentor a team of Senior and Junior data scientists / ML Engineers in delivering large scale projects Play a central role in all stages of the AI product development life cycle, including: Designing Machine Learning systems and model scaling strategies Research & Implement ML and Deep learning algorithms for production Run necessary ML tests and benchmarks for model validation Fine-tune, retrain and scale existing model deployments Extend existing ML library’s and write packages for reproducing components Partner with business leaders, domain experts, and end-users to gain business understanding, data understanding, and collect requirements Interpret results and present them to business leaders Manage production pipelines for enterprise scale projects Perform code reviews & optimization for your projects and team Lead and mentor by example, including project scrums Technical Requirements: Proven track record as a senior / lead ML engineer Expert proficiency in Python (Numpy, Pandas, Spacy, Sklearn, Pytorch/TF2, HuggingFace etc.) Excellent exposure to large scale model deployment strategies and tools Excellent knowledge of ML & Deep Learning domain Solid exposure to Information Retrieval, Web scraping and Data Extraction at scale Exposure to the following technologies - R-Shiny/Dash/Streamlit, SQL, Docker, Airflow, Redis, Celery, Flask/Django/FastAPI, PySpark, Scrapy Experience with SOTA models related to NLP and expertise in text matching techniques, including sentence transformers, word embeddings, and similarity measures Open to learning new technologies and programming languages as required A Master’s / PhD from a recognized institute in a relevant specialization Good to have: 6-7+ years of relevant experience in ML Engineering Prior substantial experience from the Economics/Financial industry Prior work to show on Github, Kaggle, StackOverflow etc. What’s In It For You? Our Purpose: Progress is not a self-starter. It requires a catalyst to be set in motion. Information, imagination, people, technology–the right combination can unlock possibility and change the world. Our world is in transition and getting more complex by the day. We push past expected observations and seek out new levels of understanding so that we can help companies, governments and individuals make an impact on tomorrow. At S&P Global we transform data into Essential Intelligence®, pinpointing risks and opening possibilities. We Accelerate Progress. Our People: We're more than 35,000 strong worldwide—so we're able to understand nuances while having a broad perspective. Our team is driven by curiosity and a shared belief that Essential Intelligence can help build a more prosperous future for us all. From finding new ways to measure sustainability to analyzing energy transition across the supply chain to building workflow solutions that make it easy to tap into insight and apply it. We are changing the way people see things and empowering them to make an impact on the world we live in. We’re committed to a more equitable future and to helping our customers find new, sustainable ways of doing business. We’re constantly seeking new solutions that have progress in mind. Join us and help create the critical insights that truly make a difference. Our Values: Integrity, Discovery, Partnership At S&P Global, we focus on Powering Global Markets. Throughout our history, the world's leading organizations have relied on us for the Essential Intelligence they need to make confident decisions about the road ahead. We start with a foundation of integrity in all we do, bring a spirit of discovery to our work, and collaborate in close partnership with each other and our customers to achieve shared goals. Benefits: We take care of you, so you can take care of business. We care about our people. That’s why we provide everything you—and your career—need to thrive at S&P Global. Our benefits include: Health & Wellness: Health care coverage designed for the mind and body. Flexible Downtime: Generous time off helps keep you energized for your time on. Continuous Learning: Access a wealth of resources to grow your career and learn valuable new skills. Invest in Your Future: Secure your financial future through competitive pay, retirement planning, a continuing education program with a company-matched student loan contribution, and financial wellness programs. Family Friendly Perks: It’s not just about you. S&P Global has perks for your partners and little ones, too, with some best-in class benefits for families. Beyond the Basics: From retail discounts to referral incentive awards—small perks can make a big difference. For more information on benefits by country visit: https://spgbenefits.com/benefit-summaries Global Hiring and Opportunity at S&P Global: At S&P Global, we are committed to fostering a connected and engaged workplace where all individuals have access to opportunities based on their skills, experience, and contributions. Our hiring practices emphasize fairness, transparency, and merit, ensuring that we attract and retain top talent. By valuing different perspectives and promoting a culture of respect and collaboration, we drive innovation and power global markets. ----------------------------------------------------------- Equal Opportunity Employer S&P Global is an equal opportunity employer and all qualified candidates will receive consideration for employment without regard to race/ethnicity, color, religion, sex, sexual orientation, gender identity, national origin, age, disability, marital status, military veteran status, unemployment status, or any other status protected by law. Only electronic job submissions will be considered for employment. If you need an accommodation during the application process due to a disability, please send an email to: EEO.Compliance@spglobal.com and your request will be forwarded to the appropriate person. US Candidates Only: The EEO is the Law Poster http://www.dol.gov/ofccp/regs/compliance/posters/pdf/eeopost.pdf describes discrimination protections under federal law. Pay Transparency Nondiscrimination Provision - https://www.dol.gov/sites/dolgov/files/ofccp/pdf/pay-transp_%20English_formattedESQA508c.pdf ----------------------------------------------------------- IFTECH202.1 - Middle Professional Tier I (EEO Job Group) Job ID: 315679 Posted On: 2025-05-20 Location: Gurgaon, Haryana, India

Posted 4 weeks ago

Apply

6.0 - 7.0 years

0 Lacs

Gurugram, Haryana

On-site

Indeed logo

Senior ML Engineer Gurgaon, India; Ahmedabad, India; Hyderabad, India; Noida, India Information Technology 315679 Job Description About The Role: Grade Level (for internal use): 10 The Team : As a member of the Data Transformation - Cognitive Engineering team you will work on building and deploying ML powered products and capabilities to power natural language understanding, data extraction, information retrieval and data sourcing solutions for S&P Global Market Intelligence and our clients. You will spearhead deployment of AI products and pipelines while leading-by-example in a highly engaging work environment. You will work in a (truly) global team and encouraged for thoughtful risk-taking and self-initiative. What’s in it for you: Be a part of a global company and build solutions at enterprise scale Lead a highly skilled and technically strong team (including leadership) Contribute to solving high complexity, high impact problems Build production ready pipelines from ideation to deployment Responsibilities: Design, Develop and Deploy ML powered products and pipelines Mentor a team of Senior and Junior data scientists / ML Engineers in delivering large scale projects Play a central role in all stages of the AI product development life cycle, including: Designing Machine Learning systems and model scaling strategies Research & Implement ML and Deep learning algorithms for production Run necessary ML tests and benchmarks for model validation Fine-tune, retrain and scale existing model deployments Extend existing ML library’s and write packages for reproducing components Partner with business leaders, domain experts, and end-users to gain business understanding, data understanding, and collect requirements Interpret results and present them to business leaders Manage production pipelines for enterprise scale projects Perform code reviews & optimization for your projects and team Lead and mentor by example, including project scrums Technical Requirements: Proven track record as a senior / lead ML engineer Expert proficiency in Python (Numpy, Pandas, Spacy, Sklearn, Pytorch/TF2, HuggingFace etc.) Excellent exposure to large scale model deployment strategies and tools Excellent knowledge of ML & Deep Learning domain Solid exposure to Information Retrieval, Web scraping and Data Extraction at scale Exposure to the following technologies - R-Shiny/Dash/Streamlit, SQL, Docker, Airflow, Redis, Celery, Flask/Django/FastAPI, PySpark, Scrapy Experience with SOTA models related to NLP and expertise in text matching techniques, including sentence transformers, word embeddings, and similarity measures Open to learning new technologies and programming languages as required A Master’s / PhD from a recognized institute in a relevant specialization Good to have: 6-7+ years of relevant experience in ML Engineering Prior substantial experience from the Economics/Financial industry Prior work to show on Github, Kaggle, StackOverflow etc. What’s In It For You? Our Purpose: Progress is not a self-starter. It requires a catalyst to be set in motion. Information, imagination, people, technology–the right combination can unlock possibility and change the world. Our world is in transition and getting more complex by the day. We push past expected observations and seek out new levels of understanding so that we can help companies, governments and individuals make an impact on tomorrow. At S&P Global we transform data into Essential Intelligence®, pinpointing risks and opening possibilities. We Accelerate Progress. Our People: We're more than 35,000 strong worldwide—so we're able to understand nuances while having a broad perspective. Our team is driven by curiosity and a shared belief that Essential Intelligence can help build a more prosperous future for us all. From finding new ways to measure sustainability to analyzing energy transition across the supply chain to building workflow solutions that make it easy to tap into insight and apply it. We are changing the way people see things and empowering them to make an impact on the world we live in. We’re committed to a more equitable future and to helping our customers find new, sustainable ways of doing business. We’re constantly seeking new solutions that have progress in mind. Join us and help create the critical insights that truly make a difference. Our Values: Integrity, Discovery, Partnership At S&P Global, we focus on Powering Global Markets. Throughout our history, the world's leading organizations have relied on us for the Essential Intelligence they need to make confident decisions about the road ahead. We start with a foundation of integrity in all we do, bring a spirit of discovery to our work, and collaborate in close partnership with each other and our customers to achieve shared goals. Benefits: We take care of you, so you can take care of business. We care about our people. That’s why we provide everything you—and your career—need to thrive at S&P Global. Our benefits include: Health & Wellness: Health care coverage designed for the mind and body. Flexible Downtime: Generous time off helps keep you energized for your time on. Continuous Learning: Access a wealth of resources to grow your career and learn valuable new skills. Invest in Your Future: Secure your financial future through competitive pay, retirement planning, a continuing education program with a company-matched student loan contribution, and financial wellness programs. Family Friendly Perks: It’s not just about you. S&P Global has perks for your partners and little ones, too, with some best-in class benefits for families. Beyond the Basics: From retail discounts to referral incentive awards—small perks can make a big difference. For more information on benefits by country visit: https://spgbenefits.com/benefit-summaries Global Hiring and Opportunity at S&P Global: At S&P Global, we are committed to fostering a connected and engaged workplace where all individuals have access to opportunities based on their skills, experience, and contributions. Our hiring practices emphasize fairness, transparency, and merit, ensuring that we attract and retain top talent. By valuing different perspectives and promoting a culture of respect and collaboration, we drive innovation and power global markets. - Equal Opportunity Employer S&P Global is an equal opportunity employer and all qualified candidates will receive consideration for employment without regard to race/ethnicity, color, religion, sex, sexual orientation, gender identity, national origin, age, disability, marital status, military veteran status, unemployment status, or any other status protected by law. Only electronic job submissions will be considered for employment. If you need an accommodation during the application process due to a disability, please send an email to: EEO.Compliance@spglobal.com and your request will be forwarded to the appropriate person. US Candidates Only: The EEO is the Law Poster http://www.dol.gov/ofccp/regs/compliance/posters/pdf/eeopost.pdf describes discrimination protections under federal law. Pay Transparency Nondiscrimination Provision - https://www.dol.gov/sites/dolgov/files/ofccp/pdf/pay-transp_%20English_formattedESQA508c.pdf - IFTECH202.1 - Middle Professional Tier I (EEO Job Group) Job ID: 315679 Posted On: 2025-05-20 Location: Gurgaon, Haryana, India

Posted 4 weeks ago

Apply

0 years

0 Lacs

India

Remote

Linkedin logo

Job Title: AI/ML & Automation Intern Location: Remote Duration: 6 Months About us Astraeus Next Gen Pvt. Ltd. is building the next generation of intelligent automation and analytics tools powered by AI/ML. Our mission is to transform how businesses access, analyze, and act on data with zero-click intelligence. Role Overview We are seeking highly motivated interns from Tier 1 institutes (IITs, NITs, IIITs, BITS, etc.) with strong foundations in AI/ML, automation, full-stack development, deployment, and ETL pipelines. As an intern, you will work on live projects that involve developing intelligent automation tools, deploying scalable solutions, and building production-grade web apps and APIs. Key Responsibilities • Design, train, and evaluate ML models for real-world applications • Build scalable ETL pipelines to ingest and process large datasets • Automate web scraping, data collection, and report generation tasks • Develop and deploy backend services (Flask/FastAPI/Django) and frontend components (React.js/Next.js) • Deploy models and APIs on cloud platforms (AWS/GCP/Azure) using Docker, CI/CD pipelines • Collaborate with senior developers and product teams on full lifecycle development Requirements • Currently enrolled in a Tier 1 engineering institute (IIT, NIT, IIIT, BITS, etc.) • Strong programming skills in Python, JavaScript (Node/React), SQL • Hands-on experience with machine learning frameworks (Scikit-learn, TensorFlow, PyTorch) • Knowledge of full-stack development and modern web frameworks • Familiarity with cloud deployment, Docker, and REST APIs • Experience with automation/scraping tools (Selenium, Scrapy, BeautifulSoup) is a plus • Knowledge of MongoDB, Postgres, or other databases • Ability to work independently and in fast-paced, agile teams Good to Have • Exposure to tools like Apache Airflow, n8n, Kafka • Experience in report generation using Pandas, ReportLab, or similar tools • Contributions to open-source or personal projects in relevant domains What You’ll Get • Real-world experience building products. • Opportunity to convert to a full-time role • Work on high-impact projects in AI, automation, and intelligence systems To Apply: Submit your resume, portfolio (if any), and GitHub/LinkedIn profile to [shreejitsen@astraeusnextgen.com] Subject Line: AI/ML & Automation Intern Application - [Your Name] Note: This is an unpaid internship Show more Show less

Posted 4 weeks ago

Apply

3.0 years

0 Lacs

Ahmedabad, Gujarat

Remote

Indeed logo

Job Title: Sr. Python Developer – Web Scraper Location: Ahmedabad, Gujarat Job Type: Full Time Experience: 3+ Years Department: Python About Simform: Simform is a premier digital engineering company specializing in Cloud, Data, AI/ML, and Experience Engineering to create seamless digital experiences and scalable products. Simform is a strong partner for Microsoft, AWS, Google Cloud, and Databricks. With a presence in 5+ countries, Simform primarily serves North America, the UK, and the Northern European market. Simform takes pride in being one of the most reputed employers in the region, having created a thriving work culture with a high work-life balance that gives a sense of freedom and opportunity to grow. Role Overview: The Sr. Python Developer – Web Scraper will be responsible for building scalable, resilient, and intelligent web scraping solutions to extract structured and unstructured data from websites, APIs, and digital platforms. This role involves handling anti-bot mechanisms, dynamic content rendering, and real-time data extraction pipelines, while ensuring high data quality and performance. The ideal candidate will have strong Python skills, deep understanding of scraping techniques, and experience working with headless browsers and cloud platforms. Key Responsibilities: Design and implement scalable web scraping frameworks to collect data from complex and dynamic websites. Develop custom spiders/crawlers using Python libraries like Playwright, Puppeteer, Selenium, Scrapy, or BeautifulSoup. Apply advanced anti-bot evasion strategies such as CAPTCHA solving, IP rotation, user-agent spoofing, browser fingerprinting, and session/cookie management. Automate scraping tasks across distributed systems using tools like Celery, Airflow, cron, and ETL orchestration platforms. Optimize scraper performance for speed, accuracy, and resilience to website structure changes. Implement network interception, DOM traversal, WebSocket handling, and headless browser control. Store and manage scraped data in cloud or local storage using PostgreSQL, MongoDB, or S3. Integrate scraping systems with APIs or microservices for data consumption and downstream workflows. Monitor scraper reliability and handle retry logic, error logging, and dynamic throttling. Write modular, well-documented, and testable Python code with proper unit testing and version control. Collaborate with engineers, data scientists, and stakeholders to understand scraping goals and align them with business needs. Required Skills & Qualifications: Bachelor’s/Master’s degree in Computer Science, Engineering, or a related field. 3+ years of experience in Python development with specialization in web scraping. Deep understanding of modern anti-scraping defenses and bypass techniques (e.g., CAPTCHA, IP bans, dynamic rendering). Proficiency with headless browser tools like Playwright, Puppeteer, or Selenium. Strong grasp of DOM manipulation, JavaScript execution, network inspection, and asynchronous scraping using asyncio, aiohttp, etc. Experience in handling large-scale data extraction and storage using SQL and NoSQL databases. Hands-on experience deploying scrapers and automation workflows on AWS, GCP, or Azure. Familiarity with containerization using Docker and optional experience with Kubernetes. Comfortable with REST API integration, job scheduling, and microservices-based environments. Strong debugging, optimization, and testing skills. Clear understanding of legal and ethical scraping boundaries. Why Join Us: Young Team, Thriving Culture Flat-hierarchical, friendly, engineering-oriented, and growth-focused culture Well-balanced learning and growth opportunities Free health insurance Office facilities with a game zone, in-office kitchen with affordable lunch service, and free snacks Sponsorship for certifications/events and library service Flexible work timing, leaves for life events, WFH, and hybrid options

Posted 1 month ago

Apply

0.0 - 2.0 years

0 Lacs

Mohali, Punjab

On-site

Indeed logo

Key Responsibilities: Build and maintain backend services using Python frameworks like Django , Flask , or FastAPI . Design and consume RESTful APIs to integrate with frontend and external systems. Perform web scraping using: Selenium BeautifulSoup Scrapy or equivalent libraries Contribute to AI/ML-based applications , including model integration, API wrapping, and data handling. Work on data preprocessing, analysis , and pipeline creation for ML applications. Collaborate with cross-functional teams including frontend developers, designers, and PMs. Participate in agile practices: daily standups, sprint planning, and code reviews. Learn and assist in projects involving no-code/low-code platforms . Requirements: Bachelor’s degree in Computer Science, Engineering, or related field. 1-2 years of Python development experience. Strong experience in web scraping using tools like: Selenium , BeautifulSoup , and Scrapy Proficient in backend frameworks: Django , Flask , or FastAPI . Good understanding of REST APIs and ORMs . Experience working with relational databases (PostgreSQL, MySQL). Familiarity with Git and collaborative development workflows. Exposure to AI/ML concepts , model integration, or use of tools like scikit-learn , Pandas , NumPy . Good problem-solving and debugging skills. Nice to Have: Experience deploying ML models or building inference APIs. Familiarity with TensorFlow , PyTorch , or other ML frameworks. Knowledge of Docker , CI/CD , and cloud platforms (AWS, GCP, Azure). Experience with No-Code/Low-Code tools (training will be provided if not experienced). Understanding of frontend basics (React.js or similar) is a plus. Working LOcation : Mohali (Punjab) Experience required - 1-2 Years Job Type: Full-time Pay: Up to ₹30,000.00 per month Benefits: Health insurance Work Location: In person

Posted 1 month ago

Apply

0 years

0 - 0 Lacs

Thiruvananthapuram, Kerala

Work from Office

Indeed logo

Data Science and AI Developer **Job Description:** We are seeking a highly skilled and motivated Data Science and AI Developer to join our dynamic team. As a Data Science and AI Developer, you will be responsible for leveraging cutting-edge technologies to develop innovative solutions that drive business insights and enhance decision-making processes. No salary will be provided during the 3-month probation period. **Key Responsibilities:** 1. Develop and deploy machine learning models for predictive analytics, classification, clustering, and anomaly detection. 2. Design and implement algorithms for data mining, pattern recognition, and natural language processing. 3. Collaborate with cross-functional teams to understand business requirements and translate them into technical solutions. 4. Utilize advanced statistical techniques to analyze complex datasets and extract actionable insights. 5. Implement scalable data pipelines for data ingestion, preprocessing, feature engineering, and model training. 6. Stay updated with the latest advancements in data science, machine learning, and artificial intelligence research. 7. Optimize model performance and scalability through experimentation and iteration. 8. Communicate findings and results to stakeholders through reports, presentations, and visualizations. 9. Ensure compliance with data privacy regulations and best practices in data handling and security. 10. Mentor junior team members and provide technical guidance and support. **Requirements:** 1. Bachelor’s or Master’s degree in Computer Science, Data Science, Statistics, or a related field. 2. Proven experience in developing and deploying machine learning models in production environments. 3. Proficiency in programming languages such as Python, R, or Scala, with strong software engineering skills. 4. Hands-on experience with machine learning libraries/frameworks such as TensorFlow, PyTorch, Scikit-learn, or Spark MLlib. 5. Solid understanding of data structures, algorithms, and computer science fundamentals. 6. Excellent problem-solving skills and the ability to think creatively to overcome challenges. 7. Strong communication and interpersonal skills, with the ability to work effectively in a collaborative team environment. 8. Certification in Data Science, Machine Learning, or Artificial Intelligence (e.g., Coursera, edX, Udacity, etc.). 9. Experience with cloud platforms such as AWS, Azure, or Google Cloud is a plus. 10. Familiarity with big data technologies (e.g., Hadoop, Spark, Kafka) is an advantage. Data Manipulation and Analysis : NumPy, Pandas Data Visualization : Matplotlib, Seaborn, Power BI Machine Learning Libraries : Scikit-learn, TensorFlow, Keras Statistical Analysis : SciPy Web Scrapping : Scrapy IDE : PyCharm, Google Colab HTML/CSS/JavaScript/React JS Proficiency in these core web development technologies is a must. Python Django Expertise: In-depth knowledge of e-commerce functionalities or deep Python Django knowledge. Theming: Proven experience in designing and implementing custom themes for Python websites. Responsive Design: Strong understanding of responsive design principles and the ability to create visually appealing and user-friendly interfaces for various devices. Problem Solving: Excellent problem-solving skills with the ability to troubleshoot and resolve issues independently. Collaboration: Ability to work closely with cross-functional teams, including marketing and design, to bring creative visions to life. interns must know about how to connect front end with datascience Also must Know to connect datascience to frontend **Benefits:** - Competitive salary package - Flexible working hours - Opportunities for career growth and professional development - Dynamic and innovative work environment Job Type: Full-time Pay: ₹8,000.00 - ₹12,000.00 per month Schedule: Day shift Ability to commute/relocate: Thiruvananthapuram, Kerala: Reliably commute or planning to relocate before starting work (Preferred) Work Location: In person

Posted 1 month ago

Apply

7 - 10 years

24 - 31 Lacs

Hyderabad

Work from Office

Naukri logo

Proficient in Python, web scraping (BeautifulSoup, Scrapy, Selenium), HTML/CSS/JS, HTTP, DevTools, IP rotation, proxies, automation, bypassing protections (e.g. Cloudflare), GCP, GraphQL, NoSQL, and Selenium Wire. Skilled in scalable data handling. Provident fund Health insurance

Posted 1 month ago

Apply

0 years

0 Lacs

Sahibzada Ajit Singh Nagar, Punjab, India

Remote

Linkedin logo

COMPANY PROFILE:Company Name: Gmware pvt. ltd. Gmware is an IT company that specializes in serving overseas clients in different verticals. We are looking for a candidate who can join our engineering team and help us develop and maintain various software products. Requirements :Proven 6month+ of relevant experience as a Web Crawling & Scraping EngineerStay updated on the latest trends and technologies in web scraping, crawling, and data processing.Strong problem-solving skillsExcellent communication and collaboration skills.Handle dynamic and complex websites, including those with anti-scraping mechanisms.Troubleshoot and resolve issues related to data extraction, including data integrity and performance challenges.Proficient in popular web scraping technologies and frameworks, such as Selenium, Beautiful soup, Scrapy Contact No. : 80555-03000

Posted 1 month ago

Apply

3 - 5 years

2 - 6 Lacs

Chennai, Pune, Delhi

Work from Office

Naukri logo

Qualifications Bachelor s degree in Computer Science, Software Engineering, Data Science, or a related f ield (or equivalent work experience) Proven experience in web scraping and data extraction techniques Strong proficiency in Python and web scraping libraries (eg, Scrapy, BeautifulSoup, Selenium, Playwright) Hands-on experience with JSON Path and XPath for structured data extraction Knowledge of task/job scheduling tools (eg, Celery, pache Airflow, Cron, Redis Queue) Experience handling large-scale data scraping, proxies, and anti-bot mechanisms Understanding of web technologies such as HTML, CSS, JavaScript, and HTTP protocols Strong problem-solving and debugging skills Ability to work independently and in a collaborative team environment Excellent communication skills, with the ability to explain technical concepts clearly Nice to Have Experience with cloud platforms (AWS, Google Cloud, Azure) for scalable scraping solutions Knowledge of database management (SQL, NoSQL) and data processing pipelines Familiarity with containerization and orchestration tools (Docker, Kubernetes) Experience with CI/CD pipelines for automated deployment of web scraping scripts Contributions to open-source web scraping projects Understanding of AI and machine learning applications in web data extraction

Posted 2 months ago

Apply

4 - 6 years

5 - 6 Lacs

Ahmedabad

Work from Office

Naukri logo

.Python syntax, data structures, Iterator, Generators, Exception Handling, File handling, OOPs, Data Structures, ORM and object-oriented programming concepts. .Proficiency in using web scraping libraries like Selenium WebDriver and Beautiful Soup.

Posted 2 months ago

Apply

0 - 1 years

1 - 3 Lacs

Bengaluru

Work from Office

Naukri logo

Title: Python Web Scraping / Web Crawling Experience Range: 0-1 Years Qualification: BE/MBA Passout Year: 2023, 2024 only Walkin Dates: 2nd, 3rd and 4th April 2025 Timings: 10AM to 4PM Address: Spire Technologies & Solutions Pvt. Ltd., Ajmera Aditya Summit, 2nd Floor, 3D, 7th C Main, 3rd Block, Koramangala, Bangalore 560034. Job Description: Seeking an experienced Web Crawling Engineer to build and optimize scalable data extraction systems. Must have strong expertise in Python scripting, web crawling, data processing, MongoDB, and Power BI. Key Responsibilities: • Develop and maintain crawlers for extracting data from websites, APIs, and complex pages. • Ensure scripts adapt to website changes over time. • Process and transform structured/unstructured data into JSON, CSV formats. • Troubleshoot scraping challenges and optimize performance. • Apply machine learning techniques, particularly time series analysis, to analyze trends and make predictive insights. • Work with MongoDB, Elasticsearch for data storage. • Strong understanding of HTTP protocols, REST APIs, JavaScript rendering, and browser automation. Required Skills: • Python (Scrapy, Selenium, BeautifulSoup, Playwright) • Regex & Shell Scripting for data extraction • NoSQL (MongoDB, Elasticsearch) & API Handling • Data Processing & Visualization (Power BI)

Posted 2 months ago

Apply

3 - 7 years

1 - 2 Lacs

Mumbai

Work from Office

Naukri logo

Key Responsibilities: Develop and maintain automated web scraping scripts using Python libraries such as BeautifulSoup, Scrapy, and Selenium. Optimize scraping pipelines for performance, scalability, and resource efficiency. Handle dynamic websites, CAPTCHA-solving, and implement IP rotation techniques for uninterrupted scraping. Process and clean raw data, ensuring accuracy and integrity in extracted datasets. Collaborate with cross-functional teams to understand data requirements and deliver actionable insights. Leverage APIs when web scraping is not feasible, managing authentication and request optimization. Document processes, pipelines, and troubleshooting steps for maintainable and reusable scraping solutions. Ensure compliance with legal and ethical web scraping practices, implementing security safeguards. Requirements: Education : Bachelors degree in Computer Science, Engineering, or a related field. Experience : 2+ years of Python development experience, with at least 1 year focused on web scraping. Technical Skills : Proficiency in Python and libraries like BeautifulSoup, Scrapy, and Selenium. Experience with regular expressions (Regex) for data parsing. Strong knowledge of HTTP protocols, cookies, headers, and user-agent rotation. Familiarity with databases (SQL and NoSQL) for storing scraped data. Hands-on experience with data manipulation libraries such as pandas and NumPy. Experience working with APIs and managing third-party integrations. Familiarity with version control systems like Git. Bonus Skills : Knowledge of containerization tools like Docker. Experience with distributed scraping solutions and task queues (e.g., Celery, RabbitMQ). Basic understanding of data visualization tools. Non-Technical Skills : Strong analytical and problem-solving skills. Excellent communication and documentation skills. Ability to work independently and collaboratively in a team environment. CANDIDATES AVAILABLE FOR FACE-TO-FACE INTERVIEWS ARE PREFERRED.

Posted 2 months ago

Apply

5 - 7 years

6 - 9 Lacs

Kolkata

Work from Office

Naukri logo

Overview We are seeking a highly experienced Web Scraping Expert (Python) specialising in Scrapy-based web scraping and large-scale data extraction. This role is focused on building and optimizing web crawlers , handling anti-scraping measures , and ensuring efficient data pipelines for structured data collection. The ideal candidate will have 5+ years of hands-on experience developing Scrapy-based scraping solutions , implementing advanced evasion techniques , and managing high-volume web data extraction . You will collaborate with a cross-functional team to design, implement, and optimize scalable scraping systems that deliver high-quality, structured data for critical business needs. Key Responsibilities Scrapy-based Web Scraping Development Develop and maintain scalable web crawlers using Scrapy to extract structured data from diverse sources. Optimize Scrapy spiders for efficiency, reliability, and speed while minimizing detection risks. Handle dynamic content using middlewares, browser-based scraping (Playwright/Selenium), and API integrations. Implement proxy rotation, user-agent switching, and CAPTCHA solving techniques to bypass anti-bot measures. Advanced Anti-Scraping Evasion Techniques Utilize AI-driven approaches to adapt to bot detection and prevent blocks. Implement headless browser automation and request-mimicking strategies to mimic human behavior. Data Processing & Pipeline Management Extract, clean, and structure large-scale web data into structured formats like JSON, CSV, and databases. Optimize Scrapy pipelines for high-speed data processing and storage in MongoDB, PostgreSQL, or cloud storage (AWS S3). Code Quality & Performance Optimization Write clean, well-structured, and maintainable Python code for scraping solutions. Implement automated testing for data accuracy and scraper reliability. Continuously improve crawler efficiency by minimizing IP bans, request delays, and resource consumption. Required Skills and Experience Technical Expertise 5+ years of professional experience in Python development with a focus on web scraping. Proficiency in using Scrapy based scraping Strong understanding of HTML, CSS, JavaScript, and browser behavior. Experience with Docker will be a plus Expertise in handling APIs (RESTful and GraphQL) for data extraction. Proficiency in database systems like MongoDB, PostgreSQL Strong knowledge of version control systems like Git and collaboration platforms like GitHub. Key Attributes Strong problem-solving and analytical skills, with a focus on efficient solutions for complex scraping challenges. Excellent communication skills, both written and verbal. A passion for data and a keen eye for detail. Why Join Us? Work on cutting-edge scraping technologies and AI-driven solutions. Collaborate with a team of talented professionals in a growth-driven environment. Opportunity to influence the development of data-driven business strategies through advanced scraping techniques. Competitive compensation and benefits.

Posted 2 months ago

Apply

0 - 1 years

0 Lacs

Ahmedabad

Work from Office

Naukri logo

Trainee Software Engineer will be a key player in helping us deliver best-in-class schema (Structured data) markup solutions for our clients. You will work closely with SEOs and project managers to understand client needs, design and develop Python scripts for web scraping, data extraction, and data manipulation using Selenium WebDriver and relevant libraries. Your code will be used to automate the process of schema (Structured data) implementation across various websites, ensuring accuracy, efficiency, and scalability Requirements: MCA or BE/BTech in Computer Science. 0-6 months Python experience Basic knowledge of Python programming Hand-on experience on web scrapping Beautiful Soup, Scrapy, Selenium will be an added advantage. Must be familiar with core Python concepts like Iterator, Generators, Exception Handling, File handling, and OOPs Must be familiar with development tools like Jupyter Notebook, IDLE, PyCharm or VS Code Must be familiar with HTML, CSS, JavaScript. Must be enthusiastic and energetic in problem solving. Knowledge of SEO like how to build structured content using LD+JSON based Schema will be an added advantage. Good verbal and written communication skills.

Posted 2 months ago

Apply
cta

Start Your Job Search Today

Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.

Job Application AI Bot

Job Application AI Bot

Apply to 20+ Portals in one click

Download Now

Download the Mobile App

Instantly access job listings, apply easily, and track applications.

Featured Companies