Warning: Don't run these, especially LinkedIn on your personal account use a burner account and have an extra layer of safety like a VPN. If linkdin detects these crapping activities they can ban you for these.
Note: I have added a text file that you might want to run to have all the required libraries called requirements.txt run code: pip install -r requirements.txt to have it all installed.
The data I collect through these codes are just for educational purposes and in no way to be sold or used for any monetary value. For any help, you can refer to the sample data attached as well.
This repository contains three Selenium-based automation scripts that extract structured data from different types of platforms. Each script is designed to work independently and saves the scraped data into CSV files for analysis or tracking.
⚠️ Disclaimer: The file names used here (e.g.,ecommerce_scraper.py
) are purposefully generic to respect platform terms of service. Actual scraping targets include Amazon, LinkedIn, and Glassdoor. Users must use their own credentials for scripts that require login. The project owner cannot be held responsible for any bans that the user can face in case of excessive scraping use at your own risk.
- Make sure you have Python and Selenium installed. Run pip install selenium to install selenium in your teminal
- You must have the Chrome WebDriver installed and added to your system PATH.
- Open the script in VS Code or any IDE of your choice.
- Make sure your browser window is not minimized — Selenium needs it visible.
- Run the script and follow on-screen prompts, if any.
- This script launches Amazon and attempts to scroll the product list.
- Ideal for learning or customizing to extract product details.
- Stores product-related data (when extended) into a .csv file.
- Customize the URL and search query manually inside the script.
- No login is required for Amazon.
Note: Update this line to target a specific page or update the search query as well driver.get("https://www.amazon.com")
This script logs into LinkedIn, navigates to the Jobs section, searches for a given role and location, and scrapes all job listings.
Features:
- Automatically scrolls and scrapes details like company name, role, location, job type, and Easy Apply status.
- Stores the data in linkedin_jobs_1.csv. (which you can edit for your own file name)
- Linkdin is a stable site so it can easily go through hundreds of job listing pages without any issue. Same code can be modified to get specific buissnes company employee or hr data as well.
Notes: Requires a LinkedIn login. Replace the email and password in the script: username.send_keys("your_email_here") password.send_keys("your_password_here")
You can modify the search keywords from:
role_input.send_keys("Analyst")
location_input.send_keys("Bengaluru")
The limit is set to 25 jobs for safety and speed.
This script logs into Glassdoor and scrapes salary data for a given role and location.
Features:
- Scrapes company name, role, salary, rating, median salary, and open job count.
- Navigates through multiple pages and saves all results in Glassdoor_full.csv (which you can edit for your own file name).
- Unlike linkdin glassdoor is not that so stable so you will encounter random 404 or crashes i have a code to tackle this 404, but not so sure about crashes. Requires a Glassdoor login. Replace the credentials in the script:
Note: Change your credentials here: send_keys("your_email_here") send_keys("your_password_here")
Modify job role and location here:
role_input.send_keys("Analyst")
location_input.send_keys("Bengaluru (India)")
Make sure to monitor and control pagination manually if needed.
These codes run well for now in April 2025 might need to tweak some stuffs in the future. Anti-bot detection version linkdin and 404 handler version of these I have you can ask me on linkdin for these at a reasonable price. Thanks