A simplified app version of the script is also available for users who prefer a more user-friendly experience.
Website Subpage Scraper is a Python script designed to scrape all internal links on a webpage. The script prompts the user for a URL, sends a GET request to retrieve the HTML content, and then parses the HTML using the BeautifulSoup module. The script filters out external links and duplicates, and prompts the user to select an output mode, which can either print the list of links to the terminal or write them to a file. In case the required modules (requests and beautifulsoup4) are not already installed, the script will install them automatically.
To run this script, you must have Python installed on your computer. Open a terminal and navigate to the directory where the script is saved. Then, enter the command "python website-subpage-scraper.py". The script will prompt you for a URL to scrape (make sure to include the http/https part of the URL), and then ask whether you want to print the links to the terminal or save them to a file. The output will be displayed on the terminal or saved to a file named "output.txt".
Please note that this tool should only be used on websites that you have permission to access. While it may not work on certain larger websites, it should work well on smaller sites such as blogs. The creator of this tool accepts no responsibility for any trouble you may encounter while using it.