This is a python3 program that works as a Wikipedia crawler and utilizes BeautifulSoup and the requests modules. It goes through the entire Wikipedia page of the user entered URL and stores the information available on the wikipedia page as a .txt file on the system locally.
- Install
Python 3.7
or above. - Install these modules:
- requests
- bs4
- Download the
Crawler.py
file. - Open the folder containing the aforementioned file in your terminal (for MacOS and Linux) or command prompt (for Windows).
- Type
python Crawler.py
and press enter to run the program.
OR
- Install any IDE.
- Create a new project, copy the
Crawler.py
code and paste it in a.py
file. - Run the program.
Pull requests are welcome for adding more feautures or fixing exisiting issues.