You can find sample data collected with scrape.py from Bentonville, London, Tokyo, Riyadh, San Francisco, Seattle, & 42+ other locations here: gumdropsteve/datasets/airbnb
Files can be downloaded or read directly from GitHub like;
import pandas as pd
pd.read_parquet('https://github.com/gumdropsteve/datasets/raw/master/airbnb/las_vegas.parquet')
git clone https://github.com/gumdropsteve/airbnb.git
cd airbnb
mkdir data
python3 scrape.py
Work in progress [WIP] Jupyter Notebooks from getting started simple in 00_building_base_scrape.ipynb
to parallelizing the worklaod with dask.delayed in 02_building_multi_location_multi_page_base_scrape.ipynb
; examples going from raw to data we care about can be found in /03_cleaning/
.