-
-
Notifications
You must be signed in to change notification settings - Fork 128
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Thank you Georgios! #52
Comments
try less cores or you can split your keywords and run synchronously i have shared a script on this closed issue #35 and make sure you are running latest version. |
@lexciobotariu what is the outcome of this? Have you managed to scrape all your keywords? |
Hello there, it did manage to scrap all the information ~500k. |
eh i misunderstood your problem. but your request seems really hard because i dont think google sends static results with requests you do. So able to resume program also need to validate returned data from google. Skipping already scraped data is more performant sure but at the end total request will be same unless only checking exact url and skip the entire results. |
I've been running the script with 5k queries for the last 10h and it got to the level where it is using over 200GB RAM and I've set it to use 35 cores.
It scraped over 300k businesses.
I'm just a bit worried that it won't finish the entire list of queries before crashing due to the lack of RAM.
Any suggestions on how to continue the scraping once it crashes, and to retake from where it left?
The text was updated successfully, but these errors were encountered: