-
Notifications
You must be signed in to change notification settings - Fork 6
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Guidance on using the tile_size parameter? #7
Comments
As a rule I use a If you expose the scatterplot as a window variable My advice for now is I think the same as before--try populating not from CSV, but from a parquet file where you can be confident of types. It's possible, e.g., that I have some North America-specific character localization choices in the way it does the Arrow CSV parser. I don't know I've every built a quadtile larger than 20 million off of a CSV file, but I've done over 1 billion happily on parquet. |
Thanks for this. I converted to Parquet (using PyArrow). This did seem to improve things somewhat - and worked on the whole dataset! 🙏 But I still think I'm getting wrong behavior during tile rendering/zooming. Rendering still takes ages, and my machine freezes almost completely when rendering. (I'm attaching a screenshot below. Done in mid-render, but still after waiting for at least 5mins or so.) I also logged the inferred extent, as you suggested. Deepscatter claims it's approx. x [ -3.433.116.378, 21.834.822.860 ], y [ -968.008.534, 4.070.507.694 ]. This is way off, but probably matches exactly with the intial viewport I'm seeing. Is there a way to fix the extent during vthe generation of the tileset? |
There shouldn't be major problems rendering 100m points quickly. I'll send you a link to a 1.5b point map that should render fairly smoothly. I suspect the rendering problems and the limits problems are somehow related. If the limits are wrong, it will attempt to draw too many points at a time. Someone else has reported. similar bug in quadfeather this week I'm investigating. Something else that can cause big performance degradation because of the way GPUs work is coincident points--it's possible adding some random noise with the |
Thanks, I'll try |
Are you running with tile_size 1m still? On further thought could conceivably slow things a great deal, because it could lead to many unnecessary points being processed on GPU. |
No, I went back to tilesize 50.000. |
Hi,
do you have any guidance on how to choose the
tile_size
parameter? I've been experimenting with our dataset of geo-coordinates (projected, EPSG 3857, with X/Y coordinate ranges between ~ [-20.000.000, 20.000.000]).This seems to work in deepscatter in principle. But rendering is extremely slow, not like the other online demos I've seen. It also appears to render all our data at the initial zoom level. I then changed the tile_size to 1.000.000 and recreated the tileset. This seemed to display the initial render much faster, but then didn't appear to render a hiqher-quality image after zooming.
Another issue I observed: the inital viewport wasn't automatically adjusted to the coordinate range of the points. Rather, the points were crammed into a tiny corner. Easy to zoom into the correct region, of course. But I'm wondering if there's something I'm doing wrong with my coordinates?
The text was updated successfully, but these errors were encountered: