Skip to content
This repository has been archived by the owner on Feb 27, 2023. It is now read-only.

How to download files from another machine in a cluster? #171

Closed
ghost opened this issue Oct 24, 2018 · 7 comments
Closed

How to download files from another machine in a cluster? #171

ghost opened this issue Oct 24, 2018 · 7 comments
Labels
kind/question all questions or confusion about this project

Comments

@ghost
Copy link

ghost commented Oct 24, 2018

The example given shows how to download a file with a given URL via the supernode. How would I go about downloading a file which exists on the supernode or any other machine on the network?

@pouchrobot pouchrobot added the kind/question all questions or confusion about this project label Oct 24, 2018
@lowzj
Copy link
Member

lowzj commented Oct 25, 2018

We cannot directly download files from machines in dragonfly cluster.
Files that exist in supernode are encoded by dragonfly, we cannot directly to use them.
And files on client machines that downloaded by dfget will be copied to $HOME/.small-dragonfly/data. These files' copies are using for transforming between dfget. They're temporary and will be deleted when no downloading request any more.

@ghost
Copy link
Author

ghost commented Oct 25, 2018

Ok so the use case is downloading files from the internet using P2P to speed it up. I for example, cannot use it for internal file sharing which I'm currently doing via rsync.

@zousheng
Copy link
Contributor

@lowzj rsync sharing is still a common use case, is it possible or meaningful to add this feature into Dragonfly in the future?

@lowzj
Copy link
Member

lowzj commented Oct 25, 2018

We should discuss this feature and find out which scenarios using rsync can be replaced by dragonfly.

@ghost
Copy link
Author

ghost commented Oct 29, 2018

To give you a use case:

I have a machine which downloads large data sets. Now this data has to be distributed to various nodes in the cluster. Not all of the data is distributed to all of the nodes.

For example I download huge data sets A, B, C & D.
Node 1 & 3 need A, B & D.
Node 2, & 4 need C & D.
In this case, I'm currently using rsync on each node to get data from the machine which downloaded all the data sets. I'd like to replace this with P2P sharing as there's an overlap of the data required by the different nodes. Obviously this is a very small example of the overall use case. I hope this helps clarify the use-case

@allencloud
Copy link
Contributor

allencloud commented Jan 7, 2019

First of all, thanks sincerely for constantly using and supporting Dragonfly. We will try our best to keep Dragonfly better, and keep community and eco-system growing.

To get more feedback of Dragonfly's adoption in industry, would you mind helping comment in issue #219 to give more information about your enterprise's Dragonfly usage. Thanks a lot in advance. @asadsm

@ghost
Copy link
Author

ghost commented Jan 7, 2019

Sure, will do.

@ghost ghost closed this as completed Jan 7, 2019
This issue was closed.
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
kind/question all questions or confusion about this project
Projects
None yet
Development

No branches or pull requests

4 participants