Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

How to implement hash sum to prefent duplicate download/upload #3018

Closed
guettli opened this issue Mar 24, 2015 · 5 comments
Closed

How to implement hash sum to prefent duplicate download/upload #3018

guettli opened this issue Mar 24, 2015 · 5 comments
Assignees

Comments

@guettli
Copy link

guettli commented Mar 24, 2015

AFAIK the client does not use hash sums to check before download/upload.

This produces unneccessary network traffic.

This has been reported before:

#779 (comment)

My question:

How to get this done?

Does the server API provide a way to get hash sums of files?

Does the webDAV API use unique ETags which could be used?

@etiess
Copy link

etiess commented Mar 24, 2015

@guettli I think it is discussed on core for now: owncloud/core#11811

It is an old discussion in OC development: if you have ideas or suggestion, I would do it there ;)

@guettli
Copy link
Author

guettli commented Mar 25, 2015

The propsal at the end of owncloud/core#11811 is about integrity. I want hashes to prevent useless up/downloads. These hashes should be used for ETags, too.

That's why I leave this ticket open.

@etiess Nevertheless, thank you very much for the link to the core ticket.

@ghost
Copy link

ghost commented Mar 25, 2015

Just a short quote from https://owncloud.org/faq/#syncspeed

The downside of this is that it is hard use 'checksums' to see if files are changed: imagine doing that on terabytes of data on an external storage! What is fast for a few gigabyte or a few thousand files is would break entirely once you add a large movie- or music database, or start pumping scientific data through the system.

@guettli
Copy link
Author

guettli commented Mar 25, 2015

Yes, scalability is important. But for me the FAQ item is FUD.

PostgreSQL can handle tables with millions of rows very well. BTree indexes scale very well :-)

I am unsure if php is turing complete and could make it possible to make something like this optional.
Leaving it up to the admin if hashes should be used or not.

@guettli
Copy link
Author

guettli commented Mar 25, 2016

Is this issue solved now, or is it a duplicate of another still open issue?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

4 participants