-
Notifications
You must be signed in to change notification settings - Fork 11
Perceptual content identification in IPFS #15
Comments
@denisnazarov will respond more later this weekend, but absolutely. solid ideas. We already have rabin fingerprinting incorporated into file chunking, and will be looking at other fingerprinting techniques too. There's lots of good ones that are content-dependent. In terms of resolving via the DHT, yeah i remember seeing a couple of papers on this. none seemed conclusively better since most diffs are actually tracked by versioning data structures anyway (git, OTs, etc). but i'll read the paper you mentioned as soon as i can, and will try to post the others i had found here. i should say that the IPFS-DHT is built to evolve, so it's is very possible we can incorporate this kind of thing into it. |
(also, for whoever has to pronounce "rabin fingerprinting", it's RAH-bin. not ra-BEEN. i had this mixed up and was recently corrected 2deg from rabin himself). |
@jbenet here's a paper describing using a LHS (Random Hyperplane Hash) for efficient similar content retrieval. The approach we're thinking of is to use a domain-specific perceptual analysis (pHash for images, MFCC for audio, etc) then use RHH to transform it into the hamming distance space as per the paper. |
This issue has been moved to https://discuss.ipfs.io/t/perceptual-content-identification-in-ipfs/366. |
The following is mostly paraphrased from the paper cited below:
Content identification in P2P networks has until now been achieved by using metadata or cryptographic hashes. However, with increasing number of duplicates in different names and formats especially in (unmanaged) P2P networks, these tools have become insufficient for proper content finding. This is especially a problem for digital images, which exist in various formats and compressions as they propagate the web. See https://github.com/mine-code/canonical-content-registry for a thorough identification of the issue.
A possible approach is to identify the content in P2P networks by using perceptual hashes (or fingerprints) extracted from the perceptual features of the content robust to typical processing. The uniform distribution of the extracted fingerprints enables the usage of existing DHT-based keyword search mechanisms for fingerprint queries. In practice, this would allows querying the DHT for canonical metadata related to an image by using perceptual features of an image instance.
Such a system has lots of implications for persistent metadata for digital media, most importantly it enables persistent attribution. Currently, attribution for digital media is explicit and is easily lost as content goes viral on the internet. This is a major disservice to content creators because they are unable to be discovered to reap the benefits of their virality. It forces them to be reliant on centralized distribution platforms such as YouTube, Twitter, Tumblr, Instagram, etc. for identity which monetizes all the content flowing through its pipes with no regard for attribution. Persistent metadata can also enable much more effective aggregation and discovery of knowledge related to digital media.
The paper Content Based Video Identification in Peer-to- Peer Networks: Requirements and a Novel Solution by Koz and Lagendijk proposes a solution using a DHT.
They identify the following:
Their proposed distributed system includes the following:
I am interested in implementing this specifically for images and wonder how an approach like the above can be implemented as a query layer in IPFS that points to metadata, ability to retrieve image instances, etc.
@jbenet @jessewalden @moudy @muneeb-ali @shea256
Related:
stacks-network/stacks-core#79
stacks-network/stacks-core#81
ipfs-inactive/sig-blockchain-data#1
The text was updated successfully, but these errors were encountered: