Hello, community,
so I was checking some videos and documents on how dedupe works in Truenas correct me if I am wrong but from what I could understand, Truenas creates a table where he puts each file hash and later on compares new file hash to this table to identify if there are duplicates or not.
Now I don't want to use the dedupe tool because it is really resource-consuming and I have a lot of unique files in my system but I'm still interested in using that hash per file that the function describes to compare maybe if there are dups every month or just compare with external storages to do parity checks.
is there any way to get those hashes directly from Truenas, is that saved in some database I can query?
or do you guys have any suggestions on how should approach this?
the idea is to reduce the time to get that hashes, right now it takes days for a separate machine to calculate each hash one by one because we are talking of hundreds of files in the range of 30 or 300 GB.
so I was checking some videos and documents on how dedupe works in Truenas correct me if I am wrong but from what I could understand, Truenas creates a table where he puts each file hash and later on compares new file hash to this table to identify if there are duplicates or not.
Now I don't want to use the dedupe tool because it is really resource-consuming and I have a lot of unique files in my system but I'm still interested in using that hash per file that the function describes to compare maybe if there are dups every month or just compare with external storages to do parity checks.
is there any way to get those hashes directly from Truenas, is that saved in some database I can query?
or do you guys have any suggestions on how should approach this?
the idea is to reduce the time to get that hashes, right now it takes days for a separate machine to calculate each hash one by one because we are talking of hundreds of files in the range of 30 or 300 GB.