a year ago
Hello everyone,
When I tried to use the PostgreSQL S3 backups template (https://github.com/railwayapp-templates/postgres-s3-backups)
to upload backups to a bucket that has object lock enabled
(a very important security feature) I got the following error message:
Error while running backup: InvalidRequest: Missing required header for this request: Content-MD5 at throwDefaultError (/app/node_modules/@smithy/smithy-client/dist-cjs/index.js:838:20)
It seems that the content MD5 is needed
if object lock is active. See here https://github.com/aws/aws-sdk-php/issues/1694
How should I calculate the MD5 content here? And does it make sense from your point of view to build this feature directly into the template,
because object lock is a standard security measure for backups.
Thank you in advance for your time
0 Replies
a year ago
so if I understand this correctly, essentially the backup service just needs to provide an md5 hash of the backup file when uploading the file to S3?
Hey Brody,
Yes, it looks like that. Also, the md5 hash apparently has to be encoded with base64 afterwards
a year ago
that part is definitely easy enough
a year ago
md5 hashing a potentially large backup file could end up being quite resource intensive so if it is implemented it would need to be off by default for the buckets that don't require it, at least that's my thinking
Yes, it can be disabled by default.
Apparently you only need it if Object Lock is enabled for the bucket.
a year ago
I'd be happy to take a crack at this, but I don't know if cloudflare R2 has object lock? I don't have access to an aws account to test with s3
I believe wasabi and backblaze have implemented object locking if you have an account there
a year ago
I do have an account with backboaze!
that is nice (:
Because i am trying to use the template with backblaze and got the error
a year ago
I'll see if I can get NodeJS to md5 hash a file efficiently, if node has to load the entire file into memory I don't think that would be worth it to implement
a year ago
no problem, I shall report back
a year ago
got node to md5 hash a large (20gb) file without loading it into memory, and it's not actually too slow, will work on integrating that into the backup service when I'm back at the computer and then I'll do some testing, if that goes well, I would like to ask you to test as well
thats sounds great! Thank you very much!
I'm happy to test when you give me the ping
a year ago
will do
a year ago
hey @Obstkompost am i supposed to be doing something special to enable object locking besides just enabling it? I have enabled it but i dont get any errors running a backup without md5 hashing enabled

a year ago
ah good idea, ill try that now
a year ago
just trying to reproduce your error before i add in the md5 hashing code
a year ago
that was it!

a year ago
indeed! next step, provide an incorrect md5 header
good idea, lets see if backblaze can detect a incorrect md5 header.
My guess is that there is a good chance, that backblaze can not detect that
a year ago
gotta cover all my bases if im doing a pr on a repo many people use!
a year ago
awsome

a year ago
and with the correct hash, it works!

a year ago
maybe someone had even forked it to add it but never pr'd it back
a year ago
could you fork my branch so that you can swap your backup's source with your fork of my branch?
a year ago
oh right, you would need to set a service variable SUPPORT_OBJECT_LOCK to true
a year ago
since md5 hashing a potentially large file isnt exactly free i left it as false by default
a year ago
awsome! is there anything else i should do before i submit a pr?
a year ago
awsome then i will go ahead and submit a pr
a year ago
happy to help!
a year ago
thank you very much, i appreciate that!
