Bug 1254422 - Do not try to set S3 ACLs r=jlund a=release
MozReview-Commit-ID: 1RyB48KhFvD
--- a/testing/mozharness/scripts/release/beet_mover.py
+++ b/testing/mozharness/scripts/release/beet_mover.py
@@ -272,24 +272,18 @@ class BeetMover(BaseScript, VirtualenvMi
key = boto.s3.key.Key(bucket) # create new key
key.key = s3_key # set key name
self.info("Checking if `{}` already exists".format(s3_key))
key = bucket.get_key(s3_key)
if not key:
self.info("Uploading to `{}`".format(s3_key))
key = bucket.new_key(s3_key)
-
# set key value
self.retry(key.set_contents_from_filename, args=[source], error_level=FATAL),
-
- # key.make_public() may lead to race conditions, because
- # it doesn't pass version_id, so it may not set permissions
- bucket.set_canned_acl(acl_str='public-read', key_name=s3_key,
- version_id=key.version_id)
else:
if not get_hash(key.get_contents_as_string()) == get_hash(open(source).read()):
# for now, let's halt. If necessary, we can revisit this and allow for overwrites
# to the same buildnum release with different bits
self.fatal("`{}` already exists with different checksum.".format(s3_key))
self.log("`{}` has the same MD5 checksum, not uploading".format(s3_key))
def scan_bits(self):