| Index: third_party/gsutil/third_party/boto/tests/integration/s3/test_key.py
|
| diff --git a/third_party/gsutil/third_party/boto/tests/integration/s3/test_key.py b/third_party/gsutil/third_party/boto/tests/integration/s3/test_key.py
|
| new file mode 100644
|
| index 0000000000000000000000000000000000000000..8d426a267b8daa9ded2a22ef7e71d1f7cc5102c6
|
| --- /dev/null
|
| +++ b/third_party/gsutil/third_party/boto/tests/integration/s3/test_key.py
|
| @@ -0,0 +1,534 @@
|
| +# -*- coding: utf-8 -*-
|
| +# Copyright (c) 2012 Mitch Garnaat http://garnaat.org/
|
| +# All rights reserved.
|
| +#
|
| +# Permission is hereby granted, free of charge, to any person obtaining a
|
| +# copy of this software and associated documentation files (the
|
| +# "Software"), to deal in the Software without restriction, including
|
| +# without limitation the rights to use, copy, modify, merge, publish, dis-
|
| +# tribute, sublicense, and/or sell copies of the Software, and to permit
|
| +# persons to whom the Software is furnished to do so, subject to the fol-
|
| +# lowing conditions:
|
| +#
|
| +# The above copyright notice and this permission notice shall be included
|
| +# in all copies or substantial portions of the Software.
|
| +#
|
| +# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS
|
| +# OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABIL-
|
| +# ITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT
|
| +# SHALL THE AUTHOR BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY,
|
| +# WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
|
| +# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
|
| +# IN THE SOFTWARE.
|
| +
|
| +"""
|
| +Some unit tests for S3 Key
|
| +"""
|
| +
|
| +from tests.unit import unittest
|
| +import time
|
| +
|
| +import boto.s3
|
| +from boto.compat import six, StringIO, urllib
|
| +from boto.s3.connection import S3Connection
|
| +from boto.s3.key import Key
|
| +from boto.exception import S3ResponseError
|
| +
|
| +
|
| +class S3KeyTest(unittest.TestCase):
|
| + s3 = True
|
| +
|
| + def setUp(self):
|
| + self.conn = S3Connection()
|
| + self.bucket_name = 'keytest-%d' % int(time.time())
|
| + self.bucket = self.conn.create_bucket(self.bucket_name)
|
| +
|
| + def tearDown(self):
|
| + for key in self.bucket:
|
| + key.delete()
|
| + self.bucket.delete()
|
| +
|
| + def test_set_contents_from_file_dataloss(self):
|
| + # Create an empty stringio and write to it.
|
| + content = "abcde"
|
| + sfp = StringIO()
|
| + sfp.write(content)
|
| + # Try set_contents_from_file() without rewinding sfp
|
| + k = self.bucket.new_key("k")
|
| + try:
|
| + k.set_contents_from_file(sfp)
|
| + self.fail("forgot to rewind so should fail.")
|
| + except AttributeError:
|
| + pass
|
| + # call with rewind and check if we wrote 5 bytes
|
| + k.set_contents_from_file(sfp, rewind=True)
|
| + self.assertEqual(k.size, 5)
|
| + # check actual contents by getting it.
|
| + kn = self.bucket.new_key("k")
|
| + ks = kn.get_contents_as_string().decode('utf-8')
|
| + self.assertEqual(ks, content)
|
| +
|
| + # finally, try with a 0 length string
|
| + sfp = StringIO()
|
| + k = self.bucket.new_key("k")
|
| + k.set_contents_from_file(sfp)
|
| + self.assertEqual(k.size, 0)
|
| + # check actual contents by getting it.
|
| + kn = self.bucket.new_key("k")
|
| + ks = kn.get_contents_as_string().decode('utf-8')
|
| + self.assertEqual(ks, "")
|
| +
|
| + def test_set_contents_as_file(self):
|
| + content="01234567890123456789"
|
| + sfp = StringIO(content)
|
| +
|
| + # fp is set at 0 for just opened (for read) files.
|
| + # set_contents should write full content to key.
|
| + k = self.bucket.new_key("k")
|
| + k.set_contents_from_file(sfp)
|
| + self.assertEqual(k.size, 20)
|
| + kn = self.bucket.new_key("k")
|
| + ks = kn.get_contents_as_string().decode('utf-8')
|
| + self.assertEqual(ks, content)
|
| +
|
| + # set fp to 5 and set contents. this should
|
| + # set "567890123456789" to the key
|
| + sfp.seek(5)
|
| + k = self.bucket.new_key("k")
|
| + k.set_contents_from_file(sfp)
|
| + self.assertEqual(k.size, 15)
|
| + kn = self.bucket.new_key("k")
|
| + ks = kn.get_contents_as_string().decode('utf-8')
|
| + self.assertEqual(ks, content[5:])
|
| +
|
| + # set fp to 5 and only set 5 bytes. this should
|
| + # write the value "56789" to the key.
|
| + sfp.seek(5)
|
| + k = self.bucket.new_key("k")
|
| + k.set_contents_from_file(sfp, size=5)
|
| + self.assertEqual(k.size, 5)
|
| + self.assertEqual(sfp.tell(), 10)
|
| + kn = self.bucket.new_key("k")
|
| + ks = kn.get_contents_as_string().decode('utf-8')
|
| + self.assertEqual(ks, content[5:10])
|
| +
|
| + def test_set_contents_with_md5(self):
|
| + content="01234567890123456789"
|
| + sfp = StringIO(content)
|
| +
|
| + # fp is set at 0 for just opened (for read) files.
|
| + # set_contents should write full content to key.
|
| + k = self.bucket.new_key("k")
|
| + good_md5 = k.compute_md5(sfp)
|
| + k.set_contents_from_file(sfp, md5=good_md5)
|
| + kn = self.bucket.new_key("k")
|
| + ks = kn.get_contents_as_string().decode('utf-8')
|
| + self.assertEqual(ks, content)
|
| +
|
| + # set fp to 5 and only set 5 bytes. this should
|
| + # write the value "56789" to the key.
|
| + sfp.seek(5)
|
| + k = self.bucket.new_key("k")
|
| + good_md5 = k.compute_md5(sfp, size=5)
|
| + k.set_contents_from_file(sfp, size=5, md5=good_md5)
|
| + self.assertEqual(sfp.tell(), 10)
|
| + kn = self.bucket.new_key("k")
|
| + ks = kn.get_contents_as_string().decode('utf-8')
|
| + self.assertEqual(ks, content[5:10])
|
| +
|
| + # let's try a wrong md5 by just altering it.
|
| + k = self.bucket.new_key("k")
|
| + sfp.seek(0)
|
| + hexdig, base64 = k.compute_md5(sfp)
|
| + bad_md5 = (hexdig, base64[3:])
|
| + try:
|
| + k.set_contents_from_file(sfp, md5=bad_md5)
|
| + self.fail("should fail with bad md5")
|
| + except S3ResponseError:
|
| + pass
|
| +
|
| + def test_get_contents_with_md5(self):
|
| + content="01234567890123456789"
|
| + sfp = StringIO(content)
|
| +
|
| + k = self.bucket.new_key("k")
|
| + k.set_contents_from_file(sfp)
|
| + kn = self.bucket.new_key("k")
|
| + s = kn.get_contents_as_string().decode('utf-8')
|
| + self.assertEqual(kn.md5, k.md5)
|
| + self.assertEqual(s, content)
|
| +
|
| + def test_file_callback(self):
|
| + def callback(wrote, total):
|
| + self.my_cb_cnt += 1
|
| + self.assertNotEqual(wrote, self.my_cb_last, "called twice with same value")
|
| + self.my_cb_last = wrote
|
| +
|
| + # Zero bytes written => 1 call
|
| + self.my_cb_cnt = 0
|
| + self.my_cb_last = None
|
| + k = self.bucket.new_key("k")
|
| + k.BufferSize = 2
|
| + sfp = StringIO("")
|
| + k.set_contents_from_file(sfp, cb=callback, num_cb=10)
|
| + self.assertEqual(self.my_cb_cnt, 1)
|
| + self.assertEqual(self.my_cb_last, 0)
|
| + sfp.close()
|
| +
|
| + # Read back zero bytes => 1 call
|
| + self.my_cb_cnt = 0
|
| + self.my_cb_last = None
|
| + s = k.get_contents_as_string(cb=callback)
|
| + self.assertEqual(self.my_cb_cnt, 1)
|
| + self.assertEqual(self.my_cb_last, 0)
|
| +
|
| + content="01234567890123456789"
|
| + sfp = StringIO(content)
|
| +
|
| + # expect 2 calls due start/finish
|
| + self.my_cb_cnt = 0
|
| + self.my_cb_last = None
|
| + k = self.bucket.new_key("k")
|
| + k.set_contents_from_file(sfp, cb=callback, num_cb=10)
|
| + self.assertEqual(self.my_cb_cnt, 2)
|
| + self.assertEqual(self.my_cb_last, 20)
|
| +
|
| + # Read back all bytes => 2 calls
|
| + self.my_cb_cnt = 0
|
| + self.my_cb_last = None
|
| + s = k.get_contents_as_string(cb=callback).decode('utf-8')
|
| + self.assertEqual(self.my_cb_cnt, 2)
|
| + self.assertEqual(self.my_cb_last, 20)
|
| + self.assertEqual(s, content)
|
| +
|
| + # rewind sfp and try upload again. -1 should call
|
| + # for every read/write so that should make 11 when bs=2
|
| + sfp.seek(0)
|
| + self.my_cb_cnt = 0
|
| + self.my_cb_last = None
|
| + k = self.bucket.new_key("k")
|
| + k.BufferSize = 2
|
| + k.set_contents_from_file(sfp, cb=callback, num_cb=-1)
|
| + self.assertEqual(self.my_cb_cnt, 11)
|
| + self.assertEqual(self.my_cb_last, 20)
|
| +
|
| + # Read back all bytes => 11 calls
|
| + self.my_cb_cnt = 0
|
| + self.my_cb_last = None
|
| + s = k.get_contents_as_string(cb=callback, num_cb=-1).decode('utf-8')
|
| + self.assertEqual(self.my_cb_cnt, 11)
|
| + self.assertEqual(self.my_cb_last, 20)
|
| + self.assertEqual(s, content)
|
| +
|
| + # no more than 1 times => 2 times
|
| + # last time always 20 bytes
|
| + sfp.seek(0)
|
| + self.my_cb_cnt = 0
|
| + self.my_cb_last = None
|
| + k = self.bucket.new_key("k")
|
| + k.BufferSize = 2
|
| + k.set_contents_from_file(sfp, cb=callback, num_cb=1)
|
| + self.assertTrue(self.my_cb_cnt <= 2)
|
| + self.assertEqual(self.my_cb_last, 20)
|
| +
|
| + # no more than 1 times => 2 times
|
| + self.my_cb_cnt = 0
|
| + self.my_cb_last = None
|
| + s = k.get_contents_as_string(cb=callback, num_cb=1).decode('utf-8')
|
| + self.assertTrue(self.my_cb_cnt <= 2)
|
| + self.assertEqual(self.my_cb_last, 20)
|
| + self.assertEqual(s, content)
|
| +
|
| + # no more than 2 times
|
| + # last time always 20 bytes
|
| + sfp.seek(0)
|
| + self.my_cb_cnt = 0
|
| + self.my_cb_last = None
|
| + k = self.bucket.new_key("k")
|
| + k.BufferSize = 2
|
| + k.set_contents_from_file(sfp, cb=callback, num_cb=2)
|
| + self.assertTrue(self.my_cb_cnt <= 2)
|
| + self.assertEqual(self.my_cb_last, 20)
|
| +
|
| + # no more than 2 times
|
| + self.my_cb_cnt = 0
|
| + self.my_cb_last = None
|
| + s = k.get_contents_as_string(cb=callback, num_cb=2).decode('utf-8')
|
| + self.assertTrue(self.my_cb_cnt <= 2)
|
| + self.assertEqual(self.my_cb_last, 20)
|
| + self.assertEqual(s, content)
|
| +
|
| + # no more than 3 times
|
| + # last time always 20 bytes
|
| + sfp.seek(0)
|
| + self.my_cb_cnt = 0
|
| + self.my_cb_last = None
|
| + k = self.bucket.new_key("k")
|
| + k.BufferSize = 2
|
| + k.set_contents_from_file(sfp, cb=callback, num_cb=3)
|
| + self.assertTrue(self.my_cb_cnt <= 3)
|
| + self.assertEqual(self.my_cb_last, 20)
|
| +
|
| + # no more than 3 times
|
| + self.my_cb_cnt = 0
|
| + self.my_cb_last = None
|
| + s = k.get_contents_as_string(cb=callback, num_cb=3).decode('utf-8')
|
| + self.assertTrue(self.my_cb_cnt <= 3)
|
| + self.assertEqual(self.my_cb_last, 20)
|
| + self.assertEqual(s, content)
|
| +
|
| + # no more than 4 times
|
| + # last time always 20 bytes
|
| + sfp.seek(0)
|
| + self.my_cb_cnt = 0
|
| + self.my_cb_last = None
|
| + k = self.bucket.new_key("k")
|
| + k.BufferSize = 2
|
| + k.set_contents_from_file(sfp, cb=callback, num_cb=4)
|
| + self.assertTrue(self.my_cb_cnt <= 4)
|
| + self.assertEqual(self.my_cb_last, 20)
|
| +
|
| + # no more than 4 times
|
| + self.my_cb_cnt = 0
|
| + self.my_cb_last = None
|
| + s = k.get_contents_as_string(cb=callback, num_cb=4).decode('utf-8')
|
| + self.assertTrue(self.my_cb_cnt <= 4)
|
| + self.assertEqual(self.my_cb_last, 20)
|
| + self.assertEqual(s, content)
|
| +
|
| + # no more than 6 times
|
| + # last time always 20 bytes
|
| + sfp.seek(0)
|
| + self.my_cb_cnt = 0
|
| + self.my_cb_last = None
|
| + k = self.bucket.new_key("k")
|
| + k.BufferSize = 2
|
| + k.set_contents_from_file(sfp, cb=callback, num_cb=6)
|
| + self.assertTrue(self.my_cb_cnt <= 6)
|
| + self.assertEqual(self.my_cb_last, 20)
|
| +
|
| + # no more than 6 times
|
| + self.my_cb_cnt = 0
|
| + self.my_cb_last = None
|
| + s = k.get_contents_as_string(cb=callback, num_cb=6).decode('utf-8')
|
| + self.assertTrue(self.my_cb_cnt <= 6)
|
| + self.assertEqual(self.my_cb_last, 20)
|
| + self.assertEqual(s, content)
|
| +
|
| + # no more than 10 times
|
| + # last time always 20 bytes
|
| + sfp.seek(0)
|
| + self.my_cb_cnt = 0
|
| + self.my_cb_last = None
|
| + k = self.bucket.new_key("k")
|
| + k.BufferSize = 2
|
| + k.set_contents_from_file(sfp, cb=callback, num_cb=10)
|
| + self.assertTrue(self.my_cb_cnt <= 10)
|
| + self.assertEqual(self.my_cb_last, 20)
|
| +
|
| + # no more than 10 times
|
| + self.my_cb_cnt = 0
|
| + self.my_cb_last = None
|
| + s = k.get_contents_as_string(cb=callback, num_cb=10).decode('utf-8')
|
| + self.assertTrue(self.my_cb_cnt <= 10)
|
| + self.assertEqual(self.my_cb_last, 20)
|
| + self.assertEqual(s, content)
|
| +
|
| + # no more than 1000 times
|
| + # last time always 20 bytes
|
| + sfp.seek(0)
|
| + self.my_cb_cnt = 0
|
| + self.my_cb_last = None
|
| + k = self.bucket.new_key("k")
|
| + k.BufferSize = 2
|
| + k.set_contents_from_file(sfp, cb=callback, num_cb=1000)
|
| + self.assertTrue(self.my_cb_cnt <= 1000)
|
| + self.assertEqual(self.my_cb_last, 20)
|
| +
|
| + # no more than 1000 times
|
| + self.my_cb_cnt = 0
|
| + self.my_cb_last = None
|
| + s = k.get_contents_as_string(cb=callback, num_cb=1000).decode('utf-8')
|
| + self.assertTrue(self.my_cb_cnt <= 1000)
|
| + self.assertEqual(self.my_cb_last, 20)
|
| + self.assertEqual(s, content)
|
| +
|
| + def test_website_redirects(self):
|
| + self.bucket.configure_website('index.html')
|
| + key = self.bucket.new_key('redirect-key')
|
| + self.assertTrue(key.set_redirect('http://www.amazon.com/'))
|
| + self.assertEqual(key.get_redirect(), 'http://www.amazon.com/')
|
| +
|
| + self.assertTrue(key.set_redirect('http://aws.amazon.com/'))
|
| + self.assertEqual(key.get_redirect(), 'http://aws.amazon.com/')
|
| +
|
| + def test_website_redirect_none_configured(self):
|
| + key = self.bucket.new_key('redirect-key')
|
| + key.set_contents_from_string('')
|
| + self.assertEqual(key.get_redirect(), None)
|
| +
|
| + def test_website_redirect_with_bad_value(self):
|
| + self.bucket.configure_website('index.html')
|
| + key = self.bucket.new_key('redirect-key')
|
| + with self.assertRaises(key.provider.storage_response_error):
|
| + # Must start with a / or http
|
| + key.set_redirect('ftp://ftp.example.org')
|
| + with self.assertRaises(key.provider.storage_response_error):
|
| + # Must start with a / or http
|
| + key.set_redirect('')
|
| +
|
| + def test_setting_date(self):
|
| + key = self.bucket.new_key('test_date')
|
| + # This should actually set x-amz-meta-date & not fail miserably.
|
| + key.set_metadata('date', '20130524T155935Z')
|
| + key.set_contents_from_string('Some text here.')
|
| +
|
| + check = self.bucket.get_key('test_date')
|
| + self.assertEqual(check.get_metadata('date'), u'20130524T155935Z')
|
| + self.assertTrue('x-amz-meta-date' in check._get_remote_metadata())
|
| +
|
| + def test_header_casing(self):
|
| + key = self.bucket.new_key('test_header_case')
|
| + # Using anything but CamelCase on ``Content-Type`` or ``Content-MD5``
|
| + # used to cause a signature error (when using ``s3`` for signing).
|
| + key.set_metadata('Content-type', 'application/json')
|
| + key.set_metadata('Content-md5', 'XmUKnus7svY1frWsVskxXg==')
|
| + key.set_contents_from_string('{"abc": 123}')
|
| +
|
| + check = self.bucket.get_key('test_header_case')
|
| + self.assertEqual(check.content_type, 'application/json')
|
| +
|
| + def test_header_encoding(self):
|
| + key = self.bucket.new_key('test_header_encoding')
|
| +
|
| + key.set_metadata('Cache-control', u'public, max-age=500')
|
| + key.set_metadata('Test-Plus', u'A plus (+)')
|
| + key.set_metadata('Content-disposition', u'filename=Schöne Zeit.txt')
|
| + key.set_metadata('Content-Encoding', 'gzip')
|
| + key.set_metadata('Content-Language', 'de')
|
| + key.set_metadata('Content-Type', 'application/pdf')
|
| + self.assertEqual(key.content_type, 'application/pdf')
|
| + key.set_metadata('X-Robots-Tag', 'all')
|
| + key.set_metadata('Expires', u'Thu, 01 Dec 1994 16:00:00 GMT')
|
| + key.set_contents_from_string('foo')
|
| +
|
| + check = self.bucket.get_key('test_header_encoding')
|
| + remote_metadata = check._get_remote_metadata()
|
| +
|
| + # TODO: investigate whether encoding ' ' as '%20' makes sense
|
| + self.assertEqual(check.cache_control, 'public,%20max-age=500')
|
| + self.assertEqual(remote_metadata['cache-control'], 'public,%20max-age=500')
|
| + self.assertEqual(check.get_metadata('test-plus'), 'A plus (+)')
|
| + self.assertEqual(check.content_disposition, 'filename=Sch%C3%B6ne%20Zeit.txt')
|
| + self.assertEqual(remote_metadata['content-disposition'], 'filename=Sch%C3%B6ne%20Zeit.txt')
|
| + self.assertEqual(check.content_encoding, 'gzip')
|
| + self.assertEqual(remote_metadata['content-encoding'], 'gzip')
|
| + self.assertEqual(check.content_language, 'de')
|
| + self.assertEqual(remote_metadata['content-language'], 'de')
|
| + self.assertEqual(check.content_type, 'application/pdf')
|
| + self.assertEqual(remote_metadata['content-type'], 'application/pdf')
|
| + self.assertEqual(check.x_robots_tag, 'all')
|
| + self.assertEqual(remote_metadata['x-robots-tag'], 'all')
|
| + self.assertEqual(check.expires, 'Thu,%2001%20Dec%201994%2016:00:00%20GMT')
|
| + self.assertEqual(remote_metadata['expires'], 'Thu,%2001%20Dec%201994%2016:00:00%20GMT')
|
| +
|
| + expected = u'filename=Schöne Zeit.txt'
|
| + if six.PY2:
|
| + # Newer versions of python default to unicode strings, but python 2
|
| + # requires encoding to UTF-8 to compare the two properly
|
| + expected = expected.encode('utf-8')
|
| +
|
| + self.assertEqual(
|
| + urllib.parse.unquote(check.content_disposition),
|
| + expected
|
| + )
|
| +
|
| + def test_set_contents_with_sse_c(self):
|
| + content="01234567890123456789"
|
| + # the plain text of customer key is "01testKeyToSSEC!"
|
| + header = {
|
| + "x-amz-server-side-encryption-customer-algorithm" :
|
| + "AES256",
|
| + "x-amz-server-side-encryption-customer-key" :
|
| + "MAAxAHQAZQBzAHQASwBlAHkAVABvAFMAUwBFAEMAIQA=",
|
| + "x-amz-server-side-encryption-customer-key-MD5" :
|
| + "fUgCZDDh6bfEMuP2bN38mg=="
|
| + }
|
| + # upload and download content with AWS specified headers
|
| + k = self.bucket.new_key("testkey_for_sse_c")
|
| + k.set_contents_from_string(content, headers=header)
|
| + kn = self.bucket.new_key("testkey_for_sse_c")
|
| + ks = kn.get_contents_as_string(headers=header)
|
| + self.assertEqual(ks, content.encode('utf-8'))
|
| +
|
| +
|
| +class S3KeySigV4Test(unittest.TestCase):
|
| + def setUp(self):
|
| + self.conn = boto.s3.connect_to_region('eu-central-1')
|
| + self.bucket_name = 'boto-sigv4-key-%d' % int(time.time())
|
| + self.bucket = self.conn.create_bucket(self.bucket_name,
|
| + location='eu-central-1')
|
| +
|
| + def tearDown(self):
|
| + for key in self.bucket:
|
| + key.delete()
|
| + self.bucket.delete()
|
| +
|
| + def test_put_get_with_non_string_headers_key(self):
|
| + k = Key(self.bucket)
|
| + k.key = 'foobar'
|
| + body = 'This is a test of S3'
|
| + # A content-length header will be added to this request since it
|
| + # has a body.
|
| + k.set_contents_from_string(body)
|
| + # Set a header that has an integer. This checks for a bug where
|
| + # the sigv4 signer assumes that all of the headers are strings.
|
| + headers = {'Content-Length': 0}
|
| + from_s3_key = self.bucket.get_key('foobar', headers=headers)
|
| + self.assertEqual(from_s3_key.get_contents_as_string().decode('utf-8'),
|
| + body)
|
| +
|
| +
|
| +class S3KeyVersionCopyTest(unittest.TestCase):
|
| + def setUp(self):
|
| + self.conn = S3Connection()
|
| + self.bucket_name = 'boto-key-version-copy-%d' % int(time.time())
|
| + self.bucket = self.conn.create_bucket(self.bucket_name)
|
| + self.bucket.configure_versioning(True)
|
| +
|
| + def tearDown(self):
|
| + for key in self.bucket.list_versions():
|
| + key.delete()
|
| + self.bucket.delete()
|
| +
|
| + def test_key_overwrite_and_copy(self):
|
| + first_content = "abcdefghijklm"
|
| + second_content = "nopqrstuvwxyz"
|
| + k = Key(self.bucket, 'testkey')
|
| + k.set_contents_from_string(first_content)
|
| + # Wait for S3's eventual consistency (may not be necessary)
|
| + while self.bucket.get_key('testkey') is None:
|
| + time.sleep(5)
|
| + # Get the first version_id
|
| + first_key = self.bucket.get_key('testkey')
|
| + first_version_id = first_key.version_id
|
| + # Overwrite the key
|
| + k = Key(self.bucket, 'testkey')
|
| + k.set_contents_from_string(second_content)
|
| + # Wait for eventual consistency
|
| + while True:
|
| + second_key = self.bucket.get_key('testkey')
|
| + if second_key is None or second_key.version_id == first_version_id:
|
| + time.sleep(5)
|
| + else:
|
| + break
|
| + # Copy first key (no longer the current version) to a new key
|
| + source_key = self.bucket.get_key('testkey',
|
| + version_id=first_version_id)
|
| + source_key.copy(self.bucket, 'copiedkey')
|
| + while self.bucket.get_key('copiedkey') is None:
|
| + time.sleep(5)
|
| + copied_key = self.bucket.get_key('copiedkey')
|
| + copied_key_contents = copied_key.get_contents_as_string()
|
| + self.assertEqual(first_content, copied_key_contents)
|
| +
|
|
|