| Index: tools/telemetry/third_party/gsutilz/third_party/apitools/samples/storage_sample/uploads_test.py
|
| diff --git a/tools/telemetry/third_party/gsutilz/third_party/apitools/samples/storage_sample/uploads_test.py b/tools/telemetry/third_party/gsutilz/third_party/apitools/samples/storage_sample/uploads_test.py
|
| deleted file mode 100644
|
| index ad4416f60748a789723ae36b37006400f7035e8d..0000000000000000000000000000000000000000
|
| --- a/tools/telemetry/third_party/gsutilz/third_party/apitools/samples/storage_sample/uploads_test.py
|
| +++ /dev/null
|
| @@ -1,146 +0,0 @@
|
| -"""Integration tests for uploading and downloading to GCS.
|
| -
|
| -These tests exercise most of the corner cases for upload/download of
|
| -files in apitools, via GCS. There are no performance tests here yet.
|
| -"""
|
| -
|
| -import json
|
| -import os
|
| -import random
|
| -import string
|
| -import unittest
|
| -
|
| -import six
|
| -
|
| -import apitools.base.py as apitools_base
|
| -import storage
|
| -
|
| -_CLIENT = None
|
| -
|
| -
|
| -def _GetClient():
|
| - global _CLIENT # pylint: disable=global-statement
|
| - if _CLIENT is None:
|
| - _CLIENT = storage.StorageV1()
|
| - return _CLIENT
|
| -
|
| -
|
| -class UploadsTest(unittest.TestCase):
|
| - _DEFAULT_BUCKET = 'apitools'
|
| - _TESTDATA_PREFIX = 'uploads'
|
| -
|
| - def setUp(self):
|
| - self.__client = _GetClient()
|
| - self.__files = []
|
| - self.__content = ''
|
| - self.__buffer = None
|
| - self.__upload = None
|
| -
|
| - def tearDown(self):
|
| - self.__DeleteFiles()
|
| -
|
| - def __ResetUpload(self, size, auto_transfer=True):
|
| - self.__content = ''.join(
|
| - random.choice(string.ascii_letters) for _ in range(size))
|
| - self.__buffer = six.StringIO(self.__content)
|
| - self.__upload = storage.Upload.FromStream(
|
| - self.__buffer, 'text/plain', auto_transfer=auto_transfer)
|
| -
|
| - def __DeleteFiles(self):
|
| - for filename in self.__files:
|
| - self.__DeleteFile(filename)
|
| -
|
| - def __DeleteFile(self, filename):
|
| - object_name = os.path.join(self._TESTDATA_PREFIX, filename)
|
| - req = storage.StorageObjectsDeleteRequest(
|
| - bucket=self._DEFAULT_BUCKET, object=object_name)
|
| - self.__client.objects.Delete(req)
|
| -
|
| - def __InsertRequest(self, filename):
|
| - object_name = os.path.join(self._TESTDATA_PREFIX, filename)
|
| - return storage.StorageObjectsInsertRequest(
|
| - name=object_name, bucket=self._DEFAULT_BUCKET)
|
| -
|
| - def __GetRequest(self, filename):
|
| - object_name = os.path.join(self._TESTDATA_PREFIX, filename)
|
| - return storage.StorageObjectsGetRequest(
|
| - object=object_name, bucket=self._DEFAULT_BUCKET)
|
| -
|
| - def __InsertFile(self, filename, request=None):
|
| - if request is None:
|
| - request = self.__InsertRequest(filename)
|
| - response = self.__client.objects.Insert(request, upload=self.__upload)
|
| - self.assertIsNotNone(response)
|
| - self.__files.append(filename)
|
| - return response
|
| -
|
| - def testZeroBytes(self):
|
| - filename = 'zero_byte_file'
|
| - self.__ResetUpload(0)
|
| - response = self.__InsertFile(filename)
|
| - self.assertEqual(0, response.size)
|
| -
|
| - def testSimpleUpload(self):
|
| - filename = 'fifteen_byte_file'
|
| - self.__ResetUpload(15)
|
| - response = self.__InsertFile(filename)
|
| - self.assertEqual(15, response.size)
|
| -
|
| - def testMultipartUpload(self):
|
| - filename = 'fifteen_byte_file'
|
| - self.__ResetUpload(15)
|
| - request = self.__InsertRequest(filename)
|
| - request.object = storage.Object(contentLanguage='en')
|
| - response = self.__InsertFile(filename, request=request)
|
| - self.assertEqual(15, response.size)
|
| - self.assertEqual('en', response.contentLanguage)
|
| -
|
| - def testAutoUpload(self):
|
| - filename = 'ten_meg_file'
|
| - size = 10 << 20
|
| - self.__ResetUpload(size)
|
| - request = self.__InsertRequest(filename)
|
| - response = self.__InsertFile(filename, request=request)
|
| - self.assertEqual(size, response.size)
|
| -
|
| - def testBreakAndResumeUpload(self):
|
| - filename = ('ten_meg_file_' +
|
| - ''.join(random.sample(string.ascii_letters, 5)))
|
| - size = 10 << 20
|
| - self.__ResetUpload(size, auto_transfer=False)
|
| - self.__upload.strategy = 'resumable'
|
| - self.__upload.total_size = size
|
| - # Start the upload
|
| - request = self.__InsertRequest(filename)
|
| - initial_response = self.__client.objects.Insert(
|
| - request, upload=self.__upload)
|
| - self.assertIsNotNone(initial_response)
|
| - self.assertEqual(0, self.__buffer.tell())
|
| - # Pretend the process died, and resume with a new attempt at the
|
| - # same upload.
|
| - upload_data = json.dumps(self.__upload.serialization_data)
|
| - second_upload_attempt = apitools_base.Upload.FromData(
|
| - self.__buffer, upload_data, self.__upload.http)
|
| - second_upload_attempt._Upload__SendChunk(0)
|
| - self.assertEqual(second_upload_attempt.chunksize, self.__buffer.tell())
|
| - # Simulate a third try, and stream from there.
|
| - final_upload_attempt = apitools_base.Upload.FromData(
|
| - self.__buffer, upload_data, self.__upload.http)
|
| - final_upload_attempt.StreamInChunks()
|
| - self.assertEqual(size, self.__buffer.tell())
|
| - # Verify the upload
|
| - object_info = self.__client.objects.Get(self.__GetRequest(filename))
|
| - self.assertEqual(size, object_info.size)
|
| - # Confirm that a new attempt successfully does nothing.
|
| - completed_upload_attempt = apitools_base.Upload.FromData(
|
| - self.__buffer, upload_data, self.__upload.http)
|
| - self.assertTrue(completed_upload_attempt.complete)
|
| - completed_upload_attempt.StreamInChunks()
|
| - # Verify the upload didn't pick up extra bytes.
|
| - object_info = self.__client.objects.Get(self.__GetRequest(filename))
|
| - self.assertEqual(size, object_info.size)
|
| - # TODO(craigcitro): Add tests for callbacks (especially around
|
| - # finish callback).
|
| -
|
| -if __name__ == '__main__':
|
| - unittest.main()
|
|
|