| Index: third_party/gsutil/gslib/tests/test_naming.py
|
| diff --git a/third_party/gsutil/gslib/tests/test_naming.py b/third_party/gsutil/gslib/tests/test_naming.py
|
| new file mode 100644
|
| index 0000000000000000000000000000000000000000..b4866084fc857eb76012bab8911473e64a91f111
|
| --- /dev/null
|
| +++ b/third_party/gsutil/gslib/tests/test_naming.py
|
| @@ -0,0 +1,1157 @@
|
| +# -*- coding: utf-8 -*-
|
| +# Copyright 2010 Google Inc. All Rights Reserved.
|
| +#
|
| +# Permission is hereby granted, free of charge, to any person obtaining a
|
| +# copy of this software and associated documentation files (the
|
| +# "Software"), to deal in the Software without restriction, including
|
| +# without limitation the rights to use, copy, modify, merge, publish, dis-
|
| +# tribute, sublicense, and/or sell copies of the Software, and to permit
|
| +# persons to whom the Software is furnished to do so, subject to the fol-
|
| +# lowing conditions:
|
| +#
|
| +# The above copyright notice and this permission notice shall be included
|
| +# in all copies or substantial portions of the Software.
|
| +#
|
| +# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS
|
| +# OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABIL-
|
| +# ITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT
|
| +# SHALL THE AUTHOR BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY,
|
| +# WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
|
| +# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
|
| +# IN THE SOFTWARE.
|
| +"""Tests for gsutil naming logic.
|
| +
|
| +The test code in this file runs against an in-memory storage service mock,
|
| +so runs very quickly. This is valuable for testing changes that impact the
|
| +naming rules, since those rules are complex and it's useful to be able to
|
| +make small incremental changes and rerun the tests frequently. Additional
|
| +end-to-end tests (which send traffic to the production Google Cloud Storage
|
| +service) are available via the gsutil test command.
|
| +"""
|
| +
|
| +from __future__ import absolute_import
|
| +
|
| +import gzip
|
| +import os
|
| +import StringIO
|
| +
|
| +from gslib import copy_helper
|
| +from gslib.cloud_api import NotFoundException
|
| +from gslib.cloud_api import ServiceException
|
| +from gslib.exception import CommandException
|
| +from gslib.exception import HashMismatchException
|
| +from gslib.storage_url import StorageUrlFromString
|
| +import gslib.tests.testcase as testcase
|
| +from gslib.tests.util import ObjectToURI as suri
|
| +from gslib.tests.util import SetBotoConfigForTest
|
| +from gslib.util import UTF8
|
| +
|
| +
|
| +def _Overwrite(fp):
|
| + """Overwrite first byte in an open file and flush contents."""
|
| + fp.seek(0)
|
| + fp.write('x')
|
| + fp.flush()
|
| +
|
| +
|
| +def _Append(fp):
|
| + """Append a byte at end of an open file and flush contents."""
|
| + fp.seek(0, 2)
|
| + fp.write('x')
|
| + fp.flush()
|
| +
|
| +
|
| +# TODO: Re-enable PerformsFileToObjectUpload decorator on tests in this file
|
| +# once we refactor to a thread-safe mock storage service implementation.
|
| +class GsutilNamingTests(testcase.GsUtilUnitTestCase):
|
| + """Unit tests for gsutil naming logic."""
|
| +
|
| + def testGetPathBeforeFinalDir(self):
|
| + """Tests GetPathBeforeFinalDir() (unit test)."""
|
| + self.assertEqual(
|
| + 'gs://', copy_helper.GetPathBeforeFinalDir(StorageUrlFromString(
|
| + 'gs://bucket/')))
|
| + self.assertEqual(
|
| + 'gs://bucket', copy_helper.GetPathBeforeFinalDir(StorageUrlFromString(
|
| + 'gs://bucket/dir/')))
|
| + self.assertEqual(
|
| + 'gs://bucket', copy_helper.GetPathBeforeFinalDir(StorageUrlFromString(
|
| + 'gs://bucket/dir')))
|
| + self.assertEqual(
|
| + 'gs://bucket/dir', copy_helper.GetPathBeforeFinalDir(
|
| + StorageUrlFromString('gs://bucket/dir/obj')))
|
| + src_dir = self.CreateTempDir()
|
| + subdir = os.path.join(src_dir, 'subdir')
|
| + os.mkdir(subdir)
|
| + self.assertEqual(suri(src_dir),
|
| + copy_helper.GetPathBeforeFinalDir(
|
| + StorageUrlFromString(suri(subdir))))
|
| +
|
| + # @PerformsFileToObjectUpload
|
| + def testCopyingTopLevelFileToBucket(self):
|
| + """Tests copying one top-level file to a bucket."""
|
| + src_file = self.CreateTempFile(file_name='f0')
|
| + dst_bucket_uri = self.CreateBucket()
|
| + self.RunCommand('cp', [src_file, suri(dst_bucket_uri)])
|
| + actual = list(self._test_wildcard_iterator(
|
| + suri(dst_bucket_uri, '**')).IterAll(expand_top_level_buckets=True))
|
| + self.assertEqual(1, len(actual))
|
| + self.assertEqual('f0', actual[0].root_object.name)
|
| +
|
| + # @PerformsFileToObjectUpload
|
| + def testCopyingMultipleFilesToBucket(self):
|
| + """Tests copying multiple files to a bucket."""
|
| + src_file0 = self.CreateTempFile(file_name='f0')
|
| + src_file1 = self.CreateTempFile(file_name='f1')
|
| + dst_bucket_uri = self.CreateBucket()
|
| + self.RunCommand('cp', [src_file0, src_file1, suri(dst_bucket_uri)])
|
| + actual = set(str(u) for u in self._test_wildcard_iterator(
|
| + suri(dst_bucket_uri, '**')).IterAll(expand_top_level_buckets=True))
|
| + expected = set([
|
| + suri(dst_bucket_uri, 'f0'),
|
| + suri(dst_bucket_uri, 'f1'),
|
| + ])
|
| + self.assertEqual(expected, actual)
|
| +
|
| + # @PerformsFileToObjectUpload
|
| + def testCopyingNestedFileToBucketSubdir(self):
|
| + """Tests copying a nested file to a bucket subdir.
|
| +
|
| + Tests that we correctly translate local FS-specific delimiters ('\' on
|
| + Windows) to bucket delimiter (/).
|
| + """
|
| + tmpdir = self.CreateTempDir()
|
| + subdir = os.path.join(tmpdir, 'subdir')
|
| + os.mkdir(subdir)
|
| + src_file = self.CreateTempFile(tmpdir=tmpdir, file_name='obj', contents='')
|
| + dst_bucket_uri = self.CreateBucket()
|
| + # Make an object under subdir so next copy will treat subdir as a subdir.
|
| + self.RunCommand('cp', [src_file, suri(dst_bucket_uri, 'subdir/a')])
|
| + self.RunCommand('cp', [src_file, suri(dst_bucket_uri, 'subdir')])
|
| + actual = set(str(u) for u in self._test_wildcard_iterator(
|
| + suri(dst_bucket_uri, '**')).IterObjects())
|
| + expected = set([
|
| + suri(dst_bucket_uri, 'subdir', 'a'),
|
| + suri(dst_bucket_uri, 'subdir', 'obj'),
|
| + ])
|
| + self.assertEqual(expected, actual)
|
| +
|
| + # @PerformsFileToObjectUpload
|
| + def testCopyingAbsolutePathDirToBucket(self):
|
| + """Tests recursively copying absolute path directory to a bucket."""
|
| + dst_bucket_uri = self.CreateBucket()
|
| + src_dir_root = self.CreateTempDir(test_files=[
|
| + 'f0', 'f1', 'f2.txt', ('dir0', 'dir1', 'nested')])
|
| + self.RunCommand('cp', ['-R', src_dir_root, suri(dst_bucket_uri)])
|
| + actual = set(str(u) for u in self._test_wildcard_iterator(
|
| + suri(dst_bucket_uri, '**')).IterAll(expand_top_level_buckets=True))
|
| + src_tmpdir = os.path.split(src_dir_root)[1]
|
| + expected = set([
|
| + suri(dst_bucket_uri, src_tmpdir, 'f0'),
|
| + suri(dst_bucket_uri, src_tmpdir, 'f1'),
|
| + suri(dst_bucket_uri, src_tmpdir, 'f2.txt'),
|
| + suri(dst_bucket_uri, src_tmpdir, 'dir0', 'dir1', 'nested')])
|
| + self.assertEqual(expected, actual)
|
| +
|
| + # @PerformsFileToObjectUpload
|
| + def testCopyingRelativePathDirToBucket(self):
|
| + """Tests recursively copying relative directory to a bucket."""
|
| + dst_bucket_uri = self.CreateBucket()
|
| + src_dir = self.CreateTempDir(test_files=[('dir0', 'f1')])
|
| + self.RunCommand('cp', ['-R', 'dir0', suri(dst_bucket_uri)], cwd=src_dir)
|
| + actual = set(str(u) for u in self._test_wildcard_iterator(
|
| + suri(dst_bucket_uri, '**')).IterAll(expand_top_level_buckets=True))
|
| + expected = set([suri(dst_bucket_uri, 'dir0', 'f1')])
|
| + self.assertEqual(expected, actual)
|
| +
|
| + # @PerformsFileToObjectUpload
|
| + def testCopyingRelPathSubDirToBucketSubdirWithDollarFolderObj(self):
|
| + """Tests recursively copying relative sub-directory to bucket subdir.
|
| +
|
| + Subdir is signified by a $folder$ object.
|
| + """
|
| + # Create a $folder$ object to simulate a folder created by GCS manager (or
|
| + # various other tools), which gsutil understands to mean there is a folder
|
| + # into which the object is being copied.
|
| + dst_bucket_uri = self.CreateBucket()
|
| + self.CreateObject(bucket_uri=dst_bucket_uri, object_name='abc_$folder$',
|
| + contents='')
|
| + src_dir = self.CreateTempDir(test_files=[('dir0', 'dir1', 'f1')])
|
| + self.RunCommand('cp', ['-R', os.path.join('dir0', 'dir1'),
|
| + suri(dst_bucket_uri, 'abc')], cwd=src_dir)
|
| + actual = set(str(u) for u in self._test_wildcard_iterator(
|
| + suri(dst_bucket_uri, '**')).IterAll(expand_top_level_buckets=True))
|
| + expected = set([suri(dst_bucket_uri, 'abc_$folder$'),
|
| + suri(dst_bucket_uri, 'abc', 'dir1', 'f1')])
|
| + self.assertEqual(expected, actual)
|
| +
|
| + # @PerformsFileToObjectUpload
|
| + def testCopyingRelativePathSubDirToBucketSubdirSignifiedBySlash(self):
|
| + """Tests recursively copying relative sub-directory to bucket subdir.
|
| +
|
| + Subdir is signified by a / object.
|
| + """
|
| + dst_bucket_uri = self.CreateBucket()
|
| + src_dir = self.CreateTempDir(test_files=[('dir0', 'dir1', 'f1')])
|
| + self.RunCommand('cp', ['-R', os.path.join('dir0', 'dir1'),
|
| + suri(dst_bucket_uri, 'abc') + '/'], cwd=src_dir)
|
| + actual = set(str(u) for u in self._test_wildcard_iterator(
|
| + suri(dst_bucket_uri, '**')).IterAll(expand_top_level_buckets=True))
|
| + expected = set([suri(dst_bucket_uri, 'abc', 'dir1', 'f1')])
|
| + self.assertEqual(expected, actual)
|
| +
|
| + # @PerformsFileToObjectUpload
|
| + def testCopyingRelativePathSubDirToBucket(self):
|
| + """Tests recursively copying relative sub-directory to a bucket."""
|
| + dst_bucket_uri = self.CreateBucket()
|
| + src_dir = self.CreateTempDir(test_files=[('dir0', 'dir1', 'f1')])
|
| + self.RunCommand('cp', ['-R', os.path.join('dir0', 'dir1'),
|
| + suri(dst_bucket_uri)], cwd=src_dir)
|
| + actual = set(str(u) for u in self._test_wildcard_iterator(
|
| + suri(dst_bucket_uri, '**')).IterAll(expand_top_level_buckets=True))
|
| + expected = set([suri(dst_bucket_uri, 'dir1', 'f1')])
|
| + self.assertEqual(expected, actual)
|
| +
|
| + # @PerformsFileToObjectUpload
|
| + def testCopyingDotSlashToBucket(self):
|
| + """Tests copying ./ to a bucket produces expected naming."""
|
| + # When running a command like gsutil cp -r . gs://dest we expect the dest
|
| + # obj names to be of the form gs://dest/abc, not gs://dest/./abc.
|
| + dst_bucket_uri = self.CreateBucket()
|
| + src_dir = self.CreateTempDir(test_files=['foo'])
|
| + for rel_src_dir in ['.', '.%s' % os.sep]:
|
| + self.RunCommand('cp', ['-R', rel_src_dir, suri(dst_bucket_uri)],
|
| + cwd=src_dir)
|
| + actual = set(str(u) for u in self._test_wildcard_iterator(
|
| + suri(dst_bucket_uri, '**')).IterAll(expand_top_level_buckets=True))
|
| + expected = set([suri(dst_bucket_uri, 'foo')])
|
| + self.assertEqual(expected, actual)
|
| +
|
| + # @PerformsFileToObjectUpload
|
| + def testCopyingDirContainingOneFileToBucket(self):
|
| + """Tests copying a directory containing 1 file to a bucket.
|
| +
|
| + We test this case to ensure that correct bucket handling isn't dependent
|
| + on the copy being treated as a multi-source copy.
|
| + """
|
| + dst_bucket_uri = self.CreateBucket()
|
| + src_dir = self.CreateTempDir(test_files=[('dir0', 'dir1', 'foo')])
|
| + self.RunCommand('cp', ['-R', os.path.join(src_dir, 'dir0', 'dir1'),
|
| + suri(dst_bucket_uri)])
|
| + actual = list((str(u) for u in self._test_wildcard_iterator(
|
| + suri(dst_bucket_uri, '**')).IterAll(expand_top_level_buckets=True)))
|
| + self.assertEqual(1, len(actual))
|
| + self.assertEqual(suri(dst_bucket_uri, 'dir1', 'foo'), actual[0])
|
| +
|
| + def testCopyingBucketToDir(self):
|
| + """Tests copying from a bucket to a directory."""
|
| + src_bucket_uri = self.CreateBucket(test_objects=['foo', 'dir/foo2'])
|
| + dst_dir = self.CreateTempDir()
|
| + # Mock objects don't support hash digestion.
|
| + with SetBotoConfigForTest([('GSUtil', 'check_hashes', 'never')]):
|
| + self.RunCommand('cp', ['-R', suri(src_bucket_uri), dst_dir])
|
| + actual = set(str(u) for u in self._test_wildcard_iterator(
|
| + '%s%s**' % (dst_dir, os.sep)).IterAll(expand_top_level_buckets=True))
|
| + expected = set([suri(dst_dir, src_bucket_uri.bucket_name, 'foo'),
|
| + suri(dst_dir, src_bucket_uri.bucket_name, 'dir', 'foo2')])
|
| + self.assertEqual(expected, actual)
|
| +
|
| + def testCopyingBucketToBucket(self):
|
| + """Tests copying from a bucket-only URI to a bucket."""
|
| + src_bucket_uri = self.CreateBucket(test_objects=['foo', 'dir/foo2'])
|
| + dst_bucket_uri = self.CreateBucket()
|
| + self.RunCommand('cp', ['-R', suri(src_bucket_uri), suri(dst_bucket_uri)])
|
| + actual = set(str(u) for u in self._test_wildcard_iterator(
|
| + suri(dst_bucket_uri, '**')).IterAll(expand_top_level_buckets=True))
|
| + expected = set([
|
| + suri(dst_bucket_uri, src_bucket_uri.bucket_name, 'foo'),
|
| + suri(dst_bucket_uri, src_bucket_uri.bucket_name, 'dir', 'foo2')])
|
| + self.assertEqual(expected, actual)
|
| +
|
| + def testCopyingDirectoryToDirectory(self):
|
| + """Tests copying from a directory to a directory."""
|
| + src_dir = self.CreateTempDir(test_files=['foo', ('dir', 'foo2')])
|
| + dst_dir = self.CreateTempDir()
|
| + self.RunCommand('cp', ['-R', src_dir, dst_dir])
|
| + actual = set(str(u) for u in self._test_wildcard_iterator(
|
| + '%s%s**' % (dst_dir, os.sep)).IterAll(expand_top_level_buckets=True))
|
| + src_dir_base = os.path.split(src_dir)[1]
|
| + expected = set([suri(dst_dir, src_dir_base, 'foo'),
|
| + suri(dst_dir, src_dir_base, 'dir', 'foo2')])
|
| + self.assertEqual(expected, actual)
|
| +
|
| + def testCopyingFilesAndDirNonRecursive(self):
|
| + """Tests copying containing files and a directory without -R."""
|
| + src_dir = self.CreateTempDir(test_files=['foo', 'bar', ('d1', 'f2'),
|
| + ('d2', 'f3'), ('d3', 'd4', 'f4')])
|
| + dst_dir = self.CreateTempDir()
|
| + self.RunCommand('cp', ['%s%s*' % (src_dir, os.sep), dst_dir])
|
| + actual = set(str(u) for u in self._test_wildcard_iterator(
|
| + '%s%s**' % (dst_dir, os.sep)).IterAll(expand_top_level_buckets=True))
|
| + expected = set([suri(dst_dir, 'foo'), suri(dst_dir, 'bar')])
|
| + self.assertEqual(expected, actual)
|
| +
|
| + def testCopyingFileToDir(self):
|
| + """Tests copying one file to a directory."""
|
| + src_file = self.CreateTempFile(file_name='foo')
|
| + dst_dir = self.CreateTempDir()
|
| + self.RunCommand('cp', [src_file, dst_dir])
|
| + actual = list(self._test_wildcard_iterator(
|
| + '%s%s*' % (dst_dir, os.sep)).IterAll(expand_top_level_buckets=True))
|
| + self.assertEqual(1, len(actual))
|
| + self.assertEqual(suri(dst_dir, 'foo'), str(actual[0]))
|
| +
|
| + # @PerformsFileToObjectUpload
|
| + def testCopyingFileToObjectWithConsecutiveSlashes(self):
|
| + """Tests copying a file to an object containing consecutive slashes."""
|
| + src_file = self.CreateTempFile(file_name='f0')
|
| + dst_bucket_uri = self.CreateBucket()
|
| + self.RunCommand('cp', [src_file, suri(dst_bucket_uri) + '//obj'])
|
| + actual = list(self._test_wildcard_iterator(
|
| + suri(dst_bucket_uri, '**')).IterAll(expand_top_level_buckets=True))
|
| + self.assertEqual(1, len(actual))
|
| + self.assertEqual('/obj', actual[0].root_object.name)
|
| +
|
| + def testCopyingCompressedFileToBucket(self):
|
| + """Tests copying one file with compression to a bucket."""
|
| + src_file = self.CreateTempFile(contents='plaintext', file_name='f2.txt')
|
| + dst_bucket_uri = self.CreateBucket()
|
| + self.RunCommand('cp', ['-z', 'txt', src_file, suri(dst_bucket_uri)],)
|
| + actual = list(self._test_wildcard_iterator(
|
| + suri(dst_bucket_uri, '*')).IterAll(expand_top_level_buckets=True))
|
| + self.assertEqual(1, len(actual))
|
| + actual_obj = actual[0].root_object
|
| + self.assertEqual('f2.txt', actual_obj.name)
|
| + self.assertEqual('gzip', actual_obj.contentEncoding)
|
| +
|
| + stdout = self.RunCommand('cat', [suri(dst_bucket_uri, 'f2.txt')],
|
| + return_stdout=True)
|
| +
|
| + f = gzip.GzipFile(fileobj=StringIO.StringIO(stdout), mode='rb')
|
| + try:
|
| + self.assertEqual(f.read(), 'plaintext')
|
| + finally:
|
| + f.close()
|
| +
|
| + def testCopyingObjectToObject(self):
|
| + """Tests copying an object to an object."""
|
| + src_bucket_uri = self.CreateBucket(test_objects=['obj'])
|
| + dst_bucket_uri = self.CreateBucket()
|
| + self.RunCommand('cp', [suri(src_bucket_uri, 'obj'), suri(dst_bucket_uri)])
|
| + actual = list(self._test_wildcard_iterator(
|
| + suri(dst_bucket_uri, '*')).IterAll(expand_top_level_buckets=True))
|
| + self.assertEqual(1, len(actual))
|
| + self.assertEqual('obj', actual[0].root_object.name)
|
| +
|
| + def testCopyingObjectToObjectUsingDestWildcard(self):
|
| + """Tests copying an object to an object using a dest wildcard."""
|
| + src_bucket_uri = self.CreateBucket(test_objects=['obj'])
|
| + dst_bucket_uri = self.CreateBucket(test_objects=['dstobj'])
|
| + self.RunCommand('cp', [suri(src_bucket_uri, 'obj'),
|
| + '%s*' % dst_bucket_uri.uri])
|
| + actual = set(str(u) for u in self._test_wildcard_iterator(
|
| + suri(dst_bucket_uri, '*')).IterAll(expand_top_level_buckets=True))
|
| + expected = set([suri(dst_bucket_uri, 'dstobj')])
|
| + self.assertEqual(expected, actual)
|
| +
|
| + def testCopyingObjsAndFilesToDir(self):
|
| + """Tests copying objects and files to a directory."""
|
| + src_bucket_uri = self.CreateBucket(test_objects=['f1'])
|
| + src_dir = self.CreateTempDir(test_files=['f2'])
|
| + dst_dir = self.CreateTempDir()
|
| + # Mock objects don't support hash digestion.
|
| + with SetBotoConfigForTest([('GSUtil', 'check_hashes', 'never')]):
|
| + self.RunCommand('cp', ['-R', suri(src_bucket_uri, '**'),
|
| + os.path.join(src_dir, '**'), dst_dir])
|
| + actual = set(str(u) for u in self._test_wildcard_iterator(
|
| + os.path.join(dst_dir, '**')).IterAll(expand_top_level_buckets=True))
|
| + expected = set([suri(dst_dir, 'f1'), suri(dst_dir, 'f2')])
|
| + self.assertEqual(expected, actual)
|
| +
|
| + def testCopyingObjToDot(self):
|
| + """Tests that copying an object to . or ./ downloads to correct name."""
|
| + src_bucket_uri = self.CreateBucket(test_objects=['f1'])
|
| + dst_dir = self.CreateTempDir()
|
| + for final_char in ('/', ''):
|
| + # Mock objects don't support hash digestion.
|
| + with SetBotoConfigForTest([('GSUtil', 'check_hashes', 'never')]):
|
| + self.RunCommand('cp', [suri(src_bucket_uri, 'f1'), '.%s' % final_char],
|
| + cwd=dst_dir)
|
| + actual = set()
|
| + for dirname, dirnames, filenames in os.walk(dst_dir):
|
| + for subdirname in dirnames:
|
| + actual.add(os.path.join(dirname, subdirname))
|
| + for filename in filenames:
|
| + actual.add(os.path.join(dirname, filename))
|
| + expected = set([os.path.join(dst_dir, 'f1')])
|
| + self.assertEqual(expected, actual)
|
| +
|
| + # @PerformsFileToObjectUpload
|
| + def testCopyingObjsAndFilesToBucket(self):
|
| + """Tests copying objects and files to a bucket."""
|
| + src_bucket_uri = self.CreateBucket(test_objects=['f1'])
|
| + src_dir = self.CreateTempDir(test_files=['f2'])
|
| + dst_bucket_uri = self.CreateBucket()
|
| + self.RunCommand('cp', ['-R', suri(src_bucket_uri, '**'),
|
| + '%s%s**' % (src_dir, os.sep), suri(dst_bucket_uri)])
|
| + actual = set(str(u) for u in self._test_wildcard_iterator(
|
| + suri(dst_bucket_uri, '**')).IterAll(expand_top_level_buckets=True))
|
| + expected = set([suri(dst_bucket_uri, 'f1'), suri(dst_bucket_uri, 'f2')])
|
| + self.assertEqual(expected, actual)
|
| +
|
| + # @PerformsFileToObjectUpload
|
| + def testCopyingSubdirRecursiveToNonexistentSubdir(self):
|
| + """Tests copying a directory with a single file recursively to a bucket.
|
| +
|
| + The file should end up in a new bucket subdirectory with the file's
|
| + directory structure starting below the recursive copy point, as in Unix cp.
|
| +
|
| + Example:
|
| + filepath: dir1/dir2/foo
|
| + cp -r dir1 dir3
|
| + Results in dir3/dir2/foo being created.
|
| + """
|
| + src_dir = self.CreateTempDir()
|
| + self.CreateTempFile(tmpdir=src_dir + '/dir1/dir2', file_name='foo')
|
| + dst_bucket_uri = self.CreateBucket()
|
| + self.RunCommand('cp', ['-R', src_dir + '/dir1',
|
| + suri(dst_bucket_uri, 'dir3')])
|
| + actual = set(str(u) for u in self._test_wildcard_iterator(
|
| + suri(dst_bucket_uri, '**')).IterAll(expand_top_level_buckets=True))
|
| + expected = set([suri(dst_bucket_uri, 'dir3/dir2/foo')])
|
| + self.assertEqual(expected, actual)
|
| +
|
| + def testAttemptDirCopyWithoutRecursion(self):
|
| + """Tests copying a directory without -R."""
|
| + src_dir = self.CreateTempDir(test_files=1)
|
| + dst_dir = self.CreateTempDir()
|
| + try:
|
| + self.RunCommand('cp', [src_dir, dst_dir])
|
| + self.fail('Did not get expected CommandException')
|
| + except CommandException, e:
|
| + self.assertIn('No URLs matched', e.reason)
|
| +
|
| + def testNonRecursiveFileAndSameNameSubdir(self):
|
| + """Tests copying a file and subdirectory of the same name without -R."""
|
| + src_bucket_uri = self.CreateBucket(test_objects=['f1', 'f1/f2'])
|
| + dst_dir = self.CreateTempDir()
|
| + # Mock objects don't support hash digestion.
|
| + with SetBotoConfigForTest([('GSUtil', 'check_hashes', 'never')]):
|
| + self.RunCommand('cp', [suri(src_bucket_uri, 'f1'), dst_dir])
|
| + actual = list(self._test_wildcard_iterator(
|
| + '%s%s*' % (dst_dir, os.sep)).IterAll(expand_top_level_buckets=True))
|
| + self.assertEqual(1, len(actual))
|
| + self.assertEqual(suri(dst_dir, 'f1'), str(actual[0]))
|
| + # TODO: Assert that we omit the prefix here when unit_testcase supports
|
| + # returning stderr.
|
| +
|
| + def testAttemptCopyingProviderOnlySrc(self):
|
| + """Attempts to copy a src specified as a provider-only URI."""
|
| + src_bucket_uri = self.CreateBucket()
|
| + try:
|
| + self.RunCommand('cp', ['gs://', suri(src_bucket_uri)])
|
| + self.fail('Did not get expected CommandException')
|
| + except CommandException, e:
|
| + self.assertIn('provider-only', e.reason)
|
| +
|
| + def testAttemptCopyingOverlappingSrcDstFile(self):
|
| + """Attempts to an object atop itself."""
|
| + src_file = self.CreateTempFile()
|
| + try:
|
| + self.RunCommand('cp', [src_file, src_file])
|
| + self.fail('Did not get expected CommandException')
|
| + except CommandException, e:
|
| + self.assertIn('are the same file - abort', e.reason)
|
| +
|
| + def testAttemptCopyingToMultiMatchWildcard(self):
|
| + """Attempts to copy where dst wildcard matches >1 obj."""
|
| + src_bucket_uri = self.CreateBucket(test_objects=2)
|
| + try:
|
| + self.RunCommand('cp', [suri(src_bucket_uri, 'obj0'),
|
| + suri(src_bucket_uri, '*')])
|
| + self.fail('Did not get expected CommandException')
|
| + except CommandException, e:
|
| + self.assertNotEqual(e.reason.find('must match exactly 1 URL'), -1)
|
| +
|
| + def testAttemptCopyingMultiObjsToFile(self):
|
| + """Attempts to copy multiple objects to a file."""
|
| + src_bucket_uri = self.CreateBucket(test_objects=2)
|
| + dst_file = self.CreateTempFile()
|
| + try:
|
| + self.RunCommand('cp', ['-R', suri(src_bucket_uri, '*'), dst_file])
|
| + self.fail('Did not get expected CommandException')
|
| + except CommandException, e:
|
| + self.assertIn('must name a directory, bucket, or', e.reason)
|
| +
|
| + def testAttemptCopyingWithFileDirConflict(self):
|
| + """Attempts to copy objects that cause a file/directory conflict."""
|
| + # Create objects with name conflicts (a/b and a). Use 'dst' bucket because
|
| + # it gets cleared after each test.
|
| + bucket_uri = self.CreateBucket()
|
| + self.CreateObject(bucket_uri=bucket_uri, object_name='a')
|
| + self.CreateObject(bucket_uri=bucket_uri, object_name='b/a')
|
| + dst_dir = self.CreateTempDir()
|
| + try:
|
| + self.RunCommand('cp', ['-R', suri(bucket_uri), dst_dir])
|
| + self.fail('Did not get expected CommandException')
|
| + except CommandException, e:
|
| + self.assertNotEqual('exists where a directory needs to be created',
|
| + e.reason)
|
| +
|
| + def testAttemptCopyingWithDirFileConflict(self):
|
| + """Attempts to copy an object that causes a directory/file conflict."""
|
| + # Create an object that conflicts with dest subdir.
|
| + tmpdir = self.CreateTempDir()
|
| + os.mkdir(os.path.join(tmpdir, 'abc'))
|
| + src_uri = self.CreateObject(object_name='abc', contents='bar')
|
| + try:
|
| + self.RunCommand('cp', [suri(src_uri), tmpdir + '/'])
|
| + self.fail('Did not get expected CommandException')
|
| + except CommandException, e:
|
| + self.assertNotEqual('where the file needs to be created', e.reason)
|
| +
|
| + def testWildcardMoveWithinBucket(self):
|
| + """Attempts to move using src wildcard that overlaps dest object.
|
| +
|
| + We want to ensure that this doesn't stomp the result data.
|
| + """
|
| + dst_bucket_uri = self.CreateBucket(test_objects=['old'])
|
| + self.RunCommand('mv', [suri(dst_bucket_uri, 'old*'),
|
| + suri(dst_bucket_uri, 'new')])
|
| + actual = set(str(u) for u in self._test_wildcard_iterator(
|
| + suri(dst_bucket_uri, '**')).IterAll(expand_top_level_buckets=True))
|
| + expected = set([suri(dst_bucket_uri, 'new')])
|
| + self.assertEqual(expected, actual)
|
| +
|
| + def testLsNonExistentObjectWithPrefixName(self):
|
| + """Test ls of non-existent obj that matches prefix of existing objs."""
|
| + # Use an object name that matches a prefix of other names at that level, to
|
| + # ensure the ls subdir handling logic doesn't pick up anything extra.
|
| + src_bucket_uri = self.CreateBucket(test_objects=['obj_with_suffix'])
|
| + try:
|
| + self.RunCommand('ls', [suri(src_bucket_uri, 'obj')])
|
| + except CommandException, e:
|
| + self.assertIn('matched no objects', e.reason)
|
| +
|
| + def testLsBucketNonRecursive(self):
|
| + """Test that ls of a bucket returns expected results."""
|
| + src_bucket_uri = self.CreateBucket(test_objects=['foo1', 'd0/foo2',
|
| + 'd1/d2/foo3'])
|
| + output = self.RunCommand('ls', [suri(src_bucket_uri, '*')],
|
| + return_stdout=True)
|
| + expected = set([suri(src_bucket_uri, 'foo1'),
|
| + suri(src_bucket_uri, 'd1', ':'),
|
| + suri(src_bucket_uri, 'd1', 'd2') + src_bucket_uri.delim,
|
| + suri(src_bucket_uri, 'd0', ':'),
|
| + suri(src_bucket_uri, 'd0', 'foo2')])
|
| + expected.add('') # Blank line between subdir listings.
|
| + actual = set(output.split('\n'))
|
| + self.assertEqual(expected, actual)
|
| +
|
| + def testLsBucketRecursive(self):
|
| + """Test that ls -R of a bucket returns expected results."""
|
| + src_bucket_uri = self.CreateBucket(test_objects=['foo1', 'd0/foo2',
|
| + 'd1/d2/foo3'])
|
| + output = self.RunCommand('ls', ['-R', suri(src_bucket_uri, '*')],
|
| + return_stdout=True)
|
| + expected = set([suri(src_bucket_uri, 'foo1'),
|
| + suri(src_bucket_uri, 'd1', ':'),
|
| + suri(src_bucket_uri, 'd1', 'd2', ':'),
|
| + suri(src_bucket_uri, 'd1', 'd2', 'foo3'),
|
| + suri(src_bucket_uri, 'd0', ':'),
|
| + suri(src_bucket_uri, 'd0', 'foo2')])
|
| + expected.add('') # Blank line between subdir listings.
|
| + actual = set(output.split('\n'))
|
| + self.assertEqual(expected, actual)
|
| +
|
| + def testLsBucketRecursiveWithLeadingSlashObjectName(self):
|
| + """Test that ls -R of a bucket with an object that has leading slash."""
|
| + dst_bucket_uri = self.CreateBucket(test_objects=['f0'])
|
| + output = self.RunCommand('ls', ['-R', suri(dst_bucket_uri) + '*'],
|
| + return_stdout=True)
|
| + expected = set([suri(dst_bucket_uri, 'f0')])
|
| + expected.add('') # Blank line between subdir listings.
|
| + actual = set(output.split('\n'))
|
| + self.assertEqual(expected, actual)
|
| +
|
| + def testLsBucketSubdirNonRecursive(self):
|
| + """Test that ls of a bucket subdir returns expected results."""
|
| + src_bucket_uri = self.CreateBucket(test_objects=['src_subdir/foo',
|
| + 'src_subdir/nested/foo2'])
|
| + output = self.RunCommand('ls', [suri(src_bucket_uri, 'src_subdir')],
|
| + return_stdout=True)
|
| + expected = set([
|
| + suri(src_bucket_uri, 'src_subdir', 'foo'),
|
| + suri(src_bucket_uri, 'src_subdir', 'nested') + src_bucket_uri.delim])
|
| + expected.add('') # Blank line between subdir listings.
|
| + actual = set(output.split('\n'))
|
| + self.assertEqual(expected, actual)
|
| +
|
| + def testLsBucketSubdirRecursive(self):
|
| + """Test that ls -R of a bucket subdir returns expected results."""
|
| + src_bucket_uri = self.CreateBucket(test_objects=['src_subdir/foo',
|
| + 'src_subdir/nested/foo2'])
|
| + for final_char in ('/', ''):
|
| + output = self.RunCommand(
|
| + 'ls', ['-R', suri(src_bucket_uri, 'src_subdir') + final_char],
|
| + return_stdout=True)
|
| + expected = set([
|
| + suri(src_bucket_uri, 'src_subdir', ':'),
|
| + suri(src_bucket_uri, 'src_subdir', 'foo'),
|
| + suri(src_bucket_uri, 'src_subdir', 'nested', ':'),
|
| + suri(src_bucket_uri, 'src_subdir', 'nested', 'foo2')])
|
| + expected.add('') # Blank line between subdir listings.
|
| + actual = set(output.split('\n'))
|
| + self.assertEqual(expected, actual)
|
| +
|
| + def testSetAclOnBucketRuns(self):
|
| + """Test that the 'acl set' command basically runs."""
|
| + # We don't test reading back the acl (via 'acl get' command) because at
|
| + # present MockStorageService doesn't translate canned ACLs into actual ACL
|
| + # XML.
|
| + src_bucket_uri = self.CreateBucket()
|
| + self.RunCommand('acl', ['set', 'private', suri(src_bucket_uri)])
|
| +
|
| + def testSetAclOnWildcardNamedBucketRuns(self):
|
| + """Test that 'acl set' basically runs against wildcard-named bucket."""
|
| + # We don't test reading back the acl (via 'acl get' command) because at
|
| + # present MockStorageService doesn't translate canned ACLs into actual ACL
|
| + # XML.
|
| + src_bucket_uri = self.CreateBucket(test_objects=['f0'])
|
| + self.RunCommand('acl', ['set', 'private', suri(src_bucket_uri)[:-2] + '*'])
|
| +
|
| + def testSetAclOnObjectRuns(self):
|
| + """Test that the 'acl set' command basically runs."""
|
| + src_bucket_uri = self.CreateBucket(test_objects=['f0'])
|
| + self.RunCommand('acl', ['set', 'private', suri(src_bucket_uri, '*')])
|
| +
|
| + def testSetDefAclOnBucketRuns(self):
|
| + """Test that the 'defacl set' command basically runs."""
|
| + src_bucket_uri = self.CreateBucket()
|
| + self.RunCommand('defacl', ['set', 'private', suri(src_bucket_uri)])
|
| +
|
| + def testSetDefAclOnObjectFails(self):
|
| + """Test that the 'defacl set' command fails when run against an object."""
|
| + src_bucket_uri = self.CreateBucket()
|
| + try:
|
| + self.RunCommand('defacl', ['set', 'private', suri(src_bucket_uri, '*')])
|
| + self.fail('Did not get expected CommandException')
|
| + except CommandException, e:
|
| + self.assertIn('URL must name a bucket', e.reason)
|
| +
|
| + # @PerformsFileToObjectUpload
|
| + def testMinusDOptionWorks(self):
|
| + """Tests using gsutil -D option."""
|
| + src_file = self.CreateTempFile(file_name='f0')
|
| + dst_bucket_uri = self.CreateBucket()
|
| + self.RunCommand('cp', [src_file, suri(dst_bucket_uri)], debug=3)
|
| + actual = list(self._test_wildcard_iterator(
|
| + suri(dst_bucket_uri, '*')).IterAll(expand_top_level_buckets=True))
|
| + self.assertEqual(1, len(actual))
|
| + self.assertEqual('f0', actual[0].root_object.name)
|
| +
|
| + def DownloadTestHelper(self, func):
|
| + """Test resumable download with custom test function.
|
| +
|
| + The custom function distorts downloaded data. We expect an exception to be
|
| + raised and the dest file to be removed.
|
| +
|
| + Args:
|
| + func: Custom test function used to distort the downloaded data.
|
| + """
|
| + object_uri = self.CreateObject(contents='foo')
|
| + # Need to explicitly tell the key to populate its etag so that hash
|
| + # validation will be performed.
|
| + object_uri.get_key().set_etag()
|
| + dst_dir = self.CreateTempDir()
|
| + got_expected_exception = False
|
| + try:
|
| + self.RunCommand('cp', [suri(object_uri), dst_dir], test_method=func)
|
| + self.fail('Did not get expected CommandException')
|
| + except HashMismatchException:
|
| + self.assertFalse(os.listdir(dst_dir))
|
| + got_expected_exception = True
|
| + except Exception, e:
|
| + self.fail('Unexpected exception raised: %s' % e)
|
| + if not got_expected_exception:
|
| + self.fail('Did not get expected CommandException')
|
| +
|
| + def testDownloadWithObjectSizeChange(self):
|
| + """Test resumable download on an object that changes size.
|
| +
|
| + Size change occurs before the downloaded file's checksum is validated.
|
| + """
|
| + self.DownloadTestHelper(_Append)
|
| +
|
| + def testDownloadWithFileContentChange(self):
|
| + """Tests resumable download on an object that changes content.
|
| +
|
| + Content change occurs before the downloaded file's checksum is validated.
|
| + """
|
| + self.DownloadTestHelper(_Overwrite)
|
| +
|
| + # @PerformsFileToObjectUpload
|
| + def testFlatCopyingObjsAndFilesToBucketSubDir(self):
|
| + """Tests copying flatly listed objects and files to bucket subdir."""
|
| + src_bucket_uri = self.CreateBucket(test_objects=['f0', 'd0/f1', 'd1/d2/f2'])
|
| + src_dir = self.CreateTempDir(test_files=['f3', ('d3', 'f4'),
|
| + ('d4', 'd5', 'f5')])
|
| + dst_bucket_uri = self.CreateBucket(test_objects=['dst_subdir0/existing',
|
| + 'dst_subdir1/existing'])
|
| + # Test with and without final slash on dest subdir.
|
| + for i, final_char in enumerate(('/', '')):
|
| + self.RunCommand(
|
| + 'cp', ['-R', suri(src_bucket_uri, '**'), os.path.join(src_dir, '**'),
|
| + suri(dst_bucket_uri, 'dst_subdir%d' % i) + final_char])
|
| +
|
| + actual = set(str(u) for u in self._test_wildcard_iterator(
|
| + suri(dst_bucket_uri, '**')).IterAll(expand_top_level_buckets=True))
|
| + expected = set()
|
| + for i in range(2):
|
| + expected.add(suri(dst_bucket_uri, 'dst_subdir%d' % i, 'existing'))
|
| + for j in range(6):
|
| + expected.add(suri(dst_bucket_uri, 'dst_subdir%d' % i, 'f%d' % j))
|
| + self.assertEqual(expected, actual)
|
| +
|
| + # @PerformsFileToObjectUpload
|
| + def testRecursiveCopyObjsAndFilesToExistingBucketSubDir(self):
|
| + """Tests recursive copy of objects and files to existing bucket subdir."""
|
| + src_bucket_uri = self.CreateBucket(test_objects=['f0', 'nested/f1'])
|
| + dst_bucket_uri = self.CreateBucket(test_objects=[
|
| + 'dst_subdir0/existing_obj', 'dst_subdir1/existing_obj'])
|
| + src_dir = self.CreateTempDir(test_files=['f2', ('nested', 'f3')])
|
| + # Test with and without final slash on dest subdir.
|
| + for i, final_char in enumerate(('/', '')):
|
| + self.RunCommand(
|
| + 'cp', ['-R', suri(src_bucket_uri), src_dir,
|
| + suri(dst_bucket_uri, 'dst_subdir%d' % i) + final_char])
|
| + actual = set(str(u) for u in self._test_wildcard_iterator(
|
| + suri(dst_bucket_uri, 'dst_subdir%d' % i, '**')).IterAll(
|
| + expand_top_level_buckets=True))
|
| + tmp_dirname = os.path.split(src_dir)[1]
|
| + bucketname = src_bucket_uri.bucket_name
|
| + expected = set([
|
| + suri(dst_bucket_uri, 'dst_subdir%d' % i, 'existing_obj'),
|
| + suri(dst_bucket_uri, 'dst_subdir%d' % i, bucketname, 'f0'),
|
| + suri(dst_bucket_uri, 'dst_subdir%d' % i, bucketname, 'nested', 'f1'),
|
| + suri(dst_bucket_uri, 'dst_subdir%d' % i, tmp_dirname, 'f2'),
|
| + suri(dst_bucket_uri, 'dst_subdir%d' % i, tmp_dirname, 'nested', 'f3')
|
| + ])
|
| + self.assertEqual(expected, actual)
|
| +
|
| + # @PerformsFileToObjectUpload
|
| + def testRecursiveCopyObjsAndFilesToNonExistentBucketSubDir(self):
|
| + """Tests recursive copy of objs + files to non-existent bucket subdir."""
|
| + src_bucket_uri = self.CreateBucket(test_objects=['f0', 'nested/f1'])
|
| + src_dir = self.CreateTempDir(test_files=['f2', ('nested', 'f3')])
|
| + dst_bucket_uri = self.CreateBucket()
|
| + self.RunCommand('cp', ['-R', src_dir, suri(src_bucket_uri),
|
| + suri(dst_bucket_uri, 'dst_subdir')])
|
| + actual = set(str(u) for u in self._test_wildcard_iterator(
|
| + suri(dst_bucket_uri, '**')).IterAll(expand_top_level_buckets=True))
|
| + expected = set([suri(dst_bucket_uri, 'dst_subdir', 'f0'),
|
| + suri(dst_bucket_uri, 'dst_subdir', 'nested', 'f1'),
|
| + suri(dst_bucket_uri, 'dst_subdir', 'f2'),
|
| + suri(dst_bucket_uri, 'dst_subdir', 'nested', 'f3')])
|
| + self.assertEqual(expected, actual)
|
| +
|
| + def testCopyingBucketSubDirToDir(self):
|
| + """Tests copying a bucket subdir to a directory."""
|
| + src_bucket_uri = self.CreateBucket(test_objects=['src_subdir/obj'])
|
| + dst_dir = self.CreateTempDir()
|
| + # Test with and without final slash on dest subdir.
|
| + for (final_src_char, final_dst_char) in (
|
| + ('', ''), ('', '/'), ('/', ''), ('/', '/')):
|
| + # Mock objects don't support hash digestion.
|
| + with SetBotoConfigForTest([('GSUtil', 'check_hashes', 'never')]):
|
| + self.RunCommand(
|
| + 'cp', ['-R', suri(src_bucket_uri, 'src_subdir') + final_src_char,
|
| + dst_dir + final_dst_char])
|
| + actual = set(str(u) for u in self._test_wildcard_iterator(
|
| + '%s%s**' % (dst_dir, os.sep)).IterAll(expand_top_level_buckets=True))
|
| + expected = set([suri(dst_dir, 'src_subdir', 'obj')])
|
| + self.assertEqual(expected, actual)
|
| +
|
| + def testCopyingWildcardSpecifiedBucketSubDirToExistingDir(self):
|
| + """Tests copying a wildcard-specified bucket subdir to a directory."""
|
| + src_bucket_uri = self.CreateBucket(
|
| + test_objects=['src_sub0dir/foo', 'src_sub1dir/foo', 'src_sub2dir/foo',
|
| + 'src_sub3dir/foo'])
|
| + dst_dir = self.CreateTempDir()
|
| + # Test with and without final slash on dest subdir.
|
| + for i, (final_src_char, final_dst_char) in enumerate((
|
| + ('', ''), ('', '/'), ('/', ''), ('/', '/'))):
|
| + # Mock objects don't support hash digestion.
|
| + with SetBotoConfigForTest([('GSUtil', 'check_hashes', 'never')]):
|
| + self.RunCommand(
|
| + 'cp', ['-R', suri(src_bucket_uri, 'src_sub%d*' % i) +
|
| + final_src_char, dst_dir + final_dst_char])
|
| + actual = set(str(u) for u in self._test_wildcard_iterator(
|
| + os.path.join(dst_dir, 'src_sub%ddir' % i, '**')).IterAll(
|
| + expand_top_level_buckets=True))
|
| + expected = set([suri(dst_dir, 'src_sub%ddir' % i, 'foo')])
|
| + self.assertEqual(expected, actual)
|
| +
|
| + def testCopyingBucketSubDirToDirFailsWithoutMinusR(self):
|
| + """Tests for failure when attempting bucket subdir copy without -R."""
|
| + src_bucket_uri = self.CreateBucket(test_objects=['src_subdir/obj'])
|
| + dst_dir = self.CreateTempDir()
|
| + try:
|
| + self.RunCommand(
|
| + 'cp', [suri(src_bucket_uri, 'src_subdir'), dst_dir])
|
| + self.fail('Did not get expected CommandException')
|
| + except CommandException, e:
|
| + self.assertIn('No URLs matched', e.reason)
|
| +
|
| + def testCopyingBucketSubDirToBucketSubDir(self):
|
| + """Tests copying a bucket subdir to another bucket subdir."""
|
| + src_bucket_uri = self.CreateBucket(
|
| + test_objects=['src_subdir_%d/obj' % i for i in range(4)])
|
| + dst_bucket_uri = self.CreateBucket(
|
| + test_objects=['dst_subdir_%d/obj2' % i for i in range(4)])
|
| + # Test with and without final slash on dest subdir.
|
| + for i, (final_src_char, final_dst_char) in enumerate((
|
| + ('', ''), ('', '/'), ('/', ''), ('/', '/'))):
|
| + self.RunCommand(
|
| + 'cp', ['-R',
|
| + suri(src_bucket_uri, 'src_subdir_%d' % i) + final_src_char,
|
| + suri(dst_bucket_uri, 'dst_subdir_%d' % i) + final_dst_char])
|
| + actual = set(str(u) for u in self._test_wildcard_iterator(
|
| + suri(dst_bucket_uri, 'dst_subdir_%d' % i, '**')).IterAll(
|
| + expand_top_level_buckets=True))
|
| + expected = set([suri(dst_bucket_uri, 'dst_subdir_%d' % i,
|
| + 'src_subdir_%d' % i, 'obj'),
|
| + suri(dst_bucket_uri, 'dst_subdir_%d' % i, 'obj2')])
|
| + self.assertEqual(expected, actual)
|
| +
|
| + def testCopyingBucketSubDirToBucketSubDirWithNested(self):
|
| + """Tests copying a bucket subdir to another bucket subdir with nesting."""
|
| + src_bucket_uri = self.CreateBucket(
|
| + test_objects=['src_subdir_%d/obj' % i for i in range(4)] +
|
| + ['src_subdir_%d/nested/obj' % i for i in range(4)])
|
| + dst_bucket_uri = self.CreateBucket(
|
| + test_objects=['dst_subdir_%d/obj2' % i for i in range(4)])
|
| + # Test with and without final slash on dest subdir.
|
| + for i, (final_src_char, final_dst_char) in enumerate((
|
| + ('', ''), ('', '/'), ('/', ''), ('/', '/'))):
|
| + self.RunCommand(
|
| + 'cp', ['-R',
|
| + suri(src_bucket_uri, 'src_subdir_%d' % i) + final_src_char,
|
| + suri(dst_bucket_uri, 'dst_subdir_%d' % i) + final_dst_char])
|
| + actual = set(str(u) for u in self._test_wildcard_iterator(
|
| + suri(dst_bucket_uri, 'dst_subdir_%d' % i, '**')).IterAll(
|
| + expand_top_level_buckets=True))
|
| + expected = set([suri(dst_bucket_uri, 'dst_subdir_%d' % i,
|
| + 'src_subdir_%d' % i, 'obj'),
|
| + suri(dst_bucket_uri, 'dst_subdir_%d' % i,
|
| + 'src_subdir_%d' % i, 'nested', 'obj'),
|
| + suri(dst_bucket_uri, 'dst_subdir_%d' % i, 'obj2')])
|
| + self.assertEqual(expected, actual)
|
| +
|
| + def testMovingBucketSubDirToExistingBucketSubDir(self):
|
| + """Tests moving a bucket subdir to a existing bucket subdir."""
|
| + src_objs = ['foo']
|
| + for i in range(4):
|
| + src_objs.extend(['src_subdir%d/foo2' % i, 'src_subdir%d/nested/foo3' % i])
|
| + src_bucket_uri = self.CreateBucket(test_objects=src_objs)
|
| + dst_bucket_uri = self.CreateBucket(
|
| + test_objects=['dst_subdir%d/existing' % i for i in range(4)])
|
| + # Test with and without final slash on dest subdir.
|
| + for i, (final_src_char, final_dst_char) in enumerate((
|
| + ('', ''), ('', '/'), ('/', ''), ('/', '/'))):
|
| + self.RunCommand(
|
| + 'mv', [suri(src_bucket_uri, 'src_subdir%d' % i) + final_src_char,
|
| + suri(dst_bucket_uri, 'dst_subdir%d' % i) + final_dst_char])
|
| +
|
| + actual = set(str(u) for u in self._test_wildcard_iterator(
|
| + suri(dst_bucket_uri, '**')).IterAll(expand_top_level_buckets=True))
|
| + expected = set()
|
| + for i in range(4):
|
| + expected.add(suri(dst_bucket_uri, 'dst_subdir%d' % i, 'existing'))
|
| + expected.add(suri(dst_bucket_uri, 'dst_subdir%d' % i, 'src_subdir%d' %i,
|
| + 'foo2'))
|
| + expected.add(suri(dst_bucket_uri, 'dst_subdir%d' % i, 'src_subdir%d' %i,
|
| + 'nested', 'foo3'))
|
| + self.assertEqual(expected, actual)
|
| +
|
| + def testCopyingObjectToBucketSubDir(self):
|
| + """Tests copying an object to a bucket subdir."""
|
| + src_bucket_uri = self.CreateBucket(test_objects=['obj0'])
|
| + dst_bucket_uri = self.CreateBucket(test_objects=['dir0/existing',
|
| + 'dir1/existing'])
|
| + # Test with and without final slash on dest subdir.
|
| + for i, final_dst_char in enumerate(('', '/')):
|
| + self.RunCommand('cp', [
|
| + suri(src_bucket_uri, 'obj0'),
|
| + suri(dst_bucket_uri, 'dir%d' % i) + final_dst_char])
|
| + actual = set(str(u) for u in self._test_wildcard_iterator(
|
| + suri(dst_bucket_uri, 'dir%d' % i, '**')).IterAll(
|
| + expand_top_level_buckets=True))
|
| + expected = set([suri(dst_bucket_uri, 'dir%d' % i, 'obj0'),
|
| + suri(dst_bucket_uri, 'dir%d' % i, 'existing')])
|
| + self.assertEqual(expected, actual)
|
| +
|
| + # @PerformsFileToObjectUpload
|
| + def testCopyingWildcardedFilesToBucketSubDir(self):
|
| + """Tests copying wildcarded files to a bucket subdir."""
|
| + dst_bucket_uri = self.CreateBucket(test_objects=['subdir0/existing',
|
| + 'subdir1/existing'])
|
| + src_dir = self.CreateTempDir(test_files=['f0', 'f1', 'f2'])
|
| + # Test with and without final slash on dest subdir.
|
| + for i, final_dst_char in enumerate(('', '/')):
|
| + self.RunCommand(
|
| + 'cp', [os.path.join(src_dir, 'f?'),
|
| + suri(dst_bucket_uri, 'subdir%d' % i) + final_dst_char])
|
| + actual = set(str(u) for u in self._test_wildcard_iterator(
|
| + suri(dst_bucket_uri, 'subdir%d' % i, '**')).IterAll(
|
| + expand_top_level_buckets=True))
|
| + expected = set([suri(dst_bucket_uri, 'subdir%d' % i, 'existing'),
|
| + suri(dst_bucket_uri, 'subdir%d' % i, 'f0'),
|
| + suri(dst_bucket_uri, 'subdir%d' % i, 'f1'),
|
| + suri(dst_bucket_uri, 'subdir%d' % i, 'f2')])
|
| + self.assertEqual(expected, actual)
|
| +
|
| + # @PerformsFileToObjectUpload
|
| + def testCopyingOneNestedFileToBucketSubDir(self):
|
| + """Tests copying one nested file to a bucket subdir."""
|
| + dst_bucket_uri = self.CreateBucket(test_objects=['d0/placeholder',
|
| + 'd1/placeholder'])
|
| + src_dir = self.CreateTempDir(test_files=[('d3', 'd4', 'nested', 'f1')])
|
| + # Test with and without final slash on dest subdir.
|
| + for i, final_dst_char in enumerate(('', '/')):
|
| + self.RunCommand('cp', ['-r', suri(src_dir, 'd3'),
|
| + suri(dst_bucket_uri, 'd%d' % i) + final_dst_char])
|
| + actual = set(str(u) for u in self._test_wildcard_iterator(
|
| + suri(dst_bucket_uri, '**')).IterAll(expand_top_level_buckets=True))
|
| + expected = set([
|
| + suri(dst_bucket_uri, 'd0', 'placeholder'),
|
| + suri(dst_bucket_uri, 'd1', 'placeholder'),
|
| + suri(dst_bucket_uri, 'd0', 'd3', 'd4', 'nested', 'f1'),
|
| + suri(dst_bucket_uri, 'd1', 'd3', 'd4', 'nested', 'f1')])
|
| + self.assertEqual(expected, actual)
|
| +
|
| + def testMovingWildcardedFilesToNonExistentBucketSubDir(self):
|
| + """Tests moving files to a non-existent bucket subdir."""
|
| + # This tests for how we allow users to do something like:
|
| + # gsutil cp *.txt gs://bucket/dir
|
| + # where *.txt matches more than 1 file and gs://bucket/dir
|
| + # doesn't exist as a subdir.
|
| + #
|
| + src_bucket_uri = self.CreateBucket(test_objects=[
|
| + 'f0f0', 'f0f1', 'f1f0', 'f1f1'])
|
| + dst_bucket_uri = self.CreateBucket(test_objects=[
|
| + 'dst_subdir0/existing_obj', 'dst_subdir1/existing_obj'])
|
| + # Test with and without final slash on dest subdir.
|
| + for i, final_dst_char in enumerate(('', '/')):
|
| + # Copy some files into place in dst bucket.
|
| + self.RunCommand(
|
| + 'cp', [suri(src_bucket_uri, 'f%df*' % i),
|
| + suri(dst_bucket_uri, 'dst_subdir%d' % i) + final_dst_char])
|
| + # Now do the move test.
|
| + self.RunCommand(
|
| + 'mv', [suri(src_bucket_uri, 'f%d*' % i),
|
| + suri(dst_bucket_uri, 'nonexisting%d' % i) + final_dst_char])
|
| +
|
| + actual = set(str(u) for u in self._test_wildcard_iterator(
|
| + suri(dst_bucket_uri, '**')).IterAll(expand_top_level_buckets=True))
|
| + expected = set([
|
| + suri(dst_bucket_uri, 'dst_subdir0', 'existing_obj'),
|
| + suri(dst_bucket_uri, 'dst_subdir0', 'f0f0'),
|
| + suri(dst_bucket_uri, 'dst_subdir0', 'f0f1'),
|
| + suri(dst_bucket_uri, 'nonexisting0', 'f0f0'),
|
| + suri(dst_bucket_uri, 'nonexisting0', 'f0f1'),
|
| + suri(dst_bucket_uri, 'dst_subdir1', 'existing_obj'),
|
| + suri(dst_bucket_uri, 'dst_subdir1', 'f1f0'),
|
| + suri(dst_bucket_uri, 'dst_subdir1', 'f1f1'),
|
| + suri(dst_bucket_uri, 'nonexisting1', 'f1f0'),
|
| + suri(dst_bucket_uri, 'nonexisting1', 'f1f1')])
|
| + self.assertEqual(expected, actual)
|
| +
|
| + def testMovingObjectToBucketSubDir(self):
|
| + """Tests moving an object to a bucket subdir."""
|
| + src_bucket_uri = self.CreateBucket(test_objects=['obj0', 'obj1'])
|
| + dst_bucket_uri = self.CreateBucket(test_objects=[
|
| + 'dst_subdir0/existing_obj', 'dst_subdir1/existing_obj'])
|
| + # Test with and without final slash on dest subdir.
|
| + for i, final_dst_char in enumerate(('', '/')):
|
| + self.RunCommand(
|
| + 'mv', [suri(src_bucket_uri, 'obj%d' % i),
|
| + suri(dst_bucket_uri, 'dst_subdir%d' % i) + final_dst_char])
|
| +
|
| + actual = set(str(u) for u in self._test_wildcard_iterator(
|
| + suri(dst_bucket_uri, '**')).IterAll(expand_top_level_buckets=True))
|
| + expected = set([
|
| + suri(dst_bucket_uri, 'dst_subdir0', 'existing_obj'),
|
| + suri(dst_bucket_uri, 'dst_subdir0', 'obj0'),
|
| + suri(dst_bucket_uri, 'dst_subdir1', 'existing_obj'),
|
| + suri(dst_bucket_uri, 'dst_subdir1', 'obj1')])
|
| + self.assertEqual(expected, actual)
|
| +
|
| + actual = set(str(u) for u in self._test_wildcard_iterator(
|
| + suri(src_bucket_uri, '**')).IterAll(expand_top_level_buckets=True))
|
| + self.assertEqual(actual, set())
|
| +
|
| + def testWildcardSrcSubDirMoveDisallowed(self):
|
| + """Tests moving a bucket subdir specified by wildcard is disallowed."""
|
| + src_bucket_uri = self.CreateBucket(test_objects=['dir/foo1'])
|
| + dst_bucket_uri = self.CreateBucket(test_objects=['dir/foo2'])
|
| + try:
|
| + self.RunCommand(
|
| + 'mv', [suri(src_bucket_uri, 'dir*'), suri(dst_bucket_uri, 'dir')])
|
| + self.fail('Did not get expected CommandException')
|
| + except CommandException, e:
|
| + self.assertIn('mv command disallows naming', e.reason)
|
| +
|
| + def testMovingBucketSubDirToNonExistentBucketSubDir(self):
|
| + """Tests moving a bucket subdir to a non-existent bucket subdir."""
|
| + src_bucket = self.CreateBucket(test_objects=[
|
| + 'foo', 'src_subdir0/foo2', 'src_subdir0/nested/foo3',
|
| + 'src_subdir1/foo2', 'src_subdir1/nested/foo3'])
|
| + dst_bucket = self.CreateBucket()
|
| + # Test with and without final slash on dest subdir.
|
| + for i, final_src_char in enumerate(('', '/')):
|
| + self.RunCommand(
|
| + 'mv', [suri(src_bucket, 'src_subdir%d' % i) + final_src_char,
|
| + suri(dst_bucket, 'dst_subdir%d' % i)])
|
| +
|
| + actual = set(str(u) for u in self._test_wildcard_iterator(
|
| + suri(dst_bucket, '**')).IterAll(expand_top_level_buckets=True))
|
| + # Unlike the case with copying, with mv we expect renaming to occur
|
| + # at the level of the src subdir, vs appending that subdir beneath the
|
| + # dst subdir like is done for copying.
|
| + expected = set([suri(dst_bucket, 'dst_subdir0', 'foo2'),
|
| + suri(dst_bucket, 'dst_subdir1', 'foo2'),
|
| + suri(dst_bucket, 'dst_subdir0', 'nested', 'foo3'),
|
| + suri(dst_bucket, 'dst_subdir1', 'nested', 'foo3')])
|
| + self.assertEqual(expected, actual)
|
| +
|
| + def testRemovingBucketSubDir(self):
|
| + """Tests removing a bucket subdir."""
|
| + dst_bucket_uri = self.CreateBucket(test_objects=[
|
| + 'f0', 'dir0/f1', 'dir0/nested/f2', 'dir1/f1', 'dir1/nested/f2'])
|
| + # Test with and without final slash on dest subdir.
|
| + for i, final_src_char in enumerate(('', '/')):
|
| + # Test removing bucket subdir.
|
| + self.RunCommand(
|
| + 'rm', ['-R', suri(dst_bucket_uri, 'dir%d' % i) + final_src_char])
|
| + actual = set(str(u) for u in self._test_wildcard_iterator(
|
| + suri(dst_bucket_uri, '**')).IterAll(expand_top_level_buckets=True))
|
| + expected = set([suri(dst_bucket_uri, 'f0')])
|
| + self.assertEqual(expected, actual)
|
| +
|
| + def testRecursiveRemoveObjsInBucket(self):
|
| + """Tests removing all objects in bucket via rm -R gs://bucket."""
|
| + bucket_uris = [
|
| + self.CreateBucket(test_objects=['f0', 'dir/f1', 'dir/nested/f2']),
|
| + self.CreateBucket(test_objects=['f0', 'dir/f1', 'dir/nested/f2'])]
|
| + # Test with and without final slash on dest subdir.
|
| + for i, final_src_char in enumerate(('', '/')):
|
| + # Test removing all objects via rm -R.
|
| + self.RunCommand('rm', ['-R', suri(bucket_uris[i]) + final_src_char])
|
| + try:
|
| + self.RunCommand('ls', [suri(bucket_uris[i])])
|
| + # Ensure exception is raised.
|
| + self.assertTrue(False)
|
| + except NotFoundException, e:
|
| + self.assertEqual(e.status, 404)
|
| +
|
| + def testUnicodeArgs(self):
|
| + """Tests that you can list an object with unicode characters."""
|
| + object_name = u'フォ'
|
| + bucket_uri = self.CreateBucket()
|
| + self.CreateObject(bucket_uri=bucket_uri, object_name=object_name,
|
| + contents='foo')
|
| + object_name_bytes = object_name.encode(UTF8)
|
| + stdout = self.RunCommand('ls', [suri(bucket_uri, object_name_bytes)],
|
| + return_stdout=True)
|
| + self.assertIn(object_name_bytes, stdout)
|
| +
|
| + def testRecursiveListTrailingSlash(self):
|
| + bucket_uri = self.CreateBucket()
|
| + obj_uri = self.CreateObject(
|
| + bucket_uri=bucket_uri, object_name='/', contents='foo')
|
| + stdout = self.RunCommand('ls', ['-R', suri(bucket_uri)], return_stdout=True)
|
| + # Note: The suri function normalizes the URI, so the double slash gets
|
| + # removed.
|
| + self.assertEqual(stdout.splitlines(), [suri(obj_uri) + '/:',
|
| + suri(obj_uri) + '/'])
|
| +
|
| + def FinalObjNameComponent(self, uri):
|
| + """For gs://bucket/abc/def/ghi returns ghi."""
|
| + return uri.uri.rpartition('/')[-1]
|
| +
|
| + def testFileContainingColon(self):
|
| + url_str = 'abc:def'
|
| + url = StorageUrlFromString(url_str)
|
| + self.assertEqual('file', url.scheme)
|
| + self.assertEqual('file://%s' % url_str, url.url_string)
|
| +
|
| +
|
| +# TODO: These should all be moved to their own test_*.py testing files.
|
| +class GsUtilCommandTests(testcase.GsUtilUnitTestCase):
|
| + """Basic sanity check tests to make sure commands run."""
|
| +
|
| + def testDisableLoggingCommandRuns(self):
|
| + """Test that the 'logging set off' command basically runs."""
|
| + src_bucket_uri = self.CreateBucket()
|
| + self.RunCommand('logging', ['set', 'off', suri(src_bucket_uri)])
|
| +
|
| + def testEnableLoggingCommandRuns(self):
|
| + """Test that the 'logging set on' command basically runs."""
|
| + src_bucket_uri = self.CreateBucket()
|
| + self.RunCommand('logging', ['set', 'on', '-b', 'gs://log_bucket',
|
| + suri(src_bucket_uri)])
|
| +
|
| + def testHelpCommandDoesntRaise(self):
|
| + """Test that the help command doesn't raise (sanity checks all help)."""
|
| + # Unset PAGER if defined, so help output paginating into $PAGER doesn't
|
| + # cause test to pause.
|
| + if 'PAGER' in os.environ:
|
| + del os.environ['PAGER']
|
| + self.RunCommand('help', [])
|
| +
|
| + def testCatCommandRuns(self):
|
| + """Test that the cat command basically runs."""
|
| + src_uri = self.CreateObject(contents='foo')
|
| + stdout = self.RunCommand('cat', [suri(src_uri)], return_stdout=True)
|
| + self.assertEqual(stdout, 'foo')
|
| +
|
| + def testGetLoggingCommandRuns(self):
|
| + """Test that the 'logging get' command basically runs."""
|
| + src_bucket_uri = self.CreateBucket()
|
| + self.RunCommand('logging', ['get', suri(src_bucket_uri)])
|
| +
|
| + def testMakeBucketsCommand(self):
|
| + """Test mb on existing bucket."""
|
| + dst_bucket_uri = self.CreateBucket()
|
| + try:
|
| + self.RunCommand('mb', [suri(dst_bucket_uri)])
|
| + self.fail('Did not get expected StorageCreateError')
|
| + except ServiceException, e:
|
| + self.assertEqual(e.status, 409)
|
| +
|
| + def testRemoveBucketsCommand(self):
|
| + """Test rb on non-existent bucket."""
|
| + dst_bucket_uri = self.CreateBucket()
|
| + try:
|
| + self.RunCommand(
|
| + 'rb', ['gs://no_exist_%s' % dst_bucket_uri.bucket_name])
|
| + self.fail('Did not get expected NotFoundException')
|
| + except NotFoundException, e:
|
| + self.assertEqual(e.status, 404)
|
| +
|
| + def testRemoveObjsCommand(self):
|
| + """Test rm command on non-existent object."""
|
| + dst_bucket_uri = self.CreateBucket()
|
| + try:
|
| + self.RunCommand('rm', [suri(dst_bucket_uri, 'non_existent')])
|
| + self.fail('Did not get expected CommandException')
|
| + except CommandException, e:
|
| + self.assertIn('No URLs matched', e.reason)
|
| +
|
| + # Now that gsutil ver computes a checksum it adds 1-3 seconds to test run
|
| + # time (for in memory mocked tests that otherwise take ~ 0.1 seconds). Since
|
| + # it provides very little test value, we're leaving this test commented out.
|
| + # def testVerCommmandRuns(self):
|
| + # """Test that the Ver command basically runs"""
|
| + # self.RunCommand('ver', [])
|
|
|