| Index: tools/telemetry/third_party/gsutilz/gslib/tests/test_ls.py
|
| diff --git a/tools/telemetry/third_party/gsutilz/gslib/tests/test_ls.py b/tools/telemetry/third_party/gsutilz/gslib/tests/test_ls.py
|
| deleted file mode 100644
|
| index 710a3e2a99f618164f8bba9d67f0fea478bf09ad..0000000000000000000000000000000000000000
|
| --- a/tools/telemetry/third_party/gsutilz/gslib/tests/test_ls.py
|
| +++ /dev/null
|
| @@ -1,439 +0,0 @@
|
| -# -*- coding: utf-8 -*-
|
| -# Copyright 2013 Google Inc. All Rights Reserved.
|
| -#
|
| -# Licensed under the Apache License, Version 2.0 (the "License");
|
| -# you may not use this file except in compliance with the License.
|
| -# You may obtain a copy of the License at
|
| -#
|
| -# http://www.apache.org/licenses/LICENSE-2.0
|
| -#
|
| -# Unless required by applicable law or agreed to in writing, software
|
| -# distributed under the License is distributed on an "AS IS" BASIS,
|
| -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
| -# See the License for the specific language governing permissions and
|
| -# limitations under the License.
|
| -"""Tests for ls command."""
|
| -
|
| -from __future__ import absolute_import
|
| -
|
| -import posixpath
|
| -import re
|
| -import subprocess
|
| -import sys
|
| -
|
| -import gslib
|
| -from gslib.cs_api_map import ApiSelector
|
| -import gslib.tests.testcase as testcase
|
| -from gslib.tests.testcase.integration_testcase import SkipForS3
|
| -from gslib.tests.util import ObjectToURI as suri
|
| -from gslib.tests.util import unittest
|
| -from gslib.util import IS_WINDOWS
|
| -from gslib.util import Retry
|
| -from gslib.util import UTF8
|
| -
|
| -
|
| -class TestLs(testcase.GsUtilIntegrationTestCase):
|
| - """Integration tests for ls command."""
|
| -
|
| - def test_blank_ls(self):
|
| - self.RunGsUtil(['ls'])
|
| -
|
| - def test_empty_bucket(self):
|
| - bucket_uri = self.CreateBucket()
|
| - self.AssertNObjectsInBucket(bucket_uri, 0)
|
| -
|
| - def test_empty_bucket_with_b(self):
|
| - bucket_uri = self.CreateBucket()
|
| - # Use @Retry as hedge against bucket listing eventual consistency.
|
| - @Retry(AssertionError, tries=3, timeout_secs=1)
|
| - def _Check1():
|
| - stdout = self.RunGsUtil(['ls', '-b', suri(bucket_uri)],
|
| - return_stdout=True)
|
| - self.assertEqual('%s/\n' % suri(bucket_uri), stdout)
|
| - _Check1()
|
| -
|
| - def test_bucket_with_Lb(self):
|
| - """Tests ls -Lb."""
|
| - bucket_uri = self.CreateBucket()
|
| - # Use @Retry as hedge against bucket listing eventual consistency.
|
| - @Retry(AssertionError, tries=3, timeout_secs=1)
|
| - def _Check1():
|
| - stdout = self.RunGsUtil(['ls', '-Lb', suri(bucket_uri)],
|
| - return_stdout=True)
|
| - self.assertIn(suri(bucket_uri), stdout)
|
| - self.assertNotIn('TOTAL:', stdout)
|
| - _Check1()
|
| -
|
| - def test_bucket_with_lb(self):
|
| - """Tests ls -lb."""
|
| - bucket_uri = self.CreateBucket()
|
| - # Use @Retry as hedge against bucket listing eventual consistency.
|
| - @Retry(AssertionError, tries=3, timeout_secs=1)
|
| - def _Check1():
|
| - stdout = self.RunGsUtil(['ls', '-lb', suri(bucket_uri)],
|
| - return_stdout=True)
|
| - self.assertIn(suri(bucket_uri), stdout)
|
| - self.assertNotIn('TOTAL:', stdout)
|
| - _Check1()
|
| -
|
| - def test_bucket_list_wildcard(self):
|
| - """Tests listing multiple buckets with a wildcard."""
|
| - random_prefix = self.MakeRandomTestString()
|
| - bucket1_name = self.MakeTempName('bucket', prefix=random_prefix)
|
| - bucket2_name = self.MakeTempName('bucket', prefix=random_prefix)
|
| - bucket1_uri = self.CreateBucket(bucket_name=bucket1_name)
|
| - bucket2_uri = self.CreateBucket(bucket_name=bucket2_name)
|
| - # This just double checks that the common prefix of the two buckets is what
|
| - # we think it should be (based on implementation detail of CreateBucket).
|
| - # We want to be careful when setting a wildcard on buckets to make sure we
|
| - # don't step outside the test buckets to affect other buckets.
|
| - common_prefix = posixpath.commonprefix([suri(bucket1_uri),
|
| - suri(bucket2_uri)])
|
| - self.assertTrue(common_prefix.startswith(
|
| - '%s://%sgsutil-test-test_bucket_list_wildcard-bucket-' %
|
| - (self.default_provider, random_prefix)))
|
| - wildcard = '%s*' % common_prefix
|
| -
|
| - # Use @Retry as hedge against bucket listing eventual consistency.
|
| - @Retry(AssertionError, tries=3, timeout_secs=1)
|
| - def _Check1():
|
| - stdout = self.RunGsUtil(['ls', '-b', wildcard], return_stdout=True)
|
| - expected = set([suri(bucket1_uri) + '/', suri(bucket2_uri) + '/'])
|
| - actual = set(stdout.split())
|
| - self.assertEqual(expected, actual)
|
| - _Check1()
|
| -
|
| - def test_nonexistent_bucket_with_ls(self):
|
| - """Tests a bucket that is known not to exist."""
|
| - stderr = self.RunGsUtil(
|
| - ['ls', '-lb', 'gs://%s' % self.nonexistent_bucket_name],
|
| - return_stderr=True, expected_status=1)
|
| - self.assertIn('404', stderr)
|
| -
|
| - stderr = self.RunGsUtil(
|
| - ['ls', '-Lb', 'gs://%s' % self.nonexistent_bucket_name],
|
| - return_stderr=True, expected_status=1)
|
| - self.assertIn('404', stderr)
|
| -
|
| - stderr = self.RunGsUtil(
|
| - ['ls', '-b', 'gs://%s' % self.nonexistent_bucket_name],
|
| - return_stderr=True, expected_status=1)
|
| - self.assertIn('404', stderr)
|
| -
|
| - def test_list_missing_object(self):
|
| - """Tests listing a non-existent object."""
|
| - bucket_uri = self.CreateBucket()
|
| - stderr = self.RunGsUtil(['ls', suri(bucket_uri, 'missing')],
|
| - return_stderr=True, expected_status=1)
|
| - self.assertIn('matched no objects', stderr)
|
| -
|
| - def test_with_one_object(self):
|
| - bucket_uri = self.CreateBucket()
|
| - obj_uri = self.CreateObject(bucket_uri=bucket_uri, contents='foo')
|
| - # Use @Retry as hedge against bucket listing eventual consistency.
|
| - @Retry(AssertionError, tries=3, timeout_secs=1)
|
| - def _Check1():
|
| - stdout = self.RunGsUtil(['ls', suri(bucket_uri)], return_stdout=True)
|
| - self.assertEqual('%s\n' % obj_uri, stdout)
|
| - _Check1()
|
| -
|
| - def test_subdir(self):
|
| - """Tests listing a bucket subdirectory."""
|
| - bucket_uri = self.CreateBucket(test_objects=1)
|
| - k1_uri = bucket_uri.clone_replace_name('foo')
|
| - k1_uri.set_contents_from_string('baz')
|
| - k2_uri = bucket_uri.clone_replace_name('dir/foo')
|
| - k2_uri.set_contents_from_string('bar')
|
| - # Use @Retry as hedge against bucket listing eventual consistency.
|
| - @Retry(AssertionError, tries=3, timeout_secs=1)
|
| - def _Check1():
|
| - stdout = self.RunGsUtil(['ls', '%s/dir' % suri(bucket_uri)],
|
| - return_stdout=True)
|
| - self.assertEqual('%s\n' % suri(k2_uri), stdout)
|
| - stdout = self.RunGsUtil(['ls', suri(k1_uri)], return_stdout=True)
|
| - self.assertEqual('%s\n' % suri(k1_uri), stdout)
|
| - _Check1()
|
| -
|
| - def test_versioning(self):
|
| - """Tests listing a versioned bucket."""
|
| - bucket1_uri = self.CreateBucket(test_objects=1)
|
| - bucket2_uri = self.CreateVersionedBucket(test_objects=1)
|
| - self.AssertNObjectsInBucket(bucket1_uri, 1, versioned=True)
|
| - bucket_list = list(bucket1_uri.list_bucket())
|
| -
|
| - objuri = [bucket1_uri.clone_replace_key(key).versionless_uri
|
| - for key in bucket_list][0]
|
| - self.RunGsUtil(['cp', objuri, suri(bucket2_uri)])
|
| - self.RunGsUtil(['cp', objuri, suri(bucket2_uri)])
|
| - # Use @Retry as hedge against bucket listing eventual consistency.
|
| - @Retry(AssertionError, tries=3, timeout_secs=1)
|
| - def _Check2():
|
| - stdout = self.RunGsUtil(['ls', '-a', suri(bucket2_uri)],
|
| - return_stdout=True)
|
| - self.assertNumLines(stdout, 3)
|
| - stdout = self.RunGsUtil(['ls', '-la', suri(bucket2_uri)],
|
| - return_stdout=True)
|
| - self.assertIn('%s#' % bucket2_uri.clone_replace_name(bucket_list[0].name),
|
| - stdout)
|
| - self.assertIn('metageneration=', stdout)
|
| - _Check2()
|
| -
|
| - def test_etag(self):
|
| - """Tests that listing an object with an etag."""
|
| - bucket_uri = self.CreateBucket()
|
| - obj_uri = self.CreateObject(bucket_uri=bucket_uri, contents='foo')
|
| - # TODO: When testcase setup can use JSON, match against the exact JSON
|
| - # etag.
|
| - etag = obj_uri.get_key().etag.strip('"\'')
|
| - # Use @Retry as hedge against bucket listing eventual consistency.
|
| - @Retry(AssertionError, tries=3, timeout_secs=1)
|
| - def _Check1():
|
| - stdout = self.RunGsUtil(['ls', '-l', suri(bucket_uri)],
|
| - return_stdout=True)
|
| - if self.test_api == ApiSelector.XML:
|
| - self.assertNotIn(etag, stdout)
|
| - else:
|
| - self.assertNotIn('etag=', stdout)
|
| - _Check1()
|
| -
|
| - def _Check2():
|
| - stdout = self.RunGsUtil(['ls', '-le', suri(bucket_uri)],
|
| - return_stdout=True)
|
| - if self.test_api == ApiSelector.XML:
|
| - self.assertIn(etag, stdout)
|
| - else:
|
| - self.assertIn('etag=', stdout)
|
| - _Check2()
|
| -
|
| - def _Check3():
|
| - stdout = self.RunGsUtil(['ls', '-ale', suri(bucket_uri)],
|
| - return_stdout=True)
|
| - if self.test_api == ApiSelector.XML:
|
| - self.assertIn(etag, stdout)
|
| - else:
|
| - self.assertIn('etag=', stdout)
|
| - _Check3()
|
| -
|
| - @SkipForS3('S3 bucket configuration values are not supported via ls.')
|
| - def test_location(self):
|
| - """Tests listing a bucket with location constraint."""
|
| - bucket_uri = self.CreateBucket()
|
| - bucket_suri = suri(bucket_uri)
|
| -
|
| - # No location info
|
| - stdout = self.RunGsUtil(['ls', '-lb', bucket_suri],
|
| - return_stdout=True)
|
| - self.assertNotIn('Location constraint', stdout)
|
| -
|
| - # Default location constraint is US
|
| - stdout = self.RunGsUtil(['ls', '-Lb', bucket_suri],
|
| - return_stdout=True)
|
| - self.assertIn('Location constraint:\t\tUS', stdout)
|
| -
|
| - @SkipForS3('S3 bucket configuration values are not supported via ls.')
|
| - def test_logging(self):
|
| - """Tests listing a bucket with logging config."""
|
| - bucket_uri = self.CreateBucket()
|
| - bucket_suri = suri(bucket_uri)
|
| -
|
| - # No logging info
|
| - stdout = self.RunGsUtil(['ls', '-lb', bucket_suri],
|
| - return_stdout=True)
|
| - self.assertNotIn('Logging configuration', stdout)
|
| -
|
| - # Logging configuration is absent by default
|
| - stdout = self.RunGsUtil(['ls', '-Lb', bucket_suri],
|
| - return_stdout=True)
|
| - self.assertIn('Logging configuration:\t\tNone', stdout)
|
| -
|
| - # Enable and check
|
| - self.RunGsUtil(['logging', 'set', 'on', '-b', bucket_suri,
|
| - bucket_suri])
|
| - stdout = self.RunGsUtil(['ls', '-Lb', bucket_suri],
|
| - return_stdout=True)
|
| - self.assertIn('Logging configuration:\t\tPresent', stdout)
|
| -
|
| - # Disable and check
|
| - self.RunGsUtil(['logging', 'set', 'off', bucket_suri])
|
| - stdout = self.RunGsUtil(['ls', '-Lb', bucket_suri],
|
| - return_stdout=True)
|
| - self.assertIn('Logging configuration:\t\tNone', stdout)
|
| -
|
| - @SkipForS3('S3 bucket configuration values are not supported via ls.')
|
| - def test_web(self):
|
| - """Tests listing a bucket with website config."""
|
| - bucket_uri = self.CreateBucket()
|
| - bucket_suri = suri(bucket_uri)
|
| -
|
| - # No website configuration
|
| - stdout = self.RunGsUtil(['ls', '-lb', bucket_suri],
|
| - return_stdout=True)
|
| - self.assertNotIn('Website configuration', stdout)
|
| -
|
| - # Website configuration is absent by default
|
| - stdout = self.RunGsUtil(['ls', '-Lb', bucket_suri],
|
| - return_stdout=True)
|
| - self.assertIn('Website configuration:\t\tNone', stdout)
|
| -
|
| - # Initialize and check
|
| - self.RunGsUtil(['web', 'set', '-m', 'google.com', bucket_suri])
|
| - stdout = self.RunGsUtil(['ls', '-Lb', bucket_suri],
|
| - return_stdout=True)
|
| - self.assertIn('Website configuration:\t\tPresent', stdout)
|
| -
|
| - # Clear and check
|
| - self.RunGsUtil(['web', 'set', bucket_suri])
|
| - stdout = self.RunGsUtil(['ls', '-Lb', bucket_suri],
|
| - return_stdout=True)
|
| - self.assertIn('Website configuration:\t\tNone', stdout)
|
| -
|
| - def test_list_sizes(self):
|
| - """Tests various size listing options."""
|
| - bucket_uri = self.CreateBucket()
|
| - self.CreateObject(bucket_uri=bucket_uri, contents='x' * 2048)
|
| -
|
| - # Use @Retry as hedge against bucket listing eventual consistency.
|
| - @Retry(AssertionError, tries=3, timeout_secs=1)
|
| - def _Check1():
|
| - stdout = self.RunGsUtil(['ls', '-l', suri(bucket_uri)],
|
| - return_stdout=True)
|
| - self.assertIn('2048', stdout)
|
| - _Check1()
|
| -
|
| - # Use @Retry as hedge against bucket listing eventual consistency.
|
| - @Retry(AssertionError, tries=3, timeout_secs=1)
|
| - def _Check2():
|
| - stdout = self.RunGsUtil(['ls', '-L', suri(bucket_uri)],
|
| - return_stdout=True)
|
| - self.assertIn('2048', stdout)
|
| - _Check2()
|
| -
|
| - # Use @Retry as hedge against bucket listing eventual consistency.
|
| - @Retry(AssertionError, tries=3, timeout_secs=1)
|
| - def _Check3():
|
| - stdout = self.RunGsUtil(['ls', '-al', suri(bucket_uri)],
|
| - return_stdout=True)
|
| - self.assertIn('2048', stdout)
|
| - _Check3()
|
| -
|
| - # Use @Retry as hedge against bucket listing eventual consistency.
|
| - @Retry(AssertionError, tries=3, timeout_secs=1)
|
| - def _Check4():
|
| - stdout = self.RunGsUtil(['ls', '-lh', suri(bucket_uri)],
|
| - return_stdout=True)
|
| - self.assertIn('2 KiB', stdout)
|
| - _Check4()
|
| -
|
| - # Use @Retry as hedge against bucket listing eventual consistency.
|
| - @Retry(AssertionError, tries=3, timeout_secs=1)
|
| - def _Check5():
|
| - stdout = self.RunGsUtil(['ls', '-alh', suri(bucket_uri)],
|
| - return_stdout=True)
|
| - self.assertIn('2 KiB', stdout)
|
| - _Check5()
|
| -
|
| - @unittest.skipIf(IS_WINDOWS,
|
| - 'Unicode handling on Windows requires mods to site-packages')
|
| - def test_list_unicode_filename(self):
|
| - """Tests listing an object with a unicode filename."""
|
| - # Note: This test fails on Windows (command.exe). I was able to get ls to
|
| - # output Unicode filenames correctly by hacking the UniStream class code
|
| - # shown at
|
| - # http://stackoverflow.com/questions/878972/windows-cmd-encoding-change-causes-python-crash/3259271
|
| - # into the start of gslib/commands/ls.py, along with no-op flush and
|
| - # isastream functions (as an experiment). However, even with that change,
|
| - # the current test still fails, since it also needs to run that
|
| - # stdout/stderr-replacement code. That UniStream class replacement really
|
| - # needs to be added to the site-packages on Windows python.
|
| - object_name = u'Аудиоархив'
|
| - object_name_bytes = object_name.encode(UTF8)
|
| - bucket_uri = self.CreateVersionedBucket()
|
| - key_uri = self.CreateObject(bucket_uri=bucket_uri, contents='foo',
|
| - object_name=object_name)
|
| - self.AssertNObjectsInBucket(bucket_uri, 1, versioned=True)
|
| - stdout = self.RunGsUtil(['ls', '-ael', suri(key_uri)],
|
| - return_stdout=True)
|
| - self.assertIn(object_name_bytes, stdout)
|
| - if self.default_provider == 'gs':
|
| - self.assertIn(str(key_uri.generation), stdout)
|
| - self.assertIn(
|
| - 'metageneration=%s' % key_uri.get_key().metageneration, stdout)
|
| - if self.test_api == ApiSelector.XML:
|
| - self.assertIn(key_uri.get_key().etag.strip('"\''), stdout)
|
| - else:
|
| - # TODO: When testcase setup can use JSON, match against the exact JSON
|
| - # etag.
|
| - self.assertIn('etag=', stdout)
|
| - elif self.default_provider == 's3':
|
| - self.assertIn(key_uri.version_id, stdout)
|
| - self.assertIn(key_uri.get_key().etag.strip('"\''), stdout)
|
| -
|
| - def test_list_gzip_content_length(self):
|
| - """Tests listing a gzipped object."""
|
| - file_size = 10000
|
| - file_contents = 'x' * file_size
|
| - fpath = self.CreateTempFile(contents=file_contents, file_name='foo.txt')
|
| - key_uri = self.CreateObject()
|
| - self.RunGsUtil(['cp', '-z', 'txt', suri(fpath), suri(key_uri)])
|
| -
|
| - # Use @Retry as hedge against bucket listing eventual consistency.
|
| - @Retry(AssertionError, tries=3, timeout_secs=1)
|
| - def _Check1():
|
| - stdout = self.RunGsUtil(['ls', '-L', suri(key_uri)], return_stdout=True)
|
| - self.assertRegexpMatches(stdout, r'Content-Encoding:\s+gzip')
|
| - find_content_length_re = r'Content-Length:\s+(?P<num>\d)'
|
| - self.assertRegexpMatches(stdout, find_content_length_re)
|
| - m = re.search(find_content_length_re, stdout)
|
| - content_length = int(m.group('num'))
|
| - self.assertGreater(content_length, 0)
|
| - self.assertLess(content_length, file_size)
|
| - _Check1()
|
| -
|
| - def test_output_chopped(self):
|
| - """Tests that gsutil still succeeds with a truncated stdout."""
|
| - bucket_uri = self.CreateBucket(test_objects=2)
|
| -
|
| - # Run Python with the -u flag so output is not buffered.
|
| - gsutil_cmd = [
|
| - sys.executable, '-u', gslib.GSUTIL_PATH, 'ls', suri(bucket_uri)]
|
| - # Set bufsize to 0 to make sure output is not buffered.
|
| - p = subprocess.Popen(gsutil_cmd, stdout=subprocess.PIPE, bufsize=0)
|
| - # Immediately close the stdout pipe so that gsutil gets a broken pipe error.
|
| - p.stdout.close()
|
| - p.wait()
|
| - # Make sure it still exited cleanly.
|
| - self.assertEqual(p.returncode, 0)
|
| -
|
| - def test_recursive_list_trailing_slash(self):
|
| - """Tests listing an object with a trailing slash."""
|
| - bucket_uri = self.CreateBucket()
|
| - self.CreateObject(bucket_uri=bucket_uri, object_name='/', contents='foo')
|
| - self.AssertNObjectsInBucket(bucket_uri, 1)
|
| - stdout = self.RunGsUtil(['ls', '-R', suri(bucket_uri)], return_stdout=True)
|
| - # Note: The suri function normalizes the URI, so the double slash gets
|
| - # removed.
|
| - self.assertIn(suri(bucket_uri) + '/', stdout)
|
| -
|
| - def test_recursive_list_trailing_two_slash(self):
|
| - """Tests listing an object with two trailing slashes."""
|
| - bucket_uri = self.CreateBucket()
|
| - self.CreateObject(bucket_uri=bucket_uri, object_name='//', contents='foo')
|
| - self.AssertNObjectsInBucket(bucket_uri, 1)
|
| - stdout = self.RunGsUtil(['ls', '-R', suri(bucket_uri)], return_stdout=True)
|
| - # Note: The suri function normalizes the URI, so the double slash gets
|
| - # removed.
|
| - self.assertIn(suri(bucket_uri) + '//', stdout)
|
| -
|
| - @SkipForS3('S3 anonymous access is not supported.')
|
| - def test_get_object_without_list_bucket_permission(self):
|
| - # Bucket is not publicly readable by default.
|
| - bucket_uri = self.CreateBucket()
|
| - object_uri = self.CreateObject(bucket_uri=bucket_uri,
|
| - object_name='permitted', contents='foo')
|
| - # Set this object to be publicly readable.
|
| - self.RunGsUtil(['acl', 'set', 'public-read', suri(object_uri)])
|
| - # Drop credentials.
|
| - with self.SetAnonymousBotoCreds():
|
| - stdout = self.RunGsUtil(['ls', '-L', suri(object_uri)],
|
| - return_stdout=True)
|
| - self.assertIn(suri(object_uri), stdout)
|
|
|