Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(67)

Unified Diff: third_party/gsutil/gslib/tests/test_ls.py

Issue 1377933002: [catapult] - Copy Telemetry's gsutilz over to third_party. (Closed) Base URL: https://github.com/catapult-project/catapult.git@master
Patch Set: Rename to gsutil. Created 5 years, 3 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View side-by-side diff with in-line comments
Download patch
« no previous file with comments | « third_party/gsutil/gslib/tests/test_logging.py ('k') | third_party/gsutil/gslib/tests/test_mb.py » ('j') | no next file with comments »
Expand Comments ('e') | Collapse Comments ('c') | Show Comments Hide Comments ('s')
Index: third_party/gsutil/gslib/tests/test_ls.py
diff --git a/third_party/gsutil/gslib/tests/test_ls.py b/third_party/gsutil/gslib/tests/test_ls.py
new file mode 100644
index 0000000000000000000000000000000000000000..710a3e2a99f618164f8bba9d67f0fea478bf09ad
--- /dev/null
+++ b/third_party/gsutil/gslib/tests/test_ls.py
@@ -0,0 +1,439 @@
+# -*- coding: utf-8 -*-
+# Copyright 2013 Google Inc. All Rights Reserved.
+#
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+"""Tests for ls command."""
+
+from __future__ import absolute_import
+
+import posixpath
+import re
+import subprocess
+import sys
+
+import gslib
+from gslib.cs_api_map import ApiSelector
+import gslib.tests.testcase as testcase
+from gslib.tests.testcase.integration_testcase import SkipForS3
+from gslib.tests.util import ObjectToURI as suri
+from gslib.tests.util import unittest
+from gslib.util import IS_WINDOWS
+from gslib.util import Retry
+from gslib.util import UTF8
+
+
+class TestLs(testcase.GsUtilIntegrationTestCase):
+ """Integration tests for ls command."""
+
+ def test_blank_ls(self):
+ self.RunGsUtil(['ls'])
+
+ def test_empty_bucket(self):
+ bucket_uri = self.CreateBucket()
+ self.AssertNObjectsInBucket(bucket_uri, 0)
+
+ def test_empty_bucket_with_b(self):
+ bucket_uri = self.CreateBucket()
+ # Use @Retry as hedge against bucket listing eventual consistency.
+ @Retry(AssertionError, tries=3, timeout_secs=1)
+ def _Check1():
+ stdout = self.RunGsUtil(['ls', '-b', suri(bucket_uri)],
+ return_stdout=True)
+ self.assertEqual('%s/\n' % suri(bucket_uri), stdout)
+ _Check1()
+
+ def test_bucket_with_Lb(self):
+ """Tests ls -Lb."""
+ bucket_uri = self.CreateBucket()
+ # Use @Retry as hedge against bucket listing eventual consistency.
+ @Retry(AssertionError, tries=3, timeout_secs=1)
+ def _Check1():
+ stdout = self.RunGsUtil(['ls', '-Lb', suri(bucket_uri)],
+ return_stdout=True)
+ self.assertIn(suri(bucket_uri), stdout)
+ self.assertNotIn('TOTAL:', stdout)
+ _Check1()
+
+ def test_bucket_with_lb(self):
+ """Tests ls -lb."""
+ bucket_uri = self.CreateBucket()
+ # Use @Retry as hedge against bucket listing eventual consistency.
+ @Retry(AssertionError, tries=3, timeout_secs=1)
+ def _Check1():
+ stdout = self.RunGsUtil(['ls', '-lb', suri(bucket_uri)],
+ return_stdout=True)
+ self.assertIn(suri(bucket_uri), stdout)
+ self.assertNotIn('TOTAL:', stdout)
+ _Check1()
+
+ def test_bucket_list_wildcard(self):
+ """Tests listing multiple buckets with a wildcard."""
+ random_prefix = self.MakeRandomTestString()
+ bucket1_name = self.MakeTempName('bucket', prefix=random_prefix)
+ bucket2_name = self.MakeTempName('bucket', prefix=random_prefix)
+ bucket1_uri = self.CreateBucket(bucket_name=bucket1_name)
+ bucket2_uri = self.CreateBucket(bucket_name=bucket2_name)
+ # This just double checks that the common prefix of the two buckets is what
+ # we think it should be (based on implementation detail of CreateBucket).
+ # We want to be careful when setting a wildcard on buckets to make sure we
+ # don't step outside the test buckets to affect other buckets.
+ common_prefix = posixpath.commonprefix([suri(bucket1_uri),
+ suri(bucket2_uri)])
+ self.assertTrue(common_prefix.startswith(
+ '%s://%sgsutil-test-test_bucket_list_wildcard-bucket-' %
+ (self.default_provider, random_prefix)))
+ wildcard = '%s*' % common_prefix
+
+ # Use @Retry as hedge against bucket listing eventual consistency.
+ @Retry(AssertionError, tries=3, timeout_secs=1)
+ def _Check1():
+ stdout = self.RunGsUtil(['ls', '-b', wildcard], return_stdout=True)
+ expected = set([suri(bucket1_uri) + '/', suri(bucket2_uri) + '/'])
+ actual = set(stdout.split())
+ self.assertEqual(expected, actual)
+ _Check1()
+
+ def test_nonexistent_bucket_with_ls(self):
+ """Tests a bucket that is known not to exist."""
+ stderr = self.RunGsUtil(
+ ['ls', '-lb', 'gs://%s' % self.nonexistent_bucket_name],
+ return_stderr=True, expected_status=1)
+ self.assertIn('404', stderr)
+
+ stderr = self.RunGsUtil(
+ ['ls', '-Lb', 'gs://%s' % self.nonexistent_bucket_name],
+ return_stderr=True, expected_status=1)
+ self.assertIn('404', stderr)
+
+ stderr = self.RunGsUtil(
+ ['ls', '-b', 'gs://%s' % self.nonexistent_bucket_name],
+ return_stderr=True, expected_status=1)
+ self.assertIn('404', stderr)
+
+ def test_list_missing_object(self):
+ """Tests listing a non-existent object."""
+ bucket_uri = self.CreateBucket()
+ stderr = self.RunGsUtil(['ls', suri(bucket_uri, 'missing')],
+ return_stderr=True, expected_status=1)
+ self.assertIn('matched no objects', stderr)
+
+ def test_with_one_object(self):
+ bucket_uri = self.CreateBucket()
+ obj_uri = self.CreateObject(bucket_uri=bucket_uri, contents='foo')
+ # Use @Retry as hedge against bucket listing eventual consistency.
+ @Retry(AssertionError, tries=3, timeout_secs=1)
+ def _Check1():
+ stdout = self.RunGsUtil(['ls', suri(bucket_uri)], return_stdout=True)
+ self.assertEqual('%s\n' % obj_uri, stdout)
+ _Check1()
+
+ def test_subdir(self):
+ """Tests listing a bucket subdirectory."""
+ bucket_uri = self.CreateBucket(test_objects=1)
+ k1_uri = bucket_uri.clone_replace_name('foo')
+ k1_uri.set_contents_from_string('baz')
+ k2_uri = bucket_uri.clone_replace_name('dir/foo')
+ k2_uri.set_contents_from_string('bar')
+ # Use @Retry as hedge against bucket listing eventual consistency.
+ @Retry(AssertionError, tries=3, timeout_secs=1)
+ def _Check1():
+ stdout = self.RunGsUtil(['ls', '%s/dir' % suri(bucket_uri)],
+ return_stdout=True)
+ self.assertEqual('%s\n' % suri(k2_uri), stdout)
+ stdout = self.RunGsUtil(['ls', suri(k1_uri)], return_stdout=True)
+ self.assertEqual('%s\n' % suri(k1_uri), stdout)
+ _Check1()
+
+ def test_versioning(self):
+ """Tests listing a versioned bucket."""
+ bucket1_uri = self.CreateBucket(test_objects=1)
+ bucket2_uri = self.CreateVersionedBucket(test_objects=1)
+ self.AssertNObjectsInBucket(bucket1_uri, 1, versioned=True)
+ bucket_list = list(bucket1_uri.list_bucket())
+
+ objuri = [bucket1_uri.clone_replace_key(key).versionless_uri
+ for key in bucket_list][0]
+ self.RunGsUtil(['cp', objuri, suri(bucket2_uri)])
+ self.RunGsUtil(['cp', objuri, suri(bucket2_uri)])
+ # Use @Retry as hedge against bucket listing eventual consistency.
+ @Retry(AssertionError, tries=3, timeout_secs=1)
+ def _Check2():
+ stdout = self.RunGsUtil(['ls', '-a', suri(bucket2_uri)],
+ return_stdout=True)
+ self.assertNumLines(stdout, 3)
+ stdout = self.RunGsUtil(['ls', '-la', suri(bucket2_uri)],
+ return_stdout=True)
+ self.assertIn('%s#' % bucket2_uri.clone_replace_name(bucket_list[0].name),
+ stdout)
+ self.assertIn('metageneration=', stdout)
+ _Check2()
+
+ def test_etag(self):
+ """Tests that listing an object with an etag."""
+ bucket_uri = self.CreateBucket()
+ obj_uri = self.CreateObject(bucket_uri=bucket_uri, contents='foo')
+ # TODO: When testcase setup can use JSON, match against the exact JSON
+ # etag.
+ etag = obj_uri.get_key().etag.strip('"\'')
+ # Use @Retry as hedge against bucket listing eventual consistency.
+ @Retry(AssertionError, tries=3, timeout_secs=1)
+ def _Check1():
+ stdout = self.RunGsUtil(['ls', '-l', suri(bucket_uri)],
+ return_stdout=True)
+ if self.test_api == ApiSelector.XML:
+ self.assertNotIn(etag, stdout)
+ else:
+ self.assertNotIn('etag=', stdout)
+ _Check1()
+
+ def _Check2():
+ stdout = self.RunGsUtil(['ls', '-le', suri(bucket_uri)],
+ return_stdout=True)
+ if self.test_api == ApiSelector.XML:
+ self.assertIn(etag, stdout)
+ else:
+ self.assertIn('etag=', stdout)
+ _Check2()
+
+ def _Check3():
+ stdout = self.RunGsUtil(['ls', '-ale', suri(bucket_uri)],
+ return_stdout=True)
+ if self.test_api == ApiSelector.XML:
+ self.assertIn(etag, stdout)
+ else:
+ self.assertIn('etag=', stdout)
+ _Check3()
+
+ @SkipForS3('S3 bucket configuration values are not supported via ls.')
+ def test_location(self):
+ """Tests listing a bucket with location constraint."""
+ bucket_uri = self.CreateBucket()
+ bucket_suri = suri(bucket_uri)
+
+ # No location info
+ stdout = self.RunGsUtil(['ls', '-lb', bucket_suri],
+ return_stdout=True)
+ self.assertNotIn('Location constraint', stdout)
+
+ # Default location constraint is US
+ stdout = self.RunGsUtil(['ls', '-Lb', bucket_suri],
+ return_stdout=True)
+ self.assertIn('Location constraint:\t\tUS', stdout)
+
+ @SkipForS3('S3 bucket configuration values are not supported via ls.')
+ def test_logging(self):
+ """Tests listing a bucket with logging config."""
+ bucket_uri = self.CreateBucket()
+ bucket_suri = suri(bucket_uri)
+
+ # No logging info
+ stdout = self.RunGsUtil(['ls', '-lb', bucket_suri],
+ return_stdout=True)
+ self.assertNotIn('Logging configuration', stdout)
+
+ # Logging configuration is absent by default
+ stdout = self.RunGsUtil(['ls', '-Lb', bucket_suri],
+ return_stdout=True)
+ self.assertIn('Logging configuration:\t\tNone', stdout)
+
+ # Enable and check
+ self.RunGsUtil(['logging', 'set', 'on', '-b', bucket_suri,
+ bucket_suri])
+ stdout = self.RunGsUtil(['ls', '-Lb', bucket_suri],
+ return_stdout=True)
+ self.assertIn('Logging configuration:\t\tPresent', stdout)
+
+ # Disable and check
+ self.RunGsUtil(['logging', 'set', 'off', bucket_suri])
+ stdout = self.RunGsUtil(['ls', '-Lb', bucket_suri],
+ return_stdout=True)
+ self.assertIn('Logging configuration:\t\tNone', stdout)
+
+ @SkipForS3('S3 bucket configuration values are not supported via ls.')
+ def test_web(self):
+ """Tests listing a bucket with website config."""
+ bucket_uri = self.CreateBucket()
+ bucket_suri = suri(bucket_uri)
+
+ # No website configuration
+ stdout = self.RunGsUtil(['ls', '-lb', bucket_suri],
+ return_stdout=True)
+ self.assertNotIn('Website configuration', stdout)
+
+ # Website configuration is absent by default
+ stdout = self.RunGsUtil(['ls', '-Lb', bucket_suri],
+ return_stdout=True)
+ self.assertIn('Website configuration:\t\tNone', stdout)
+
+ # Initialize and check
+ self.RunGsUtil(['web', 'set', '-m', 'google.com', bucket_suri])
+ stdout = self.RunGsUtil(['ls', '-Lb', bucket_suri],
+ return_stdout=True)
+ self.assertIn('Website configuration:\t\tPresent', stdout)
+
+ # Clear and check
+ self.RunGsUtil(['web', 'set', bucket_suri])
+ stdout = self.RunGsUtil(['ls', '-Lb', bucket_suri],
+ return_stdout=True)
+ self.assertIn('Website configuration:\t\tNone', stdout)
+
+ def test_list_sizes(self):
+ """Tests various size listing options."""
+ bucket_uri = self.CreateBucket()
+ self.CreateObject(bucket_uri=bucket_uri, contents='x' * 2048)
+
+ # Use @Retry as hedge against bucket listing eventual consistency.
+ @Retry(AssertionError, tries=3, timeout_secs=1)
+ def _Check1():
+ stdout = self.RunGsUtil(['ls', '-l', suri(bucket_uri)],
+ return_stdout=True)
+ self.assertIn('2048', stdout)
+ _Check1()
+
+ # Use @Retry as hedge against bucket listing eventual consistency.
+ @Retry(AssertionError, tries=3, timeout_secs=1)
+ def _Check2():
+ stdout = self.RunGsUtil(['ls', '-L', suri(bucket_uri)],
+ return_stdout=True)
+ self.assertIn('2048', stdout)
+ _Check2()
+
+ # Use @Retry as hedge against bucket listing eventual consistency.
+ @Retry(AssertionError, tries=3, timeout_secs=1)
+ def _Check3():
+ stdout = self.RunGsUtil(['ls', '-al', suri(bucket_uri)],
+ return_stdout=True)
+ self.assertIn('2048', stdout)
+ _Check3()
+
+ # Use @Retry as hedge against bucket listing eventual consistency.
+ @Retry(AssertionError, tries=3, timeout_secs=1)
+ def _Check4():
+ stdout = self.RunGsUtil(['ls', '-lh', suri(bucket_uri)],
+ return_stdout=True)
+ self.assertIn('2 KiB', stdout)
+ _Check4()
+
+ # Use @Retry as hedge against bucket listing eventual consistency.
+ @Retry(AssertionError, tries=3, timeout_secs=1)
+ def _Check5():
+ stdout = self.RunGsUtil(['ls', '-alh', suri(bucket_uri)],
+ return_stdout=True)
+ self.assertIn('2 KiB', stdout)
+ _Check5()
+
+ @unittest.skipIf(IS_WINDOWS,
+ 'Unicode handling on Windows requires mods to site-packages')
+ def test_list_unicode_filename(self):
+ """Tests listing an object with a unicode filename."""
+ # Note: This test fails on Windows (command.exe). I was able to get ls to
+ # output Unicode filenames correctly by hacking the UniStream class code
+ # shown at
+ # http://stackoverflow.com/questions/878972/windows-cmd-encoding-change-causes-python-crash/3259271
+ # into the start of gslib/commands/ls.py, along with no-op flush and
+ # isastream functions (as an experiment). However, even with that change,
+ # the current test still fails, since it also needs to run that
+ # stdout/stderr-replacement code. That UniStream class replacement really
+ # needs to be added to the site-packages on Windows python.
+ object_name = u'Аудиоархив'
+ object_name_bytes = object_name.encode(UTF8)
+ bucket_uri = self.CreateVersionedBucket()
+ key_uri = self.CreateObject(bucket_uri=bucket_uri, contents='foo',
+ object_name=object_name)
+ self.AssertNObjectsInBucket(bucket_uri, 1, versioned=True)
+ stdout = self.RunGsUtil(['ls', '-ael', suri(key_uri)],
+ return_stdout=True)
+ self.assertIn(object_name_bytes, stdout)
+ if self.default_provider == 'gs':
+ self.assertIn(str(key_uri.generation), stdout)
+ self.assertIn(
+ 'metageneration=%s' % key_uri.get_key().metageneration, stdout)
+ if self.test_api == ApiSelector.XML:
+ self.assertIn(key_uri.get_key().etag.strip('"\''), stdout)
+ else:
+ # TODO: When testcase setup can use JSON, match against the exact JSON
+ # etag.
+ self.assertIn('etag=', stdout)
+ elif self.default_provider == 's3':
+ self.assertIn(key_uri.version_id, stdout)
+ self.assertIn(key_uri.get_key().etag.strip('"\''), stdout)
+
+ def test_list_gzip_content_length(self):
+ """Tests listing a gzipped object."""
+ file_size = 10000
+ file_contents = 'x' * file_size
+ fpath = self.CreateTempFile(contents=file_contents, file_name='foo.txt')
+ key_uri = self.CreateObject()
+ self.RunGsUtil(['cp', '-z', 'txt', suri(fpath), suri(key_uri)])
+
+ # Use @Retry as hedge against bucket listing eventual consistency.
+ @Retry(AssertionError, tries=3, timeout_secs=1)
+ def _Check1():
+ stdout = self.RunGsUtil(['ls', '-L', suri(key_uri)], return_stdout=True)
+ self.assertRegexpMatches(stdout, r'Content-Encoding:\s+gzip')
+ find_content_length_re = r'Content-Length:\s+(?P<num>\d)'
+ self.assertRegexpMatches(stdout, find_content_length_re)
+ m = re.search(find_content_length_re, stdout)
+ content_length = int(m.group('num'))
+ self.assertGreater(content_length, 0)
+ self.assertLess(content_length, file_size)
+ _Check1()
+
+ def test_output_chopped(self):
+ """Tests that gsutil still succeeds with a truncated stdout."""
+ bucket_uri = self.CreateBucket(test_objects=2)
+
+ # Run Python with the -u flag so output is not buffered.
+ gsutil_cmd = [
+ sys.executable, '-u', gslib.GSUTIL_PATH, 'ls', suri(bucket_uri)]
+ # Set bufsize to 0 to make sure output is not buffered.
+ p = subprocess.Popen(gsutil_cmd, stdout=subprocess.PIPE, bufsize=0)
+ # Immediately close the stdout pipe so that gsutil gets a broken pipe error.
+ p.stdout.close()
+ p.wait()
+ # Make sure it still exited cleanly.
+ self.assertEqual(p.returncode, 0)
+
+ def test_recursive_list_trailing_slash(self):
+ """Tests listing an object with a trailing slash."""
+ bucket_uri = self.CreateBucket()
+ self.CreateObject(bucket_uri=bucket_uri, object_name='/', contents='foo')
+ self.AssertNObjectsInBucket(bucket_uri, 1)
+ stdout = self.RunGsUtil(['ls', '-R', suri(bucket_uri)], return_stdout=True)
+ # Note: The suri function normalizes the URI, so the double slash gets
+ # removed.
+ self.assertIn(suri(bucket_uri) + '/', stdout)
+
+ def test_recursive_list_trailing_two_slash(self):
+ """Tests listing an object with two trailing slashes."""
+ bucket_uri = self.CreateBucket()
+ self.CreateObject(bucket_uri=bucket_uri, object_name='//', contents='foo')
+ self.AssertNObjectsInBucket(bucket_uri, 1)
+ stdout = self.RunGsUtil(['ls', '-R', suri(bucket_uri)], return_stdout=True)
+ # Note: The suri function normalizes the URI, so the double slash gets
+ # removed.
+ self.assertIn(suri(bucket_uri) + '//', stdout)
+
+ @SkipForS3('S3 anonymous access is not supported.')
+ def test_get_object_without_list_bucket_permission(self):
+ # Bucket is not publicly readable by default.
+ bucket_uri = self.CreateBucket()
+ object_uri = self.CreateObject(bucket_uri=bucket_uri,
+ object_name='permitted', contents='foo')
+ # Set this object to be publicly readable.
+ self.RunGsUtil(['acl', 'set', 'public-read', suri(object_uri)])
+ # Drop credentials.
+ with self.SetAnonymousBotoCreds():
+ stdout = self.RunGsUtil(['ls', '-L', suri(object_uri)],
+ return_stdout=True)
+ self.assertIn(suri(object_uri), stdout)
« no previous file with comments | « third_party/gsutil/gslib/tests/test_logging.py ('k') | third_party/gsutil/gslib/tests/test_mb.py » ('j') | no next file with comments »

Powered by Google App Engine
This is Rietveld 408576698