| Index: tools/telemetry/third_party/gsutilz/third_party/boto/tests/integration/dynamodb/test_layer2.py
|
| diff --git a/tools/telemetry/third_party/gsutilz/third_party/boto/tests/integration/dynamodb/test_layer2.py b/tools/telemetry/third_party/gsutilz/third_party/boto/tests/integration/dynamodb/test_layer2.py
|
| deleted file mode 100644
|
| index b56562a7151380788e226155c49f0343caff3baf..0000000000000000000000000000000000000000
|
| --- a/tools/telemetry/third_party/gsutilz/third_party/boto/tests/integration/dynamodb/test_layer2.py
|
| +++ /dev/null
|
| @@ -1,496 +0,0 @@
|
| -# Copyright (c) 2012 Mitch Garnaat http://garnaat.org/
|
| -# All rights reserved.
|
| -#
|
| -# Permission is hereby granted, free of charge, to any person obtaining a
|
| -# copy of this software and associated documentation files (the
|
| -# "Software"), to deal in the Software without restriction, including
|
| -# without limitation the rights to use, copy, modify, merge, publish, dis-
|
| -# tribute, sublicense, and/or sell copies of the Software, and to permit
|
| -# persons to whom the Software is furnished to do so, subject to the fol-
|
| -# lowing conditions:
|
| -#
|
| -# The above copyright notice and this permission notice shall be included
|
| -# in all copies or substantial portions of the Software.
|
| -#
|
| -# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS
|
| -# OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABIL-
|
| -# ITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT
|
| -# SHALL THE AUTHOR BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY,
|
| -# WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
|
| -# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
|
| -# IN THE SOFTWARE.
|
| -
|
| -"""
|
| -Tests for Layer2 of Amazon DynamoDB
|
| -"""
|
| -import time
|
| -import uuid
|
| -from decimal import Decimal
|
| -
|
| -from tests.unit import unittest
|
| -from boto.dynamodb.exceptions import DynamoDBKeyNotFoundError
|
| -from boto.dynamodb.exceptions import DynamoDBConditionalCheckFailedError
|
| -from boto.dynamodb.layer2 import Layer2
|
| -from boto.dynamodb.types import get_dynamodb_type, Binary
|
| -from boto.dynamodb.condition import BEGINS_WITH, CONTAINS, GT
|
| -from boto.compat import six, long_type
|
| -
|
| -
|
| -class DynamoDBLayer2Test(unittest.TestCase):
|
| - dynamodb = True
|
| -
|
| - def setUp(self):
|
| - self.dynamodb = Layer2()
|
| - self.hash_key_name = 'forum_name'
|
| - self.hash_key_proto_value = ''
|
| - self.range_key_name = 'subject'
|
| - self.range_key_proto_value = ''
|
| - self.table_name = 'sample_data_%s' % int(time.time())
|
| -
|
| - def create_sample_table(self):
|
| - schema = self.dynamodb.create_schema(
|
| - self.hash_key_name, self.hash_key_proto_value,
|
| - self.range_key_name,
|
| - self.range_key_proto_value)
|
| - table = self.create_table(self.table_name, schema, 5, 5)
|
| - table.refresh(wait_for_active=True)
|
| - return table
|
| -
|
| - def create_table(self, table_name, schema, read_units, write_units):
|
| - result = self.dynamodb.create_table(table_name, schema, read_units, write_units)
|
| - self.addCleanup(self.dynamodb.delete_table, result)
|
| - return result
|
| -
|
| - def test_layer2_basic(self):
|
| - print('--- running Amazon DynamoDB Layer2 tests ---')
|
| - c = self.dynamodb
|
| -
|
| - # First create a schema for the table
|
| - schema = c.create_schema(self.hash_key_name, self.hash_key_proto_value,
|
| - self.range_key_name,
|
| - self.range_key_proto_value)
|
| -
|
| - # Create another schema without a range key
|
| - schema2 = c.create_schema('post_id', '')
|
| -
|
| - # Now create a table
|
| - index = int(time.time())
|
| - table_name = 'test-%d' % index
|
| - read_units = 5
|
| - write_units = 5
|
| - table = self.create_table(table_name, schema, read_units, write_units)
|
| - assert table.name == table_name
|
| - assert table.schema.hash_key_name == self.hash_key_name
|
| - assert table.schema.hash_key_type == get_dynamodb_type(self.hash_key_proto_value)
|
| - assert table.schema.range_key_name == self.range_key_name
|
| - assert table.schema.range_key_type == get_dynamodb_type(self.range_key_proto_value)
|
| - assert table.read_units == read_units
|
| - assert table.write_units == write_units
|
| - assert table.item_count == 0
|
| - assert table.size_bytes == 0
|
| -
|
| - # Create the second table
|
| - table2_name = 'test-%d' % (index + 1)
|
| - table2 = self.create_table(table2_name, schema2, read_units, write_units)
|
| -
|
| - # Wait for table to become active
|
| - table.refresh(wait_for_active=True)
|
| - table2.refresh(wait_for_active=True)
|
| -
|
| - # List tables and make sure new one is there
|
| - table_names = c.list_tables()
|
| - assert table_name in table_names
|
| - assert table2_name in table_names
|
| -
|
| - # Update the tables ProvisionedThroughput
|
| - new_read_units = 10
|
| - new_write_units = 5
|
| - table.update_throughput(new_read_units, new_write_units)
|
| -
|
| - # Wait for table to be updated
|
| - table.refresh(wait_for_active=True)
|
| - assert table.read_units == new_read_units
|
| - assert table.write_units == new_write_units
|
| -
|
| - # Put an item
|
| - item1_key = 'Amazon DynamoDB'
|
| - item1_range = 'DynamoDB Thread 1'
|
| - item1_attrs = {
|
| - 'Message': 'DynamoDB thread 1 message text',
|
| - 'LastPostedBy': 'User A',
|
| - 'Views': 0,
|
| - 'Replies': 0,
|
| - 'Answered': 0,
|
| - 'Public': True,
|
| - 'Tags': set(['index', 'primarykey', 'table']),
|
| - 'LastPostDateTime': '12/9/2011 11:36:03 PM'}
|
| -
|
| - # Test a few corner cases with new_item
|
| -
|
| - # Try supplying a hash_key as an arg and as an item in attrs
|
| - item1_attrs[self.hash_key_name] = 'foo'
|
| - foobar_item = table.new_item(item1_key, item1_range, item1_attrs)
|
| - assert foobar_item.hash_key == item1_key
|
| -
|
| - # Try supplying a range_key as an arg and as an item in attrs
|
| - item1_attrs[self.range_key_name] = 'bar'
|
| - foobar_item = table.new_item(item1_key, item1_range, item1_attrs)
|
| - assert foobar_item.range_key == item1_range
|
| -
|
| - # Try supplying hash and range key in attrs dict
|
| - foobar_item = table.new_item(attrs=item1_attrs)
|
| - assert foobar_item.hash_key == 'foo'
|
| - assert foobar_item.range_key == 'bar'
|
| -
|
| - del item1_attrs[self.hash_key_name]
|
| - del item1_attrs[self.range_key_name]
|
| -
|
| - item1 = table.new_item(item1_key, item1_range, item1_attrs)
|
| - # make sure the put() succeeds
|
| - try:
|
| - item1.put()
|
| - except c.layer1.ResponseError as e:
|
| - raise Exception("Item put failed: %s" % e)
|
| -
|
| - # Try to get an item that does not exist.
|
| - self.assertRaises(DynamoDBKeyNotFoundError,
|
| - table.get_item, 'bogus_key', item1_range)
|
| -
|
| - # Now do a consistent read and check results
|
| - item1_copy = table.get_item(item1_key, item1_range,
|
| - consistent_read=True)
|
| - assert item1_copy.hash_key == item1.hash_key
|
| - assert item1_copy.range_key == item1.range_key
|
| - for attr_name in item1_attrs:
|
| - val = item1_copy[attr_name]
|
| - if isinstance(val, (int, long_type, float, six.string_types)):
|
| - assert val == item1[attr_name]
|
| -
|
| - # Try retrieving only select attributes
|
| - attributes = ['Message', 'Views']
|
| - item1_small = table.get_item(item1_key, item1_range,
|
| - attributes_to_get=attributes,
|
| - consistent_read=True)
|
| - for attr_name in item1_small:
|
| - # The item will include the attributes we asked for as
|
| - # well as the hashkey and rangekey, so filter those out.
|
| - if attr_name not in (item1_small.hash_key_name,
|
| - item1_small.range_key_name):
|
| - assert attr_name in attributes
|
| -
|
| - self.assertTrue(table.has_item(item1_key, range_key=item1_range,
|
| - consistent_read=True))
|
| -
|
| - # Try to delete the item with the wrong Expected value
|
| - expected = {'Views': 1}
|
| - self.assertRaises(DynamoDBConditionalCheckFailedError,
|
| - item1.delete, expected_value=expected)
|
| -
|
| - # Try to delete a value while expecting a non-existant attribute
|
| - expected = {'FooBar': True}
|
| - try:
|
| - item1.delete(expected_value=expected)
|
| - except c.layer1.ResponseError:
|
| - pass
|
| -
|
| - # Now update the existing object
|
| - item1.add_attribute('Replies', 2)
|
| -
|
| - removed_attr = 'Public'
|
| - item1.delete_attribute(removed_attr)
|
| -
|
| - removed_tag = item1_attrs['Tags'].copy().pop()
|
| - item1.delete_attribute('Tags', set([removed_tag]))
|
| -
|
| - replies_by_set = set(['Adam', 'Arnie'])
|
| - item1.put_attribute('RepliesBy', replies_by_set)
|
| - retvals = item1.save(return_values='ALL_OLD')
|
| - # Need more tests here for variations on return_values
|
| - assert 'Attributes' in retvals
|
| -
|
| - # Check for correct updates
|
| - item1_updated = table.get_item(item1_key, item1_range,
|
| - consistent_read=True)
|
| - assert item1_updated['Replies'] == item1_attrs['Replies'] + 2
|
| - self.assertFalse(removed_attr in item1_updated)
|
| - self.assertTrue(removed_tag not in item1_updated['Tags'])
|
| - self.assertTrue('RepliesBy' in item1_updated)
|
| - self.assertTrue(item1_updated['RepliesBy'] == replies_by_set)
|
| -
|
| - # Put a few more items into the table
|
| - item2_key = 'Amazon DynamoDB'
|
| - item2_range = 'DynamoDB Thread 2'
|
| - item2_attrs = {
|
| - 'Message': 'DynamoDB thread 2 message text',
|
| - 'LastPostedBy': 'User A',
|
| - 'Views': 0,
|
| - 'Replies': 0,
|
| - 'Answered': 0,
|
| - 'Tags': set(["index", "primarykey", "table"]),
|
| - 'LastPost2DateTime': '12/9/2011 11:36:03 PM'}
|
| - item2 = table.new_item(item2_key, item2_range, item2_attrs)
|
| - item2.put()
|
| -
|
| - item3_key = 'Amazon S3'
|
| - item3_range = 'S3 Thread 1'
|
| - item3_attrs = {
|
| - 'Message': 'S3 Thread 1 message text',
|
| - 'LastPostedBy': 'User A',
|
| - 'Views': 0,
|
| - 'Replies': 0,
|
| - 'Answered': 0,
|
| - 'Tags': set(['largeobject', 'multipart upload']),
|
| - 'LastPostDateTime': '12/9/2011 11:36:03 PM'
|
| - }
|
| - item3 = table.new_item(item3_key, item3_range, item3_attrs)
|
| - item3.put()
|
| -
|
| - # Put an item into the second table
|
| - table2_item1_key = uuid.uuid4().hex
|
| - table2_item1_attrs = {
|
| - 'DateTimePosted': '25/1/2011 12:34:56 PM',
|
| - 'Text': 'I think boto rocks and so does DynamoDB'
|
| - }
|
| - table2_item1 = table2.new_item(table2_item1_key,
|
| - attrs=table2_item1_attrs)
|
| - table2_item1.put()
|
| -
|
| - # Try a few queries
|
| - items = table.query('Amazon DynamoDB', range_key_condition=BEGINS_WITH('DynamoDB'))
|
| - n = 0
|
| - for item in items:
|
| - n += 1
|
| - assert n == 2
|
| - assert items.consumed_units > 0
|
| -
|
| - items = table.query('Amazon DynamoDB', range_key_condition=BEGINS_WITH('DynamoDB'),
|
| - request_limit=1, max_results=1)
|
| - n = 0
|
| - for item in items:
|
| - n += 1
|
| - assert n == 1
|
| - assert items.consumed_units > 0
|
| -
|
| - # Try a few scans
|
| - items = table.scan()
|
| - n = 0
|
| - for item in items:
|
| - n += 1
|
| - assert n == 3
|
| - assert items.consumed_units > 0
|
| -
|
| - items = table.scan(scan_filter={'Replies': GT(0)})
|
| - n = 0
|
| - for item in items:
|
| - n += 1
|
| - assert n == 1
|
| - assert items.consumed_units > 0
|
| -
|
| - # Test some integer and float attributes
|
| - integer_value = 42
|
| - float_value = 345.678
|
| - item3['IntAttr'] = integer_value
|
| - item3['FloatAttr'] = float_value
|
| -
|
| - # Test booleans
|
| - item3['TrueBoolean'] = True
|
| - item3['FalseBoolean'] = False
|
| -
|
| - # Test some set values
|
| - integer_set = set([1, 2, 3, 4, 5])
|
| - float_set = set([1.1, 2.2, 3.3, 4.4, 5.5])
|
| - mixed_set = set([1, 2, 3.3, 4, 5.555])
|
| - str_set = set(['foo', 'bar', 'fie', 'baz'])
|
| - item3['IntSetAttr'] = integer_set
|
| - item3['FloatSetAttr'] = float_set
|
| - item3['MixedSetAttr'] = mixed_set
|
| - item3['StrSetAttr'] = str_set
|
| - item3.put()
|
| -
|
| - # Now do a consistent read
|
| - item4 = table.get_item(item3_key, item3_range, consistent_read=True)
|
| - assert item4['IntAttr'] == integer_value
|
| - assert item4['FloatAttr'] == float_value
|
| - assert bool(item4['TrueBoolean']) is True
|
| - assert bool(item4['FalseBoolean']) is False
|
| - # The values will not necessarily be in the same order as when
|
| - # we wrote them to the DB.
|
| - for i in item4['IntSetAttr']:
|
| - assert i in integer_set
|
| - for i in item4['FloatSetAttr']:
|
| - assert i in float_set
|
| - for i in item4['MixedSetAttr']:
|
| - assert i in mixed_set
|
| - for i in item4['StrSetAttr']:
|
| - assert i in str_set
|
| -
|
| - # Try a batch get
|
| - batch_list = c.new_batch_list()
|
| - batch_list.add_batch(table, [(item2_key, item2_range),
|
| - (item3_key, item3_range)])
|
| - response = batch_list.submit()
|
| - assert len(response['Responses'][table.name]['Items']) == 2
|
| -
|
| - # Try an empty batch get
|
| - batch_list = c.new_batch_list()
|
| - batch_list.add_batch(table, [])
|
| - response = batch_list.submit()
|
| - assert response == {}
|
| -
|
| - # Try a few batch write operations
|
| - item4_key = 'Amazon S3'
|
| - item4_range = 'S3 Thread 2'
|
| - item4_attrs = {
|
| - 'Message': 'S3 Thread 2 message text',
|
| - 'LastPostedBy': 'User A',
|
| - 'Views': 0,
|
| - 'Replies': 0,
|
| - 'Answered': 0,
|
| - 'Tags': set(['largeobject', 'multipart upload']),
|
| - 'LastPostDateTime': '12/9/2011 11:36:03 PM'
|
| - }
|
| - item5_key = 'Amazon S3'
|
| - item5_range = 'S3 Thread 3'
|
| - item5_attrs = {
|
| - 'Message': 'S3 Thread 3 message text',
|
| - 'LastPostedBy': 'User A',
|
| - 'Views': 0,
|
| - 'Replies': 0,
|
| - 'Answered': 0,
|
| - 'Tags': set(['largeobject', 'multipart upload']),
|
| - 'LastPostDateTime': '12/9/2011 11:36:03 PM'
|
| - }
|
| - item4 = table.new_item(item4_key, item4_range, item4_attrs)
|
| - item5 = table.new_item(item5_key, item5_range, item5_attrs)
|
| - batch_list = c.new_batch_write_list()
|
| - batch_list.add_batch(table, puts=[item4, item5])
|
| - response = batch_list.submit()
|
| - # should really check for unprocessed items
|
| -
|
| - # Do some generator gymnastics
|
| - results = table.scan(scan_filter={'Tags': CONTAINS('table')})
|
| - assert results.scanned_count == 5
|
| - results = table.scan(request_limit=2, max_results=5)
|
| - assert results.count == 2
|
| - for item in results:
|
| - if results.count == 2:
|
| - assert results.remaining == 4
|
| - results.remaining -= 2
|
| - results.next_response()
|
| - else:
|
| - assert results.count == 4
|
| - assert results.remaining in (0, 1)
|
| - assert results.count == 4
|
| - results = table.scan(request_limit=6, max_results=4)
|
| - assert len(list(results)) == 4
|
| - assert results.count == 4
|
| -
|
| - batch_list = c.new_batch_write_list()
|
| - batch_list.add_batch(table, deletes=[(item4_key, item4_range),
|
| - (item5_key, item5_range)])
|
| - response = batch_list.submit()
|
| -
|
| - # Try queries
|
| - results = table.query('Amazon DynamoDB', range_key_condition=BEGINS_WITH('DynamoDB'))
|
| - n = 0
|
| - for item in results:
|
| - n += 1
|
| - assert n == 2
|
| -
|
| - # Try to delete the item with the right Expected value
|
| - expected = {'Views': 0}
|
| - item1.delete(expected_value=expected)
|
| -
|
| - self.assertFalse(table.has_item(item1_key, range_key=item1_range,
|
| - consistent_read=True))
|
| - # Now delete the remaining items
|
| - ret_vals = item2.delete(return_values='ALL_OLD')
|
| - # some additional checks here would be useful
|
| - assert ret_vals['Attributes'][self.hash_key_name] == item2_key
|
| - assert ret_vals['Attributes'][self.range_key_name] == item2_range
|
| -
|
| - item3.delete()
|
| - table2_item1.delete()
|
| - print('--- tests completed ---')
|
| -
|
| - def test_binary_attrs(self):
|
| - c = self.dynamodb
|
| - schema = c.create_schema(self.hash_key_name, self.hash_key_proto_value,
|
| - self.range_key_name,
|
| - self.range_key_proto_value)
|
| - index = int(time.time())
|
| - table_name = 'test-%d' % index
|
| - read_units = 5
|
| - write_units = 5
|
| - table = self.create_table(table_name, schema, read_units, write_units)
|
| - table.refresh(wait_for_active=True)
|
| - item1_key = 'Amazon S3'
|
| - item1_range = 'S3 Thread 1'
|
| - item1_attrs = {
|
| - 'Message': 'S3 Thread 1 message text',
|
| - 'LastPostedBy': 'User A',
|
| - 'Views': 0,
|
| - 'Replies': 0,
|
| - 'Answered': 0,
|
| - 'BinaryData': Binary(b'\x01\x02\x03\x04'),
|
| - 'BinarySequence': set([Binary(b'\x01\x02'), Binary(b'\x03\x04')]),
|
| - 'Tags': set(['largeobject', 'multipart upload']),
|
| - 'LastPostDateTime': '12/9/2011 11:36:03 PM'
|
| - }
|
| - item1 = table.new_item(item1_key, item1_range, item1_attrs)
|
| - item1.put()
|
| -
|
| - retrieved = table.get_item(item1_key, item1_range, consistent_read=True)
|
| - self.assertEqual(retrieved['Message'], 'S3 Thread 1 message text')
|
| - self.assertEqual(retrieved['Views'], 0)
|
| - self.assertEqual(retrieved['Tags'],
|
| - set(['largeobject', 'multipart upload']))
|
| - self.assertEqual(retrieved['BinaryData'], Binary(b'\x01\x02\x03\x04'))
|
| - # Also comparable directly to bytes:
|
| - self.assertEqual(retrieved['BinaryData'], b'\x01\x02\x03\x04')
|
| - self.assertEqual(retrieved['BinarySequence'],
|
| - set([Binary(b'\x01\x02'), Binary(b'\x03\x04')]))
|
| -
|
| - def test_put_decimal_attrs(self):
|
| - self.dynamodb.use_decimals()
|
| - table = self.create_sample_table()
|
| - item = table.new_item('foo', 'bar')
|
| - item['decimalvalue'] = Decimal('1.12345678912345')
|
| - item.put()
|
| - retrieved = table.get_item('foo', 'bar')
|
| - self.assertEqual(retrieved['decimalvalue'], Decimal('1.12345678912345'))
|
| -
|
| - @unittest.skipIf(six.PY3, "skipping lossy_float_conversion test for Python 3.x")
|
| - def test_lossy_float_conversion(self):
|
| - table = self.create_sample_table()
|
| - item = table.new_item('foo', 'bar')
|
| - item['floatvalue'] = 1.12345678912345
|
| - item.put()
|
| - retrieved = table.get_item('foo', 'bar')['floatvalue']
|
| - # Notice how this is not equal to the original value.
|
| - self.assertNotEqual(1.12345678912345, retrieved)
|
| - # Instead, it's truncated:
|
| - self.assertEqual(1.12345678912, retrieved)
|
| -
|
| - def test_large_integers(self):
|
| - # It's not just floating point numbers, large integers
|
| - # can trigger rouding issues.
|
| - self.dynamodb.use_decimals()
|
| - table = self.create_sample_table()
|
| - item = table.new_item('foo', 'bar')
|
| - item['decimalvalue'] = Decimal('129271300103398600')
|
| - item.put()
|
| - retrieved = table.get_item('foo', 'bar')
|
| - self.assertEqual(retrieved['decimalvalue'], Decimal('129271300103398600'))
|
| - # Also comparable directly to an int.
|
| - self.assertEqual(retrieved['decimalvalue'], 129271300103398600)
|
| -
|
| - def test_put_single_letter_attr(self):
|
| - # When an attr is added that is a single letter, if it overlaps with
|
| - # the built-in "types", the decoding used to fall down. Assert that
|
| - # it's now working correctly.
|
| - table = self.create_sample_table()
|
| - item = table.new_item('foo', 'foo1')
|
| - item.put_attribute('b', 4)
|
| - stored = item.save(return_values='UPDATED_NEW')
|
| - self.assertEqual(stored['Attributes'], {'b': 4})
|
|
|