| Index: third_party/gsutil/third_party/boto/boto/logs/layer1.py
|
| diff --git a/third_party/gsutil/third_party/boto/boto/logs/layer1.py b/third_party/gsutil/third_party/boto/boto/logs/layer1.py
|
| new file mode 100644
|
| index 0000000000000000000000000000000000000000..26f7aff7cd23a30b32d6a7c8a7579cfce5c765ed
|
| --- /dev/null
|
| +++ b/third_party/gsutil/third_party/boto/boto/logs/layer1.py
|
| @@ -0,0 +1,576 @@
|
| +# Copyright (c) 2014 Amazon.com, Inc. or its affiliates. All Rights Reserved
|
| +#
|
| +# Permission is hereby granted, free of charge, to any person obtaining a
|
| +# copy of this software and associated documentation files (the
|
| +# "Software"), to deal in the Software without restriction, including
|
| +# without limitation the rights to use, copy, modify, merge, publish, dis-
|
| +# tribute, sublicense, and/or sell copies of the Software, and to permit
|
| +# persons to whom the Software is furnished to do so, subject to the fol-
|
| +# lowing conditions:
|
| +#
|
| +# The above copyright notice and this permission notice shall be included
|
| +# in all copies or substantial portions of the Software.
|
| +#
|
| +# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS
|
| +# OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABIL-
|
| +# ITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT
|
| +# SHALL THE AUTHOR BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY,
|
| +# WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
|
| +# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
|
| +# IN THE SOFTWARE.
|
| +#
|
| +
|
| +import boto
|
| +from boto.connection import AWSQueryConnection
|
| +from boto.regioninfo import RegionInfo
|
| +from boto.exception import JSONResponseError
|
| +from boto.logs import exceptions
|
| +from boto.compat import json
|
| +
|
| +
|
| +class CloudWatchLogsConnection(AWSQueryConnection):
|
| + """
|
| + Amazon CloudWatch Logs Service API Reference
|
| + This is the Amazon CloudWatch Logs API Reference . Amazon
|
| + CloudWatch Logs is a managed service for real time monitoring and
|
| + archival of application logs. This guide provides detailed
|
| + information about Amazon CloudWatch Logs actions, data types,
|
| + parameters, and errors. For detailed information about Amazon
|
| + CloudWatch Logs features and their associated API calls, go to the
|
| + `Amazon CloudWatch Logs Developer Guide`_.
|
| +
|
| + Use the following links to get started using the Amazon CloudWatch
|
| + API Reference :
|
| +
|
| +
|
| + + `Actions`_: An alphabetical list of all Amazon CloudWatch Logs
|
| + actions.
|
| + + `Data Types`_: An alphabetical list of all Amazon CloudWatch
|
| + Logs data types.
|
| + + `Common Parameters`_: Parameters that all Query actions can use.
|
| + + `Common Errors`_: Client and server errors that all actions can
|
| + return.
|
| + + `Regions and Endpoints`_: Itemized regions and endpoints for all
|
| + AWS products.
|
| +
|
| +
|
| + In addition to using the Amazon CloudWatch Logs API, you can also
|
| + use the following SDKs and third-party libraries to access Amazon
|
| + CloudWatch Logs programmatically.
|
| +
|
| +
|
| + + `AWS SDK for Java Documentation`_
|
| + + `AWS SDK for .NET Documentation`_
|
| + + `AWS SDK for PHP Documentation`_
|
| + + `AWS SDK for Ruby Documentation`_
|
| +
|
| +
|
| + Developers in the AWS developer community also provide their own
|
| + libraries, which you can find at the following AWS developer
|
| + centers:
|
| +
|
| +
|
| + + `AWS Java Developer Center`_
|
| + + `AWS PHP Developer Center`_
|
| + + `AWS Python Developer Center`_
|
| + + `AWS Ruby Developer Center`_
|
| + + `AWS Windows and .NET Developer Center`_
|
| + """
|
| + APIVersion = "2014-03-28"
|
| + DefaultRegionName = "us-east-1"
|
| + DefaultRegionEndpoint = "logs.us-east-1.amazonaws.com"
|
| + ServiceName = "CloudWatchLogs"
|
| + TargetPrefix = "Logs_20140328"
|
| + ResponseError = JSONResponseError
|
| +
|
| + _faults = {
|
| + "LimitExceededException": exceptions.LimitExceededException,
|
| + "DataAlreadyAcceptedException": exceptions.DataAlreadyAcceptedException,
|
| + "ResourceInUseException": exceptions.ResourceInUseException,
|
| + "ServiceUnavailableException": exceptions.ServiceUnavailableException,
|
| + "InvalidParameterException": exceptions.InvalidParameterException,
|
| + "ResourceNotFoundException": exceptions.ResourceNotFoundException,
|
| + "ResourceAlreadyExistsException": exceptions.ResourceAlreadyExistsException,
|
| + "OperationAbortedException": exceptions.OperationAbortedException,
|
| + "InvalidSequenceTokenException": exceptions.InvalidSequenceTokenException,
|
| + }
|
| +
|
| + def __init__(self, **kwargs):
|
| + region = kwargs.pop('region', None)
|
| + if not region:
|
| + region = RegionInfo(self, self.DefaultRegionName,
|
| + self.DefaultRegionEndpoint)
|
| +
|
| + if 'host' not in kwargs or kwargs['host'] is None:
|
| + kwargs['host'] = region.endpoint
|
| +
|
| + super(CloudWatchLogsConnection, self).__init__(**kwargs)
|
| + self.region = region
|
| +
|
| + def _required_auth_capability(self):
|
| + return ['hmac-v4']
|
| +
|
| + def create_log_group(self, log_group_name):
|
| + """
|
| + Creates a new log group with the specified name. The name of
|
| + the log group must be unique within a region for an AWS
|
| + account. You can create up to 100 log groups per account.
|
| +
|
| + You must use the following guidelines when naming a log group:
|
| +
|
| + + Log group names can be between 1 and 512 characters long.
|
| + + Allowed characters are az, AZ, 09, '_' (underscore), '-'
|
| + (hyphen), '/' (forward slash), and '.' (period).
|
| +
|
| +
|
| +
|
| + Log groups are created with a default retention of 14 days.
|
| + The retention attribute allow you to configure the number of
|
| + days you want to retain log events in the specified log group.
|
| + See the `SetRetention` operation on how to modify the
|
| + retention of your log groups.
|
| +
|
| + :type log_group_name: string
|
| + :param log_group_name:
|
| +
|
| + """
|
| + params = {'logGroupName': log_group_name, }
|
| + return self.make_request(action='CreateLogGroup',
|
| + body=json.dumps(params))
|
| +
|
| + def create_log_stream(self, log_group_name, log_stream_name):
|
| + """
|
| + Creates a new log stream in the specified log group. The name
|
| + of the log stream must be unique within the log group. There
|
| + is no limit on the number of log streams that can exist in a
|
| + log group.
|
| +
|
| + You must use the following guidelines when naming a log
|
| + stream:
|
| +
|
| + + Log stream names can be between 1 and 512 characters long.
|
| + + The ':' colon character is not allowed.
|
| +
|
| + :type log_group_name: string
|
| + :param log_group_name:
|
| +
|
| + :type log_stream_name: string
|
| + :param log_stream_name:
|
| +
|
| + """
|
| + params = {
|
| + 'logGroupName': log_group_name,
|
| + 'logStreamName': log_stream_name,
|
| + }
|
| + return self.make_request(action='CreateLogStream',
|
| + body=json.dumps(params))
|
| +
|
| + def delete_log_group(self, log_group_name):
|
| + """
|
| + Deletes the log group with the specified name. Amazon
|
| + CloudWatch Logs will delete a log group only if there are no
|
| + log streams and no metric filters associated with the log
|
| + group. If this condition is not satisfied, the request will
|
| + fail and the log group will not be deleted.
|
| +
|
| + :type log_group_name: string
|
| + :param log_group_name:
|
| +
|
| + """
|
| + params = {'logGroupName': log_group_name, }
|
| + return self.make_request(action='DeleteLogGroup',
|
| + body=json.dumps(params))
|
| +
|
| + def delete_log_stream(self, log_group_name, log_stream_name):
|
| + """
|
| + Deletes a log stream and permanently deletes all the archived
|
| + log events associated with it.
|
| +
|
| + :type log_group_name: string
|
| + :param log_group_name:
|
| +
|
| + :type log_stream_name: string
|
| + :param log_stream_name:
|
| +
|
| + """
|
| + params = {
|
| + 'logGroupName': log_group_name,
|
| + 'logStreamName': log_stream_name,
|
| + }
|
| + return self.make_request(action='DeleteLogStream',
|
| + body=json.dumps(params))
|
| +
|
| + def delete_metric_filter(self, log_group_name, filter_name):
|
| + """
|
| + Deletes a metric filter associated with the specified log
|
| + group.
|
| +
|
| + :type log_group_name: string
|
| + :param log_group_name:
|
| +
|
| + :type filter_name: string
|
| + :param filter_name: The name of the metric filter.
|
| +
|
| + """
|
| + params = {
|
| + 'logGroupName': log_group_name,
|
| + 'filterName': filter_name,
|
| + }
|
| + return self.make_request(action='DeleteMetricFilter',
|
| + body=json.dumps(params))
|
| +
|
| + def delete_retention_policy(self, log_group_name):
|
| + """
|
| +
|
| +
|
| + :type log_group_name: string
|
| + :param log_group_name:
|
| +
|
| + """
|
| + params = {'logGroupName': log_group_name, }
|
| + return self.make_request(action='DeleteRetentionPolicy',
|
| + body=json.dumps(params))
|
| +
|
| + def describe_log_groups(self, log_group_name_prefix=None,
|
| + next_token=None, limit=None):
|
| + """
|
| + Returns all the log groups that are associated with the AWS
|
| + account making the request. The list returned in the response
|
| + is ASCII-sorted by log group name.
|
| +
|
| + By default, this operation returns up to 50 log groups. If
|
| + there are more log groups to list, the response would contain
|
| + a `nextToken` value in the response body. You can also limit
|
| + the number of log groups returned in the response by
|
| + specifying the `limit` parameter in the request.
|
| +
|
| + :type log_group_name_prefix: string
|
| + :param log_group_name_prefix:
|
| +
|
| + :type next_token: string
|
| + :param next_token: A string token used for pagination that points to
|
| + the next page of results. It must be a value obtained from the
|
| + response of the previous `DescribeLogGroups` request.
|
| +
|
| + :type limit: integer
|
| + :param limit: The maximum number of items returned in the response. If
|
| + you don't specify a value, the request would return up to 50 items.
|
| +
|
| + """
|
| + params = {}
|
| + if log_group_name_prefix is not None:
|
| + params['logGroupNamePrefix'] = log_group_name_prefix
|
| + if next_token is not None:
|
| + params['nextToken'] = next_token
|
| + if limit is not None:
|
| + params['limit'] = limit
|
| + return self.make_request(action='DescribeLogGroups',
|
| + body=json.dumps(params))
|
| +
|
| + def describe_log_streams(self, log_group_name,
|
| + log_stream_name_prefix=None, next_token=None,
|
| + limit=None):
|
| + """
|
| + Returns all the log streams that are associated with the
|
| + specified log group. The list returned in the response is
|
| + ASCII-sorted by log stream name.
|
| +
|
| + By default, this operation returns up to 50 log streams. If
|
| + there are more log streams to list, the response would contain
|
| + a `nextToken` value in the response body. You can also limit
|
| + the number of log streams returned in the response by
|
| + specifying the `limit` parameter in the request.
|
| +
|
| + :type log_group_name: string
|
| + :param log_group_name:
|
| +
|
| + :type log_stream_name_prefix: string
|
| + :param log_stream_name_prefix:
|
| +
|
| + :type next_token: string
|
| + :param next_token: A string token used for pagination that points to
|
| + the next page of results. It must be a value obtained from the
|
| + response of the previous `DescribeLogStreams` request.
|
| +
|
| + :type limit: integer
|
| + :param limit: The maximum number of items returned in the response. If
|
| + you don't specify a value, the request would return up to 50 items.
|
| +
|
| + """
|
| + params = {'logGroupName': log_group_name, }
|
| + if log_stream_name_prefix is not None:
|
| + params['logStreamNamePrefix'] = log_stream_name_prefix
|
| + if next_token is not None:
|
| + params['nextToken'] = next_token
|
| + if limit is not None:
|
| + params['limit'] = limit
|
| + return self.make_request(action='DescribeLogStreams',
|
| + body=json.dumps(params))
|
| +
|
| + def describe_metric_filters(self, log_group_name,
|
| + filter_name_prefix=None, next_token=None,
|
| + limit=None):
|
| + """
|
| + Returns all the metrics filters associated with the specified
|
| + log group. The list returned in the response is ASCII-sorted
|
| + by filter name.
|
| +
|
| + By default, this operation returns up to 50 metric filters. If
|
| + there are more metric filters to list, the response would
|
| + contain a `nextToken` value in the response body. You can also
|
| + limit the number of metric filters returned in the response by
|
| + specifying the `limit` parameter in the request.
|
| +
|
| + :type log_group_name: string
|
| + :param log_group_name:
|
| +
|
| + :type filter_name_prefix: string
|
| + :param filter_name_prefix: The name of the metric filter.
|
| +
|
| + :type next_token: string
|
| + :param next_token: A string token used for pagination that points to
|
| + the next page of results. It must be a value obtained from the
|
| + response of the previous `DescribeMetricFilters` request.
|
| +
|
| + :type limit: integer
|
| + :param limit: The maximum number of items returned in the response. If
|
| + you don't specify a value, the request would return up to 50 items.
|
| +
|
| + """
|
| + params = {'logGroupName': log_group_name, }
|
| + if filter_name_prefix is not None:
|
| + params['filterNamePrefix'] = filter_name_prefix
|
| + if next_token is not None:
|
| + params['nextToken'] = next_token
|
| + if limit is not None:
|
| + params['limit'] = limit
|
| + return self.make_request(action='DescribeMetricFilters',
|
| + body=json.dumps(params))
|
| +
|
| + def get_log_events(self, log_group_name, log_stream_name,
|
| + start_time=None, end_time=None, next_token=None,
|
| + limit=None, start_from_head=None):
|
| + """
|
| + Retrieves log events from the specified log stream. You can
|
| + provide an optional time range to filter the results on the
|
| + event `timestamp`.
|
| +
|
| + By default, this operation returns as much log events as can
|
| + fit in a response size of 1MB, up to 10,000 log events. The
|
| + response will always include a `nextForwardToken` and a
|
| + `nextBackwardToken` in the response body. You can use any of
|
| + these tokens in subsequent `GetLogEvents` requests to paginate
|
| + through events in either forward or backward direction. You
|
| + can also limit the number of log events returned in the
|
| + response by specifying the `limit` parameter in the request.
|
| +
|
| + :type log_group_name: string
|
| + :param log_group_name:
|
| +
|
| + :type log_stream_name: string
|
| + :param log_stream_name:
|
| +
|
| + :type start_time: long
|
| + :param start_time: A point in time expressed as the number milliseconds
|
| + since Jan 1, 1970 00:00:00 UTC.
|
| +
|
| + :type end_time: long
|
| + :param end_time: A point in time expressed as the number milliseconds
|
| + since Jan 1, 1970 00:00:00 UTC.
|
| +
|
| + :type next_token: string
|
| + :param next_token: A string token used for pagination that points to
|
| + the next page of results. It must be a value obtained from the
|
| + `nextForwardToken` or `nextBackwardToken` fields in the response of
|
| + the previous `GetLogEvents` request.
|
| +
|
| + :type limit: integer
|
| + :param limit: The maximum number of log events returned in the
|
| + response. If you don't specify a value, the request would return as
|
| + much log events as can fit in a response size of 1MB, up to 10,000
|
| + log events.
|
| +
|
| + :type start_from_head: boolean
|
| + :param start_from_head:
|
| +
|
| + """
|
| + params = {
|
| + 'logGroupName': log_group_name,
|
| + 'logStreamName': log_stream_name,
|
| + }
|
| + if start_time is not None:
|
| + params['startTime'] = start_time
|
| + if end_time is not None:
|
| + params['endTime'] = end_time
|
| + if next_token is not None:
|
| + params['nextToken'] = next_token
|
| + if limit is not None:
|
| + params['limit'] = limit
|
| + if start_from_head is not None:
|
| + params['startFromHead'] = start_from_head
|
| + return self.make_request(action='GetLogEvents',
|
| + body=json.dumps(params))
|
| +
|
| + def put_log_events(self, log_group_name, log_stream_name, log_events,
|
| + sequence_token=None):
|
| + """
|
| + Uploads a batch of log events to the specified log stream.
|
| +
|
| + Every PutLogEvents request must include the `sequenceToken`
|
| + obtained from the response of the previous request. An upload
|
| + in a newly created log stream does not require a
|
| + `sequenceToken`.
|
| +
|
| + The batch of events must satisfy the following constraints:
|
| +
|
| + + The maximum batch size is 32,768 bytes, and this size is
|
| + calculated as the sum of all event messages in UTF-8, plus 26
|
| + bytes for each log event.
|
| + + None of the log events in the batch can be more than 2 hours
|
| + in the future.
|
| + + None of the log events in the batch can be older than 14
|
| + days or the retention period of the log group.
|
| + + The log events in the batch must be in chronological ordered
|
| + by their `timestamp`.
|
| + + The maximum number of log events in a batch is 1,000.
|
| +
|
| + :type log_group_name: string
|
| + :param log_group_name:
|
| +
|
| + :type log_stream_name: string
|
| + :param log_stream_name:
|
| +
|
| + :type log_events: list
|
| + :param log_events: A list of events belonging to a log stream.
|
| +
|
| + :type sequence_token: string
|
| + :param sequence_token: A string token that must be obtained from the
|
| + response of the previous `PutLogEvents` request.
|
| +
|
| + """
|
| + params = {
|
| + 'logGroupName': log_group_name,
|
| + 'logStreamName': log_stream_name,
|
| + 'logEvents': log_events,
|
| + }
|
| + if sequence_token is not None:
|
| + params['sequenceToken'] = sequence_token
|
| + return self.make_request(action='PutLogEvents',
|
| + body=json.dumps(params))
|
| +
|
| + def put_metric_filter(self, log_group_name, filter_name, filter_pattern,
|
| + metric_transformations):
|
| + """
|
| + Creates or updates a metric filter and associates it with the
|
| + specified log group. Metric filters allow you to configure
|
| + rules to extract metric data from log events ingested through
|
| + `PutLogEvents` requests.
|
| +
|
| + :type log_group_name: string
|
| + :param log_group_name:
|
| +
|
| + :type filter_name: string
|
| + :param filter_name: The name of the metric filter.
|
| +
|
| + :type filter_pattern: string
|
| + :param filter_pattern:
|
| +
|
| + :type metric_transformations: list
|
| + :param metric_transformations:
|
| +
|
| + """
|
| + params = {
|
| + 'logGroupName': log_group_name,
|
| + 'filterName': filter_name,
|
| + 'filterPattern': filter_pattern,
|
| + 'metricTransformations': metric_transformations,
|
| + }
|
| + return self.make_request(action='PutMetricFilter',
|
| + body=json.dumps(params))
|
| +
|
| + def put_retention_policy(self, log_group_name, retention_in_days):
|
| + """
|
| +
|
| +
|
| + :type log_group_name: string
|
| + :param log_group_name:
|
| +
|
| + :type retention_in_days: integer
|
| + :param retention_in_days: Specifies the number of days you want to
|
| + retain log events in the specified log group. Possible values are:
|
| + 1, 3, 5, 7, 14, 30, 60, 90, 120, 150, 180, 365, 400, 547, 730.
|
| +
|
| + """
|
| + params = {
|
| + 'logGroupName': log_group_name,
|
| + 'retentionInDays': retention_in_days,
|
| + }
|
| + return self.make_request(action='PutRetentionPolicy',
|
| + body=json.dumps(params))
|
| +
|
| + def set_retention(self, log_group_name, retention_in_days):
|
| + """
|
| + Sets the retention of the specified log group. Log groups are
|
| + created with a default retention of 14 days. The retention
|
| + attribute allow you to configure the number of days you want
|
| + to retain log events in the specified log group.
|
| +
|
| + :type log_group_name: string
|
| + :param log_group_name:
|
| +
|
| + :type retention_in_days: integer
|
| + :param retention_in_days: Specifies the number of days you want to
|
| + retain log events in the specified log group. Possible values are:
|
| + 1, 3, 5, 7, 14, 30, 60, 90, 120, 150, 180, 365, 400, 547, 730.
|
| +
|
| + """
|
| + params = {
|
| + 'logGroupName': log_group_name,
|
| + 'retentionInDays': retention_in_days,
|
| + }
|
| + return self.make_request(action='SetRetention',
|
| + body=json.dumps(params))
|
| +
|
| + def test_metric_filter(self, filter_pattern, log_event_messages):
|
| + """
|
| + Tests the filter pattern of a metric filter against a sample
|
| + of log event messages. You can use this operation to validate
|
| + the correctness of a metric filter pattern.
|
| +
|
| + :type filter_pattern: string
|
| + :param filter_pattern:
|
| +
|
| + :type log_event_messages: list
|
| + :param log_event_messages:
|
| +
|
| + """
|
| + params = {
|
| + 'filterPattern': filter_pattern,
|
| + 'logEventMessages': log_event_messages,
|
| + }
|
| + return self.make_request(action='TestMetricFilter',
|
| + body=json.dumps(params))
|
| +
|
| + def make_request(self, action, body):
|
| + headers = {
|
| + 'X-Amz-Target': '%s.%s' % (self.TargetPrefix, action),
|
| + 'Host': self.region.endpoint,
|
| + 'Content-Type': 'application/x-amz-json-1.1',
|
| + 'Content-Length': str(len(body)),
|
| + }
|
| + http_request = self.build_base_http_request(
|
| + method='POST', path='/', auth_path='/', params={},
|
| + headers=headers, data=body)
|
| + response = self._mexe(http_request, sender=None,
|
| + override_num_retries=10)
|
| + response_body = response.read().decode('utf-8')
|
| + boto.log.debug(response_body)
|
| + if response.status == 200:
|
| + if response_body:
|
| + return json.loads(response_body)
|
| + else:
|
| + json_body = json.loads(response_body)
|
| + fault_name = json_body.get('__type', None)
|
| + exception_class = self._faults.get(fault_name, self.ResponseError)
|
| + raise exception_class(response.status, response.reason,
|
| + body=json_body)
|
|
|