GCP Stackdriver Logging (version v2.*.*)

create_exclusion_in_billing_account#

Creates a new exclusion in a specified parent resource. Only log entries belonging to that resource can be excluded. You can have up to 10 exclusions in a resource.

Parameters

billingAccountsId (required)#

Type: string

$body#

Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.

Type: object

{
"filter" : "Required. An advanced logs filter that matches the log entries to be excluded. By using the sample function, you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:\"resource.type=gcs_bucket severity",
"createTime" : "Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.",
"name" : "Required. A client-assigned identifier, such as \"load-balancer-exclusion\". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.",
"description" : "Optional. A description of this exclusion.",
"disabled" : "Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.",
"updateTime" : "Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions."
}

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

create_exclusion_in_folder#

Creates a new exclusion in a specified parent resource. Only log entries belonging to that resource can be excluded. You can have up to 10 exclusions in a resource.

Parameters

foldersId (required)#

Type: string

$body#

Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.

Type: object

{
"filter" : "Required. An advanced logs filter that matches the log entries to be excluded. By using the sample function, you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:\"resource.type=gcs_bucket severity",
"createTime" : "Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.",
"name" : "Required. A client-assigned identifier, such as \"load-balancer-exclusion\". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.",
"description" : "Optional. A description of this exclusion.",
"disabled" : "Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.",
"updateTime" : "Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions."
}

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

create_exclusion_in_organization#

Creates a new exclusion in a specified parent resource. Only log entries belonging to that resource can be excluded. You can have up to 10 exclusions in a resource.

Parameters

organizationsId (required)#

Type: string

$body#

Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.

Type: object

{
"filter" : "Required. An advanced logs filter that matches the log entries to be excluded. By using the sample function, you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:\"resource.type=gcs_bucket severity",
"createTime" : "Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.",
"name" : "Required. A client-assigned identifier, such as \"load-balancer-exclusion\". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.",
"description" : "Optional. A description of this exclusion.",
"disabled" : "Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.",
"updateTime" : "Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions."
}

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

create_exclusion_in_project#

Creates a new exclusion in a specified parent resource. Only log entries belonging to that resource can be excluded. You can have up to 10 exclusions in a resource.

Parameters

projectsId (required)#

Type: string

$body#

Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.

Type: object

{
"filter" : "Required. An advanced logs filter that matches the log entries to be excluded. By using the sample function, you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:\"resource.type=gcs_bucket severity",
"createTime" : "Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.",
"name" : "Required. A client-assigned identifier, such as \"load-balancer-exclusion\". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.",
"description" : "Optional. A description of this exclusion.",
"disabled" : "Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.",
"updateTime" : "Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions."
}

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

create_metric_in_project#

Creates a logs-based metric.

Parameters

projectsId (required)#

Type: string

$body#

Describes a logs-based metric. The value of the metric is the number of log entries that match a logs filter in a given time interval.Logs-based metric can also be used to extract values from logs and create a a distribution of the values. The distribution records the statistics of the extracted values along with an optional histogram of the values as specified by the bucket options.

Type: object

{
"filter" : "Required. An advanced logs filter which is used to match log entries. Example:\n\"resource.type=gae_app AND severity>=ERROR\"\nThe maximum length of the filter is 20000 characters.",
"labelExtractors" : "Optional. A map from a label key string to an extractor expression which is used to extract data from a log entry field and assign as the label value. Each label key specified in the LabelDescriptor must have an associated extractor expression in this map. The syntax of the extractor expression is the same as for the value_extractor field.The extracted value is converted to the type defined in the label descriptor. If the either the extraction or the type conversion fails, the label will have a default value. The default value for a string label is an empty string, for an integer label its 0, and for a boolean label its false.Note that there are upper bounds on the maximum number of labels and the number of active time series that are allowed in a project.",
"valueExtractor" : "Optional. A value_extractor is required when using a distribution logs-based metric to extract the values to record from a log entry. Two functions are supported for value extraction: EXTRACT(field) or REGEXP_EXTRACT(field, regex). The argument are: 1. field: The name of the log entry field from which the value is to be extracted. 2. regex: A regular expression using the Google RE2 syntax (https://github.com/google/re2/wiki/Syntax) with a single capture group to extract data from the specified log entry field. The value of the field is converted to a string before applying the regex. It is an error to specify a regex that does not include exactly one capture group.The result of the extraction must be convertible to a double type, as the distribution always records double values. If either the extraction or the conversion to double fails, then those values are not recorded in the distribution.Example: REGEXP_EXTRACT(jsonPayload.request, \".*quantity=(\\d+).*\")",
"createTime" : "Output only. The creation timestamp of the metric.This field may not be present for older metrics.",
"metricDescriptor" : {
"monitoredResourceTypes" : [ "string" ],
"unit" : "The units in which the metric value is reported. It is only applicable if the value_type is INT64, DOUBLE, or DISTRIBUTION. The unit defines the representation of the stored metric values.Different systems may scale the values to be more easily displayed (so a value of 0.02KBy might be displayed as 20By, and a value of 3523KBy might be displayed as 3.5MBy). However, if the unit is KBy, then the value of the metric is always in thousands of bytes, no matter how it may be displayed..If you want a custom metric to record the exact number of CPU-seconds used by a job, you can create an INT64 CUMULATIVE metric whose unit is s{CPU} (or equivalently 1s{CPU} or just s). If the job uses 12,005 CPU-seconds, then the value is written as 12005.Alternatively, if you want a custom metric to record data in a more granular way, you can create a DOUBLE CUMULATIVE metric whose unit is ks{CPU}, and then write the value 12.005 (which is 12005/1000), or use Kis{CPU} and write 11.723 (which is 12005/1024).The supported units are a subset of The Unified Code for Units of Measure (http://unitsofmeasure.org/ucum.html) standard:Basic units (UNIT)\nbit bit\nBy byte\ns second\nmin minute\nh hour\nd dayPrefixes (PREFIX)\nk kilo (10^3)\nM mega (10^6)\nG giga (10^9)\nT tera (10^12)\nP peta (10^15)\nE exa (10^18)\nZ zetta (10^21)\nY yotta (10^24)\nm milli (10^-3)\nu micro (10^-6)\nn nano (10^-9)\np pico (10^-12)\nf femto (10^-15)\na atto (10^-18)\nz zepto (10^-21)\ny yocto (10^-24)\nKi kibi (2^10)\nMi mebi (2^20)\nGi gibi (2^30)\nTi tebi (2^40)\nPi pebi (2^50)GrammarThe grammar also includes these connectors:\n/ division or ratio (as an infix operator). For examples, kBy/{email} or MiBy/10ms (although you should almost never have /s in a metric unit; rates should always be computed at query time from the underlying cumulative or delta value).\n. multiplication or composition (as an infix operator). For examples, GBy.d or k{watt}.h.The grammar for a unit is as follows:\nExpression = Component { \".\" Component } { \"/\" Component } ;\n\nComponent = ( [ PREFIX ] UNIT | \"%\" ) [ Annotation ]\n | Annotation\n | \"1\"\n ;\n\nAnnotation = \"{\" NAME \"}\" ;\nNotes:\nAnnotation is just a comment if it follows a UNIT. If the annotation is used alone, then the unit is equivalent to 1. For examples, {request}/s == 1/s, By{transmitted}/s == By/s.\nNAME is a sequence of non-blank printable ASCII characters not containing { or }.\n1 represents a unitary dimensionless unit (https://en.wikipedia.org/wiki/Dimensionless_quantity) of 1, such as in 1/s. It is typically used when none of the basic units are appropriate. For example, \"new users per day\" can be represented as 1/d or {new-users}/d (and a metric value 5 would mean \"5 new users). Alternatively, \"thousands of page views per day\" would be represented as 1000/d or k1/d or k{page_views}/d (and a metric value of 5.3 would mean \"5300 page views per day\").\n% represents dimensionless value of 1/100, and annotates values giving a percentage (so the metric values are typically in the range of 0..100, and a metric value 3 means \"3 percent\").\n10^2.% indicates a metric contains a ratio, typically in the range 0..1, that will be multiplied by 100 and displayed as a percentage (so a metric value 0.03 means \"3 percent\").",
"metadata" : {
"ingestDelay" : "The delay of data points caused by ingestion. Data points older than this age are guaranteed to be ingested and available to be read, excluding data loss due to errors.",
"launchStage" : "Deprecated. Must use the MetricDescriptor.launch_stage instead.",
"samplePeriod" : "The sampling period of metric data points. For metrics which are written periodically, consecutive data points are stored at this time interval, excluding data loss due to errors. Metrics with a higher granularity have a smaller sampling period."
},
"metricKind" : "Whether the metric records instantaneous values, changes to a value, etc. Some combinations of metric_kind and value_type might not be supported.",
"displayName" : "A concise name for the metric, which can be displayed in user interfaces. Use sentence case without an ending period, for example \"Request count\". This field is optional but it is recommended to be set for any metrics associated with user-visible concepts, such as Quota.",
"valueType" : "Whether the measurement is an integer, a floating-point number, etc. Some combinations of metric_kind and value_type might not be supported.",
"name" : "The resource name of the metric descriptor.",
"description" : "A detailed description of the metric, which can be used in documentation.",
"launchStage" : "Optional. The launch stage of the metric definition.",
"type" : "The metric type, including its DNS name prefix. The type is not URL-encoded. All user-defined metric types have the DNS name custom.googleapis.com or external.googleapis.com. Metric types should use a natural hierarchical grouping. For example:\n\"custom.googleapis.com/invoice/paid/amount\"\n\"external.googleapis.com/prometheus/up\"\n\"appengine.googleapis.com/http/server/response_latencies\"\n",
"labels" : [ {
"valueType" : "The type of data that can be assigned to the label.",
"description" : "A human-readable description for the label.",
"key" : "The label key."
} ]
},
"bucketOptions" : {
"explicitBuckets" : {
"bounds" : [ "number" ]
},
"linearBuckets" : {
"offset" : "Lower bound of the first bucket.",
"width" : "Must be greater than 0.",
"numFiniteBuckets" : "Must be greater than 0."
},
"exponentialBuckets" : {
"growthFactor" : "Must be greater than 1.",
"scale" : "Must be greater than 0.",
"numFiniteBuckets" : "Must be greater than 0."
}
},
"name" : "Required. The client-assigned metric identifier. Examples: \"error_count\", \"nginx/requests\".Metric identifiers are limited to 100 characters and can include only the following characters: A-Z, a-z, 0-9, and the special characters _-.,+!*',()%/. The forward-slash character (/) denotes a hierarchy of name pieces, and it cannot be the first character of the name.The metric identifier in this field must not be URL-encoded (https://en.wikipedia.org/wiki/Percent-encoding). However, when the metric identifier appears as the [METRIC_ID] part of a metric_name API parameter, then the metric identifier must be URL-encoded. Example: \"projects/my-project/metrics/nginx%2Frequests\".",
"description" : "Optional. A description of this metric, which is used in documentation. The maximum length of the description is 8000 characters.",
"updateTime" : "Output only. The last update timestamp of the metric.This field may not be present for older metrics.",
"version" : "Deprecated. The API version that created or updated this metric. The v2 format is used by default and cannot be changed."
}

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

create_sink_in_billing_account#

Creates a sink that exports specified log entries to a destination. The export of newly-ingested log entries begins immediately, unless the sink's writer_identity is not permitted to write to the destination. A sink can export log entries only from the resource owning the sink.

Parameters

billingAccountsId (required)#

Type: string

$body#

Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, or a Cloud Pub/Sub topic. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.

Type: object

{
"filter" : "Optional. An advanced logs filter. The only exported log entries are those that are in the resource owning the sink and that match the filter. For example:\nlogName=\"projects/[PROJECT_ID]/logs/[LOG_ID]\" AND severity>=ERROR\n",
"writerIdentity" : "Output only. An IAM identity—a service account or group—under which Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource. Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.",
"createTime" : "Output only. The creation timestamp of the sink.This field may not be present for older sinks.",
"includeChildren" : "Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then logs from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression. For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent. To only export entries from certain child projects, filter on the project part of the log name:\nlogName:(\"projects/test-project1/\" OR \"projects/test-project2/\") AND\nresource.type=gce_instance\n",
"name" : "Required. The client-assigned sink identifier, unique within the project. Example: \"my-syslog-errors-to-pubsub\". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.",
"bigqueryOptions" : {
"usesTimestampColumnPartitioning" : "Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use. All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.",
"usePartitionedTables" : "Optional. Whether to use BigQuery's partition tables. By default, Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax has to be used instead. In both cases, tables are sharded based on UTC timezone."
},
"destination" : "Required. The export destination:\n\"storage.googleapis.com/[GCS_BUCKET]\"\n\"bigquery.googleapis.com/projects/[PROJECT_ID]/datasets/[DATASET]\"\n\"pubsub.googleapis.com/projects/[PROJECT_ID]/topics/[TOPIC_ID]\"\nThe sink's writer_identity, set when the sink is created, must have permission to write to the destination or else the log entries are not exported. For more information, see Exporting Logs with Sinks.",
"description" : "Optional. A description of this sink. The maximum length of the description is 8000 characters.",
"disabled" : "Optional. If set to True, then this sink is disabled and it does not export any log entries.",
"updateTime" : "Output only. The last update timestamp of the sink.This field may not be present for older sinks.",
"outputVersionFormat" : "Deprecated. The log entry format to use for this sink's exported log entries. The v2 format is used by default and cannot be changed."
}

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uniqueWriterIdentity#

Optional. Determines the kind of IAM identity returned as writer_identity in the new sink. If this value is omitted or set to false, and if the sink's parent is a project, then the value returned as writer_identity is the same group or service account used by Logging before the addition of writer identities to this API. The sink's destination must be in the same project as the sink itself.If this field is set to true, or if the sink is owned by a non-project resource such as an organization, then the value of writer_identity will be a unique service account used only for exports from the new sink. For more information, see writer_identity in LogSink.

Type: boolean

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

create_sink_in_folder#

Creates a sink that exports specified log entries to a destination. The export of newly-ingested log entries begins immediately, unless the sink's writer_identity is not permitted to write to the destination. A sink can export log entries only from the resource owning the sink.

Parameters

foldersId (required)#

Type: string

$body#

Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, or a Cloud Pub/Sub topic. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.

Type: object

{
"filter" : "Optional. An advanced logs filter. The only exported log entries are those that are in the resource owning the sink and that match the filter. For example:\nlogName=\"projects/[PROJECT_ID]/logs/[LOG_ID]\" AND severity>=ERROR\n",
"writerIdentity" : "Output only. An IAM identity—a service account or group—under which Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource. Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.",
"createTime" : "Output only. The creation timestamp of the sink.This field may not be present for older sinks.",
"includeChildren" : "Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then logs from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression. For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent. To only export entries from certain child projects, filter on the project part of the log name:\nlogName:(\"projects/test-project1/\" OR \"projects/test-project2/\") AND\nresource.type=gce_instance\n",
"name" : "Required. The client-assigned sink identifier, unique within the project. Example: \"my-syslog-errors-to-pubsub\". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.",
"bigqueryOptions" : {
"usesTimestampColumnPartitioning" : "Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use. All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.",
"usePartitionedTables" : "Optional. Whether to use BigQuery's partition tables. By default, Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax has to be used instead. In both cases, tables are sharded based on UTC timezone."
},
"destination" : "Required. The export destination:\n\"storage.googleapis.com/[GCS_BUCKET]\"\n\"bigquery.googleapis.com/projects/[PROJECT_ID]/datasets/[DATASET]\"\n\"pubsub.googleapis.com/projects/[PROJECT_ID]/topics/[TOPIC_ID]\"\nThe sink's writer_identity, set when the sink is created, must have permission to write to the destination or else the log entries are not exported. For more information, see Exporting Logs with Sinks.",
"description" : "Optional. A description of this sink. The maximum length of the description is 8000 characters.",
"disabled" : "Optional. If set to True, then this sink is disabled and it does not export any log entries.",
"updateTime" : "Output only. The last update timestamp of the sink.This field may not be present for older sinks.",
"outputVersionFormat" : "Deprecated. The log entry format to use for this sink's exported log entries. The v2 format is used by default and cannot be changed."
}

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uniqueWriterIdentity#

Optional. Determines the kind of IAM identity returned as writer_identity in the new sink. If this value is omitted or set to false, and if the sink's parent is a project, then the value returned as writer_identity is the same group or service account used by Logging before the addition of writer identities to this API. The sink's destination must be in the same project as the sink itself.If this field is set to true, or if the sink is owned by a non-project resource such as an organization, then the value of writer_identity will be a unique service account used only for exports from the new sink. For more information, see writer_identity in LogSink.

Type: boolean

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

create_sink_in_organization#

Creates a sink that exports specified log entries to a destination. The export of newly-ingested log entries begins immediately, unless the sink's writer_identity is not permitted to write to the destination. A sink can export log entries only from the resource owning the sink.

Parameters

organizationsId (required)#

Type: string

$body#

Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, or a Cloud Pub/Sub topic. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.

Type: object

{
"filter" : "Optional. An advanced logs filter. The only exported log entries are those that are in the resource owning the sink and that match the filter. For example:\nlogName=\"projects/[PROJECT_ID]/logs/[LOG_ID]\" AND severity>=ERROR\n",
"writerIdentity" : "Output only. An IAM identity—a service account or group—under which Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource. Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.",
"createTime" : "Output only. The creation timestamp of the sink.This field may not be present for older sinks.",
"includeChildren" : "Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then logs from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression. For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent. To only export entries from certain child projects, filter on the project part of the log name:\nlogName:(\"projects/test-project1/\" OR \"projects/test-project2/\") AND\nresource.type=gce_instance\n",
"name" : "Required. The client-assigned sink identifier, unique within the project. Example: \"my-syslog-errors-to-pubsub\". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.",
"bigqueryOptions" : {
"usesTimestampColumnPartitioning" : "Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use. All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.",
"usePartitionedTables" : "Optional. Whether to use BigQuery's partition tables. By default, Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax has to be used instead. In both cases, tables are sharded based on UTC timezone."
},
"destination" : "Required. The export destination:\n\"storage.googleapis.com/[GCS_BUCKET]\"\n\"bigquery.googleapis.com/projects/[PROJECT_ID]/datasets/[DATASET]\"\n\"pubsub.googleapis.com/projects/[PROJECT_ID]/topics/[TOPIC_ID]\"\nThe sink's writer_identity, set when the sink is created, must have permission to write to the destination or else the log entries are not exported. For more information, see Exporting Logs with Sinks.",
"description" : "Optional. A description of this sink. The maximum length of the description is 8000 characters.",
"disabled" : "Optional. If set to True, then this sink is disabled and it does not export any log entries.",
"updateTime" : "Output only. The last update timestamp of the sink.This field may not be present for older sinks.",
"outputVersionFormat" : "Deprecated. The log entry format to use for this sink's exported log entries. The v2 format is used by default and cannot be changed."
}

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uniqueWriterIdentity#

Optional. Determines the kind of IAM identity returned as writer_identity in the new sink. If this value is omitted or set to false, and if the sink's parent is a project, then the value returned as writer_identity is the same group or service account used by Logging before the addition of writer identities to this API. The sink's destination must be in the same project as the sink itself.If this field is set to true, or if the sink is owned by a non-project resource such as an organization, then the value of writer_identity will be a unique service account used only for exports from the new sink. For more information, see writer_identity in LogSink.

Type: boolean

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

create_sink_in_project#

Creates a sink that exports specified log entries to a destination. The export of newly-ingested log entries begins immediately, unless the sink's writer_identity is not permitted to write to the destination. A sink can export log entries only from the resource owning the sink.

Parameters

projectsId (required)#

Type: string

$body#

Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, or a Cloud Pub/Sub topic. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.

Type: object

{
"filter" : "Optional. An advanced logs filter. The only exported log entries are those that are in the resource owning the sink and that match the filter. For example:\nlogName=\"projects/[PROJECT_ID]/logs/[LOG_ID]\" AND severity>=ERROR\n",
"writerIdentity" : "Output only. An IAM identity—a service account or group—under which Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource. Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.",
"createTime" : "Output only. The creation timestamp of the sink.This field may not be present for older sinks.",
"includeChildren" : "Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then logs from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression. For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent. To only export entries from certain child projects, filter on the project part of the log name:\nlogName:(\"projects/test-project1/\" OR \"projects/test-project2/\") AND\nresource.type=gce_instance\n",
"name" : "Required. The client-assigned sink identifier, unique within the project. Example: \"my-syslog-errors-to-pubsub\". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.",
"bigqueryOptions" : {
"usesTimestampColumnPartitioning" : "Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use. All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.",
"usePartitionedTables" : "Optional. Whether to use BigQuery's partition tables. By default, Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax has to be used instead. In both cases, tables are sharded based on UTC timezone."
},
"destination" : "Required. The export destination:\n\"storage.googleapis.com/[GCS_BUCKET]\"\n\"bigquery.googleapis.com/projects/[PROJECT_ID]/datasets/[DATASET]\"\n\"pubsub.googleapis.com/projects/[PROJECT_ID]/topics/[TOPIC_ID]\"\nThe sink's writer_identity, set when the sink is created, must have permission to write to the destination or else the log entries are not exported. For more information, see Exporting Logs with Sinks.",
"description" : "Optional. A description of this sink. The maximum length of the description is 8000 characters.",
"disabled" : "Optional. If set to True, then this sink is disabled and it does not export any log entries.",
"updateTime" : "Output only. The last update timestamp of the sink.This field may not be present for older sinks.",
"outputVersionFormat" : "Deprecated. The log entry format to use for this sink's exported log entries. The v2 format is used by default and cannot be changed."
}

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uniqueWriterIdentity#

Optional. Determines the kind of IAM identity returned as writer_identity in the new sink. If this value is omitted or set to false, and if the sink's parent is a project, then the value returned as writer_identity is the same group or service account used by Logging before the addition of writer identities to this API. The sink's destination must be in the same project as the sink itself.If this field is set to true, or if the sink is owned by a non-project resource such as an organization, then the value of writer_identity will be a unique service account used only for exports from the new sink. For more information, see writer_identity in LogSink.

Type: boolean

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

delete_exclusion_in_billing_account#

Deletes an exclusion.

Parameters

billingAccountsId (required)#

Type: string

exclusionsId (required)#

Type: string

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

delete_exclusion_in_folder#

Deletes an exclusion.

Parameters

exclusionsId (required)#

Type: string

foldersId (required)#

Type: string

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

delete_exclusion_in_organization#

Deletes an exclusion.

Parameters

exclusionsId (required)#

Type: string

organizationsId (required)#

Type: string

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

delete_exclusion_in_project#

Deletes an exclusion.

Parameters

exclusionsId (required)#

Type: string

projectsId (required)#

Type: string

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

delete_log_in_billing_account#

Deletes all the log entries in a log. The log reappears if it receives new entries. Log entries written shortly before the delete operation might not be deleted. Entries received after the delete operation with a timestamp before the operation will be deleted.

Parameters

billingAccountsId (required)#

Type: string

logsId (required)#

Type: string

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

delete_log_in_folder#

Deletes all the log entries in a log. The log reappears if it receives new entries. Log entries written shortly before the delete operation might not be deleted. Entries received after the delete operation with a timestamp before the operation will be deleted.

Parameters

foldersId (required)#

Type: string

logsId (required)#

Type: string

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

delete_log_in_organization#

Deletes all the log entries in a log. The log reappears if it receives new entries. Log entries written shortly before the delete operation might not be deleted. Entries received after the delete operation with a timestamp before the operation will be deleted.

Parameters

logsId (required)#

Type: string

organizationsId (required)#

Type: string

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

delete_log_in_project#

Deletes all the log entries in a log. The log reappears if it receives new entries. Log entries written shortly before the delete operation might not be deleted. Entries received after the delete operation with a timestamp before the operation will be deleted.

Parameters

logsId (required)#

Type: string

projectsId (required)#

Type: string

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

delete_metric_in_project#

Deletes a logs-based metric.

Parameters

metricsId (required)#

Type: string

projectsId (required)#

Type: string

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

delete_sink_in_billing_account#

Deletes a sink. If the sink has a unique writer_identity, then that service account is also deleted.

Parameters

billingAccountsId (required)#

Type: string

sinksId (required)#

Type: string

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

delete_sink_in_folder#

Deletes a sink. If the sink has a unique writer_identity, then that service account is also deleted.

Parameters

foldersId (required)#

Type: string

sinksId (required)#

Type: string

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

delete_sink_in_organization#

Deletes a sink. If the sink has a unique writer_identity, then that service account is also deleted.

Parameters

organizationsId (required)#

Type: string

sinksId (required)#

Type: string

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

delete_sink_in_project#

Deletes a sink. If the sink has a unique writer_identity, then that service account is also deleted.

Parameters

projectsId (required)#

Type: string

sinksId (required)#

Type: string

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

get_bucket_in_billing_account#

Gets a bucket (Beta).

Parameters

billingAccountsId (required)#

Type: string

bucketsId (required)#

Type: string

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

get_bucket_in_location_in_folder#

Gets a bucket (Beta).

Parameters

bucketsId (required)#

Type: string

foldersId (required)#

Type: string

locationsId (required)#

Type: string

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

get_bucket_in_location_in_organization#

Gets a bucket (Beta).

Parameters

bucketsId (required)#

Type: string

locationsId (required)#

Type: string

organizationsId (required)#

Type: string

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

get_bucket_in_location_in_project#

Gets a bucket (Beta).

Parameters

bucketsId (required)#

Type: string

locationsId (required)#

Type: string

projectsId (required)#

Type: string

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

get_cmek_setting_in_organization#

Gets the Logs Router CMEK settings for the given resource.Note: CMEK for the Logs Router can currently only be configured for GCP organizations. Once configured, it applies to all projects and folders in the GCP organization.See Enabling CMEK for Logs Router for more information.

Parameters

organizationsId (required)#

Type: string

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

get_exclusion_in_billing_account#

Gets the description of an exclusion.

Parameters

billingAccountsId (required)#

Type: string

exclusionsId (required)#

Type: string

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

get_exclusion_in_folder#

Gets the description of an exclusion.

Parameters

exclusionsId (required)#

Type: string

foldersId (required)#

Type: string

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

get_exclusion_in_organization#

Gets the description of an exclusion.

Parameters

exclusionsId (required)#

Type: string

organizationsId (required)#

Type: string

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

get_exclusion_in_project#

Gets the description of an exclusion.

Parameters

exclusionsId (required)#

Type: string

projectsId (required)#

Type: string

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

get_metric_in_project#

Gets a logs-based metric.

Parameters

metricsId (required)#

Type: string

projectsId (required)#

Type: string

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

get_sink_in_billing_account#

Gets a sink.

Parameters

billingAccountsId (required)#

Type: string

sinksId (required)#

Type: string

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

get_sink_in_folder#

Gets a sink.

Parameters

foldersId (required)#

Type: string

sinksId (required)#

Type: string

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

get_sink_in_organization#

Gets a sink.

Parameters

organizationsId (required)#

Type: string

sinksId (required)#

Type: string

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

get_sink_in_project#

Gets a sink.

Parameters

projectsId (required)#

Type: string

sinksId (required)#

Type: string

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

list_buckets_in_locations_in_billing_account#

Lists buckets (Beta).

Parameters

billingAccountsId (required)#

Type: string

locationsId (required)#

Type: string

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

list_buckets_in_locations_in_folder#

Lists buckets (Beta).

Parameters

foldersId (required)#

Type: string

locationsId (required)#

Type: string

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

list_buckets_in_locations_in_organization#

Lists buckets (Beta).

Parameters

locationsId (required)#

Type: string

organizationsId (required)#

Type: string

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

list_buckets_in_locations_in_project#

Lists buckets (Beta).

Parameters

locationsId (required)#

Type: string

projectsId (required)#

Type: string

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

list_entries#

Lists log entries. Use this method to retrieve log entries that originated from a project/folder/organization/billing account. For ways to export log entries, see Exporting Logs.

Parameters

$body#

The parameters to ListLogEntries.

Type: object

{
"filter" : "Optional. A filter that chooses which log entries to return. See Advanced Logs Queries. Only log entries that match the filter are returned. An empty filter matches all log entries in the resources listed in resource_names. Referencing a parent resource that is not listed in resource_names will cause the filter to return no results. The maximum length of the filter is 20000 characters.",
"resourceNames" : [ "string" ],
"orderBy" : "Optional. How the results should be sorted. Presently, the only permitted values are \"timestamp asc\" (default) and \"timestamp desc\". The first option returns entries in order of increasing values of LogEntry.timestamp (oldest first), and the second option returns entries in order of decreasing timestamps (newest first). Entries with equal timestamps are returned in order of their insert_id values.",
"pageSize" : "Optional. The maximum number of results to return from this request. Non-positive values are ignored. The presence of next_page_token in the response indicates that more results might be available.",
"projectIds" : [ "string" ],
"pageToken" : "Optional. If present, then retrieve the next batch of results from the preceding call to this method. page_token must be the value of next_page_token from the previous response. The values of other method parameters should be identical to those in the previous call."
}

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

list_exclusions_in_billing_account#

Lists all the exclusions in a parent resource.

Parameters

billingAccountsId (required)#

Type: string

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

list_exclusions_in_folder#

Lists all the exclusions in a parent resource.

Parameters

foldersId (required)#

Type: string

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

list_exclusions_in_organization#

Lists all the exclusions in a parent resource.

Parameters

organizationsId (required)#

Type: string

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

list_exclusions_in_project#

Lists all the exclusions in a parent resource.

Parameters

projectsId (required)#

Type: string

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

list_logs_in_billing_account#

Lists the logs in projects, organizations, folders, or billing accounts. Only logs that have entries are listed.

Parameters

billingAccountsId (required)#

Type: string

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

list_logs_in_folder#

Lists the logs in projects, organizations, folders, or billing accounts. Only logs that have entries are listed.

Parameters

foldersId (required)#

Type: string

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

list_logs_in_organization#

Lists the logs in projects, organizations, folders, or billing accounts. Only logs that have entries are listed.

Parameters

organizationsId (required)#

Type: string

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

list_logs_in_project#

Lists the logs in projects, organizations, folders, or billing accounts. Only logs that have entries are listed.

Parameters

projectsId (required)#

Type: string

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

list_metrics_in_project#

Lists logs-based metrics.

Parameters

projectsId (required)#

Type: string

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

list_monitored_resource_descriptors#

Lists the descriptors for monitored resource types used by Logging.

Parameters

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

list_sinks_in_billing_account#

Lists sinks.

Parameters

billingAccountsId (required)#

Type: string

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

list_sinks_in_folder#

Lists sinks.

Parameters

foldersId (required)#

Type: string

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

list_sinks_in_organization#

Lists sinks.

Parameters

organizationsId (required)#

Type: string

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

list_sinks_in_project#

Lists sinks.

Parameters

projectsId (required)#

Type: string

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

patch_bucket_in_location_in_billing_account#

Updates a bucket. This method replaces the following fields in the existing bucket with values from the new bucket: retention_periodIf the retention period is decreased and the bucket is locked, FAILED_PRECONDITION will be returned.If the bucket has a LifecycleState of DELETE_REQUESTED, FAILED_PRECONDITION will be returned.A buckets region may not be modified after it is created. This method is in Beta.

Parameters

billingAccountsId (required)#

Type: string

bucketsId (required)#

Type: string

locationsId (required)#

Type: string

$body#

Describes a repository of logs (Beta).

Type: object

{
"lifecycleState" : "Output only. The bucket lifecycle state.",
"createTime" : "Output only. The creation timestamp of the bucket. This is not set for any of the default buckets.",
"retentionDays" : "Logs will be retained by default for this amount of time, after which they will automatically be deleted. The minimum retention period is 1 day. If this value is set to zero at bucket creation time, the default time of 30 days will be used.",
"name" : "The resource name of the bucket. For example: \"projects/my-project-id/locations/my-location/buckets/my-bucket-id The supported locations are: \"global\" \"us-central1\"For the location of global it is unspecified where logs are actually stored. Once a bucket has been created, the location can not be changed.",
"description" : "Describes this bucket.",
"updateTime" : "Output only. The last update timestamp of the bucket."
}

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

updateMask#

Required. Field mask that specifies the fields in bucket that need an update. A bucket field will be overwritten if, and only if, it is in the update mask. name and output only fields cannot be updated.For a detailed FieldMask definition, see https://developers.google.com/protocol-buffers/docs/reference/google.protobuf#google.protobuf.FieldMaskExample: updateMask=retention_days.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

patch_bucket_in_location_in_folder#

Updates a bucket. This method replaces the following fields in the existing bucket with values from the new bucket: retention_periodIf the retention period is decreased and the bucket is locked, FAILED_PRECONDITION will be returned.If the bucket has a LifecycleState of DELETE_REQUESTED, FAILED_PRECONDITION will be returned.A buckets region may not be modified after it is created. This method is in Beta.

Parameters

bucketsId (required)#

Type: string

foldersId (required)#

Type: string

locationsId (required)#

Type: string

$body#

Describes a repository of logs (Beta).

Type: object

{
"lifecycleState" : "Output only. The bucket lifecycle state.",
"createTime" : "Output only. The creation timestamp of the bucket. This is not set for any of the default buckets.",
"retentionDays" : "Logs will be retained by default for this amount of time, after which they will automatically be deleted. The minimum retention period is 1 day. If this value is set to zero at bucket creation time, the default time of 30 days will be used.",
"name" : "The resource name of the bucket. For example: \"projects/my-project-id/locations/my-location/buckets/my-bucket-id The supported locations are: \"global\" \"us-central1\"For the location of global it is unspecified where logs are actually stored. Once a bucket has been created, the location can not be changed.",
"description" : "Describes this bucket.",
"updateTime" : "Output only. The last update timestamp of the bucket."
}

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

updateMask#

Required. Field mask that specifies the fields in bucket that need an update. A bucket field will be overwritten if, and only if, it is in the update mask. name and output only fields cannot be updated.For a detailed FieldMask definition, see https://developers.google.com/protocol-buffers/docs/reference/google.protobuf#google.protobuf.FieldMaskExample: updateMask=retention_days.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

patch_bucket_in_location_in_organization#

Updates a bucket. This method replaces the following fields in the existing bucket with values from the new bucket: retention_periodIf the retention period is decreased and the bucket is locked, FAILED_PRECONDITION will be returned.If the bucket has a LifecycleState of DELETE_REQUESTED, FAILED_PRECONDITION will be returned.A buckets region may not be modified after it is created. This method is in Beta.

Parameters

bucketsId (required)#

Type: string

locationsId (required)#

Type: string

organizationsId (required)#

Type: string

$body#

Describes a repository of logs (Beta).

Type: object

{
"lifecycleState" : "Output only. The bucket lifecycle state.",
"createTime" : "Output only. The creation timestamp of the bucket. This is not set for any of the default buckets.",
"retentionDays" : "Logs will be retained by default for this amount of time, after which they will automatically be deleted. The minimum retention period is 1 day. If this value is set to zero at bucket creation time, the default time of 30 days will be used.",
"name" : "The resource name of the bucket. For example: \"projects/my-project-id/locations/my-location/buckets/my-bucket-id The supported locations are: \"global\" \"us-central1\"For the location of global it is unspecified where logs are actually stored. Once a bucket has been created, the location can not be changed.",
"description" : "Describes this bucket.",
"updateTime" : "Output only. The last update timestamp of the bucket."
}

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

updateMask#

Required. Field mask that specifies the fields in bucket that need an update. A bucket field will be overwritten if, and only if, it is in the update mask. name and output only fields cannot be updated.For a detailed FieldMask definition, see https://developers.google.com/protocol-buffers/docs/reference/google.protobuf#google.protobuf.FieldMaskExample: updateMask=retention_days.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

patch_bucket_in_location_in_project#

Updates a bucket. This method replaces the following fields in the existing bucket with values from the new bucket: retention_periodIf the retention period is decreased and the bucket is locked, FAILED_PRECONDITION will be returned.If the bucket has a LifecycleState of DELETE_REQUESTED, FAILED_PRECONDITION will be returned.A buckets region may not be modified after it is created. This method is in Beta.

Parameters

bucketsId (required)#

Type: string

locationsId (required)#

Type: string

projectsId (required)#

Type: string

$body#

Describes a repository of logs (Beta).

Type: object

{
"lifecycleState" : "Output only. The bucket lifecycle state.",
"createTime" : "Output only. The creation timestamp of the bucket. This is not set for any of the default buckets.",
"retentionDays" : "Logs will be retained by default for this amount of time, after which they will automatically be deleted. The minimum retention period is 1 day. If this value is set to zero at bucket creation time, the default time of 30 days will be used.",
"name" : "The resource name of the bucket. For example: \"projects/my-project-id/locations/my-location/buckets/my-bucket-id The supported locations are: \"global\" \"us-central1\"For the location of global it is unspecified where logs are actually stored. Once a bucket has been created, the location can not be changed.",
"description" : "Describes this bucket.",
"updateTime" : "Output only. The last update timestamp of the bucket."
}

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

updateMask#

Required. Field mask that specifies the fields in bucket that need an update. A bucket field will be overwritten if, and only if, it is in the update mask. name and output only fields cannot be updated.For a detailed FieldMask definition, see https://developers.google.com/protocol-buffers/docs/reference/google.protobuf#google.protobuf.FieldMaskExample: updateMask=retention_days.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

patch_exclusion_in_billing_account#

Changes one or more properties of an existing exclusion.

Parameters

billingAccountsId (required)#

Type: string

exclusionsId (required)#

Type: string

$body#

Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.

Type: object

{
"filter" : "Required. An advanced logs filter that matches the log entries to be excluded. By using the sample function, you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:\"resource.type=gcs_bucket severity",
"createTime" : "Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.",
"name" : "Required. A client-assigned identifier, such as \"load-balancer-exclusion\". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.",
"description" : "Optional. A description of this exclusion.",
"disabled" : "Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.",
"updateTime" : "Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions."
}

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

updateMask#

Required. A non-empty list of fields to change in the existing exclusion. New values for the fields are taken from the corresponding fields in the LogExclusion included in this request. Fields not mentioned in update_mask are not changed and are ignored in the request.For example, to change the filter and description of an exclusion, specify an update_mask of "filter,description".

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

patch_exclusion_in_folder#

Changes one or more properties of an existing exclusion.

Parameters

exclusionsId (required)#

Type: string

foldersId (required)#

Type: string

$body#

Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.

Type: object

{
"filter" : "Required. An advanced logs filter that matches the log entries to be excluded. By using the sample function, you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:\"resource.type=gcs_bucket severity",
"createTime" : "Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.",
"name" : "Required. A client-assigned identifier, such as \"load-balancer-exclusion\". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.",
"description" : "Optional. A description of this exclusion.",
"disabled" : "Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.",
"updateTime" : "Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions."
}

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

updateMask#

Required. A non-empty list of fields to change in the existing exclusion. New values for the fields are taken from the corresponding fields in the LogExclusion included in this request. Fields not mentioned in update_mask are not changed and are ignored in the request.For example, to change the filter and description of an exclusion, specify an update_mask of "filter,description".

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

patch_exclusion_in_organization#

Changes one or more properties of an existing exclusion.

Parameters

exclusionsId (required)#

Type: string

organizationsId (required)#

Type: string

$body#

Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.

Type: object

{
"filter" : "Required. An advanced logs filter that matches the log entries to be excluded. By using the sample function, you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:\"resource.type=gcs_bucket severity",
"createTime" : "Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.",
"name" : "Required. A client-assigned identifier, such as \"load-balancer-exclusion\". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.",
"description" : "Optional. A description of this exclusion.",
"disabled" : "Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.",
"updateTime" : "Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions."
}

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

updateMask#

Required. A non-empty list of fields to change in the existing exclusion. New values for the fields are taken from the corresponding fields in the LogExclusion included in this request. Fields not mentioned in update_mask are not changed and are ignored in the request.For example, to change the filter and description of an exclusion, specify an update_mask of "filter,description".

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

patch_exclusion_in_project#

Changes one or more properties of an existing exclusion.

Parameters

exclusionsId (required)#

Type: string

projectsId (required)#

Type: string

$body#

Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.

Type: object

{
"filter" : "Required. An advanced logs filter that matches the log entries to be excluded. By using the sample function, you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:\"resource.type=gcs_bucket severity",
"createTime" : "Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.",
"name" : "Required. A client-assigned identifier, such as \"load-balancer-exclusion\". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.",
"description" : "Optional. A description of this exclusion.",
"disabled" : "Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.",
"updateTime" : "Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions."
}

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

updateMask#

Required. A non-empty list of fields to change in the existing exclusion. New values for the fields are taken from the corresponding fields in the LogExclusion included in this request. Fields not mentioned in update_mask are not changed and are ignored in the request.For example, to change the filter and description of an exclusion, specify an update_mask of "filter,description".

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

patch_sink_in_billing_account#

Updates a sink. This method replaces the following fields in the existing sink with values from the new sink: destination, and filter.The updated sink might also have a new writer_identity; see the unique_writer_identity field.

Parameters

billingAccountsId (required)#

Type: string

sinksId (required)#

Type: string

$body#

Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, or a Cloud Pub/Sub topic. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.

Type: object

{
"filter" : "Optional. An advanced logs filter. The only exported log entries are those that are in the resource owning the sink and that match the filter. For example:\nlogName=\"projects/[PROJECT_ID]/logs/[LOG_ID]\" AND severity>=ERROR\n",
"writerIdentity" : "Output only. An IAM identity—a service account or group—under which Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource. Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.",
"createTime" : "Output only. The creation timestamp of the sink.This field may not be present for older sinks.",
"includeChildren" : "Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then logs from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression. For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent. To only export entries from certain child projects, filter on the project part of the log name:\nlogName:(\"projects/test-project1/\" OR \"projects/test-project2/\") AND\nresource.type=gce_instance\n",
"name" : "Required. The client-assigned sink identifier, unique within the project. Example: \"my-syslog-errors-to-pubsub\". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.",
"bigqueryOptions" : {
"usesTimestampColumnPartitioning" : "Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use. All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.",
"usePartitionedTables" : "Optional. Whether to use BigQuery's partition tables. By default, Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax has to be used instead. In both cases, tables are sharded based on UTC timezone."
},
"destination" : "Required. The export destination:\n\"storage.googleapis.com/[GCS_BUCKET]\"\n\"bigquery.googleapis.com/projects/[PROJECT_ID]/datasets/[DATASET]\"\n\"pubsub.googleapis.com/projects/[PROJECT_ID]/topics/[TOPIC_ID]\"\nThe sink's writer_identity, set when the sink is created, must have permission to write to the destination or else the log entries are not exported. For more information, see Exporting Logs with Sinks.",
"description" : "Optional. A description of this sink. The maximum length of the description is 8000 characters.",
"disabled" : "Optional. If set to True, then this sink is disabled and it does not export any log entries.",
"updateTime" : "Output only. The last update timestamp of the sink.This field may not be present for older sinks.",
"outputVersionFormat" : "Deprecated. The log entry format to use for this sink's exported log entries. The v2 format is used by default and cannot be changed."
}

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uniqueWriterIdentity#

Optional. See sinks.create for a description of this field. When updating a sink, the effect of this field on the value of writer_identity in the updated sink depends on both the old and new values of this field:
If the old and new values of this field are both false or both true, then there is no change to the sink's writer_identity.
If the old value is false and the new value is true, then writer_identity is changed to a unique service account.
It is an error if the old value is true and the new value is set to false or defaulted to false.

Type: boolean

updateMask#

Optional. Field mask that specifies the fields in sink that need an update. A sink field will be overwritten if, and only if, it is in the update mask. name and output only fields cannot be updated.An empty updateMask is temporarily treated as using the following mask for backwards compatibility purposes: destination,filter,includeChildren At some point in the future, behavior will be removed and specifying an empty updateMask will be an error.For a detailed FieldMask definition, see https://developers.google.com/protocol-buffers/docs/reference/google.protobuf#google.protobuf.FieldMaskExample: updateMask=filter.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

patch_sink_in_folder#

Updates a sink. This method replaces the following fields in the existing sink with values from the new sink: destination, and filter.The updated sink might also have a new writer_identity; see the unique_writer_identity field.

Parameters

foldersId (required)#

Type: string

sinksId (required)#

Type: string

$body#

Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, or a Cloud Pub/Sub topic. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.

Type: object

{
"filter" : "Optional. An advanced logs filter. The only exported log entries are those that are in the resource owning the sink and that match the filter. For example:\nlogName=\"projects/[PROJECT_ID]/logs/[LOG_ID]\" AND severity>=ERROR\n",
"writerIdentity" : "Output only. An IAM identity—a service account or group—under which Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource. Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.",
"createTime" : "Output only. The creation timestamp of the sink.This field may not be present for older sinks.",
"includeChildren" : "Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then logs from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression. For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent. To only export entries from certain child projects, filter on the project part of the log name:\nlogName:(\"projects/test-project1/\" OR \"projects/test-project2/\") AND\nresource.type=gce_instance\n",
"name" : "Required. The client-assigned sink identifier, unique within the project. Example: \"my-syslog-errors-to-pubsub\". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.",
"bigqueryOptions" : {
"usesTimestampColumnPartitioning" : "Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use. All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.",
"usePartitionedTables" : "Optional. Whether to use BigQuery's partition tables. By default, Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax has to be used instead. In both cases, tables are sharded based on UTC timezone."
},
"destination" : "Required. The export destination:\n\"storage.googleapis.com/[GCS_BUCKET]\"\n\"bigquery.googleapis.com/projects/[PROJECT_ID]/datasets/[DATASET]\"\n\"pubsub.googleapis.com/projects/[PROJECT_ID]/topics/[TOPIC_ID]\"\nThe sink's writer_identity, set when the sink is created, must have permission to write to the destination or else the log entries are not exported. For more information, see Exporting Logs with Sinks.",
"description" : "Optional. A description of this sink. The maximum length of the description is 8000 characters.",
"disabled" : "Optional. If set to True, then this sink is disabled and it does not export any log entries.",
"updateTime" : "Output only. The last update timestamp of the sink.This field may not be present for older sinks.",
"outputVersionFormat" : "Deprecated. The log entry format to use for this sink's exported log entries. The v2 format is used by default and cannot be changed."
}

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uniqueWriterIdentity#

Optional. See sinks.create for a description of this field. When updating a sink, the effect of this field on the value of writer_identity in the updated sink depends on both the old and new values of this field:
If the old and new values of this field are both false or both true, then there is no change to the sink's writer_identity.
If the old value is false and the new value is true, then writer_identity is changed to a unique service account.
It is an error if the old value is true and the new value is set to false or defaulted to false.

Type: boolean

updateMask#

Optional. Field mask that specifies the fields in sink that need an update. A sink field will be overwritten if, and only if, it is in the update mask. name and output only fields cannot be updated.An empty updateMask is temporarily treated as using the following mask for backwards compatibility purposes: destination,filter,includeChildren At some point in the future, behavior will be removed and specifying an empty updateMask will be an error.For a detailed FieldMask definition, see https://developers.google.com/protocol-buffers/docs/reference/google.protobuf#google.protobuf.FieldMaskExample: updateMask=filter.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

patch_sink_in_organization#

Updates a sink. This method replaces the following fields in the existing sink with values from the new sink: destination, and filter.The updated sink might also have a new writer_identity; see the unique_writer_identity field.

Parameters

organizationsId (required)#

Type: string

sinksId (required)#

Type: string

$body#

Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, or a Cloud Pub/Sub topic. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.

Type: object

{
"filter" : "Optional. An advanced logs filter. The only exported log entries are those that are in the resource owning the sink and that match the filter. For example:\nlogName=\"projects/[PROJECT_ID]/logs/[LOG_ID]\" AND severity>=ERROR\n",
"writerIdentity" : "Output only. An IAM identity—a service account or group—under which Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource. Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.",
"createTime" : "Output only. The creation timestamp of the sink.This field may not be present for older sinks.",
"includeChildren" : "Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then logs from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression. For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent. To only export entries from certain child projects, filter on the project part of the log name:\nlogName:(\"projects/test-project1/\" OR \"projects/test-project2/\") AND\nresource.type=gce_instance\n",
"name" : "Required. The client-assigned sink identifier, unique within the project. Example: \"my-syslog-errors-to-pubsub\". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.",
"bigqueryOptions" : {
"usesTimestampColumnPartitioning" : "Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use. All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.",
"usePartitionedTables" : "Optional. Whether to use BigQuery's partition tables. By default, Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax has to be used instead. In both cases, tables are sharded based on UTC timezone."
},
"destination" : "Required. The export destination:\n\"storage.googleapis.com/[GCS_BUCKET]\"\n\"bigquery.googleapis.com/projects/[PROJECT_ID]/datasets/[DATASET]\"\n\"pubsub.googleapis.com/projects/[PROJECT_ID]/topics/[TOPIC_ID]\"\nThe sink's writer_identity, set when the sink is created, must have permission to write to the destination or else the log entries are not exported. For more information, see Exporting Logs with Sinks.",
"description" : "Optional. A description of this sink. The maximum length of the description is 8000 characters.",
"disabled" : "Optional. If set to True, then this sink is disabled and it does not export any log entries.",
"updateTime" : "Output only. The last update timestamp of the sink.This field may not be present for older sinks.",
"outputVersionFormat" : "Deprecated. The log entry format to use for this sink's exported log entries. The v2 format is used by default and cannot be changed."
}

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uniqueWriterIdentity#

Optional. See sinks.create for a description of this field. When updating a sink, the effect of this field on the value of writer_identity in the updated sink depends on both the old and new values of this field:
If the old and new values of this field are both false or both true, then there is no change to the sink's writer_identity.
If the old value is false and the new value is true, then writer_identity is changed to a unique service account.
It is an error if the old value is true and the new value is set to false or defaulted to false.

Type: boolean

updateMask#

Optional. Field mask that specifies the fields in sink that need an update. A sink field will be overwritten if, and only if, it is in the update mask. name and output only fields cannot be updated.An empty updateMask is temporarily treated as using the following mask for backwards compatibility purposes: destination,filter,includeChildren At some point in the future, behavior will be removed and specifying an empty updateMask will be an error.For a detailed FieldMask definition, see https://developers.google.com/protocol-buffers/docs/reference/google.protobuf#google.protobuf.FieldMaskExample: updateMask=filter.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

patch_sink_in_project#

Updates a sink. This method replaces the following fields in the existing sink with values from the new sink: destination, and filter.The updated sink might also have a new writer_identity; see the unique_writer_identity field.

Parameters

projectsId (required)#

Type: string

sinksId (required)#

Type: string

$body#

Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, or a Cloud Pub/Sub topic. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.

Type: object

{
"filter" : "Optional. An advanced logs filter. The only exported log entries are those that are in the resource owning the sink and that match the filter. For example:\nlogName=\"projects/[PROJECT_ID]/logs/[LOG_ID]\" AND severity>=ERROR\n",
"writerIdentity" : "Output only. An IAM identity—a service account or group—under which Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource. Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.",
"createTime" : "Output only. The creation timestamp of the sink.This field may not be present for older sinks.",
"includeChildren" : "Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then logs from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression. For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent. To only export entries from certain child projects, filter on the project part of the log name:\nlogName:(\"projects/test-project1/\" OR \"projects/test-project2/\") AND\nresource.type=gce_instance\n",
"name" : "Required. The client-assigned sink identifier, unique within the project. Example: \"my-syslog-errors-to-pubsub\". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.",
"bigqueryOptions" : {
"usesTimestampColumnPartitioning" : "Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use. All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.",
"usePartitionedTables" : "Optional. Whether to use BigQuery's partition tables. By default, Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax has to be used instead. In both cases, tables are sharded based on UTC timezone."
},
"destination" : "Required. The export destination:\n\"storage.googleapis.com/[GCS_BUCKET]\"\n\"bigquery.googleapis.com/projects/[PROJECT_ID]/datasets/[DATASET]\"\n\"pubsub.googleapis.com/projects/[PROJECT_ID]/topics/[TOPIC_ID]\"\nThe sink's writer_identity, set when the sink is created, must have permission to write to the destination or else the log entries are not exported. For more information, see Exporting Logs with Sinks.",
"description" : "Optional. A description of this sink. The maximum length of the description is 8000 characters.",
"disabled" : "Optional. If set to True, then this sink is disabled and it does not export any log entries.",
"updateTime" : "Output only. The last update timestamp of the sink.This field may not be present for older sinks.",
"outputVersionFormat" : "Deprecated. The log entry format to use for this sink's exported log entries. The v2 format is used by default and cannot be changed."
}

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uniqueWriterIdentity#

Optional. See sinks.create for a description of this field. When updating a sink, the effect of this field on the value of writer_identity in the updated sink depends on both the old and new values of this field:
If the old and new values of this field are both false or both true, then there is no change to the sink's writer_identity.
If the old value is false and the new value is true, then writer_identity is changed to a unique service account.
It is an error if the old value is true and the new value is set to false or defaulted to false.

Type: boolean

updateMask#

Optional. Field mask that specifies the fields in sink that need an update. A sink field will be overwritten if, and only if, it is in the update mask. name and output only fields cannot be updated.An empty updateMask is temporarily treated as using the following mask for backwards compatibility purposes: destination,filter,includeChildren At some point in the future, behavior will be removed and specifying an empty updateMask will be an error.For a detailed FieldMask definition, see https://developers.google.com/protocol-buffers/docs/reference/google.protobuf#google.protobuf.FieldMaskExample: updateMask=filter.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

update_cmek_setting_in_organization#

Updates the Logs Router CMEK settings for the given resource.Note: CMEK for the Logs Router can currently only be configured for GCP organizations. Once configured, it applies to all projects and folders in the GCP organization.UpdateCmekSettings will fail if 1) kms_key_name is invalid, or 2) the associated service account does not have the required roles/cloudkms.cryptoKeyEncrypterDecrypter role assigned for the key, or 3) access to the key is disabled.See Enabling CMEK for Logs Router for more information.

Parameters

organizationsId (required)#

Type: string

$body#

Describes the customer-managed encryption key (CMEK) settings associated with a project, folder, organization, billing account, or flexible resource.Note: CMEK for the Logs Router can currently only be configured for GCP organizations. Once configured, it applies to all projects and folders in the GCP organization.See Enabling CMEK for Logs Router for more information.

Type: object

{
"serviceAccountId" : "Output only. The service account that will be used by the Logs Router to access your Cloud KMS key.Before enabling CMEK for Logs Router, you must first assign the role roles/cloudkms.cryptoKeyEncrypterDecrypter to the service account that the Logs Router will use to access your Cloud KMS key. Use GetCmekSettings to obtain the service account ID.See Enabling CMEK for Logs Router for more information.",
"name" : "Output only. The resource name of the CMEK settings.",
"kmsKeyName" : "The resource name for the configured Cloud KMS key.KMS key name format: \"projects/PROJECT_ID/locations/LOCATION/keyRings/KEYRING/cryptoKeys/KEY\"For example: \"projects/my-project-id/locations/my-region/keyRings/key-ring-name/cryptoKeys/key-name\"To enable CMEK for the Logs Router, set this field to a valid kms_key_name for which the associated service account has the required roles/cloudkms.cryptoKeyEncrypterDecrypter role assigned for the key.The Cloud KMS key used by the Log Router can be updated by changing the kms_key_name to a new valid key name. Encryption operations that are in progress will be completed with the key that was in use when they started. Decryption operations will be completed using the key that was used at the time of encryption unless access to that key has been revoked.To disable CMEK for the Logs Router, set this field to an empty string.See Enabling CMEK for Logs Router for more information."
}

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

updateMask#

Optional. Field mask identifying which fields from cmek_settings should be updated. A field will be overwritten if and only if it is in the update mask. Output only fields cannot be updated.See FieldMask for more information.Example: "updateMask=kmsKeyName"

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

update_metric_in_project#

Creates or updates a logs-based metric.

Parameters

metricsId (required)#

Type: string

projectsId (required)#

Type: string

$body#

Describes a logs-based metric. The value of the metric is the number of log entries that match a logs filter in a given time interval.Logs-based metric can also be used to extract values from logs and create a a distribution of the values. The distribution records the statistics of the extracted values along with an optional histogram of the values as specified by the bucket options.

Type: object

{
"filter" : "Required. An advanced logs filter which is used to match log entries. Example:\n\"resource.type=gae_app AND severity>=ERROR\"\nThe maximum length of the filter is 20000 characters.",
"labelExtractors" : "Optional. A map from a label key string to an extractor expression which is used to extract data from a log entry field and assign as the label value. Each label key specified in the LabelDescriptor must have an associated extractor expression in this map. The syntax of the extractor expression is the same as for the value_extractor field.The extracted value is converted to the type defined in the label descriptor. If the either the extraction or the type conversion fails, the label will have a default value. The default value for a string label is an empty string, for an integer label its 0, and for a boolean label its false.Note that there are upper bounds on the maximum number of labels and the number of active time series that are allowed in a project.",
"valueExtractor" : "Optional. A value_extractor is required when using a distribution logs-based metric to extract the values to record from a log entry. Two functions are supported for value extraction: EXTRACT(field) or REGEXP_EXTRACT(field, regex). The argument are: 1. field: The name of the log entry field from which the value is to be extracted. 2. regex: A regular expression using the Google RE2 syntax (https://github.com/google/re2/wiki/Syntax) with a single capture group to extract data from the specified log entry field. The value of the field is converted to a string before applying the regex. It is an error to specify a regex that does not include exactly one capture group.The result of the extraction must be convertible to a double type, as the distribution always records double values. If either the extraction or the conversion to double fails, then those values are not recorded in the distribution.Example: REGEXP_EXTRACT(jsonPayload.request, \".*quantity=(\\d+).*\")",
"createTime" : "Output only. The creation timestamp of the metric.This field may not be present for older metrics.",
"metricDescriptor" : {
"monitoredResourceTypes" : [ "string" ],
"unit" : "The units in which the metric value is reported. It is only applicable if the value_type is INT64, DOUBLE, or DISTRIBUTION. The unit defines the representation of the stored metric values.Different systems may scale the values to be more easily displayed (so a value of 0.02KBy might be displayed as 20By, and a value of 3523KBy might be displayed as 3.5MBy). However, if the unit is KBy, then the value of the metric is always in thousands of bytes, no matter how it may be displayed..If you want a custom metric to record the exact number of CPU-seconds used by a job, you can create an INT64 CUMULATIVE metric whose unit is s{CPU} (or equivalently 1s{CPU} or just s). If the job uses 12,005 CPU-seconds, then the value is written as 12005.Alternatively, if you want a custom metric to record data in a more granular way, you can create a DOUBLE CUMULATIVE metric whose unit is ks{CPU}, and then write the value 12.005 (which is 12005/1000), or use Kis{CPU} and write 11.723 (which is 12005/1024).The supported units are a subset of The Unified Code for Units of Measure (http://unitsofmeasure.org/ucum.html) standard:Basic units (UNIT)\nbit bit\nBy byte\ns second\nmin minute\nh hour\nd dayPrefixes (PREFIX)\nk kilo (10^3)\nM mega (10^6)\nG giga (10^9)\nT tera (10^12)\nP peta (10^15)\nE exa (10^18)\nZ zetta (10^21)\nY yotta (10^24)\nm milli (10^-3)\nu micro (10^-6)\nn nano (10^-9)\np pico (10^-12)\nf femto (10^-15)\na atto (10^-18)\nz zepto (10^-21)\ny yocto (10^-24)\nKi kibi (2^10)\nMi mebi (2^20)\nGi gibi (2^30)\nTi tebi (2^40)\nPi pebi (2^50)GrammarThe grammar also includes these connectors:\n/ division or ratio (as an infix operator). For examples, kBy/{email} or MiBy/10ms (although you should almost never have /s in a metric unit; rates should always be computed at query time from the underlying cumulative or delta value).\n. multiplication or composition (as an infix operator). For examples, GBy.d or k{watt}.h.The grammar for a unit is as follows:\nExpression = Component { \".\" Component } { \"/\" Component } ;\n\nComponent = ( [ PREFIX ] UNIT | \"%\" ) [ Annotation ]\n | Annotation\n | \"1\"\n ;\n\nAnnotation = \"{\" NAME \"}\" ;\nNotes:\nAnnotation is just a comment if it follows a UNIT. If the annotation is used alone, then the unit is equivalent to 1. For examples, {request}/s == 1/s, By{transmitted}/s == By/s.\nNAME is a sequence of non-blank printable ASCII characters not containing { or }.\n1 represents a unitary dimensionless unit (https://en.wikipedia.org/wiki/Dimensionless_quantity) of 1, such as in 1/s. It is typically used when none of the basic units are appropriate. For example, \"new users per day\" can be represented as 1/d or {new-users}/d (and a metric value 5 would mean \"5 new users). Alternatively, \"thousands of page views per day\" would be represented as 1000/d or k1/d or k{page_views}/d (and a metric value of 5.3 would mean \"5300 page views per day\").\n% represents dimensionless value of 1/100, and annotates values giving a percentage (so the metric values are typically in the range of 0..100, and a metric value 3 means \"3 percent\").\n10^2.% indicates a metric contains a ratio, typically in the range 0..1, that will be multiplied by 100 and displayed as a percentage (so a metric value 0.03 means \"3 percent\").",
"metadata" : {
"ingestDelay" : "The delay of data points caused by ingestion. Data points older than this age are guaranteed to be ingested and available to be read, excluding data loss due to errors.",
"launchStage" : "Deprecated. Must use the MetricDescriptor.launch_stage instead.",
"samplePeriod" : "The sampling period of metric data points. For metrics which are written periodically, consecutive data points are stored at this time interval, excluding data loss due to errors. Metrics with a higher granularity have a smaller sampling period."
},
"metricKind" : "Whether the metric records instantaneous values, changes to a value, etc. Some combinations of metric_kind and value_type might not be supported.",
"displayName" : "A concise name for the metric, which can be displayed in user interfaces. Use sentence case without an ending period, for example \"Request count\". This field is optional but it is recommended to be set for any metrics associated with user-visible concepts, such as Quota.",
"valueType" : "Whether the measurement is an integer, a floating-point number, etc. Some combinations of metric_kind and value_type might not be supported.",
"name" : "The resource name of the metric descriptor.",
"description" : "A detailed description of the metric, which can be used in documentation.",
"launchStage" : "Optional. The launch stage of the metric definition.",
"type" : "The metric type, including its DNS name prefix. The type is not URL-encoded. All user-defined metric types have the DNS name custom.googleapis.com or external.googleapis.com. Metric types should use a natural hierarchical grouping. For example:\n\"custom.googleapis.com/invoice/paid/amount\"\n\"external.googleapis.com/prometheus/up\"\n\"appengine.googleapis.com/http/server/response_latencies\"\n",
"labels" : [ {
"valueType" : "The type of data that can be assigned to the label.",
"description" : "A human-readable description for the label.",
"key" : "The label key."
} ]
},
"bucketOptions" : {
"explicitBuckets" : {
"bounds" : [ "number" ]
},
"linearBuckets" : {
"offset" : "Lower bound of the first bucket.",
"width" : "Must be greater than 0.",
"numFiniteBuckets" : "Must be greater than 0."
},
"exponentialBuckets" : {
"growthFactor" : "Must be greater than 1.",
"scale" : "Must be greater than 0.",
"numFiniteBuckets" : "Must be greater than 0."
}
},
"name" : "Required. The client-assigned metric identifier. Examples: \"error_count\", \"nginx/requests\".Metric identifiers are limited to 100 characters and can include only the following characters: A-Z, a-z, 0-9, and the special characters _-.,+!*',()%/. The forward-slash character (/) denotes a hierarchy of name pieces, and it cannot be the first character of the name.The metric identifier in this field must not be URL-encoded (https://en.wikipedia.org/wiki/Percent-encoding). However, when the metric identifier appears as the [METRIC_ID] part of a metric_name API parameter, then the metric identifier must be URL-encoded. Example: \"projects/my-project/metrics/nginx%2Frequests\".",
"description" : "Optional. A description of this metric, which is used in documentation. The maximum length of the description is 8000 characters.",
"updateTime" : "Output only. The last update timestamp of the metric.This field may not be present for older metrics.",
"version" : "Deprecated. The API version that created or updated this metric. The v2 format is used by default and cannot be changed."
}

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

update_sink_in_billing_account#

Updates a sink. This method replaces the following fields in the existing sink with values from the new sink: destination, and filter.The updated sink might also have a new writer_identity; see the unique_writer_identity field.

Parameters

billingAccountsId (required)#

Type: string

sinksId (required)#

Type: string

$body#

Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, or a Cloud Pub/Sub topic. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.

Type: object

{
"filter" : "Optional. An advanced logs filter. The only exported log entries are those that are in the resource owning the sink and that match the filter. For example:\nlogName=\"projects/[PROJECT_ID]/logs/[LOG_ID]\" AND severity>=ERROR\n",
"writerIdentity" : "Output only. An IAM identity—a service account or group—under which Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource. Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.",
"createTime" : "Output only. The creation timestamp of the sink.This field may not be present for older sinks.",
"includeChildren" : "Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then logs from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression. For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent. To only export entries from certain child projects, filter on the project part of the log name:\nlogName:(\"projects/test-project1/\" OR \"projects/test-project2/\") AND\nresource.type=gce_instance\n",
"name" : "Required. The client-assigned sink identifier, unique within the project. Example: \"my-syslog-errors-to-pubsub\". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.",
"bigqueryOptions" : {
"usesTimestampColumnPartitioning" : "Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use. All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.",
"usePartitionedTables" : "Optional. Whether to use BigQuery's partition tables. By default, Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax has to be used instead. In both cases, tables are sharded based on UTC timezone."
},
"destination" : "Required. The export destination:\n\"storage.googleapis.com/[GCS_BUCKET]\"\n\"bigquery.googleapis.com/projects/[PROJECT_ID]/datasets/[DATASET]\"\n\"pubsub.googleapis.com/projects/[PROJECT_ID]/topics/[TOPIC_ID]\"\nThe sink's writer_identity, set when the sink is created, must have permission to write to the destination or else the log entries are not exported. For more information, see Exporting Logs with Sinks.",
"description" : "Optional. A description of this sink. The maximum length of the description is 8000 characters.",
"disabled" : "Optional. If set to True, then this sink is disabled and it does not export any log entries.",
"updateTime" : "Output only. The last update timestamp of the sink.This field may not be present for older sinks.",
"outputVersionFormat" : "Deprecated. The log entry format to use for this sink's exported log entries. The v2 format is used by default and cannot be changed."
}

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uniqueWriterIdentity#

Optional. See sinks.create for a description of this field. When updating a sink, the effect of this field on the value of writer_identity in the updated sink depends on both the old and new values of this field:
If the old and new values of this field are both false or both true, then there is no change to the sink's writer_identity.
If the old value is false and the new value is true, then writer_identity is changed to a unique service account.
It is an error if the old value is true and the new value is set to false or defaulted to false.

Type: boolean

updateMask#

Optional. Field mask that specifies the fields in sink that need an update. A sink field will be overwritten if, and only if, it is in the update mask. name and output only fields cannot be updated.An empty updateMask is temporarily treated as using the following mask for backwards compatibility purposes: destination,filter,includeChildren At some point in the future, behavior will be removed and specifying an empty updateMask will be an error.For a detailed FieldMask definition, see https://developers.google.com/protocol-buffers/docs/reference/google.protobuf#google.protobuf.FieldMaskExample: updateMask=filter.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

update_sink_in_folder#

Updates a sink. This method replaces the following fields in the existing sink with values from the new sink: destination, and filter.The updated sink might also have a new writer_identity; see the unique_writer_identity field.

Parameters

foldersId (required)#

Type: string

sinksId (required)#

Type: string

$body#

Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, or a Cloud Pub/Sub topic. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.

Type: object

{
"filter" : "Optional. An advanced logs filter. The only exported log entries are those that are in the resource owning the sink and that match the filter. For example:\nlogName=\"projects/[PROJECT_ID]/logs/[LOG_ID]\" AND severity>=ERROR\n",
"writerIdentity" : "Output only. An IAM identity—a service account or group—under which Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource. Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.",
"createTime" : "Output only. The creation timestamp of the sink.This field may not be present for older sinks.",
"includeChildren" : "Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then logs from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression. For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent. To only export entries from certain child projects, filter on the project part of the log name:\nlogName:(\"projects/test-project1/\" OR \"projects/test-project2/\") AND\nresource.type=gce_instance\n",
"name" : "Required. The client-assigned sink identifier, unique within the project. Example: \"my-syslog-errors-to-pubsub\". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.",
"bigqueryOptions" : {
"usesTimestampColumnPartitioning" : "Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use. All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.",
"usePartitionedTables" : "Optional. Whether to use BigQuery's partition tables. By default, Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax has to be used instead. In both cases, tables are sharded based on UTC timezone."
},
"destination" : "Required. The export destination:\n\"storage.googleapis.com/[GCS_BUCKET]\"\n\"bigquery.googleapis.com/projects/[PROJECT_ID]/datasets/[DATASET]\"\n\"pubsub.googleapis.com/projects/[PROJECT_ID]/topics/[TOPIC_ID]\"\nThe sink's writer_identity, set when the sink is created, must have permission to write to the destination or else the log entries are not exported. For more information, see Exporting Logs with Sinks.",
"description" : "Optional. A description of this sink. The maximum length of the description is 8000 characters.",
"disabled" : "Optional. If set to True, then this sink is disabled and it does not export any log entries.",
"updateTime" : "Output only. The last update timestamp of the sink.This field may not be present for older sinks.",
"outputVersionFormat" : "Deprecated. The log entry format to use for this sink's exported log entries. The v2 format is used by default and cannot be changed."
}

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uniqueWriterIdentity#

Optional. See sinks.create for a description of this field. When updating a sink, the effect of this field on the value of writer_identity in the updated sink depends on both the old and new values of this field:
If the old and new values of this field are both false or both true, then there is no change to the sink's writer_identity.
If the old value is false and the new value is true, then writer_identity is changed to a unique service account.
It is an error if the old value is true and the new value is set to false or defaulted to false.

Type: boolean

updateMask#

Optional. Field mask that specifies the fields in sink that need an update. A sink field will be overwritten if, and only if, it is in the update mask. name and output only fields cannot be updated.An empty updateMask is temporarily treated as using the following mask for backwards compatibility purposes: destination,filter,includeChildren At some point in the future, behavior will be removed and specifying an empty updateMask will be an error.For a detailed FieldMask definition, see https://developers.google.com/protocol-buffers/docs/reference/google.protobuf#google.protobuf.FieldMaskExample: updateMask=filter.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

update_sink_in_organization#

Updates a sink. This method replaces the following fields in the existing sink with values from the new sink: destination, and filter.The updated sink might also have a new writer_identity; see the unique_writer_identity field.

Parameters

organizationsId (required)#

Type: string

sinksId (required)#

Type: string

$body#

Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, or a Cloud Pub/Sub topic. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.

Type: object

{
"filter" : "Optional. An advanced logs filter. The only exported log entries are those that are in the resource owning the sink and that match the filter. For example:\nlogName=\"projects/[PROJECT_ID]/logs/[LOG_ID]\" AND severity>=ERROR\n",
"writerIdentity" : "Output only. An IAM identity—a service account or group—under which Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource. Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.",
"createTime" : "Output only. The creation timestamp of the sink.This field may not be present for older sinks.",
"includeChildren" : "Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then logs from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression. For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent. To only export entries from certain child projects, filter on the project part of the log name:\nlogName:(\"projects/test-project1/\" OR \"projects/test-project2/\") AND\nresource.type=gce_instance\n",
"name" : "Required. The client-assigned sink identifier, unique within the project. Example: \"my-syslog-errors-to-pubsub\". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.",
"bigqueryOptions" : {
"usesTimestampColumnPartitioning" : "Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use. All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.",
"usePartitionedTables" : "Optional. Whether to use BigQuery's partition tables. By default, Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax has to be used instead. In both cases, tables are sharded based on UTC timezone."
},
"destination" : "Required. The export destination:\n\"storage.googleapis.com/[GCS_BUCKET]\"\n\"bigquery.googleapis.com/projects/[PROJECT_ID]/datasets/[DATASET]\"\n\"pubsub.googleapis.com/projects/[PROJECT_ID]/topics/[TOPIC_ID]\"\nThe sink's writer_identity, set when the sink is created, must have permission to write to the destination or else the log entries are not exported. For more information, see Exporting Logs with Sinks.",
"description" : "Optional. A description of this sink. The maximum length of the description is 8000 characters.",
"disabled" : "Optional. If set to True, then this sink is disabled and it does not export any log entries.",
"updateTime" : "Output only. The last update timestamp of the sink.This field may not be present for older sinks.",
"outputVersionFormat" : "Deprecated. The log entry format to use for this sink's exported log entries. The v2 format is used by default and cannot be changed."
}

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uniqueWriterIdentity#

Optional. See sinks.create for a description of this field. When updating a sink, the effect of this field on the value of writer_identity in the updated sink depends on both the old and new values of this field:
If the old and new values of this field are both false or both true, then there is no change to the sink's writer_identity.
If the old value is false and the new value is true, then writer_identity is changed to a unique service account.
It is an error if the old value is true and the new value is set to false or defaulted to false.

Type: boolean

updateMask#

Optional. Field mask that specifies the fields in sink that need an update. A sink field will be overwritten if, and only if, it is in the update mask. name and output only fields cannot be updated.An empty updateMask is temporarily treated as using the following mask for backwards compatibility purposes: destination,filter,includeChildren At some point in the future, behavior will be removed and specifying an empty updateMask will be an error.For a detailed FieldMask definition, see https://developers.google.com/protocol-buffers/docs/reference/google.protobuf#google.protobuf.FieldMaskExample: updateMask=filter.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

update_sink_in_project#

Updates a sink. This method replaces the following fields in the existing sink with values from the new sink: destination, and filter.The updated sink might also have a new writer_identity; see the unique_writer_identity field.

Parameters

projectsId (required)#

Type: string

sinksId (required)#

Type: string

$body#

Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, or a Cloud Pub/Sub topic. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.

Type: object

{
"filter" : "Optional. An advanced logs filter. The only exported log entries are those that are in the resource owning the sink and that match the filter. For example:\nlogName=\"projects/[PROJECT_ID]/logs/[LOG_ID]\" AND severity>=ERROR\n",
"writerIdentity" : "Output only. An IAM identity—a service account or group—under which Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource. Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.",
"createTime" : "Output only. The creation timestamp of the sink.This field may not be present for older sinks.",
"includeChildren" : "Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then logs from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression. For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent. To only export entries from certain child projects, filter on the project part of the log name:\nlogName:(\"projects/test-project1/\" OR \"projects/test-project2/\") AND\nresource.type=gce_instance\n",
"name" : "Required. The client-assigned sink identifier, unique within the project. Example: \"my-syslog-errors-to-pubsub\". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.",
"bigqueryOptions" : {
"usesTimestampColumnPartitioning" : "Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use. All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.",
"usePartitionedTables" : "Optional. Whether to use BigQuery's partition tables. By default, Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax has to be used instead. In both cases, tables are sharded based on UTC timezone."
},
"destination" : "Required. The export destination:\n\"storage.googleapis.com/[GCS_BUCKET]\"\n\"bigquery.googleapis.com/projects/[PROJECT_ID]/datasets/[DATASET]\"\n\"pubsub.googleapis.com/projects/[PROJECT_ID]/topics/[TOPIC_ID]\"\nThe sink's writer_identity, set when the sink is created, must have permission to write to the destination or else the log entries are not exported. For more information, see Exporting Logs with Sinks.",
"description" : "Optional. A description of this sink. The maximum length of the description is 8000 characters.",
"disabled" : "Optional. If set to True, then this sink is disabled and it does not export any log entries.",
"updateTime" : "Output only. The last update timestamp of the sink.This field may not be present for older sinks.",
"outputVersionFormat" : "Deprecated. The log entry format to use for this sink's exported log entries. The v2 format is used by default and cannot be changed."
}

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uniqueWriterIdentity#

Optional. See sinks.create for a description of this field. When updating a sink, the effect of this field on the value of writer_identity in the updated sink depends on both the old and new values of this field:
If the old and new values of this field are both false or both true, then there is no change to the sink's writer_identity.
If the old value is false and the new value is true, then writer_identity is changed to a unique service account.
It is an error if the old value is true and the new value is set to false or defaulted to false.

Type: boolean

updateMask#

Optional. Field mask that specifies the fields in sink that need an update. A sink field will be overwritten if, and only if, it is in the update mask. name and output only fields cannot be updated.An empty updateMask is temporarily treated as using the following mask for backwards compatibility purposes: destination,filter,includeChildren At some point in the future, behavior will be removed and specifying an empty updateMask will be an error.For a detailed FieldMask definition, see https://developers.google.com/protocol-buffers/docs/reference/google.protobuf#google.protobuf.FieldMaskExample: updateMask=filter.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string

write_entry#

Writes log entries to Logging. This API method is the only way to send log entries to Logging. This method is used, directly or indirectly, by the Logging agent (fluentd) and all logging libraries configured to use Logging. A single request may contain log entries for a maximum of 1000 different resources (projects, organizations, billing accounts or folders)

Parameters

$body#

The parameters to WriteLogEntries.

Type: object

{
"entries" : [ {
"severity" : "Optional. The severity of the log entry. The default value is LogSeverity.DEFAULT.",
"metadata" : {
"systemLabels" : "Output only. Values for predefined system metadata labels. System labels are a kind of metadata extracted by Google, including \"machine_image\", \"vpc\", \"subnet_id\", \"security_group\", \"name\", etc. System label values can be only strings, Boolean values, or a list of strings. For example:\n{ \"name\": \"my-test-instance\",\n \"security_group\": [\"a\", \"b\", \"c\"],\n \"spot_instance\": false }\n",
"userLabels" : "Output only. A map of user-defined metadata labels."
},
"traceSampled" : "Optional. The sampling decision of the trace associated with the log entry.True means that the trace resource name in the trace field was sampled for storage in a trace backend. False means that the trace was not sampled for storage when this log entry was written, or the sampling decision was unknown at the time. A non-sampled trace value is still useful as a request correlation identifier. The default is False.",
"resource" : {
"type" : "Required. The monitored resource type. This field must match the type field of a MonitoredResourceDescriptor object. For example, the type of a Compute Engine VM instance is gce_instance.",
"labels" : "Required. Values for all of the labels listed in the associated monitored resource descriptor. For example, Compute Engine VM instances use the labels \"project_id\", \"instance_id\", and \"zone\"."
},
"protoPayload" : "The log entry payload, represented as a protocol buffer. Some Google Cloud Platform services use this field for their log entry payloads.The following protocol buffer types are supported; user-defined types are not supported:\"type.googleapis.com/google.cloud.audit.AuditLog\" \"type.googleapis.com/google.appengine.logging.v1.RequestLog\"",
"textPayload" : "The log entry payload, represented as a Unicode string (UTF-8).",
"receiveTimestamp" : "Output only. The time the log entry was received by Logging.",
"labels" : "Optional. A set of user-defined (key, value) data that provides additional information about the log entry.",
"spanId" : "Optional. The span ID within the trace associated with the log entry.For Trace spans, this is the same format that the Trace API v2 uses: a 16-character hexadecimal encoding of an 8-byte array, such as 000000000000004a.",
"trace" : "Optional. Resource name of the trace associated with the log entry, if any. If it contains a relative resource name, the name is assumed to be relative to //tracing.googleapis.com. Example: projects/my-projectid/traces/06796866738c859f2f19b7cfb3214824",
"logName" : "Required. The resource name of the log to which this log entry belongs:\n\"projects/[PROJECT_ID]/logs/[LOG_ID]\"\n\"organizations/[ORGANIZATION_ID]/logs/[LOG_ID]\"\n\"billingAccounts/[BILLING_ACCOUNT_ID]/logs/[LOG_ID]\"\n\"folders/[FOLDER_ID]/logs/[LOG_ID]\"\nA project number may be used in place of PROJECT_ID. The project number is translated to its corresponding PROJECT_ID internally and the log_name field will contain PROJECT_ID in queries and exports.[LOG_ID] must be URL-encoded within log_name. Example: \"organizations/1234567890/logs/cloudresourcemanager.googleapis.com%2Factivity\". [LOG_ID] must be less than 512 characters long and can only include the following characters: upper and lower case alphanumeric characters, forward-slash, underscore, hyphen, and period.For backward compatibility, if log_name begins with a forward-slash, such as /projects/..., then the log entry is ingested as usual but the forward-slash is removed. Listing the log entry will not show the leading slash and filtering for a log name with a leading slash will never return any results.",
"jsonPayload" : "The log entry payload, represented as a structure that is expressed as a JSON object.",
"httpRequest" : {
"referer" : "The referer URL of the request, as defined in HTTP/1.1 Header Field Definitions (http://www.w3.org/Protocols/rfc2616/rfc2616-sec14.html).",
"remoteIp" : "The IP address (IPv4 or IPv6) of the client that issued the HTTP request. Examples: \"192.168.1.1\", \"FE80::0202:B3FF:FE1E:8329\".",
"latency" : "The request processing latency on the server, from the time the request was received until the response was sent.",
"requestMethod" : "The request method. Examples: \"GET\", \"HEAD\", \"PUT\", \"POST\".",
"userAgent" : "The user agent sent by the client. Example: \"Mozilla/4.0 (compatible; MSIE 6.0; Windows 98; Q312461; .NET\nCLR 1.0.3705)\".",
"responseSize" : "The size of the HTTP response message sent back to the client, in bytes, including the response headers and the response body.",
"cacheLookup" : "Whether or not a cache lookup was attempted.",
"protocol" : "Protocol used for the request. Examples: \"HTTP/1.1\", \"HTTP/2\", \"websocket\"",
"requestUrl" : "The scheme (http, https), the host name, the path and the query portion of the URL that was requested. Example: \"http://example.com/some/info?color=red\".",
"cacheHit" : "Whether or not an entity was served from cache (with or without validation).",
"serverIp" : "The IP address (IPv4 or IPv6) of the origin server that the request was sent to.",
"cacheValidatedWithOriginServer" : "Whether or not the response was validated with the origin server before being served from cache. This field is only meaningful if cache_hit is True.",
"requestSize" : "The size of the HTTP request message in bytes, including the request headers and the request body.",
"cacheFillBytes" : "The number of HTTP response bytes inserted into cache. Set only when a cache fill was attempted.",
"status" : "The response code indicating the status of response. Examples: 200, 404."
},
"sourceLocation" : {
"file" : "Optional. Source file name. Depending on the runtime environment, this might be a simple name or a fully-qualified name.",
"line" : "Optional. Line within the source file. 1-based; 0 indicates no line number available.",
"function" : "Optional. Human-readable name of the function or method being invoked, with optional context such as the class or package name. This information may be used in contexts such as the logs viewer, where a file and line number are less meaningful. The format can vary by language. For example: qual.if.ied.Class.method (Java), dir/package.func (Go), function (Python)."
},
"operation" : {
"last" : "Optional. Set this to True if this is the last log entry in the operation.",
"producer" : "Optional. An arbitrary producer identifier. The combination of id and producer must be globally unique. Examples for producer: \"MyDivision.MyBigCompany.com\", \"github.com/MyProject/MyApplication\".",
"id" : "Optional. An arbitrary operation identifier. Log entries with the same identifier are assumed to be part of the same operation.",
"first" : "Optional. Set this to True if this is the first log entry in the operation."
},
"insertId" : "Optional. A unique identifier for the log entry. If you provide a value, then Logging considers other log entries in the same project, with the same timestamp, and with the same insert_id to be duplicates which are removed in a single query result. However, there are no guarantees of de-duplication in the export of logs.If the insert_id is omitted when writing a log entry, the Logging API assigns its own unique identifier in this field.In queries, the insert_id is also used to order log entries that have the same log_name and timestamp values.",
"timestamp" : "Optional. The time the event described by the log entry occurred. This time is used to compute the log entry's age and to enforce the logs retention period. If this field is omitted in a new log entry, then Logging assigns it the current time. Timestamps have nanosecond accuracy, but trailing zeros in the fractional seconds might be omitted when the timestamp is displayed.Incoming log entries should have timestamps that are no more than the logs retention period in the past, and no more than 24 hours in the future. Log entries outside those time boundaries will not be available when calling entries.list, but those log entries can still be exported with LogSinks."
} ],
"dryRun" : "Optional. If true, the request should expect normal response, but the entries won't be persisted nor exported. Useful for checking whether the logging API endpoints are working properly before sending valuable data.",
"logName" : "Optional. A default log resource name that is assigned to all log entries in entries that do not specify a value for log_name:\n\"projects/[PROJECT_ID]/logs/[LOG_ID]\"\n\"organizations/[ORGANIZATION_ID]/logs/[LOG_ID]\"\n\"billingAccounts/[BILLING_ACCOUNT_ID]/logs/[LOG_ID]\"\n\"folders/[FOLDER_ID]/logs/[LOG_ID]\"\n[LOG_ID] must be URL-encoded. For example:\n\"projects/my-project-id/logs/syslog\"\n\"organizations/1234567890/logs/cloudresourcemanager.googleapis.com%2Factivity\"\nThe permission logging.logEntries.create is needed on each project, organization, billing account, or folder that is receiving new log entries, whether the resource is specified in logName or in an individual log entry.",
"resource" : {
"type" : "Required. The monitored resource type. This field must match the type field of a MonitoredResourceDescriptor object. For example, the type of a Compute Engine VM instance is gce_instance.",
"labels" : "Required. Values for all of the labels listed in the associated monitored resource descriptor. For example, Compute Engine VM instances use the labels \"project_id\", \"instance_id\", and \"zone\"."
},
"partialSuccess" : "Optional. Whether valid entries should be written even if some other entries fail due to INVALID_ARGUMENT or PERMISSION_DENIED errors. If any entry is not written, then the response status is the error associated with one of the failed entries and the response includes error details keyed by the entries' zero-based index in the entries.write method.",
"labels" : "Optional. Default labels that are added to the labels field of all log entries in entries. If a log entry already has a label with the same key as a label in this parameter, then the log entry's label is not changed. See LogEntry."
}

access_token#

OAuth access token.

Type: string

alt#

Data format for response.

Type: string

Potential values: json, media, proto

callback#

JSONP

Type: string

fields#

Selector specifying which fields to include in a partial response.

Type: string

key#

API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.

Type: string

oauth_token#

OAuth 2.0 token for the current user.

Type: string

prettyPrint#

Returns response with indentations and line breaks.

Type: boolean

quotaUser#

Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.

Type: string

uploadType#

Legacy upload protocol for media (e.g. "media", "multipart").

Type: string

upload_protocol#

Upload protocol for media (e.g. "raw", "multipart").

Type: string