From 63d369a60feb3036d50300bc15f762fc7d622caf Mon Sep 17 00:00:00 2001 From: "gcf-owl-bot[bot]" <78513119+gcf-owl-bot[bot]@users.noreply.github.com> Date: Fri, 20 Jan 2023 15:30:46 -0500 Subject: [PATCH 1/2] docs: Add documentation for enums (#166) MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit * docs: Add documentation for enums fix: Add context manager return types chore: Update gapic-generator-python to v1.8.1 PiperOrigin-RevId: 503210727 Source-Link: https://github.com/googleapis/googleapis/commit/a391fd1dac18dfdfa00c18c8404f2c3a6ff8e98e Source-Link: https://github.com/googleapis/googleapis-gen/commit/0080f830dec37c3384157082bce279e37079ea58 Copy-Tag: eyJwIjoiLmdpdGh1Yi8uT3dsQm90LnlhbWwiLCJoIjoiMDA4MGY4MzBkZWMzN2MzMzg0MTU3MDgyYmNlMjc5ZTM3MDc5ZWE1OCJ9 * 🦉 Updates from OwlBot post-processor See https://github.com/googleapis/repo-automation-bots/blob/main/packages/owl-bot/README.md Co-authored-by: Owl Bot --- .../services/flex_templates_service/client.py | 2 +- .../services/jobs_v1_beta3/client.py | 2 +- .../services/messages_v1_beta3/client.py | 2 +- .../services/metrics_v1_beta3/client.py | 2 +- .../services/snapshots_v1_beta3/client.py | 2 +- .../services/templates_service/client.py | 2 +- .../dataflow_v1beta3/types/environment.py | 74 +++++++++ google/cloud/dataflow_v1beta3/types/jobs.py | 152 +++++++++++++++++- .../cloud/dataflow_v1beta3/types/messages.py | 71 +++++++- .../cloud/dataflow_v1beta3/types/metrics.py | 18 ++- .../cloud/dataflow_v1beta3/types/snapshots.py | 19 ++- .../cloud/dataflow_v1beta3/types/templates.py | 59 ++++++- ...ppet_metadata_google.dataflow.v1beta3.json | 2 +- 13 files changed, 391 insertions(+), 16 deletions(-) diff --git a/google/cloud/dataflow_v1beta3/services/flex_templates_service/client.py b/google/cloud/dataflow_v1beta3/services/flex_templates_service/client.py index 98ff5a4..3319093 100644 --- a/google/cloud/dataflow_v1beta3/services/flex_templates_service/client.py +++ b/google/cloud/dataflow_v1beta3/services/flex_templates_service/client.py @@ -507,7 +507,7 @@ def sample_launch_flex_template(): # Done; return the response. return response - def __enter__(self): + def __enter__(self) -> "FlexTemplatesServiceClient": return self def __exit__(self, type, value, traceback): diff --git a/google/cloud/dataflow_v1beta3/services/jobs_v1_beta3/client.py b/google/cloud/dataflow_v1beta3/services/jobs_v1_beta3/client.py index d756776..f532fa4 100644 --- a/google/cloud/dataflow_v1beta3/services/jobs_v1_beta3/client.py +++ b/google/cloud/dataflow_v1beta3/services/jobs_v1_beta3/client.py @@ -1071,7 +1071,7 @@ def sample_snapshot_job(): # Done; return the response. return response - def __enter__(self): + def __enter__(self) -> "JobsV1Beta3Client": return self def __exit__(self, type, value, traceback): diff --git a/google/cloud/dataflow_v1beta3/services/messages_v1_beta3/client.py b/google/cloud/dataflow_v1beta3/services/messages_v1_beta3/client.py index 8934744..cf649ec 100644 --- a/google/cloud/dataflow_v1beta3/services/messages_v1_beta3/client.py +++ b/google/cloud/dataflow_v1beta3/services/messages_v1_beta3/client.py @@ -533,7 +533,7 @@ def sample_list_job_messages(): # Done; return the response. return response - def __enter__(self): + def __enter__(self) -> "MessagesV1Beta3Client": return self def __exit__(self, type, value, traceback): diff --git a/google/cloud/dataflow_v1beta3/services/metrics_v1_beta3/client.py b/google/cloud/dataflow_v1beta3/services/metrics_v1_beta3/client.py index 94b0450..7debffe 100644 --- a/google/cloud/dataflow_v1beta3/services/metrics_v1_beta3/client.py +++ b/google/cloud/dataflow_v1beta3/services/metrics_v1_beta3/client.py @@ -738,7 +738,7 @@ def sample_get_stage_execution_details(): # Done; return the response. return response - def __enter__(self): + def __enter__(self) -> "MetricsV1Beta3Client": return self def __exit__(self, type, value, traceback): diff --git a/google/cloud/dataflow_v1beta3/services/snapshots_v1_beta3/client.py b/google/cloud/dataflow_v1beta3/services/snapshots_v1_beta3/client.py index 3aa0946..26e50bd 100644 --- a/google/cloud/dataflow_v1beta3/services/snapshots_v1_beta3/client.py +++ b/google/cloud/dataflow_v1beta3/services/snapshots_v1_beta3/client.py @@ -675,7 +675,7 @@ def sample_list_snapshots(): # Done; return the response. return response - def __enter__(self): + def __enter__(self) -> "SnapshotsV1Beta3Client": return self def __exit__(self, type, value, traceback): diff --git a/google/cloud/dataflow_v1beta3/services/templates_service/client.py b/google/cloud/dataflow_v1beta3/services/templates_service/client.py index ec8be74..36da1ad 100644 --- a/google/cloud/dataflow_v1beta3/services/templates_service/client.py +++ b/google/cloud/dataflow_v1beta3/services/templates_service/client.py @@ -682,7 +682,7 @@ def sample_get_template(): # Done; return the response. return response - def __enter__(self): + def __enter__(self) -> "TemplatesServiceClient": return self def __exit__(self, type, value, traceback): diff --git a/google/cloud/dataflow_v1beta3/types/environment.py b/google/cloud/dataflow_v1beta3/types/environment.py index 60d03cb..c3f563b 100644 --- a/google/cloud/dataflow_v1beta3/types/environment.py +++ b/google/cloud/dataflow_v1beta3/types/environment.py @@ -47,6 +47,19 @@ class JobType(proto.Enum): [google.dataflow.v1beta3.Job], which determines the way the Job is managed by the Cloud Dataflow service (how workers are scheduled, how inputs are sharded, etc). + + Values: + JOB_TYPE_UNKNOWN (0): + The type of the job is unspecified, or + unknown. + JOB_TYPE_BATCH (1): + A batch job with a well-defined end point: + data is read, data is processed, data is + written, and the job is done. + JOB_TYPE_STREAMING (2): + A continuously streaming job with no end: + data is read, processed, and written + continuously. """ JOB_TYPE_UNKNOWN = 0 JOB_TYPE_BATCH = 1 @@ -56,6 +69,14 @@ class JobType(proto.Enum): class FlexResourceSchedulingGoal(proto.Enum): r"""Specifies the resource to optimize for in Flexible Resource Scheduling. + + Values: + FLEXRS_UNSPECIFIED (0): + Run in the default mode. + FLEXRS_SPEED_OPTIMIZED (1): + Optimize for lower execution time. + FLEXRS_COST_OPTIMIZED (2): + Optimize for lower cost. """ FLEXRS_UNSPECIFIED = 0 FLEXRS_SPEED_OPTIMIZED = 1 @@ -66,6 +87,19 @@ class TeardownPolicy(proto.Enum): r"""Specifies what happens to a resource when a Cloud Dataflow [google.dataflow.v1beta3.Job][google.dataflow.v1beta3.Job] has completed. + + Values: + TEARDOWN_POLICY_UNKNOWN (0): + The teardown policy isn't specified, or is + unknown. + TEARDOWN_ALWAYS (1): + Always teardown the resource. + TEARDOWN_ON_SUCCESS (2): + Teardown the resource on success. This is + useful for debugging failures. + TEARDOWN_NEVER (3): + Never teardown the resource. This is useful + for debugging and development. """ TEARDOWN_POLICY_UNKNOWN = 0 TEARDOWN_ALWAYS = 1 @@ -76,6 +110,21 @@ class TeardownPolicy(proto.Enum): class DefaultPackageSet(proto.Enum): r"""The default set of packages to be staged on a pool of workers. + + Values: + DEFAULT_PACKAGE_SET_UNKNOWN (0): + The default set of packages to stage is + unknown, or unspecified. + DEFAULT_PACKAGE_SET_NONE (1): + Indicates that no packages should be staged + at the worker unless explicitly specified by the + job. + DEFAULT_PACKAGE_SET_JAVA (2): + Stage packages typically useful to workers + written in Java. + DEFAULT_PACKAGE_SET_PYTHON (3): + Stage packages typically useful to workers + written in Python. """ DEFAULT_PACKAGE_SET_UNKNOWN = 0 DEFAULT_PACKAGE_SET_NONE = 1 @@ -88,6 +137,15 @@ class AutoscalingAlgorithm(proto.Enum): worker processes to run at any given point in time, based on the amount of data left to process, the number of workers, and how quickly existing workers are processing data. + + Values: + AUTOSCALING_ALGORITHM_UNKNOWN (0): + The algorithm is unknown, or unspecified. + AUTOSCALING_ALGORITHM_NONE (1): + Disable autoscaling. + AUTOSCALING_ALGORITHM_BASIC (2): + Increase worker count over time to reduce job + execution time. """ AUTOSCALING_ALGORITHM_UNKNOWN = 0 AUTOSCALING_ALGORITHM_NONE = 1 @@ -97,6 +155,14 @@ class AutoscalingAlgorithm(proto.Enum): class WorkerIPAddressConfiguration(proto.Enum): r"""Specifies how IP addresses should be allocated to the worker machines. + + Values: + WORKER_IP_UNSPECIFIED (0): + The configuration is unknown, or unspecified. + WORKER_IP_PUBLIC (1): + Workers should have public IP addresses. + WORKER_IP_PRIVATE (2): + Workers should have private IP addresses. """ WORKER_IP_UNSPECIFIED = 0 WORKER_IP_PUBLIC = 1 @@ -108,6 +174,14 @@ class ShuffleMode(proto.Enum): which determines the approach data is shuffled during processing. More details in: https://cloud.google.com/dataflow/docs/guides/deploying-a-pipeline#dataflow-shuffle + + Values: + SHUFFLE_MODE_UNSPECIFIED (0): + Shuffle mode information is not available. + VM_BASED (1): + Shuffle is done on the worker VMs. + SERVICE_BASED (2): + Shuffle is done on the service side. """ SHUFFLE_MODE_UNSPECIFIED = 0 VM_BASED = 1 diff --git a/google/cloud/dataflow_v1beta3/types/jobs.py b/google/cloud/dataflow_v1beta3/types/jobs.py index 0d2e538..26382c3 100644 --- a/google/cloud/dataflow_v1beta3/types/jobs.py +++ b/google/cloud/dataflow_v1beta3/types/jobs.py @@ -59,7 +59,30 @@ class KindType(proto.Enum): - r"""Type of transform or stage operation.""" + r"""Type of transform or stage operation. + + Values: + UNKNOWN_KIND (0): + Unrecognized transform type. + PAR_DO_KIND (1): + ParDo transform. + GROUP_BY_KEY_KIND (2): + Group By Key transform. + FLATTEN_KIND (3): + Flatten transform. + READ_KIND (4): + Read transform. + WRITE_KIND (5): + Write transform. + CONSTANT_KIND (6): + Constructs from a constant value, such as + with Create.of. + SINGLETON_KIND (7): + Creates a Singleton view of a collection. + SHUFFLE_KIND (8): + Opening or closing a shuffle session, often + as part of a GroupByKey. + """ UNKNOWN_KIND = 0 PAR_DO_KIND = 1 GROUP_BY_KEY_KIND = 2 @@ -74,6 +97,79 @@ class KindType(proto.Enum): class JobState(proto.Enum): r"""Describes the overall state of a [google.dataflow.v1beta3.Job][google.dataflow.v1beta3.Job]. + + Values: + JOB_STATE_UNKNOWN (0): + The job's run state isn't specified. + JOB_STATE_STOPPED (1): + ``JOB_STATE_STOPPED`` indicates that the job has not yet + started to run. + JOB_STATE_RUNNING (2): + ``JOB_STATE_RUNNING`` indicates that the job is currently + running. + JOB_STATE_DONE (3): + ``JOB_STATE_DONE`` indicates that the job has successfully + completed. This is a terminal job state. This state may be + set by the Cloud Dataflow service, as a transition from + ``JOB_STATE_RUNNING``. It may also be set via a Cloud + Dataflow ``UpdateJob`` call, if the job has not yet reached + a terminal state. + JOB_STATE_FAILED (4): + ``JOB_STATE_FAILED`` indicates that the job has failed. This + is a terminal job state. This state may only be set by the + Cloud Dataflow service, and only as a transition from + ``JOB_STATE_RUNNING``. + JOB_STATE_CANCELLED (5): + ``JOB_STATE_CANCELLED`` indicates that the job has been + explicitly cancelled. This is a terminal job state. This + state may only be set via a Cloud Dataflow ``UpdateJob`` + call, and only if the job has not yet reached another + terminal state. + JOB_STATE_UPDATED (6): + ``JOB_STATE_UPDATED`` indicates that the job was + successfully updated, meaning that this job was stopped and + another job was started, inheriting state from this one. + This is a terminal job state. This state may only be set by + the Cloud Dataflow service, and only as a transition from + ``JOB_STATE_RUNNING``. + JOB_STATE_DRAINING (7): + ``JOB_STATE_DRAINING`` indicates that the job is in the + process of draining. A draining job has stopped pulling from + its input sources and is processing any data that remains + in-flight. This state may be set via a Cloud Dataflow + ``UpdateJob`` call, but only as a transition from + ``JOB_STATE_RUNNING``. Jobs that are draining may only + transition to ``JOB_STATE_DRAINED``, + ``JOB_STATE_CANCELLED``, or ``JOB_STATE_FAILED``. + JOB_STATE_DRAINED (8): + ``JOB_STATE_DRAINED`` indicates that the job has been + drained. A drained job terminated by stopping pulling from + its input sources and processing any data that remained + in-flight when draining was requested. This state is a + terminal state, may only be set by the Cloud Dataflow + service, and only as a transition from + ``JOB_STATE_DRAINING``. + JOB_STATE_PENDING (9): + ``JOB_STATE_PENDING`` indicates that the job has been + created but is not yet running. Jobs that are pending may + only transition to ``JOB_STATE_RUNNING``, or + ``JOB_STATE_FAILED``. + JOB_STATE_CANCELLING (10): + ``JOB_STATE_CANCELLING`` indicates that the job has been + explicitly cancelled and is in the process of stopping. Jobs + that are cancelling may only transition to + ``JOB_STATE_CANCELLED`` or ``JOB_STATE_FAILED``. + JOB_STATE_QUEUED (11): + ``JOB_STATE_QUEUED`` indicates that the job has been created + but is being delayed until launch. Jobs that are queued may + only transition to ``JOB_STATE_PENDING`` or + ``JOB_STATE_CANCELLED``. + JOB_STATE_RESOURCE_CLEANING_UP (12): + ``JOB_STATE_RESOURCE_CLEANING_UP`` indicates that the batch + job's associated resources are currently being cleaned up + after a successful run. Currently, this is an opt-in + feature, please reach out to Cloud support team if you are + interested. """ JOB_STATE_UNKNOWN = 0 JOB_STATE_STOPPED = 1 @@ -93,6 +189,24 @@ class JobState(proto.Enum): class JobView(proto.Enum): r"""Selector for how much information is returned in Job responses. + + Values: + JOB_VIEW_UNKNOWN (0): + The job view to return isn't specified, or is unknown. + Responses will contain at least the ``JOB_VIEW_SUMMARY`` + information, and may contain additional information. + JOB_VIEW_SUMMARY (1): + Request summary information only: + Project ID, Job ID, job name, job type, job + status, start/end time, and Cloud SDK version + details. + JOB_VIEW_ALL (2): + Request all information available for this + job. + JOB_VIEW_DESCRIPTION (3): + Request summary info and limited job + description data for steps, labels and + environment. """ JOB_VIEW_UNKNOWN = 0 JOB_VIEW_SUMMARY = 1 @@ -522,7 +636,24 @@ class SdkVersion(proto.Message): """ class SdkSupportStatus(proto.Enum): - r"""The support status of the SDK used to run the job.""" + r"""The support status of the SDK used to run the job. + + Values: + UNKNOWN (0): + Cloud Dataflow is unaware of this version. + SUPPORTED (1): + This is a known version of an SDK, and is + supported. + STALE (2): + A newer version of the SDK family exists, and + an update is recommended. + DEPRECATED (3): + This version of the SDK is deprecated and + will eventually be unsupported. + UNSUPPORTED (4): + Support for this SDK version has ended and it + should no longer be used. + """ UNKNOWN = 0 SUPPORTED = 1 STALE = 2 @@ -1259,6 +1390,23 @@ class Filter(proto.Enum): r"""This field filters out and returns jobs in the specified job state. The order of data returned is determined by the filter used, and is subject to change. + + Values: + UNKNOWN (0): + The filter isn't specified, or is unknown. This returns all + jobs ordered on descending ``JobUuid``. + ALL (1): + Returns all running jobs first ordered on + creation timestamp, then returns all terminated + jobs ordered on the termination timestamp. + TERMINATED (2): + Filters the jobs that have a terminated state, ordered on + the termination timestamp. Example terminated states: + ``JOB_STATE_STOPPED``, ``JOB_STATE_UPDATED``, + ``JOB_STATE_DRAINED``, etc. + ACTIVE (3): + Filters the jobs that are running ordered on + the creation timestamp. """ UNKNOWN = 0 ALL = 1 diff --git a/google/cloud/dataflow_v1beta3/types/messages.py b/google/cloud/dataflow_v1beta3/types/messages.py index e3da083..e2eb88b 100644 --- a/google/cloud/dataflow_v1beta3/types/messages.py +++ b/google/cloud/dataflow_v1beta3/types/messages.py @@ -33,7 +33,48 @@ class JobMessageImportance(proto.Enum): - r"""Indicates the importance of the message.""" + r"""Indicates the importance of the message. + + Values: + JOB_MESSAGE_IMPORTANCE_UNKNOWN (0): + The message importance isn't specified, or is + unknown. + JOB_MESSAGE_DEBUG (1): + The message is at the 'debug' level: + typically only useful for software engineers + working on the code the job is running. + Typically, Dataflow pipeline runners do not + display log messages at this level by default. + JOB_MESSAGE_DETAILED (2): + The message is at the 'detailed' level: + somewhat verbose, but potentially useful to + users. Typically, Dataflow pipeline runners do + not display log messages at this level by + default. These messages are displayed by default + in the Dataflow monitoring UI. + JOB_MESSAGE_BASIC (5): + The message is at the 'basic' level: useful + for keeping track of the execution of a Dataflow + pipeline. Typically, Dataflow pipeline runners + display log messages at this level by default, + and these messages are displayed by default in + the Dataflow monitoring UI. + JOB_MESSAGE_WARNING (3): + The message is at the 'warning' level: + indicating a condition pertaining to a job which + may require human intervention. Typically, + Dataflow pipeline runners display log messages + at this level by default, and these messages are + displayed by default in the Dataflow monitoring + UI. + JOB_MESSAGE_ERROR (4): + The message is at the 'error' level: + indicating a condition preventing a job from + succeeding. Typically, Dataflow pipeline + runners display log messages at this level by + default, and these messages are displayed by + default in the Dataflow monitoring UI. + """ JOB_MESSAGE_IMPORTANCE_UNKNOWN = 0 JOB_MESSAGE_DEBUG = 1 JOB_MESSAGE_DETAILED = 2 @@ -154,7 +195,33 @@ class AutoscalingEvent(proto.Message): """ class AutoscalingEventType(proto.Enum): - r"""Indicates the type of autoscaling event.""" + r"""Indicates the type of autoscaling event. + + Values: + TYPE_UNKNOWN (0): + Default type for the enum. Value should + never be returned. + TARGET_NUM_WORKERS_CHANGED (1): + The TARGET_NUM_WORKERS_CHANGED type should be used when the + target worker pool size has changed at the start of an + actuation. An event should always be specified as + TARGET_NUM_WORKERS_CHANGED if it reflects a change in the + target_num_workers. + CURRENT_NUM_WORKERS_CHANGED (2): + The CURRENT_NUM_WORKERS_CHANGED type should be used when + actual worker pool size has been changed, but the + target_num_workers has not changed. + ACTUATION_FAILURE (3): + The ACTUATION_FAILURE type should be used when we want to + report an error to the user indicating why the current + number of workers in the pool could not be changed. + Displayed in the current status and history widgets. + NO_CHANGE (4): + Used when we want to report to the user a reason why we are + not currently adjusting the number of workers. Should + specify both target_num_workers, current_num_workers and a + decision_message. + """ TYPE_UNKNOWN = 0 TARGET_NUM_WORKERS_CHANGED = 1 CURRENT_NUM_WORKERS_CHANGED = 2 diff --git a/google/cloud/dataflow_v1beta3/types/metrics.py b/google/cloud/dataflow_v1beta3/types/metrics.py index c52f0ad..4e6b45b 100644 --- a/google/cloud/dataflow_v1beta3/types/metrics.py +++ b/google/cloud/dataflow_v1beta3/types/metrics.py @@ -40,7 +40,23 @@ class ExecutionState(proto.Enum): - r"""The state of some component of job execution.""" + r"""The state of some component of job execution. + + Values: + EXECUTION_STATE_UNKNOWN (0): + The component state is unknown or + unspecified. + EXECUTION_STATE_NOT_STARTED (1): + The component is not yet running. + EXECUTION_STATE_RUNNING (2): + The component is currently running. + EXECUTION_STATE_SUCCEEDED (3): + The component succeeded. + EXECUTION_STATE_FAILED (4): + The component failed. + EXECUTION_STATE_CANCELLED (5): + Execution of the component was cancelled. + """ EXECUTION_STATE_UNKNOWN = 0 EXECUTION_STATE_NOT_STARTED = 1 EXECUTION_STATE_RUNNING = 2 diff --git a/google/cloud/dataflow_v1beta3/types/snapshots.py b/google/cloud/dataflow_v1beta3/types/snapshots.py index 926e44b..a9432dd 100644 --- a/google/cloud/dataflow_v1beta3/types/snapshots.py +++ b/google/cloud/dataflow_v1beta3/types/snapshots.py @@ -35,7 +35,24 @@ class SnapshotState(proto.Enum): - r"""Snapshot state.""" + r"""Snapshot state. + + Values: + UNKNOWN_SNAPSHOT_STATE (0): + Unknown state. + PENDING (1): + Snapshot intent to create has been persisted, + snapshotting of state has not yet started. + RUNNING (2): + Snapshotting is being performed. + READY (3): + Snapshot has been created and is ready to be + used. + FAILED (4): + Snapshot failed to be created. + DELETED (5): + Snapshot has been deleted. + """ UNKNOWN_SNAPSHOT_STATE = 0 PENDING = 1 RUNNING = 2 diff --git a/google/cloud/dataflow_v1beta3/types/templates.py b/google/cloud/dataflow_v1beta3/types/templates.py index 1f01262..4118d9e 100644 --- a/google/cloud/dataflow_v1beta3/types/templates.py +++ b/google/cloud/dataflow_v1beta3/types/templates.py @@ -50,6 +50,35 @@ class ParameterType(proto.Enum): r"""ParameterType specifies what kind of input we need for this parameter. + + Values: + DEFAULT (0): + Default input type. + TEXT (1): + The parameter specifies generic text input. + GCS_READ_BUCKET (2): + The parameter specifies a Cloud Storage + Bucket to read from. + GCS_WRITE_BUCKET (3): + The parameter specifies a Cloud Storage + Bucket to write to. + GCS_READ_FILE (4): + The parameter specifies a Cloud Storage file + path to read from. + GCS_WRITE_FILE (5): + The parameter specifies a Cloud Storage file + path to write to. + GCS_READ_FOLDER (6): + The parameter specifies a Cloud Storage + folder path to read from. + GCS_WRITE_FOLDER (7): + The parameter specifies a Cloud Storage + folder to write to. + PUBSUB_TOPIC (8): + The parameter specifies a Pub/Sub Topic. + PUBSUB_SUBSCRIPTION (9): + The parameter specifies a Pub/Sub + Subscription. """ DEFAULT = 0 TEXT = 1 @@ -681,7 +710,16 @@ class SDKInfo(proto.Message): """ class Language(proto.Enum): - r"""SDK Language.""" + r"""SDK Language. + + Values: + UNKNOWN (0): + UNKNOWN Language. + JAVA (1): + Java. + PYTHON (2): + Python. + """ UNKNOWN = 0 JAVA = 1 PYTHON = 2 @@ -801,7 +839,13 @@ class GetTemplateRequest(proto.Message): """ class TemplateView(proto.Enum): - r"""The various views of a template that may be retrieved.""" + r"""The various views of a template that may be retrieved. + + Values: + METADATA_ONLY (0): + Template view that retrieves only the + metadata associated with the template. + """ METADATA_ONLY = 0 project_id: str = proto.Field( @@ -842,7 +886,16 @@ class GetTemplateResponse(proto.Message): """ class TemplateType(proto.Enum): - r"""Template Type.""" + r"""Template Type. + + Values: + UNKNOWN (0): + Unknown Template Type. + LEGACY (1): + Legacy Template. + FLEX (2): + Flex Template. + """ UNKNOWN = 0 LEGACY = 1 FLEX = 2 diff --git a/samples/generated_samples/snippet_metadata_google.dataflow.v1beta3.json b/samples/generated_samples/snippet_metadata_google.dataflow.v1beta3.json index 654cdc2..07ae5fa 100644 --- a/samples/generated_samples/snippet_metadata_google.dataflow.v1beta3.json +++ b/samples/generated_samples/snippet_metadata_google.dataflow.v1beta3.json @@ -8,7 +8,7 @@ ], "language": "PYTHON", "name": "google-cloud-dataflow-client", - "version": "0.8.0" + "version": "0.1.0" }, "snippets": [ { From 8f7d71087c24c9a8f7aa3e706768440d154891df Mon Sep 17 00:00:00 2001 From: "release-please[bot]" <55107282+release-please[bot]@users.noreply.github.com> Date: Mon, 23 Jan 2023 10:34:18 -0500 Subject: [PATCH 2/2] chore(main): release 0.8.1 (#167) Co-authored-by: release-please[bot] <55107282+release-please[bot]@users.noreply.github.com> --- .release-please-manifest.json | 2 +- CHANGELOG.md | 12 ++++++++++++ google/cloud/dataflow/gapic_version.py | 2 +- google/cloud/dataflow_v1beta3/gapic_version.py | 2 +- .../snippet_metadata_google.dataflow.v1beta3.json | 2 +- 5 files changed, 16 insertions(+), 4 deletions(-) diff --git a/.release-please-manifest.json b/.release-please-manifest.json index 64f3cdd..02f17d9 100644 --- a/.release-please-manifest.json +++ b/.release-please-manifest.json @@ -1,3 +1,3 @@ { - ".": "0.8.0" + ".": "0.8.1" } diff --git a/CHANGELOG.md b/CHANGELOG.md index 147e14f..f64032c 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,17 @@ # Changelog +## [0.8.1](https://github.com/googleapis/python-dataflow-client/compare/v0.8.0...v0.8.1) (2023-01-20) + + +### Bug Fixes + +* Add context manager return types ([63d369a](https://github.com/googleapis/python-dataflow-client/commit/63d369a60feb3036d50300bc15f762fc7d622caf)) + + +### Documentation + +* Add documentation for enums ([63d369a](https://github.com/googleapis/python-dataflow-client/commit/63d369a60feb3036d50300bc15f762fc7d622caf)) + ## [0.8.0](https://github.com/googleapis/python-dataflow-client/compare/v0.7.0...v0.8.0) (2023-01-10) diff --git a/google/cloud/dataflow/gapic_version.py b/google/cloud/dataflow/gapic_version.py index b88d7a7..48d7dde 100644 --- a/google/cloud/dataflow/gapic_version.py +++ b/google/cloud/dataflow/gapic_version.py @@ -13,4 +13,4 @@ # See the License for the specific language governing permissions and # limitations under the License. # -__version__ = "0.8.0" # {x-release-please-version} +__version__ = "0.8.1" # {x-release-please-version} diff --git a/google/cloud/dataflow_v1beta3/gapic_version.py b/google/cloud/dataflow_v1beta3/gapic_version.py index b88d7a7..48d7dde 100644 --- a/google/cloud/dataflow_v1beta3/gapic_version.py +++ b/google/cloud/dataflow_v1beta3/gapic_version.py @@ -13,4 +13,4 @@ # See the License for the specific language governing permissions and # limitations under the License. # -__version__ = "0.8.0" # {x-release-please-version} +__version__ = "0.8.1" # {x-release-please-version} diff --git a/samples/generated_samples/snippet_metadata_google.dataflow.v1beta3.json b/samples/generated_samples/snippet_metadata_google.dataflow.v1beta3.json index 07ae5fa..a8534ee 100644 --- a/samples/generated_samples/snippet_metadata_google.dataflow.v1beta3.json +++ b/samples/generated_samples/snippet_metadata_google.dataflow.v1beta3.json @@ -8,7 +8,7 @@ ], "language": "PYTHON", "name": "google-cloud-dataflow-client", - "version": "0.1.0" + "version": "0.8.1" }, "snippets": [ {