From 64d9c481df1c2737189dcb575c69f2968c0aa034 Mon Sep 17 00:00:00 2001 From: "gcf-owl-bot[bot]" <78513119+gcf-owl-bot[bot]@users.noreply.github.com> Date: Fri, 30 Sep 2022 10:45:07 -0400 Subject: [PATCH 01/10] feat: Add support for notebook tasks (#100) MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit * feat: Add support for notebook tasks PiperOrigin-RevId: 477935449 Source-Link: https://github.com/googleapis/googleapis/commit/9e3a1dd591e5a5ace521726f79a265eaab52843b Source-Link: https://github.com/googleapis/googleapis-gen/commit/a5443ab48e30bdd608ae7ae1d6cf2fabc17b6dca Copy-Tag: eyJwIjoiLmdpdGh1Yi8uT3dsQm90LnlhbWwiLCJoIjoiYTU0NDNhYjQ4ZTMwYmRkNjA4YWU3YWUxZDZjZjJmYWJjMTdiNmRjYSJ9 * 🦉 Updates from OwlBot post-processor See https://github.com/googleapis/repo-automation-bots/blob/main/packages/owl-bot/README.md Co-authored-by: Owl Bot --- google/cloud/dataplex_v1/types/logs.py | 24 ++++++++++- google/cloud/dataplex_v1/types/tasks.py | 56 +++++++++++++++++++++++++ 2 files changed, 79 insertions(+), 1 deletion(-) diff --git a/google/cloud/dataplex_v1/types/logs.py b/google/cloud/dataplex_v1/types/logs.py index 4259b06..626667f 100644 --- a/google/cloud/dataplex_v1/types/logs.py +++ b/google/cloud/dataplex_v1/types/logs.py @@ -328,7 +328,8 @@ class SessionEvent(proto.Message): The log message. user_id (str): The information about the user that created - the session. + the session. It will be the email address of the + user. session_id (str): Unique identifier for the session. type_ (google.cloud.dataplex_v1.types.SessionEvent.EventType): @@ -337,6 +338,13 @@ class SessionEvent(proto.Message): The execution details of the query. This field is a member of `oneof`_ ``detail``. + event_succeeded (bool): + The status of the event. + warm_pool_enabled (bool): + If the session is a warm pooled session. + unassigned_duration (google.protobuf.duration_pb2.Duration): + The idle duration of a warm pooled session + before it is assigned to user. """ class EventType(proto.Enum): @@ -345,6 +353,7 @@ class EventType(proto.Enum): START = 1 STOP = 2 QUERY = 3 + CREATE = 4 class QueryDetail(proto.Message): r"""Execution details of the query. @@ -420,6 +429,19 @@ class Engine(proto.Enum): oneof="detail", message=QueryDetail, ) + event_succeeded = proto.Field( + proto.BOOL, + number=6, + ) + warm_pool_enabled = proto.Field( + proto.BOOL, + number=7, + ) + unassigned_duration = proto.Field( + proto.MESSAGE, + number=8, + message=duration_pb2.Duration, + ) __all__ = tuple(sorted(__protobuf__.manifest)) diff --git a/google/cloud/dataplex_v1/types/tasks.py b/google/cloud/dataplex_v1/types/tasks.py index 059a6df..431e693 100644 --- a/google/cloud/dataplex_v1/types/tasks.py +++ b/google/cloud/dataplex_v1/types/tasks.py @@ -31,6 +31,11 @@ class Task(proto.Message): r"""A task represents a user-visible job. + This message has `oneof`_ fields (mutually exclusive fields). + For each oneof, at most one member field can be set at the same time. + Setting any member of the oneof automatically clears all other + members. + .. _oneof: https://proto-plus-python.readthedocs.io/en/stable/fields.html#oneofs-mutually-exclusive-fields Attributes: @@ -70,6 +75,11 @@ class Task(proto.Message): spark (google.cloud.dataplex_v1.types.Task.SparkTaskConfig): Config related to running custom Spark tasks. + This field is a member of `oneof`_ ``config``. + notebook (google.cloud.dataplex_v1.types.Task.NotebookTaskConfig): + Config related to running scheduled + Notebooks. + This field is a member of `oneof`_ ``config``. """ @@ -445,6 +455,46 @@ class SparkTaskConfig(proto.Message): message="Task.InfrastructureSpec", ) + class NotebookTaskConfig(proto.Message): + r"""Config for running scheduled notebooks. + + Attributes: + notebook (str): + Required. Path to input notebook. This can be the GCS URI of + the notebook file or the path to a Notebook Content. The + execution args are accessible as environment variables + (``TASK_key=value``). + infrastructure_spec (google.cloud.dataplex_v1.types.Task.InfrastructureSpec): + Optional. Infrastructure specification for + the execution. + file_uris (Sequence[str]): + Optional. GCS URIs of files to be placed in + the working directory of each executor. + archive_uris (Sequence[str]): + Optional. GCS URIs of archives to be + extracted into the working directory of each + executor. Supported file types: .jar, .tar, + .tar.gz, .tgz, and .zip. + """ + + notebook = proto.Field( + proto.STRING, + number=4, + ) + infrastructure_spec = proto.Field( + proto.MESSAGE, + number=3, + message="Task.InfrastructureSpec", + ) + file_uris = proto.RepeatedField( + proto.STRING, + number=5, + ) + archive_uris = proto.RepeatedField( + proto.STRING, + number=6, + ) + class ExecutionStatus(proto.Message): r"""Status of the task execution (e.g. Jobs). @@ -523,6 +573,12 @@ class ExecutionStatus(proto.Message): oneof="config", message=SparkTaskConfig, ) + notebook = proto.Field( + proto.MESSAGE, + number=302, + oneof="config", + message=NotebookTaskConfig, + ) class Job(proto.Message): From cb0adf1108e26df775b65ddf13832c7bf9a7e2b8 Mon Sep 17 00:00:00 2001 From: "gcf-owl-bot[bot]" <78513119+gcf-owl-bot[bot]@users.noreply.github.com> Date: Thu, 6 Oct 2022 09:50:57 -0400 Subject: [PATCH 02/10] chore: rename warm_pool_enabled to fast_startup_enabled (#102) MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit * chore: rename warm_pool_enabled to fast_startup_enabled While this would normally be a breaking change, libraries haven't been released yet with the old value warm_pool_enabled. PiperOrigin-RevId: 479215286 Source-Link: https://github.com/googleapis/googleapis/commit/767dc45050047fd0e6a9ceb3a56ed6cdb4cb3ed2 Source-Link: https://github.com/googleapis/googleapis-gen/commit/d497faaa460ae68eb9e87f1471f2ea300d426145 Copy-Tag: eyJwIjoiLmdpdGh1Yi8uT3dsQm90LnlhbWwiLCJoIjoiZDQ5N2ZhYWE0NjBhZTY4ZWI5ZTg3ZjE0NzFmMmVhMzAwZDQyNjE0NSJ9 * 🦉 Updates from OwlBot post-processor See https://github.com/googleapis/repo-automation-bots/blob/main/packages/owl-bot/README.md Co-authored-by: Owl Bot --- google/cloud/dataplex_v1/types/logs.py | 8 +++++--- google/cloud/dataplex_v1/types/tasks.py | 17 +++++++++-------- 2 files changed, 14 insertions(+), 11 deletions(-) diff --git a/google/cloud/dataplex_v1/types/logs.py b/google/cloud/dataplex_v1/types/logs.py index 626667f..39db8b3 100644 --- a/google/cloud/dataplex_v1/types/logs.py +++ b/google/cloud/dataplex_v1/types/logs.py @@ -340,8 +340,10 @@ class SessionEvent(proto.Message): This field is a member of `oneof`_ ``detail``. event_succeeded (bool): The status of the event. - warm_pool_enabled (bool): - If the session is a warm pooled session. + fast_startup_enabled (bool): + If the session is associated with an + Environment with fast startup enabled, and was + pre-created before being assigned to a user. unassigned_duration (google.protobuf.duration_pb2.Duration): The idle duration of a warm pooled session before it is assigned to user. @@ -433,7 +435,7 @@ class Engine(proto.Enum): proto.BOOL, number=6, ) - warm_pool_enabled = proto.Field( + fast_startup_enabled = proto.Field( proto.BOOL, number=7, ) diff --git a/google/cloud/dataplex_v1/types/tasks.py b/google/cloud/dataplex_v1/types/tasks.py index 431e693..288ba38 100644 --- a/google/cloud/dataplex_v1/types/tasks.py +++ b/google/cloud/dataplex_v1/types/tasks.py @@ -460,19 +460,20 @@ class NotebookTaskConfig(proto.Message): Attributes: notebook (str): - Required. Path to input notebook. This can be the GCS URI of - the notebook file or the path to a Notebook Content. The - execution args are accessible as environment variables - (``TASK_key=value``). + Required. Path to input notebook. This can be the Cloud + Storage URI of the notebook file or the path to a Notebook + Content. The execution args are accessible as environment + variables (``TASK_key=value``). infrastructure_spec (google.cloud.dataplex_v1.types.Task.InfrastructureSpec): Optional. Infrastructure specification for the execution. file_uris (Sequence[str]): - Optional. GCS URIs of files to be placed in - the working directory of each executor. + Optional. Cloud Storage URIs of files to be + placed in the working directory of each + executor. archive_uris (Sequence[str]): - Optional. GCS URIs of archives to be - extracted into the working directory of each + Optional. Cloud Storage URIs of archives to + be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip. """ From 65adbb31c94794f27a78b309550c519734a7b030 Mon Sep 17 00:00:00 2001 From: Anthonios Partheniou Date: Fri, 7 Oct 2022 18:14:27 -0400 Subject: [PATCH 03/10] fix(deps): allow protobuf 3.19.5 (#103) * fix(deps): allow protobuf 3.19.5 * explicitly exclude protobuf 4.21.0 --- setup.py | 2 +- testing/constraints-3.7.txt | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/setup.py b/setup.py index 83df24c..3fc72c3 100644 --- a/setup.py +++ b/setup.py @@ -27,7 +27,7 @@ dependencies = [ "google-api-core[grpc] >= 1.32.0, <3.0.0dev,!=2.0.*,!=2.1.*,!=2.2.*,!=2.3.*,!=2.4.*,!=2.5.*,!=2.6.*,!=2.7.*", "proto-plus >= 1.22.0, <2.0.0dev", - "protobuf >= 3.20.2, <5.0.0dev", + "protobuf>=3.19.5,<5.0.0dev,!=3.20.0,!=3.20.1,!=4.21.0,!=4.21.1,!=4.21.2,!=4.21.3,!=4.21.4,!=4.21.5", "grpc-google-iam-v1 >=0.12.4, <1.0.0dev", ] diff --git a/testing/constraints-3.7.txt b/testing/constraints-3.7.txt index f6f583b..b0cdb69 100644 --- a/testing/constraints-3.7.txt +++ b/testing/constraints-3.7.txt @@ -6,5 +6,5 @@ # Then this file should have google-cloud-foo==1.14.0 google-api-core==1.32.0 proto-plus==1.22.0 -protobuf==3.20.2 +protobuf==3.19.5 grpc-google-iam-v1==0.12.4 From fa75c80939bf3d872092c64ad08809b29e295d16 Mon Sep 17 00:00:00 2001 From: "gcf-owl-bot[bot]" <78513119+gcf-owl-bot[bot]@users.noreply.github.com> Date: Tue, 8 Nov 2022 02:02:25 +0000 Subject: [PATCH 04/10] chore(python): update dependencies in .kokoro/requirements.txt [autoapprove] (#108) Source-Link: https://togithub.com/googleapis/synthtool/commit/e3a1277ac35fc88c09db1930533e24292b132ced Post-Processor: gcr.io/cloud-devrel-public-resources/owlbot-python:latest@sha256:452901c74a22f9b9a3bd02bce780b8e8805c97270d424684bff809ce5be8c2a2 --- .github/.OwlBot.lock.yaml | 2 +- .kokoro/requirements.txt | 325 +++++++++++++++++++++----------------- noxfile.py | 11 +- 3 files changed, 187 insertions(+), 151 deletions(-) diff --git a/.github/.OwlBot.lock.yaml b/.github/.OwlBot.lock.yaml index 3815c98..12edee7 100644 --- a/.github/.OwlBot.lock.yaml +++ b/.github/.OwlBot.lock.yaml @@ -13,4 +13,4 @@ # limitations under the License. docker: image: gcr.io/cloud-devrel-public-resources/owlbot-python:latest - digest: sha256:7a40313731a7cb1454eef6b33d3446ebb121836738dc3ab3d2d3ded5268c35b6 + digest: sha256:452901c74a22f9b9a3bd02bce780b8e8805c97270d424684bff809ce5be8c2a2 diff --git a/.kokoro/requirements.txt b/.kokoro/requirements.txt index d15994b..31425f1 100644 --- a/.kokoro/requirements.txt +++ b/.kokoro/requirements.txt @@ -20,9 +20,9 @@ cachetools==5.2.0 \ --hash=sha256:6a94c6402995a99c3970cc7e4884bb60b4a8639938157eeed436098bf9831757 \ --hash=sha256:f9f17d2aec496a9aa6b76f53e3b614c965223c061982d434d160f930c698a9db # via google-auth -certifi==2022.6.15 \ - --hash=sha256:84c85a9078b11105f04f3036a9482ae10e4621616db313fe045dd24743a0820d \ - --hash=sha256:fe86415d55e84719d75f8b69414f6438ac3547d2078ab91b67e779ef69378412 +certifi==2022.9.24 \ + --hash=sha256:0d9c601124e5a6ba9712dbc60d9c53c21e34f5f641fe83002317394311bdce14 \ + --hash=sha256:90c1a32f1d68f940488354e36370f6cca89f0f106db09518524c88d6ed83f382 # via requests cffi==1.15.1 \ --hash=sha256:00a9ed42e88df81ffae7a8ab6d9356b371399b91dbdf0c3cb1e84c03a13aceb5 \ @@ -110,29 +110,33 @@ commonmark==0.9.1 \ --hash=sha256:452f9dc859be7f06631ddcb328b6919c67984aca654e5fefb3914d54691aed60 \ --hash=sha256:da2f38c92590f83de410ba1a3cbceafbc74fee9def35f9251ba9a971d6d66fd9 # via rich -cryptography==37.0.4 \ - --hash=sha256:190f82f3e87033821828f60787cfa42bff98404483577b591429ed99bed39d59 \ - --hash=sha256:2be53f9f5505673eeda5f2736bea736c40f051a739bfae2f92d18aed1eb54596 \ - --hash=sha256:30788e070800fec9bbcf9faa71ea6d8068f5136f60029759fd8c3efec3c9dcb3 \ - --hash=sha256:3d41b965b3380f10e4611dbae366f6dc3cefc7c9ac4e8842a806b9672ae9add5 \ - --hash=sha256:4c590ec31550a724ef893c50f9a97a0c14e9c851c85621c5650d699a7b88f7ab \ - --hash=sha256:549153378611c0cca1042f20fd9c5030d37a72f634c9326e225c9f666d472884 \ - --hash=sha256:63f9c17c0e2474ccbebc9302ce2f07b55b3b3fcb211ded18a42d5764f5c10a82 \ - --hash=sha256:6bc95ed67b6741b2607298f9ea4932ff157e570ef456ef7ff0ef4884a134cc4b \ - --hash=sha256:7099a8d55cd49b737ffc99c17de504f2257e3787e02abe6d1a6d136574873441 \ - --hash=sha256:75976c217f10d48a8b5a8de3d70c454c249e4b91851f6838a4e48b8f41eb71aa \ - --hash=sha256:7bc997818309f56c0038a33b8da5c0bfbb3f1f067f315f9abd6fc07ad359398d \ - --hash=sha256:80f49023dd13ba35f7c34072fa17f604d2f19bf0989f292cedf7ab5770b87a0b \ - --hash=sha256:91ce48d35f4e3d3f1d83e29ef4a9267246e6a3be51864a5b7d2247d5086fa99a \ - --hash=sha256:a958c52505c8adf0d3822703078580d2c0456dd1d27fabfb6f76fe63d2971cd6 \ - --hash=sha256:b62439d7cd1222f3da897e9a9fe53bbf5c104fff4d60893ad1355d4c14a24157 \ - --hash=sha256:b7f8dd0d4c1f21759695c05a5ec8536c12f31611541f8904083f3dc582604280 \ - --hash=sha256:d204833f3c8a33bbe11eda63a54b1aad7aa7456ed769a982f21ec599ba5fa282 \ - --hash=sha256:e007f052ed10cc316df59bc90fbb7ff7950d7e2919c9757fd42a2b8ecf8a5f67 \ - --hash=sha256:f2dcb0b3b63afb6df7fd94ec6fbddac81b5492513f7b0436210d390c14d46ee8 \ - --hash=sha256:f721d1885ecae9078c3f6bbe8a88bc0786b6e749bf32ccec1ef2b18929a05046 \ - --hash=sha256:f7a6de3e98771e183645181b3627e2563dcde3ce94a9e42a3f427d2255190327 \ - --hash=sha256:f8c0a6e9e1dd3eb0414ba320f85da6b0dcbd543126e30fcc546e7372a7fbf3b9 +cryptography==38.0.3 \ + --hash=sha256:068147f32fa662c81aebab95c74679b401b12b57494872886eb5c1139250ec5d \ + --hash=sha256:06fc3cc7b6f6cca87bd56ec80a580c88f1da5306f505876a71c8cfa7050257dd \ + --hash=sha256:25c1d1f19729fb09d42e06b4bf9895212292cb27bb50229f5aa64d039ab29146 \ + --hash=sha256:402852a0aea73833d982cabb6d0c3bb582c15483d29fb7085ef2c42bfa7e38d7 \ + --hash=sha256:4e269dcd9b102c5a3d72be3c45d8ce20377b8076a43cbed6f660a1afe365e436 \ + --hash=sha256:5419a127426084933076132d317911e3c6eb77568a1ce23c3ac1e12d111e61e0 \ + --hash=sha256:554bec92ee7d1e9d10ded2f7e92a5d70c1f74ba9524947c0ba0c850c7b011828 \ + --hash=sha256:5e89468fbd2fcd733b5899333bc54d0d06c80e04cd23d8c6f3e0542358c6060b \ + --hash=sha256:65535bc550b70bd6271984d9863a37741352b4aad6fb1b3344a54e6950249b55 \ + --hash=sha256:6ab9516b85bebe7aa83f309bacc5f44a61eeb90d0b4ec125d2d003ce41932d36 \ + --hash=sha256:6addc3b6d593cd980989261dc1cce38263c76954d758c3c94de51f1e010c9a50 \ + --hash=sha256:728f2694fa743a996d7784a6194da430f197d5c58e2f4e278612b359f455e4a2 \ + --hash=sha256:785e4056b5a8b28f05a533fab69febf5004458e20dad7e2e13a3120d8ecec75a \ + --hash=sha256:78cf5eefac2b52c10398a42765bfa981ce2372cbc0457e6bf9658f41ec3c41d8 \ + --hash=sha256:7f836217000342d448e1c9a342e9163149e45d5b5eca76a30e84503a5a96cab0 \ + --hash=sha256:8d41a46251bf0634e21fac50ffd643216ccecfaf3701a063257fe0b2be1b6548 \ + --hash=sha256:984fe150f350a3c91e84de405fe49e688aa6092b3525f407a18b9646f6612320 \ + --hash=sha256:9b24bcff7853ed18a63cfb0c2b008936a9554af24af2fb146e16d8e1aed75748 \ + --hash=sha256:b1b35d9d3a65542ed2e9d90115dfd16bbc027b3f07ee3304fc83580f26e43249 \ + --hash=sha256:b1b52c9e5f8aa2b802d48bd693190341fae201ea51c7a167d69fc48b60e8a959 \ + --hash=sha256:bbf203f1a814007ce24bd4d51362991d5cb90ba0c177a9c08825f2cc304d871f \ + --hash=sha256:be243c7e2bfcf6cc4cb350c0d5cdf15ca6383bbcb2a8ef51d3c9411a9d4386f0 \ + --hash=sha256:bfbe6ee19615b07a98b1d2287d6a6073f734735b49ee45b11324d85efc4d5cbd \ + --hash=sha256:c46837ea467ed1efea562bbeb543994c2d1f6e800785bd5a2c98bc096f5cb220 \ + --hash=sha256:dfb4f4dd568de1b6af9f4cda334adf7d72cf5bc052516e1b2608b683375dd95c \ + --hash=sha256:ed7b00096790213e09eb11c97cc6e2b757f15f3d2f85833cd2d3ec3fe37c1722 # via # gcp-releasetool # secretstorage @@ -148,23 +152,23 @@ filelock==3.8.0 \ --hash=sha256:55447caa666f2198c5b6b13a26d2084d26fa5b115c00d065664b2124680c4edc \ --hash=sha256:617eb4e5eedc82fc5f47b6d61e4d11cb837c56cb4544e39081099fa17ad109d4 # via virtualenv -gcp-docuploader==0.6.3 \ - --hash=sha256:ba8c9d76b3bbac54b0311c503a373b00edc2dc02d6d54ea9507045adb8e870f7 \ - --hash=sha256:c0f5aaa82ce1854a386197e4e359b120ad6d4e57ae2c812fce42219a3288026b +gcp-docuploader==0.6.4 \ + --hash=sha256:01486419e24633af78fd0167db74a2763974765ee8078ca6eb6964d0ebd388af \ + --hash=sha256:70861190c123d907b3b067da896265ead2eeb9263969d6955c9e0bb091b5ccbf # via -r requirements.in -gcp-releasetool==1.8.7 \ - --hash=sha256:3d2a67c9db39322194afb3b427e9cb0476ce8f2a04033695f0aeb63979fc2b37 \ - --hash=sha256:5e4d28f66e90780d77f3ecf1e9155852b0c3b13cbccb08ab07e66b2357c8da8d +gcp-releasetool==1.9.1 \ + --hash=sha256:952f4055d5d986b070ae2a71c4410b250000f9cc5a1e26398fcd55a5bbc5a15f \ + --hash=sha256:d0d3c814a97c1a237517e837d8cfa668ced8df4b882452578ecef4a4e79c583b # via -r requirements.in -google-api-core==2.8.2 \ - --hash=sha256:06f7244c640322b508b125903bb5701bebabce8832f85aba9335ec00b3d02edc \ - --hash=sha256:93c6a91ccac79079ac6bbf8b74ee75db970cc899278b97d53bc012f35908cf50 +google-api-core==2.10.2 \ + --hash=sha256:10c06f7739fe57781f87523375e8e1a3a4674bf6392cd6131a3222182b971320 \ + --hash=sha256:34f24bd1d5f72a8c4519773d99ca6bf080a6c4e041b4e9f024fe230191dda62e # via # google-cloud-core # google-cloud-storage -google-auth==2.11.0 \ - --hash=sha256:be62acaae38d0049c21ca90f27a23847245c9f161ff54ede13af2cb6afecbac9 \ - --hash=sha256:ed65ecf9f681832298e29328e1ef0a3676e3732b2e56f41532d45f70a22de0fb +google-auth==2.14.0 \ + --hash=sha256:1ad5b0e6eba5f69645971abb3d2c197537d5914070a8c6d30299dfdb07c5c700 \ + --hash=sha256:cf24817855d874ede2efd071aa22125445f555de1685b739a9782fcf408c2a3d # via # gcp-releasetool # google-api-core @@ -178,72 +182,97 @@ google-cloud-storage==2.5.0 \ --hash=sha256:19a26c66c317ce542cea0830b7e787e8dac2588b6bfa4d3fd3b871ba16305ab0 \ --hash=sha256:382f34b91de2212e3c2e7b40ec079d27ee2e3dbbae99b75b1bcd8c63063ce235 # via gcp-docuploader -google-crc32c==1.3.0 \ - --hash=sha256:04e7c220798a72fd0f08242bc8d7a05986b2a08a0573396187fd32c1dcdd58b3 \ - --hash=sha256:05340b60bf05b574159e9bd940152a47d38af3fb43803ffe71f11d704b7696a6 \ - --hash=sha256:12674a4c3b56b706153a358eaa1018c4137a5a04635b92b4652440d3d7386206 \ - --hash=sha256:127f9cc3ac41b6a859bd9dc4321097b1a4f6aa7fdf71b4f9227b9e3ebffb4422 \ - --hash=sha256:13af315c3a0eec8bb8b8d80b8b128cb3fcd17d7e4edafc39647846345a3f003a \ - --hash=sha256:1926fd8de0acb9d15ee757175ce7242e235482a783cd4ec711cc999fc103c24e \ - --hash=sha256:226f2f9b8e128a6ca6a9af9b9e8384f7b53a801907425c9a292553a3a7218ce0 \ - --hash=sha256:276de6273eb074a35bc598f8efbc00c7869c5cf2e29c90748fccc8c898c244df \ - --hash=sha256:318f73f5484b5671f0c7f5f63741ab020a599504ed81d209b5c7129ee4667407 \ - --hash=sha256:3bbce1be3687bbfebe29abdb7631b83e6b25da3f4e1856a1611eb21854b689ea \ - --hash=sha256:42ae4781333e331a1743445931b08ebdad73e188fd554259e772556fc4937c48 \ - --hash=sha256:58be56ae0529c664cc04a9c76e68bb92b091e0194d6e3c50bea7e0f266f73713 \ - --hash=sha256:5da2c81575cc3ccf05d9830f9e8d3c70954819ca9a63828210498c0774fda1a3 \ - --hash=sha256:6311853aa2bba4064d0c28ca54e7b50c4d48e3de04f6770f6c60ebda1e975267 \ - --hash=sha256:650e2917660e696041ab3dcd7abac160b4121cd9a484c08406f24c5964099829 \ - --hash=sha256:6a4db36f9721fdf391646685ecffa404eb986cbe007a3289499020daf72e88a2 \ - --hash=sha256:779cbf1ce375b96111db98fca913c1f5ec11b1d870e529b1dc7354b2681a8c3a \ - --hash=sha256:7f6fe42536d9dcd3e2ffb9d3053f5d05221ae3bbcefbe472bdf2c71c793e3183 \ - --hash=sha256:891f712ce54e0d631370e1f4997b3f182f3368179198efc30d477c75d1f44942 \ - --hash=sha256:95c68a4b9b7828ba0428f8f7e3109c5d476ca44996ed9a5f8aac6269296e2d59 \ - --hash=sha256:96a8918a78d5d64e07c8ea4ed2bc44354e3f93f46a4866a40e8db934e4c0d74b \ - --hash=sha256:9c3cf890c3c0ecfe1510a452a165431b5831e24160c5fcf2071f0f85ca5a47cd \ - --hash=sha256:9f58099ad7affc0754ae42e6d87443299f15d739b0ce03c76f515153a5cda06c \ - --hash=sha256:a0b9e622c3b2b8d0ce32f77eba617ab0d6768b82836391e4f8f9e2074582bf02 \ - --hash=sha256:a7f9cbea4245ee36190f85fe1814e2d7b1e5f2186381b082f5d59f99b7f11328 \ - --hash=sha256:bab4aebd525218bab4ee615786c4581952eadc16b1ff031813a2fd51f0cc7b08 \ - --hash=sha256:c124b8c8779bf2d35d9b721e52d4adb41c9bfbde45e6a3f25f0820caa9aba73f \ - --hash=sha256:c9da0a39b53d2fab3e5467329ed50e951eb91386e9d0d5b12daf593973c3b168 \ - --hash=sha256:ca60076c388728d3b6ac3846842474f4250c91efbfe5afa872d3ffd69dd4b318 \ - --hash=sha256:cb6994fff247987c66a8a4e550ef374671c2b82e3c0d2115e689d21e511a652d \ - --hash=sha256:d1c1d6236feab51200272d79b3d3e0f12cf2cbb12b208c835b175a21efdb0a73 \ - --hash=sha256:dd7760a88a8d3d705ff562aa93f8445ead54f58fd482e4f9e2bafb7e177375d4 \ - --hash=sha256:dda4d8a3bb0b50f540f6ff4b6033f3a74e8bf0bd5320b70fab2c03e512a62812 \ - --hash=sha256:e0f1ff55dde0ebcfbef027edc21f71c205845585fffe30d4ec4979416613e9b3 \ - --hash=sha256:e7a539b9be7b9c00f11ef16b55486141bc2cdb0c54762f84e3c6fc091917436d \ - --hash=sha256:eb0b14523758e37802f27b7f8cd973f5f3d33be7613952c0df904b68c4842f0e \ - --hash=sha256:ed447680ff21c14aaceb6a9f99a5f639f583ccfe4ce1a5e1d48eb41c3d6b3217 \ - --hash=sha256:f52a4ad2568314ee713715b1e2d79ab55fab11e8b304fd1462ff5cccf4264b3e \ - --hash=sha256:fbd60c6aaa07c31d7754edbc2334aef50601b7f1ada67a96eb1eb57c7c72378f \ - --hash=sha256:fc28e0db232c62ca0c3600884933178f0825c99be4474cdd645e378a10588125 \ - --hash=sha256:fe31de3002e7b08eb20823b3735b97c86c5926dd0581c7710a680b418a8709d4 \ - --hash=sha256:fec221a051150eeddfdfcff162e6db92c65ecf46cb0f7bb1bf812a1520ec026b \ - --hash=sha256:ff71073ebf0e42258a42a0b34f2c09ec384977e7f6808999102eedd5b49920e3 +google-crc32c==1.5.0 \ + --hash=sha256:024894d9d3cfbc5943f8f230e23950cd4906b2fe004c72e29b209420a1e6b05a \ + --hash=sha256:02c65b9817512edc6a4ae7c7e987fea799d2e0ee40c53ec573a692bee24de876 \ + --hash=sha256:02ebb8bf46c13e36998aeaad1de9b48f4caf545e91d14041270d9dca767b780c \ + --hash=sha256:07eb3c611ce363c51a933bf6bd7f8e3878a51d124acfc89452a75120bc436289 \ + --hash=sha256:1034d91442ead5a95b5aaef90dbfaca8633b0247d1e41621d1e9f9db88c36298 \ + --hash=sha256:116a7c3c616dd14a3de8c64a965828b197e5f2d121fedd2f8c5585c547e87b02 \ + --hash=sha256:19e0a019d2c4dcc5e598cd4a4bc7b008546b0358bd322537c74ad47a5386884f \ + --hash=sha256:1c7abdac90433b09bad6c43a43af253e688c9cfc1c86d332aed13f9a7c7f65e2 \ + --hash=sha256:1e986b206dae4476f41bcec1faa057851f3889503a70e1bdb2378d406223994a \ + --hash=sha256:272d3892a1e1a2dbc39cc5cde96834c236d5327e2122d3aaa19f6614531bb6eb \ + --hash=sha256:278d2ed7c16cfc075c91378c4f47924c0625f5fc84b2d50d921b18b7975bd210 \ + --hash=sha256:2ad40e31093a4af319dadf503b2467ccdc8f67c72e4bcba97f8c10cb078207b5 \ + --hash=sha256:2e920d506ec85eb4ba50cd4228c2bec05642894d4c73c59b3a2fe20346bd00ee \ + --hash=sha256:3359fc442a743e870f4588fcf5dcbc1bf929df1fad8fb9905cd94e5edb02e84c \ + --hash=sha256:37933ec6e693e51a5b07505bd05de57eee12f3e8c32b07da7e73669398e6630a \ + --hash=sha256:398af5e3ba9cf768787eef45c803ff9614cc3e22a5b2f7d7ae116df8b11e3314 \ + --hash=sha256:3b747a674c20a67343cb61d43fdd9207ce5da6a99f629c6e2541aa0e89215bcd \ + --hash=sha256:461665ff58895f508e2866824a47bdee72497b091c730071f2b7575d5762ab65 \ + --hash=sha256:4c6fdd4fccbec90cc8a01fc00773fcd5fa28db683c116ee3cb35cd5da9ef6c37 \ + --hash=sha256:5829b792bf5822fd0a6f6eb34c5f81dd074f01d570ed7f36aa101d6fc7a0a6e4 \ + --hash=sha256:596d1f98fc70232fcb6590c439f43b350cb762fb5d61ce7b0e9db4539654cc13 \ + --hash=sha256:5ae44e10a8e3407dbe138984f21e536583f2bba1be9491239f942c2464ac0894 \ + --hash=sha256:635f5d4dd18758a1fbd1049a8e8d2fee4ffed124462d837d1a02a0e009c3ab31 \ + --hash=sha256:64e52e2b3970bd891309c113b54cf0e4384762c934d5ae56e283f9a0afcd953e \ + --hash=sha256:66741ef4ee08ea0b2cc3c86916ab66b6aef03768525627fd6a1b34968b4e3709 \ + --hash=sha256:67b741654b851abafb7bc625b6d1cdd520a379074e64b6a128e3b688c3c04740 \ + --hash=sha256:6ac08d24c1f16bd2bf5eca8eaf8304812f44af5cfe5062006ec676e7e1d50afc \ + --hash=sha256:6f998db4e71b645350b9ac28a2167e6632c239963ca9da411523bb439c5c514d \ + --hash=sha256:72218785ce41b9cfd2fc1d6a017dc1ff7acfc4c17d01053265c41a2c0cc39b8c \ + --hash=sha256:74dea7751d98034887dbd821b7aae3e1d36eda111d6ca36c206c44478035709c \ + --hash=sha256:759ce4851a4bb15ecabae28f4d2e18983c244eddd767f560165563bf9aefbc8d \ + --hash=sha256:77e2fd3057c9d78e225fa0a2160f96b64a824de17840351b26825b0848022906 \ + --hash=sha256:7c074fece789b5034b9b1404a1f8208fc2d4c6ce9decdd16e8220c5a793e6f61 \ + --hash=sha256:7c42c70cd1d362284289c6273adda4c6af8039a8ae12dc451dcd61cdabb8ab57 \ + --hash=sha256:7f57f14606cd1dd0f0de396e1e53824c371e9544a822648cd76c034d209b559c \ + --hash=sha256:83c681c526a3439b5cf94f7420471705bbf96262f49a6fe546a6db5f687a3d4a \ + --hash=sha256:8485b340a6a9e76c62a7dce3c98e5f102c9219f4cfbf896a00cf48caf078d438 \ + --hash=sha256:84e6e8cd997930fc66d5bb4fde61e2b62ba19d62b7abd7a69920406f9ecca946 \ + --hash=sha256:89284716bc6a5a415d4eaa11b1726d2d60a0cd12aadf5439828353662ede9dd7 \ + --hash=sha256:8b87e1a59c38f275c0e3676fc2ab6d59eccecfd460be267ac360cc31f7bcde96 \ + --hash=sha256:8f24ed114432de109aa9fd317278518a5af2d31ac2ea6b952b2f7782b43da091 \ + --hash=sha256:98cb4d057f285bd80d8778ebc4fde6b4d509ac3f331758fb1528b733215443ae \ + --hash=sha256:998679bf62b7fb599d2878aa3ed06b9ce688b8974893e7223c60db155f26bd8d \ + --hash=sha256:9ba053c5f50430a3fcfd36f75aff9caeba0440b2d076afdb79a318d6ca245f88 \ + --hash=sha256:9c99616c853bb585301df6de07ca2cadad344fd1ada6d62bb30aec05219c45d2 \ + --hash=sha256:a1fd716e7a01f8e717490fbe2e431d2905ab8aa598b9b12f8d10abebb36b04dd \ + --hash=sha256:a2355cba1f4ad8b6988a4ca3feed5bff33f6af2d7f134852cf279c2aebfde541 \ + --hash=sha256:b1f8133c9a275df5613a451e73f36c2aea4fe13c5c8997e22cf355ebd7bd0728 \ + --hash=sha256:b8667b48e7a7ef66afba2c81e1094ef526388d35b873966d8a9a447974ed9178 \ + --hash=sha256:ba1eb1843304b1e5537e1fca632fa894d6f6deca8d6389636ee5b4797affb968 \ + --hash=sha256:be82c3c8cfb15b30f36768797a640e800513793d6ae1724aaaafe5bf86f8f346 \ + --hash=sha256:c02ec1c5856179f171e032a31d6f8bf84e5a75c45c33b2e20a3de353b266ebd8 \ + --hash=sha256:c672d99a345849301784604bfeaeba4db0c7aae50b95be04dd651fd2a7310b93 \ + --hash=sha256:c6c777a480337ac14f38564ac88ae82d4cd238bf293f0a22295b66eb89ffced7 \ + --hash=sha256:cae0274952c079886567f3f4f685bcaf5708f0a23a5f5216fdab71f81a6c0273 \ + --hash=sha256:cd67cf24a553339d5062eff51013780a00d6f97a39ca062781d06b3a73b15462 \ + --hash=sha256:d3515f198eaa2f0ed49f8819d5732d70698c3fa37384146079b3799b97667a94 \ + --hash=sha256:d5280312b9af0976231f9e317c20e4a61cd2f9629b7bfea6a693d1878a264ebd \ + --hash=sha256:de06adc872bcd8c2a4e0dc51250e9e65ef2ca91be023b9d13ebd67c2ba552e1e \ + --hash=sha256:e1674e4307fa3024fc897ca774e9c7562c957af85df55efe2988ed9056dc4e57 \ + --hash=sha256:e2096eddb4e7c7bdae4bd69ad364e55e07b8316653234a56552d9c988bd2d61b \ + --hash=sha256:e560628513ed34759456a416bf86b54b2476c59144a9138165c9a1575801d0d9 \ + --hash=sha256:edfedb64740750e1a3b16152620220f51d58ff1b4abceb339ca92e934775c27a \ + --hash=sha256:f13cae8cc389a440def0c8c52057f37359014ccbc9dc1f0827936bcd367c6100 \ + --hash=sha256:f314013e7dcd5cf45ab1945d92e713eec788166262ae8deb2cfacd53def27325 \ + --hash=sha256:f583edb943cf2e09c60441b910d6a20b4d9d626c75a36c8fcac01a6c96c01183 \ + --hash=sha256:fd8536e902db7e365f49e7d9029283403974ccf29b13fc7028b97e2295b33556 \ + --hash=sha256:fe70e325aa68fa4b5edf7d1a4b6f691eb04bbccac0ace68e34820d283b5f80d4 # via google-resumable-media -google-resumable-media==2.3.3 \ - --hash=sha256:27c52620bd364d1c8116eaac4ea2afcbfb81ae9139fb3199652fcac1724bfb6c \ - --hash=sha256:5b52774ea7a829a8cdaa8bd2d4c3d4bc660c91b30857ab2668d0eb830f4ea8c5 +google-resumable-media==2.4.0 \ + --hash=sha256:2aa004c16d295c8f6c33b2b4788ba59d366677c0a25ae7382436cb30f776deaa \ + --hash=sha256:8d5518502f92b9ecc84ac46779bd4f09694ecb3ba38a3e7ca737a86d15cbca1f # via google-cloud-storage googleapis-common-protos==1.56.4 \ --hash=sha256:8eb2cbc91b69feaf23e32452a7ae60e791e09967d81d4fcc7fc388182d1bd394 \ --hash=sha256:c25873c47279387cfdcbdafa36149887901d36202cb645a0e4f29686bf6e4417 # via google-api-core -idna==3.3 \ - --hash=sha256:84d9dd047ffa80596e0f246e2eab0b391788b0503584e8945f2368256d2735ff \ - --hash=sha256:9d643ff0a55b762d5cdb124b8eaa99c66322e2157b69160bc32796e824360e6d +idna==3.4 \ + --hash=sha256:814f528e8dead7d329833b91c5faa87d60bf71824cd12a7530b5526063d02cb4 \ + --hash=sha256:90b77e79eaa3eba6de819a0c442c0b4ceefc341a7a2ab77d7562bf49f425c5c2 # via requests -importlib-metadata==4.12.0 \ - --hash=sha256:637245b8bab2b6502fcbc752cc4b7a6f6243bb02b31c5c26156ad103d3d45670 \ - --hash=sha256:7401a975809ea1fdc658c3aa4f78cc2195a0e019c5cbc4c06122884e9ae80c23 +importlib-metadata==5.0.0 \ + --hash=sha256:da31db32b304314d044d3c12c79bd59e307889b287ad12ff387b3500835fc2ab \ + --hash=sha256:ddb0e35065e8938f867ed4928d0ae5bf2a53b7773871bfe6bcc7e4fcdc7dea43 # via # -r requirements.in # twine -jaraco-classes==3.2.2 \ - --hash=sha256:6745f113b0b588239ceb49532aa09c3ebb947433ce311ef2f8e3ad64ebb74594 \ - --hash=sha256:e6ef6fd3fcf4579a7a019d87d1e56a883f4e4c35cfe925f86731abc58804e647 +jaraco-classes==3.2.3 \ + --hash=sha256:2353de3288bc6b82120752201c6b1c1a14b058267fa424ed5ce5984e3b922158 \ + --hash=sha256:89559fa5c1d3c34eff6f631ad80bb21f378dbcbb35dd161fd2c6b93f5be2f98a # via keyring jeepney==0.8.0 \ --hash=sha256:5efe48d255973902f6badc3ce55e2aa6c5c3b3bc642059ef3a91247bcfcc5806 \ @@ -255,9 +284,9 @@ jinja2==3.1.2 \ --hash=sha256:31351a702a408a9e7595a8fc6150fc3f43bb6bf7e319770cbc0db9df9437e852 \ --hash=sha256:6088930bfe239f0e6710546ab9c19c9ef35e29792895fed6e6e31a023a182a61 # via gcp-releasetool -keyring==23.9.0 \ - --hash=sha256:4c32a31174faaee48f43a7e2c7e9c3216ec5e95acf22a2bebfb4a1d05056ee44 \ - --hash=sha256:98f060ec95ada2ab910c195a2d4317be6ef87936a766b239c46aa3c7aac4f0db +keyring==23.9.3 \ + --hash=sha256:69732a15cb1433bdfbc3b980a8a36a04878a6cfd7cb99f497b573f31618001c0 \ + --hash=sha256:69b01dd83c42f590250fe7a1f503fc229b14de83857314b1933a3ddbf595c4a5 # via # gcp-releasetool # twine @@ -303,9 +332,9 @@ markupsafe==2.1.1 \ --hash=sha256:f121a1420d4e173a5d96e47e9a0c0dcff965afdf1626d28de1460815f7c4ee7a \ --hash=sha256:fc7b548b17d238737688817ab67deebb30e8073c95749d55538ed473130ec0c7 # via jinja2 -more-itertools==8.14.0 \ - --hash=sha256:1bc4f91ee5b1b31ac7ceacc17c09befe6a40a503907baf9c839c229b5095cfd2 \ - --hash=sha256:c09443cd3d5438b8dafccd867a6bc1cb0894389e90cb53d227456b0b0bccb750 +more-itertools==9.0.0 \ + --hash=sha256:250e83d7e81d0c87ca6bd942e6aeab8cc9daa6096d12c5308f3f92fa5e5c1f41 \ + --hash=sha256:5a6257e40878ef0520b1803990e3e22303a41b5714006c32a3fd8304b26ea1ab # via jaraco-classes nox==2022.8.7 \ --hash=sha256:1b894940551dc5c389f9271d197ca5d655d40bdc6ccf93ed6880e4042760a34b \ @@ -325,34 +354,34 @@ platformdirs==2.5.2 \ --hash=sha256:027d8e83a2d7de06bbac4e5ef7e023c02b863d7ea5d079477e722bb41ab25788 \ --hash=sha256:58c8abb07dcb441e6ee4b11d8df0ac856038f944ab98b7be6b27b2a3c7feef19 # via virtualenv -protobuf==3.20.2 \ - --hash=sha256:03d76b7bd42ac4a6e109742a4edf81ffe26ffd87c5993126d894fe48a120396a \ - --hash=sha256:09e25909c4297d71d97612f04f41cea8fa8510096864f2835ad2f3b3df5a5559 \ - --hash=sha256:18e34a10ae10d458b027d7638a599c964b030c1739ebd035a1dfc0e22baa3bfe \ - --hash=sha256:291fb4307094bf5ccc29f424b42268640e00d5240bf0d9b86bf3079f7576474d \ - --hash=sha256:2c0b040d0b5d5d207936ca2d02f00f765906622c07d3fa19c23a16a8ca71873f \ - --hash=sha256:384164994727f274cc34b8abd41a9e7e0562801361ee77437099ff6dfedd024b \ - --hash=sha256:3cb608e5a0eb61b8e00fe641d9f0282cd0eedb603be372f91f163cbfbca0ded0 \ - --hash=sha256:5d9402bf27d11e37801d1743eada54372f986a372ec9679673bfcc5c60441151 \ - --hash=sha256:712dca319eee507a1e7df3591e639a2b112a2f4a62d40fe7832a16fd19151750 \ - --hash=sha256:7a5037af4e76c975b88c3becdf53922b5ffa3f2cddf657574a4920a3b33b80f3 \ - --hash=sha256:8228e56a865c27163d5d1d1771d94b98194aa6917bcfb6ce139cbfa8e3c27334 \ - --hash=sha256:84a1544252a933ef07bb0b5ef13afe7c36232a774affa673fc3636f7cee1db6c \ - --hash=sha256:84fe5953b18a383fd4495d375fe16e1e55e0a3afe7b4f7b4d01a3a0649fcda9d \ - --hash=sha256:9c673c8bfdf52f903081816b9e0e612186684f4eb4c17eeb729133022d6032e3 \ - --hash=sha256:9f876a69ca55aed879b43c295a328970306e8e80a263ec91cf6e9189243c613b \ - --hash=sha256:a9e5ae5a8e8985c67e8944c23035a0dff2c26b0f5070b2f55b217a1c33bbe8b1 \ - --hash=sha256:b4fdb29c5a7406e3f7ef176b2a7079baa68b5b854f364c21abe327bbeec01cdb \ - --hash=sha256:c184485e0dfba4dfd451c3bd348c2e685d6523543a0f91b9fd4ae90eb09e8422 \ - --hash=sha256:c9cdf251c582c16fd6a9f5e95836c90828d51b0069ad22f463761d27c6c19019 \ - --hash=sha256:e39cf61bb8582bda88cdfebc0db163b774e7e03364bbf9ce1ead13863e81e359 \ - --hash=sha256:e8fbc522303e09036c752a0afcc5c0603e917222d8bedc02813fd73b4b4ed804 \ - --hash=sha256:f34464ab1207114e73bba0794d1257c150a2b89b7a9faf504e00af7c9fd58978 \ - --hash=sha256:f52dabc96ca99ebd2169dadbe018824ebda08a795c7684a0b7d203a290f3adb0 +protobuf==3.20.3 \ + --hash=sha256:03038ac1cfbc41aa21f6afcbcd357281d7521b4157926f30ebecc8d4ea59dcb7 \ + --hash=sha256:28545383d61f55b57cf4df63eebd9827754fd2dc25f80c5253f9184235db242c \ + --hash=sha256:2e3427429c9cffebf259491be0af70189607f365c2f41c7c3764af6f337105f2 \ + --hash=sha256:398a9e0c3eaceb34ec1aee71894ca3299605fa8e761544934378bbc6c97de23b \ + --hash=sha256:44246bab5dd4b7fbd3c0c80b6f16686808fab0e4aca819ade6e8d294a29c7050 \ + --hash=sha256:447d43819997825d4e71bf5769d869b968ce96848b6479397e29fc24c4a5dfe9 \ + --hash=sha256:67a3598f0a2dcbc58d02dd1928544e7d88f764b47d4a286202913f0b2801c2e7 \ + --hash=sha256:74480f79a023f90dc6e18febbf7b8bac7508420f2006fabd512013c0c238f454 \ + --hash=sha256:819559cafa1a373b7096a482b504ae8a857c89593cf3a25af743ac9ecbd23480 \ + --hash=sha256:899dc660cd599d7352d6f10d83c95df430a38b410c1b66b407a6b29265d66469 \ + --hash=sha256:8c0c984a1b8fef4086329ff8dd19ac77576b384079247c770f29cc8ce3afa06c \ + --hash=sha256:9aae4406ea63d825636cc11ffb34ad3379335803216ee3a856787bcf5ccc751e \ + --hash=sha256:a7ca6d488aa8ff7f329d4c545b2dbad8ac31464f1d8b1c87ad1346717731e4db \ + --hash=sha256:b6cc7ba72a8850621bfec987cb72623e703b7fe2b9127a161ce61e61558ad905 \ + --hash=sha256:bf01b5720be110540be4286e791db73f84a2b721072a3711efff6c324cdf074b \ + --hash=sha256:c02ce36ec760252242a33967d51c289fd0e1c0e6e5cc9397e2279177716add86 \ + --hash=sha256:d9e4432ff660d67d775c66ac42a67cf2453c27cb4d738fc22cb53b5d84c135d4 \ + --hash=sha256:daa564862dd0d39c00f8086f88700fdbe8bc717e993a21e90711acfed02f2402 \ + --hash=sha256:de78575669dddf6099a8a0f46a27e82a1783c557ccc38ee620ed8cc96d3be7d7 \ + --hash=sha256:e64857f395505ebf3d2569935506ae0dfc4a15cb80dc25261176c784662cdcc4 \ + --hash=sha256:f4bd856d702e5b0d96a00ec6b307b0f51c1982c2bf9c0052cf9019e9a544ba99 \ + --hash=sha256:f4c42102bc82a51108e449cbb32b19b180022941c727bac0cfd50170341f16ee # via # gcp-docuploader # gcp-releasetool # google-api-core + # googleapis-common-protos py==1.11.0 \ --hash=sha256:51c75c4126074b472f746a24399ad32f6053d1b34b68d2fa41e558e6f4a98719 \ --hash=sha256:607c53218732647dff4acdfcd50cb62615cedf612e72d1724fb1a0cc6405b378 @@ -377,9 +406,9 @@ pygments==2.13.0 \ # via # readme-renderer # rich -pyjwt==2.4.0 \ - --hash=sha256:72d1d253f32dbd4f5c88eaf1fdc62f3a19f676ccbadb9dbc5d07e951b2b26daf \ - --hash=sha256:d42908208c699b3b973cbeb01a969ba6a96c821eefb1c5bfe4c390c01d67abba +pyjwt==2.6.0 \ + --hash=sha256:69285c7e31fc44f68a1feb309e948e0df53259d579295e6cfe2b1792329f05fd \ + --hash=sha256:d83c3d892a77bbb74d3e1a2cfa90afaadb60945205d1095d9221f04466f64c14 # via gcp-releasetool pyparsing==3.0.9 \ --hash=sha256:2b020ecf7d21b687f219b71ecad3631f644a47f01403fa1d1036b0c6416d70fb \ @@ -392,9 +421,9 @@ python-dateutil==2.8.2 \ --hash=sha256:0123cacc1627ae19ddf3c27a5de5bd67ee4586fbdd6440d9748f8abb483d3e86 \ --hash=sha256:961d03dc3453ebbc59dbdea9e4e11c5651520a876d0f4db161e8674aae935da9 # via gcp-releasetool -readme-renderer==37.0 \ - --hash=sha256:07b7ea234e03e58f77cc222e206e6abb8f4c0435becce5104794ee591f9301c5 \ - --hash=sha256:9fa416704703e509eeb900696751c908ddeb2011319d93700d8f18baff887a69 +readme-renderer==37.3 \ + --hash=sha256:cd653186dfc73055656f090f227f5cb22a046d7f71a841dfa305f55c9a513273 \ + --hash=sha256:f67a16caedfa71eef48a31b39708637a6f4664c4394801a7b0d6432d13907343 # via twine requests==2.28.1 \ --hash=sha256:7c5599b102feddaa661c826c56ab4fee28bfd17f5abca1ebbe3e7f19d7c97983 \ @@ -405,17 +434,17 @@ requests==2.28.1 \ # google-cloud-storage # requests-toolbelt # twine -requests-toolbelt==0.9.1 \ - --hash=sha256:380606e1d10dc85c3bd47bf5a6095f815ec007be7a8b69c878507068df059e6f \ - --hash=sha256:968089d4584ad4ad7c171454f0a5c6dac23971e9472521ea3b6d49d610aa6fc0 +requests-toolbelt==0.10.1 \ + --hash=sha256:18565aa58116d9951ac39baa288d3adb5b3ff975c4f25eee78555d89e8f247f7 \ + --hash=sha256:62e09f7ff5ccbda92772a29f394a49c3ad6cb181d568b1337626b2abb628a63d # via twine rfc3986==2.0.0 \ --hash=sha256:50b1502b60e289cb37883f3dfd34532b8873c7de9f49bb546641ce9cbd256ebd \ --hash=sha256:97aacf9dbd4bfd829baad6e6309fa6573aaf1be3f6fa735c8ab05e46cecb261c # via twine -rich==12.5.1 \ - --hash=sha256:2eb4e6894cde1e017976d2975ac210ef515d7548bc595ba20e195fb9628acdeb \ - --hash=sha256:63a5c5ce3673d3d5fbbf23cd87e11ab84b6b451436f1b7f19ec54b6bc36ed7ca +rich==12.6.0 \ + --hash=sha256:a4eb26484f2c82589bd9a17c73d32a010b1e29d89f1604cd9bf3a2097b81bb5e \ + --hash=sha256:ba3a3775974105c221d31141f2c116f4fd65c5ceb0698657a11e9f295ec93fd0 # via twine rsa==4.9 \ --hash=sha256:90260d9058e514786967344d0ef75fa8727eed8a7d2e43ce9f4bcf1b536174f7 \ @@ -437,9 +466,9 @@ twine==4.0.1 \ --hash=sha256:42026c18e394eac3e06693ee52010baa5313e4811d5a11050e7d48436cf41b9e \ --hash=sha256:96b1cf12f7ae611a4a40b6ae8e9570215daff0611828f5fe1f37a16255ab24a0 # via -r requirements.in -typing-extensions==4.3.0 \ - --hash=sha256:25642c956049920a5aa49edcdd6ab1e06d7e5d467fc00e0506c44ac86fbfca02 \ - --hash=sha256:e6d2677a32f47fc7eb2795db1dd15c1f34eff616bcaf2cfb5e997f854fa1c4a6 +typing-extensions==4.4.0 \ + --hash=sha256:1511434bb92bf8dd198c12b1cc812e800d4181cfcb867674e0f8279cc93087aa \ + --hash=sha256:16fa4864408f655d35ec496218b85f79b3437c829e93320c7c9215ccfd92489e # via -r requirements.in urllib3==1.26.12 \ --hash=sha256:3fa96cf423e6987997fc326ae8df396db2a8b7c667747d47ddd8ecba91f4a74e \ @@ -447,9 +476,9 @@ urllib3==1.26.12 \ # via # requests # twine -virtualenv==20.16.4 \ - --hash=sha256:014f766e4134d0008dcaa1f95bafa0fb0f575795d07cae50b1bee514185d6782 \ - --hash=sha256:035ed57acce4ac35c82c9d8802202b0e71adac011a511ff650cbcf9635006a22 +virtualenv==20.16.6 \ + --hash=sha256:186ca84254abcbde98180fd17092f9628c5fe742273c02724972a1d8a2035108 \ + --hash=sha256:530b850b523c6449406dfba859d6345e48ef19b8439606c5d74d7d3c9e14d76e # via nox webencodings==0.5.1 \ --hash=sha256:a0af1213f3c2226497a97e2b3aa01a7e4bee4f403f95be16fc9acd2947514a78 \ @@ -459,13 +488,13 @@ wheel==0.37.1 \ --hash=sha256:4bdcd7d840138086126cd09254dc6195fb4fc6f01c050a1d7236f2630db1d22a \ --hash=sha256:e9a504e793efbca1b8e0e9cb979a249cf4a0a7b5b8c9e8b65a5e39d49529c1c4 # via -r requirements.in -zipp==3.8.1 \ - --hash=sha256:05b45f1ee8f807d0cc928485ca40a07cb491cf092ff587c0df9cb1fd154848d2 \ - --hash=sha256:47c40d7fe183a6f21403a199b3e4192cca5774656965b0a4988ad2f8feb5f009 +zipp==3.10.0 \ + --hash=sha256:4fcb6f278987a6605757302a6e40e896257570d11c51628968ccb2a47e80c6c1 \ + --hash=sha256:7a7262fd930bd3e36c50b9a64897aec3fafff3dfdeec9623ae22b40e93f99bb8 # via importlib-metadata # The following packages are considered to be unsafe in a requirements file: -setuptools==65.2.0 \ - --hash=sha256:7f4bc85450898a09f76ebf28b72fa25bc7111f6c7d665d514a60bba9c75ef2a9 \ - --hash=sha256:a3ca5857c89f82f5c9410e8508cb32f4872a3bafd4aa7ae122a24ca33bccc750 +setuptools==65.5.0 \ + --hash=sha256:512e5536220e38146176efb833d4a62aa726b7bbff82cfbc8ba9eaa3996e0b17 \ + --hash=sha256:f62ea9da9ed6289bfe868cd6845968a2c854d1427f8548d52cae02a42b4f0356 # via -r requirements.in diff --git a/noxfile.py b/noxfile.py index 51cd1cb..240d3ec 100644 --- a/noxfile.py +++ b/noxfile.py @@ -278,7 +278,11 @@ def docs(session): """Build the docs for this library.""" session.install("-e", ".") - session.install("sphinx==4.0.1", "alabaster", "recommonmark") + session.install( + "sphinx==4.0.1", + "alabaster", + "recommonmark", + ) shutil.rmtree(os.path.join("docs", "_build"), ignore_errors=True) session.run( @@ -301,7 +305,10 @@ def docfx(session): session.install("-e", ".") session.install( - "sphinx==4.0.1", "alabaster", "recommonmark", "gcp-sphinx-docfx-yaml" + "sphinx==4.0.1", + "alabaster", + "recommonmark", + "gcp-sphinx-docfx-yaml", ) shutil.rmtree(os.path.join("docs", "_build"), ignore_errors=True) From 2bc8fe68e9cbb0968fad92e7f6ce98c028bccb7e Mon Sep 17 00:00:00 2001 From: "gcf-owl-bot[bot]" <78513119+gcf-owl-bot[bot]@users.noreply.github.com> Date: Wed, 16 Nov 2022 17:10:44 +0000 Subject: [PATCH 05/10] chore(python): update release script dependencies [autoapprove] (#110) Source-Link: https://togithub.com/googleapis/synthtool/commit/25083af347468dd5f90f69627420f7d452b6c50e Post-Processor: gcr.io/cloud-devrel-public-resources/owlbot-python:latest@sha256:e6cbd61f1838d9ff6a31436dfc13717f372a7482a82fc1863ca954ec47bff8c8 --- .github/.OwlBot.lock.yaml | 2 +- .github/workflows/docs.yml | 4 +-- .github/workflows/lint.yml | 2 +- .github/workflows/unittest.yml | 2 +- .kokoro/docker/docs/Dockerfile | 12 +++---- .kokoro/requirements.in | 4 ++- .kokoro/requirements.txt | 61 ++++++++++++++++++---------------- noxfile.py | 4 +-- 8 files changed, 48 insertions(+), 43 deletions(-) diff --git a/.github/.OwlBot.lock.yaml b/.github/.OwlBot.lock.yaml index 12edee7..3f1ccc0 100644 --- a/.github/.OwlBot.lock.yaml +++ b/.github/.OwlBot.lock.yaml @@ -13,4 +13,4 @@ # limitations under the License. docker: image: gcr.io/cloud-devrel-public-resources/owlbot-python:latest - digest: sha256:452901c74a22f9b9a3bd02bce780b8e8805c97270d424684bff809ce5be8c2a2 + digest: sha256:e6cbd61f1838d9ff6a31436dfc13717f372a7482a82fc1863ca954ec47bff8c8 diff --git a/.github/workflows/docs.yml b/.github/workflows/docs.yml index 7092a13..e97d89e 100644 --- a/.github/workflows/docs.yml +++ b/.github/workflows/docs.yml @@ -12,7 +12,7 @@ jobs: - name: Setup Python uses: actions/setup-python@v4 with: - python-version: "3.10" + python-version: "3.9" - name: Install nox run: | python -m pip install --upgrade setuptools pip wheel @@ -28,7 +28,7 @@ jobs: - name: Setup Python uses: actions/setup-python@v4 with: - python-version: "3.10" + python-version: "3.9" - name: Install nox run: | python -m pip install --upgrade setuptools pip wheel diff --git a/.github/workflows/lint.yml b/.github/workflows/lint.yml index d2aee5b..16d5a9e 100644 --- a/.github/workflows/lint.yml +++ b/.github/workflows/lint.yml @@ -12,7 +12,7 @@ jobs: - name: Setup Python uses: actions/setup-python@v4 with: - python-version: "3.10" + python-version: "3.8" - name: Install nox run: | python -m pip install --upgrade setuptools pip wheel diff --git a/.github/workflows/unittest.yml b/.github/workflows/unittest.yml index 24e4655..c0dca03 100644 --- a/.github/workflows/unittest.yml +++ b/.github/workflows/unittest.yml @@ -41,7 +41,7 @@ jobs: - name: Setup Python uses: actions/setup-python@v4 with: - python-version: "3.10" + python-version: "3.8" - name: Install coverage run: | python -m pip install --upgrade setuptools pip wheel diff --git a/.kokoro/docker/docs/Dockerfile b/.kokoro/docker/docs/Dockerfile index 238b87b..f8137d0 100644 --- a/.kokoro/docker/docs/Dockerfile +++ b/.kokoro/docker/docs/Dockerfile @@ -60,16 +60,16 @@ RUN apt-get update \ && rm -rf /var/lib/apt/lists/* \ && rm -f /var/cache/apt/archives/*.deb -###################### Install python 3.8.11 +###################### Install python 3.9.13 -# Download python 3.8.11 -RUN wget https://www.python.org/ftp/python/3.8.11/Python-3.8.11.tgz +# Download python 3.9.13 +RUN wget https://www.python.org/ftp/python/3.9.13/Python-3.9.13.tgz # Extract files -RUN tar -xvf Python-3.8.11.tgz +RUN tar -xvf Python-3.9.13.tgz -# Install python 3.8.11 -RUN ./Python-3.8.11/configure --enable-optimizations +# Install python 3.9.13 +RUN ./Python-3.9.13/configure --enable-optimizations RUN make altinstall ###################### Install pip diff --git a/.kokoro/requirements.in b/.kokoro/requirements.in index 7718391..cbd7e77 100644 --- a/.kokoro/requirements.in +++ b/.kokoro/requirements.in @@ -5,4 +5,6 @@ typing-extensions twine wheel setuptools -nox \ No newline at end of file +nox +charset-normalizer<3 +click<8.1.0 diff --git a/.kokoro/requirements.txt b/.kokoro/requirements.txt index 31425f1..9c1b9be 100644 --- a/.kokoro/requirements.txt +++ b/.kokoro/requirements.txt @@ -93,11 +93,14 @@ cffi==1.15.1 \ charset-normalizer==2.1.1 \ --hash=sha256:5a3d016c7c547f69d6f81fb0db9449ce888b418b5b9952cc5e6e66843e9dd845 \ --hash=sha256:83e9a75d1911279afd89352c68b45348559d1fc0506b054b346651b5e7fee29f - # via requests + # via + # -r requirements.in + # requests click==8.0.4 \ --hash=sha256:6a7a62563bbfabfda3a38f3023a1db4a35978c0abd76f6c9605ecd6554d6d9b1 \ --hash=sha256:8458d7b1287c5fb128c90e23381cf99dcde74beaf6c7ff6384ce84d6fe090adb # via + # -r requirements.in # gcp-docuploader # gcp-releasetool colorlog==6.7.0 \ @@ -156,9 +159,9 @@ gcp-docuploader==0.6.4 \ --hash=sha256:01486419e24633af78fd0167db74a2763974765ee8078ca6eb6964d0ebd388af \ --hash=sha256:70861190c123d907b3b067da896265ead2eeb9263969d6955c9e0bb091b5ccbf # via -r requirements.in -gcp-releasetool==1.9.1 \ - --hash=sha256:952f4055d5d986b070ae2a71c4410b250000f9cc5a1e26398fcd55a5bbc5a15f \ - --hash=sha256:d0d3c814a97c1a237517e837d8cfa668ced8df4b882452578ecef4a4e79c583b +gcp-releasetool==1.10.0 \ + --hash=sha256:72a38ca91b59c24f7e699e9227c90cbe4dd71b789383cb0164b088abae294c83 \ + --hash=sha256:8c7c99320208383d4bb2b808c6880eb7a81424afe7cdba3c8d84b25f4f0e097d # via -r requirements.in google-api-core==2.10.2 \ --hash=sha256:10c06f7739fe57781f87523375e8e1a3a4674bf6392cd6131a3222182b971320 \ @@ -166,9 +169,9 @@ google-api-core==2.10.2 \ # via # google-cloud-core # google-cloud-storage -google-auth==2.14.0 \ - --hash=sha256:1ad5b0e6eba5f69645971abb3d2c197537d5914070a8c6d30299dfdb07c5c700 \ - --hash=sha256:cf24817855d874ede2efd071aa22125445f555de1685b739a9782fcf408c2a3d +google-auth==2.14.1 \ + --hash=sha256:ccaa901f31ad5cbb562615eb8b664b3dd0bf5404a67618e642307f00613eda4d \ + --hash=sha256:f5d8701633bebc12e0deea4df8abd8aff31c28b355360597f7f2ee60f2e4d016 # via # gcp-releasetool # google-api-core @@ -178,9 +181,9 @@ google-cloud-core==2.3.2 \ --hash=sha256:8417acf6466be2fa85123441696c4badda48db314c607cf1e5d543fa8bdc22fe \ --hash=sha256:b9529ee7047fd8d4bf4a2182de619154240df17fbe60ead399078c1ae152af9a # via google-cloud-storage -google-cloud-storage==2.5.0 \ - --hash=sha256:19a26c66c317ce542cea0830b7e787e8dac2588b6bfa4d3fd3b871ba16305ab0 \ - --hash=sha256:382f34b91de2212e3c2e7b40ec079d27ee2e3dbbae99b75b1bcd8c63063ce235 +google-cloud-storage==2.6.0 \ + --hash=sha256:104ca28ae61243b637f2f01455cc8a05e8f15a2a18ced96cb587241cdd3820f5 \ + --hash=sha256:4ad0415ff61abdd8bb2ae81c1f8f7ec7d91a1011613f2db87c614c550f97bfe9 # via gcp-docuploader google-crc32c==1.5.0 \ --hash=sha256:024894d9d3cfbc5943f8f230e23950cd4906b2fe004c72e29b209420a1e6b05a \ @@ -256,9 +259,9 @@ google-resumable-media==2.4.0 \ --hash=sha256:2aa004c16d295c8f6c33b2b4788ba59d366677c0a25ae7382436cb30f776deaa \ --hash=sha256:8d5518502f92b9ecc84ac46779bd4f09694ecb3ba38a3e7ca737a86d15cbca1f # via google-cloud-storage -googleapis-common-protos==1.56.4 \ - --hash=sha256:8eb2cbc91b69feaf23e32452a7ae60e791e09967d81d4fcc7fc388182d1bd394 \ - --hash=sha256:c25873c47279387cfdcbdafa36149887901d36202cb645a0e4f29686bf6e4417 +googleapis-common-protos==1.57.0 \ + --hash=sha256:27a849d6205838fb6cc3c1c21cb9800707a661bb21c6ce7fb13e99eb1f8a0c46 \ + --hash=sha256:a9f4a1d7f6d9809657b7f1316a1aa527f6664891531bcfcc13b6696e685f443c # via google-api-core idna==3.4 \ --hash=sha256:814f528e8dead7d329833b91c5faa87d60bf71824cd12a7530b5526063d02cb4 \ @@ -269,6 +272,7 @@ importlib-metadata==5.0.0 \ --hash=sha256:ddb0e35065e8938f867ed4928d0ae5bf2a53b7773871bfe6bcc7e4fcdc7dea43 # via # -r requirements.in + # keyring # twine jaraco-classes==3.2.3 \ --hash=sha256:2353de3288bc6b82120752201c6b1c1a14b058267fa424ed5ce5984e3b922158 \ @@ -284,9 +288,9 @@ jinja2==3.1.2 \ --hash=sha256:31351a702a408a9e7595a8fc6150fc3f43bb6bf7e319770cbc0db9df9437e852 \ --hash=sha256:6088930bfe239f0e6710546ab9c19c9ef35e29792895fed6e6e31a023a182a61 # via gcp-releasetool -keyring==23.9.3 \ - --hash=sha256:69732a15cb1433bdfbc3b980a8a36a04878a6cfd7cb99f497b573f31618001c0 \ - --hash=sha256:69b01dd83c42f590250fe7a1f503fc229b14de83857314b1933a3ddbf595c4a5 +keyring==23.11.0 \ + --hash=sha256:3dd30011d555f1345dec2c262f0153f2f0ca6bca041fb1dc4588349bb4c0ac1e \ + --hash=sha256:ad192263e2cdd5f12875dedc2da13534359a7e760e77f8d04b50968a821c2361 # via # gcp-releasetool # twine @@ -350,9 +354,9 @@ pkginfo==1.8.3 \ --hash=sha256:848865108ec99d4901b2f7e84058b6e7660aae8ae10164e015a6dcf5b242a594 \ --hash=sha256:a84da4318dd86f870a9447a8c98340aa06216bfc6f2b7bdc4b8766984ae1867c # via twine -platformdirs==2.5.2 \ - --hash=sha256:027d8e83a2d7de06bbac4e5ef7e023c02b863d7ea5d079477e722bb41ab25788 \ - --hash=sha256:58c8abb07dcb441e6ee4b11d8df0ac856038f944ab98b7be6b27b2a3c7feef19 +platformdirs==2.5.4 \ + --hash=sha256:1006647646d80f16130f052404c6b901e80ee4ed6bef6792e1f238a8969106f7 \ + --hash=sha256:af0276409f9a02373d540bf8480021a048711d572745aef4b7842dad245eba10 # via virtualenv protobuf==3.20.3 \ --hash=sha256:03038ac1cfbc41aa21f6afcbcd357281d7521b4157926f30ebecc8d4ea59dcb7 \ @@ -381,7 +385,6 @@ protobuf==3.20.3 \ # gcp-docuploader # gcp-releasetool # google-api-core - # googleapis-common-protos py==1.11.0 \ --hash=sha256:51c75c4126074b472f746a24399ad32f6053d1b34b68d2fa41e558e6f4a98719 \ --hash=sha256:607c53218732647dff4acdfcd50cb62615cedf612e72d1724fb1a0cc6405b378 @@ -476,17 +479,17 @@ urllib3==1.26.12 \ # via # requests # twine -virtualenv==20.16.6 \ - --hash=sha256:186ca84254abcbde98180fd17092f9628c5fe742273c02724972a1d8a2035108 \ - --hash=sha256:530b850b523c6449406dfba859d6345e48ef19b8439606c5d74d7d3c9e14d76e +virtualenv==20.16.7 \ + --hash=sha256:8691e3ff9387f743e00f6bb20f70121f5e4f596cae754531f2b3b3a1b1ac696e \ + --hash=sha256:efd66b00386fdb7dbe4822d172303f40cd05e50e01740b19ea42425cbe653e29 # via nox webencodings==0.5.1 \ --hash=sha256:a0af1213f3c2226497a97e2b3aa01a7e4bee4f403f95be16fc9acd2947514a78 \ --hash=sha256:b36a1c245f2d304965eb4e0a82848379241dc04b865afcc4aab16748587e1923 # via bleach -wheel==0.37.1 \ - --hash=sha256:4bdcd7d840138086126cd09254dc6195fb4fc6f01c050a1d7236f2630db1d22a \ - --hash=sha256:e9a504e793efbca1b8e0e9cb979a249cf4a0a7b5b8c9e8b65a5e39d49529c1c4 +wheel==0.38.4 \ + --hash=sha256:965f5259b566725405b05e7cf774052044b1ed30119b5d586b2703aafe8719ac \ + --hash=sha256:b60533f3f5d530e971d6737ca6d58681ee434818fab630c83a734bb10c083ce8 # via -r requirements.in zipp==3.10.0 \ --hash=sha256:4fcb6f278987a6605757302a6e40e896257570d11c51628968ccb2a47e80c6c1 \ @@ -494,7 +497,7 @@ zipp==3.10.0 \ # via importlib-metadata # The following packages are considered to be unsafe in a requirements file: -setuptools==65.5.0 \ - --hash=sha256:512e5536220e38146176efb833d4a62aa726b7bbff82cfbc8ba9eaa3996e0b17 \ - --hash=sha256:f62ea9da9ed6289bfe868cd6845968a2c854d1427f8548d52cae02a42b4f0356 +setuptools==65.5.1 \ + --hash=sha256:d0b9a8433464d5800cbe05094acf5c6d52a91bfac9b52bcfc4d41382be5d5d31 \ + --hash=sha256:e197a19aa8ec9722928f2206f8de752def0e4c9fc6953527360d1c36d94ddb2f # via -r requirements.in diff --git a/noxfile.py b/noxfile.py index 240d3ec..30133c9 100644 --- a/noxfile.py +++ b/noxfile.py @@ -273,7 +273,7 @@ def cover(session): session.run("coverage", "erase") -@nox.session(python=DEFAULT_PYTHON_VERSION) +@nox.session(python="3.9") def docs(session): """Build the docs for this library.""" @@ -299,7 +299,7 @@ def docs(session): ) -@nox.session(python=DEFAULT_PYTHON_VERSION) +@nox.session(python="3.9") def docfx(session): """Build the docfx yaml files for this library.""" From 18e2a32c425f6a7ca0684392a796e18547ea408a Mon Sep 17 00:00:00 2001 From: "gcf-owl-bot[bot]" <78513119+gcf-owl-bot[bot]@users.noreply.github.com> Date: Tue, 22 Nov 2022 15:09:56 -0500 Subject: [PATCH 06/10] chore: Update gapic-generator-python to v1.6.1 (#107) MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit * chore: update to gapic-generator-python 1.5.0 feat: add support for `google.cloud..__version__` PiperOrigin-RevId: 484665853 Source-Link: https://github.com/googleapis/googleapis/commit/8eb249a19db926c2fbc4ecf1dc09c0e521a88b22 Source-Link: https://github.com/googleapis/googleapis-gen/commit/c8aa327b5f478865fc3fd91e3c2768e54e26ad44 Copy-Tag: eyJwIjoiLmdpdGh1Yi8uT3dsQm90LnlhbWwiLCJoIjoiYzhhYTMyN2I1ZjQ3ODg2NWZjM2ZkOTFlM2MyNzY4ZTU0ZTI2YWQ0NCJ9 * 🦉 Updates from OwlBot post-processor See https://github.com/googleapis/repo-automation-bots/blob/main/packages/owl-bot/README.md * update version in gapic_version.py * add .release-please-manifest.json with correct version * set manifest to true in .github/release-please.yml * add release-please-config.json * 🦉 Updates from OwlBot post-processor See https://github.com/googleapis/repo-automation-bots/blob/main/packages/owl-bot/README.md * chore: Update to gapic-generator-python 1.6.0 feat(python): Add typing to proto.Message based class attributes feat(python): Snippetgen handling of repeated enum field PiperOrigin-RevId: 487326846 Source-Link: https://github.com/googleapis/googleapis/commit/da380c77bb87ba0f752baf07605dd1db30e1f7e1 Source-Link: https://github.com/googleapis/googleapis-gen/commit/61ef5762ee6731a0cbbfea22fd0eecee51ab1c8e Copy-Tag: eyJwIjoiLmdpdGh1Yi8uT3dsQm90LnlhbWwiLCJoIjoiNjFlZjU3NjJlZTY3MzFhMGNiYmZlYTIyZmQwZWVjZWU1MWFiMWM4ZSJ9 * 🦉 Updates from OwlBot post-processor See https://github.com/googleapis/repo-automation-bots/blob/main/packages/owl-bot/README.md * feat: new APIs added to reflect updates to the filestore service - Add ENTERPRISE Tier - Add snapshot APIs: RevertInstance, ListSnapshots, CreateSnapshot, DeleteSnapshot, UpdateSnapshot - Add multi-share APIs: ListShares, GetShare, CreateShare, DeleteShare, UpdateShare - Add ConnectMode to NetworkConfig (for Private Service Access support) - New status codes (SUSPENDED/SUSPENDING, REVERTING/RESUMING) - Add SuspensionReason (for KMS related suspension) - Add new fields to Instance information: max_capacity_gb, capacity_step_size_gb, max_share_count, capacity_gb, multi_share_enabled PiperOrigin-RevId: 487492758 Source-Link: https://github.com/googleapis/googleapis/commit/5be5981f50322cf0c7388595e0f31ac5d0693469 Source-Link: https://github.com/googleapis/googleapis-gen/commit/ab0e217f560cc2c1afc11441c2eab6b6950efd2b Copy-Tag: eyJwIjoiLmdpdGh1Yi8uT3dsQm90LnlhbWwiLCJoIjoiYWIwZTIxN2Y1NjBjYzJjMWFmYzExNDQxYzJlYWI2YjY5NTBlZmQyYiJ9 * 🦉 Updates from OwlBot post-processor See https://github.com/googleapis/repo-automation-bots/blob/main/packages/owl-bot/README.md * update path to snippet metadata json * chore: Update gapic-generator-python to v1.6.1 PiperOrigin-RevId: 488036204 Source-Link: https://github.com/googleapis/googleapis/commit/08f275f5c1c0d99056e1cb68376323414459ee19 Source-Link: https://github.com/googleapis/googleapis-gen/commit/555c0945e60649e38739ae64bc45719cdf72178f Copy-Tag: eyJwIjoiLmdpdGh1Yi8uT3dsQm90LnlhbWwiLCJoIjoiNTU1YzA5NDVlNjA2NDllMzg3MzlhZTY0YmM0NTcxOWNkZjcyMTc4ZiJ9 * 🦉 Updates from OwlBot post-processor See https://github.com/googleapis/repo-automation-bots/blob/main/packages/owl-bot/README.md * use templated owlbot.py and setup.py Co-authored-by: Owl Bot Co-authored-by: Anthonios Partheniou --- .github/release-please.yml | 1 + .github/workflows/unittest.yml | 2 +- .release-please-manifest.json | 3 + docs/dataplex_v1/types.rst | 1 - google/cloud/dataplex/__init__.py | 4 + google/cloud/dataplex/gapic_version.py | 16 + google/cloud/dataplex_v1/__init__.py | 4 + .../services/content_service/async_client.py | 104 +++--- .../services/content_service/client.py | 94 ++--- .../content_service/transports/base.py | 2 +- .../content_service/transports/grpc.py | 20 +- .../transports/grpc_asyncio.py | 16 +- .../services/dataplex_service/async_client.py | 326 +++++++++--------- .../services/dataplex_service/client.py | 268 +++++++------- .../dataplex_service/transports/base.py | 2 +- .../dataplex_service/transports/grpc.py | 20 +- .../transports/grpc_asyncio.py | 16 +- .../services/metadata_service/async_client.py | 114 +++--- .../services/metadata_service/client.py | 102 +++--- .../metadata_service/transports/base.py | 2 +- .../metadata_service/transports/grpc.py | 20 +- .../transports/grpc_asyncio.py | 16 +- google/cloud/dataplex_v1/types/analyze.py | 96 +++--- google/cloud/dataplex_v1/types/content.py | 34 +- google/cloud/dataplex_v1/types/logs.py | 88 ++--- google/cloud/dataplex_v1/types/metadata_.py | 160 ++++----- google/cloud/dataplex_v1/types/resources.py | 258 +++++++------- google/cloud/dataplex_v1/types/service.py | 248 ++++++------- google/cloud/dataplex_v1/types/tasks.py | 138 ++++---- noxfile.py | 2 +- owlbot.py | 22 +- release-please-config.json | 23 ++ ...et_metadata_google.cloud.dataplex.v1.json} | 3 +- setup.py | 35 +- testing/constraints-3.10.txt | 7 + testing/constraints-3.11.txt | 7 + testing/constraints-3.7.txt | 2 +- testing/constraints-3.8.txt | 7 + testing/constraints-3.9.txt | 7 + .../dataplex_v1/test_dataplex_service.py | 1 + 40 files changed, 1236 insertions(+), 1055 deletions(-) create mode 100644 .release-please-manifest.json create mode 100644 google/cloud/dataplex/gapic_version.py create mode 100644 release-please-config.json rename samples/generated_samples/{snippet_metadata_dataplex_v1.json => snippet_metadata_google.cloud.dataplex.v1.json} (99%) diff --git a/.github/release-please.yml b/.github/release-please.yml index 6def37a..e9a4f00 100644 --- a/.github/release-please.yml +++ b/.github/release-please.yml @@ -1,5 +1,6 @@ releaseType: python handleGHRelease: true +manifest: true # NOTE: this section is generated by synthtool.languages.python # See https://github.com/googleapis/synthtool/blob/master/synthtool/languages/python.py branches: diff --git a/.github/workflows/unittest.yml b/.github/workflows/unittest.yml index c0dca03..23000c0 100644 --- a/.github/workflows/unittest.yml +++ b/.github/workflows/unittest.yml @@ -54,4 +54,4 @@ jobs: - name: Report coverage results run: | coverage combine .coverage-results/.coverage* - coverage report --show-missing --fail-under=99 + coverage report --show-missing --fail-under=100 diff --git a/.release-please-manifest.json b/.release-please-manifest.json new file mode 100644 index 0000000..53b7bc9 --- /dev/null +++ b/.release-please-manifest.json @@ -0,0 +1,3 @@ +{ + ".": "1.1.2" +} diff --git a/docs/dataplex_v1/types.rst b/docs/dataplex_v1/types.rst index d71129b..391acd5 100644 --- a/docs/dataplex_v1/types.rst +++ b/docs/dataplex_v1/types.rst @@ -3,5 +3,4 @@ Types for Google Cloud Dataplex v1 API .. automodule:: google.cloud.dataplex_v1.types :members: - :undoc-members: :show-inheritance: diff --git a/google/cloud/dataplex/__init__.py b/google/cloud/dataplex/__init__.py index ed37721..c63eb5f 100644 --- a/google/cloud/dataplex/__init__.py +++ b/google/cloud/dataplex/__init__.py @@ -13,6 +13,10 @@ # See the License for the specific language governing permissions and # limitations under the License. # +from google.cloud.dataplex import gapic_version as package_version + +__version__ = package_version.__version__ + from google.cloud.dataplex_v1.services.content_service.async_client import ( ContentServiceAsyncClient, diff --git a/google/cloud/dataplex/gapic_version.py b/google/cloud/dataplex/gapic_version.py new file mode 100644 index 0000000..19bb9db --- /dev/null +++ b/google/cloud/dataplex/gapic_version.py @@ -0,0 +1,16 @@ +# -*- coding: utf-8 -*- +# Copyright 2022 Google LLC +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +# +__version__ = "1.1.2" # {x-release-please-version} diff --git a/google/cloud/dataplex_v1/__init__.py b/google/cloud/dataplex_v1/__init__.py index 72a3b30..442a201 100644 --- a/google/cloud/dataplex_v1/__init__.py +++ b/google/cloud/dataplex_v1/__init__.py @@ -13,6 +13,10 @@ # See the License for the specific language governing permissions and # limitations under the License. # +from google.cloud.dataplex import gapic_version as package_version + +__version__ = package_version.__version__ + from .services.content_service import ContentServiceAsyncClient, ContentServiceClient from .services.dataplex_service import DataplexServiceAsyncClient, DataplexServiceClient diff --git a/google/cloud/dataplex_v1/services/content_service/async_client.py b/google/cloud/dataplex_v1/services/content_service/async_client.py index 7342dd8..1a4dcc8 100644 --- a/google/cloud/dataplex_v1/services/content_service/async_client.py +++ b/google/cloud/dataplex_v1/services/content_service/async_client.py @@ -16,7 +16,17 @@ from collections import OrderedDict import functools import re -from typing import Dict, Mapping, Optional, Sequence, Tuple, Type, Union +from typing import ( + Dict, + Mapping, + MutableMapping, + MutableSequence, + Optional, + Sequence, + Tuple, + Type, + Union, +) from google.api_core import exceptions as core_exceptions from google.api_core import gapic_v1 @@ -170,9 +180,9 @@ def transport(self) -> ContentServiceTransport: def __init__( self, *, - credentials: ga_credentials.Credentials = None, + credentials: Optional[ga_credentials.Credentials] = None, transport: Union[str, ContentServiceTransport] = "grpc_asyncio", - client_options: ClientOptions = None, + client_options: Optional[ClientOptions] = None, client_info: gapic_v1.client_info.ClientInfo = DEFAULT_CLIENT_INFO, ) -> None: """Instantiates the content service client. @@ -216,12 +226,12 @@ def __init__( async def create_content( self, - request: Union[gcd_content.CreateContentRequest, dict] = None, + request: Optional[Union[gcd_content.CreateContentRequest, dict]] = None, *, - parent: str = None, - content: analyze.Content = None, + parent: Optional[str] = None, + content: Optional[analyze.Content] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> analyze.Content: r"""Create a content. @@ -259,7 +269,7 @@ async def sample_create_content(): print(response) Args: - request (Union[google.cloud.dataplex_v1.types.CreateContentRequest, dict]): + request (Optional[Union[google.cloud.dataplex_v1.types.CreateContentRequest, dict]]): The request object. Create content request. parent (:class:`str`): Required. The resource name of the parent lake: @@ -331,12 +341,12 @@ async def sample_create_content(): async def update_content( self, - request: Union[gcd_content.UpdateContentRequest, dict] = None, + request: Optional[Union[gcd_content.UpdateContentRequest, dict]] = None, *, - content: analyze.Content = None, - update_mask: field_mask_pb2.FieldMask = None, + content: Optional[analyze.Content] = None, + update_mask: Optional[field_mask_pb2.FieldMask] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> analyze.Content: r"""Update a content. Only supports full resource update. @@ -373,7 +383,7 @@ async def sample_update_content(): print(response) Args: - request (Union[google.cloud.dataplex_v1.types.UpdateContentRequest, dict]): + request (Optional[Union[google.cloud.dataplex_v1.types.UpdateContentRequest, dict]]): The request object. Update content request. content (:class:`google.cloud.dataplex_v1.types.Content`): Required. Update description. Only fields specified in @@ -447,11 +457,11 @@ async def sample_update_content(): async def delete_content( self, - request: Union[content.DeleteContentRequest, dict] = None, + request: Optional[Union[content.DeleteContentRequest, dict]] = None, *, - name: str = None, + name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> None: r"""Delete a content. @@ -480,7 +490,7 @@ async def sample_delete_content(): await client.delete_content(request=request) Args: - request (Union[google.cloud.dataplex_v1.types.DeleteContentRequest, dict]): + request (Optional[Union[google.cloud.dataplex_v1.types.DeleteContentRequest, dict]]): The request object. Delete content request. name (:class:`str`): Required. The resource name of the content: @@ -536,11 +546,11 @@ async def sample_delete_content(): async def get_content( self, - request: Union[content.GetContentRequest, dict] = None, + request: Optional[Union[content.GetContentRequest, dict]] = None, *, - name: str = None, + name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> analyze.Content: r"""Get a content resource. @@ -572,7 +582,7 @@ async def sample_get_content(): print(response) Args: - request (Union[google.cloud.dataplex_v1.types.GetContentRequest, dict]): + request (Optional[Union[google.cloud.dataplex_v1.types.GetContentRequest, dict]]): The request object. Get content request. name (:class:`str`): Required. The resource name of the content: @@ -646,11 +656,11 @@ async def sample_get_content(): async def get_iam_policy( self, - request: Union[iam_policy_pb2.GetIamPolicyRequest, dict] = None, + request: Optional[Union[iam_policy_pb2.GetIamPolicyRequest, dict]] = None, *, - resource: str = None, + resource: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> policy_pb2.Policy: r"""Gets the access control policy for a contentitem resource. A @@ -689,7 +699,7 @@ async def sample_get_iam_policy(): print(response) Args: - request (Union[google.iam.v1.iam_policy_pb2.GetIamPolicyRequest, dict]): + request (Optional[Union[google.iam.v1.iam_policy_pb2.GetIamPolicyRequest, dict]]): The request object. Request message for `GetIamPolicy` method. resource (:class:`str`): @@ -827,10 +837,10 @@ async def sample_get_iam_policy(): async def set_iam_policy( self, - request: Union[iam_policy_pb2.SetIamPolicyRequest, dict] = None, + request: Optional[Union[iam_policy_pb2.SetIamPolicyRequest, dict]] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> policy_pb2.Policy: r"""Sets the access control policy on the specified contentitem @@ -867,7 +877,7 @@ async def sample_set_iam_policy(): print(response) Args: - request (Union[google.iam.v1.iam_policy_pb2.SetIamPolicyRequest, dict]): + request (Optional[Union[google.iam.v1.iam_policy_pb2.SetIamPolicyRequest, dict]]): The request object. Request message for `SetIamPolicy` method. retry (google.api_core.retry.Retry): Designation of what errors, if any, @@ -974,10 +984,10 @@ async def sample_set_iam_policy(): async def test_iam_permissions( self, - request: Union[iam_policy_pb2.TestIamPermissionsRequest, dict] = None, + request: Optional[Union[iam_policy_pb2.TestIamPermissionsRequest, dict]] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> iam_policy_pb2.TestIamPermissionsResponse: r"""Returns the caller's permissions on a resource. If the resource @@ -1021,7 +1031,7 @@ async def sample_test_iam_permissions(): print(response) Args: - request (Union[google.iam.v1.iam_policy_pb2.TestIamPermissionsRequest, dict]): + request (Optional[Union[google.iam.v1.iam_policy_pb2.TestIamPermissionsRequest, dict]]): The request object. Request message for `TestIamPermissions` method. retry (google.api_core.retry.Retry): Designation of what errors, if any, @@ -1076,11 +1086,11 @@ async def sample_test_iam_permissions(): async def list_content( self, - request: Union[content.ListContentRequest, dict] = None, + request: Optional[Union[content.ListContentRequest, dict]] = None, *, - parent: str = None, + parent: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.ListContentAsyncPager: r"""List content. @@ -1113,7 +1123,7 @@ async def sample_list_content(): print(response) Args: - request (Union[google.cloud.dataplex_v1.types.ListContentRequest, dict]): + request (Optional[Union[google.cloud.dataplex_v1.types.ListContentRequest, dict]]): The request object. List content request. Returns the BASIC Content view. parent (:class:`str`): @@ -1199,10 +1209,10 @@ async def sample_list_content(): async def list_operations( self, - request: operations_pb2.ListOperationsRequest = None, + request: Optional[operations_pb2.ListOperationsRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> operations_pb2.ListOperationsResponse: r"""Lists operations that match the specified filter in the request. @@ -1253,10 +1263,10 @@ async def list_operations( async def get_operation( self, - request: operations_pb2.GetOperationRequest = None, + request: Optional[operations_pb2.GetOperationRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> operations_pb2.Operation: r"""Gets the latest state of a long-running operation. @@ -1307,10 +1317,10 @@ async def get_operation( async def delete_operation( self, - request: operations_pb2.DeleteOperationRequest = None, + request: Optional[operations_pb2.DeleteOperationRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> None: r"""Deletes a long-running operation. @@ -1362,10 +1372,10 @@ async def delete_operation( async def cancel_operation( self, - request: operations_pb2.CancelOperationRequest = None, + request: Optional[operations_pb2.CancelOperationRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> None: r"""Starts asynchronous cancellation on a long-running operation. @@ -1416,10 +1426,10 @@ async def cancel_operation( async def get_location( self, - request: locations_pb2.GetLocationRequest = None, + request: Optional[locations_pb2.GetLocationRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> locations_pb2.Location: r"""Gets information about a location. @@ -1470,10 +1480,10 @@ async def get_location( async def list_locations( self, - request: locations_pb2.ListLocationsRequest = None, + request: Optional[locations_pb2.ListLocationsRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> locations_pb2.ListLocationsResponse: r"""Lists information about the supported locations for this service. diff --git a/google/cloud/dataplex_v1/services/content_service/client.py b/google/cloud/dataplex_v1/services/content_service/client.py index a08d232..79b51fc 100644 --- a/google/cloud/dataplex_v1/services/content_service/client.py +++ b/google/cloud/dataplex_v1/services/content_service/client.py @@ -16,7 +16,18 @@ from collections import OrderedDict import os import re -from typing import Dict, Mapping, Optional, Sequence, Tuple, Type, Union +from typing import ( + Dict, + Mapping, + MutableMapping, + MutableSequence, + Optional, + Sequence, + Tuple, + Type, + Union, + cast, +) from google.api_core import client_options as client_options_lib from google.api_core import exceptions as core_exceptions @@ -67,7 +78,7 @@ class ContentServiceClientMeta(type): def get_transport_class( cls, - label: str = None, + label: Optional[str] = None, ) -> Type[ContentServiceTransport]: """Returns an appropriate transport class. @@ -366,8 +377,8 @@ def __init__( self, *, credentials: Optional[ga_credentials.Credentials] = None, - transport: Union[str, ContentServiceTransport, None] = None, - client_options: Optional[client_options_lib.ClientOptions] = None, + transport: Optional[Union[str, ContentServiceTransport]] = None, + client_options: Optional[Union[client_options_lib.ClientOptions, dict]] = None, client_info: gapic_v1.client_info.ClientInfo = DEFAULT_CLIENT_INFO, ) -> None: """Instantiates the content service client. @@ -381,7 +392,7 @@ def __init__( transport (Union[str, ContentServiceTransport]): The transport to use. If set to None, a transport is chosen automatically. - client_options (google.api_core.client_options.ClientOptions): Custom options for the + client_options (Optional[Union[google.api_core.client_options.ClientOptions, dict]]): Custom options for the client. It won't take effect if a ``transport`` instance is provided. (1) The ``api_endpoint`` property can be used to override the default endpoint provided by the client. GOOGLE_API_USE_MTLS_ENDPOINT @@ -411,6 +422,7 @@ def __init__( client_options = client_options_lib.from_dict(client_options) if client_options is None: client_options = client_options_lib.ClientOptions() + client_options = cast(client_options_lib.ClientOptions, client_options) api_endpoint, client_cert_source_func = self.get_mtls_endpoint_and_cert_source( client_options @@ -463,12 +475,12 @@ def __init__( def create_content( self, - request: Union[gcd_content.CreateContentRequest, dict] = None, + request: Optional[Union[gcd_content.CreateContentRequest, dict]] = None, *, - parent: str = None, - content: analyze.Content = None, + parent: Optional[str] = None, + content: Optional[analyze.Content] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> analyze.Content: r"""Create a content. @@ -578,12 +590,12 @@ def sample_create_content(): def update_content( self, - request: Union[gcd_content.UpdateContentRequest, dict] = None, + request: Optional[Union[gcd_content.UpdateContentRequest, dict]] = None, *, - content: analyze.Content = None, - update_mask: field_mask_pb2.FieldMask = None, + content: Optional[analyze.Content] = None, + update_mask: Optional[field_mask_pb2.FieldMask] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> analyze.Content: r"""Update a content. Only supports full resource update. @@ -694,11 +706,11 @@ def sample_update_content(): def delete_content( self, - request: Union[content.DeleteContentRequest, dict] = None, + request: Optional[Union[content.DeleteContentRequest, dict]] = None, *, - name: str = None, + name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> None: r"""Delete a content. @@ -783,11 +795,11 @@ def sample_delete_content(): def get_content( self, - request: Union[content.GetContentRequest, dict] = None, + request: Optional[Union[content.GetContentRequest, dict]] = None, *, - name: str = None, + name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> analyze.Content: r"""Get a content resource. @@ -884,11 +896,11 @@ def sample_get_content(): def get_iam_policy( self, - request: Union[iam_policy_pb2.GetIamPolicyRequest, dict] = None, + request: Optional[Union[iam_policy_pb2.GetIamPolicyRequest, dict]] = None, *, - resource: str = None, + resource: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> policy_pb2.Policy: r"""Gets the access control policy for a contentitem resource. A @@ -1053,10 +1065,10 @@ def sample_get_iam_policy(): def set_iam_policy( self, - request: Union[iam_policy_pb2.SetIamPolicyRequest, dict] = None, + request: Optional[Union[iam_policy_pb2.SetIamPolicyRequest, dict]] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> policy_pb2.Policy: r"""Sets the access control policy on the specified contentitem @@ -1199,10 +1211,10 @@ def sample_set_iam_policy(): def test_iam_permissions( self, - request: Union[iam_policy_pb2.TestIamPermissionsRequest, dict] = None, + request: Optional[Union[iam_policy_pb2.TestIamPermissionsRequest, dict]] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> iam_policy_pb2.TestIamPermissionsResponse: r"""Returns the caller's permissions on a resource. If the resource @@ -1291,11 +1303,11 @@ def sample_test_iam_permissions(): def list_content( self, - request: Union[content.ListContentRequest, dict] = None, + request: Optional[Union[content.ListContentRequest, dict]] = None, *, - parent: str = None, + parent: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.ListContentPager: r"""List content. @@ -1418,10 +1430,10 @@ def __exit__(self, type, value, traceback): def list_operations( self, - request: operations_pb2.ListOperationsRequest = None, + request: Optional[operations_pb2.ListOperationsRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> operations_pb2.ListOperationsResponse: r"""Lists operations that match the specified filter in the request. @@ -1472,10 +1484,10 @@ def list_operations( def get_operation( self, - request: operations_pb2.GetOperationRequest = None, + request: Optional[operations_pb2.GetOperationRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> operations_pb2.Operation: r"""Gets the latest state of a long-running operation. @@ -1526,10 +1538,10 @@ def get_operation( def delete_operation( self, - request: operations_pb2.DeleteOperationRequest = None, + request: Optional[operations_pb2.DeleteOperationRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> None: r"""Deletes a long-running operation. @@ -1581,10 +1593,10 @@ def delete_operation( def cancel_operation( self, - request: operations_pb2.CancelOperationRequest = None, + request: Optional[operations_pb2.CancelOperationRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> None: r"""Starts asynchronous cancellation on a long-running operation. @@ -1635,10 +1647,10 @@ def cancel_operation( def get_location( self, - request: locations_pb2.GetLocationRequest = None, + request: Optional[locations_pb2.GetLocationRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> locations_pb2.Location: r"""Gets information about a location. @@ -1689,10 +1701,10 @@ def get_location( def list_locations( self, - request: locations_pb2.ListLocationsRequest = None, + request: Optional[locations_pb2.ListLocationsRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> locations_pb2.ListLocationsResponse: r"""Lists information about the supported locations for this service. diff --git a/google/cloud/dataplex_v1/services/content_service/transports/base.py b/google/cloud/dataplex_v1/services/content_service/transports/base.py index f89cd34..9c42e86 100644 --- a/google/cloud/dataplex_v1/services/content_service/transports/base.py +++ b/google/cloud/dataplex_v1/services/content_service/transports/base.py @@ -55,7 +55,7 @@ def __init__( self, *, host: str = DEFAULT_HOST, - credentials: ga_credentials.Credentials = None, + credentials: Optional[ga_credentials.Credentials] = None, credentials_file: Optional[str] = None, scopes: Optional[Sequence[str]] = None, quota_project_id: Optional[str] = None, diff --git a/google/cloud/dataplex_v1/services/content_service/transports/grpc.py b/google/cloud/dataplex_v1/services/content_service/transports/grpc.py index 81f4bed..ca00d75 100644 --- a/google/cloud/dataplex_v1/services/content_service/transports/grpc.py +++ b/google/cloud/dataplex_v1/services/content_service/transports/grpc.py @@ -53,14 +53,14 @@ def __init__( self, *, host: str = "dataplex.googleapis.com", - credentials: ga_credentials.Credentials = None, - credentials_file: str = None, - scopes: Sequence[str] = None, - channel: grpc.Channel = None, - api_mtls_endpoint: str = None, - client_cert_source: Callable[[], Tuple[bytes, bytes]] = None, - ssl_channel_credentials: grpc.ChannelCredentials = None, - client_cert_source_for_mtls: Callable[[], Tuple[bytes, bytes]] = None, + credentials: Optional[ga_credentials.Credentials] = None, + credentials_file: Optional[str] = None, + scopes: Optional[Sequence[str]] = None, + channel: Optional[grpc.Channel] = None, + api_mtls_endpoint: Optional[str] = None, + client_cert_source: Optional[Callable[[], Tuple[bytes, bytes]]] = None, + ssl_channel_credentials: Optional[grpc.ChannelCredentials] = None, + client_cert_source_for_mtls: Optional[Callable[[], Tuple[bytes, bytes]]] = None, quota_project_id: Optional[str] = None, client_info: gapic_v1.client_info.ClientInfo = DEFAULT_CLIENT_INFO, always_use_jwt_access: Optional[bool] = False, @@ -187,8 +187,8 @@ def __init__( def create_channel( cls, host: str = "dataplex.googleapis.com", - credentials: ga_credentials.Credentials = None, - credentials_file: str = None, + credentials: Optional[ga_credentials.Credentials] = None, + credentials_file: Optional[str] = None, scopes: Optional[Sequence[str]] = None, quota_project_id: Optional[str] = None, **kwargs, diff --git a/google/cloud/dataplex_v1/services/content_service/transports/grpc_asyncio.py b/google/cloud/dataplex_v1/services/content_service/transports/grpc_asyncio.py index 6a9c315..edf53a0 100644 --- a/google/cloud/dataplex_v1/services/content_service/transports/grpc_asyncio.py +++ b/google/cloud/dataplex_v1/services/content_service/transports/grpc_asyncio.py @@ -55,7 +55,7 @@ class ContentServiceGrpcAsyncIOTransport(ContentServiceTransport): def create_channel( cls, host: str = "dataplex.googleapis.com", - credentials: ga_credentials.Credentials = None, + credentials: Optional[ga_credentials.Credentials] = None, credentials_file: Optional[str] = None, scopes: Optional[Sequence[str]] = None, quota_project_id: Optional[str] = None, @@ -98,15 +98,15 @@ def __init__( self, *, host: str = "dataplex.googleapis.com", - credentials: ga_credentials.Credentials = None, + credentials: Optional[ga_credentials.Credentials] = None, credentials_file: Optional[str] = None, scopes: Optional[Sequence[str]] = None, - channel: aio.Channel = None, - api_mtls_endpoint: str = None, - client_cert_source: Callable[[], Tuple[bytes, bytes]] = None, - ssl_channel_credentials: grpc.ChannelCredentials = None, - client_cert_source_for_mtls: Callable[[], Tuple[bytes, bytes]] = None, - quota_project_id=None, + channel: Optional[aio.Channel] = None, + api_mtls_endpoint: Optional[str] = None, + client_cert_source: Optional[Callable[[], Tuple[bytes, bytes]]] = None, + ssl_channel_credentials: Optional[grpc.ChannelCredentials] = None, + client_cert_source_for_mtls: Optional[Callable[[], Tuple[bytes, bytes]]] = None, + quota_project_id: Optional[str] = None, client_info: gapic_v1.client_info.ClientInfo = DEFAULT_CLIENT_INFO, always_use_jwt_access: Optional[bool] = False, api_audience: Optional[str] = None, diff --git a/google/cloud/dataplex_v1/services/dataplex_service/async_client.py b/google/cloud/dataplex_v1/services/dataplex_service/async_client.py index a783749..5e261d7 100644 --- a/google/cloud/dataplex_v1/services/dataplex_service/async_client.py +++ b/google/cloud/dataplex_v1/services/dataplex_service/async_client.py @@ -16,7 +16,17 @@ from collections import OrderedDict import functools import re -from typing import Dict, Mapping, Optional, Sequence, Tuple, Type, Union +from typing import ( + Dict, + Mapping, + MutableMapping, + MutableSequence, + Optional, + Sequence, + Tuple, + Type, + Union, +) from google.api_core import exceptions as core_exceptions from google.api_core import gapic_v1 @@ -189,9 +199,9 @@ def transport(self) -> DataplexServiceTransport: def __init__( self, *, - credentials: ga_credentials.Credentials = None, + credentials: Optional[ga_credentials.Credentials] = None, transport: Union[str, DataplexServiceTransport] = "grpc_asyncio", - client_options: ClientOptions = None, + client_options: Optional[ClientOptions] = None, client_info: gapic_v1.client_info.ClientInfo = DEFAULT_CLIENT_INFO, ) -> None: """Instantiates the dataplex service client. @@ -235,13 +245,13 @@ def __init__( async def create_lake( self, - request: Union[service.CreateLakeRequest, dict] = None, + request: Optional[Union[service.CreateLakeRequest, dict]] = None, *, - parent: str = None, - lake: resources.Lake = None, - lake_id: str = None, + parent: Optional[str] = None, + lake: Optional[resources.Lake] = None, + lake_id: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> operation_async.AsyncOperation: r"""Creates a lake resource. @@ -278,7 +288,7 @@ async def sample_create_lake(): print(response) Args: - request (Union[google.cloud.dataplex_v1.types.CreateLakeRequest, dict]): + request (Optional[Union[google.cloud.dataplex_v1.types.CreateLakeRequest, dict]]): The request object. Create lake request. parent (:class:`str`): Required. The resource name of the lake location, of the @@ -389,12 +399,12 @@ async def sample_create_lake(): async def update_lake( self, - request: Union[service.UpdateLakeRequest, dict] = None, + request: Optional[Union[service.UpdateLakeRequest, dict]] = None, *, - lake: resources.Lake = None, - update_mask: field_mask_pb2.FieldMask = None, + lake: Optional[resources.Lake] = None, + update_mask: Optional[field_mask_pb2.FieldMask] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> operation_async.AsyncOperation: r"""Updates a lake resource. @@ -429,7 +439,7 @@ async def sample_update_lake(): print(response) Args: - request (Union[google.cloud.dataplex_v1.types.UpdateLakeRequest, dict]): + request (Optional[Union[google.cloud.dataplex_v1.types.UpdateLakeRequest, dict]]): The request object. Update lake request. lake (:class:`google.cloud.dataplex_v1.types.Lake`): Required. Update description. Only fields specified in @@ -523,11 +533,11 @@ async def sample_update_lake(): async def delete_lake( self, - request: Union[service.DeleteLakeRequest, dict] = None, + request: Optional[Union[service.DeleteLakeRequest, dict]] = None, *, - name: str = None, + name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> operation_async.AsyncOperation: r"""Deletes a lake resource. All zones within the lake @@ -564,7 +574,7 @@ async def sample_delete_lake(): print(response) Args: - request (Union[google.cloud.dataplex_v1.types.DeleteLakeRequest, dict]): + request (Optional[Union[google.cloud.dataplex_v1.types.DeleteLakeRequest, dict]]): The request object. Delete lake request. name (:class:`str`): Required. The resource name of the lake: @@ -647,11 +657,11 @@ async def sample_delete_lake(): async def list_lakes( self, - request: Union[service.ListLakesRequest, dict] = None, + request: Optional[Union[service.ListLakesRequest, dict]] = None, *, - parent: str = None, + parent: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.ListLakesAsyncPager: r"""Lists lake resources in a project and location. @@ -684,7 +694,7 @@ async def sample_list_lakes(): print(response) Args: - request (Union[google.cloud.dataplex_v1.types.ListLakesRequest, dict]): + request (Optional[Union[google.cloud.dataplex_v1.types.ListLakesRequest, dict]]): The request object. List lakes request. parent (:class:`str`): Required. The resource name of the lake location, of the @@ -771,11 +781,11 @@ async def sample_list_lakes(): async def get_lake( self, - request: Union[service.GetLakeRequest, dict] = None, + request: Optional[Union[service.GetLakeRequest, dict]] = None, *, - name: str = None, + name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> resources.Lake: r"""Retrieves a lake resource. @@ -807,7 +817,7 @@ async def sample_get_lake(): print(response) Args: - request (Union[google.cloud.dataplex_v1.types.GetLakeRequest, dict]): + request (Optional[Union[google.cloud.dataplex_v1.types.GetLakeRequest, dict]]): The request object. Get lake request. name (:class:`str`): Required. The resource name of the lake: @@ -895,11 +905,11 @@ async def sample_get_lake(): async def list_lake_actions( self, - request: Union[service.ListLakeActionsRequest, dict] = None, + request: Optional[Union[service.ListLakeActionsRequest, dict]] = None, *, - parent: str = None, + parent: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.ListLakeActionsAsyncPager: r"""Lists action resources in a lake. @@ -932,7 +942,7 @@ async def sample_list_lake_actions(): print(response) Args: - request (Union[google.cloud.dataplex_v1.types.ListLakeActionsRequest, dict]): + request (Optional[Union[google.cloud.dataplex_v1.types.ListLakeActionsRequest, dict]]): The request object. List lake actions request. parent (:class:`str`): Required. The resource name of the parent lake: @@ -1017,13 +1027,13 @@ async def sample_list_lake_actions(): async def create_zone( self, - request: Union[service.CreateZoneRequest, dict] = None, + request: Optional[Union[service.CreateZoneRequest, dict]] = None, *, - parent: str = None, - zone: resources.Zone = None, - zone_id: str = None, + parent: Optional[str] = None, + zone: Optional[resources.Zone] = None, + zone_id: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> operation_async.AsyncOperation: r"""Creates a zone resource within a lake. @@ -1065,7 +1075,7 @@ async def sample_create_zone(): print(response) Args: - request (Union[google.cloud.dataplex_v1.types.CreateZoneRequest, dict]): + request (Optional[Union[google.cloud.dataplex_v1.types.CreateZoneRequest, dict]]): The request object. Create zone request. parent (:class:`str`): Required. The resource name of the parent lake: @@ -1170,12 +1180,12 @@ async def sample_create_zone(): async def update_zone( self, - request: Union[service.UpdateZoneRequest, dict] = None, + request: Optional[Union[service.UpdateZoneRequest, dict]] = None, *, - zone: resources.Zone = None, - update_mask: field_mask_pb2.FieldMask = None, + zone: Optional[resources.Zone] = None, + update_mask: Optional[field_mask_pb2.FieldMask] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> operation_async.AsyncOperation: r"""Updates a zone resource. @@ -1215,7 +1225,7 @@ async def sample_update_zone(): print(response) Args: - request (Union[google.cloud.dataplex_v1.types.UpdateZoneRequest, dict]): + request (Optional[Union[google.cloud.dataplex_v1.types.UpdateZoneRequest, dict]]): The request object. Update zone request. zone (:class:`google.cloud.dataplex_v1.types.Zone`): Required. Update description. Only fields specified in @@ -1302,11 +1312,11 @@ async def sample_update_zone(): async def delete_zone( self, - request: Union[service.DeleteZoneRequest, dict] = None, + request: Optional[Union[service.DeleteZoneRequest, dict]] = None, *, - name: str = None, + name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> operation_async.AsyncOperation: r"""Deletes a zone resource. All assets within a zone @@ -1343,7 +1353,7 @@ async def sample_delete_zone(): print(response) Args: - request (Union[google.cloud.dataplex_v1.types.DeleteZoneRequest, dict]): + request (Optional[Union[google.cloud.dataplex_v1.types.DeleteZoneRequest, dict]]): The request object. Delete zone request. name (:class:`str`): Required. The resource name of the zone: @@ -1426,11 +1436,11 @@ async def sample_delete_zone(): async def list_zones( self, - request: Union[service.ListZonesRequest, dict] = None, + request: Optional[Union[service.ListZonesRequest, dict]] = None, *, - parent: str = None, + parent: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.ListZonesAsyncPager: r"""Lists zone resources in a lake. @@ -1463,7 +1473,7 @@ async def sample_list_zones(): print(response) Args: - request (Union[google.cloud.dataplex_v1.types.ListZonesRequest, dict]): + request (Optional[Union[google.cloud.dataplex_v1.types.ListZonesRequest, dict]]): The request object. List zones request. parent (:class:`str`): Required. The resource name of the parent lake: @@ -1548,11 +1558,11 @@ async def sample_list_zones(): async def get_zone( self, - request: Union[service.GetZoneRequest, dict] = None, + request: Optional[Union[service.GetZoneRequest, dict]] = None, *, - name: str = None, + name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> resources.Zone: r"""Retrieves a zone resource. @@ -1584,7 +1594,7 @@ async def sample_get_zone(): print(response) Args: - request (Union[google.cloud.dataplex_v1.types.GetZoneRequest, dict]): + request (Optional[Union[google.cloud.dataplex_v1.types.GetZoneRequest, dict]]): The request object. Get zone request. name (:class:`str`): Required. The resource name of the zone: @@ -1664,11 +1674,11 @@ async def sample_get_zone(): async def list_zone_actions( self, - request: Union[service.ListZoneActionsRequest, dict] = None, + request: Optional[Union[service.ListZoneActionsRequest, dict]] = None, *, - parent: str = None, + parent: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.ListZoneActionsAsyncPager: r"""Lists action resources in a zone. @@ -1701,7 +1711,7 @@ async def sample_list_zone_actions(): print(response) Args: - request (Union[google.cloud.dataplex_v1.types.ListZoneActionsRequest, dict]): + request (Optional[Union[google.cloud.dataplex_v1.types.ListZoneActionsRequest, dict]]): The request object. List zone actions request. parent (:class:`str`): Required. The resource name of the parent zone: @@ -1786,13 +1796,13 @@ async def sample_list_zone_actions(): async def create_asset( self, - request: Union[service.CreateAssetRequest, dict] = None, + request: Optional[Union[service.CreateAssetRequest, dict]] = None, *, - parent: str = None, - asset: resources.Asset = None, - asset_id: str = None, + parent: Optional[str] = None, + asset: Optional[resources.Asset] = None, + asset_id: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> operation_async.AsyncOperation: r"""Creates an asset resource. @@ -1833,7 +1843,7 @@ async def sample_create_asset(): print(response) Args: - request (Union[google.cloud.dataplex_v1.types.CreateAssetRequest, dict]): + request (Optional[Union[google.cloud.dataplex_v1.types.CreateAssetRequest, dict]]): The request object. Create asset request. parent (:class:`str`): Required. The resource name of the parent zone: @@ -1932,12 +1942,12 @@ async def sample_create_asset(): async def update_asset( self, - request: Union[service.UpdateAssetRequest, dict] = None, + request: Optional[Union[service.UpdateAssetRequest, dict]] = None, *, - asset: resources.Asset = None, - update_mask: field_mask_pb2.FieldMask = None, + asset: Optional[resources.Asset] = None, + update_mask: Optional[field_mask_pb2.FieldMask] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> operation_async.AsyncOperation: r"""Updates an asset resource. @@ -1976,7 +1986,7 @@ async def sample_update_asset(): print(response) Args: - request (Union[google.cloud.dataplex_v1.types.UpdateAssetRequest, dict]): + request (Optional[Union[google.cloud.dataplex_v1.types.UpdateAssetRequest, dict]]): The request object. Update asset request. asset (:class:`google.cloud.dataplex_v1.types.Asset`): Required. Update description. Only fields specified in @@ -2060,11 +2070,11 @@ async def sample_update_asset(): async def delete_asset( self, - request: Union[service.DeleteAssetRequest, dict] = None, + request: Optional[Union[service.DeleteAssetRequest, dict]] = None, *, - name: str = None, + name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> operation_async.AsyncOperation: r"""Deletes an asset resource. The referenced storage @@ -2102,7 +2112,7 @@ async def sample_delete_asset(): print(response) Args: - request (Union[google.cloud.dataplex_v1.types.DeleteAssetRequest, dict]): + request (Optional[Union[google.cloud.dataplex_v1.types.DeleteAssetRequest, dict]]): The request object. Delete asset request. name (:class:`str`): Required. The resource name of the asset: @@ -2185,11 +2195,11 @@ async def sample_delete_asset(): async def list_assets( self, - request: Union[service.ListAssetsRequest, dict] = None, + request: Optional[Union[service.ListAssetsRequest, dict]] = None, *, - parent: str = None, + parent: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.ListAssetsAsyncPager: r"""Lists asset resources in a zone. @@ -2222,7 +2232,7 @@ async def sample_list_assets(): print(response) Args: - request (Union[google.cloud.dataplex_v1.types.ListAssetsRequest, dict]): + request (Optional[Union[google.cloud.dataplex_v1.types.ListAssetsRequest, dict]]): The request object. List assets request. parent (:class:`str`): Required. The resource name of the parent zone: @@ -2307,11 +2317,11 @@ async def sample_list_assets(): async def get_asset( self, - request: Union[service.GetAssetRequest, dict] = None, + request: Optional[Union[service.GetAssetRequest, dict]] = None, *, - name: str = None, + name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> resources.Asset: r"""Retrieves an asset resource. @@ -2343,7 +2353,7 @@ async def sample_get_asset(): print(response) Args: - request (Union[google.cloud.dataplex_v1.types.GetAssetRequest, dict]): + request (Optional[Union[google.cloud.dataplex_v1.types.GetAssetRequest, dict]]): The request object. Get asset request. name (:class:`str`): Required. The resource name of the asset: @@ -2418,11 +2428,11 @@ async def sample_get_asset(): async def list_asset_actions( self, - request: Union[service.ListAssetActionsRequest, dict] = None, + request: Optional[Union[service.ListAssetActionsRequest, dict]] = None, *, - parent: str = None, + parent: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.ListAssetActionsAsyncPager: r"""Lists action resources in an asset. @@ -2455,7 +2465,7 @@ async def sample_list_asset_actions(): print(response) Args: - request (Union[google.cloud.dataplex_v1.types.ListAssetActionsRequest, dict]): + request (Optional[Union[google.cloud.dataplex_v1.types.ListAssetActionsRequest, dict]]): The request object. List asset actions request. parent (:class:`str`): Required. The resource name of the parent asset: @@ -2540,13 +2550,13 @@ async def sample_list_asset_actions(): async def create_task( self, - request: Union[service.CreateTaskRequest, dict] = None, + request: Optional[Union[service.CreateTaskRequest, dict]] = None, *, - parent: str = None, - task: tasks.Task = None, - task_id: str = None, + parent: Optional[str] = None, + task: Optional[tasks.Task] = None, + task_id: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> operation_async.AsyncOperation: r"""Creates a task resource within a lake. @@ -2590,7 +2600,7 @@ async def sample_create_task(): print(response) Args: - request (Union[google.cloud.dataplex_v1.types.CreateTaskRequest, dict]): + request (Optional[Union[google.cloud.dataplex_v1.types.CreateTaskRequest, dict]]): The request object. Create task request. parent (:class:`str`): Required. The resource name of the parent lake: @@ -2680,12 +2690,12 @@ async def sample_create_task(): async def update_task( self, - request: Union[service.UpdateTaskRequest, dict] = None, + request: Optional[Union[service.UpdateTaskRequest, dict]] = None, *, - task: tasks.Task = None, - update_mask: field_mask_pb2.FieldMask = None, + task: Optional[tasks.Task] = None, + update_mask: Optional[field_mask_pb2.FieldMask] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> operation_async.AsyncOperation: r"""Update the task resource. @@ -2727,7 +2737,7 @@ async def sample_update_task(): print(response) Args: - request (Union[google.cloud.dataplex_v1.types.UpdateTaskRequest, dict]): + request (Optional[Union[google.cloud.dataplex_v1.types.UpdateTaskRequest, dict]]): The request object. Update task request. task (:class:`google.cloud.dataplex_v1.types.Task`): Required. Update description. Only fields specified in @@ -2812,11 +2822,11 @@ async def sample_update_task(): async def delete_task( self, - request: Union[service.DeleteTaskRequest, dict] = None, + request: Optional[Union[service.DeleteTaskRequest, dict]] = None, *, - name: str = None, + name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> operation_async.AsyncOperation: r"""Delete the task resource. @@ -2852,7 +2862,7 @@ async def sample_delete_task(): print(response) Args: - request (Union[google.cloud.dataplex_v1.types.DeleteTaskRequest, dict]): + request (Optional[Union[google.cloud.dataplex_v1.types.DeleteTaskRequest, dict]]): The request object. Delete task request. name (:class:`str`): Required. The resource name of the task: @@ -2935,11 +2945,11 @@ async def sample_delete_task(): async def list_tasks( self, - request: Union[service.ListTasksRequest, dict] = None, + request: Optional[Union[service.ListTasksRequest, dict]] = None, *, - parent: str = None, + parent: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.ListTasksAsyncPager: r"""Lists tasks under the given lake. @@ -2972,7 +2982,7 @@ async def sample_list_tasks(): print(response) Args: - request (Union[google.cloud.dataplex_v1.types.ListTasksRequest, dict]): + request (Optional[Union[google.cloud.dataplex_v1.types.ListTasksRequest, dict]]): The request object. List tasks request. parent (:class:`str`): Required. The resource name of the parent lake: @@ -3057,11 +3067,11 @@ async def sample_list_tasks(): async def get_task( self, - request: Union[service.GetTaskRequest, dict] = None, + request: Optional[Union[service.GetTaskRequest, dict]] = None, *, - name: str = None, + name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> tasks.Task: r"""Get task resource. @@ -3093,7 +3103,7 @@ async def sample_get_task(): print(response) Args: - request (Union[google.cloud.dataplex_v1.types.GetTaskRequest, dict]): + request (Optional[Union[google.cloud.dataplex_v1.types.GetTaskRequest, dict]]): The request object. Get task request. name (:class:`str`): Required. The resource name of the task: @@ -3165,11 +3175,11 @@ async def sample_get_task(): async def list_jobs( self, - request: Union[service.ListJobsRequest, dict] = None, + request: Optional[Union[service.ListJobsRequest, dict]] = None, *, - parent: str = None, + parent: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.ListJobsAsyncPager: r"""Lists Jobs under the given task. @@ -3202,7 +3212,7 @@ async def sample_list_jobs(): print(response) Args: - request (Union[google.cloud.dataplex_v1.types.ListJobsRequest, dict]): + request (Optional[Union[google.cloud.dataplex_v1.types.ListJobsRequest, dict]]): The request object. List jobs request. parent (:class:`str`): Required. The resource name of the parent environment: @@ -3287,11 +3297,11 @@ async def sample_list_jobs(): async def get_job( self, - request: Union[service.GetJobRequest, dict] = None, + request: Optional[Union[service.GetJobRequest, dict]] = None, *, - name: str = None, + name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> tasks.Job: r"""Get job resource. @@ -3323,7 +3333,7 @@ async def sample_get_job(): print(response) Args: - request (Union[google.cloud.dataplex_v1.types.GetJobRequest, dict]): + request (Optional[Union[google.cloud.dataplex_v1.types.GetJobRequest, dict]]): The request object. Get job request. name (:class:`str`): Required. The resource name of the job: @@ -3397,11 +3407,11 @@ async def sample_get_job(): async def cancel_job( self, - request: Union[service.CancelJobRequest, dict] = None, + request: Optional[Union[service.CancelJobRequest, dict]] = None, *, - name: str = None, + name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> None: r"""Cancel jobs running for the task resource. @@ -3430,7 +3440,7 @@ async def sample_cancel_job(): await client.cancel_job(request=request) Args: - request (Union[google.cloud.dataplex_v1.types.CancelJobRequest, dict]): + request (Optional[Union[google.cloud.dataplex_v1.types.CancelJobRequest, dict]]): The request object. Cancel task jobs. name (:class:`str`): Required. The resource name of the job: @@ -3486,13 +3496,13 @@ async def sample_cancel_job(): async def create_environment( self, - request: Union[service.CreateEnvironmentRequest, dict] = None, + request: Optional[Union[service.CreateEnvironmentRequest, dict]] = None, *, - parent: str = None, - environment: analyze.Environment = None, - environment_id: str = None, + parent: Optional[str] = None, + environment: Optional[analyze.Environment] = None, + environment_id: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> operation_async.AsyncOperation: r"""Create an environment resource. @@ -3533,7 +3543,7 @@ async def sample_create_environment(): print(response) Args: - request (Union[google.cloud.dataplex_v1.types.CreateEnvironmentRequest, dict]): + request (Optional[Union[google.cloud.dataplex_v1.types.CreateEnvironmentRequest, dict]]): The request object. Create environment request. parent (:class:`str`): Required. The resource name of the parent lake: @@ -3630,12 +3640,12 @@ async def sample_create_environment(): async def update_environment( self, - request: Union[service.UpdateEnvironmentRequest, dict] = None, + request: Optional[Union[service.UpdateEnvironmentRequest, dict]] = None, *, - environment: analyze.Environment = None, - update_mask: field_mask_pb2.FieldMask = None, + environment: Optional[analyze.Environment] = None, + update_mask: Optional[field_mask_pb2.FieldMask] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> operation_async.AsyncOperation: r"""Update the environment resource. @@ -3674,7 +3684,7 @@ async def sample_update_environment(): print(response) Args: - request (Union[google.cloud.dataplex_v1.types.UpdateEnvironmentRequest, dict]): + request (Optional[Union[google.cloud.dataplex_v1.types.UpdateEnvironmentRequest, dict]]): The request object. Update environment request. environment (:class:`google.cloud.dataplex_v1.types.Environment`): Required. Update description. Only fields specified in @@ -3758,11 +3768,11 @@ async def sample_update_environment(): async def delete_environment( self, - request: Union[service.DeleteEnvironmentRequest, dict] = None, + request: Optional[Union[service.DeleteEnvironmentRequest, dict]] = None, *, - name: str = None, + name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> operation_async.AsyncOperation: r"""Delete the environment resource. All the child @@ -3800,7 +3810,7 @@ async def sample_delete_environment(): print(response) Args: - request (Union[google.cloud.dataplex_v1.types.DeleteEnvironmentRequest, dict]): + request (Optional[Union[google.cloud.dataplex_v1.types.DeleteEnvironmentRequest, dict]]): The request object. Delete environment request. name (:class:`str`): Required. The resource name of the environment: @@ -3883,11 +3893,11 @@ async def sample_delete_environment(): async def list_environments( self, - request: Union[service.ListEnvironmentsRequest, dict] = None, + request: Optional[Union[service.ListEnvironmentsRequest, dict]] = None, *, - parent: str = None, + parent: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.ListEnvironmentsAsyncPager: r"""Lists environments under the given lake. @@ -3920,7 +3930,7 @@ async def sample_list_environments(): print(response) Args: - request (Union[google.cloud.dataplex_v1.types.ListEnvironmentsRequest, dict]): + request (Optional[Union[google.cloud.dataplex_v1.types.ListEnvironmentsRequest, dict]]): The request object. List environments request. parent (:class:`str`): Required. The resource name of the parent lake: @@ -4005,11 +4015,11 @@ async def sample_list_environments(): async def get_environment( self, - request: Union[service.GetEnvironmentRequest, dict] = None, + request: Optional[Union[service.GetEnvironmentRequest, dict]] = None, *, - name: str = None, + name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> analyze.Environment: r"""Get environment resource. @@ -4041,7 +4051,7 @@ async def sample_get_environment(): print(response) Args: - request (Union[google.cloud.dataplex_v1.types.GetEnvironmentRequest, dict]): + request (Optional[Union[google.cloud.dataplex_v1.types.GetEnvironmentRequest, dict]]): The request object. Get environment request. name (:class:`str`): Required. The resource name of the environment: @@ -4116,11 +4126,11 @@ async def sample_get_environment(): async def list_sessions( self, - request: Union[service.ListSessionsRequest, dict] = None, + request: Optional[Union[service.ListSessionsRequest, dict]] = None, *, - parent: str = None, + parent: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.ListSessionsAsyncPager: r"""Lists session resources in an environment. @@ -4153,7 +4163,7 @@ async def sample_list_sessions(): print(response) Args: - request (Union[google.cloud.dataplex_v1.types.ListSessionsRequest, dict]): + request (Optional[Union[google.cloud.dataplex_v1.types.ListSessionsRequest, dict]]): The request object. List sessions request. parent (:class:`str`): Required. The resource name of the parent environment: @@ -4229,10 +4239,10 @@ async def sample_list_sessions(): async def list_operations( self, - request: operations_pb2.ListOperationsRequest = None, + request: Optional[operations_pb2.ListOperationsRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> operations_pb2.ListOperationsResponse: r"""Lists operations that match the specified filter in the request. @@ -4283,10 +4293,10 @@ async def list_operations( async def get_operation( self, - request: operations_pb2.GetOperationRequest = None, + request: Optional[operations_pb2.GetOperationRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> operations_pb2.Operation: r"""Gets the latest state of a long-running operation. @@ -4337,10 +4347,10 @@ async def get_operation( async def delete_operation( self, - request: operations_pb2.DeleteOperationRequest = None, + request: Optional[operations_pb2.DeleteOperationRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> None: r"""Deletes a long-running operation. @@ -4392,10 +4402,10 @@ async def delete_operation( async def cancel_operation( self, - request: operations_pb2.CancelOperationRequest = None, + request: Optional[operations_pb2.CancelOperationRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> None: r"""Starts asynchronous cancellation on a long-running operation. @@ -4446,10 +4456,10 @@ async def cancel_operation( async def get_location( self, - request: locations_pb2.GetLocationRequest = None, + request: Optional[locations_pb2.GetLocationRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> locations_pb2.Location: r"""Gets information about a location. @@ -4500,10 +4510,10 @@ async def get_location( async def list_locations( self, - request: locations_pb2.ListLocationsRequest = None, + request: Optional[locations_pb2.ListLocationsRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> locations_pb2.ListLocationsResponse: r"""Lists information about the supported locations for this service. diff --git a/google/cloud/dataplex_v1/services/dataplex_service/client.py b/google/cloud/dataplex_v1/services/dataplex_service/client.py index b1f40ae..2b289b5 100644 --- a/google/cloud/dataplex_v1/services/dataplex_service/client.py +++ b/google/cloud/dataplex_v1/services/dataplex_service/client.py @@ -16,7 +16,18 @@ from collections import OrderedDict import os import re -from typing import Dict, Mapping, Optional, Sequence, Tuple, Type, Union +from typing import ( + Dict, + Mapping, + MutableMapping, + MutableSequence, + Optional, + Sequence, + Tuple, + Type, + Union, + cast, +) from google.api_core import client_options as client_options_lib from google.api_core import exceptions as core_exceptions @@ -68,7 +79,7 @@ class DataplexServiceClientMeta(type): def get_transport_class( cls, - label: str = None, + label: Optional[str] = None, ) -> Type[DataplexServiceTransport]: """Returns an appropriate transport class. @@ -527,8 +538,8 @@ def __init__( self, *, credentials: Optional[ga_credentials.Credentials] = None, - transport: Union[str, DataplexServiceTransport, None] = None, - client_options: Optional[client_options_lib.ClientOptions] = None, + transport: Optional[Union[str, DataplexServiceTransport]] = None, + client_options: Optional[Union[client_options_lib.ClientOptions, dict]] = None, client_info: gapic_v1.client_info.ClientInfo = DEFAULT_CLIENT_INFO, ) -> None: """Instantiates the dataplex service client. @@ -542,7 +553,7 @@ def __init__( transport (Union[str, DataplexServiceTransport]): The transport to use. If set to None, a transport is chosen automatically. - client_options (google.api_core.client_options.ClientOptions): Custom options for the + client_options (Optional[Union[google.api_core.client_options.ClientOptions, dict]]): Custom options for the client. It won't take effect if a ``transport`` instance is provided. (1) The ``api_endpoint`` property can be used to override the default endpoint provided by the client. GOOGLE_API_USE_MTLS_ENDPOINT @@ -572,6 +583,7 @@ def __init__( client_options = client_options_lib.from_dict(client_options) if client_options is None: client_options = client_options_lib.ClientOptions() + client_options = cast(client_options_lib.ClientOptions, client_options) api_endpoint, client_cert_source_func = self.get_mtls_endpoint_and_cert_source( client_options @@ -624,13 +636,13 @@ def __init__( def create_lake( self, - request: Union[service.CreateLakeRequest, dict] = None, + request: Optional[Union[service.CreateLakeRequest, dict]] = None, *, - parent: str = None, - lake: resources.Lake = None, - lake_id: str = None, + parent: Optional[str] = None, + lake: Optional[resources.Lake] = None, + lake_id: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> operation.Operation: r"""Creates a lake resource. @@ -778,12 +790,12 @@ def sample_create_lake(): def update_lake( self, - request: Union[service.UpdateLakeRequest, dict] = None, + request: Optional[Union[service.UpdateLakeRequest, dict]] = None, *, - lake: resources.Lake = None, - update_mask: field_mask_pb2.FieldMask = None, + lake: Optional[resources.Lake] = None, + update_mask: Optional[field_mask_pb2.FieldMask] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> operation.Operation: r"""Updates a lake resource. @@ -912,11 +924,11 @@ def sample_update_lake(): def delete_lake( self, - request: Union[service.DeleteLakeRequest, dict] = None, + request: Optional[Union[service.DeleteLakeRequest, dict]] = None, *, - name: str = None, + name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> operation.Operation: r"""Deletes a lake resource. All zones within the lake @@ -1036,11 +1048,11 @@ def sample_delete_lake(): def list_lakes( self, - request: Union[service.ListLakesRequest, dict] = None, + request: Optional[Union[service.ListLakesRequest, dict]] = None, *, - parent: str = None, + parent: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.ListLakesPager: r"""Lists lake resources in a project and location. @@ -1151,11 +1163,11 @@ def sample_list_lakes(): def get_lake( self, - request: Union[service.GetLakeRequest, dict] = None, + request: Optional[Union[service.GetLakeRequest, dict]] = None, *, - name: str = None, + name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> resources.Lake: r"""Retrieves a lake resource. @@ -1266,11 +1278,11 @@ def sample_get_lake(): def list_lake_actions( self, - request: Union[service.ListLakeActionsRequest, dict] = None, + request: Optional[Union[service.ListLakeActionsRequest, dict]] = None, *, - parent: str = None, + parent: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.ListLakeActionsPager: r"""Lists action resources in a lake. @@ -1379,13 +1391,13 @@ def sample_list_lake_actions(): def create_zone( self, - request: Union[service.CreateZoneRequest, dict] = None, + request: Optional[Union[service.CreateZoneRequest, dict]] = None, *, - parent: str = None, - zone: resources.Zone = None, - zone_id: str = None, + parent: Optional[str] = None, + zone: Optional[resources.Zone] = None, + zone_id: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> operation.Operation: r"""Creates a zone resource within a lake. @@ -1532,12 +1544,12 @@ def sample_create_zone(): def update_zone( self, - request: Union[service.UpdateZoneRequest, dict] = None, + request: Optional[Union[service.UpdateZoneRequest, dict]] = None, *, - zone: resources.Zone = None, - update_mask: field_mask_pb2.FieldMask = None, + zone: Optional[resources.Zone] = None, + update_mask: Optional[field_mask_pb2.FieldMask] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> operation.Operation: r"""Updates a zone resource. @@ -1664,11 +1676,11 @@ def sample_update_zone(): def delete_zone( self, - request: Union[service.DeleteZoneRequest, dict] = None, + request: Optional[Union[service.DeleteZoneRequest, dict]] = None, *, - name: str = None, + name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> operation.Operation: r"""Deletes a zone resource. All assets within a zone @@ -1788,11 +1800,11 @@ def sample_delete_zone(): def list_zones( self, - request: Union[service.ListZonesRequest, dict] = None, + request: Optional[Union[service.ListZonesRequest, dict]] = None, *, - parent: str = None, + parent: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.ListZonesPager: r"""Lists zone resources in a lake. @@ -1901,11 +1913,11 @@ def sample_list_zones(): def get_zone( self, - request: Union[service.GetZoneRequest, dict] = None, + request: Optional[Union[service.GetZoneRequest, dict]] = None, *, - name: str = None, + name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> resources.Zone: r"""Retrieves a zone resource. @@ -2008,11 +2020,11 @@ def sample_get_zone(): def list_zone_actions( self, - request: Union[service.ListZoneActionsRequest, dict] = None, + request: Optional[Union[service.ListZoneActionsRequest, dict]] = None, *, - parent: str = None, + parent: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.ListZoneActionsPager: r"""Lists action resources in a zone. @@ -2121,13 +2133,13 @@ def sample_list_zone_actions(): def create_asset( self, - request: Union[service.CreateAssetRequest, dict] = None, + request: Optional[Union[service.CreateAssetRequest, dict]] = None, *, - parent: str = None, - asset: resources.Asset = None, - asset_id: str = None, + parent: Optional[str] = None, + asset: Optional[resources.Asset] = None, + asset_id: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> operation.Operation: r"""Creates an asset resource. @@ -2267,12 +2279,12 @@ def sample_create_asset(): def update_asset( self, - request: Union[service.UpdateAssetRequest, dict] = None, + request: Optional[Union[service.UpdateAssetRequest, dict]] = None, *, - asset: resources.Asset = None, - update_mask: field_mask_pb2.FieldMask = None, + asset: Optional[resources.Asset] = None, + update_mask: Optional[field_mask_pb2.FieldMask] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> operation.Operation: r"""Updates an asset resource. @@ -2395,11 +2407,11 @@ def sample_update_asset(): def delete_asset( self, - request: Union[service.DeleteAssetRequest, dict] = None, + request: Optional[Union[service.DeleteAssetRequest, dict]] = None, *, - name: str = None, + name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> operation.Operation: r"""Deletes an asset resource. The referenced storage @@ -2520,11 +2532,11 @@ def sample_delete_asset(): def list_assets( self, - request: Union[service.ListAssetsRequest, dict] = None, + request: Optional[Union[service.ListAssetsRequest, dict]] = None, *, - parent: str = None, + parent: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.ListAssetsPager: r"""Lists asset resources in a zone. @@ -2633,11 +2645,11 @@ def sample_list_assets(): def get_asset( self, - request: Union[service.GetAssetRequest, dict] = None, + request: Optional[Union[service.GetAssetRequest, dict]] = None, *, - name: str = None, + name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> resources.Asset: r"""Retrieves an asset resource. @@ -2735,11 +2747,11 @@ def sample_get_asset(): def list_asset_actions( self, - request: Union[service.ListAssetActionsRequest, dict] = None, + request: Optional[Union[service.ListAssetActionsRequest, dict]] = None, *, - parent: str = None, + parent: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.ListAssetActionsPager: r"""Lists action resources in an asset. @@ -2848,13 +2860,13 @@ def sample_list_asset_actions(): def create_task( self, - request: Union[service.CreateTaskRequest, dict] = None, + request: Optional[Union[service.CreateTaskRequest, dict]] = None, *, - parent: str = None, - task: tasks.Task = None, - task_id: str = None, + parent: Optional[str] = None, + task: Optional[tasks.Task] = None, + task_id: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> operation.Operation: r"""Creates a task resource within a lake. @@ -2988,12 +3000,12 @@ def sample_create_task(): def update_task( self, - request: Union[service.UpdateTaskRequest, dict] = None, + request: Optional[Union[service.UpdateTaskRequest, dict]] = None, *, - task: tasks.Task = None, - update_mask: field_mask_pb2.FieldMask = None, + task: Optional[tasks.Task] = None, + update_mask: Optional[field_mask_pb2.FieldMask] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> operation.Operation: r"""Update the task resource. @@ -3120,11 +3132,11 @@ def sample_update_task(): def delete_task( self, - request: Union[service.DeleteTaskRequest, dict] = None, + request: Optional[Union[service.DeleteTaskRequest, dict]] = None, *, - name: str = None, + name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> operation.Operation: r"""Delete the task resource. @@ -3243,11 +3255,11 @@ def sample_delete_task(): def list_tasks( self, - request: Union[service.ListTasksRequest, dict] = None, + request: Optional[Union[service.ListTasksRequest, dict]] = None, *, - parent: str = None, + parent: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.ListTasksPager: r"""Lists tasks under the given lake. @@ -3356,11 +3368,11 @@ def sample_list_tasks(): def get_task( self, - request: Union[service.GetTaskRequest, dict] = None, + request: Optional[Union[service.GetTaskRequest, dict]] = None, *, - name: str = None, + name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> tasks.Task: r"""Get task resource. @@ -3455,11 +3467,11 @@ def sample_get_task(): def list_jobs( self, - request: Union[service.ListJobsRequest, dict] = None, + request: Optional[Union[service.ListJobsRequest, dict]] = None, *, - parent: str = None, + parent: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.ListJobsPager: r"""Lists Jobs under the given task. @@ -3568,11 +3580,11 @@ def sample_list_jobs(): def get_job( self, - request: Union[service.GetJobRequest, dict] = None, + request: Optional[Union[service.GetJobRequest, dict]] = None, *, - name: str = None, + name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> tasks.Job: r"""Get job resource. @@ -3669,11 +3681,11 @@ def sample_get_job(): def cancel_job( self, - request: Union[service.CancelJobRequest, dict] = None, + request: Optional[Union[service.CancelJobRequest, dict]] = None, *, - name: str = None, + name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> None: r"""Cancel jobs running for the task resource. @@ -3758,13 +3770,13 @@ def sample_cancel_job(): def create_environment( self, - request: Union[service.CreateEnvironmentRequest, dict] = None, + request: Optional[Union[service.CreateEnvironmentRequest, dict]] = None, *, - parent: str = None, - environment: analyze.Environment = None, - environment_id: str = None, + parent: Optional[str] = None, + environment: Optional[analyze.Environment] = None, + environment_id: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> operation.Operation: r"""Create an environment resource. @@ -3902,12 +3914,12 @@ def sample_create_environment(): def update_environment( self, - request: Union[service.UpdateEnvironmentRequest, dict] = None, + request: Optional[Union[service.UpdateEnvironmentRequest, dict]] = None, *, - environment: analyze.Environment = None, - update_mask: field_mask_pb2.FieldMask = None, + environment: Optional[analyze.Environment] = None, + update_mask: Optional[field_mask_pb2.FieldMask] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> operation.Operation: r"""Update the environment resource. @@ -4030,11 +4042,11 @@ def sample_update_environment(): def delete_environment( self, - request: Union[service.DeleteEnvironmentRequest, dict] = None, + request: Optional[Union[service.DeleteEnvironmentRequest, dict]] = None, *, - name: str = None, + name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> operation.Operation: r"""Delete the environment resource. All the child @@ -4155,11 +4167,11 @@ def sample_delete_environment(): def list_environments( self, - request: Union[service.ListEnvironmentsRequest, dict] = None, + request: Optional[Union[service.ListEnvironmentsRequest, dict]] = None, *, - parent: str = None, + parent: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.ListEnvironmentsPager: r"""Lists environments under the given lake. @@ -4268,11 +4280,11 @@ def sample_list_environments(): def get_environment( self, - request: Union[service.GetEnvironmentRequest, dict] = None, + request: Optional[Union[service.GetEnvironmentRequest, dict]] = None, *, - name: str = None, + name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> analyze.Environment: r"""Get environment resource. @@ -4370,11 +4382,11 @@ def sample_get_environment(): def list_sessions( self, - request: Union[service.ListSessionsRequest, dict] = None, + request: Optional[Union[service.ListSessionsRequest, dict]] = None, *, - parent: str = None, + parent: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.ListSessionsPager: r"""Lists session resources in an environment. @@ -4496,10 +4508,10 @@ def __exit__(self, type, value, traceback): def list_operations( self, - request: operations_pb2.ListOperationsRequest = None, + request: Optional[operations_pb2.ListOperationsRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> operations_pb2.ListOperationsResponse: r"""Lists operations that match the specified filter in the request. @@ -4550,10 +4562,10 @@ def list_operations( def get_operation( self, - request: operations_pb2.GetOperationRequest = None, + request: Optional[operations_pb2.GetOperationRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> operations_pb2.Operation: r"""Gets the latest state of a long-running operation. @@ -4604,10 +4616,10 @@ def get_operation( def delete_operation( self, - request: operations_pb2.DeleteOperationRequest = None, + request: Optional[operations_pb2.DeleteOperationRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> None: r"""Deletes a long-running operation. @@ -4659,10 +4671,10 @@ def delete_operation( def cancel_operation( self, - request: operations_pb2.CancelOperationRequest = None, + request: Optional[operations_pb2.CancelOperationRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> None: r"""Starts asynchronous cancellation on a long-running operation. @@ -4713,10 +4725,10 @@ def cancel_operation( def get_location( self, - request: locations_pb2.GetLocationRequest = None, + request: Optional[locations_pb2.GetLocationRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> locations_pb2.Location: r"""Gets information about a location. @@ -4767,10 +4779,10 @@ def get_location( def list_locations( self, - request: locations_pb2.ListLocationsRequest = None, + request: Optional[locations_pb2.ListLocationsRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> locations_pb2.ListLocationsResponse: r"""Lists information about the supported locations for this service. diff --git a/google/cloud/dataplex_v1/services/dataplex_service/transports/base.py b/google/cloud/dataplex_v1/services/dataplex_service/transports/base.py index 0100bd6..525bfc9 100644 --- a/google/cloud/dataplex_v1/services/dataplex_service/transports/base.py +++ b/google/cloud/dataplex_v1/services/dataplex_service/transports/base.py @@ -53,7 +53,7 @@ def __init__( self, *, host: str = DEFAULT_HOST, - credentials: ga_credentials.Credentials = None, + credentials: Optional[ga_credentials.Credentials] = None, credentials_file: Optional[str] = None, scopes: Optional[Sequence[str]] = None, quota_project_id: Optional[str] = None, diff --git a/google/cloud/dataplex_v1/services/dataplex_service/transports/grpc.py b/google/cloud/dataplex_v1/services/dataplex_service/transports/grpc.py index d46563e..dcbb77c 100644 --- a/google/cloud/dataplex_v1/services/dataplex_service/transports/grpc.py +++ b/google/cloud/dataplex_v1/services/dataplex_service/transports/grpc.py @@ -56,14 +56,14 @@ def __init__( self, *, host: str = "dataplex.googleapis.com", - credentials: ga_credentials.Credentials = None, - credentials_file: str = None, - scopes: Sequence[str] = None, - channel: grpc.Channel = None, - api_mtls_endpoint: str = None, - client_cert_source: Callable[[], Tuple[bytes, bytes]] = None, - ssl_channel_credentials: grpc.ChannelCredentials = None, - client_cert_source_for_mtls: Callable[[], Tuple[bytes, bytes]] = None, + credentials: Optional[ga_credentials.Credentials] = None, + credentials_file: Optional[str] = None, + scopes: Optional[Sequence[str]] = None, + channel: Optional[grpc.Channel] = None, + api_mtls_endpoint: Optional[str] = None, + client_cert_source: Optional[Callable[[], Tuple[bytes, bytes]]] = None, + ssl_channel_credentials: Optional[grpc.ChannelCredentials] = None, + client_cert_source_for_mtls: Optional[Callable[[], Tuple[bytes, bytes]]] = None, quota_project_id: Optional[str] = None, client_info: gapic_v1.client_info.ClientInfo = DEFAULT_CLIENT_INFO, always_use_jwt_access: Optional[bool] = False, @@ -191,8 +191,8 @@ def __init__( def create_channel( cls, host: str = "dataplex.googleapis.com", - credentials: ga_credentials.Credentials = None, - credentials_file: str = None, + credentials: Optional[ga_credentials.Credentials] = None, + credentials_file: Optional[str] = None, scopes: Optional[Sequence[str]] = None, quota_project_id: Optional[str] = None, **kwargs, diff --git a/google/cloud/dataplex_v1/services/dataplex_service/transports/grpc_asyncio.py b/google/cloud/dataplex_v1/services/dataplex_service/transports/grpc_asyncio.py index 654572e..b5114bb 100644 --- a/google/cloud/dataplex_v1/services/dataplex_service/transports/grpc_asyncio.py +++ b/google/cloud/dataplex_v1/services/dataplex_service/transports/grpc_asyncio.py @@ -58,7 +58,7 @@ class DataplexServiceGrpcAsyncIOTransport(DataplexServiceTransport): def create_channel( cls, host: str = "dataplex.googleapis.com", - credentials: ga_credentials.Credentials = None, + credentials: Optional[ga_credentials.Credentials] = None, credentials_file: Optional[str] = None, scopes: Optional[Sequence[str]] = None, quota_project_id: Optional[str] = None, @@ -101,15 +101,15 @@ def __init__( self, *, host: str = "dataplex.googleapis.com", - credentials: ga_credentials.Credentials = None, + credentials: Optional[ga_credentials.Credentials] = None, credentials_file: Optional[str] = None, scopes: Optional[Sequence[str]] = None, - channel: aio.Channel = None, - api_mtls_endpoint: str = None, - client_cert_source: Callable[[], Tuple[bytes, bytes]] = None, - ssl_channel_credentials: grpc.ChannelCredentials = None, - client_cert_source_for_mtls: Callable[[], Tuple[bytes, bytes]] = None, - quota_project_id=None, + channel: Optional[aio.Channel] = None, + api_mtls_endpoint: Optional[str] = None, + client_cert_source: Optional[Callable[[], Tuple[bytes, bytes]]] = None, + ssl_channel_credentials: Optional[grpc.ChannelCredentials] = None, + client_cert_source_for_mtls: Optional[Callable[[], Tuple[bytes, bytes]]] = None, + quota_project_id: Optional[str] = None, client_info: gapic_v1.client_info.ClientInfo = DEFAULT_CLIENT_INFO, always_use_jwt_access: Optional[bool] = False, api_audience: Optional[str] = None, diff --git a/google/cloud/dataplex_v1/services/metadata_service/async_client.py b/google/cloud/dataplex_v1/services/metadata_service/async_client.py index 9446701..e6a7b23 100644 --- a/google/cloud/dataplex_v1/services/metadata_service/async_client.py +++ b/google/cloud/dataplex_v1/services/metadata_service/async_client.py @@ -16,7 +16,17 @@ from collections import OrderedDict import functools import re -from typing import Dict, Mapping, Optional, Sequence, Tuple, Type, Union +from typing import ( + Dict, + Mapping, + MutableMapping, + MutableSequence, + Optional, + Sequence, + Tuple, + Type, + Union, +) from google.api_core import exceptions as core_exceptions from google.api_core import gapic_v1 @@ -171,9 +181,9 @@ def transport(self) -> MetadataServiceTransport: def __init__( self, *, - credentials: ga_credentials.Credentials = None, + credentials: Optional[ga_credentials.Credentials] = None, transport: Union[str, MetadataServiceTransport] = "grpc_asyncio", - client_options: ClientOptions = None, + client_options: Optional[ClientOptions] = None, client_info: gapic_v1.client_info.ClientInfo = DEFAULT_CLIENT_INFO, ) -> None: """Instantiates the metadata service client. @@ -217,12 +227,12 @@ def __init__( async def create_entity( self, - request: Union[metadata_.CreateEntityRequest, dict] = None, + request: Optional[Union[metadata_.CreateEntityRequest, dict]] = None, *, - parent: str = None, - entity: metadata_.Entity = None, + parent: Optional[str] = None, + entity: Optional[metadata_.Entity] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> metadata_.Entity: r"""Create a metadata entity. @@ -264,7 +274,7 @@ async def sample_create_entity(): print(response) Args: - request (Union[google.cloud.dataplex_v1.types.CreateEntityRequest, dict]): + request (Optional[Union[google.cloud.dataplex_v1.types.CreateEntityRequest, dict]]): The request object. Create a metadata entity request. parent (:class:`str`): Required. The resource name of the parent zone: @@ -336,10 +346,10 @@ async def sample_create_entity(): async def update_entity( self, - request: Union[metadata_.UpdateEntityRequest, dict] = None, + request: Optional[Union[metadata_.UpdateEntityRequest, dict]] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> metadata_.Entity: r"""Update a metadata entity. Only supports full resource @@ -381,7 +391,7 @@ async def sample_update_entity(): print(response) Args: - request (Union[google.cloud.dataplex_v1.types.UpdateEntityRequest, dict]): + request (Optional[Union[google.cloud.dataplex_v1.types.UpdateEntityRequest, dict]]): The request object. Update a metadata entity request. The exiting entity will be fully replaced by the entity in the request. The entity ID is mutable. To modify the @@ -431,11 +441,11 @@ async def sample_update_entity(): async def delete_entity( self, - request: Union[metadata_.DeleteEntityRequest, dict] = None, + request: Optional[Union[metadata_.DeleteEntityRequest, dict]] = None, *, - name: str = None, + name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> None: r"""Delete a metadata entity. @@ -465,7 +475,7 @@ async def sample_delete_entity(): await client.delete_entity(request=request) Args: - request (Union[google.cloud.dataplex_v1.types.DeleteEntityRequest, dict]): + request (Optional[Union[google.cloud.dataplex_v1.types.DeleteEntityRequest, dict]]): The request object. Delete a metadata entity request. name (:class:`str`): Required. The resource name of the entity: @@ -521,11 +531,11 @@ async def sample_delete_entity(): async def get_entity( self, - request: Union[metadata_.GetEntityRequest, dict] = None, + request: Optional[Union[metadata_.GetEntityRequest, dict]] = None, *, - name: str = None, + name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> metadata_.Entity: r"""Get a metadata entity. @@ -557,7 +567,7 @@ async def sample_get_entity(): print(response) Args: - request (Union[google.cloud.dataplex_v1.types.GetEntityRequest, dict]): + request (Optional[Union[google.cloud.dataplex_v1.types.GetEntityRequest, dict]]): The request object. Get metadata entity request. name (:class:`str`): Required. The resource name of the entity: @@ -631,11 +641,11 @@ async def sample_get_entity(): async def list_entities( self, - request: Union[metadata_.ListEntitiesRequest, dict] = None, + request: Optional[Union[metadata_.ListEntitiesRequest, dict]] = None, *, - parent: str = None, + parent: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.ListEntitiesAsyncPager: r"""List metadata entities in a zone. @@ -669,7 +679,7 @@ async def sample_list_entities(): print(response) Args: - request (Union[google.cloud.dataplex_v1.types.ListEntitiesRequest, dict]): + request (Optional[Union[google.cloud.dataplex_v1.types.ListEntitiesRequest, dict]]): The request object. List metadata entities request. parent (:class:`str`): Required. The resource name of the parent zone: @@ -754,12 +764,12 @@ async def sample_list_entities(): async def create_partition( self, - request: Union[metadata_.CreatePartitionRequest, dict] = None, + request: Optional[Union[metadata_.CreatePartitionRequest, dict]] = None, *, - parent: str = None, - partition: metadata_.Partition = None, + parent: Optional[str] = None, + partition: Optional[metadata_.Partition] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> metadata_.Partition: r"""Create a metadata partition. @@ -796,7 +806,7 @@ async def sample_create_partition(): print(response) Args: - request (Union[google.cloud.dataplex_v1.types.CreatePartitionRequest, dict]): + request (Optional[Union[google.cloud.dataplex_v1.types.CreatePartitionRequest, dict]]): The request object. Create metadata partition request. parent (:class:`str`): Required. The resource name of the parent zone: @@ -868,11 +878,11 @@ async def sample_create_partition(): async def delete_partition( self, - request: Union[metadata_.DeletePartitionRequest, dict] = None, + request: Optional[Union[metadata_.DeletePartitionRequest, dict]] = None, *, - name: str = None, + name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> None: r"""Delete a metadata partition. @@ -901,7 +911,7 @@ async def sample_delete_partition(): await client.delete_partition(request=request) Args: - request (Union[google.cloud.dataplex_v1.types.DeletePartitionRequest, dict]): + request (Optional[Union[google.cloud.dataplex_v1.types.DeletePartitionRequest, dict]]): The request object. Delete metadata partition request. name (:class:`str`): Required. The resource name of the partition. format: @@ -960,11 +970,11 @@ async def sample_delete_partition(): async def get_partition( self, - request: Union[metadata_.GetPartitionRequest, dict] = None, + request: Optional[Union[metadata_.GetPartitionRequest, dict]] = None, *, - name: str = None, + name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> metadata_.Partition: r"""Get a metadata partition of an entity. @@ -996,7 +1006,7 @@ async def sample_get_partition(): print(response) Args: - request (Union[google.cloud.dataplex_v1.types.GetPartitionRequest, dict]): + request (Optional[Union[google.cloud.dataplex_v1.types.GetPartitionRequest, dict]]): The request object. Get metadata partition request. name (:class:`str`): Required. The resource name of the partition: @@ -1073,11 +1083,11 @@ async def sample_get_partition(): async def list_partitions( self, - request: Union[metadata_.ListPartitionsRequest, dict] = None, + request: Optional[Union[metadata_.ListPartitionsRequest, dict]] = None, *, - parent: str = None, + parent: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.ListPartitionsAsyncPager: r"""List metadata partitions of an entity. @@ -1110,7 +1120,7 @@ async def sample_list_partitions(): print(response) Args: - request (Union[google.cloud.dataplex_v1.types.ListPartitionsRequest, dict]): + request (Optional[Union[google.cloud.dataplex_v1.types.ListPartitionsRequest, dict]]): The request object. List metadata partitions request. parent (:class:`str`): Required. The resource name of the parent entity: @@ -1195,10 +1205,10 @@ async def sample_list_partitions(): async def list_operations( self, - request: operations_pb2.ListOperationsRequest = None, + request: Optional[operations_pb2.ListOperationsRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> operations_pb2.ListOperationsResponse: r"""Lists operations that match the specified filter in the request. @@ -1249,10 +1259,10 @@ async def list_operations( async def get_operation( self, - request: operations_pb2.GetOperationRequest = None, + request: Optional[operations_pb2.GetOperationRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> operations_pb2.Operation: r"""Gets the latest state of a long-running operation. @@ -1303,10 +1313,10 @@ async def get_operation( async def delete_operation( self, - request: operations_pb2.DeleteOperationRequest = None, + request: Optional[operations_pb2.DeleteOperationRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> None: r"""Deletes a long-running operation. @@ -1358,10 +1368,10 @@ async def delete_operation( async def cancel_operation( self, - request: operations_pb2.CancelOperationRequest = None, + request: Optional[operations_pb2.CancelOperationRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> None: r"""Starts asynchronous cancellation on a long-running operation. @@ -1412,10 +1422,10 @@ async def cancel_operation( async def get_location( self, - request: locations_pb2.GetLocationRequest = None, + request: Optional[locations_pb2.GetLocationRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> locations_pb2.Location: r"""Gets information about a location. @@ -1466,10 +1476,10 @@ async def get_location( async def list_locations( self, - request: locations_pb2.ListLocationsRequest = None, + request: Optional[locations_pb2.ListLocationsRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> locations_pb2.ListLocationsResponse: r"""Lists information about the supported locations for this service. diff --git a/google/cloud/dataplex_v1/services/metadata_service/client.py b/google/cloud/dataplex_v1/services/metadata_service/client.py index e1fce01..4438af0 100644 --- a/google/cloud/dataplex_v1/services/metadata_service/client.py +++ b/google/cloud/dataplex_v1/services/metadata_service/client.py @@ -16,7 +16,18 @@ from collections import OrderedDict import os import re -from typing import Dict, Mapping, Optional, Sequence, Tuple, Type, Union +from typing import ( + Dict, + Mapping, + MutableMapping, + MutableSequence, + Optional, + Sequence, + Tuple, + Type, + Union, + cast, +) from google.api_core import client_options as client_options_lib from google.api_core import exceptions as core_exceptions @@ -64,7 +75,7 @@ class MetadataServiceClientMeta(type): def get_transport_class( cls, - label: str = None, + label: Optional[str] = None, ) -> Type[MetadataServiceTransport]: """Returns an appropriate transport class. @@ -399,8 +410,8 @@ def __init__( self, *, credentials: Optional[ga_credentials.Credentials] = None, - transport: Union[str, MetadataServiceTransport, None] = None, - client_options: Optional[client_options_lib.ClientOptions] = None, + transport: Optional[Union[str, MetadataServiceTransport]] = None, + client_options: Optional[Union[client_options_lib.ClientOptions, dict]] = None, client_info: gapic_v1.client_info.ClientInfo = DEFAULT_CLIENT_INFO, ) -> None: """Instantiates the metadata service client. @@ -414,7 +425,7 @@ def __init__( transport (Union[str, MetadataServiceTransport]): The transport to use. If set to None, a transport is chosen automatically. - client_options (google.api_core.client_options.ClientOptions): Custom options for the + client_options (Optional[Union[google.api_core.client_options.ClientOptions, dict]]): Custom options for the client. It won't take effect if a ``transport`` instance is provided. (1) The ``api_endpoint`` property can be used to override the default endpoint provided by the client. GOOGLE_API_USE_MTLS_ENDPOINT @@ -444,6 +455,7 @@ def __init__( client_options = client_options_lib.from_dict(client_options) if client_options is None: client_options = client_options_lib.ClientOptions() + client_options = cast(client_options_lib.ClientOptions, client_options) api_endpoint, client_cert_source_func = self.get_mtls_endpoint_and_cert_source( client_options @@ -496,12 +508,12 @@ def __init__( def create_entity( self, - request: Union[metadata_.CreateEntityRequest, dict] = None, + request: Optional[Union[metadata_.CreateEntityRequest, dict]] = None, *, - parent: str = None, - entity: metadata_.Entity = None, + parent: Optional[str] = None, + entity: Optional[metadata_.Entity] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> metadata_.Entity: r"""Create a metadata entity. @@ -615,10 +627,10 @@ def sample_create_entity(): def update_entity( self, - request: Union[metadata_.UpdateEntityRequest, dict] = None, + request: Optional[Union[metadata_.UpdateEntityRequest, dict]] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> metadata_.Entity: r"""Update a metadata entity. Only supports full resource @@ -711,11 +723,11 @@ def sample_update_entity(): def delete_entity( self, - request: Union[metadata_.DeleteEntityRequest, dict] = None, + request: Optional[Union[metadata_.DeleteEntityRequest, dict]] = None, *, - name: str = None, + name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> None: r"""Delete a metadata entity. @@ -801,11 +813,11 @@ def sample_delete_entity(): def get_entity( self, - request: Union[metadata_.GetEntityRequest, dict] = None, + request: Optional[Union[metadata_.GetEntityRequest, dict]] = None, *, - name: str = None, + name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> metadata_.Entity: r"""Get a metadata entity. @@ -902,11 +914,11 @@ def sample_get_entity(): def list_entities( self, - request: Union[metadata_.ListEntitiesRequest, dict] = None, + request: Optional[Union[metadata_.ListEntitiesRequest, dict]] = None, *, - parent: str = None, + parent: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.ListEntitiesPager: r"""List metadata entities in a zone. @@ -1016,12 +1028,12 @@ def sample_list_entities(): def create_partition( self, - request: Union[metadata_.CreatePartitionRequest, dict] = None, + request: Optional[Union[metadata_.CreatePartitionRequest, dict]] = None, *, - parent: str = None, - partition: metadata_.Partition = None, + parent: Optional[str] = None, + partition: Optional[metadata_.Partition] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> metadata_.Partition: r"""Create a metadata partition. @@ -1130,11 +1142,11 @@ def sample_create_partition(): def delete_partition( self, - request: Union[metadata_.DeletePartitionRequest, dict] = None, + request: Optional[Union[metadata_.DeletePartitionRequest, dict]] = None, *, - name: str = None, + name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> None: r"""Delete a metadata partition. @@ -1222,11 +1234,11 @@ def sample_delete_partition(): def get_partition( self, - request: Union[metadata_.GetPartitionRequest, dict] = None, + request: Optional[Union[metadata_.GetPartitionRequest, dict]] = None, *, - name: str = None, + name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> metadata_.Partition: r"""Get a metadata partition of an entity. @@ -1326,11 +1338,11 @@ def sample_get_partition(): def list_partitions( self, - request: Union[metadata_.ListPartitionsRequest, dict] = None, + request: Optional[Union[metadata_.ListPartitionsRequest, dict]] = None, *, - parent: str = None, + parent: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.ListPartitionsPager: r"""List metadata partitions of an entity. @@ -1452,10 +1464,10 @@ def __exit__(self, type, value, traceback): def list_operations( self, - request: operations_pb2.ListOperationsRequest = None, + request: Optional[operations_pb2.ListOperationsRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> operations_pb2.ListOperationsResponse: r"""Lists operations that match the specified filter in the request. @@ -1506,10 +1518,10 @@ def list_operations( def get_operation( self, - request: operations_pb2.GetOperationRequest = None, + request: Optional[operations_pb2.GetOperationRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> operations_pb2.Operation: r"""Gets the latest state of a long-running operation. @@ -1560,10 +1572,10 @@ def get_operation( def delete_operation( self, - request: operations_pb2.DeleteOperationRequest = None, + request: Optional[operations_pb2.DeleteOperationRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> None: r"""Deletes a long-running operation. @@ -1615,10 +1627,10 @@ def delete_operation( def cancel_operation( self, - request: operations_pb2.CancelOperationRequest = None, + request: Optional[operations_pb2.CancelOperationRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> None: r"""Starts asynchronous cancellation on a long-running operation. @@ -1669,10 +1681,10 @@ def cancel_operation( def get_location( self, - request: locations_pb2.GetLocationRequest = None, + request: Optional[locations_pb2.GetLocationRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> locations_pb2.Location: r"""Gets information about a location. @@ -1723,10 +1735,10 @@ def get_location( def list_locations( self, - request: locations_pb2.ListLocationsRequest = None, + request: Optional[locations_pb2.ListLocationsRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> locations_pb2.ListLocationsResponse: r"""Lists information about the supported locations for this service. diff --git a/google/cloud/dataplex_v1/services/metadata_service/transports/base.py b/google/cloud/dataplex_v1/services/metadata_service/transports/base.py index ebbb658..8d10f83 100644 --- a/google/cloud/dataplex_v1/services/metadata_service/transports/base.py +++ b/google/cloud/dataplex_v1/services/metadata_service/transports/base.py @@ -53,7 +53,7 @@ def __init__( self, *, host: str = DEFAULT_HOST, - credentials: ga_credentials.Credentials = None, + credentials: Optional[ga_credentials.Credentials] = None, credentials_file: Optional[str] = None, scopes: Optional[Sequence[str]] = None, quota_project_id: Optional[str] = None, diff --git a/google/cloud/dataplex_v1/services/metadata_service/transports/grpc.py b/google/cloud/dataplex_v1/services/metadata_service/transports/grpc.py index 91348e4..03284d6 100644 --- a/google/cloud/dataplex_v1/services/metadata_service/transports/grpc.py +++ b/google/cloud/dataplex_v1/services/metadata_service/transports/grpc.py @@ -52,14 +52,14 @@ def __init__( self, *, host: str = "dataplex.googleapis.com", - credentials: ga_credentials.Credentials = None, - credentials_file: str = None, - scopes: Sequence[str] = None, - channel: grpc.Channel = None, - api_mtls_endpoint: str = None, - client_cert_source: Callable[[], Tuple[bytes, bytes]] = None, - ssl_channel_credentials: grpc.ChannelCredentials = None, - client_cert_source_for_mtls: Callable[[], Tuple[bytes, bytes]] = None, + credentials: Optional[ga_credentials.Credentials] = None, + credentials_file: Optional[str] = None, + scopes: Optional[Sequence[str]] = None, + channel: Optional[grpc.Channel] = None, + api_mtls_endpoint: Optional[str] = None, + client_cert_source: Optional[Callable[[], Tuple[bytes, bytes]]] = None, + ssl_channel_credentials: Optional[grpc.ChannelCredentials] = None, + client_cert_source_for_mtls: Optional[Callable[[], Tuple[bytes, bytes]]] = None, quota_project_id: Optional[str] = None, client_info: gapic_v1.client_info.ClientInfo = DEFAULT_CLIENT_INFO, always_use_jwt_access: Optional[bool] = False, @@ -186,8 +186,8 @@ def __init__( def create_channel( cls, host: str = "dataplex.googleapis.com", - credentials: ga_credentials.Credentials = None, - credentials_file: str = None, + credentials: Optional[ga_credentials.Credentials] = None, + credentials_file: Optional[str] = None, scopes: Optional[Sequence[str]] = None, quota_project_id: Optional[str] = None, **kwargs, diff --git a/google/cloud/dataplex_v1/services/metadata_service/transports/grpc_asyncio.py b/google/cloud/dataplex_v1/services/metadata_service/transports/grpc_asyncio.py index 6ab2609..a8539d1 100644 --- a/google/cloud/dataplex_v1/services/metadata_service/transports/grpc_asyncio.py +++ b/google/cloud/dataplex_v1/services/metadata_service/transports/grpc_asyncio.py @@ -54,7 +54,7 @@ class MetadataServiceGrpcAsyncIOTransport(MetadataServiceTransport): def create_channel( cls, host: str = "dataplex.googleapis.com", - credentials: ga_credentials.Credentials = None, + credentials: Optional[ga_credentials.Credentials] = None, credentials_file: Optional[str] = None, scopes: Optional[Sequence[str]] = None, quota_project_id: Optional[str] = None, @@ -97,15 +97,15 @@ def __init__( self, *, host: str = "dataplex.googleapis.com", - credentials: ga_credentials.Credentials = None, + credentials: Optional[ga_credentials.Credentials] = None, credentials_file: Optional[str] = None, scopes: Optional[Sequence[str]] = None, - channel: aio.Channel = None, - api_mtls_endpoint: str = None, - client_cert_source: Callable[[], Tuple[bytes, bytes]] = None, - ssl_channel_credentials: grpc.ChannelCredentials = None, - client_cert_source_for_mtls: Callable[[], Tuple[bytes, bytes]] = None, - quota_project_id=None, + channel: Optional[aio.Channel] = None, + api_mtls_endpoint: Optional[str] = None, + client_cert_source: Optional[Callable[[], Tuple[bytes, bytes]]] = None, + ssl_channel_credentials: Optional[grpc.ChannelCredentials] = None, + client_cert_source_for_mtls: Optional[Callable[[], Tuple[bytes, bytes]]] = None, + quota_project_id: Optional[str] = None, client_info: gapic_v1.client_info.ClientInfo = DEFAULT_CLIENT_INFO, always_use_jwt_access: Optional[bool] = False, api_audience: Optional[str] = None, diff --git a/google/cloud/dataplex_v1/types/analyze.py b/google/cloud/dataplex_v1/types/analyze.py index e676ed0..798e670 100644 --- a/google/cloud/dataplex_v1/types/analyze.py +++ b/google/cloud/dataplex_v1/types/analyze.py @@ -13,6 +13,8 @@ # See the License for the specific language governing permissions and # limitations under the License. # +from typing import MutableMapping, MutableSequence + from google.protobuf import duration_pb2 # type: ignore from google.protobuf import timestamp_pb2 # type: ignore import proto # type: ignore @@ -50,7 +52,7 @@ class Environment(proto.Message): update_time (google.protobuf.timestamp_pb2.Timestamp): Output only. The time when the environment was last updated. - labels (Mapping[str, str]): + labels (MutableMapping[str, str]): Optional. User defined labels for the environment. description (str): @@ -108,15 +110,15 @@ class ComputeResources(proto.Message): node_count, then auto-scaling is enabled. """ - disk_size_gb = proto.Field( + disk_size_gb: int = proto.Field( proto.INT32, number=1, ) - node_count = proto.Field( + node_count: int = proto.Field( proto.INT32, number=2, ) - max_node_count = proto.Field( + max_node_count: int = proto.Field( proto.INT32, number=3, ) @@ -127,48 +129,48 @@ class OsImageRuntime(proto.Message): Attributes: image_version (str): Required. Dataplex Image version. - java_libraries (Sequence[str]): + java_libraries (MutableSequence[str]): Optional. List of Java jars to be included in the runtime environment. Valid input includes Cloud Storage URIs to Jar binaries. For example, gs://bucket-name/my/path/to/file.jar - python_packages (Sequence[str]): + python_packages (MutableSequence[str]): Optional. A list of python packages to be installed. Valid formats include Cloud Storage URI to a PIP installable library. For example, gs://bucket-name/my/path/to/lib.tar.gz - properties (Mapping[str, str]): + properties (MutableMapping[str, str]): Optional. Spark properties to provide configuration for use in sessions created for this environment. The properties to set on daemon config files. Property keys are specified in ``prefix:property`` format. The prefix must be "spark". """ - image_version = proto.Field( + image_version: str = proto.Field( proto.STRING, number=1, ) - java_libraries = proto.RepeatedField( + java_libraries: MutableSequence[str] = proto.RepeatedField( proto.STRING, number=2, ) - python_packages = proto.RepeatedField( + python_packages: MutableSequence[str] = proto.RepeatedField( proto.STRING, number=3, ) - properties = proto.MapField( + properties: MutableMapping[str, str] = proto.MapField( proto.STRING, proto.STRING, number=4, ) - compute = proto.Field( + compute: "Environment.InfrastructureSpec.ComputeResources" = proto.Field( proto.MESSAGE, number=50, oneof="resources", message="Environment.InfrastructureSpec.ComputeResources", ) - os_image = proto.Field( + os_image: "Environment.InfrastructureSpec.OsImageRuntime" = proto.Field( proto.MESSAGE, number=100, oneof="runtime", @@ -194,12 +196,12 @@ class SessionSpec(proto.Message): default configuration. """ - max_idle_duration = proto.Field( + max_idle_duration: duration_pb2.Duration = proto.Field( proto.MESSAGE, number=1, message=duration_pb2.Duration, ) - enable_fast_startup = proto.Field( + enable_fast_startup: bool = proto.Field( proto.BOOL, number=2, ) @@ -214,7 +216,7 @@ class SessionStatus(proto.Message): not """ - active = proto.Field( + active: bool = proto.Field( proto.BOOL, number=1, ) @@ -229,67 +231,67 @@ class Endpoints(proto.Message): Output only. URI to serve SQL APIs """ - notebooks = proto.Field( + notebooks: str = proto.Field( proto.STRING, number=1, ) - sql = proto.Field( + sql: str = proto.Field( proto.STRING, number=2, ) - name = proto.Field( + name: str = proto.Field( proto.STRING, number=1, ) - display_name = proto.Field( + display_name: str = proto.Field( proto.STRING, number=2, ) - uid = proto.Field( + uid: str = proto.Field( proto.STRING, number=3, ) - create_time = proto.Field( + create_time: timestamp_pb2.Timestamp = proto.Field( proto.MESSAGE, number=4, message=timestamp_pb2.Timestamp, ) - update_time = proto.Field( + update_time: timestamp_pb2.Timestamp = proto.Field( proto.MESSAGE, number=5, message=timestamp_pb2.Timestamp, ) - labels = proto.MapField( + labels: MutableMapping[str, str] = proto.MapField( proto.STRING, proto.STRING, number=6, ) - description = proto.Field( + description: str = proto.Field( proto.STRING, number=7, ) - state = proto.Field( + state: resources.State = proto.Field( proto.ENUM, number=8, enum=resources.State, ) - infrastructure_spec = proto.Field( + infrastructure_spec: InfrastructureSpec = proto.Field( proto.MESSAGE, number=100, message=InfrastructureSpec, ) - session_spec = proto.Field( + session_spec: SessionSpec = proto.Field( proto.MESSAGE, number=101, message=SessionSpec, ) - session_status = proto.Field( + session_status: SessionStatus = proto.Field( proto.MESSAGE, number=102, message=SessionStatus, ) - endpoints = proto.Field( + endpoints: Endpoints = proto.Field( proto.MESSAGE, number=200, message=Endpoints, @@ -326,7 +328,7 @@ class Content(proto.Message): update_time (google.protobuf.timestamp_pb2.Timestamp): Output only. The time when the content was last updated. - labels (Mapping[str, str]): + labels (MutableMapping[str, str]): Optional. User defined labels for the content. description (str): @@ -359,7 +361,7 @@ class QueryEngine(proto.Enum): QUERY_ENGINE_UNSPECIFIED = 0 SPARK = 2 - engine = proto.Field( + engine: "Content.SqlScript.QueryEngine" = proto.Field( proto.ENUM, number=1, enum="Content.SqlScript.QueryEngine", @@ -378,55 +380,55 @@ class KernelType(proto.Enum): KERNEL_TYPE_UNSPECIFIED = 0 PYTHON3 = 1 - kernel_type = proto.Field( + kernel_type: "Content.Notebook.KernelType" = proto.Field( proto.ENUM, number=1, enum="Content.Notebook.KernelType", ) - name = proto.Field( + name: str = proto.Field( proto.STRING, number=1, ) - uid = proto.Field( + uid: str = proto.Field( proto.STRING, number=2, ) - path = proto.Field( + path: str = proto.Field( proto.STRING, number=3, ) - create_time = proto.Field( + create_time: timestamp_pb2.Timestamp = proto.Field( proto.MESSAGE, number=4, message=timestamp_pb2.Timestamp, ) - update_time = proto.Field( + update_time: timestamp_pb2.Timestamp = proto.Field( proto.MESSAGE, number=5, message=timestamp_pb2.Timestamp, ) - labels = proto.MapField( + labels: MutableMapping[str, str] = proto.MapField( proto.STRING, proto.STRING, number=6, ) - description = proto.Field( + description: str = proto.Field( proto.STRING, number=7, ) - data_text = proto.Field( + data_text: str = proto.Field( proto.STRING, number=9, oneof="data", ) - sql_script = proto.Field( + sql_script: SqlScript = proto.Field( proto.MESSAGE, number=100, oneof="content", message=SqlScript, ) - notebook = proto.Field( + notebook: Notebook = proto.Field( proto.MESSAGE, number=101, oneof="content", @@ -451,20 +453,20 @@ class Session(proto.Message): """ - name = proto.Field( + name: str = proto.Field( proto.STRING, number=1, ) - user_id = proto.Field( + user_id: str = proto.Field( proto.STRING, number=2, ) - create_time = proto.Field( + create_time: timestamp_pb2.Timestamp = proto.Field( proto.MESSAGE, number=3, message=timestamp_pb2.Timestamp, ) - state = proto.Field( + state: resources.State = proto.Field( proto.ENUM, number=4, enum=resources.State, diff --git a/google/cloud/dataplex_v1/types/content.py b/google/cloud/dataplex_v1/types/content.py index 3f589f4..2127b49 100644 --- a/google/cloud/dataplex_v1/types/content.py +++ b/google/cloud/dataplex_v1/types/content.py @@ -13,6 +13,8 @@ # See the License for the specific language governing permissions and # limitations under the License. # +from typing import MutableMapping, MutableSequence + from google.protobuf import field_mask_pb2 # type: ignore import proto # type: ignore @@ -45,16 +47,16 @@ class CreateContentRequest(proto.Message): not perform mutations. The default is false. """ - parent = proto.Field( + parent: str = proto.Field( proto.STRING, number=1, ) - content = proto.Field( + content: analyze.Content = proto.Field( proto.MESSAGE, number=2, message=analyze.Content, ) - validate_only = proto.Field( + validate_only: bool = proto.Field( proto.BOOL, number=3, ) @@ -74,17 +76,17 @@ class UpdateContentRequest(proto.Message): not perform mutations. The default is false. """ - update_mask = proto.Field( + update_mask: field_mask_pb2.FieldMask = proto.Field( proto.MESSAGE, number=1, message=field_mask_pb2.FieldMask, ) - content = proto.Field( + content: analyze.Content = proto.Field( proto.MESSAGE, number=2, message=analyze.Content, ) - validate_only = proto.Field( + validate_only: bool = proto.Field( proto.BOOL, number=3, ) @@ -99,7 +101,7 @@ class DeleteContentRequest(proto.Message): projects/{project_id}/locations/{location_id}/lakes/{lake_id}/content/{content_id} """ - name = proto.Field( + name: str = proto.Field( proto.STRING, number=1, ) @@ -135,19 +137,19 @@ class ListContentRequest(proto.Message): conjunctions. """ - parent = proto.Field( + parent: str = proto.Field( proto.STRING, number=1, ) - page_size = proto.Field( + page_size: int = proto.Field( proto.INT32, number=2, ) - page_token = proto.Field( + page_token: str = proto.Field( proto.STRING, number=3, ) - filter = proto.Field( + filter: str = proto.Field( proto.STRING, number=4, ) @@ -157,7 +159,7 @@ class ListContentResponse(proto.Message): r"""List content response. Attributes: - content (Sequence[google.cloud.dataplex_v1.types.Content]): + content (MutableSequence[google.cloud.dataplex_v1.types.Content]): Content under the given parent lake. next_page_token (str): Token to retrieve the next page of results, @@ -169,12 +171,12 @@ class ListContentResponse(proto.Message): def raw_page(self): return self - content = proto.RepeatedField( + content: MutableSequence[analyze.Content] = proto.RepeatedField( proto.MESSAGE, number=1, message=analyze.Content, ) - next_page_token = proto.Field( + next_page_token: str = proto.Field( proto.STRING, number=2, ) @@ -200,11 +202,11 @@ class ContentView(proto.Enum): BASIC = 1 FULL = 2 - name = proto.Field( + name: str = proto.Field( proto.STRING, number=1, ) - view = proto.Field( + view: ContentView = proto.Field( proto.ENUM, number=2, enum=ContentView, diff --git a/google/cloud/dataplex_v1/types/logs.py b/google/cloud/dataplex_v1/types/logs.py index 39db8b3..809bad9 100644 --- a/google/cloud/dataplex_v1/types/logs.py +++ b/google/cloud/dataplex_v1/types/logs.py @@ -13,6 +13,8 @@ # See the License for the specific language governing permissions and # limitations under the License. # +from typing import MutableMapping, MutableSequence + from google.protobuf import duration_pb2 # type: ignore from google.protobuf import timestamp_pb2 # type: ignore import proto # type: ignore @@ -93,7 +95,7 @@ class ConfigDetails(proto.Message): r"""Details about configuration events. Attributes: - parameters (Mapping[str, str]): + parameters (MutableMapping[str, str]): A list of discovery configuration parameters in effect. The keys are the field paths within DiscoverySpec. Eg. includePatterns, @@ -101,7 +103,7 @@ class ConfigDetails(proto.Message): csvOptions.disableTypeInference, etc. """ - parameters = proto.MapField( + parameters: MutableMapping[str, str] = proto.MapField( proto.STRING, proto.STRING, number=1, @@ -118,11 +120,11 @@ class EntityDetails(proto.Message): The type of the entity resource. """ - entity = proto.Field( + entity: str = proto.Field( proto.STRING, number=1, ) - type_ = proto.Field( + type_: "DiscoveryEvent.EntityType" = proto.Field( proto.ENUM, number=2, enum="DiscoveryEvent.EntityType", @@ -140,26 +142,26 @@ class PartitionDetails(proto.Message): The name is the fully-qualified resource name. type_ (google.cloud.dataplex_v1.types.DiscoveryEvent.EntityType): The type of the containing entity resource. - sampled_data_locations (Sequence[str]): + sampled_data_locations (MutableSequence[str]): The locations of the data items (e.g., a Cloud Storage objects) sampled for metadata inference. """ - partition = proto.Field( + partition: str = proto.Field( proto.STRING, number=1, ) - entity = proto.Field( + entity: str = proto.Field( proto.STRING, number=2, ) - type_ = proto.Field( + type_: "DiscoveryEvent.EntityType" = proto.Field( proto.ENUM, number=3, enum="DiscoveryEvent.EntityType", ) - sampled_data_locations = proto.RepeatedField( + sampled_data_locations: MutableSequence[str] = proto.RepeatedField( proto.STRING, number=4, ) @@ -173,55 +175,55 @@ class ActionDetails(proto.Message): Eg. IncompatibleDataSchema, InvalidDataFormat """ - type_ = proto.Field( + type_: str = proto.Field( proto.STRING, number=1, ) - message = proto.Field( + message: str = proto.Field( proto.STRING, number=1, ) - lake_id = proto.Field( + lake_id: str = proto.Field( proto.STRING, number=2, ) - zone_id = proto.Field( + zone_id: str = proto.Field( proto.STRING, number=3, ) - asset_id = proto.Field( + asset_id: str = proto.Field( proto.STRING, number=4, ) - data_location = proto.Field( + data_location: str = proto.Field( proto.STRING, number=5, ) - type_ = proto.Field( + type_: EventType = proto.Field( proto.ENUM, number=10, enum=EventType, ) - config = proto.Field( + config: ConfigDetails = proto.Field( proto.MESSAGE, number=20, oneof="details", message=ConfigDetails, ) - entity = proto.Field( + entity: EntityDetails = proto.Field( proto.MESSAGE, number=21, oneof="details", message=EntityDetails, ) - partition = proto.Field( + partition: PartitionDetails = proto.Field( proto.MESSAGE, number=22, oneof="details", message=PartitionDetails, ) - action = proto.Field( + action: ActionDetails = proto.Field( proto.MESSAGE, number=23, oneof="details", @@ -273,44 +275,44 @@ class Service(proto.Enum): SERVICE_UNSPECIFIED = 0 DATAPROC = 1 - message = proto.Field( + message: str = proto.Field( proto.STRING, number=1, ) - job_id = proto.Field( + job_id: str = proto.Field( proto.STRING, number=2, ) - start_time = proto.Field( + start_time: timestamp_pb2.Timestamp = proto.Field( proto.MESSAGE, number=3, message=timestamp_pb2.Timestamp, ) - end_time = proto.Field( + end_time: timestamp_pb2.Timestamp = proto.Field( proto.MESSAGE, number=4, message=timestamp_pb2.Timestamp, ) - state = proto.Field( + state: State = proto.Field( proto.ENUM, number=5, enum=State, ) - retries = proto.Field( + retries: int = proto.Field( proto.INT32, number=6, ) - type_ = proto.Field( + type_: Type = proto.Field( proto.ENUM, number=7, enum=Type, ) - service = proto.Field( + service: Service = proto.Field( proto.ENUM, number=8, enum=Service, ) - service_job = proto.Field( + service_job: str = proto.Field( proto.STRING, number=9, ) @@ -381,65 +383,65 @@ class Engine(proto.Enum): SPARK_SQL = 1 BIGQUERY = 2 - query_id = proto.Field( + query_id: str = proto.Field( proto.STRING, number=1, ) - query_text = proto.Field( + query_text: str = proto.Field( proto.STRING, number=2, ) - engine = proto.Field( + engine: "SessionEvent.QueryDetail.Engine" = proto.Field( proto.ENUM, number=3, enum="SessionEvent.QueryDetail.Engine", ) - duration = proto.Field( + duration: duration_pb2.Duration = proto.Field( proto.MESSAGE, number=4, message=duration_pb2.Duration, ) - result_size_bytes = proto.Field( + result_size_bytes: int = proto.Field( proto.INT64, number=5, ) - data_processed_bytes = proto.Field( + data_processed_bytes: int = proto.Field( proto.INT64, number=6, ) - message = proto.Field( + message: str = proto.Field( proto.STRING, number=1, ) - user_id = proto.Field( + user_id: str = proto.Field( proto.STRING, number=2, ) - session_id = proto.Field( + session_id: str = proto.Field( proto.STRING, number=3, ) - type_ = proto.Field( + type_: EventType = proto.Field( proto.ENUM, number=4, enum=EventType, ) - query = proto.Field( + query: QueryDetail = proto.Field( proto.MESSAGE, number=5, oneof="detail", message=QueryDetail, ) - event_succeeded = proto.Field( + event_succeeded: bool = proto.Field( proto.BOOL, number=6, ) - fast_startup_enabled = proto.Field( + fast_startup_enabled: bool = proto.Field( proto.BOOL, number=7, ) - unassigned_duration = proto.Field( + unassigned_duration: duration_pb2.Duration = proto.Field( proto.MESSAGE, number=8, message=duration_pb2.Duration, diff --git a/google/cloud/dataplex_v1/types/metadata_.py b/google/cloud/dataplex_v1/types/metadata_.py index e0e7e75..e1bf939 100644 --- a/google/cloud/dataplex_v1/types/metadata_.py +++ b/google/cloud/dataplex_v1/types/metadata_.py @@ -13,6 +13,8 @@ # See the License for the specific language governing permissions and # limitations under the License. # +from typing import MutableMapping, MutableSequence + from google.protobuf import timestamp_pb2 # type: ignore import proto # type: ignore @@ -60,16 +62,16 @@ class CreateEntityRequest(proto.Message): not perform mutations. The default is false. """ - parent = proto.Field( + parent: str = proto.Field( proto.STRING, number=1, ) - entity = proto.Field( + entity: "Entity" = proto.Field( proto.MESSAGE, number=3, message="Entity", ) - validate_only = proto.Field( + validate_only: bool = proto.Field( proto.BOOL, number=4, ) @@ -90,12 +92,12 @@ class UpdateEntityRequest(proto.Message): not perform mutations. The default is false. """ - entity = proto.Field( + entity: "Entity" = proto.Field( proto.MESSAGE, number=2, message="Entity", ) - validate_only = proto.Field( + validate_only: bool = proto.Field( proto.BOOL, number=3, ) @@ -113,11 +115,11 @@ class DeleteEntityRequest(proto.Message): retrieved with a [GetEntity][] request. """ - name = proto.Field( + name: str = proto.Field( proto.STRING, number=1, ) - etag = proto.Field( + etag: str = proto.Field( proto.STRING, number=2, ) @@ -163,24 +165,24 @@ class EntityView(proto.Enum): TABLES = 1 FILESETS = 2 - parent = proto.Field( + parent: str = proto.Field( proto.STRING, number=1, ) - view = proto.Field( + view: EntityView = proto.Field( proto.ENUM, number=2, enum=EntityView, ) - page_size = proto.Field( + page_size: int = proto.Field( proto.INT32, number=3, ) - page_token = proto.Field( + page_token: str = proto.Field( proto.STRING, number=4, ) - filter = proto.Field( + filter: str = proto.Field( proto.STRING, number=5, ) @@ -190,7 +192,7 @@ class ListEntitiesResponse(proto.Message): r"""List metadata entities response. Attributes: - entities (Sequence[google.cloud.dataplex_v1.types.Entity]): + entities (MutableSequence[google.cloud.dataplex_v1.types.Entity]): Entities in the specified parent zone. next_page_token (str): Token to retrieve the next page of results, @@ -202,12 +204,12 @@ class ListEntitiesResponse(proto.Message): def raw_page(self): return self - entities = proto.RepeatedField( + entities: MutableSequence["Entity"] = proto.RepeatedField( proto.MESSAGE, number=1, message="Entity", ) - next_page_token = proto.Field( + next_page_token: str = proto.Field( proto.STRING, number=2, ) @@ -232,11 +234,11 @@ class EntityView(proto.Enum): SCHEMA = 2 FULL = 4 - name = proto.Field( + name: str = proto.Field( proto.STRING, number=1, ) - view = proto.Field( + view: EntityView = proto.Field( proto.ENUM, number=2, enum=EntityView, @@ -290,19 +292,19 @@ class ListPartitionsRequest(proto.Message): used as a filter. """ - parent = proto.Field( + parent: str = proto.Field( proto.STRING, number=1, ) - page_size = proto.Field( + page_size: int = proto.Field( proto.INT32, number=2, ) - page_token = proto.Field( + page_token: str = proto.Field( proto.STRING, number=3, ) - filter = proto.Field( + filter: str = proto.Field( proto.STRING, number=4, ) @@ -322,16 +324,16 @@ class CreatePartitionRequest(proto.Message): not perform mutations. The default is false. """ - parent = proto.Field( + parent: str = proto.Field( proto.STRING, number=1, ) - partition = proto.Field( + partition: "Partition" = proto.Field( proto.MESSAGE, number=3, message="Partition", ) - validate_only = proto.Field( + validate_only: bool = proto.Field( proto.BOOL, number=4, ) @@ -352,11 +354,11 @@ class DeletePartitionRequest(proto.Message): partition. """ - name = proto.Field( + name: str = proto.Field( proto.STRING, number=1, ) - etag = proto.Field( + etag: str = proto.Field( proto.STRING, number=2, ) @@ -366,7 +368,7 @@ class ListPartitionsResponse(proto.Message): r"""List metadata partitions response. Attributes: - partitions (Sequence[google.cloud.dataplex_v1.types.Partition]): + partitions (MutableSequence[google.cloud.dataplex_v1.types.Partition]): Partitions under the specified parent entity. next_page_token (str): Token to retrieve the next page of results, @@ -378,12 +380,12 @@ class ListPartitionsResponse(proto.Message): def raw_page(self): return self - partitions = proto.RepeatedField( + partitions: MutableSequence["Partition"] = proto.RepeatedField( proto.MESSAGE, number=1, message="Partition", ) - next_page_token = proto.Field( + next_page_token: str = proto.Field( proto.STRING, number=2, ) @@ -401,7 +403,7 @@ class GetPartitionRequest(proto.Message): must be provided. """ - name = proto.Field( + name: str = proto.Field( proto.STRING, number=1, ) @@ -511,93 +513,93 @@ class Compatibility(proto.Message): store. """ - compatible = proto.Field( + compatible: bool = proto.Field( proto.BOOL, number=1, ) - reason = proto.Field( + reason: str = proto.Field( proto.STRING, number=2, ) - hive_metastore = proto.Field( + hive_metastore: "Entity.CompatibilityStatus.Compatibility" = proto.Field( proto.MESSAGE, number=1, message="Entity.CompatibilityStatus.Compatibility", ) - bigquery = proto.Field( + bigquery: "Entity.CompatibilityStatus.Compatibility" = proto.Field( proto.MESSAGE, number=2, message="Entity.CompatibilityStatus.Compatibility", ) - name = proto.Field( + name: str = proto.Field( proto.STRING, number=1, ) - display_name = proto.Field( + display_name: str = proto.Field( proto.STRING, number=2, ) - description = proto.Field( + description: str = proto.Field( proto.STRING, number=3, ) - create_time = proto.Field( + create_time: timestamp_pb2.Timestamp = proto.Field( proto.MESSAGE, number=5, message=timestamp_pb2.Timestamp, ) - update_time = proto.Field( + update_time: timestamp_pb2.Timestamp = proto.Field( proto.MESSAGE, number=6, message=timestamp_pb2.Timestamp, ) - id = proto.Field( + id: str = proto.Field( proto.STRING, number=7, ) - etag = proto.Field( + etag: str = proto.Field( proto.STRING, number=8, ) - type_ = proto.Field( + type_: Type = proto.Field( proto.ENUM, number=10, enum=Type, ) - asset = proto.Field( + asset: str = proto.Field( proto.STRING, number=11, ) - data_path = proto.Field( + data_path: str = proto.Field( proto.STRING, number=12, ) - data_path_pattern = proto.Field( + data_path_pattern: str = proto.Field( proto.STRING, number=13, ) - catalog_entry = proto.Field( + catalog_entry: str = proto.Field( proto.STRING, number=14, ) - system = proto.Field( + system: "StorageSystem" = proto.Field( proto.ENUM, number=15, enum="StorageSystem", ) - format_ = proto.Field( + format_: "StorageFormat" = proto.Field( proto.MESSAGE, number=16, message="StorageFormat", ) - compatibility = proto.Field( + compatibility: CompatibilityStatus = proto.Field( proto.MESSAGE, number=19, message=CompatibilityStatus, ) - schema = proto.Field( + schema: "Schema" = proto.Field( proto.MESSAGE, number=50, message="Schema", @@ -616,7 +618,7 @@ class Partition(proto.Message): "US:CA/CA#Sunnyvale so that the request URL ends with "/partitions/US%253ACA/CA%2523Sunnyvale". The name field in the response retains the encoded format. - values (Sequence[str]): + values (MutableSequence[str]): Required. Immutable. The set of values representing the partition, which correspond to the partition schema defined in the parent @@ -630,19 +632,19 @@ class Partition(proto.Message): Optional. The etag for this partition. """ - name = proto.Field( + name: str = proto.Field( proto.STRING, number=1, ) - values = proto.RepeatedField( + values: MutableSequence[str] = proto.RepeatedField( proto.STRING, number=2, ) - location = proto.Field( + location: str = proto.Field( proto.STRING, number=3, ) - etag = proto.Field( + etag: str = proto.Field( proto.STRING, number=4, ) @@ -669,10 +671,10 @@ class Schema(proto.Message): - Set to ``true`` to fully manage the entity schema. This setting guarantees that Dataplex will not change schema fields. - fields (Sequence[google.cloud.dataplex_v1.types.Schema.SchemaField]): + fields (MutableSequence[google.cloud.dataplex_v1.types.Schema.SchemaField]): Optional. The sequence of fields describing data in table entities. **Note:** BigQuery SchemaFields are immutable. - partition_fields (Sequence[google.cloud.dataplex_v1.types.Schema.PartitionField]): + partition_fields (MutableSequence[google.cloud.dataplex_v1.types.Schema.PartitionField]): Optional. The sequence of fields describing the partition structure in entities. If this field is empty, there are no partitions within @@ -731,29 +733,29 @@ class SchemaField(proto.Message): Required. The type of field. mode (google.cloud.dataplex_v1.types.Schema.Mode): Required. Additional field semantics. - fields (Sequence[google.cloud.dataplex_v1.types.Schema.SchemaField]): + fields (MutableSequence[google.cloud.dataplex_v1.types.Schema.SchemaField]): Optional. Any nested field for complex types. """ - name = proto.Field( + name: str = proto.Field( proto.STRING, number=1, ) - description = proto.Field( + description: str = proto.Field( proto.STRING, number=2, ) - type_ = proto.Field( + type_: "Schema.Type" = proto.Field( proto.ENUM, number=3, enum="Schema.Type", ) - mode = proto.Field( + mode: "Schema.Mode" = proto.Field( proto.ENUM, number=4, enum="Schema.Mode", ) - fields = proto.RepeatedField( + fields: MutableSequence["Schema.SchemaField"] = proto.RepeatedField( proto.MESSAGE, number=10, message="Schema.SchemaField", @@ -775,31 +777,31 @@ class PartitionField(proto.Message): Required. Immutable. The type of field. """ - name = proto.Field( + name: str = proto.Field( proto.STRING, number=1, ) - type_ = proto.Field( + type_: "Schema.Type" = proto.Field( proto.ENUM, number=2, enum="Schema.Type", ) - user_managed = proto.Field( + user_managed: bool = proto.Field( proto.BOOL, number=1, ) - fields = proto.RepeatedField( + fields: MutableSequence[SchemaField] = proto.RepeatedField( proto.MESSAGE, number=2, message=SchemaField, ) - partition_fields = proto.RepeatedField( + partition_fields: MutableSequence[PartitionField] = proto.RepeatedField( proto.MESSAGE, number=3, message=PartitionField, ) - partition_style = proto.Field( + partition_style: PartitionStyle = proto.Field( proto.ENUM, number=4, enum=PartitionStyle, @@ -896,19 +898,19 @@ class CsvOptions(proto.Message): (double quotation mark) if unspecified. """ - encoding = proto.Field( + encoding: str = proto.Field( proto.STRING, number=1, ) - header_rows = proto.Field( + header_rows: int = proto.Field( proto.INT32, number=2, ) - delimiter = proto.Field( + delimiter: str = proto.Field( proto.STRING, number=3, ) - quote = proto.Field( + quote: str = proto.Field( proto.STRING, number=4, ) @@ -923,32 +925,32 @@ class JsonOptions(proto.Message): Defaults to UTF-8 if not specified. """ - encoding = proto.Field( + encoding: str = proto.Field( proto.STRING, number=1, ) - format_ = proto.Field( + format_: Format = proto.Field( proto.ENUM, number=1, enum=Format, ) - compression_format = proto.Field( + compression_format: CompressionFormat = proto.Field( proto.ENUM, number=2, enum=CompressionFormat, ) - mime_type = proto.Field( + mime_type: str = proto.Field( proto.STRING, number=3, ) - csv = proto.Field( + csv: CsvOptions = proto.Field( proto.MESSAGE, number=10, oneof="options", message=CsvOptions, ) - json = proto.Field( + json: JsonOptions = proto.Field( proto.MESSAGE, number=11, oneof="options", diff --git a/google/cloud/dataplex_v1/types/resources.py b/google/cloud/dataplex_v1/types/resources.py index 64039d8..15326cb 100644 --- a/google/cloud/dataplex_v1/types/resources.py +++ b/google/cloud/dataplex_v1/types/resources.py @@ -13,6 +13,8 @@ # See the License for the specific language governing permissions and # limitations under the License. # +from typing import MutableMapping, MutableSequence + from google.protobuf import duration_pb2 # type: ignore from google.protobuf import timestamp_pb2 # type: ignore import proto # type: ignore @@ -69,7 +71,7 @@ class Lake(proto.Message): update_time (google.protobuf.timestamp_pb2.Timestamp): Output only. The time when the lake was last updated. - labels (Mapping[str, str]): + labels (MutableMapping[str, str]): Optional. User-defined labels for the lake. description (str): Optional. Description of the lake. @@ -102,7 +104,7 @@ class Metastore(proto.Message): ``projects/{project_id}/locations/{location_id}/services/{service_id}`` """ - service = proto.Field( + service: str = proto.Field( proto.STRING, number=1, ) @@ -133,76 +135,76 @@ class State(proto.Enum): UPDATING = 3 ERROR = 4 - state = proto.Field( + state: "Lake.MetastoreStatus.State" = proto.Field( proto.ENUM, number=1, enum="Lake.MetastoreStatus.State", ) - message = proto.Field( + message: str = proto.Field( proto.STRING, number=2, ) - update_time = proto.Field( + update_time: timestamp_pb2.Timestamp = proto.Field( proto.MESSAGE, number=3, message=timestamp_pb2.Timestamp, ) - endpoint = proto.Field( + endpoint: str = proto.Field( proto.STRING, number=4, ) - name = proto.Field( + name: str = proto.Field( proto.STRING, number=1, ) - display_name = proto.Field( + display_name: str = proto.Field( proto.STRING, number=2, ) - uid = proto.Field( + uid: str = proto.Field( proto.STRING, number=3, ) - create_time = proto.Field( + create_time: timestamp_pb2.Timestamp = proto.Field( proto.MESSAGE, number=4, message=timestamp_pb2.Timestamp, ) - update_time = proto.Field( + update_time: timestamp_pb2.Timestamp = proto.Field( proto.MESSAGE, number=5, message=timestamp_pb2.Timestamp, ) - labels = proto.MapField( + labels: MutableMapping[str, str] = proto.MapField( proto.STRING, proto.STRING, number=6, ) - description = proto.Field( + description: str = proto.Field( proto.STRING, number=7, ) - state = proto.Field( + state: "State" = proto.Field( proto.ENUM, number=8, enum="State", ) - service_account = proto.Field( + service_account: str = proto.Field( proto.STRING, number=9, ) - metastore = proto.Field( + metastore: Metastore = proto.Field( proto.MESSAGE, number=102, message=Metastore, ) - asset_status = proto.Field( + asset_status: "AssetStatus" = proto.Field( proto.MESSAGE, number=103, message="AssetStatus", ) - metastore_status = proto.Field( + metastore_status: MetastoreStatus = proto.Field( proto.MESSAGE, number=104, message=MetastoreStatus, @@ -223,16 +225,16 @@ class AssetStatus(proto.Message): resources. """ - update_time = proto.Field( + update_time: timestamp_pb2.Timestamp = proto.Field( proto.MESSAGE, number=1, message=timestamp_pb2.Timestamp, ) - active_assets = proto.Field( + active_assets: int = proto.Field( proto.INT32, number=2, ) - security_policy_applying_assets = proto.Field( + security_policy_applying_assets: int = proto.Field( proto.INT32, number=3, ) @@ -263,7 +265,7 @@ class Zone(proto.Message): update_time (google.protobuf.timestamp_pb2.Timestamp): Output only. The time when the zone was last updated. - labels (Mapping[str, str]): + labels (MutableMapping[str, str]): Optional. User defined labels for the zone. description (str): Optional. Description of the zone. @@ -304,7 +306,7 @@ class LocationType(proto.Enum): SINGLE_REGION = 1 MULTI_REGION = 2 - location_type = proto.Field( + location_type: "Zone.ResourceSpec.LocationType" = proto.Field( proto.ENUM, number=1, enum="Zone.ResourceSpec.LocationType", @@ -320,7 +322,7 @@ class DiscoverySpec(proto.Message): Attributes: enabled (bool): Required. Whether discovery is enabled. - include_patterns (Sequence[str]): + include_patterns (MutableSequence[str]): Optional. The list of patterns to apply for selecting data to include during discovery if only a subset of the data should considered. For @@ -329,7 +331,7 @@ class DiscoverySpec(proto.Message): object names. For BigQuery dataset assets, these are interpreted as patterns to match table names. - exclude_patterns (Sequence[str]): + exclude_patterns (MutableSequence[str]): Optional. The list of patterns to apply for selecting data to exclude during discovery. For Cloud Storage bucket assets, these are @@ -376,19 +378,19 @@ class CsvOptions(proto.Message): will be registered as strings. """ - header_rows = proto.Field( + header_rows: int = proto.Field( proto.INT32, number=1, ) - delimiter = proto.Field( + delimiter: str = proto.Field( proto.STRING, number=2, ) - encoding = proto.Field( + encoding: str = proto.Field( proto.STRING, number=3, ) - disable_type_inference = proto.Field( + disable_type_inference: bool = proto.Field( proto.BOOL, number=4, ) @@ -407,95 +409,95 @@ class JsonOptions(proto.Message): (strings, number or boolean). """ - encoding = proto.Field( + encoding: str = proto.Field( proto.STRING, number=1, ) - disable_type_inference = proto.Field( + disable_type_inference: bool = proto.Field( proto.BOOL, number=2, ) - enabled = proto.Field( + enabled: bool = proto.Field( proto.BOOL, number=1, ) - include_patterns = proto.RepeatedField( + include_patterns: MutableSequence[str] = proto.RepeatedField( proto.STRING, number=2, ) - exclude_patterns = proto.RepeatedField( + exclude_patterns: MutableSequence[str] = proto.RepeatedField( proto.STRING, number=3, ) - csv_options = proto.Field( + csv_options: "Zone.DiscoverySpec.CsvOptions" = proto.Field( proto.MESSAGE, number=4, message="Zone.DiscoverySpec.CsvOptions", ) - json_options = proto.Field( + json_options: "Zone.DiscoverySpec.JsonOptions" = proto.Field( proto.MESSAGE, number=5, message="Zone.DiscoverySpec.JsonOptions", ) - schedule = proto.Field( + schedule: str = proto.Field( proto.STRING, number=10, oneof="trigger", ) - name = proto.Field( + name: str = proto.Field( proto.STRING, number=1, ) - display_name = proto.Field( + display_name: str = proto.Field( proto.STRING, number=2, ) - uid = proto.Field( + uid: str = proto.Field( proto.STRING, number=3, ) - create_time = proto.Field( + create_time: timestamp_pb2.Timestamp = proto.Field( proto.MESSAGE, number=4, message=timestamp_pb2.Timestamp, ) - update_time = proto.Field( + update_time: timestamp_pb2.Timestamp = proto.Field( proto.MESSAGE, number=5, message=timestamp_pb2.Timestamp, ) - labels = proto.MapField( + labels: MutableMapping[str, str] = proto.MapField( proto.STRING, proto.STRING, number=6, ) - description = proto.Field( + description: str = proto.Field( proto.STRING, number=7, ) - state = proto.Field( + state: "State" = proto.Field( proto.ENUM, number=8, enum="State", ) - type_ = proto.Field( + type_: Type = proto.Field( proto.ENUM, number=9, enum=Type, ) - discovery_spec = proto.Field( + discovery_spec: DiscoverySpec = proto.Field( proto.MESSAGE, number=103, message=DiscoverySpec, ) - resource_spec = proto.Field( + resource_spec: ResourceSpec = proto.Field( proto.MESSAGE, number=104, message=ResourceSpec, ) - asset_status = proto.Field( + asset_status: "AssetStatus" = proto.Field( proto.MESSAGE, number=105, message="AssetStatus", @@ -540,7 +542,7 @@ class Action(proto.Message): Output only. The relative resource name of the asset, of the form: ``projects/{project_number}/locations/{location_id}/lakes/{lake_id}/zones/{zone_id}/assets/{asset_id}``. - data_locations (Sequence[str]): + data_locations (MutableSequence[str]): The list of data locations associated with this action. Cloud Storage locations are represented as URI paths(E.g. ``gs://bucket/table1/year=2020/month=Jan/``). BigQuery @@ -622,7 +624,7 @@ class FailedSecurityPolicyApply(proto.Message): for a lake or zone resource only. """ - asset = proto.Field( + asset: str = proto.Field( proto.STRING, number=1, ) @@ -632,7 +634,7 @@ class InvalidDataFormat(proto.Message): by discovery. Attributes: - sampled_data_locations (Sequence[str]): + sampled_data_locations (MutableSequence[str]): The list of data locations sampled and used for format/schema inference. expected_format (str): @@ -642,15 +644,15 @@ class InvalidDataFormat(proto.Message): entity. """ - sampled_data_locations = proto.RepeatedField( + sampled_data_locations: MutableSequence[str] = proto.RepeatedField( proto.STRING, number=1, ) - expected_format = proto.Field( + expected_format: str = proto.Field( proto.STRING, number=2, ) - new_format = proto.Field( + new_format: str = proto.Field( proto.STRING, number=3, ) @@ -673,7 +675,7 @@ class IncompatibleDataSchema(proto.Message): table. The schema is provided as a JSON formatted structured listing columns and data types. - sampled_data_locations (Sequence[str]): + sampled_data_locations (MutableSequence[str]): The list of data locations sampled and used for format/schema inference. schema_change (google.cloud.dataplex_v1.types.Action.IncompatibleDataSchema.SchemaChange): @@ -689,23 +691,23 @@ class SchemaChange(proto.Enum): INCOMPATIBLE = 1 MODIFIED = 2 - table = proto.Field( + table: str = proto.Field( proto.STRING, number=1, ) - existing_schema = proto.Field( + existing_schema: str = proto.Field( proto.STRING, number=2, ) - new_schema = proto.Field( + new_schema: str = proto.Field( proto.STRING, number=3, ) - sampled_data_locations = proto.RepeatedField( + sampled_data_locations: MutableSequence[str] = proto.RepeatedField( proto.STRING, number=4, ) - schema_change = proto.Field( + schema_change: "Action.IncompatibleDataSchema.SchemaChange" = proto.Field( proto.ENUM, number=5, enum="Action.IncompatibleDataSchema.SchemaChange", @@ -726,10 +728,12 @@ class PartitionStructure(proto.Enum): CONSISTENT_KEYS = 1 HIVE_STYLE_KEYS = 2 - expected_structure = proto.Field( - proto.ENUM, - number=1, - enum="Action.InvalidDataPartition.PartitionStructure", + expected_structure: "Action.InvalidDataPartition.PartitionStructure" = ( + proto.Field( + proto.ENUM, + number=1, + enum="Action.InvalidDataPartition.PartitionStructure", + ) ) class MissingData(proto.Message): @@ -738,83 +742,83 @@ class MissingData(proto.Message): class InvalidDataOrganization(proto.Message): r"""Action details for invalid data arrangement.""" - category = proto.Field( + category: Category = proto.Field( proto.ENUM, number=1, enum=Category, ) - issue = proto.Field( + issue: str = proto.Field( proto.STRING, number=2, ) - detect_time = proto.Field( + detect_time: timestamp_pb2.Timestamp = proto.Field( proto.MESSAGE, number=4, message=timestamp_pb2.Timestamp, ) - name = proto.Field( + name: str = proto.Field( proto.STRING, number=5, ) - lake = proto.Field( + lake: str = proto.Field( proto.STRING, number=6, ) - zone = proto.Field( + zone: str = proto.Field( proto.STRING, number=7, ) - asset = proto.Field( + asset: str = proto.Field( proto.STRING, number=8, ) - data_locations = proto.RepeatedField( + data_locations: MutableSequence[str] = proto.RepeatedField( proto.STRING, number=9, ) - invalid_data_format = proto.Field( + invalid_data_format: InvalidDataFormat = proto.Field( proto.MESSAGE, number=10, oneof="details", message=InvalidDataFormat, ) - incompatible_data_schema = proto.Field( + incompatible_data_schema: IncompatibleDataSchema = proto.Field( proto.MESSAGE, number=11, oneof="details", message=IncompatibleDataSchema, ) - invalid_data_partition = proto.Field( + invalid_data_partition: InvalidDataPartition = proto.Field( proto.MESSAGE, number=12, oneof="details", message=InvalidDataPartition, ) - missing_data = proto.Field( + missing_data: MissingData = proto.Field( proto.MESSAGE, number=13, oneof="details", message=MissingData, ) - missing_resource = proto.Field( + missing_resource: MissingResource = proto.Field( proto.MESSAGE, number=14, oneof="details", message=MissingResource, ) - unauthorized_resource = proto.Field( + unauthorized_resource: UnauthorizedResource = proto.Field( proto.MESSAGE, number=15, oneof="details", message=UnauthorizedResource, ) - failed_security_policy_apply = proto.Field( + failed_security_policy_apply: FailedSecurityPolicyApply = proto.Field( proto.MESSAGE, number=21, oneof="details", message=FailedSecurityPolicyApply, ) - invalid_data_organization = proto.Field( + invalid_data_organization: InvalidDataOrganization = proto.Field( proto.MESSAGE, number=22, oneof="details", @@ -844,7 +848,7 @@ class Asset(proto.Message): update_time (google.protobuf.timestamp_pb2.Timestamp): Output only. The time when the asset was last updated. - labels (Mapping[str, str]): + labels (MutableMapping[str, str]): Optional. User defined labels for the asset. description (str): Optional. Description of the asset. @@ -892,16 +896,16 @@ class State(proto.Enum): APPLYING = 2 ERROR = 3 - state = proto.Field( + state: "Asset.SecurityStatus.State" = proto.Field( proto.ENUM, number=1, enum="Asset.SecurityStatus.State", ) - message = proto.Field( + message: str = proto.Field( proto.STRING, number=2, ) - update_time = proto.Field( + update_time: timestamp_pb2.Timestamp = proto.Field( proto.MESSAGE, number=3, message=timestamp_pb2.Timestamp, @@ -917,7 +921,7 @@ class DiscoverySpec(proto.Message): Attributes: enabled (bool): Optional. Whether discovery is enabled. - include_patterns (Sequence[str]): + include_patterns (MutableSequence[str]): Optional. The list of patterns to apply for selecting data to include during discovery if only a subset of the data should considered. @@ -926,7 +930,7 @@ class DiscoverySpec(proto.Message): object names. For BigQuery dataset assets, these are interpreted as patterns to match table names. - exclude_patterns (Sequence[str]): + exclude_patterns (MutableSequence[str]): Optional. The list of patterns to apply for selecting data to exclude during discovery. For Cloud Storage bucket assets, these are @@ -973,19 +977,19 @@ class CsvOptions(proto.Message): will be registered as strings. """ - header_rows = proto.Field( + header_rows: int = proto.Field( proto.INT32, number=1, ) - delimiter = proto.Field( + delimiter: str = proto.Field( proto.STRING, number=2, ) - encoding = proto.Field( + encoding: str = proto.Field( proto.STRING, number=3, ) - disable_type_inference = proto.Field( + disable_type_inference: bool = proto.Field( proto.BOOL, number=4, ) @@ -1004,38 +1008,38 @@ class JsonOptions(proto.Message): (strings, number or boolean). """ - encoding = proto.Field( + encoding: str = proto.Field( proto.STRING, number=1, ) - disable_type_inference = proto.Field( + disable_type_inference: bool = proto.Field( proto.BOOL, number=2, ) - enabled = proto.Field( + enabled: bool = proto.Field( proto.BOOL, number=1, ) - include_patterns = proto.RepeatedField( + include_patterns: MutableSequence[str] = proto.RepeatedField( proto.STRING, number=2, ) - exclude_patterns = proto.RepeatedField( + exclude_patterns: MutableSequence[str] = proto.RepeatedField( proto.STRING, number=3, ) - csv_options = proto.Field( + csv_options: "Asset.DiscoverySpec.CsvOptions" = proto.Field( proto.MESSAGE, number=4, message="Asset.DiscoverySpec.CsvOptions", ) - json_options = proto.Field( + json_options: "Asset.DiscoverySpec.JsonOptions" = proto.Field( proto.MESSAGE, number=5, message="Asset.DiscoverySpec.JsonOptions", ) - schedule = proto.Field( + schedule: str = proto.Field( proto.STRING, number=10, oneof="trigger", @@ -1061,11 +1065,11 @@ class Type(proto.Enum): STORAGE_BUCKET = 1 BIGQUERY_DATASET = 2 - name = proto.Field( + name: str = proto.Field( proto.STRING, number=1, ) - type_ = proto.Field( + type_: "Asset.ResourceSpec.Type" = proto.Field( proto.ENUM, number=2, enum="Asset.ResourceSpec.Type", @@ -1090,16 +1094,16 @@ class State(proto.Enum): READY = 1 ERROR = 2 - state = proto.Field( + state: "Asset.ResourceStatus.State" = proto.Field( proto.ENUM, number=1, enum="Asset.ResourceStatus.State", ) - message = proto.Field( + message: str = proto.Field( proto.STRING, number=2, ) - update_time = proto.Field( + update_time: timestamp_pb2.Timestamp = proto.Field( proto.MESSAGE, number=3, message=timestamp_pb2.Timestamp, @@ -1152,110 +1156,110 @@ class Stats(proto.Message): referenced resource. """ - data_items = proto.Field( + data_items: int = proto.Field( proto.INT64, number=1, ) - data_size = proto.Field( + data_size: int = proto.Field( proto.INT64, number=2, ) - tables = proto.Field( + tables: int = proto.Field( proto.INT64, number=3, ) - filesets = proto.Field( + filesets: int = proto.Field( proto.INT64, number=4, ) - state = proto.Field( + state: "Asset.DiscoveryStatus.State" = proto.Field( proto.ENUM, number=1, enum="Asset.DiscoveryStatus.State", ) - message = proto.Field( + message: str = proto.Field( proto.STRING, number=2, ) - update_time = proto.Field( + update_time: timestamp_pb2.Timestamp = proto.Field( proto.MESSAGE, number=3, message=timestamp_pb2.Timestamp, ) - last_run_time = proto.Field( + last_run_time: timestamp_pb2.Timestamp = proto.Field( proto.MESSAGE, number=4, message=timestamp_pb2.Timestamp, ) - stats = proto.Field( + stats: "Asset.DiscoveryStatus.Stats" = proto.Field( proto.MESSAGE, number=6, message="Asset.DiscoveryStatus.Stats", ) - last_run_duration = proto.Field( + last_run_duration: duration_pb2.Duration = proto.Field( proto.MESSAGE, number=7, message=duration_pb2.Duration, ) - name = proto.Field( + name: str = proto.Field( proto.STRING, number=1, ) - display_name = proto.Field( + display_name: str = proto.Field( proto.STRING, number=2, ) - uid = proto.Field( + uid: str = proto.Field( proto.STRING, number=3, ) - create_time = proto.Field( + create_time: timestamp_pb2.Timestamp = proto.Field( proto.MESSAGE, number=4, message=timestamp_pb2.Timestamp, ) - update_time = proto.Field( + update_time: timestamp_pb2.Timestamp = proto.Field( proto.MESSAGE, number=5, message=timestamp_pb2.Timestamp, ) - labels = proto.MapField( + labels: MutableMapping[str, str] = proto.MapField( proto.STRING, proto.STRING, number=6, ) - description = proto.Field( + description: str = proto.Field( proto.STRING, number=7, ) - state = proto.Field( + state: "State" = proto.Field( proto.ENUM, number=8, enum="State", ) - resource_spec = proto.Field( + resource_spec: ResourceSpec = proto.Field( proto.MESSAGE, number=100, message=ResourceSpec, ) - resource_status = proto.Field( + resource_status: ResourceStatus = proto.Field( proto.MESSAGE, number=101, message=ResourceStatus, ) - security_status = proto.Field( + security_status: SecurityStatus = proto.Field( proto.MESSAGE, number=103, message=SecurityStatus, ) - discovery_spec = proto.Field( + discovery_spec: DiscoverySpec = proto.Field( proto.MESSAGE, number=106, message=DiscoverySpec, ) - discovery_status = proto.Field( + discovery_status: DiscoveryStatus = proto.Field( proto.MESSAGE, number=107, message=DiscoveryStatus, diff --git a/google/cloud/dataplex_v1/types/service.py b/google/cloud/dataplex_v1/types/service.py index 0ae8601..e968c56 100644 --- a/google/cloud/dataplex_v1/types/service.py +++ b/google/cloud/dataplex_v1/types/service.py @@ -13,6 +13,8 @@ # See the License for the specific language governing permissions and # limitations under the License. # +from typing import MutableMapping, MutableSequence + from google.protobuf import field_mask_pb2 # type: ignore from google.protobuf import timestamp_pb2 # type: ignore import proto # type: ignore @@ -93,20 +95,20 @@ class CreateLakeRequest(proto.Message): not perform mutations. The default is false. """ - parent = proto.Field( + parent: str = proto.Field( proto.STRING, number=1, ) - lake_id = proto.Field( + lake_id: str = proto.Field( proto.STRING, number=2, ) - lake = proto.Field( + lake: resources.Lake = proto.Field( proto.MESSAGE, number=3, message=resources.Lake, ) - validate_only = proto.Field( + validate_only: bool = proto.Field( proto.BOOL, number=4, ) @@ -126,17 +128,17 @@ class UpdateLakeRequest(proto.Message): not perform mutations. The default is false. """ - update_mask = proto.Field( + update_mask: field_mask_pb2.FieldMask = proto.Field( proto.MESSAGE, number=1, message=field_mask_pb2.FieldMask, ) - lake = proto.Field( + lake: resources.Lake = proto.Field( proto.MESSAGE, number=2, message=resources.Lake, ) - validate_only = proto.Field( + validate_only: bool = proto.Field( proto.BOOL, number=3, ) @@ -151,7 +153,7 @@ class DeleteLakeRequest(proto.Message): ``projects/{project_number}/locations/{location_id}/lakes/{lake_id}``. """ - name = proto.Field( + name: str = proto.Field( proto.STRING, number=1, ) @@ -182,23 +184,23 @@ class ListLakesRequest(proto.Message): Optional. Order by fields for the result. """ - parent = proto.Field( + parent: str = proto.Field( proto.STRING, number=1, ) - page_size = proto.Field( + page_size: int = proto.Field( proto.INT32, number=2, ) - page_token = proto.Field( + page_token: str = proto.Field( proto.STRING, number=3, ) - filter = proto.Field( + filter: str = proto.Field( proto.STRING, number=4, ) - order_by = proto.Field( + order_by: str = proto.Field( proto.STRING, number=5, ) @@ -208,13 +210,13 @@ class ListLakesResponse(proto.Message): r"""List lakes response. Attributes: - lakes (Sequence[google.cloud.dataplex_v1.types.Lake]): + lakes (MutableSequence[google.cloud.dataplex_v1.types.Lake]): Lakes under the given parent location. next_page_token (str): Token to retrieve the next page of results, or empty if there are no more results in the list. - unreachable_locations (Sequence[str]): + unreachable_locations (MutableSequence[str]): Locations that could not be reached. """ @@ -222,16 +224,16 @@ class ListLakesResponse(proto.Message): def raw_page(self): return self - lakes = proto.RepeatedField( + lakes: MutableSequence[resources.Lake] = proto.RepeatedField( proto.MESSAGE, number=1, message=resources.Lake, ) - next_page_token = proto.Field( + next_page_token: str = proto.Field( proto.STRING, number=2, ) - unreachable_locations = proto.RepeatedField( + unreachable_locations: MutableSequence[str] = proto.RepeatedField( proto.STRING, number=3, ) @@ -258,15 +260,15 @@ class ListLakeActionsRequest(proto.Message): provided the page token. """ - parent = proto.Field( + parent: str = proto.Field( proto.STRING, number=1, ) - page_size = proto.Field( + page_size: int = proto.Field( proto.INT32, number=2, ) - page_token = proto.Field( + page_token: str = proto.Field( proto.STRING, number=3, ) @@ -276,7 +278,7 @@ class ListActionsResponse(proto.Message): r"""List actions response. Attributes: - actions (Sequence[google.cloud.dataplex_v1.types.Action]): + actions (MutableSequence[google.cloud.dataplex_v1.types.Action]): Actions under the given parent lake/zone/asset. next_page_token (str): @@ -289,12 +291,12 @@ class ListActionsResponse(proto.Message): def raw_page(self): return self - actions = proto.RepeatedField( + actions: MutableSequence[resources.Action] = proto.RepeatedField( proto.MESSAGE, number=1, message=resources.Action, ) - next_page_token = proto.Field( + next_page_token: str = proto.Field( proto.STRING, number=2, ) @@ -309,7 +311,7 @@ class GetLakeRequest(proto.Message): ``projects/{project_number}/locations/{location_id}/lakes/{lake_id}``. """ - name = proto.Field( + name: str = proto.Field( proto.STRING, number=1, ) @@ -342,20 +344,20 @@ class CreateZoneRequest(proto.Message): not perform mutations. The default is false. """ - parent = proto.Field( + parent: str = proto.Field( proto.STRING, number=1, ) - zone_id = proto.Field( + zone_id: str = proto.Field( proto.STRING, number=2, ) - zone = proto.Field( + zone: resources.Zone = proto.Field( proto.MESSAGE, number=3, message=resources.Zone, ) - validate_only = proto.Field( + validate_only: bool = proto.Field( proto.BOOL, number=4, ) @@ -375,17 +377,17 @@ class UpdateZoneRequest(proto.Message): not perform mutations. The default is false. """ - update_mask = proto.Field( + update_mask: field_mask_pb2.FieldMask = proto.Field( proto.MESSAGE, number=1, message=field_mask_pb2.FieldMask, ) - zone = proto.Field( + zone: resources.Zone = proto.Field( proto.MESSAGE, number=2, message=resources.Zone, ) - validate_only = proto.Field( + validate_only: bool = proto.Field( proto.BOOL, number=3, ) @@ -400,7 +402,7 @@ class DeleteZoneRequest(proto.Message): ``projects/{project_number}/locations/{location_id}/lakes/{lake_id}/zones/{zone_id}``. """ - name = proto.Field( + name: str = proto.Field( proto.STRING, number=1, ) @@ -430,23 +432,23 @@ class ListZonesRequest(proto.Message): Optional. Order by fields for the result. """ - parent = proto.Field( + parent: str = proto.Field( proto.STRING, number=1, ) - page_size = proto.Field( + page_size: int = proto.Field( proto.INT32, number=2, ) - page_token = proto.Field( + page_token: str = proto.Field( proto.STRING, number=3, ) - filter = proto.Field( + filter: str = proto.Field( proto.STRING, number=4, ) - order_by = proto.Field( + order_by: str = proto.Field( proto.STRING, number=5, ) @@ -456,7 +458,7 @@ class ListZonesResponse(proto.Message): r"""List zones response. Attributes: - zones (Sequence[google.cloud.dataplex_v1.types.Zone]): + zones (MutableSequence[google.cloud.dataplex_v1.types.Zone]): Zones under the given parent lake. next_page_token (str): Token to retrieve the next page of results, @@ -468,12 +470,12 @@ class ListZonesResponse(proto.Message): def raw_page(self): return self - zones = proto.RepeatedField( + zones: MutableSequence[resources.Zone] = proto.RepeatedField( proto.MESSAGE, number=1, message=resources.Zone, ) - next_page_token = proto.Field( + next_page_token: str = proto.Field( proto.STRING, number=2, ) @@ -500,15 +502,15 @@ class ListZoneActionsRequest(proto.Message): provided the page token. """ - parent = proto.Field( + parent: str = proto.Field( proto.STRING, number=1, ) - page_size = proto.Field( + page_size: int = proto.Field( proto.INT32, number=2, ) - page_token = proto.Field( + page_token: str = proto.Field( proto.STRING, number=3, ) @@ -523,7 +525,7 @@ class GetZoneRequest(proto.Message): ``projects/{project_number}/locations/{location_id}/lakes/{lake_id}/zones/{zone_id}``. """ - name = proto.Field( + name: str = proto.Field( proto.STRING, number=1, ) @@ -553,20 +555,20 @@ class CreateAssetRequest(proto.Message): not perform mutations. The default is false. """ - parent = proto.Field( + parent: str = proto.Field( proto.STRING, number=1, ) - asset_id = proto.Field( + asset_id: str = proto.Field( proto.STRING, number=2, ) - asset = proto.Field( + asset: resources.Asset = proto.Field( proto.MESSAGE, number=3, message=resources.Asset, ) - validate_only = proto.Field( + validate_only: bool = proto.Field( proto.BOOL, number=4, ) @@ -586,17 +588,17 @@ class UpdateAssetRequest(proto.Message): not perform mutations. The default is false. """ - update_mask = proto.Field( + update_mask: field_mask_pb2.FieldMask = proto.Field( proto.MESSAGE, number=1, message=field_mask_pb2.FieldMask, ) - asset = proto.Field( + asset: resources.Asset = proto.Field( proto.MESSAGE, number=2, message=resources.Asset, ) - validate_only = proto.Field( + validate_only: bool = proto.Field( proto.BOOL, number=3, ) @@ -611,7 +613,7 @@ class DeleteAssetRequest(proto.Message): ``projects/{project_number}/locations/{location_id}/lakes/{lake_id}/zones/{zone_id}/assets/{asset_id}``. """ - name = proto.Field( + name: str = proto.Field( proto.STRING, number=1, ) @@ -641,23 +643,23 @@ class ListAssetsRequest(proto.Message): Optional. Order by fields for the result. """ - parent = proto.Field( + parent: str = proto.Field( proto.STRING, number=1, ) - page_size = proto.Field( + page_size: int = proto.Field( proto.INT32, number=2, ) - page_token = proto.Field( + page_token: str = proto.Field( proto.STRING, number=3, ) - filter = proto.Field( + filter: str = proto.Field( proto.STRING, number=4, ) - order_by = proto.Field( + order_by: str = proto.Field( proto.STRING, number=5, ) @@ -667,7 +669,7 @@ class ListAssetsResponse(proto.Message): r"""List assets response. Attributes: - assets (Sequence[google.cloud.dataplex_v1.types.Asset]): + assets (MutableSequence[google.cloud.dataplex_v1.types.Asset]): Asset under the given parent zone. next_page_token (str): Token to retrieve the next page of results, @@ -679,12 +681,12 @@ class ListAssetsResponse(proto.Message): def raw_page(self): return self - assets = proto.RepeatedField( + assets: MutableSequence[resources.Asset] = proto.RepeatedField( proto.MESSAGE, number=1, message=resources.Asset, ) - next_page_token = proto.Field( + next_page_token: str = proto.Field( proto.STRING, number=2, ) @@ -711,15 +713,15 @@ class ListAssetActionsRequest(proto.Message): provided the page token. """ - parent = proto.Field( + parent: str = proto.Field( proto.STRING, number=1, ) - page_size = proto.Field( + page_size: int = proto.Field( proto.INT32, number=2, ) - page_token = proto.Field( + page_token: str = proto.Field( proto.STRING, number=3, ) @@ -734,7 +736,7 @@ class GetAssetRequest(proto.Message): ``projects/{project_number}/locations/{location_id}/lakes/{lake_id}/zones/{zone_id}/assets/{asset_id}``. """ - name = proto.Field( + name: str = proto.Field( proto.STRING, number=1, ) @@ -770,33 +772,33 @@ class OperationMetadata(proto.Message): operation. """ - create_time = proto.Field( + create_time: timestamp_pb2.Timestamp = proto.Field( proto.MESSAGE, number=1, message=timestamp_pb2.Timestamp, ) - end_time = proto.Field( + end_time: timestamp_pb2.Timestamp = proto.Field( proto.MESSAGE, number=2, message=timestamp_pb2.Timestamp, ) - target = proto.Field( + target: str = proto.Field( proto.STRING, number=3, ) - verb = proto.Field( + verb: str = proto.Field( proto.STRING, number=4, ) - status_message = proto.Field( + status_message: str = proto.Field( proto.STRING, number=5, ) - requested_cancellation = proto.Field( + requested_cancellation: bool = proto.Field( proto.BOOL, number=6, ) - api_version = proto.Field( + api_version: str = proto.Field( proto.STRING, number=7, ) @@ -818,20 +820,20 @@ class CreateTaskRequest(proto.Message): not perform mutations. The default is false. """ - parent = proto.Field( + parent: str = proto.Field( proto.STRING, number=1, ) - task_id = proto.Field( + task_id: str = proto.Field( proto.STRING, number=2, ) - task = proto.Field( + task: gcd_tasks.Task = proto.Field( proto.MESSAGE, number=3, message=gcd_tasks.Task, ) - validate_only = proto.Field( + validate_only: bool = proto.Field( proto.BOOL, number=4, ) @@ -851,17 +853,17 @@ class UpdateTaskRequest(proto.Message): not perform mutations. The default is false. """ - update_mask = proto.Field( + update_mask: field_mask_pb2.FieldMask = proto.Field( proto.MESSAGE, number=1, message=field_mask_pb2.FieldMask, ) - task = proto.Field( + task: gcd_tasks.Task = proto.Field( proto.MESSAGE, number=2, message=gcd_tasks.Task, ) - validate_only = proto.Field( + validate_only: bool = proto.Field( proto.BOOL, number=3, ) @@ -876,7 +878,7 @@ class DeleteTaskRequest(proto.Message): ``projects/{project_number}/locations/{location_id}/lakes/{lake_id}/task/{task_id}``. """ - name = proto.Field( + name: str = proto.Field( proto.STRING, number=1, ) @@ -906,23 +908,23 @@ class ListTasksRequest(proto.Message): Optional. Order by fields for the result. """ - parent = proto.Field( + parent: str = proto.Field( proto.STRING, number=1, ) - page_size = proto.Field( + page_size: int = proto.Field( proto.INT32, number=2, ) - page_token = proto.Field( + page_token: str = proto.Field( proto.STRING, number=3, ) - filter = proto.Field( + filter: str = proto.Field( proto.STRING, number=4, ) - order_by = proto.Field( + order_by: str = proto.Field( proto.STRING, number=5, ) @@ -932,13 +934,13 @@ class ListTasksResponse(proto.Message): r"""List tasks response. Attributes: - tasks (Sequence[google.cloud.dataplex_v1.types.Task]): + tasks (MutableSequence[google.cloud.dataplex_v1.types.Task]): Tasks under the given parent lake. next_page_token (str): Token to retrieve the next page of results, or empty if there are no more results in the list. - unreachable_locations (Sequence[str]): + unreachable_locations (MutableSequence[str]): Locations that could not be reached. """ @@ -946,16 +948,16 @@ class ListTasksResponse(proto.Message): def raw_page(self): return self - tasks = proto.RepeatedField( + tasks: MutableSequence[gcd_tasks.Task] = proto.RepeatedField( proto.MESSAGE, number=1, message=gcd_tasks.Task, ) - next_page_token = proto.Field( + next_page_token: str = proto.Field( proto.STRING, number=2, ) - unreachable_locations = proto.RepeatedField( + unreachable_locations: MutableSequence[str] = proto.RepeatedField( proto.STRING, number=3, ) @@ -970,7 +972,7 @@ class GetTaskRequest(proto.Message): ``projects/{project_number}/locations/{location_id}/lakes/{lake_id}/tasks/{tasks_id}``. """ - name = proto.Field( + name: str = proto.Field( proto.STRING, number=1, ) @@ -985,7 +987,7 @@ class GetJobRequest(proto.Message): ``projects/{project_number}/locations/{location_id}/lakes/{lake_id}/tasks/{task_id}/jobs/{job_id}``. """ - name = proto.Field( + name: str = proto.Field( proto.STRING, number=1, ) @@ -1011,15 +1013,15 @@ class ListJobsRequest(proto.Message): must match the call that provided the page token. """ - parent = proto.Field( + parent: str = proto.Field( proto.STRING, number=1, ) - page_size = proto.Field( + page_size: int = proto.Field( proto.INT32, number=2, ) - page_token = proto.Field( + page_token: str = proto.Field( proto.STRING, number=3, ) @@ -1029,7 +1031,7 @@ class ListJobsResponse(proto.Message): r"""List jobs response. Attributes: - jobs (Sequence[google.cloud.dataplex_v1.types.Job]): + jobs (MutableSequence[google.cloud.dataplex_v1.types.Job]): Jobs under a given task. next_page_token (str): Token to retrieve the next page of results, @@ -1041,12 +1043,12 @@ class ListJobsResponse(proto.Message): def raw_page(self): return self - jobs = proto.RepeatedField( + jobs: MutableSequence[gcd_tasks.Job] = proto.RepeatedField( proto.MESSAGE, number=1, message=gcd_tasks.Job, ) - next_page_token = proto.Field( + next_page_token: str = proto.Field( proto.STRING, number=2, ) @@ -1061,7 +1063,7 @@ class CancelJobRequest(proto.Message): ``projects/{project_number}/locations/{location_id}/lakes/{lake_id}/task/{task_id}/job/{job_id}``. """ - name = proto.Field( + name: str = proto.Field( proto.STRING, number=1, ) @@ -1089,20 +1091,20 @@ class CreateEnvironmentRequest(proto.Message): not perform mutations. The default is false. """ - parent = proto.Field( + parent: str = proto.Field( proto.STRING, number=1, ) - environment_id = proto.Field( + environment_id: str = proto.Field( proto.STRING, number=2, ) - environment = proto.Field( + environment: analyze.Environment = proto.Field( proto.MESSAGE, number=3, message=analyze.Environment, ) - validate_only = proto.Field( + validate_only: bool = proto.Field( proto.BOOL, number=4, ) @@ -1122,17 +1124,17 @@ class UpdateEnvironmentRequest(proto.Message): not perform mutations. The default is false. """ - update_mask = proto.Field( + update_mask: field_mask_pb2.FieldMask = proto.Field( proto.MESSAGE, number=1, message=field_mask_pb2.FieldMask, ) - environment = proto.Field( + environment: analyze.Environment = proto.Field( proto.MESSAGE, number=2, message=analyze.Environment, ) - validate_only = proto.Field( + validate_only: bool = proto.Field( proto.BOOL, number=3, ) @@ -1147,7 +1149,7 @@ class DeleteEnvironmentRequest(proto.Message): ``projects/{project_id}/locations/{location_id}/lakes/{lake_id}/environments/{environment_id}``. """ - name = proto.Field( + name: str = proto.Field( proto.STRING, number=1, ) @@ -1178,23 +1180,23 @@ class ListEnvironmentsRequest(proto.Message): Optional. Order by fields for the result. """ - parent = proto.Field( + parent: str = proto.Field( proto.STRING, number=1, ) - page_size = proto.Field( + page_size: int = proto.Field( proto.INT32, number=2, ) - page_token = proto.Field( + page_token: str = proto.Field( proto.STRING, number=3, ) - filter = proto.Field( + filter: str = proto.Field( proto.STRING, number=4, ) - order_by = proto.Field( + order_by: str = proto.Field( proto.STRING, number=5, ) @@ -1204,7 +1206,7 @@ class ListEnvironmentsResponse(proto.Message): r"""List environments response. Attributes: - environments (Sequence[google.cloud.dataplex_v1.types.Environment]): + environments (MutableSequence[google.cloud.dataplex_v1.types.Environment]): Environments under the given parent lake. next_page_token (str): Token to retrieve the next page of results, @@ -1216,12 +1218,12 @@ class ListEnvironmentsResponse(proto.Message): def raw_page(self): return self - environments = proto.RepeatedField( + environments: MutableSequence[analyze.Environment] = proto.RepeatedField( proto.MESSAGE, number=1, message=analyze.Environment, ) - next_page_token = proto.Field( + next_page_token: str = proto.Field( proto.STRING, number=2, ) @@ -1236,7 +1238,7 @@ class GetEnvironmentRequest(proto.Message): ``projects/{project_id}/locations/{location_id}/lakes/{lake_id}/environments/{environment_id}``. """ - name = proto.Field( + name: str = proto.Field( proto.STRING, number=1, ) @@ -1275,19 +1277,19 @@ class ListSessionsRequest(proto.Message): mode = ADMIN \| USER """ - parent = proto.Field( + parent: str = proto.Field( proto.STRING, number=1, ) - page_size = proto.Field( + page_size: int = proto.Field( proto.INT32, number=2, ) - page_token = proto.Field( + page_token: str = proto.Field( proto.STRING, number=3, ) - filter = proto.Field( + filter: str = proto.Field( proto.STRING, number=4, ) @@ -1297,7 +1299,7 @@ class ListSessionsResponse(proto.Message): r"""List sessions response. Attributes: - sessions (Sequence[google.cloud.dataplex_v1.types.Session]): + sessions (MutableSequence[google.cloud.dataplex_v1.types.Session]): Sessions under a given environment. next_page_token (str): Token to retrieve the next page of results, @@ -1309,12 +1311,12 @@ class ListSessionsResponse(proto.Message): def raw_page(self): return self - sessions = proto.RepeatedField( + sessions: MutableSequence[analyze.Session] = proto.RepeatedField( proto.MESSAGE, number=1, message=analyze.Session, ) - next_page_token = proto.Field( + next_page_token: str = proto.Field( proto.STRING, number=2, ) diff --git a/google/cloud/dataplex_v1/types/tasks.py b/google/cloud/dataplex_v1/types/tasks.py index 288ba38..0f09b9b 100644 --- a/google/cloud/dataplex_v1/types/tasks.py +++ b/google/cloud/dataplex_v1/types/tasks.py @@ -13,6 +13,8 @@ # See the License for the specific language governing permissions and # limitations under the License. # +from typing import MutableMapping, MutableSequence + from google.protobuf import duration_pb2 # type: ignore from google.protobuf import timestamp_pb2 # type: ignore import proto # type: ignore @@ -61,7 +63,7 @@ class Task(proto.Message): Optional. User friendly display name. state (google.cloud.dataplex_v1.types.State): Output only. Current state of the task. - labels (Mapping[str, str]): + labels (MutableMapping[str, str]): Optional. User-defined labels for the task. trigger_spec (google.cloud.dataplex_v1.types.Task.TriggerSpec): Required. Spec related to how often and when @@ -119,11 +121,11 @@ class BatchComputeResources(proto.Message): Executor Count should be between 2 and 1000. [Default=1000] """ - executors_count = proto.Field( + executors_count: int = proto.Field( proto.INT32, number=1, ) - max_executors_count = proto.Field( + max_executors_count: int = proto.Field( proto.INT32, number=2, ) @@ -135,17 +137,17 @@ class ContainerImageRuntime(proto.Message): Attributes: image (str): Optional. Container image to use. - java_jars (Sequence[str]): + java_jars (MutableSequence[str]): Optional. A list of Java JARS to add to the classpath. Valid input includes Cloud Storage URIs to Jar binaries. For example, gs://bucket-name/my/path/to/file.jar - python_packages (Sequence[str]): + python_packages (MutableSequence[str]): Optional. A list of python packages to be installed. Valid formats include Cloud Storage URI to a PIP installable library. For example, gs://bucket-name/my/path/to/lib.tar.gz - properties (Mapping[str, str]): + properties (MutableMapping[str, str]): Optional. Override to common configuration of open source components installed on the Dataproc cluster. The properties to set on daemon config files. Property keys are specified @@ -154,19 +156,19 @@ class ContainerImageRuntime(proto.Message): properties `__. """ - image = proto.Field( + image: str = proto.Field( proto.STRING, number=1, ) - java_jars = proto.RepeatedField( + java_jars: MutableSequence[str] = proto.RepeatedField( proto.STRING, number=2, ) - python_packages = proto.RepeatedField( + python_packages: MutableSequence[str] = proto.RepeatedField( proto.STRING, number=3, ) - properties = proto.MapField( + properties: MutableMapping[str, str] = proto.MapField( proto.STRING, proto.STRING, number=4, @@ -194,39 +196,39 @@ class VpcNetwork(proto.Message): the job is run. This field is a member of `oneof`_ ``network_name``. - network_tags (Sequence[str]): + network_tags (MutableSequence[str]): Optional. List of network tags to apply to the job. """ - network = proto.Field( + network: str = proto.Field( proto.STRING, number=1, oneof="network_name", ) - sub_network = proto.Field( + sub_network: str = proto.Field( proto.STRING, number=2, oneof="network_name", ) - network_tags = proto.RepeatedField( + network_tags: MutableSequence[str] = proto.RepeatedField( proto.STRING, number=3, ) - batch = proto.Field( + batch: "Task.InfrastructureSpec.BatchComputeResources" = proto.Field( proto.MESSAGE, number=52, oneof="resources", message="Task.InfrastructureSpec.BatchComputeResources", ) - container_image = proto.Field( + container_image: "Task.InfrastructureSpec.ContainerImageRuntime" = proto.Field( proto.MESSAGE, number=101, oneof="runtime", message="Task.InfrastructureSpec.ContainerImageRuntime", ) - vpc_network = proto.Field( + vpc_network: "Task.InfrastructureSpec.VpcNetwork" = proto.Field( proto.MESSAGE, number=150, oneof="network", @@ -275,25 +277,25 @@ class Type(proto.Enum): ON_DEMAND = 1 RECURRING = 2 - type_ = proto.Field( + type_: "Task.TriggerSpec.Type" = proto.Field( proto.ENUM, number=5, enum="Task.TriggerSpec.Type", ) - start_time = proto.Field( + start_time: timestamp_pb2.Timestamp = proto.Field( proto.MESSAGE, number=6, message=timestamp_pb2.Timestamp, ) - disabled = proto.Field( + disabled: bool = proto.Field( proto.BOOL, number=4, ) - max_retries = proto.Field( + max_retries: int = proto.Field( proto.INT32, number=7, ) - schedule = proto.Field( + schedule: str = proto.Field( proto.STRING, number=100, oneof="trigger", @@ -303,7 +305,7 @@ class ExecutionSpec(proto.Message): r"""Execution related settings, like retry and service_account. Attributes: - args (Mapping[str, str]): + args (MutableMapping[str, str]): Optional. The arguments to pass to the task. The args can use placeholders of the format ${placeholder} as part of key/value string. These will be interpolated before passing @@ -336,25 +338,25 @@ class ExecutionSpec(proto.Message): ``projects/{project_number}/locations/{location_id}/keyRings/{key-ring-name}/cryptoKeys/{key-name}``. """ - args = proto.MapField( + args: MutableMapping[str, str] = proto.MapField( proto.STRING, proto.STRING, number=4, ) - service_account = proto.Field( + service_account: str = proto.Field( proto.STRING, number=5, ) - project = proto.Field( + project: str = proto.Field( proto.STRING, number=7, ) - max_job_execution_lifetime = proto.Field( + max_job_execution_lifetime: duration_pb2.Duration = proto.Field( proto.MESSAGE, number=8, message=duration_pb2.Duration, ) - kms_key = proto.Field( + kms_key: str = proto.Field( proto.STRING, number=9, ) @@ -402,11 +404,11 @@ class SparkTaskConfig(proto.Message): of script variables (``set key="value";``). This field is a member of `oneof`_ ``driver``. - file_uris (Sequence[str]): + file_uris (MutableSequence[str]): Optional. Cloud Storage URIs of files to be placed in the working directory of each executor. - archive_uris (Sequence[str]): + archive_uris (MutableSequence[str]): Optional. Cloud Storage URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, @@ -416,40 +418,40 @@ class SparkTaskConfig(proto.Message): the execution. """ - main_jar_file_uri = proto.Field( + main_jar_file_uri: str = proto.Field( proto.STRING, number=100, oneof="driver", ) - main_class = proto.Field( + main_class: str = proto.Field( proto.STRING, number=101, oneof="driver", ) - python_script_file = proto.Field( + python_script_file: str = proto.Field( proto.STRING, number=102, oneof="driver", ) - sql_script_file = proto.Field( + sql_script_file: str = proto.Field( proto.STRING, number=104, oneof="driver", ) - sql_script = proto.Field( + sql_script: str = proto.Field( proto.STRING, number=105, oneof="driver", ) - file_uris = proto.RepeatedField( + file_uris: MutableSequence[str] = proto.RepeatedField( proto.STRING, number=3, ) - archive_uris = proto.RepeatedField( + archive_uris: MutableSequence[str] = proto.RepeatedField( proto.STRING, number=4, ) - infrastructure_spec = proto.Field( + infrastructure_spec: "Task.InfrastructureSpec" = proto.Field( proto.MESSAGE, number=6, message="Task.InfrastructureSpec", @@ -467,31 +469,31 @@ class NotebookTaskConfig(proto.Message): infrastructure_spec (google.cloud.dataplex_v1.types.Task.InfrastructureSpec): Optional. Infrastructure specification for the execution. - file_uris (Sequence[str]): + file_uris (MutableSequence[str]): Optional. Cloud Storage URIs of files to be placed in the working directory of each executor. - archive_uris (Sequence[str]): + archive_uris (MutableSequence[str]): Optional. Cloud Storage URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip. """ - notebook = proto.Field( + notebook: str = proto.Field( proto.STRING, number=4, ) - infrastructure_spec = proto.Field( + infrastructure_spec: "Task.InfrastructureSpec" = proto.Field( proto.MESSAGE, number=3, message="Task.InfrastructureSpec", ) - file_uris = proto.RepeatedField( + file_uris: MutableSequence[str] = proto.RepeatedField( proto.STRING, number=5, ) - archive_uris = proto.RepeatedField( + archive_uris: MutableSequence[str] = proto.RepeatedField( proto.STRING, number=6, ) @@ -506,75 +508,75 @@ class ExecutionStatus(proto.Message): Output only. latest job execution """ - update_time = proto.Field( + update_time: timestamp_pb2.Timestamp = proto.Field( proto.MESSAGE, number=3, message=timestamp_pb2.Timestamp, ) - latest_job = proto.Field( + latest_job: "Job" = proto.Field( proto.MESSAGE, number=9, message="Job", ) - name = proto.Field( + name: str = proto.Field( proto.STRING, number=1, ) - uid = proto.Field( + uid: str = proto.Field( proto.STRING, number=2, ) - create_time = proto.Field( + create_time: timestamp_pb2.Timestamp = proto.Field( proto.MESSAGE, number=3, message=timestamp_pb2.Timestamp, ) - update_time = proto.Field( + update_time: timestamp_pb2.Timestamp = proto.Field( proto.MESSAGE, number=4, message=timestamp_pb2.Timestamp, ) - description = proto.Field( + description: str = proto.Field( proto.STRING, number=5, ) - display_name = proto.Field( + display_name: str = proto.Field( proto.STRING, number=6, ) - state = proto.Field( + state: resources.State = proto.Field( proto.ENUM, number=7, enum=resources.State, ) - labels = proto.MapField( + labels: MutableMapping[str, str] = proto.MapField( proto.STRING, proto.STRING, number=8, ) - trigger_spec = proto.Field( + trigger_spec: TriggerSpec = proto.Field( proto.MESSAGE, number=100, message=TriggerSpec, ) - execution_spec = proto.Field( + execution_spec: ExecutionSpec = proto.Field( proto.MESSAGE, number=101, message=ExecutionSpec, ) - execution_status = proto.Field( + execution_status: ExecutionStatus = proto.Field( proto.MESSAGE, number=201, message=ExecutionStatus, ) - spark = proto.Field( + spark: SparkTaskConfig = proto.Field( proto.MESSAGE, number=300, oneof="config", message=SparkTaskConfig, ) - notebook = proto.Field( + notebook: NotebookTaskConfig = proto.Field( proto.MESSAGE, number=302, oneof="config", @@ -629,43 +631,43 @@ class State(proto.Enum): FAILED = 5 ABORTED = 6 - name = proto.Field( + name: str = proto.Field( proto.STRING, number=1, ) - uid = proto.Field( + uid: str = proto.Field( proto.STRING, number=2, ) - start_time = proto.Field( + start_time: timestamp_pb2.Timestamp = proto.Field( proto.MESSAGE, number=3, message=timestamp_pb2.Timestamp, ) - end_time = proto.Field( + end_time: timestamp_pb2.Timestamp = proto.Field( proto.MESSAGE, number=4, message=timestamp_pb2.Timestamp, ) - state = proto.Field( + state: State = proto.Field( proto.ENUM, number=5, enum=State, ) - retry_count = proto.Field( + retry_count: int = proto.Field( proto.UINT32, number=6, ) - service = proto.Field( + service: Service = proto.Field( proto.ENUM, number=7, enum=Service, ) - service_job = proto.Field( + service_job: str = proto.Field( proto.STRING, number=8, ) - message = proto.Field( + message: str = proto.Field( proto.STRING, number=9, ) diff --git a/noxfile.py b/noxfile.py index 30133c9..d8440c0 100644 --- a/noxfile.py +++ b/noxfile.py @@ -268,7 +268,7 @@ def cover(session): test runs (not system test runs), and then erases coverage data. """ session.install("coverage", "pytest-cov") - session.run("coverage", "report", "--show-missing", "--fail-under=99") + session.run("coverage", "report", "--show-missing", "--fail-under=100") session.run("coverage", "erase") diff --git a/owlbot.py b/owlbot.py index e738816..908909d 100644 --- a/owlbot.py +++ b/owlbot.py @@ -12,7 +12,9 @@ # See the License for the specific language governing permissions and # limitations under the License. +import json from pathlib import Path +import shutil import synthtool as s import synthtool.gcp as gcp @@ -22,10 +24,19 @@ # Copy the generated client from the owl-bot staging directory # ---------------------------------------------------------------------------- -default_version = "v1" +clean_up_generated_samples = True + +# Load the default version defined in .repo-metadata.json. +default_version = json.load(open(".repo-metadata.json", "rt")).get( + "default_version" +) for library in s.get_staging_dirs(default_version): - s.move(library, excludes=["setup.py"]) + if clean_up_generated_samples: + shutil.rmtree("samples/generated_samples", ignore_errors=True) + clean_up_generated_samples = False + + s.move([library], excludes=["**/gapic_version.py"]) s.remove_staging_dirs() # ---------------------------------------------------------------------------- @@ -33,17 +44,14 @@ # ---------------------------------------------------------------------------- templated_files = gcp.CommonTemplates().py_library( + cov_level=100, microgenerator=True, versions=gcp.common.detect_versions(path="./google", default_first=True), - cov_level=99, ) -s.move(templated_files, excludes=[".coveragerc"]) # the microgenerator has a good coveragerc file - -python.configure_previous_major_version_branches() +s.move(templated_files, excludes=[".coveragerc", ".github/release-please.yml"]) python.py_samples(skip_readmes=True) - # run format session for all directories which have a noxfile for noxfile in Path(".").glob("**/noxfile.py"): s.shell.run(["nox", "-s", "format"], cwd=noxfile.parent, hide_output=False) diff --git a/release-please-config.json b/release-please-config.json new file mode 100644 index 0000000..0db768f --- /dev/null +++ b/release-please-config.json @@ -0,0 +1,23 @@ +{ + "$schema": "https://raw.githubusercontent.com/googleapis/release-please/main/schemas/config.json", + "packages": { + ".": { + "release-type": "python", + "extra-files": [ + "google/cloud/dataplex/gapic_version.py", + { + "type": "json", + "path": "samples/generated_samples/snippet_metadata_google.cloud.dataplex.v1.json", + "jsonpath": "$.clientLibrary.version" + } + ] + } + }, + "release-type": "python", + "plugins": [ + { + "type": "sentence-case" + } + ], + "initial-version": "0.1.0" +} diff --git a/samples/generated_samples/snippet_metadata_dataplex_v1.json b/samples/generated_samples/snippet_metadata_google.cloud.dataplex.v1.json similarity index 99% rename from samples/generated_samples/snippet_metadata_dataplex_v1.json rename to samples/generated_samples/snippet_metadata_google.cloud.dataplex.v1.json index dfad3b6..370f7c5 100644 --- a/samples/generated_samples/snippet_metadata_dataplex_v1.json +++ b/samples/generated_samples/snippet_metadata_google.cloud.dataplex.v1.json @@ -7,7 +7,8 @@ } ], "language": "PYTHON", - "name": "google-cloud-dataplex" + "name": "google-cloud-dataplex", + "version": "0.1.0" }, "snippets": [ { diff --git a/setup.py b/setup.py index 3fc72c3..39fac0f 100644 --- a/setup.py +++ b/setup.py @@ -1,35 +1,47 @@ # -*- coding: utf-8 -*- -# -# Copyright 2020 Google LLC +# Copyright 2022 Google LLC # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # -# https://www.apache.org/licenses/LICENSE-2.0 +# http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. - +# import io import os -import setuptools +import setuptools # type: ignore + +package_root = os.path.abspath(os.path.dirname(__file__)) name = "google-cloud-dataplex" -description = "Dataplex API client library" -url = "https://github.com/googleapis/python-dataplex" -version = "1.1.2" -release_status = "Development Status :: 5 - Production/Stable" + + +description = "Google Cloud Dataplex API client library" + +version = {} +with open(os.path.join(package_root, "google/cloud/dataplex/gapic_version.py")) as fp: + exec(fp.read(), version) +version = version["__version__"] + +if version[0] == "0": + release_status = "Development Status :: 4 - Beta" +else: + release_status = "Development Status :: 5 - Production/Stable" + dependencies = [ - "google-api-core[grpc] >= 1.32.0, <3.0.0dev,!=2.0.*,!=2.1.*,!=2.2.*,!=2.3.*,!=2.4.*,!=2.5.*,!=2.6.*,!=2.7.*", + "google-api-core[grpc] >= 1.33.2, <3.0.0dev,!=2.0.*,!=2.1.*,!=2.2.*,!=2.3.*,!=2.4.*,!=2.5.*,!=2.6.*,!=2.7.*", "proto-plus >= 1.22.0, <2.0.0dev", "protobuf>=3.19.5,<5.0.0dev,!=3.20.0,!=3.20.1,!=4.21.0,!=4.21.1,!=4.21.2,!=4.21.3,!=4.21.4,!=4.21.5", - "grpc-google-iam-v1 >=0.12.4, <1.0.0dev", + "grpc-google-iam-v1 >= 0.12.4, < 1.0.0dev", ] +url = "https://github.com/googleapis/python-dataplex" package_root = os.path.abspath(os.path.dirname(__file__)) @@ -47,7 +59,6 @@ if "google.cloud" in packages: namespaces.append("google.cloud") - setuptools.setup( name=name, version=version, diff --git a/testing/constraints-3.10.txt b/testing/constraints-3.10.txt index e69de29..ad3f0fa 100644 --- a/testing/constraints-3.10.txt +++ b/testing/constraints-3.10.txt @@ -0,0 +1,7 @@ +# -*- coding: utf-8 -*- +# This constraints file is required for unit tests. +# List all library dependencies and extras in this file. +google-api-core +proto-plus +protobuf +grpc-google-iam-v1 diff --git a/testing/constraints-3.11.txt b/testing/constraints-3.11.txt index e69de29..ad3f0fa 100644 --- a/testing/constraints-3.11.txt +++ b/testing/constraints-3.11.txt @@ -0,0 +1,7 @@ +# -*- coding: utf-8 -*- +# This constraints file is required for unit tests. +# List all library dependencies and extras in this file. +google-api-core +proto-plus +protobuf +grpc-google-iam-v1 diff --git a/testing/constraints-3.7.txt b/testing/constraints-3.7.txt index b0cdb69..883a87a 100644 --- a/testing/constraints-3.7.txt +++ b/testing/constraints-3.7.txt @@ -4,7 +4,7 @@ # Pin the version to the lower bound. # e.g., if setup.py has "google-cloud-foo >= 1.14.0, < 2.0.0dev", # Then this file should have google-cloud-foo==1.14.0 -google-api-core==1.32.0 +google-api-core==1.33.2 proto-plus==1.22.0 protobuf==3.19.5 grpc-google-iam-v1==0.12.4 diff --git a/testing/constraints-3.8.txt b/testing/constraints-3.8.txt index e69de29..ad3f0fa 100644 --- a/testing/constraints-3.8.txt +++ b/testing/constraints-3.8.txt @@ -0,0 +1,7 @@ +# -*- coding: utf-8 -*- +# This constraints file is required for unit tests. +# List all library dependencies and extras in this file. +google-api-core +proto-plus +protobuf +grpc-google-iam-v1 diff --git a/testing/constraints-3.9.txt b/testing/constraints-3.9.txt index e69de29..ad3f0fa 100644 --- a/testing/constraints-3.9.txt +++ b/testing/constraints-3.9.txt @@ -0,0 +1,7 @@ +# -*- coding: utf-8 -*- +# This constraints file is required for unit tests. +# List all library dependencies and extras in this file. +google-api-core +proto-plus +protobuf +grpc-google-iam-v1 diff --git a/tests/unit/gapic/dataplex_v1/test_dataplex_service.py b/tests/unit/gapic/dataplex_v1/test_dataplex_service.py index 8843469..3c48d9f 100644 --- a/tests/unit/gapic/dataplex_v1/test_dataplex_service.py +++ b/tests/unit/gapic/dataplex_v1/test_dataplex_service.py @@ -46,6 +46,7 @@ from google.longrunning import operations_pb2 from google.oauth2 import service_account from google.protobuf import duration_pb2 # type: ignore +from google.protobuf import empty_pb2 # type: ignore from google.protobuf import field_mask_pb2 # type: ignore from google.protobuf import timestamp_pb2 # type: ignore import grpc From 1e94a024024638d5d7d31f7bba4408b3f0b3d5d1 Mon Sep 17 00:00:00 2001 From: "gcf-owl-bot[bot]" <78513119+gcf-owl-bot[bot]@users.noreply.github.com> Date: Wed, 7 Dec 2022 10:58:34 -0500 Subject: [PATCH 07/10] fix(deps): Require google-api-core >=1.34.0, >=2.11.0 (#111) MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit * fix(deps): Require google-api-core >=1.34.0, >=2.11.0 fix: Drop usage of pkg_resources fix: Fix timeout default values docs(samples): Snippetgen should call await on the operation coroutine before calling result PiperOrigin-RevId: 493260409 Source-Link: https://github.com/googleapis/googleapis/commit/fea43879f83a8d0dacc9353b3f75f8f46d37162f Source-Link: https://github.com/googleapis/googleapis-gen/commit/387b7344c7529ee44be84e613b19a820508c612b Copy-Tag: eyJwIjoiLmdpdGh1Yi8uT3dsQm90LnlhbWwiLCJoIjoiMzg3YjczNDRjNzUyOWVlNDRiZTg0ZTYxM2IxOWE4MjA1MDhjNjEyYiJ9 * 🦉 Updates from OwlBot post-processor See https://github.com/googleapis/repo-automation-bots/blob/main/packages/owl-bot/README.md * add gapic_version.py Co-authored-by: Owl Bot Co-authored-by: Anthonios Partheniou --- .coveragerc | 5 - google/cloud/dataplex_v1/gapic_version.py | 16 +++ .../services/content_service/async_client.py | 42 +++--- .../services/content_service/client.py | 42 +++--- .../content_service/transports/base.py | 13 +- .../services/dataplex_service/async_client.py | 120 +++++++++--------- .../services/dataplex_service/client.py | 90 +++++++------ .../dataplex_service/transports/base.py | 13 +- .../services/metadata_service/async_client.py | 44 +++---- .../services/metadata_service/client.py | 44 +++---- .../metadata_service/transports/base.py | 13 +- release-please-config.json | 1 + ...ted_dataplex_service_create_asset_async.py | 2 +- ...taplex_service_create_environment_async.py | 2 +- ...ated_dataplex_service_create_lake_async.py | 2 +- ...ated_dataplex_service_create_task_async.py | 2 +- ...ated_dataplex_service_create_zone_async.py | 2 +- ...ted_dataplex_service_delete_asset_async.py | 2 +- ...taplex_service_delete_environment_async.py | 2 +- ...ated_dataplex_service_delete_lake_async.py | 2 +- ...ated_dataplex_service_delete_task_async.py | 2 +- ...ated_dataplex_service_delete_zone_async.py | 2 +- ...ted_dataplex_service_update_asset_async.py | 2 +- ...taplex_service_update_environment_async.py | 2 +- ...ated_dataplex_service_update_lake_async.py | 2 +- ...ated_dataplex_service_update_task_async.py | 2 +- ...ated_dataplex_service_update_zone_async.py | 2 +- setup.py | 2 +- testing/constraints-3.7.txt | 2 +- 29 files changed, 225 insertions(+), 252 deletions(-) create mode 100644 google/cloud/dataplex_v1/gapic_version.py diff --git a/.coveragerc b/.coveragerc index 0a5725d..662231b 100644 --- a/.coveragerc +++ b/.coveragerc @@ -10,8 +10,3 @@ exclude_lines = pragma: NO COVER # Ignore debug-only repr def __repr__ - # Ignore pkg_resources exceptions. - # This is added at the module level as a safeguard for if someone - # generates the code and tries to run it without pip installing. This - # makes it virtually impossible to test properly. - except pkg_resources.DistributionNotFound diff --git a/google/cloud/dataplex_v1/gapic_version.py b/google/cloud/dataplex_v1/gapic_version.py new file mode 100644 index 0000000..19bb9db --- /dev/null +++ b/google/cloud/dataplex_v1/gapic_version.py @@ -0,0 +1,16 @@ +# -*- coding: utf-8 -*- +# Copyright 2022 Google LLC +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +# +__version__ = "1.1.2" # {x-release-please-version} diff --git a/google/cloud/dataplex_v1/services/content_service/async_client.py b/google/cloud/dataplex_v1/services/content_service/async_client.py index 1a4dcc8..bda519c 100644 --- a/google/cloud/dataplex_v1/services/content_service/async_client.py +++ b/google/cloud/dataplex_v1/services/content_service/async_client.py @@ -34,7 +34,8 @@ from google.api_core.client_options import ClientOptions from google.auth import credentials as ga_credentials # type: ignore from google.oauth2 import service_account # type: ignore -import pkg_resources + +from google.cloud.dataplex_v1 import gapic_version as package_version try: OptionalRetry = Union[retries.Retry, gapic_v1.method._MethodDefault] @@ -231,7 +232,7 @@ async def create_content( parent: Optional[str] = None, content: Optional[analyze.Content] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> analyze.Content: r"""Create a content. @@ -346,7 +347,7 @@ async def update_content( content: Optional[analyze.Content] = None, update_mask: Optional[field_mask_pb2.FieldMask] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> analyze.Content: r"""Update a content. Only supports full resource update. @@ -461,7 +462,7 @@ async def delete_content( *, name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> None: r"""Delete a content. @@ -550,7 +551,7 @@ async def get_content( *, name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> analyze.Content: r"""Get a content resource. @@ -660,7 +661,7 @@ async def get_iam_policy( *, resource: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> policy_pb2.Policy: r"""Gets the access control policy for a contentitem resource. A @@ -840,7 +841,7 @@ async def set_iam_policy( request: Optional[Union[iam_policy_pb2.SetIamPolicyRequest, dict]] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> policy_pb2.Policy: r"""Sets the access control policy on the specified contentitem @@ -987,7 +988,7 @@ async def test_iam_permissions( request: Optional[Union[iam_policy_pb2.TestIamPermissionsRequest, dict]] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> iam_policy_pb2.TestIamPermissionsResponse: r"""Returns the caller's permissions on a resource. If the resource @@ -1090,7 +1091,7 @@ async def list_content( *, parent: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.ListContentAsyncPager: r"""List content. @@ -1212,7 +1213,7 @@ async def list_operations( request: Optional[operations_pb2.ListOperationsRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> operations_pb2.ListOperationsResponse: r"""Lists operations that match the specified filter in the request. @@ -1266,7 +1267,7 @@ async def get_operation( request: Optional[operations_pb2.GetOperationRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> operations_pb2.Operation: r"""Gets the latest state of a long-running operation. @@ -1320,7 +1321,7 @@ async def delete_operation( request: Optional[operations_pb2.DeleteOperationRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> None: r"""Deletes a long-running operation. @@ -1375,7 +1376,7 @@ async def cancel_operation( request: Optional[operations_pb2.CancelOperationRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> None: r"""Starts asynchronous cancellation on a long-running operation. @@ -1429,7 +1430,7 @@ async def get_location( request: Optional[locations_pb2.GetLocationRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> locations_pb2.Location: r"""Gets information about a location. @@ -1483,7 +1484,7 @@ async def list_locations( request: Optional[locations_pb2.ListLocationsRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> locations_pb2.ListLocationsResponse: r"""Lists information about the supported locations for this service. @@ -1539,14 +1540,9 @@ async def __aexit__(self, exc_type, exc, tb): await self.transport.close() -try: - DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( - gapic_version=pkg_resources.get_distribution( - "google-cloud-dataplex", - ).version, - ) -except pkg_resources.DistributionNotFound: - DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo() +DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( + gapic_version=package_version.__version__ +) __all__ = ("ContentServiceAsyncClient",) diff --git a/google/cloud/dataplex_v1/services/content_service/client.py b/google/cloud/dataplex_v1/services/content_service/client.py index 79b51fc..4e93b26 100644 --- a/google/cloud/dataplex_v1/services/content_service/client.py +++ b/google/cloud/dataplex_v1/services/content_service/client.py @@ -38,7 +38,8 @@ from google.auth.transport import mtls # type: ignore from google.auth.transport.grpc import SslCredentials # type: ignore from google.oauth2 import service_account # type: ignore -import pkg_resources + +from google.cloud.dataplex_v1 import gapic_version as package_version try: OptionalRetry = Union[retries.Retry, gapic_v1.method._MethodDefault] @@ -480,7 +481,7 @@ def create_content( parent: Optional[str] = None, content: Optional[analyze.Content] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> analyze.Content: r"""Create a content. @@ -595,7 +596,7 @@ def update_content( content: Optional[analyze.Content] = None, update_mask: Optional[field_mask_pb2.FieldMask] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> analyze.Content: r"""Update a content. Only supports full resource update. @@ -710,7 +711,7 @@ def delete_content( *, name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> None: r"""Delete a content. @@ -799,7 +800,7 @@ def get_content( *, name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> analyze.Content: r"""Get a content resource. @@ -900,7 +901,7 @@ def get_iam_policy( *, resource: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> policy_pb2.Policy: r"""Gets the access control policy for a contentitem resource. A @@ -1068,7 +1069,7 @@ def set_iam_policy( request: Optional[Union[iam_policy_pb2.SetIamPolicyRequest, dict]] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> policy_pb2.Policy: r"""Sets the access control policy on the specified contentitem @@ -1214,7 +1215,7 @@ def test_iam_permissions( request: Optional[Union[iam_policy_pb2.TestIamPermissionsRequest, dict]] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> iam_policy_pb2.TestIamPermissionsResponse: r"""Returns the caller's permissions on a resource. If the resource @@ -1307,7 +1308,7 @@ def list_content( *, parent: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.ListContentPager: r"""List content. @@ -1433,7 +1434,7 @@ def list_operations( request: Optional[operations_pb2.ListOperationsRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> operations_pb2.ListOperationsResponse: r"""Lists operations that match the specified filter in the request. @@ -1487,7 +1488,7 @@ def get_operation( request: Optional[operations_pb2.GetOperationRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> operations_pb2.Operation: r"""Gets the latest state of a long-running operation. @@ -1541,7 +1542,7 @@ def delete_operation( request: Optional[operations_pb2.DeleteOperationRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> None: r"""Deletes a long-running operation. @@ -1596,7 +1597,7 @@ def cancel_operation( request: Optional[operations_pb2.CancelOperationRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> None: r"""Starts asynchronous cancellation on a long-running operation. @@ -1650,7 +1651,7 @@ def get_location( request: Optional[locations_pb2.GetLocationRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> locations_pb2.Location: r"""Gets information about a location. @@ -1704,7 +1705,7 @@ def list_locations( request: Optional[locations_pb2.ListLocationsRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> locations_pb2.ListLocationsResponse: r"""Lists information about the supported locations for this service. @@ -1754,14 +1755,9 @@ def list_locations( return response -try: - DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( - gapic_version=pkg_resources.get_distribution( - "google-cloud-dataplex", - ).version, - ) -except pkg_resources.DistributionNotFound: - DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo() +DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( + gapic_version=package_version.__version__ +) __all__ = ("ContentServiceClient",) diff --git a/google/cloud/dataplex_v1/services/content_service/transports/base.py b/google/cloud/dataplex_v1/services/content_service/transports/base.py index 9c42e86..9523955 100644 --- a/google/cloud/dataplex_v1/services/content_service/transports/base.py +++ b/google/cloud/dataplex_v1/services/content_service/transports/base.py @@ -28,20 +28,15 @@ from google.longrunning import operations_pb2 from google.oauth2 import service_account # type: ignore from google.protobuf import empty_pb2 # type: ignore -import pkg_resources +from google.cloud.dataplex_v1 import gapic_version as package_version from google.cloud.dataplex_v1.types import analyze from google.cloud.dataplex_v1.types import content from google.cloud.dataplex_v1.types import content as gcd_content -try: - DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( - gapic_version=pkg_resources.get_distribution( - "google-cloud-dataplex", - ).version, - ) -except pkg_resources.DistributionNotFound: - DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo() +DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( + gapic_version=package_version.__version__ +) class ContentServiceTransport(abc.ABC): diff --git a/google/cloud/dataplex_v1/services/dataplex_service/async_client.py b/google/cloud/dataplex_v1/services/dataplex_service/async_client.py index 5e261d7..5fbc667 100644 --- a/google/cloud/dataplex_v1/services/dataplex_service/async_client.py +++ b/google/cloud/dataplex_v1/services/dataplex_service/async_client.py @@ -34,7 +34,8 @@ from google.api_core.client_options import ClientOptions from google.auth import credentials as ga_credentials # type: ignore from google.oauth2 import service_account # type: ignore -import pkg_resources + +from google.cloud.dataplex_v1 import gapic_version as package_version try: OptionalRetry = Union[retries.Retry, gapic_v1.method._MethodDefault] @@ -251,7 +252,7 @@ async def create_lake( lake: Optional[resources.Lake] = None, lake_id: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> operation_async.AsyncOperation: r"""Creates a lake resource. @@ -282,7 +283,7 @@ async def sample_create_lake(): print("Waiting for operation to complete...") - response = await operation.result() + response = (await operation).result() # Handle the response print(response) @@ -404,7 +405,7 @@ async def update_lake( lake: Optional[resources.Lake] = None, update_mask: Optional[field_mask_pb2.FieldMask] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> operation_async.AsyncOperation: r"""Updates a lake resource. @@ -433,7 +434,7 @@ async def sample_update_lake(): print("Waiting for operation to complete...") - response = await operation.result() + response = (await operation).result() # Handle the response print(response) @@ -537,7 +538,7 @@ async def delete_lake( *, name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> operation_async.AsyncOperation: r"""Deletes a lake resource. All zones within the lake @@ -568,7 +569,7 @@ async def sample_delete_lake(): print("Waiting for operation to complete...") - response = await operation.result() + response = (await operation).result() # Handle the response print(response) @@ -661,7 +662,7 @@ async def list_lakes( *, parent: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.ListLakesAsyncPager: r"""Lists lake resources in a project and location. @@ -785,7 +786,7 @@ async def get_lake( *, name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> resources.Lake: r"""Retrieves a lake resource. @@ -909,7 +910,7 @@ async def list_lake_actions( *, parent: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.ListLakeActionsAsyncPager: r"""Lists action resources in a lake. @@ -1033,7 +1034,7 @@ async def create_zone( zone: Optional[resources.Zone] = None, zone_id: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> operation_async.AsyncOperation: r"""Creates a zone resource within a lake. @@ -1069,7 +1070,7 @@ async def sample_create_zone(): print("Waiting for operation to complete...") - response = await operation.result() + response = (await operation).result() # Handle the response print(response) @@ -1185,7 +1186,7 @@ async def update_zone( zone: Optional[resources.Zone] = None, update_mask: Optional[field_mask_pb2.FieldMask] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> operation_async.AsyncOperation: r"""Updates a zone resource. @@ -1219,7 +1220,7 @@ async def sample_update_zone(): print("Waiting for operation to complete...") - response = await operation.result() + response = (await operation).result() # Handle the response print(response) @@ -1316,7 +1317,7 @@ async def delete_zone( *, name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> operation_async.AsyncOperation: r"""Deletes a zone resource. All assets within a zone @@ -1347,7 +1348,7 @@ async def sample_delete_zone(): print("Waiting for operation to complete...") - response = await operation.result() + response = (await operation).result() # Handle the response print(response) @@ -1440,7 +1441,7 @@ async def list_zones( *, parent: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.ListZonesAsyncPager: r"""Lists zone resources in a lake. @@ -1562,7 +1563,7 @@ async def get_zone( *, name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> resources.Zone: r"""Retrieves a zone resource. @@ -1678,7 +1679,7 @@ async def list_zone_actions( *, parent: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.ListZoneActionsAsyncPager: r"""Lists action resources in a zone. @@ -1802,7 +1803,7 @@ async def create_asset( asset: Optional[resources.Asset] = None, asset_id: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> operation_async.AsyncOperation: r"""Creates an asset resource. @@ -1837,7 +1838,7 @@ async def sample_create_asset(): print("Waiting for operation to complete...") - response = await operation.result() + response = (await operation).result() # Handle the response print(response) @@ -1947,7 +1948,7 @@ async def update_asset( asset: Optional[resources.Asset] = None, update_mask: Optional[field_mask_pb2.FieldMask] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> operation_async.AsyncOperation: r"""Updates an asset resource. @@ -1980,7 +1981,7 @@ async def sample_update_asset(): print("Waiting for operation to complete...") - response = await operation.result() + response = (await operation).result() # Handle the response print(response) @@ -2074,7 +2075,7 @@ async def delete_asset( *, name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> operation_async.AsyncOperation: r"""Deletes an asset resource. The referenced storage @@ -2106,7 +2107,7 @@ async def sample_delete_asset(): print("Waiting for operation to complete...") - response = await operation.result() + response = (await operation).result() # Handle the response print(response) @@ -2199,7 +2200,7 @@ async def list_assets( *, parent: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.ListAssetsAsyncPager: r"""Lists asset resources in a zone. @@ -2321,7 +2322,7 @@ async def get_asset( *, name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> resources.Asset: r"""Retrieves an asset resource. @@ -2432,7 +2433,7 @@ async def list_asset_actions( *, parent: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.ListAssetActionsAsyncPager: r"""Lists action resources in an asset. @@ -2556,7 +2557,7 @@ async def create_task( task: Optional[tasks.Task] = None, task_id: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> operation_async.AsyncOperation: r"""Creates a task resource within a lake. @@ -2594,7 +2595,7 @@ async def sample_create_task(): print("Waiting for operation to complete...") - response = await operation.result() + response = (await operation).result() # Handle the response print(response) @@ -2695,7 +2696,7 @@ async def update_task( task: Optional[tasks.Task] = None, update_mask: Optional[field_mask_pb2.FieldMask] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> operation_async.AsyncOperation: r"""Update the task resource. @@ -2731,7 +2732,7 @@ async def sample_update_task(): print("Waiting for operation to complete...") - response = await operation.result() + response = (await operation).result() # Handle the response print(response) @@ -2826,7 +2827,7 @@ async def delete_task( *, name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> operation_async.AsyncOperation: r"""Delete the task resource. @@ -2856,7 +2857,7 @@ async def sample_delete_task(): print("Waiting for operation to complete...") - response = await operation.result() + response = (await operation).result() # Handle the response print(response) @@ -2949,7 +2950,7 @@ async def list_tasks( *, parent: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.ListTasksAsyncPager: r"""Lists tasks under the given lake. @@ -3071,7 +3072,7 @@ async def get_task( *, name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> tasks.Task: r"""Get task resource. @@ -3179,7 +3180,7 @@ async def list_jobs( *, parent: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.ListJobsAsyncPager: r"""Lists Jobs under the given task. @@ -3301,7 +3302,7 @@ async def get_job( *, name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> tasks.Job: r"""Get job resource. @@ -3411,7 +3412,7 @@ async def cancel_job( *, name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> None: r"""Cancel jobs running for the task resource. @@ -3502,7 +3503,7 @@ async def create_environment( environment: Optional[analyze.Environment] = None, environment_id: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> operation_async.AsyncOperation: r"""Create an environment resource. @@ -3537,7 +3538,7 @@ async def sample_create_environment(): print("Waiting for operation to complete...") - response = await operation.result() + response = (await operation).result() # Handle the response print(response) @@ -3645,7 +3646,7 @@ async def update_environment( environment: Optional[analyze.Environment] = None, update_mask: Optional[field_mask_pb2.FieldMask] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> operation_async.AsyncOperation: r"""Update the environment resource. @@ -3678,7 +3679,7 @@ async def sample_update_environment(): print("Waiting for operation to complete...") - response = await operation.result() + response = (await operation).result() # Handle the response print(response) @@ -3772,7 +3773,7 @@ async def delete_environment( *, name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> operation_async.AsyncOperation: r"""Delete the environment resource. All the child @@ -3804,7 +3805,7 @@ async def sample_delete_environment(): print("Waiting for operation to complete...") - response = await operation.result() + response = (await operation).result() # Handle the response print(response) @@ -3897,7 +3898,7 @@ async def list_environments( *, parent: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.ListEnvironmentsAsyncPager: r"""Lists environments under the given lake. @@ -4019,7 +4020,7 @@ async def get_environment( *, name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> analyze.Environment: r"""Get environment resource. @@ -4130,7 +4131,7 @@ async def list_sessions( *, parent: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.ListSessionsAsyncPager: r"""Lists session resources in an environment. @@ -4242,7 +4243,7 @@ async def list_operations( request: Optional[operations_pb2.ListOperationsRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> operations_pb2.ListOperationsResponse: r"""Lists operations that match the specified filter in the request. @@ -4296,7 +4297,7 @@ async def get_operation( request: Optional[operations_pb2.GetOperationRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> operations_pb2.Operation: r"""Gets the latest state of a long-running operation. @@ -4350,7 +4351,7 @@ async def delete_operation( request: Optional[operations_pb2.DeleteOperationRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> None: r"""Deletes a long-running operation. @@ -4405,7 +4406,7 @@ async def cancel_operation( request: Optional[operations_pb2.CancelOperationRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> None: r"""Starts asynchronous cancellation on a long-running operation. @@ -4459,7 +4460,7 @@ async def get_location( request: Optional[locations_pb2.GetLocationRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> locations_pb2.Location: r"""Gets information about a location. @@ -4513,7 +4514,7 @@ async def list_locations( request: Optional[locations_pb2.ListLocationsRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> locations_pb2.ListLocationsResponse: r"""Lists information about the supported locations for this service. @@ -4569,14 +4570,9 @@ async def __aexit__(self, exc_type, exc, tb): await self.transport.close() -try: - DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( - gapic_version=pkg_resources.get_distribution( - "google-cloud-dataplex", - ).version, - ) -except pkg_resources.DistributionNotFound: - DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo() +DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( + gapic_version=package_version.__version__ +) __all__ = ("DataplexServiceAsyncClient",) diff --git a/google/cloud/dataplex_v1/services/dataplex_service/client.py b/google/cloud/dataplex_v1/services/dataplex_service/client.py index 2b289b5..67a6a3a 100644 --- a/google/cloud/dataplex_v1/services/dataplex_service/client.py +++ b/google/cloud/dataplex_v1/services/dataplex_service/client.py @@ -38,7 +38,8 @@ from google.auth.transport import mtls # type: ignore from google.auth.transport.grpc import SslCredentials # type: ignore from google.oauth2 import service_account # type: ignore -import pkg_resources + +from google.cloud.dataplex_v1 import gapic_version as package_version try: OptionalRetry = Union[retries.Retry, gapic_v1.method._MethodDefault] @@ -642,7 +643,7 @@ def create_lake( lake: Optional[resources.Lake] = None, lake_id: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> operation.Operation: r"""Creates a lake resource. @@ -795,7 +796,7 @@ def update_lake( lake: Optional[resources.Lake] = None, update_mask: Optional[field_mask_pb2.FieldMask] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> operation.Operation: r"""Updates a lake resource. @@ -928,7 +929,7 @@ def delete_lake( *, name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> operation.Operation: r"""Deletes a lake resource. All zones within the lake @@ -1052,7 +1053,7 @@ def list_lakes( *, parent: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.ListLakesPager: r"""Lists lake resources in a project and location. @@ -1167,7 +1168,7 @@ def get_lake( *, name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> resources.Lake: r"""Retrieves a lake resource. @@ -1282,7 +1283,7 @@ def list_lake_actions( *, parent: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.ListLakeActionsPager: r"""Lists action resources in a lake. @@ -1397,7 +1398,7 @@ def create_zone( zone: Optional[resources.Zone] = None, zone_id: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> operation.Operation: r"""Creates a zone resource within a lake. @@ -1549,7 +1550,7 @@ def update_zone( zone: Optional[resources.Zone] = None, update_mask: Optional[field_mask_pb2.FieldMask] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> operation.Operation: r"""Updates a zone resource. @@ -1680,7 +1681,7 @@ def delete_zone( *, name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> operation.Operation: r"""Deletes a zone resource. All assets within a zone @@ -1804,7 +1805,7 @@ def list_zones( *, parent: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.ListZonesPager: r"""Lists zone resources in a lake. @@ -1917,7 +1918,7 @@ def get_zone( *, name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> resources.Zone: r"""Retrieves a zone resource. @@ -2024,7 +2025,7 @@ def list_zone_actions( *, parent: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.ListZoneActionsPager: r"""Lists action resources in a zone. @@ -2139,7 +2140,7 @@ def create_asset( asset: Optional[resources.Asset] = None, asset_id: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> operation.Operation: r"""Creates an asset resource. @@ -2284,7 +2285,7 @@ def update_asset( asset: Optional[resources.Asset] = None, update_mask: Optional[field_mask_pb2.FieldMask] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> operation.Operation: r"""Updates an asset resource. @@ -2411,7 +2412,7 @@ def delete_asset( *, name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> operation.Operation: r"""Deletes an asset resource. The referenced storage @@ -2536,7 +2537,7 @@ def list_assets( *, parent: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.ListAssetsPager: r"""Lists asset resources in a zone. @@ -2649,7 +2650,7 @@ def get_asset( *, name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> resources.Asset: r"""Retrieves an asset resource. @@ -2751,7 +2752,7 @@ def list_asset_actions( *, parent: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.ListAssetActionsPager: r"""Lists action resources in an asset. @@ -2866,7 +2867,7 @@ def create_task( task: Optional[tasks.Task] = None, task_id: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> operation.Operation: r"""Creates a task resource within a lake. @@ -3005,7 +3006,7 @@ def update_task( task: Optional[tasks.Task] = None, update_mask: Optional[field_mask_pb2.FieldMask] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> operation.Operation: r"""Update the task resource. @@ -3136,7 +3137,7 @@ def delete_task( *, name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> operation.Operation: r"""Delete the task resource. @@ -3259,7 +3260,7 @@ def list_tasks( *, parent: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.ListTasksPager: r"""Lists tasks under the given lake. @@ -3372,7 +3373,7 @@ def get_task( *, name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> tasks.Task: r"""Get task resource. @@ -3471,7 +3472,7 @@ def list_jobs( *, parent: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.ListJobsPager: r"""Lists Jobs under the given task. @@ -3584,7 +3585,7 @@ def get_job( *, name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> tasks.Job: r"""Get job resource. @@ -3685,7 +3686,7 @@ def cancel_job( *, name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> None: r"""Cancel jobs running for the task resource. @@ -3776,7 +3777,7 @@ def create_environment( environment: Optional[analyze.Environment] = None, environment_id: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> operation.Operation: r"""Create an environment resource. @@ -3919,7 +3920,7 @@ def update_environment( environment: Optional[analyze.Environment] = None, update_mask: Optional[field_mask_pb2.FieldMask] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> operation.Operation: r"""Update the environment resource. @@ -4046,7 +4047,7 @@ def delete_environment( *, name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> operation.Operation: r"""Delete the environment resource. All the child @@ -4171,7 +4172,7 @@ def list_environments( *, parent: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.ListEnvironmentsPager: r"""Lists environments under the given lake. @@ -4284,7 +4285,7 @@ def get_environment( *, name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> analyze.Environment: r"""Get environment resource. @@ -4386,7 +4387,7 @@ def list_sessions( *, parent: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.ListSessionsPager: r"""Lists session resources in an environment. @@ -4511,7 +4512,7 @@ def list_operations( request: Optional[operations_pb2.ListOperationsRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> operations_pb2.ListOperationsResponse: r"""Lists operations that match the specified filter in the request. @@ -4565,7 +4566,7 @@ def get_operation( request: Optional[operations_pb2.GetOperationRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> operations_pb2.Operation: r"""Gets the latest state of a long-running operation. @@ -4619,7 +4620,7 @@ def delete_operation( request: Optional[operations_pb2.DeleteOperationRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> None: r"""Deletes a long-running operation. @@ -4674,7 +4675,7 @@ def cancel_operation( request: Optional[operations_pb2.CancelOperationRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> None: r"""Starts asynchronous cancellation on a long-running operation. @@ -4728,7 +4729,7 @@ def get_location( request: Optional[locations_pb2.GetLocationRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> locations_pb2.Location: r"""Gets information about a location. @@ -4782,7 +4783,7 @@ def list_locations( request: Optional[locations_pb2.ListLocationsRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> locations_pb2.ListLocationsResponse: r"""Lists information about the supported locations for this service. @@ -4832,14 +4833,9 @@ def list_locations( return response -try: - DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( - gapic_version=pkg_resources.get_distribution( - "google-cloud-dataplex", - ).version, - ) -except pkg_resources.DistributionNotFound: - DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo() +DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( + gapic_version=package_version.__version__ +) __all__ = ("DataplexServiceClient",) diff --git a/google/cloud/dataplex_v1/services/dataplex_service/transports/base.py b/google/cloud/dataplex_v1/services/dataplex_service/transports/base.py index 525bfc9..2a7504e 100644 --- a/google/cloud/dataplex_v1/services/dataplex_service/transports/base.py +++ b/google/cloud/dataplex_v1/services/dataplex_service/transports/base.py @@ -28,18 +28,13 @@ from google.longrunning import operations_pb2 # type: ignore from google.oauth2 import service_account # type: ignore from google.protobuf import empty_pb2 # type: ignore -import pkg_resources +from google.cloud.dataplex_v1 import gapic_version as package_version from google.cloud.dataplex_v1.types import analyze, resources, service, tasks -try: - DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( - gapic_version=pkg_resources.get_distribution( - "google-cloud-dataplex", - ).version, - ) -except pkg_resources.DistributionNotFound: - DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo() +DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( + gapic_version=package_version.__version__ +) class DataplexServiceTransport(abc.ABC): diff --git a/google/cloud/dataplex_v1/services/metadata_service/async_client.py b/google/cloud/dataplex_v1/services/metadata_service/async_client.py index e6a7b23..f2945b9 100644 --- a/google/cloud/dataplex_v1/services/metadata_service/async_client.py +++ b/google/cloud/dataplex_v1/services/metadata_service/async_client.py @@ -34,7 +34,8 @@ from google.api_core.client_options import ClientOptions from google.auth import credentials as ga_credentials # type: ignore from google.oauth2 import service_account # type: ignore -import pkg_resources + +from google.cloud.dataplex_v1 import gapic_version as package_version try: OptionalRetry = Union[retries.Retry, gapic_v1.method._MethodDefault] @@ -232,7 +233,7 @@ async def create_entity( parent: Optional[str] = None, entity: Optional[metadata_.Entity] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> metadata_.Entity: r"""Create a metadata entity. @@ -349,7 +350,7 @@ async def update_entity( request: Optional[Union[metadata_.UpdateEntityRequest, dict]] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> metadata_.Entity: r"""Update a metadata entity. Only supports full resource @@ -445,7 +446,7 @@ async def delete_entity( *, name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> None: r"""Delete a metadata entity. @@ -535,7 +536,7 @@ async def get_entity( *, name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> metadata_.Entity: r"""Get a metadata entity. @@ -645,7 +646,7 @@ async def list_entities( *, parent: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.ListEntitiesAsyncPager: r"""List metadata entities in a zone. @@ -769,7 +770,7 @@ async def create_partition( parent: Optional[str] = None, partition: Optional[metadata_.Partition] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> metadata_.Partition: r"""Create a metadata partition. @@ -882,7 +883,7 @@ async def delete_partition( *, name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> None: r"""Delete a metadata partition. @@ -974,7 +975,7 @@ async def get_partition( *, name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> metadata_.Partition: r"""Get a metadata partition of an entity. @@ -1087,7 +1088,7 @@ async def list_partitions( *, parent: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.ListPartitionsAsyncPager: r"""List metadata partitions of an entity. @@ -1208,7 +1209,7 @@ async def list_operations( request: Optional[operations_pb2.ListOperationsRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> operations_pb2.ListOperationsResponse: r"""Lists operations that match the specified filter in the request. @@ -1262,7 +1263,7 @@ async def get_operation( request: Optional[operations_pb2.GetOperationRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> operations_pb2.Operation: r"""Gets the latest state of a long-running operation. @@ -1316,7 +1317,7 @@ async def delete_operation( request: Optional[operations_pb2.DeleteOperationRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> None: r"""Deletes a long-running operation. @@ -1371,7 +1372,7 @@ async def cancel_operation( request: Optional[operations_pb2.CancelOperationRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> None: r"""Starts asynchronous cancellation on a long-running operation. @@ -1425,7 +1426,7 @@ async def get_location( request: Optional[locations_pb2.GetLocationRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> locations_pb2.Location: r"""Gets information about a location. @@ -1479,7 +1480,7 @@ async def list_locations( request: Optional[locations_pb2.ListLocationsRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> locations_pb2.ListLocationsResponse: r"""Lists information about the supported locations for this service. @@ -1535,14 +1536,9 @@ async def __aexit__(self, exc_type, exc, tb): await self.transport.close() -try: - DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( - gapic_version=pkg_resources.get_distribution( - "google-cloud-dataplex", - ).version, - ) -except pkg_resources.DistributionNotFound: - DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo() +DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( + gapic_version=package_version.__version__ +) __all__ = ("MetadataServiceAsyncClient",) diff --git a/google/cloud/dataplex_v1/services/metadata_service/client.py b/google/cloud/dataplex_v1/services/metadata_service/client.py index 4438af0..38b34e7 100644 --- a/google/cloud/dataplex_v1/services/metadata_service/client.py +++ b/google/cloud/dataplex_v1/services/metadata_service/client.py @@ -38,7 +38,8 @@ from google.auth.transport import mtls # type: ignore from google.auth.transport.grpc import SslCredentials # type: ignore from google.oauth2 import service_account # type: ignore -import pkg_resources + +from google.cloud.dataplex_v1 import gapic_version as package_version try: OptionalRetry = Union[retries.Retry, gapic_v1.method._MethodDefault] @@ -513,7 +514,7 @@ def create_entity( parent: Optional[str] = None, entity: Optional[metadata_.Entity] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> metadata_.Entity: r"""Create a metadata entity. @@ -630,7 +631,7 @@ def update_entity( request: Optional[Union[metadata_.UpdateEntityRequest, dict]] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> metadata_.Entity: r"""Update a metadata entity. Only supports full resource @@ -727,7 +728,7 @@ def delete_entity( *, name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> None: r"""Delete a metadata entity. @@ -817,7 +818,7 @@ def get_entity( *, name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> metadata_.Entity: r"""Get a metadata entity. @@ -918,7 +919,7 @@ def list_entities( *, parent: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.ListEntitiesPager: r"""List metadata entities in a zone. @@ -1033,7 +1034,7 @@ def create_partition( parent: Optional[str] = None, partition: Optional[metadata_.Partition] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> metadata_.Partition: r"""Create a metadata partition. @@ -1146,7 +1147,7 @@ def delete_partition( *, name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> None: r"""Delete a metadata partition. @@ -1238,7 +1239,7 @@ def get_partition( *, name: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> metadata_.Partition: r"""Get a metadata partition of an entity. @@ -1342,7 +1343,7 @@ def list_partitions( *, parent: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.ListPartitionsPager: r"""List metadata partitions of an entity. @@ -1467,7 +1468,7 @@ def list_operations( request: Optional[operations_pb2.ListOperationsRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> operations_pb2.ListOperationsResponse: r"""Lists operations that match the specified filter in the request. @@ -1521,7 +1522,7 @@ def get_operation( request: Optional[operations_pb2.GetOperationRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> operations_pb2.Operation: r"""Gets the latest state of a long-running operation. @@ -1575,7 +1576,7 @@ def delete_operation( request: Optional[operations_pb2.DeleteOperationRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> None: r"""Deletes a long-running operation. @@ -1630,7 +1631,7 @@ def cancel_operation( request: Optional[operations_pb2.CancelOperationRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> None: r"""Starts asynchronous cancellation on a long-running operation. @@ -1684,7 +1685,7 @@ def get_location( request: Optional[locations_pb2.GetLocationRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> locations_pb2.Location: r"""Gets information about a location. @@ -1738,7 +1739,7 @@ def list_locations( request: Optional[locations_pb2.ListLocationsRequest] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> locations_pb2.ListLocationsResponse: r"""Lists information about the supported locations for this service. @@ -1788,14 +1789,9 @@ def list_locations( return response -try: - DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( - gapic_version=pkg_resources.get_distribution( - "google-cloud-dataplex", - ).version, - ) -except pkg_resources.DistributionNotFound: - DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo() +DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( + gapic_version=package_version.__version__ +) __all__ = ("MetadataServiceClient",) diff --git a/google/cloud/dataplex_v1/services/metadata_service/transports/base.py b/google/cloud/dataplex_v1/services/metadata_service/transports/base.py index 8d10f83..059effe 100644 --- a/google/cloud/dataplex_v1/services/metadata_service/transports/base.py +++ b/google/cloud/dataplex_v1/services/metadata_service/transports/base.py @@ -28,18 +28,13 @@ from google.longrunning import operations_pb2 from google.oauth2 import service_account # type: ignore from google.protobuf import empty_pb2 # type: ignore -import pkg_resources +from google.cloud.dataplex_v1 import gapic_version as package_version from google.cloud.dataplex_v1.types import metadata_ -try: - DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( - gapic_version=pkg_resources.get_distribution( - "google-cloud-dataplex", - ).version, - ) -except pkg_resources.DistributionNotFound: - DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo() +DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( + gapic_version=package_version.__version__ +) class MetadataServiceTransport(abc.ABC): diff --git a/release-please-config.json b/release-please-config.json index 0db768f..b69acb3 100644 --- a/release-please-config.json +++ b/release-please-config.json @@ -5,6 +5,7 @@ "release-type": "python", "extra-files": [ "google/cloud/dataplex/gapic_version.py", + "google/cloud/dataplex_v1/gapic_version.py", { "type": "json", "path": "samples/generated_samples/snippet_metadata_google.cloud.dataplex.v1.json", diff --git a/samples/generated_samples/dataplex_v1_generated_dataplex_service_create_asset_async.py b/samples/generated_samples/dataplex_v1_generated_dataplex_service_create_asset_async.py index 14d7f3f..2ba577c 100644 --- a/samples/generated_samples/dataplex_v1_generated_dataplex_service_create_asset_async.py +++ b/samples/generated_samples/dataplex_v1_generated_dataplex_service_create_asset_async.py @@ -53,7 +53,7 @@ async def sample_create_asset(): print("Waiting for operation to complete...") - response = await operation.result() + response = (await operation).result() # Handle the response print(response) diff --git a/samples/generated_samples/dataplex_v1_generated_dataplex_service_create_environment_async.py b/samples/generated_samples/dataplex_v1_generated_dataplex_service_create_environment_async.py index 42d995c..9089677 100644 --- a/samples/generated_samples/dataplex_v1_generated_dataplex_service_create_environment_async.py +++ b/samples/generated_samples/dataplex_v1_generated_dataplex_service_create_environment_async.py @@ -53,7 +53,7 @@ async def sample_create_environment(): print("Waiting for operation to complete...") - response = await operation.result() + response = (await operation).result() # Handle the response print(response) diff --git a/samples/generated_samples/dataplex_v1_generated_dataplex_service_create_lake_async.py b/samples/generated_samples/dataplex_v1_generated_dataplex_service_create_lake_async.py index c2604d2..7e04801 100644 --- a/samples/generated_samples/dataplex_v1_generated_dataplex_service_create_lake_async.py +++ b/samples/generated_samples/dataplex_v1_generated_dataplex_service_create_lake_async.py @@ -49,7 +49,7 @@ async def sample_create_lake(): print("Waiting for operation to complete...") - response = await operation.result() + response = (await operation).result() # Handle the response print(response) diff --git a/samples/generated_samples/dataplex_v1_generated_dataplex_service_create_task_async.py b/samples/generated_samples/dataplex_v1_generated_dataplex_service_create_task_async.py index 4d7bc54..43d27ac 100644 --- a/samples/generated_samples/dataplex_v1_generated_dataplex_service_create_task_async.py +++ b/samples/generated_samples/dataplex_v1_generated_dataplex_service_create_task_async.py @@ -56,7 +56,7 @@ async def sample_create_task(): print("Waiting for operation to complete...") - response = await operation.result() + response = (await operation).result() # Handle the response print(response) diff --git a/samples/generated_samples/dataplex_v1_generated_dataplex_service_create_zone_async.py b/samples/generated_samples/dataplex_v1_generated_dataplex_service_create_zone_async.py index bb15bd0..c656a5b 100644 --- a/samples/generated_samples/dataplex_v1_generated_dataplex_service_create_zone_async.py +++ b/samples/generated_samples/dataplex_v1_generated_dataplex_service_create_zone_async.py @@ -54,7 +54,7 @@ async def sample_create_zone(): print("Waiting for operation to complete...") - response = await operation.result() + response = (await operation).result() # Handle the response print(response) diff --git a/samples/generated_samples/dataplex_v1_generated_dataplex_service_delete_asset_async.py b/samples/generated_samples/dataplex_v1_generated_dataplex_service_delete_asset_async.py index 3113c89..c228721 100644 --- a/samples/generated_samples/dataplex_v1_generated_dataplex_service_delete_asset_async.py +++ b/samples/generated_samples/dataplex_v1_generated_dataplex_service_delete_asset_async.py @@ -48,7 +48,7 @@ async def sample_delete_asset(): print("Waiting for operation to complete...") - response = await operation.result() + response = (await operation).result() # Handle the response print(response) diff --git a/samples/generated_samples/dataplex_v1_generated_dataplex_service_delete_environment_async.py b/samples/generated_samples/dataplex_v1_generated_dataplex_service_delete_environment_async.py index 1b49fe7..31f7de0 100644 --- a/samples/generated_samples/dataplex_v1_generated_dataplex_service_delete_environment_async.py +++ b/samples/generated_samples/dataplex_v1_generated_dataplex_service_delete_environment_async.py @@ -48,7 +48,7 @@ async def sample_delete_environment(): print("Waiting for operation to complete...") - response = await operation.result() + response = (await operation).result() # Handle the response print(response) diff --git a/samples/generated_samples/dataplex_v1_generated_dataplex_service_delete_lake_async.py b/samples/generated_samples/dataplex_v1_generated_dataplex_service_delete_lake_async.py index f19f4ce..88de8af 100644 --- a/samples/generated_samples/dataplex_v1_generated_dataplex_service_delete_lake_async.py +++ b/samples/generated_samples/dataplex_v1_generated_dataplex_service_delete_lake_async.py @@ -48,7 +48,7 @@ async def sample_delete_lake(): print("Waiting for operation to complete...") - response = await operation.result() + response = (await operation).result() # Handle the response print(response) diff --git a/samples/generated_samples/dataplex_v1_generated_dataplex_service_delete_task_async.py b/samples/generated_samples/dataplex_v1_generated_dataplex_service_delete_task_async.py index b64c416..f5415ee 100644 --- a/samples/generated_samples/dataplex_v1_generated_dataplex_service_delete_task_async.py +++ b/samples/generated_samples/dataplex_v1_generated_dataplex_service_delete_task_async.py @@ -48,7 +48,7 @@ async def sample_delete_task(): print("Waiting for operation to complete...") - response = await operation.result() + response = (await operation).result() # Handle the response print(response) diff --git a/samples/generated_samples/dataplex_v1_generated_dataplex_service_delete_zone_async.py b/samples/generated_samples/dataplex_v1_generated_dataplex_service_delete_zone_async.py index 5ac9c4a..19a4edb 100644 --- a/samples/generated_samples/dataplex_v1_generated_dataplex_service_delete_zone_async.py +++ b/samples/generated_samples/dataplex_v1_generated_dataplex_service_delete_zone_async.py @@ -48,7 +48,7 @@ async def sample_delete_zone(): print("Waiting for operation to complete...") - response = await operation.result() + response = (await operation).result() # Handle the response print(response) diff --git a/samples/generated_samples/dataplex_v1_generated_dataplex_service_update_asset_async.py b/samples/generated_samples/dataplex_v1_generated_dataplex_service_update_asset_async.py index 00f2ffa..46572ff 100644 --- a/samples/generated_samples/dataplex_v1_generated_dataplex_service_update_asset_async.py +++ b/samples/generated_samples/dataplex_v1_generated_dataplex_service_update_asset_async.py @@ -51,7 +51,7 @@ async def sample_update_asset(): print("Waiting for operation to complete...") - response = await operation.result() + response = (await operation).result() # Handle the response print(response) diff --git a/samples/generated_samples/dataplex_v1_generated_dataplex_service_update_environment_async.py b/samples/generated_samples/dataplex_v1_generated_dataplex_service_update_environment_async.py index 6bdcdde..6842474 100644 --- a/samples/generated_samples/dataplex_v1_generated_dataplex_service_update_environment_async.py +++ b/samples/generated_samples/dataplex_v1_generated_dataplex_service_update_environment_async.py @@ -51,7 +51,7 @@ async def sample_update_environment(): print("Waiting for operation to complete...") - response = await operation.result() + response = (await operation).result() # Handle the response print(response) diff --git a/samples/generated_samples/dataplex_v1_generated_dataplex_service_update_lake_async.py b/samples/generated_samples/dataplex_v1_generated_dataplex_service_update_lake_async.py index 94c9435..f927720 100644 --- a/samples/generated_samples/dataplex_v1_generated_dataplex_service_update_lake_async.py +++ b/samples/generated_samples/dataplex_v1_generated_dataplex_service_update_lake_async.py @@ -47,7 +47,7 @@ async def sample_update_lake(): print("Waiting for operation to complete...") - response = await operation.result() + response = (await operation).result() # Handle the response print(response) diff --git a/samples/generated_samples/dataplex_v1_generated_dataplex_service_update_task_async.py b/samples/generated_samples/dataplex_v1_generated_dataplex_service_update_task_async.py index fbd9652..c26ad13 100644 --- a/samples/generated_samples/dataplex_v1_generated_dataplex_service_update_task_async.py +++ b/samples/generated_samples/dataplex_v1_generated_dataplex_service_update_task_async.py @@ -54,7 +54,7 @@ async def sample_update_task(): print("Waiting for operation to complete...") - response = await operation.result() + response = (await operation).result() # Handle the response print(response) diff --git a/samples/generated_samples/dataplex_v1_generated_dataplex_service_update_zone_async.py b/samples/generated_samples/dataplex_v1_generated_dataplex_service_update_zone_async.py index bafa0e3..281f76d 100644 --- a/samples/generated_samples/dataplex_v1_generated_dataplex_service_update_zone_async.py +++ b/samples/generated_samples/dataplex_v1_generated_dataplex_service_update_zone_async.py @@ -52,7 +52,7 @@ async def sample_update_zone(): print("Waiting for operation to complete...") - response = await operation.result() + response = (await operation).result() # Handle the response print(response) diff --git a/setup.py b/setup.py index 39fac0f..d994e54 100644 --- a/setup.py +++ b/setup.py @@ -36,7 +36,7 @@ release_status = "Development Status :: 5 - Production/Stable" dependencies = [ - "google-api-core[grpc] >= 1.33.2, <3.0.0dev,!=2.0.*,!=2.1.*,!=2.2.*,!=2.3.*,!=2.4.*,!=2.5.*,!=2.6.*,!=2.7.*", + "google-api-core[grpc] >= 1.34.0, <3.0.0dev,!=2.0.*,!=2.1.*,!=2.2.*,!=2.3.*,!=2.4.*,!=2.5.*,!=2.6.*,!=2.7.*,!=2.8.*,!=2.9.*,!=2.10.*", "proto-plus >= 1.22.0, <2.0.0dev", "protobuf>=3.19.5,<5.0.0dev,!=3.20.0,!=3.20.1,!=4.21.0,!=4.21.1,!=4.21.2,!=4.21.3,!=4.21.4,!=4.21.5", "grpc-google-iam-v1 >= 0.12.4, < 1.0.0dev", diff --git a/testing/constraints-3.7.txt b/testing/constraints-3.7.txt index 883a87a..2beecf9 100644 --- a/testing/constraints-3.7.txt +++ b/testing/constraints-3.7.txt @@ -4,7 +4,7 @@ # Pin the version to the lower bound. # e.g., if setup.py has "google-cloud-foo >= 1.14.0, < 2.0.0dev", # Then this file should have google-cloud-foo==1.14.0 -google-api-core==1.33.2 +google-api-core==1.34.0 proto-plus==1.22.0 protobuf==3.19.5 grpc-google-iam-v1==0.12.4 From 651a42b15cb23e59b8454c73cf090e3d7aaf4ff0 Mon Sep 17 00:00:00 2001 From: "gcf-owl-bot[bot]" <78513119+gcf-owl-bot[bot]@users.noreply.github.com> Date: Thu, 8 Dec 2022 18:54:19 +0000 Subject: [PATCH 08/10] build(deps): bump certifi from 2022.9.24 to 2022.12.7 [autoapprove] (#112) Source-Link: https://togithub.com/googleapis/synthtool/commit/b4fe62efb5114b6738ad4b13d6f654f2bf4b7cc0 Post-Processor: gcr.io/cloud-devrel-public-resources/owlbot-python:latest@sha256:3bf87e47c2173d7eed42714589dc4da2c07c3268610f1e47f8e1a30decbfc7f1 --- .github/.OwlBot.lock.yaml | 2 +- .kokoro/requirements.txt | 6 +++--- .pre-commit-config.yaml | 2 +- 3 files changed, 5 insertions(+), 5 deletions(-) diff --git a/.github/.OwlBot.lock.yaml b/.github/.OwlBot.lock.yaml index 3f1ccc0..fccaa8e 100644 --- a/.github/.OwlBot.lock.yaml +++ b/.github/.OwlBot.lock.yaml @@ -13,4 +13,4 @@ # limitations under the License. docker: image: gcr.io/cloud-devrel-public-resources/owlbot-python:latest - digest: sha256:e6cbd61f1838d9ff6a31436dfc13717f372a7482a82fc1863ca954ec47bff8c8 + digest: sha256:3bf87e47c2173d7eed42714589dc4da2c07c3268610f1e47f8e1a30decbfc7f1 diff --git a/.kokoro/requirements.txt b/.kokoro/requirements.txt index 9c1b9be..05dc467 100644 --- a/.kokoro/requirements.txt +++ b/.kokoro/requirements.txt @@ -20,9 +20,9 @@ cachetools==5.2.0 \ --hash=sha256:6a94c6402995a99c3970cc7e4884bb60b4a8639938157eeed436098bf9831757 \ --hash=sha256:f9f17d2aec496a9aa6b76f53e3b614c965223c061982d434d160f930c698a9db # via google-auth -certifi==2022.9.24 \ - --hash=sha256:0d9c601124e5a6ba9712dbc60d9c53c21e34f5f641fe83002317394311bdce14 \ - --hash=sha256:90c1a32f1d68f940488354e36370f6cca89f0f106db09518524c88d6ed83f382 +certifi==2022.12.7 \ + --hash=sha256:35824b4c3a97115964b408844d64aa14db1cc518f6562e8d7261699d1350a9e3 \ + --hash=sha256:4ad3232f5e926d6718ec31cfc1fcadfde020920e278684144551c91769c7bc18 # via requests cffi==1.15.1 \ --hash=sha256:00a9ed42e88df81ffae7a8ab6d9356b371399b91dbdf0c3cb1e84c03a13aceb5 \ diff --git a/.pre-commit-config.yaml b/.pre-commit-config.yaml index 46d2371..5405cc8 100644 --- a/.pre-commit-config.yaml +++ b/.pre-commit-config.yaml @@ -25,7 +25,7 @@ repos: rev: 22.3.0 hooks: - id: black -- repo: https://gitlab.com/pycqa/flake8 +- repo: https://github.com/pycqa/flake8 rev: 3.9.2 hooks: - id: flake8 From 0dc28b3c0f2c5be59a279c9ff859607d25906e84 Mon Sep 17 00:00:00 2001 From: "gcf-owl-bot[bot]" <78513119+gcf-owl-bot[bot]@users.noreply.github.com> Date: Wed, 14 Dec 2022 13:48:36 -0500 Subject: [PATCH 09/10] docs: fix minor docstring formatting (#113) MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit * chore: reformatted the protos PiperOrigin-RevId: 494247971 Source-Link: https://github.com/googleapis/googleapis/commit/f7f499371afaa6fc236b67148aa4bd41943725a6 Source-Link: https://github.com/googleapis/googleapis-gen/commit/f2b7d97b7ba4713a324aab058cf56c9d7305eb21 Copy-Tag: eyJwIjoiLmdpdGh1Yi8uT3dsQm90LnlhbWwiLCJoIjoiZjJiN2Q5N2I3YmE0NzEzYTMyNGFhYjA1OGNmNTZjOWQ3MzA1ZWIyMSJ9 * 🦉 Updates from OwlBot post-processor See https://github.com/googleapis/repo-automation-bots/blob/main/packages/owl-bot/README.md * docs: fix minor docstring formatting PiperOrigin-RevId: 495317762 Source-Link: https://github.com/googleapis/googleapis/commit/d9dc42bf24866ac28c09489feb58590c838ed970 Source-Link: https://github.com/googleapis/googleapis-gen/commit/6947bc43478b20e5650b62894d0c64137ef39909 Copy-Tag: eyJwIjoiLmdpdGh1Yi8uT3dsQm90LnlhbWwiLCJoIjoiNjk0N2JjNDM0NzhiMjBlNTY1MGI2Mjg5NGQwYzY0MTM3ZWYzOTkwOSJ9 * 🦉 Updates from OwlBot post-processor See https://github.com/googleapis/repo-automation-bots/blob/main/packages/owl-bot/README.md Co-authored-by: Owl Bot --- google/cloud/dataplex_v1/types/analyze.py | 3 +-- google/cloud/dataplex_v1/types/resources.py | 8 ++++---- google/cloud/dataplex_v1/types/tasks.py | 7 ++++--- 3 files changed, 9 insertions(+), 9 deletions(-) diff --git a/google/cloud/dataplex_v1/types/analyze.py b/google/cloud/dataplex_v1/types/analyze.py index 798e670..1480350 100644 --- a/google/cloud/dataplex_v1/types/analyze.py +++ b/google/cloud/dataplex_v1/types/analyze.py @@ -190,8 +190,7 @@ class SessionSpec(proto.Message): pre-created and available for faster startup to enable interactive exploration use-cases. This defaults to False to avoid additional billed - charges. - These can only be set to True for the + charges. These can only be set to True for the environment with name set to "default", and with default configuration. """ diff --git a/google/cloud/dataplex_v1/types/resources.py b/google/cloud/dataplex_v1/types/resources.py index 15326cb..699d51e 100644 --- a/google/cloud/dataplex_v1/types/resources.py +++ b/google/cloud/dataplex_v1/types/resources.py @@ -352,8 +352,8 @@ class DiscoverySpec(proto.Message): the cron tab: "CRON_TZ=${IANA_TIME_ZONE}" or TZ=${IANA_TIME_ZONE}". The ${IANA_TIME_ZONE} may only be a valid string from IANA time zone database. For example, - "CRON_TZ=America/New_York 1 \* \* \* \*", or - "TZ=America/New_York 1 \* \* \* \*". + ``CRON_TZ=America/New_York 1 * * * *``, or + ``TZ=America/New_York 1 * * * *``. This field is a member of `oneof`_ ``trigger``. """ @@ -951,8 +951,8 @@ class DiscoverySpec(proto.Message): the cron tab: "CRON_TZ=${IANA_TIME_ZONE}" or TZ=${IANA_TIME_ZONE}". The ${IANA_TIME_ZONE} may only be a valid string from IANA time zone database. For example, - "CRON_TZ=America/New_York 1 \* \* \* \*", or - "TZ=America/New_York 1 \* \* \* \*". + ``CRON_TZ=America/New_York 1 * * * *``, or + ``TZ=America/New_York 1 * * * *``. This field is a member of `oneof`_ ``trigger``. """ diff --git a/google/cloud/dataplex_v1/types/tasks.py b/google/cloud/dataplex_v1/types/tasks.py index 0f09b9b..8e68a90 100644 --- a/google/cloud/dataplex_v1/types/tasks.py +++ b/google/cloud/dataplex_v1/types/tasks.py @@ -264,9 +264,10 @@ class TriggerSpec(proto.Message): to the cron tab, apply a prefix in the cron tab: "CRON_TZ=${IANA_TIME_ZONE}" or "TZ=${IANA_TIME_ZONE}". The ${IANA_TIME_ZONE} may only be a valid string from IANA time - zone database. For example, "CRON_TZ=America/New_York 1 \* - \* \* \*", or "TZ=America/New_York 1 \* \* \* \*". This - field is required for RECURRING tasks. + zone database. For example, + ``CRON_TZ=America/New_York 1 * * * *``, or + ``TZ=America/New_York 1 * * * *``. This field is required + for RECURRING tasks. This field is a member of `oneof`_ ``trigger``. """ From 0c1c55732b79383470a59fbdb7cddc31539fe228 Mon Sep 17 00:00:00 2001 From: "release-please[bot]" <55107282+release-please[bot]@users.noreply.github.com> Date: Thu, 15 Dec 2022 15:11:46 -0500 Subject: [PATCH 10/10] chore(main): release 1.2.0 (#101) Co-authored-by: release-please[bot] <55107282+release-please[bot]@users.noreply.github.com> --- .release-please-manifest.json | 2 +- CHANGELOG.md | 25 +++++++++++++++++++ google/cloud/dataplex/gapic_version.py | 2 +- google/cloud/dataplex_v1/gapic_version.py | 2 +- ...pet_metadata_google.cloud.dataplex.v1.json | 2 +- 5 files changed, 29 insertions(+), 4 deletions(-) diff --git a/.release-please-manifest.json b/.release-please-manifest.json index 53b7bc9..c3f1463 100644 --- a/.release-please-manifest.json +++ b/.release-please-manifest.json @@ -1,3 +1,3 @@ { - ".": "1.1.2" + ".": "1.2.0" } diff --git a/CHANGELOG.md b/CHANGELOG.md index 2bb840a..9576124 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,30 @@ # Changelog +## [1.2.0](https://github.com/googleapis/python-dataplex/compare/v1.1.2...v1.2.0) (2022-12-15) + + +### Features + +* Add support for `google.cloud.dataplex.__version__` ([18e2a32](https://github.com/googleapis/python-dataplex/commit/18e2a32c425f6a7ca0684392a796e18547ea408a)) +* Add support for notebook tasks ([#100](https://github.com/googleapis/python-dataplex/issues/100)) ([64d9c48](https://github.com/googleapis/python-dataplex/commit/64d9c481df1c2737189dcb575c69f2968c0aa034)) +* Add typing to proto.Message based class attributes ([18e2a32](https://github.com/googleapis/python-dataplex/commit/18e2a32c425f6a7ca0684392a796e18547ea408a)) + + +### Bug Fixes + +* Add dict typing for client_options ([18e2a32](https://github.com/googleapis/python-dataplex/commit/18e2a32c425f6a7ca0684392a796e18547ea408a)) +* **deps:** Allow protobuf 3.19.5 ([#103](https://github.com/googleapis/python-dataplex/issues/103)) ([65adbb3](https://github.com/googleapis/python-dataplex/commit/65adbb31c94794f27a78b309550c519734a7b030)) +* **deps:** Require google-api-core >=1.34.0, >=2.11.0 ([1e94a02](https://github.com/googleapis/python-dataplex/commit/1e94a024024638d5d7d31f7bba4408b3f0b3d5d1)) +* Drop usage of pkg_resources ([1e94a02](https://github.com/googleapis/python-dataplex/commit/1e94a024024638d5d7d31f7bba4408b3f0b3d5d1)) +* Fix timeout default values ([1e94a02](https://github.com/googleapis/python-dataplex/commit/1e94a024024638d5d7d31f7bba4408b3f0b3d5d1)) + + +### Documentation + +* Fix minor docstring formatting ([#113](https://github.com/googleapis/python-dataplex/issues/113)) ([0dc28b3](https://github.com/googleapis/python-dataplex/commit/0dc28b3c0f2c5be59a279c9ff859607d25906e84)) +* **samples:** Snippetgen handling of repeated enum field ([18e2a32](https://github.com/googleapis/python-dataplex/commit/18e2a32c425f6a7ca0684392a796e18547ea408a)) +* **samples:** Snippetgen should call await on the operation coroutine before calling result ([1e94a02](https://github.com/googleapis/python-dataplex/commit/1e94a024024638d5d7d31f7bba4408b3f0b3d5d1)) + ## [1.1.2](https://github.com/googleapis/python-dataplex/compare/v1.1.1...v1.1.2) (2022-09-29) diff --git a/google/cloud/dataplex/gapic_version.py b/google/cloud/dataplex/gapic_version.py index 19bb9db..8f031bd 100644 --- a/google/cloud/dataplex/gapic_version.py +++ b/google/cloud/dataplex/gapic_version.py @@ -13,4 +13,4 @@ # See the License for the specific language governing permissions and # limitations under the License. # -__version__ = "1.1.2" # {x-release-please-version} +__version__ = "1.2.0" # {x-release-please-version} diff --git a/google/cloud/dataplex_v1/gapic_version.py b/google/cloud/dataplex_v1/gapic_version.py index 19bb9db..8f031bd 100644 --- a/google/cloud/dataplex_v1/gapic_version.py +++ b/google/cloud/dataplex_v1/gapic_version.py @@ -13,4 +13,4 @@ # See the License for the specific language governing permissions and # limitations under the License. # -__version__ = "1.1.2" # {x-release-please-version} +__version__ = "1.2.0" # {x-release-please-version} diff --git a/samples/generated_samples/snippet_metadata_google.cloud.dataplex.v1.json b/samples/generated_samples/snippet_metadata_google.cloud.dataplex.v1.json index 370f7c5..2b0f270 100644 --- a/samples/generated_samples/snippet_metadata_google.cloud.dataplex.v1.json +++ b/samples/generated_samples/snippet_metadata_google.cloud.dataplex.v1.json @@ -8,7 +8,7 @@ ], "language": "PYTHON", "name": "google-cloud-dataplex", - "version": "0.1.0" + "version": "1.2.0" }, "snippets": [ {