diff --git a/.coveragerc b/.coveragerc index da7caa1..bcaea1e 100644 --- a/.coveragerc +++ b/.coveragerc @@ -10,8 +10,3 @@ exclude_lines = pragma: NO COVER # Ignore debug-only repr def __repr__ - # Ignore pkg_resources exceptions. - # This is added at the module level as a safeguard for if someone - # generates the code and tries to run it without pip installing. This - # makes it virtually impossible to test properly. - except pkg_resources.DistributionNotFound diff --git a/.github/.OwlBot.lock.yaml b/.github/.OwlBot.lock.yaml index 3815c98..fccaa8e 100644 --- a/.github/.OwlBot.lock.yaml +++ b/.github/.OwlBot.lock.yaml @@ -13,4 +13,4 @@ # limitations under the License. docker: image: gcr.io/cloud-devrel-public-resources/owlbot-python:latest - digest: sha256:7a40313731a7cb1454eef6b33d3446ebb121836738dc3ab3d2d3ded5268c35b6 + digest: sha256:3bf87e47c2173d7eed42714589dc4da2c07c3268610f1e47f8e1a30decbfc7f1 diff --git a/.github/release-please.yml b/.github/release-please.yml index 466597e..528b274 100644 --- a/.github/release-please.yml +++ b/.github/release-please.yml @@ -1,2 +1,3 @@ releaseType: python handleGHRelease: true +manifest: true diff --git a/.github/workflows/docs.yml b/.github/workflows/docs.yml index 7092a13..e97d89e 100644 --- a/.github/workflows/docs.yml +++ b/.github/workflows/docs.yml @@ -12,7 +12,7 @@ jobs: - name: Setup Python uses: actions/setup-python@v4 with: - python-version: "3.10" + python-version: "3.9" - name: Install nox run: | python -m pip install --upgrade setuptools pip wheel @@ -28,7 +28,7 @@ jobs: - name: Setup Python uses: actions/setup-python@v4 with: - python-version: "3.10" + python-version: "3.9" - name: Install nox run: | python -m pip install --upgrade setuptools pip wheel diff --git a/.github/workflows/lint.yml b/.github/workflows/lint.yml index d2aee5b..16d5a9e 100644 --- a/.github/workflows/lint.yml +++ b/.github/workflows/lint.yml @@ -12,7 +12,7 @@ jobs: - name: Setup Python uses: actions/setup-python@v4 with: - python-version: "3.10" + python-version: "3.8" - name: Install nox run: | python -m pip install --upgrade setuptools pip wheel diff --git a/.github/workflows/unittest.yml b/.github/workflows/unittest.yml index 87ade4d..23000c0 100644 --- a/.github/workflows/unittest.yml +++ b/.github/workflows/unittest.yml @@ -41,7 +41,7 @@ jobs: - name: Setup Python uses: actions/setup-python@v4 with: - python-version: "3.10" + python-version: "3.8" - name: Install coverage run: | python -m pip install --upgrade setuptools pip wheel diff --git a/.kokoro/docker/docs/Dockerfile b/.kokoro/docker/docs/Dockerfile index 238b87b..f8137d0 100644 --- a/.kokoro/docker/docs/Dockerfile +++ b/.kokoro/docker/docs/Dockerfile @@ -60,16 +60,16 @@ RUN apt-get update \ && rm -rf /var/lib/apt/lists/* \ && rm -f /var/cache/apt/archives/*.deb -###################### Install python 3.8.11 +###################### Install python 3.9.13 -# Download python 3.8.11 -RUN wget https://www.python.org/ftp/python/3.8.11/Python-3.8.11.tgz +# Download python 3.9.13 +RUN wget https://www.python.org/ftp/python/3.9.13/Python-3.9.13.tgz # Extract files -RUN tar -xvf Python-3.8.11.tgz +RUN tar -xvf Python-3.9.13.tgz -# Install python 3.8.11 -RUN ./Python-3.8.11/configure --enable-optimizations +# Install python 3.9.13 +RUN ./Python-3.9.13/configure --enable-optimizations RUN make altinstall ###################### Install pip diff --git a/.kokoro/requirements.in b/.kokoro/requirements.in index 7718391..cbd7e77 100644 --- a/.kokoro/requirements.in +++ b/.kokoro/requirements.in @@ -5,4 +5,6 @@ typing-extensions twine wheel setuptools -nox \ No newline at end of file +nox +charset-normalizer<3 +click<8.1.0 diff --git a/.kokoro/requirements.txt b/.kokoro/requirements.txt index d15994b..05dc467 100644 --- a/.kokoro/requirements.txt +++ b/.kokoro/requirements.txt @@ -20,9 +20,9 @@ cachetools==5.2.0 \ --hash=sha256:6a94c6402995a99c3970cc7e4884bb60b4a8639938157eeed436098bf9831757 \ --hash=sha256:f9f17d2aec496a9aa6b76f53e3b614c965223c061982d434d160f930c698a9db # via google-auth -certifi==2022.6.15 \ - --hash=sha256:84c85a9078b11105f04f3036a9482ae10e4621616db313fe045dd24743a0820d \ - --hash=sha256:fe86415d55e84719d75f8b69414f6438ac3547d2078ab91b67e779ef69378412 +certifi==2022.12.7 \ + --hash=sha256:35824b4c3a97115964b408844d64aa14db1cc518f6562e8d7261699d1350a9e3 \ + --hash=sha256:4ad3232f5e926d6718ec31cfc1fcadfde020920e278684144551c91769c7bc18 # via requests cffi==1.15.1 \ --hash=sha256:00a9ed42e88df81ffae7a8ab6d9356b371399b91dbdf0c3cb1e84c03a13aceb5 \ @@ -93,11 +93,14 @@ cffi==1.15.1 \ charset-normalizer==2.1.1 \ --hash=sha256:5a3d016c7c547f69d6f81fb0db9449ce888b418b5b9952cc5e6e66843e9dd845 \ --hash=sha256:83e9a75d1911279afd89352c68b45348559d1fc0506b054b346651b5e7fee29f - # via requests + # via + # -r requirements.in + # requests click==8.0.4 \ --hash=sha256:6a7a62563bbfabfda3a38f3023a1db4a35978c0abd76f6c9605ecd6554d6d9b1 \ --hash=sha256:8458d7b1287c5fb128c90e23381cf99dcde74beaf6c7ff6384ce84d6fe090adb # via + # -r requirements.in # gcp-docuploader # gcp-releasetool colorlog==6.7.0 \ @@ -110,29 +113,33 @@ commonmark==0.9.1 \ --hash=sha256:452f9dc859be7f06631ddcb328b6919c67984aca654e5fefb3914d54691aed60 \ --hash=sha256:da2f38c92590f83de410ba1a3cbceafbc74fee9def35f9251ba9a971d6d66fd9 # via rich -cryptography==37.0.4 \ - --hash=sha256:190f82f3e87033821828f60787cfa42bff98404483577b591429ed99bed39d59 \ - --hash=sha256:2be53f9f5505673eeda5f2736bea736c40f051a739bfae2f92d18aed1eb54596 \ - --hash=sha256:30788e070800fec9bbcf9faa71ea6d8068f5136f60029759fd8c3efec3c9dcb3 \ - --hash=sha256:3d41b965b3380f10e4611dbae366f6dc3cefc7c9ac4e8842a806b9672ae9add5 \ - --hash=sha256:4c590ec31550a724ef893c50f9a97a0c14e9c851c85621c5650d699a7b88f7ab \ - --hash=sha256:549153378611c0cca1042f20fd9c5030d37a72f634c9326e225c9f666d472884 \ - --hash=sha256:63f9c17c0e2474ccbebc9302ce2f07b55b3b3fcb211ded18a42d5764f5c10a82 \ - --hash=sha256:6bc95ed67b6741b2607298f9ea4932ff157e570ef456ef7ff0ef4884a134cc4b \ - --hash=sha256:7099a8d55cd49b737ffc99c17de504f2257e3787e02abe6d1a6d136574873441 \ - --hash=sha256:75976c217f10d48a8b5a8de3d70c454c249e4b91851f6838a4e48b8f41eb71aa \ - --hash=sha256:7bc997818309f56c0038a33b8da5c0bfbb3f1f067f315f9abd6fc07ad359398d \ - --hash=sha256:80f49023dd13ba35f7c34072fa17f604d2f19bf0989f292cedf7ab5770b87a0b \ - --hash=sha256:91ce48d35f4e3d3f1d83e29ef4a9267246e6a3be51864a5b7d2247d5086fa99a \ - --hash=sha256:a958c52505c8adf0d3822703078580d2c0456dd1d27fabfb6f76fe63d2971cd6 \ - --hash=sha256:b62439d7cd1222f3da897e9a9fe53bbf5c104fff4d60893ad1355d4c14a24157 \ - --hash=sha256:b7f8dd0d4c1f21759695c05a5ec8536c12f31611541f8904083f3dc582604280 \ - --hash=sha256:d204833f3c8a33bbe11eda63a54b1aad7aa7456ed769a982f21ec599ba5fa282 \ - --hash=sha256:e007f052ed10cc316df59bc90fbb7ff7950d7e2919c9757fd42a2b8ecf8a5f67 \ - --hash=sha256:f2dcb0b3b63afb6df7fd94ec6fbddac81b5492513f7b0436210d390c14d46ee8 \ - --hash=sha256:f721d1885ecae9078c3f6bbe8a88bc0786b6e749bf32ccec1ef2b18929a05046 \ - --hash=sha256:f7a6de3e98771e183645181b3627e2563dcde3ce94a9e42a3f427d2255190327 \ - --hash=sha256:f8c0a6e9e1dd3eb0414ba320f85da6b0dcbd543126e30fcc546e7372a7fbf3b9 +cryptography==38.0.3 \ + --hash=sha256:068147f32fa662c81aebab95c74679b401b12b57494872886eb5c1139250ec5d \ + --hash=sha256:06fc3cc7b6f6cca87bd56ec80a580c88f1da5306f505876a71c8cfa7050257dd \ + --hash=sha256:25c1d1f19729fb09d42e06b4bf9895212292cb27bb50229f5aa64d039ab29146 \ + --hash=sha256:402852a0aea73833d982cabb6d0c3bb582c15483d29fb7085ef2c42bfa7e38d7 \ + --hash=sha256:4e269dcd9b102c5a3d72be3c45d8ce20377b8076a43cbed6f660a1afe365e436 \ + --hash=sha256:5419a127426084933076132d317911e3c6eb77568a1ce23c3ac1e12d111e61e0 \ + --hash=sha256:554bec92ee7d1e9d10ded2f7e92a5d70c1f74ba9524947c0ba0c850c7b011828 \ + --hash=sha256:5e89468fbd2fcd733b5899333bc54d0d06c80e04cd23d8c6f3e0542358c6060b \ + --hash=sha256:65535bc550b70bd6271984d9863a37741352b4aad6fb1b3344a54e6950249b55 \ + --hash=sha256:6ab9516b85bebe7aa83f309bacc5f44a61eeb90d0b4ec125d2d003ce41932d36 \ + --hash=sha256:6addc3b6d593cd980989261dc1cce38263c76954d758c3c94de51f1e010c9a50 \ + --hash=sha256:728f2694fa743a996d7784a6194da430f197d5c58e2f4e278612b359f455e4a2 \ + --hash=sha256:785e4056b5a8b28f05a533fab69febf5004458e20dad7e2e13a3120d8ecec75a \ + --hash=sha256:78cf5eefac2b52c10398a42765bfa981ce2372cbc0457e6bf9658f41ec3c41d8 \ + --hash=sha256:7f836217000342d448e1c9a342e9163149e45d5b5eca76a30e84503a5a96cab0 \ + --hash=sha256:8d41a46251bf0634e21fac50ffd643216ccecfaf3701a063257fe0b2be1b6548 \ + --hash=sha256:984fe150f350a3c91e84de405fe49e688aa6092b3525f407a18b9646f6612320 \ + --hash=sha256:9b24bcff7853ed18a63cfb0c2b008936a9554af24af2fb146e16d8e1aed75748 \ + --hash=sha256:b1b35d9d3a65542ed2e9d90115dfd16bbc027b3f07ee3304fc83580f26e43249 \ + --hash=sha256:b1b52c9e5f8aa2b802d48bd693190341fae201ea51c7a167d69fc48b60e8a959 \ + --hash=sha256:bbf203f1a814007ce24bd4d51362991d5cb90ba0c177a9c08825f2cc304d871f \ + --hash=sha256:be243c7e2bfcf6cc4cb350c0d5cdf15ca6383bbcb2a8ef51d3c9411a9d4386f0 \ + --hash=sha256:bfbe6ee19615b07a98b1d2287d6a6073f734735b49ee45b11324d85efc4d5cbd \ + --hash=sha256:c46837ea467ed1efea562bbeb543994c2d1f6e800785bd5a2c98bc096f5cb220 \ + --hash=sha256:dfb4f4dd568de1b6af9f4cda334adf7d72cf5bc052516e1b2608b683375dd95c \ + --hash=sha256:ed7b00096790213e09eb11c97cc6e2b757f15f3d2f85833cd2d3ec3fe37c1722 # via # gcp-releasetool # secretstorage @@ -148,23 +155,23 @@ filelock==3.8.0 \ --hash=sha256:55447caa666f2198c5b6b13a26d2084d26fa5b115c00d065664b2124680c4edc \ --hash=sha256:617eb4e5eedc82fc5f47b6d61e4d11cb837c56cb4544e39081099fa17ad109d4 # via virtualenv -gcp-docuploader==0.6.3 \ - --hash=sha256:ba8c9d76b3bbac54b0311c503a373b00edc2dc02d6d54ea9507045adb8e870f7 \ - --hash=sha256:c0f5aaa82ce1854a386197e4e359b120ad6d4e57ae2c812fce42219a3288026b +gcp-docuploader==0.6.4 \ + --hash=sha256:01486419e24633af78fd0167db74a2763974765ee8078ca6eb6964d0ebd388af \ + --hash=sha256:70861190c123d907b3b067da896265ead2eeb9263969d6955c9e0bb091b5ccbf # via -r requirements.in -gcp-releasetool==1.8.7 \ - --hash=sha256:3d2a67c9db39322194afb3b427e9cb0476ce8f2a04033695f0aeb63979fc2b37 \ - --hash=sha256:5e4d28f66e90780d77f3ecf1e9155852b0c3b13cbccb08ab07e66b2357c8da8d +gcp-releasetool==1.10.0 \ + --hash=sha256:72a38ca91b59c24f7e699e9227c90cbe4dd71b789383cb0164b088abae294c83 \ + --hash=sha256:8c7c99320208383d4bb2b808c6880eb7a81424afe7cdba3c8d84b25f4f0e097d # via -r requirements.in -google-api-core==2.8.2 \ - --hash=sha256:06f7244c640322b508b125903bb5701bebabce8832f85aba9335ec00b3d02edc \ - --hash=sha256:93c6a91ccac79079ac6bbf8b74ee75db970cc899278b97d53bc012f35908cf50 +google-api-core==2.10.2 \ + --hash=sha256:10c06f7739fe57781f87523375e8e1a3a4674bf6392cd6131a3222182b971320 \ + --hash=sha256:34f24bd1d5f72a8c4519773d99ca6bf080a6c4e041b4e9f024fe230191dda62e # via # google-cloud-core # google-cloud-storage -google-auth==2.11.0 \ - --hash=sha256:be62acaae38d0049c21ca90f27a23847245c9f161ff54ede13af2cb6afecbac9 \ - --hash=sha256:ed65ecf9f681832298e29328e1ef0a3676e3732b2e56f41532d45f70a22de0fb +google-auth==2.14.1 \ + --hash=sha256:ccaa901f31ad5cbb562615eb8b664b3dd0bf5404a67618e642307f00613eda4d \ + --hash=sha256:f5d8701633bebc12e0deea4df8abd8aff31c28b355360597f7f2ee60f2e4d016 # via # gcp-releasetool # google-api-core @@ -174,76 +181,102 @@ google-cloud-core==2.3.2 \ --hash=sha256:8417acf6466be2fa85123441696c4badda48db314c607cf1e5d543fa8bdc22fe \ --hash=sha256:b9529ee7047fd8d4bf4a2182de619154240df17fbe60ead399078c1ae152af9a # via google-cloud-storage -google-cloud-storage==2.5.0 \ - --hash=sha256:19a26c66c317ce542cea0830b7e787e8dac2588b6bfa4d3fd3b871ba16305ab0 \ - --hash=sha256:382f34b91de2212e3c2e7b40ec079d27ee2e3dbbae99b75b1bcd8c63063ce235 +google-cloud-storage==2.6.0 \ + --hash=sha256:104ca28ae61243b637f2f01455cc8a05e8f15a2a18ced96cb587241cdd3820f5 \ + --hash=sha256:4ad0415ff61abdd8bb2ae81c1f8f7ec7d91a1011613f2db87c614c550f97bfe9 # via gcp-docuploader -google-crc32c==1.3.0 \ - --hash=sha256:04e7c220798a72fd0f08242bc8d7a05986b2a08a0573396187fd32c1dcdd58b3 \ - --hash=sha256:05340b60bf05b574159e9bd940152a47d38af3fb43803ffe71f11d704b7696a6 \ - --hash=sha256:12674a4c3b56b706153a358eaa1018c4137a5a04635b92b4652440d3d7386206 \ - --hash=sha256:127f9cc3ac41b6a859bd9dc4321097b1a4f6aa7fdf71b4f9227b9e3ebffb4422 \ - --hash=sha256:13af315c3a0eec8bb8b8d80b8b128cb3fcd17d7e4edafc39647846345a3f003a \ - --hash=sha256:1926fd8de0acb9d15ee757175ce7242e235482a783cd4ec711cc999fc103c24e \ - --hash=sha256:226f2f9b8e128a6ca6a9af9b9e8384f7b53a801907425c9a292553a3a7218ce0 \ - --hash=sha256:276de6273eb074a35bc598f8efbc00c7869c5cf2e29c90748fccc8c898c244df \ - --hash=sha256:318f73f5484b5671f0c7f5f63741ab020a599504ed81d209b5c7129ee4667407 \ - --hash=sha256:3bbce1be3687bbfebe29abdb7631b83e6b25da3f4e1856a1611eb21854b689ea \ - --hash=sha256:42ae4781333e331a1743445931b08ebdad73e188fd554259e772556fc4937c48 \ - --hash=sha256:58be56ae0529c664cc04a9c76e68bb92b091e0194d6e3c50bea7e0f266f73713 \ - --hash=sha256:5da2c81575cc3ccf05d9830f9e8d3c70954819ca9a63828210498c0774fda1a3 \ - --hash=sha256:6311853aa2bba4064d0c28ca54e7b50c4d48e3de04f6770f6c60ebda1e975267 \ - --hash=sha256:650e2917660e696041ab3dcd7abac160b4121cd9a484c08406f24c5964099829 \ - --hash=sha256:6a4db36f9721fdf391646685ecffa404eb986cbe007a3289499020daf72e88a2 \ - --hash=sha256:779cbf1ce375b96111db98fca913c1f5ec11b1d870e529b1dc7354b2681a8c3a \ - --hash=sha256:7f6fe42536d9dcd3e2ffb9d3053f5d05221ae3bbcefbe472bdf2c71c793e3183 \ - --hash=sha256:891f712ce54e0d631370e1f4997b3f182f3368179198efc30d477c75d1f44942 \ - --hash=sha256:95c68a4b9b7828ba0428f8f7e3109c5d476ca44996ed9a5f8aac6269296e2d59 \ - --hash=sha256:96a8918a78d5d64e07c8ea4ed2bc44354e3f93f46a4866a40e8db934e4c0d74b \ - --hash=sha256:9c3cf890c3c0ecfe1510a452a165431b5831e24160c5fcf2071f0f85ca5a47cd \ - --hash=sha256:9f58099ad7affc0754ae42e6d87443299f15d739b0ce03c76f515153a5cda06c \ - --hash=sha256:a0b9e622c3b2b8d0ce32f77eba617ab0d6768b82836391e4f8f9e2074582bf02 \ - --hash=sha256:a7f9cbea4245ee36190f85fe1814e2d7b1e5f2186381b082f5d59f99b7f11328 \ - --hash=sha256:bab4aebd525218bab4ee615786c4581952eadc16b1ff031813a2fd51f0cc7b08 \ - --hash=sha256:c124b8c8779bf2d35d9b721e52d4adb41c9bfbde45e6a3f25f0820caa9aba73f \ - --hash=sha256:c9da0a39b53d2fab3e5467329ed50e951eb91386e9d0d5b12daf593973c3b168 \ - --hash=sha256:ca60076c388728d3b6ac3846842474f4250c91efbfe5afa872d3ffd69dd4b318 \ - --hash=sha256:cb6994fff247987c66a8a4e550ef374671c2b82e3c0d2115e689d21e511a652d \ - --hash=sha256:d1c1d6236feab51200272d79b3d3e0f12cf2cbb12b208c835b175a21efdb0a73 \ - --hash=sha256:dd7760a88a8d3d705ff562aa93f8445ead54f58fd482e4f9e2bafb7e177375d4 \ - --hash=sha256:dda4d8a3bb0b50f540f6ff4b6033f3a74e8bf0bd5320b70fab2c03e512a62812 \ - --hash=sha256:e0f1ff55dde0ebcfbef027edc21f71c205845585fffe30d4ec4979416613e9b3 \ - --hash=sha256:e7a539b9be7b9c00f11ef16b55486141bc2cdb0c54762f84e3c6fc091917436d \ - --hash=sha256:eb0b14523758e37802f27b7f8cd973f5f3d33be7613952c0df904b68c4842f0e \ - --hash=sha256:ed447680ff21c14aaceb6a9f99a5f639f583ccfe4ce1a5e1d48eb41c3d6b3217 \ - --hash=sha256:f52a4ad2568314ee713715b1e2d79ab55fab11e8b304fd1462ff5cccf4264b3e \ - --hash=sha256:fbd60c6aaa07c31d7754edbc2334aef50601b7f1ada67a96eb1eb57c7c72378f \ - --hash=sha256:fc28e0db232c62ca0c3600884933178f0825c99be4474cdd645e378a10588125 \ - --hash=sha256:fe31de3002e7b08eb20823b3735b97c86c5926dd0581c7710a680b418a8709d4 \ - --hash=sha256:fec221a051150eeddfdfcff162e6db92c65ecf46cb0f7bb1bf812a1520ec026b \ - --hash=sha256:ff71073ebf0e42258a42a0b34f2c09ec384977e7f6808999102eedd5b49920e3 +google-crc32c==1.5.0 \ + --hash=sha256:024894d9d3cfbc5943f8f230e23950cd4906b2fe004c72e29b209420a1e6b05a \ + --hash=sha256:02c65b9817512edc6a4ae7c7e987fea799d2e0ee40c53ec573a692bee24de876 \ + --hash=sha256:02ebb8bf46c13e36998aeaad1de9b48f4caf545e91d14041270d9dca767b780c \ + --hash=sha256:07eb3c611ce363c51a933bf6bd7f8e3878a51d124acfc89452a75120bc436289 \ + --hash=sha256:1034d91442ead5a95b5aaef90dbfaca8633b0247d1e41621d1e9f9db88c36298 \ + --hash=sha256:116a7c3c616dd14a3de8c64a965828b197e5f2d121fedd2f8c5585c547e87b02 \ + --hash=sha256:19e0a019d2c4dcc5e598cd4a4bc7b008546b0358bd322537c74ad47a5386884f \ + --hash=sha256:1c7abdac90433b09bad6c43a43af253e688c9cfc1c86d332aed13f9a7c7f65e2 \ + --hash=sha256:1e986b206dae4476f41bcec1faa057851f3889503a70e1bdb2378d406223994a \ + --hash=sha256:272d3892a1e1a2dbc39cc5cde96834c236d5327e2122d3aaa19f6614531bb6eb \ + --hash=sha256:278d2ed7c16cfc075c91378c4f47924c0625f5fc84b2d50d921b18b7975bd210 \ + --hash=sha256:2ad40e31093a4af319dadf503b2467ccdc8f67c72e4bcba97f8c10cb078207b5 \ + --hash=sha256:2e920d506ec85eb4ba50cd4228c2bec05642894d4c73c59b3a2fe20346bd00ee \ + --hash=sha256:3359fc442a743e870f4588fcf5dcbc1bf929df1fad8fb9905cd94e5edb02e84c \ + --hash=sha256:37933ec6e693e51a5b07505bd05de57eee12f3e8c32b07da7e73669398e6630a \ + --hash=sha256:398af5e3ba9cf768787eef45c803ff9614cc3e22a5b2f7d7ae116df8b11e3314 \ + --hash=sha256:3b747a674c20a67343cb61d43fdd9207ce5da6a99f629c6e2541aa0e89215bcd \ + --hash=sha256:461665ff58895f508e2866824a47bdee72497b091c730071f2b7575d5762ab65 \ + --hash=sha256:4c6fdd4fccbec90cc8a01fc00773fcd5fa28db683c116ee3cb35cd5da9ef6c37 \ + --hash=sha256:5829b792bf5822fd0a6f6eb34c5f81dd074f01d570ed7f36aa101d6fc7a0a6e4 \ + --hash=sha256:596d1f98fc70232fcb6590c439f43b350cb762fb5d61ce7b0e9db4539654cc13 \ + --hash=sha256:5ae44e10a8e3407dbe138984f21e536583f2bba1be9491239f942c2464ac0894 \ + --hash=sha256:635f5d4dd18758a1fbd1049a8e8d2fee4ffed124462d837d1a02a0e009c3ab31 \ + --hash=sha256:64e52e2b3970bd891309c113b54cf0e4384762c934d5ae56e283f9a0afcd953e \ + --hash=sha256:66741ef4ee08ea0b2cc3c86916ab66b6aef03768525627fd6a1b34968b4e3709 \ + --hash=sha256:67b741654b851abafb7bc625b6d1cdd520a379074e64b6a128e3b688c3c04740 \ + --hash=sha256:6ac08d24c1f16bd2bf5eca8eaf8304812f44af5cfe5062006ec676e7e1d50afc \ + --hash=sha256:6f998db4e71b645350b9ac28a2167e6632c239963ca9da411523bb439c5c514d \ + --hash=sha256:72218785ce41b9cfd2fc1d6a017dc1ff7acfc4c17d01053265c41a2c0cc39b8c \ + --hash=sha256:74dea7751d98034887dbd821b7aae3e1d36eda111d6ca36c206c44478035709c \ + --hash=sha256:759ce4851a4bb15ecabae28f4d2e18983c244eddd767f560165563bf9aefbc8d \ + --hash=sha256:77e2fd3057c9d78e225fa0a2160f96b64a824de17840351b26825b0848022906 \ + --hash=sha256:7c074fece789b5034b9b1404a1f8208fc2d4c6ce9decdd16e8220c5a793e6f61 \ + --hash=sha256:7c42c70cd1d362284289c6273adda4c6af8039a8ae12dc451dcd61cdabb8ab57 \ + --hash=sha256:7f57f14606cd1dd0f0de396e1e53824c371e9544a822648cd76c034d209b559c \ + --hash=sha256:83c681c526a3439b5cf94f7420471705bbf96262f49a6fe546a6db5f687a3d4a \ + --hash=sha256:8485b340a6a9e76c62a7dce3c98e5f102c9219f4cfbf896a00cf48caf078d438 \ + --hash=sha256:84e6e8cd997930fc66d5bb4fde61e2b62ba19d62b7abd7a69920406f9ecca946 \ + --hash=sha256:89284716bc6a5a415d4eaa11b1726d2d60a0cd12aadf5439828353662ede9dd7 \ + --hash=sha256:8b87e1a59c38f275c0e3676fc2ab6d59eccecfd460be267ac360cc31f7bcde96 \ + --hash=sha256:8f24ed114432de109aa9fd317278518a5af2d31ac2ea6b952b2f7782b43da091 \ + --hash=sha256:98cb4d057f285bd80d8778ebc4fde6b4d509ac3f331758fb1528b733215443ae \ + --hash=sha256:998679bf62b7fb599d2878aa3ed06b9ce688b8974893e7223c60db155f26bd8d \ + --hash=sha256:9ba053c5f50430a3fcfd36f75aff9caeba0440b2d076afdb79a318d6ca245f88 \ + --hash=sha256:9c99616c853bb585301df6de07ca2cadad344fd1ada6d62bb30aec05219c45d2 \ + --hash=sha256:a1fd716e7a01f8e717490fbe2e431d2905ab8aa598b9b12f8d10abebb36b04dd \ + --hash=sha256:a2355cba1f4ad8b6988a4ca3feed5bff33f6af2d7f134852cf279c2aebfde541 \ + --hash=sha256:b1f8133c9a275df5613a451e73f36c2aea4fe13c5c8997e22cf355ebd7bd0728 \ + --hash=sha256:b8667b48e7a7ef66afba2c81e1094ef526388d35b873966d8a9a447974ed9178 \ + --hash=sha256:ba1eb1843304b1e5537e1fca632fa894d6f6deca8d6389636ee5b4797affb968 \ + --hash=sha256:be82c3c8cfb15b30f36768797a640e800513793d6ae1724aaaafe5bf86f8f346 \ + --hash=sha256:c02ec1c5856179f171e032a31d6f8bf84e5a75c45c33b2e20a3de353b266ebd8 \ + --hash=sha256:c672d99a345849301784604bfeaeba4db0c7aae50b95be04dd651fd2a7310b93 \ + --hash=sha256:c6c777a480337ac14f38564ac88ae82d4cd238bf293f0a22295b66eb89ffced7 \ + --hash=sha256:cae0274952c079886567f3f4f685bcaf5708f0a23a5f5216fdab71f81a6c0273 \ + --hash=sha256:cd67cf24a553339d5062eff51013780a00d6f97a39ca062781d06b3a73b15462 \ + --hash=sha256:d3515f198eaa2f0ed49f8819d5732d70698c3fa37384146079b3799b97667a94 \ + --hash=sha256:d5280312b9af0976231f9e317c20e4a61cd2f9629b7bfea6a693d1878a264ebd \ + --hash=sha256:de06adc872bcd8c2a4e0dc51250e9e65ef2ca91be023b9d13ebd67c2ba552e1e \ + --hash=sha256:e1674e4307fa3024fc897ca774e9c7562c957af85df55efe2988ed9056dc4e57 \ + --hash=sha256:e2096eddb4e7c7bdae4bd69ad364e55e07b8316653234a56552d9c988bd2d61b \ + --hash=sha256:e560628513ed34759456a416bf86b54b2476c59144a9138165c9a1575801d0d9 \ + --hash=sha256:edfedb64740750e1a3b16152620220f51d58ff1b4abceb339ca92e934775c27a \ + --hash=sha256:f13cae8cc389a440def0c8c52057f37359014ccbc9dc1f0827936bcd367c6100 \ + --hash=sha256:f314013e7dcd5cf45ab1945d92e713eec788166262ae8deb2cfacd53def27325 \ + --hash=sha256:f583edb943cf2e09c60441b910d6a20b4d9d626c75a36c8fcac01a6c96c01183 \ + --hash=sha256:fd8536e902db7e365f49e7d9029283403974ccf29b13fc7028b97e2295b33556 \ + --hash=sha256:fe70e325aa68fa4b5edf7d1a4b6f691eb04bbccac0ace68e34820d283b5f80d4 # via google-resumable-media -google-resumable-media==2.3.3 \ - --hash=sha256:27c52620bd364d1c8116eaac4ea2afcbfb81ae9139fb3199652fcac1724bfb6c \ - --hash=sha256:5b52774ea7a829a8cdaa8bd2d4c3d4bc660c91b30857ab2668d0eb830f4ea8c5 +google-resumable-media==2.4.0 \ + --hash=sha256:2aa004c16d295c8f6c33b2b4788ba59d366677c0a25ae7382436cb30f776deaa \ + --hash=sha256:8d5518502f92b9ecc84ac46779bd4f09694ecb3ba38a3e7ca737a86d15cbca1f # via google-cloud-storage -googleapis-common-protos==1.56.4 \ - --hash=sha256:8eb2cbc91b69feaf23e32452a7ae60e791e09967d81d4fcc7fc388182d1bd394 \ - --hash=sha256:c25873c47279387cfdcbdafa36149887901d36202cb645a0e4f29686bf6e4417 +googleapis-common-protos==1.57.0 \ + --hash=sha256:27a849d6205838fb6cc3c1c21cb9800707a661bb21c6ce7fb13e99eb1f8a0c46 \ + --hash=sha256:a9f4a1d7f6d9809657b7f1316a1aa527f6664891531bcfcc13b6696e685f443c # via google-api-core -idna==3.3 \ - --hash=sha256:84d9dd047ffa80596e0f246e2eab0b391788b0503584e8945f2368256d2735ff \ - --hash=sha256:9d643ff0a55b762d5cdb124b8eaa99c66322e2157b69160bc32796e824360e6d +idna==3.4 \ + --hash=sha256:814f528e8dead7d329833b91c5faa87d60bf71824cd12a7530b5526063d02cb4 \ + --hash=sha256:90b77e79eaa3eba6de819a0c442c0b4ceefc341a7a2ab77d7562bf49f425c5c2 # via requests -importlib-metadata==4.12.0 \ - --hash=sha256:637245b8bab2b6502fcbc752cc4b7a6f6243bb02b31c5c26156ad103d3d45670 \ - --hash=sha256:7401a975809ea1fdc658c3aa4f78cc2195a0e019c5cbc4c06122884e9ae80c23 +importlib-metadata==5.0.0 \ + --hash=sha256:da31db32b304314d044d3c12c79bd59e307889b287ad12ff387b3500835fc2ab \ + --hash=sha256:ddb0e35065e8938f867ed4928d0ae5bf2a53b7773871bfe6bcc7e4fcdc7dea43 # via # -r requirements.in + # keyring # twine -jaraco-classes==3.2.2 \ - --hash=sha256:6745f113b0b588239ceb49532aa09c3ebb947433ce311ef2f8e3ad64ebb74594 \ - --hash=sha256:e6ef6fd3fcf4579a7a019d87d1e56a883f4e4c35cfe925f86731abc58804e647 +jaraco-classes==3.2.3 \ + --hash=sha256:2353de3288bc6b82120752201c6b1c1a14b058267fa424ed5ce5984e3b922158 \ + --hash=sha256:89559fa5c1d3c34eff6f631ad80bb21f378dbcbb35dd161fd2c6b93f5be2f98a # via keyring jeepney==0.8.0 \ --hash=sha256:5efe48d255973902f6badc3ce55e2aa6c5c3b3bc642059ef3a91247bcfcc5806 \ @@ -255,9 +288,9 @@ jinja2==3.1.2 \ --hash=sha256:31351a702a408a9e7595a8fc6150fc3f43bb6bf7e319770cbc0db9df9437e852 \ --hash=sha256:6088930bfe239f0e6710546ab9c19c9ef35e29792895fed6e6e31a023a182a61 # via gcp-releasetool -keyring==23.9.0 \ - --hash=sha256:4c32a31174faaee48f43a7e2c7e9c3216ec5e95acf22a2bebfb4a1d05056ee44 \ - --hash=sha256:98f060ec95ada2ab910c195a2d4317be6ef87936a766b239c46aa3c7aac4f0db +keyring==23.11.0 \ + --hash=sha256:3dd30011d555f1345dec2c262f0153f2f0ca6bca041fb1dc4588349bb4c0ac1e \ + --hash=sha256:ad192263e2cdd5f12875dedc2da13534359a7e760e77f8d04b50968a821c2361 # via # gcp-releasetool # twine @@ -303,9 +336,9 @@ markupsafe==2.1.1 \ --hash=sha256:f121a1420d4e173a5d96e47e9a0c0dcff965afdf1626d28de1460815f7c4ee7a \ --hash=sha256:fc7b548b17d238737688817ab67deebb30e8073c95749d55538ed473130ec0c7 # via jinja2 -more-itertools==8.14.0 \ - --hash=sha256:1bc4f91ee5b1b31ac7ceacc17c09befe6a40a503907baf9c839c229b5095cfd2 \ - --hash=sha256:c09443cd3d5438b8dafccd867a6bc1cb0894389e90cb53d227456b0b0bccb750 +more-itertools==9.0.0 \ + --hash=sha256:250e83d7e81d0c87ca6bd942e6aeab8cc9daa6096d12c5308f3f92fa5e5c1f41 \ + --hash=sha256:5a6257e40878ef0520b1803990e3e22303a41b5714006c32a3fd8304b26ea1ab # via jaraco-classes nox==2022.8.7 \ --hash=sha256:1b894940551dc5c389f9271d197ca5d655d40bdc6ccf93ed6880e4042760a34b \ @@ -321,34 +354,33 @@ pkginfo==1.8.3 \ --hash=sha256:848865108ec99d4901b2f7e84058b6e7660aae8ae10164e015a6dcf5b242a594 \ --hash=sha256:a84da4318dd86f870a9447a8c98340aa06216bfc6f2b7bdc4b8766984ae1867c # via twine -platformdirs==2.5.2 \ - --hash=sha256:027d8e83a2d7de06bbac4e5ef7e023c02b863d7ea5d079477e722bb41ab25788 \ - --hash=sha256:58c8abb07dcb441e6ee4b11d8df0ac856038f944ab98b7be6b27b2a3c7feef19 +platformdirs==2.5.4 \ + --hash=sha256:1006647646d80f16130f052404c6b901e80ee4ed6bef6792e1f238a8969106f7 \ + --hash=sha256:af0276409f9a02373d540bf8480021a048711d572745aef4b7842dad245eba10 # via virtualenv -protobuf==3.20.2 \ - --hash=sha256:03d76b7bd42ac4a6e109742a4edf81ffe26ffd87c5993126d894fe48a120396a \ - --hash=sha256:09e25909c4297d71d97612f04f41cea8fa8510096864f2835ad2f3b3df5a5559 \ - --hash=sha256:18e34a10ae10d458b027d7638a599c964b030c1739ebd035a1dfc0e22baa3bfe \ - --hash=sha256:291fb4307094bf5ccc29f424b42268640e00d5240bf0d9b86bf3079f7576474d \ - --hash=sha256:2c0b040d0b5d5d207936ca2d02f00f765906622c07d3fa19c23a16a8ca71873f \ - --hash=sha256:384164994727f274cc34b8abd41a9e7e0562801361ee77437099ff6dfedd024b \ - --hash=sha256:3cb608e5a0eb61b8e00fe641d9f0282cd0eedb603be372f91f163cbfbca0ded0 \ - --hash=sha256:5d9402bf27d11e37801d1743eada54372f986a372ec9679673bfcc5c60441151 \ - --hash=sha256:712dca319eee507a1e7df3591e639a2b112a2f4a62d40fe7832a16fd19151750 \ - --hash=sha256:7a5037af4e76c975b88c3becdf53922b5ffa3f2cddf657574a4920a3b33b80f3 \ - --hash=sha256:8228e56a865c27163d5d1d1771d94b98194aa6917bcfb6ce139cbfa8e3c27334 \ - --hash=sha256:84a1544252a933ef07bb0b5ef13afe7c36232a774affa673fc3636f7cee1db6c \ - --hash=sha256:84fe5953b18a383fd4495d375fe16e1e55e0a3afe7b4f7b4d01a3a0649fcda9d \ - --hash=sha256:9c673c8bfdf52f903081816b9e0e612186684f4eb4c17eeb729133022d6032e3 \ - --hash=sha256:9f876a69ca55aed879b43c295a328970306e8e80a263ec91cf6e9189243c613b \ - --hash=sha256:a9e5ae5a8e8985c67e8944c23035a0dff2c26b0f5070b2f55b217a1c33bbe8b1 \ - --hash=sha256:b4fdb29c5a7406e3f7ef176b2a7079baa68b5b854f364c21abe327bbeec01cdb \ - --hash=sha256:c184485e0dfba4dfd451c3bd348c2e685d6523543a0f91b9fd4ae90eb09e8422 \ - --hash=sha256:c9cdf251c582c16fd6a9f5e95836c90828d51b0069ad22f463761d27c6c19019 \ - --hash=sha256:e39cf61bb8582bda88cdfebc0db163b774e7e03364bbf9ce1ead13863e81e359 \ - --hash=sha256:e8fbc522303e09036c752a0afcc5c0603e917222d8bedc02813fd73b4b4ed804 \ - --hash=sha256:f34464ab1207114e73bba0794d1257c150a2b89b7a9faf504e00af7c9fd58978 \ - --hash=sha256:f52dabc96ca99ebd2169dadbe018824ebda08a795c7684a0b7d203a290f3adb0 +protobuf==3.20.3 \ + --hash=sha256:03038ac1cfbc41aa21f6afcbcd357281d7521b4157926f30ebecc8d4ea59dcb7 \ + --hash=sha256:28545383d61f55b57cf4df63eebd9827754fd2dc25f80c5253f9184235db242c \ + --hash=sha256:2e3427429c9cffebf259491be0af70189607f365c2f41c7c3764af6f337105f2 \ + --hash=sha256:398a9e0c3eaceb34ec1aee71894ca3299605fa8e761544934378bbc6c97de23b \ + --hash=sha256:44246bab5dd4b7fbd3c0c80b6f16686808fab0e4aca819ade6e8d294a29c7050 \ + --hash=sha256:447d43819997825d4e71bf5769d869b968ce96848b6479397e29fc24c4a5dfe9 \ + --hash=sha256:67a3598f0a2dcbc58d02dd1928544e7d88f764b47d4a286202913f0b2801c2e7 \ + --hash=sha256:74480f79a023f90dc6e18febbf7b8bac7508420f2006fabd512013c0c238f454 \ + --hash=sha256:819559cafa1a373b7096a482b504ae8a857c89593cf3a25af743ac9ecbd23480 \ + --hash=sha256:899dc660cd599d7352d6f10d83c95df430a38b410c1b66b407a6b29265d66469 \ + --hash=sha256:8c0c984a1b8fef4086329ff8dd19ac77576b384079247c770f29cc8ce3afa06c \ + --hash=sha256:9aae4406ea63d825636cc11ffb34ad3379335803216ee3a856787bcf5ccc751e \ + --hash=sha256:a7ca6d488aa8ff7f329d4c545b2dbad8ac31464f1d8b1c87ad1346717731e4db \ + --hash=sha256:b6cc7ba72a8850621bfec987cb72623e703b7fe2b9127a161ce61e61558ad905 \ + --hash=sha256:bf01b5720be110540be4286e791db73f84a2b721072a3711efff6c324cdf074b \ + --hash=sha256:c02ce36ec760252242a33967d51c289fd0e1c0e6e5cc9397e2279177716add86 \ + --hash=sha256:d9e4432ff660d67d775c66ac42a67cf2453c27cb4d738fc22cb53b5d84c135d4 \ + --hash=sha256:daa564862dd0d39c00f8086f88700fdbe8bc717e993a21e90711acfed02f2402 \ + --hash=sha256:de78575669dddf6099a8a0f46a27e82a1783c557ccc38ee620ed8cc96d3be7d7 \ + --hash=sha256:e64857f395505ebf3d2569935506ae0dfc4a15cb80dc25261176c784662cdcc4 \ + --hash=sha256:f4bd856d702e5b0d96a00ec6b307b0f51c1982c2bf9c0052cf9019e9a544ba99 \ + --hash=sha256:f4c42102bc82a51108e449cbb32b19b180022941c727bac0cfd50170341f16ee # via # gcp-docuploader # gcp-releasetool @@ -377,9 +409,9 @@ pygments==2.13.0 \ # via # readme-renderer # rich -pyjwt==2.4.0 \ - --hash=sha256:72d1d253f32dbd4f5c88eaf1fdc62f3a19f676ccbadb9dbc5d07e951b2b26daf \ - --hash=sha256:d42908208c699b3b973cbeb01a969ba6a96c821eefb1c5bfe4c390c01d67abba +pyjwt==2.6.0 \ + --hash=sha256:69285c7e31fc44f68a1feb309e948e0df53259d579295e6cfe2b1792329f05fd \ + --hash=sha256:d83c3d892a77bbb74d3e1a2cfa90afaadb60945205d1095d9221f04466f64c14 # via gcp-releasetool pyparsing==3.0.9 \ --hash=sha256:2b020ecf7d21b687f219b71ecad3631f644a47f01403fa1d1036b0c6416d70fb \ @@ -392,9 +424,9 @@ python-dateutil==2.8.2 \ --hash=sha256:0123cacc1627ae19ddf3c27a5de5bd67ee4586fbdd6440d9748f8abb483d3e86 \ --hash=sha256:961d03dc3453ebbc59dbdea9e4e11c5651520a876d0f4db161e8674aae935da9 # via gcp-releasetool -readme-renderer==37.0 \ - --hash=sha256:07b7ea234e03e58f77cc222e206e6abb8f4c0435becce5104794ee591f9301c5 \ - --hash=sha256:9fa416704703e509eeb900696751c908ddeb2011319d93700d8f18baff887a69 +readme-renderer==37.3 \ + --hash=sha256:cd653186dfc73055656f090f227f5cb22a046d7f71a841dfa305f55c9a513273 \ + --hash=sha256:f67a16caedfa71eef48a31b39708637a6f4664c4394801a7b0d6432d13907343 # via twine requests==2.28.1 \ --hash=sha256:7c5599b102feddaa661c826c56ab4fee28bfd17f5abca1ebbe3e7f19d7c97983 \ @@ -405,17 +437,17 @@ requests==2.28.1 \ # google-cloud-storage # requests-toolbelt # twine -requests-toolbelt==0.9.1 \ - --hash=sha256:380606e1d10dc85c3bd47bf5a6095f815ec007be7a8b69c878507068df059e6f \ - --hash=sha256:968089d4584ad4ad7c171454f0a5c6dac23971e9472521ea3b6d49d610aa6fc0 +requests-toolbelt==0.10.1 \ + --hash=sha256:18565aa58116d9951ac39baa288d3adb5b3ff975c4f25eee78555d89e8f247f7 \ + --hash=sha256:62e09f7ff5ccbda92772a29f394a49c3ad6cb181d568b1337626b2abb628a63d # via twine rfc3986==2.0.0 \ --hash=sha256:50b1502b60e289cb37883f3dfd34532b8873c7de9f49bb546641ce9cbd256ebd \ --hash=sha256:97aacf9dbd4bfd829baad6e6309fa6573aaf1be3f6fa735c8ab05e46cecb261c # via twine -rich==12.5.1 \ - --hash=sha256:2eb4e6894cde1e017976d2975ac210ef515d7548bc595ba20e195fb9628acdeb \ - --hash=sha256:63a5c5ce3673d3d5fbbf23cd87e11ab84b6b451436f1b7f19ec54b6bc36ed7ca +rich==12.6.0 \ + --hash=sha256:a4eb26484f2c82589bd9a17c73d32a010b1e29d89f1604cd9bf3a2097b81bb5e \ + --hash=sha256:ba3a3775974105c221d31141f2c116f4fd65c5ceb0698657a11e9f295ec93fd0 # via twine rsa==4.9 \ --hash=sha256:90260d9058e514786967344d0ef75fa8727eed8a7d2e43ce9f4bcf1b536174f7 \ @@ -437,9 +469,9 @@ twine==4.0.1 \ --hash=sha256:42026c18e394eac3e06693ee52010baa5313e4811d5a11050e7d48436cf41b9e \ --hash=sha256:96b1cf12f7ae611a4a40b6ae8e9570215daff0611828f5fe1f37a16255ab24a0 # via -r requirements.in -typing-extensions==4.3.0 \ - --hash=sha256:25642c956049920a5aa49edcdd6ab1e06d7e5d467fc00e0506c44ac86fbfca02 \ - --hash=sha256:e6d2677a32f47fc7eb2795db1dd15c1f34eff616bcaf2cfb5e997f854fa1c4a6 +typing-extensions==4.4.0 \ + --hash=sha256:1511434bb92bf8dd198c12b1cc812e800d4181cfcb867674e0f8279cc93087aa \ + --hash=sha256:16fa4864408f655d35ec496218b85f79b3437c829e93320c7c9215ccfd92489e # via -r requirements.in urllib3==1.26.12 \ --hash=sha256:3fa96cf423e6987997fc326ae8df396db2a8b7c667747d47ddd8ecba91f4a74e \ @@ -447,25 +479,25 @@ urllib3==1.26.12 \ # via # requests # twine -virtualenv==20.16.4 \ - --hash=sha256:014f766e4134d0008dcaa1f95bafa0fb0f575795d07cae50b1bee514185d6782 \ - --hash=sha256:035ed57acce4ac35c82c9d8802202b0e71adac011a511ff650cbcf9635006a22 +virtualenv==20.16.7 \ + --hash=sha256:8691e3ff9387f743e00f6bb20f70121f5e4f596cae754531f2b3b3a1b1ac696e \ + --hash=sha256:efd66b00386fdb7dbe4822d172303f40cd05e50e01740b19ea42425cbe653e29 # via nox webencodings==0.5.1 \ --hash=sha256:a0af1213f3c2226497a97e2b3aa01a7e4bee4f403f95be16fc9acd2947514a78 \ --hash=sha256:b36a1c245f2d304965eb4e0a82848379241dc04b865afcc4aab16748587e1923 # via bleach -wheel==0.37.1 \ - --hash=sha256:4bdcd7d840138086126cd09254dc6195fb4fc6f01c050a1d7236f2630db1d22a \ - --hash=sha256:e9a504e793efbca1b8e0e9cb979a249cf4a0a7b5b8c9e8b65a5e39d49529c1c4 +wheel==0.38.4 \ + --hash=sha256:965f5259b566725405b05e7cf774052044b1ed30119b5d586b2703aafe8719ac \ + --hash=sha256:b60533f3f5d530e971d6737ca6d58681ee434818fab630c83a734bb10c083ce8 # via -r requirements.in -zipp==3.8.1 \ - --hash=sha256:05b45f1ee8f807d0cc928485ca40a07cb491cf092ff587c0df9cb1fd154848d2 \ - --hash=sha256:47c40d7fe183a6f21403a199b3e4192cca5774656965b0a4988ad2f8feb5f009 +zipp==3.10.0 \ + --hash=sha256:4fcb6f278987a6605757302a6e40e896257570d11c51628968ccb2a47e80c6c1 \ + --hash=sha256:7a7262fd930bd3e36c50b9a64897aec3fafff3dfdeec9623ae22b40e93f99bb8 # via importlib-metadata # The following packages are considered to be unsafe in a requirements file: -setuptools==65.2.0 \ - --hash=sha256:7f4bc85450898a09f76ebf28b72fa25bc7111f6c7d665d514a60bba9c75ef2a9 \ - --hash=sha256:a3ca5857c89f82f5c9410e8508cb32f4872a3bafd4aa7ae122a24ca33bccc750 +setuptools==65.5.1 \ + --hash=sha256:d0b9a8433464d5800cbe05094acf5c6d52a91bfac9b52bcfc4d41382be5d5d31 \ + --hash=sha256:e197a19aa8ec9722928f2206f8de752def0e4c9fc6953527360d1c36d94ddb2f # via -r requirements.in diff --git a/.pre-commit-config.yaml b/.pre-commit-config.yaml index 46d2371..5405cc8 100644 --- a/.pre-commit-config.yaml +++ b/.pre-commit-config.yaml @@ -25,7 +25,7 @@ repos: rev: 22.3.0 hooks: - id: black -- repo: https://gitlab.com/pycqa/flake8 +- repo: https://github.com/pycqa/flake8 rev: 3.9.2 hooks: - id: flake8 diff --git a/.release-please-manifest.json b/.release-please-manifest.json new file mode 100644 index 0000000..e7ca613 --- /dev/null +++ b/.release-please-manifest.json @@ -0,0 +1,3 @@ +{ + ".": "0.7.0" +} diff --git a/CHANGELOG.md b/CHANGELOG.md index fb9b915..19e3df8 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,27 @@ # Changelog +## [0.7.0](https://github.com/googleapis/python-dataflow-client/compare/v0.6.2...v0.7.0) (2022-12-15) + + +### Features + +* Add support for `google.cloud.dataflow.__version__` ([5f36251](https://github.com/googleapis/python-dataflow-client/commit/5f362512a1c36b1c5ce27fa175afb57fc5b375bc)) +* Add typing to proto.Message based class attributes ([5f36251](https://github.com/googleapis/python-dataflow-client/commit/5f362512a1c36b1c5ce27fa175afb57fc5b375bc)) + + +### Bug Fixes + +* Add dict typing for client_options ([5f36251](https://github.com/googleapis/python-dataflow-client/commit/5f362512a1c36b1c5ce27fa175afb57fc5b375bc)) +* **deps:** Require google-api-core >=1.34.0, >=2.11.0 ([9b9083c](https://github.com/googleapis/python-dataflow-client/commit/9b9083c177dd2b19cf218a4c0574159b9c162135)) +* Drop usage of pkg_resources ([9b9083c](https://github.com/googleapis/python-dataflow-client/commit/9b9083c177dd2b19cf218a4c0574159b9c162135)) +* Fix timeout default values ([9b9083c](https://github.com/googleapis/python-dataflow-client/commit/9b9083c177dd2b19cf218a4c0574159b9c162135)) + + +### Documentation + +* **samples:** Snippetgen handling of repeated enum field ([5f36251](https://github.com/googleapis/python-dataflow-client/commit/5f362512a1c36b1c5ce27fa175afb57fc5b375bc)) +* **samples:** Snippetgen should call await on the operation coroutine before calling result ([9b9083c](https://github.com/googleapis/python-dataflow-client/commit/9b9083c177dd2b19cf218a4c0574159b9c162135)) + ## [0.6.2](https://github.com/googleapis/python-dataflow-client/compare/v0.6.1...v0.6.2) (2022-10-10) diff --git a/docs/dataflow_v1beta3/types.rst b/docs/dataflow_v1beta3/types.rst index 368642a..a768d4a 100644 --- a/docs/dataflow_v1beta3/types.rst +++ b/docs/dataflow_v1beta3/types.rst @@ -3,5 +3,4 @@ Types for Google Cloud Dataflow v1beta3 API .. automodule:: google.cloud.dataflow_v1beta3.types :members: - :undoc-members: :show-inheritance: diff --git a/google/cloud/dataflow/__init__.py b/google/cloud/dataflow/__init__.py index 1c7d3b1..61e9177 100644 --- a/google/cloud/dataflow/__init__.py +++ b/google/cloud/dataflow/__init__.py @@ -13,6 +13,10 @@ # See the License for the specific language governing permissions and # limitations under the License. # +from google.cloud.dataflow import gapic_version as package_version + +__version__ = package_version.__version__ + from google.cloud.dataflow_v1beta3.services.flex_templates_service.async_client import ( FlexTemplatesServiceAsyncClient, diff --git a/google/cloud/dataflow/gapic_version.py b/google/cloud/dataflow/gapic_version.py new file mode 100644 index 0000000..e341813 --- /dev/null +++ b/google/cloud/dataflow/gapic_version.py @@ -0,0 +1,16 @@ +# -*- coding: utf-8 -*- +# Copyright 2022 Google LLC +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +# +__version__ = "0.7.0" # {x-release-please-version} diff --git a/google/cloud/dataflow_v1beta3/__init__.py b/google/cloud/dataflow_v1beta3/__init__.py index aa72412..db7fb4f 100644 --- a/google/cloud/dataflow_v1beta3/__init__.py +++ b/google/cloud/dataflow_v1beta3/__init__.py @@ -13,6 +13,10 @@ # See the License for the specific language governing permissions and # limitations under the License. # +from google.cloud.dataflow import gapic_version as package_version + +__version__ = package_version.__version__ + from .services.flex_templates_service import ( FlexTemplatesServiceAsyncClient, diff --git a/google/cloud/dataflow_v1beta3/gapic_version.py b/google/cloud/dataflow_v1beta3/gapic_version.py new file mode 100644 index 0000000..e341813 --- /dev/null +++ b/google/cloud/dataflow_v1beta3/gapic_version.py @@ -0,0 +1,16 @@ +# -*- coding: utf-8 -*- +# Copyright 2022 Google LLC +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +# +__version__ = "0.7.0" # {x-release-please-version} diff --git a/google/cloud/dataflow_v1beta3/services/flex_templates_service/async_client.py b/google/cloud/dataflow_v1beta3/services/flex_templates_service/async_client.py index a7a4f46..9e9ae8a 100644 --- a/google/cloud/dataflow_v1beta3/services/flex_templates_service/async_client.py +++ b/google/cloud/dataflow_v1beta3/services/flex_templates_service/async_client.py @@ -16,7 +16,17 @@ from collections import OrderedDict import functools import re -from typing import Dict, Mapping, Optional, Sequence, Tuple, Type, Union +from typing import ( + Dict, + Mapping, + MutableMapping, + MutableSequence, + Optional, + Sequence, + Tuple, + Type, + Union, +) from google.api_core import exceptions as core_exceptions from google.api_core import gapic_v1 @@ -24,7 +34,8 @@ from google.api_core.client_options import ClientOptions from google.auth import credentials as ga_credentials # type: ignore from google.oauth2 import service_account # type: ignore -import pkg_resources + +from google.cloud.dataflow_v1beta3 import gapic_version as package_version try: OptionalRetry = Union[retries.Retry, gapic_v1.method._MethodDefault] @@ -159,9 +170,9 @@ def transport(self) -> FlexTemplatesServiceTransport: def __init__( self, *, - credentials: ga_credentials.Credentials = None, + credentials: Optional[ga_credentials.Credentials] = None, transport: Union[str, FlexTemplatesServiceTransport] = "grpc_asyncio", - client_options: ClientOptions = None, + client_options: Optional[ClientOptions] = None, client_info: gapic_v1.client_info.ClientInfo = DEFAULT_CLIENT_INFO, ) -> None: """Instantiates the flex templates service client. @@ -205,10 +216,10 @@ def __init__( async def launch_flex_template( self, - request: Union[templates.LaunchFlexTemplateRequest, dict] = None, + request: Optional[Union[templates.LaunchFlexTemplateRequest, dict]] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> templates.LaunchFlexTemplateResponse: r"""Launch a job with a FlexTemplate. @@ -239,7 +250,7 @@ async def sample_launch_flex_template(): print(response) Args: - request (Union[google.cloud.dataflow_v1beta3.types.LaunchFlexTemplateRequest, dict]): + request (Optional[Union[google.cloud.dataflow_v1beta3.types.LaunchFlexTemplateRequest, dict]]): The request object. A request to launch a Cloud Dataflow job from a FlexTemplate. retry (google.api_core.retry.Retry): Designation of what errors, if any, @@ -294,14 +305,9 @@ async def __aexit__(self, exc_type, exc, tb): await self.transport.close() -try: - DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( - gapic_version=pkg_resources.get_distribution( - "google-cloud-dataflow-client", - ).version, - ) -except pkg_resources.DistributionNotFound: - DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo() +DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( + gapic_version=package_version.__version__ +) __all__ = ("FlexTemplatesServiceAsyncClient",) diff --git a/google/cloud/dataflow_v1beta3/services/flex_templates_service/client.py b/google/cloud/dataflow_v1beta3/services/flex_templates_service/client.py index a443b47..c816ef2 100644 --- a/google/cloud/dataflow_v1beta3/services/flex_templates_service/client.py +++ b/google/cloud/dataflow_v1beta3/services/flex_templates_service/client.py @@ -16,7 +16,18 @@ from collections import OrderedDict import os import re -from typing import Dict, Mapping, Optional, Sequence, Tuple, Type, Union +from typing import ( + Dict, + Mapping, + MutableMapping, + MutableSequence, + Optional, + Sequence, + Tuple, + Type, + Union, + cast, +) from google.api_core import client_options as client_options_lib from google.api_core import exceptions as core_exceptions @@ -27,7 +38,8 @@ from google.auth.transport import mtls # type: ignore from google.auth.transport.grpc import SslCredentials # type: ignore from google.oauth2 import service_account # type: ignore -import pkg_resources + +from google.cloud.dataflow_v1beta3 import gapic_version as package_version try: OptionalRetry = Union[retries.Retry, gapic_v1.method._MethodDefault] @@ -59,7 +71,7 @@ class FlexTemplatesServiceClientMeta(type): def get_transport_class( cls, - label: str = None, + label: Optional[str] = None, ) -> Type[FlexTemplatesServiceTransport]: """Returns an appropriate transport class. @@ -314,8 +326,8 @@ def __init__( self, *, credentials: Optional[ga_credentials.Credentials] = None, - transport: Union[str, FlexTemplatesServiceTransport, None] = None, - client_options: Optional[client_options_lib.ClientOptions] = None, + transport: Optional[Union[str, FlexTemplatesServiceTransport]] = None, + client_options: Optional[Union[client_options_lib.ClientOptions, dict]] = None, client_info: gapic_v1.client_info.ClientInfo = DEFAULT_CLIENT_INFO, ) -> None: """Instantiates the flex templates service client. @@ -329,10 +341,7 @@ def __init__( transport (Union[str, FlexTemplatesServiceTransport]): The transport to use. If set to None, a transport is chosen automatically. - NOTE: "rest" transport functionality is currently in a - beta state (preview). We welcome your feedback via an - issue in this library's source repository. - client_options (google.api_core.client_options.ClientOptions): Custom options for the + client_options (Optional[Union[google.api_core.client_options.ClientOptions, dict]]): Custom options for the client. It won't take effect if a ``transport`` instance is provided. (1) The ``api_endpoint`` property can be used to override the default endpoint provided by the client. GOOGLE_API_USE_MTLS_ENDPOINT @@ -362,6 +371,7 @@ def __init__( client_options = client_options_lib.from_dict(client_options) if client_options is None: client_options = client_options_lib.ClientOptions() + client_options = cast(client_options_lib.ClientOptions, client_options) api_endpoint, client_cert_source_func = self.get_mtls_endpoint_and_cert_source( client_options @@ -414,10 +424,10 @@ def __init__( def launch_flex_template( self, - request: Union[templates.LaunchFlexTemplateRequest, dict] = None, + request: Optional[Union[templates.LaunchFlexTemplateRequest, dict]] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> templates.LaunchFlexTemplateResponse: r"""Launch a job with a FlexTemplate. @@ -511,14 +521,9 @@ def __exit__(self, type, value, traceback): self.transport.close() -try: - DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( - gapic_version=pkg_resources.get_distribution( - "google-cloud-dataflow-client", - ).version, - ) -except pkg_resources.DistributionNotFound: - DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo() +DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( + gapic_version=package_version.__version__ +) __all__ = ("FlexTemplatesServiceClient",) diff --git a/google/cloud/dataflow_v1beta3/services/flex_templates_service/transports/base.py b/google/cloud/dataflow_v1beta3/services/flex_templates_service/transports/base.py index 577cc0f..15579b0 100644 --- a/google/cloud/dataflow_v1beta3/services/flex_templates_service/transports/base.py +++ b/google/cloud/dataflow_v1beta3/services/flex_templates_service/transports/base.py @@ -23,18 +23,13 @@ import google.auth # type: ignore from google.auth import credentials as ga_credentials # type: ignore from google.oauth2 import service_account # type: ignore -import pkg_resources +from google.cloud.dataflow_v1beta3 import gapic_version as package_version from google.cloud.dataflow_v1beta3.types import templates -try: - DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( - gapic_version=pkg_resources.get_distribution( - "google-cloud-dataflow-client", - ).version, - ) -except pkg_resources.DistributionNotFound: - DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo() +DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( + gapic_version=package_version.__version__ +) class FlexTemplatesServiceTransport(abc.ABC): @@ -53,7 +48,7 @@ def __init__( self, *, host: str = DEFAULT_HOST, - credentials: ga_credentials.Credentials = None, + credentials: Optional[ga_credentials.Credentials] = None, credentials_file: Optional[str] = None, scopes: Optional[Sequence[str]] = None, quota_project_id: Optional[str] = None, diff --git a/google/cloud/dataflow_v1beta3/services/flex_templates_service/transports/grpc.py b/google/cloud/dataflow_v1beta3/services/flex_templates_service/transports/grpc.py index 4e3b008..172ea31 100644 --- a/google/cloud/dataflow_v1beta3/services/flex_templates_service/transports/grpc.py +++ b/google/cloud/dataflow_v1beta3/services/flex_templates_service/transports/grpc.py @@ -47,14 +47,14 @@ def __init__( self, *, host: str = "dataflow.googleapis.com", - credentials: ga_credentials.Credentials = None, - credentials_file: str = None, - scopes: Sequence[str] = None, - channel: grpc.Channel = None, - api_mtls_endpoint: str = None, - client_cert_source: Callable[[], Tuple[bytes, bytes]] = None, - ssl_channel_credentials: grpc.ChannelCredentials = None, - client_cert_source_for_mtls: Callable[[], Tuple[bytes, bytes]] = None, + credentials: Optional[ga_credentials.Credentials] = None, + credentials_file: Optional[str] = None, + scopes: Optional[Sequence[str]] = None, + channel: Optional[grpc.Channel] = None, + api_mtls_endpoint: Optional[str] = None, + client_cert_source: Optional[Callable[[], Tuple[bytes, bytes]]] = None, + ssl_channel_credentials: Optional[grpc.ChannelCredentials] = None, + client_cert_source_for_mtls: Optional[Callable[[], Tuple[bytes, bytes]]] = None, quota_project_id: Optional[str] = None, client_info: gapic_v1.client_info.ClientInfo = DEFAULT_CLIENT_INFO, always_use_jwt_access: Optional[bool] = False, @@ -181,8 +181,8 @@ def __init__( def create_channel( cls, host: str = "dataflow.googleapis.com", - credentials: ga_credentials.Credentials = None, - credentials_file: str = None, + credentials: Optional[ga_credentials.Credentials] = None, + credentials_file: Optional[str] = None, scopes: Optional[Sequence[str]] = None, quota_project_id: Optional[str] = None, **kwargs, diff --git a/google/cloud/dataflow_v1beta3/services/flex_templates_service/transports/grpc_asyncio.py b/google/cloud/dataflow_v1beta3/services/flex_templates_service/transports/grpc_asyncio.py index 719719f..8ed3f3d 100644 --- a/google/cloud/dataflow_v1beta3/services/flex_templates_service/transports/grpc_asyncio.py +++ b/google/cloud/dataflow_v1beta3/services/flex_templates_service/transports/grpc_asyncio.py @@ -49,7 +49,7 @@ class FlexTemplatesServiceGrpcAsyncIOTransport(FlexTemplatesServiceTransport): def create_channel( cls, host: str = "dataflow.googleapis.com", - credentials: ga_credentials.Credentials = None, + credentials: Optional[ga_credentials.Credentials] = None, credentials_file: Optional[str] = None, scopes: Optional[Sequence[str]] = None, quota_project_id: Optional[str] = None, @@ -92,15 +92,15 @@ def __init__( self, *, host: str = "dataflow.googleapis.com", - credentials: ga_credentials.Credentials = None, + credentials: Optional[ga_credentials.Credentials] = None, credentials_file: Optional[str] = None, scopes: Optional[Sequence[str]] = None, - channel: aio.Channel = None, - api_mtls_endpoint: str = None, - client_cert_source: Callable[[], Tuple[bytes, bytes]] = None, - ssl_channel_credentials: grpc.ChannelCredentials = None, - client_cert_source_for_mtls: Callable[[], Tuple[bytes, bytes]] = None, - quota_project_id=None, + channel: Optional[aio.Channel] = None, + api_mtls_endpoint: Optional[str] = None, + client_cert_source: Optional[Callable[[], Tuple[bytes, bytes]]] = None, + ssl_channel_credentials: Optional[grpc.ChannelCredentials] = None, + client_cert_source_for_mtls: Optional[Callable[[], Tuple[bytes, bytes]]] = None, + quota_project_id: Optional[str] = None, client_info: gapic_v1.client_info.ClientInfo = DEFAULT_CLIENT_INFO, always_use_jwt_access: Optional[bool] = False, api_audience: Optional[str] = None, diff --git a/google/cloud/dataflow_v1beta3/services/flex_templates_service/transports/rest.py b/google/cloud/dataflow_v1beta3/services/flex_templates_service/transports/rest.py index 1986e26..9d55045 100644 --- a/google/cloud/dataflow_v1beta3/services/flex_templates_service/transports/rest.py +++ b/google/cloud/dataflow_v1beta3/services/flex_templates_service/transports/rest.py @@ -63,12 +63,13 @@ class FlexTemplatesServiceRestInterceptor: .. code-block:: python class MyCustomFlexTemplatesServiceInterceptor(FlexTemplatesServiceRestInterceptor): - def pre_launch_flex_template(request, metadata): + def pre_launch_flex_template(self, request, metadata): logging.log(f"Received request: {request}") return request, metadata - def post_launch_flex_template(response): + def post_launch_flex_template(self, response): logging.log(f"Received response: {response}") + return response transport = FlexTemplatesServiceRestTransport(interceptor=MyCustomFlexTemplatesServiceInterceptor()) client = FlexTemplatesServiceClient(transport=transport) @@ -119,19 +120,16 @@ class FlexTemplatesServiceRestTransport(FlexTemplatesServiceTransport): It sends JSON representations of protocol buffers over HTTP/1.1 - NOTE: This REST transport functionality is currently in a beta - state (preview). We welcome your feedback via an issue in this - library's source repository. Thank you! """ def __init__( self, *, host: str = "dataflow.googleapis.com", - credentials: ga_credentials.Credentials = None, - credentials_file: str = None, - scopes: Sequence[str] = None, - client_cert_source_for_mtls: Callable[[], Tuple[bytes, bytes]] = None, + credentials: Optional[ga_credentials.Credentials] = None, + credentials_file: Optional[str] = None, + scopes: Optional[Sequence[str]] = None, + client_cert_source_for_mtls: Optional[Callable[[], Tuple[bytes, bytes]]] = None, quota_project_id: Optional[str] = None, client_info: gapic_v1.client_info.ClientInfo = DEFAULT_CLIENT_INFO, always_use_jwt_access: Optional[bool] = False, @@ -141,39 +139,35 @@ def __init__( ) -> None: """Instantiate the transport. - NOTE: This REST transport functionality is currently in a beta - state (preview). We welcome your feedback via a GitHub issue in - this library's repository. Thank you! - - Args: - host (Optional[str]): - The hostname to connect to. - credentials (Optional[google.auth.credentials.Credentials]): The - authorization credentials to attach to requests. These - credentials identify the application to the service; if none - are specified, the client will attempt to ascertain the - credentials from the environment. - - credentials_file (Optional[str]): A file with credentials that can - be loaded with :func:`google.auth.load_credentials_from_file`. - This argument is ignored if ``channel`` is provided. - scopes (Optional(Sequence[str])): A list of scopes. This argument is - ignored if ``channel`` is provided. - client_cert_source_for_mtls (Callable[[], Tuple[bytes, bytes]]): Client - certificate to configure mutual TLS HTTP channel. It is ignored - if ``channel`` is provided. - quota_project_id (Optional[str]): An optional project to use for billing - and quota. - client_info (google.api_core.gapic_v1.client_info.ClientInfo): - The client info used to send a user-agent string along with - API requests. If ``None``, then default info will be used. - Generally, you only need to set this if you are developing - your own client library. - always_use_jwt_access (Optional[bool]): Whether self signed JWT should - be used for service account credentials. - url_scheme: the protocol scheme for the API endpoint. Normally - "https", but for testing or local servers, - "http" can be specified. + Args: + host (Optional[str]): + The hostname to connect to. + credentials (Optional[google.auth.credentials.Credentials]): The + authorization credentials to attach to requests. These + credentials identify the application to the service; if none + are specified, the client will attempt to ascertain the + credentials from the environment. + + credentials_file (Optional[str]): A file with credentials that can + be loaded with :func:`google.auth.load_credentials_from_file`. + This argument is ignored if ``channel`` is provided. + scopes (Optional(Sequence[str])): A list of scopes. This argument is + ignored if ``channel`` is provided. + client_cert_source_for_mtls (Callable[[], Tuple[bytes, bytes]]): Client + certificate to configure mutual TLS HTTP channel. It is ignored + if ``channel`` is provided. + quota_project_id (Optional[str]): An optional project to use for billing + and quota. + client_info (google.api_core.gapic_v1.client_info.ClientInfo): + The client info used to send a user-agent string along with + API requests. If ``None``, then default info will be used. + Generally, you only need to set this if you are developing + your own client library. + always_use_jwt_access (Optional[bool]): Whether self signed JWT should + be used for service account credentials. + url_scheme: the protocol scheme for the API endpoint. Normally + "https", but for testing or local servers, + "http" can be specified. """ # Run the base constructor # TODO(yon-mg): resolve other ctor params i.e. scopes, quota, etc. @@ -213,7 +207,7 @@ def __call__( request: templates.LaunchFlexTemplateRequest, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> templates.LaunchFlexTemplateResponse: r"""Call the launch flex template method over HTTP. @@ -254,7 +248,7 @@ def __call__( body = json_format.MessageToJson( transcoded_request["body"], including_default_value_fields=False, - use_integers_for_enums=False, + use_integers_for_enums=True, ) uri = transcoded_request["uri"] method = transcoded_request["method"] @@ -264,10 +258,12 @@ def __call__( json_format.MessageToJson( transcoded_request["query_params"], including_default_value_fields=False, - use_integers_for_enums=False, + use_integers_for_enums=True, ) ) + query_params["$alt"] = "json;enum-encoding=int" + # Send the request headers = dict(metadata) headers["Content-Type"] = "application/json" diff --git a/google/cloud/dataflow_v1beta3/services/jobs_v1_beta3/async_client.py b/google/cloud/dataflow_v1beta3/services/jobs_v1_beta3/async_client.py index 19973e7..abf20fb 100644 --- a/google/cloud/dataflow_v1beta3/services/jobs_v1_beta3/async_client.py +++ b/google/cloud/dataflow_v1beta3/services/jobs_v1_beta3/async_client.py @@ -16,7 +16,17 @@ from collections import OrderedDict import functools import re -from typing import Dict, Mapping, Optional, Sequence, Tuple, Type, Union +from typing import ( + Dict, + Mapping, + MutableMapping, + MutableSequence, + Optional, + Sequence, + Tuple, + Type, + Union, +) from google.api_core import exceptions as core_exceptions from google.api_core import gapic_v1 @@ -24,7 +34,8 @@ from google.api_core.client_options import ClientOptions from google.auth import credentials as ga_credentials # type: ignore from google.oauth2 import service_account # type: ignore -import pkg_resources + +from google.cloud.dataflow_v1beta3 import gapic_version as package_version try: OptionalRetry = Union[retries.Retry, gapic_v1.method._MethodDefault] @@ -159,9 +170,9 @@ def transport(self) -> JobsV1Beta3Transport: def __init__( self, *, - credentials: ga_credentials.Credentials = None, + credentials: Optional[ga_credentials.Credentials] = None, transport: Union[str, JobsV1Beta3Transport] = "grpc_asyncio", - client_options: ClientOptions = None, + client_options: Optional[ClientOptions] = None, client_info: gapic_v1.client_info.ClientInfo = DEFAULT_CLIENT_INFO, ) -> None: """Instantiates the jobs v1 beta3 client. @@ -205,10 +216,10 @@ def __init__( async def create_job( self, - request: Union[jobs.CreateJobRequest, dict] = None, + request: Optional[Union[jobs.CreateJobRequest, dict]] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> jobs.Job: r"""Creates a Cloud Dataflow job. @@ -245,7 +256,7 @@ async def sample_create_job(): print(response) Args: - request (Union[google.cloud.dataflow_v1beta3.types.CreateJobRequest, dict]): + request (Optional[Union[google.cloud.dataflow_v1beta3.types.CreateJobRequest, dict]]): The request object. Request to create a Cloud Dataflow job. retry (google.api_core.retry.Retry): Designation of what errors, if any, @@ -295,10 +306,10 @@ async def sample_create_job(): async def get_job( self, - request: Union[jobs.GetJobRequest, dict] = None, + request: Optional[Union[jobs.GetJobRequest, dict]] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> jobs.Job: r"""Gets the state of the specified Cloud Dataflow job. @@ -335,7 +346,7 @@ async def sample_get_job(): print(response) Args: - request (Union[google.cloud.dataflow_v1beta3.types.GetJobRequest, dict]): + request (Optional[Union[google.cloud.dataflow_v1beta3.types.GetJobRequest, dict]]): The request object. Request to get the state of a Cloud Dataflow job. retry (google.api_core.retry.Retry): Designation of what errors, if any, @@ -386,10 +397,10 @@ async def sample_get_job(): async def update_job( self, - request: Union[jobs.UpdateJobRequest, dict] = None, + request: Optional[Union[jobs.UpdateJobRequest, dict]] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> jobs.Job: r"""Updates the state of an existing Cloud Dataflow job. @@ -427,7 +438,7 @@ async def sample_update_job(): print(response) Args: - request (Union[google.cloud.dataflow_v1beta3.types.UpdateJobRequest, dict]): + request (Optional[Union[google.cloud.dataflow_v1beta3.types.UpdateJobRequest, dict]]): The request object. Request to update a Cloud Dataflow job. retry (google.api_core.retry.Retry): Designation of what errors, if any, @@ -478,10 +489,10 @@ async def sample_update_job(): async def list_jobs( self, - request: Union[jobs.ListJobsRequest, dict] = None, + request: Optional[Union[jobs.ListJobsRequest, dict]] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.ListJobsAsyncPager: r"""List the jobs of a project. @@ -521,7 +532,7 @@ async def sample_list_jobs(): print(response) Args: - request (Union[google.cloud.dataflow_v1beta3.types.ListJobsRequest, dict]): + request (Optional[Union[google.cloud.dataflow_v1beta3.types.ListJobsRequest, dict]]): The request object. Request to list Cloud Dataflow jobs. retry (google.api_core.retry.Retry): Designation of what errors, if any, should be retried. @@ -589,10 +600,10 @@ async def sample_list_jobs(): async def aggregated_list_jobs( self, - request: Union[jobs.ListJobsRequest, dict] = None, + request: Optional[Union[jobs.ListJobsRequest, dict]] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.AggregatedListJobsAsyncPager: r"""List the jobs of a project across all regions. @@ -624,7 +635,7 @@ async def sample_aggregated_list_jobs(): print(response) Args: - request (Union[google.cloud.dataflow_v1beta3.types.ListJobsRequest, dict]): + request (Optional[Union[google.cloud.dataflow_v1beta3.types.ListJobsRequest, dict]]): The request object. Request to list Cloud Dataflow jobs. retry (google.api_core.retry.Retry): Designation of what errors, if any, should be retried. @@ -689,10 +700,10 @@ async def sample_aggregated_list_jobs(): async def check_active_jobs( self, - request: Union[jobs.CheckActiveJobsRequest, dict] = None, + request: Optional[Union[jobs.CheckActiveJobsRequest, dict]] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> jobs.CheckActiveJobsResponse: r"""Check for existence of active jobs in the given @@ -724,7 +735,7 @@ async def sample_check_active_jobs(): print(response) Args: - request (Union[google.cloud.dataflow_v1beta3.types.CheckActiveJobsRequest, dict]): + request (Optional[Union[google.cloud.dataflow_v1beta3.types.CheckActiveJobsRequest, dict]]): The request object. Request to check is active jobs exists for a project retry (google.api_core.retry.Retry): Designation of what errors, if any, @@ -761,10 +772,10 @@ async def sample_check_active_jobs(): async def snapshot_job( self, - request: Union[jobs.SnapshotJobRequest, dict] = None, + request: Optional[Union[jobs.SnapshotJobRequest, dict]] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> snapshots.Snapshot: r"""Snapshot the state of a streaming job. @@ -795,7 +806,7 @@ async def sample_snapshot_job(): print(response) Args: - request (Union[google.cloud.dataflow_v1beta3.types.SnapshotJobRequest, dict]): + request (Optional[Union[google.cloud.dataflow_v1beta3.types.SnapshotJobRequest, dict]]): The request object. Request to create a snapshot of a job. retry (google.api_core.retry.Retry): Designation of what errors, if any, @@ -849,14 +860,9 @@ async def __aexit__(self, exc_type, exc, tb): await self.transport.close() -try: - DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( - gapic_version=pkg_resources.get_distribution( - "google-cloud-dataflow-client", - ).version, - ) -except pkg_resources.DistributionNotFound: - DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo() +DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( + gapic_version=package_version.__version__ +) __all__ = ("JobsV1Beta3AsyncClient",) diff --git a/google/cloud/dataflow_v1beta3/services/jobs_v1_beta3/client.py b/google/cloud/dataflow_v1beta3/services/jobs_v1_beta3/client.py index d216313..b278651 100644 --- a/google/cloud/dataflow_v1beta3/services/jobs_v1_beta3/client.py +++ b/google/cloud/dataflow_v1beta3/services/jobs_v1_beta3/client.py @@ -16,7 +16,18 @@ from collections import OrderedDict import os import re -from typing import Dict, Mapping, Optional, Sequence, Tuple, Type, Union +from typing import ( + Dict, + Mapping, + MutableMapping, + MutableSequence, + Optional, + Sequence, + Tuple, + Type, + Union, + cast, +) from google.api_core import client_options as client_options_lib from google.api_core import exceptions as core_exceptions @@ -27,7 +38,8 @@ from google.auth.transport import mtls # type: ignore from google.auth.transport.grpc import SslCredentials # type: ignore from google.oauth2 import service_account # type: ignore -import pkg_resources + +from google.cloud.dataflow_v1beta3 import gapic_version as package_version try: OptionalRetry = Union[retries.Retry, gapic_v1.method._MethodDefault] @@ -61,7 +73,7 @@ class JobsV1Beta3ClientMeta(type): def get_transport_class( cls, - label: str = None, + label: Optional[str] = None, ) -> Type[JobsV1Beta3Transport]: """Returns an appropriate transport class. @@ -317,8 +329,8 @@ def __init__( self, *, credentials: Optional[ga_credentials.Credentials] = None, - transport: Union[str, JobsV1Beta3Transport, None] = None, - client_options: Optional[client_options_lib.ClientOptions] = None, + transport: Optional[Union[str, JobsV1Beta3Transport]] = None, + client_options: Optional[Union[client_options_lib.ClientOptions, dict]] = None, client_info: gapic_v1.client_info.ClientInfo = DEFAULT_CLIENT_INFO, ) -> None: """Instantiates the jobs v1 beta3 client. @@ -332,10 +344,7 @@ def __init__( transport (Union[str, JobsV1Beta3Transport]): The transport to use. If set to None, a transport is chosen automatically. - NOTE: "rest" transport functionality is currently in a - beta state (preview). We welcome your feedback via an - issue in this library's source repository. - client_options (google.api_core.client_options.ClientOptions): Custom options for the + client_options (Optional[Union[google.api_core.client_options.ClientOptions, dict]]): Custom options for the client. It won't take effect if a ``transport`` instance is provided. (1) The ``api_endpoint`` property can be used to override the default endpoint provided by the client. GOOGLE_API_USE_MTLS_ENDPOINT @@ -365,6 +374,7 @@ def __init__( client_options = client_options_lib.from_dict(client_options) if client_options is None: client_options = client_options_lib.ClientOptions() + client_options = cast(client_options_lib.ClientOptions, client_options) api_endpoint, client_cert_source_func = self.get_mtls_endpoint_and_cert_source( client_options @@ -417,10 +427,10 @@ def __init__( def create_job( self, - request: Union[jobs.CreateJobRequest, dict] = None, + request: Optional[Union[jobs.CreateJobRequest, dict]] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> jobs.Job: r"""Creates a Cloud Dataflow job. @@ -508,10 +518,10 @@ def sample_create_job(): def get_job( self, - request: Union[jobs.GetJobRequest, dict] = None, + request: Optional[Union[jobs.GetJobRequest, dict]] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> jobs.Job: r"""Gets the state of the specified Cloud Dataflow job. @@ -600,10 +610,10 @@ def sample_get_job(): def update_job( self, - request: Union[jobs.UpdateJobRequest, dict] = None, + request: Optional[Union[jobs.UpdateJobRequest, dict]] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> jobs.Job: r"""Updates the state of an existing Cloud Dataflow job. @@ -693,10 +703,10 @@ def sample_update_job(): def list_jobs( self, - request: Union[jobs.ListJobsRequest, dict] = None, + request: Optional[Union[jobs.ListJobsRequest, dict]] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.ListJobsPager: r"""List the jobs of a project. @@ -805,10 +815,10 @@ def sample_list_jobs(): def aggregated_list_jobs( self, - request: Union[jobs.ListJobsRequest, dict] = None, + request: Optional[Union[jobs.ListJobsRequest, dict]] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.AggregatedListJobsPager: r"""List the jobs of a project across all regions. @@ -906,10 +916,10 @@ def sample_aggregated_list_jobs(): def check_active_jobs( self, - request: Union[jobs.CheckActiveJobsRequest, dict] = None, + request: Optional[Union[jobs.CheckActiveJobsRequest, dict]] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> jobs.CheckActiveJobsResponse: r"""Check for existence of active jobs in the given @@ -979,10 +989,10 @@ def sample_check_active_jobs(): def snapshot_job( self, - request: Union[jobs.SnapshotJobRequest, dict] = None, + request: Optional[Union[jobs.SnapshotJobRequest, dict]] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> snapshots.Snapshot: r"""Snapshot the state of a streaming job. @@ -1075,14 +1085,9 @@ def __exit__(self, type, value, traceback): self.transport.close() -try: - DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( - gapic_version=pkg_resources.get_distribution( - "google-cloud-dataflow-client", - ).version, - ) -except pkg_resources.DistributionNotFound: - DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo() +DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( + gapic_version=package_version.__version__ +) __all__ = ("JobsV1Beta3Client",) diff --git a/google/cloud/dataflow_v1beta3/services/jobs_v1_beta3/transports/base.py b/google/cloud/dataflow_v1beta3/services/jobs_v1_beta3/transports/base.py index 453fc1f..9e3e0c8 100644 --- a/google/cloud/dataflow_v1beta3/services/jobs_v1_beta3/transports/base.py +++ b/google/cloud/dataflow_v1beta3/services/jobs_v1_beta3/transports/base.py @@ -23,18 +23,13 @@ import google.auth # type: ignore from google.auth import credentials as ga_credentials # type: ignore from google.oauth2 import service_account # type: ignore -import pkg_resources +from google.cloud.dataflow_v1beta3 import gapic_version as package_version from google.cloud.dataflow_v1beta3.types import jobs, snapshots -try: - DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( - gapic_version=pkg_resources.get_distribution( - "google-cloud-dataflow-client", - ).version, - ) -except pkg_resources.DistributionNotFound: - DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo() +DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( + gapic_version=package_version.__version__ +) class JobsV1Beta3Transport(abc.ABC): @@ -53,7 +48,7 @@ def __init__( self, *, host: str = DEFAULT_HOST, - credentials: ga_credentials.Credentials = None, + credentials: Optional[ga_credentials.Credentials] = None, credentials_file: Optional[str] = None, scopes: Optional[Sequence[str]] = None, quota_project_id: Optional[str] = None, diff --git a/google/cloud/dataflow_v1beta3/services/jobs_v1_beta3/transports/grpc.py b/google/cloud/dataflow_v1beta3/services/jobs_v1_beta3/transports/grpc.py index 27131eb..cbe441f 100644 --- a/google/cloud/dataflow_v1beta3/services/jobs_v1_beta3/transports/grpc.py +++ b/google/cloud/dataflow_v1beta3/services/jobs_v1_beta3/transports/grpc.py @@ -48,14 +48,14 @@ def __init__( self, *, host: str = "dataflow.googleapis.com", - credentials: ga_credentials.Credentials = None, - credentials_file: str = None, - scopes: Sequence[str] = None, - channel: grpc.Channel = None, - api_mtls_endpoint: str = None, - client_cert_source: Callable[[], Tuple[bytes, bytes]] = None, - ssl_channel_credentials: grpc.ChannelCredentials = None, - client_cert_source_for_mtls: Callable[[], Tuple[bytes, bytes]] = None, + credentials: Optional[ga_credentials.Credentials] = None, + credentials_file: Optional[str] = None, + scopes: Optional[Sequence[str]] = None, + channel: Optional[grpc.Channel] = None, + api_mtls_endpoint: Optional[str] = None, + client_cert_source: Optional[Callable[[], Tuple[bytes, bytes]]] = None, + ssl_channel_credentials: Optional[grpc.ChannelCredentials] = None, + client_cert_source_for_mtls: Optional[Callable[[], Tuple[bytes, bytes]]] = None, quota_project_id: Optional[str] = None, client_info: gapic_v1.client_info.ClientInfo = DEFAULT_CLIENT_INFO, always_use_jwt_access: Optional[bool] = False, @@ -182,8 +182,8 @@ def __init__( def create_channel( cls, host: str = "dataflow.googleapis.com", - credentials: ga_credentials.Credentials = None, - credentials_file: str = None, + credentials: Optional[ga_credentials.Credentials] = None, + credentials_file: Optional[str] = None, scopes: Optional[Sequence[str]] = None, quota_project_id: Optional[str] = None, **kwargs, diff --git a/google/cloud/dataflow_v1beta3/services/jobs_v1_beta3/transports/grpc_asyncio.py b/google/cloud/dataflow_v1beta3/services/jobs_v1_beta3/transports/grpc_asyncio.py index e7856e2..094b22b 100644 --- a/google/cloud/dataflow_v1beta3/services/jobs_v1_beta3/transports/grpc_asyncio.py +++ b/google/cloud/dataflow_v1beta3/services/jobs_v1_beta3/transports/grpc_asyncio.py @@ -50,7 +50,7 @@ class JobsV1Beta3GrpcAsyncIOTransport(JobsV1Beta3Transport): def create_channel( cls, host: str = "dataflow.googleapis.com", - credentials: ga_credentials.Credentials = None, + credentials: Optional[ga_credentials.Credentials] = None, credentials_file: Optional[str] = None, scopes: Optional[Sequence[str]] = None, quota_project_id: Optional[str] = None, @@ -93,15 +93,15 @@ def __init__( self, *, host: str = "dataflow.googleapis.com", - credentials: ga_credentials.Credentials = None, + credentials: Optional[ga_credentials.Credentials] = None, credentials_file: Optional[str] = None, scopes: Optional[Sequence[str]] = None, - channel: aio.Channel = None, - api_mtls_endpoint: str = None, - client_cert_source: Callable[[], Tuple[bytes, bytes]] = None, - ssl_channel_credentials: grpc.ChannelCredentials = None, - client_cert_source_for_mtls: Callable[[], Tuple[bytes, bytes]] = None, - quota_project_id=None, + channel: Optional[aio.Channel] = None, + api_mtls_endpoint: Optional[str] = None, + client_cert_source: Optional[Callable[[], Tuple[bytes, bytes]]] = None, + ssl_channel_credentials: Optional[grpc.ChannelCredentials] = None, + client_cert_source_for_mtls: Optional[Callable[[], Tuple[bytes, bytes]]] = None, + quota_project_id: Optional[str] = None, client_info: gapic_v1.client_info.ClientInfo = DEFAULT_CLIENT_INFO, always_use_jwt_access: Optional[bool] = False, api_audience: Optional[str] = None, diff --git a/google/cloud/dataflow_v1beta3/services/jobs_v1_beta3/transports/rest.py b/google/cloud/dataflow_v1beta3/services/jobs_v1_beta3/transports/rest.py index 1e8598a..18eaf06 100644 --- a/google/cloud/dataflow_v1beta3/services/jobs_v1_beta3/transports/rest.py +++ b/google/cloud/dataflow_v1beta3/services/jobs_v1_beta3/transports/rest.py @@ -63,54 +63,61 @@ class JobsV1Beta3RestInterceptor: .. code-block:: python class MyCustomJobsV1Beta3Interceptor(JobsV1Beta3RestInterceptor): - def pre_aggregated_list_jobs(request, metadata): + def pre_aggregated_list_jobs(self, request, metadata): logging.log(f"Received request: {request}") return request, metadata - def post_aggregated_list_jobs(response): + def post_aggregated_list_jobs(self, response): logging.log(f"Received response: {response}") + return response - def pre_check_active_jobs(request, metadata): + def pre_check_active_jobs(self, request, metadata): logging.log(f"Received request: {request}") return request, metadata - def post_check_active_jobs(response): + def post_check_active_jobs(self, response): logging.log(f"Received response: {response}") + return response - def pre_create_job(request, metadata): + def pre_create_job(self, request, metadata): logging.log(f"Received request: {request}") return request, metadata - def post_create_job(response): + def post_create_job(self, response): logging.log(f"Received response: {response}") + return response - def pre_get_job(request, metadata): + def pre_get_job(self, request, metadata): logging.log(f"Received request: {request}") return request, metadata - def post_get_job(response): + def post_get_job(self, response): logging.log(f"Received response: {response}") + return response - def pre_list_jobs(request, metadata): + def pre_list_jobs(self, request, metadata): logging.log(f"Received request: {request}") return request, metadata - def post_list_jobs(response): + def post_list_jobs(self, response): logging.log(f"Received response: {response}") + return response - def pre_snapshot_job(request, metadata): + def pre_snapshot_job(self, request, metadata): logging.log(f"Received request: {request}") return request, metadata - def post_snapshot_job(response): + def post_snapshot_job(self, response): logging.log(f"Received response: {response}") + return response - def pre_update_job(request, metadata): + def pre_update_job(self, request, metadata): logging.log(f"Received request: {request}") return request, metadata - def post_update_job(response): + def post_update_job(self, response): logging.log(f"Received response: {response}") + return response transport = JobsV1Beta3RestTransport(interceptor=MyCustomJobsV1Beta3Interceptor()) client = JobsV1Beta3Client(transport=transport) @@ -255,19 +262,16 @@ class JobsV1Beta3RestTransport(JobsV1Beta3Transport): It sends JSON representations of protocol buffers over HTTP/1.1 - NOTE: This REST transport functionality is currently in a beta - state (preview). We welcome your feedback via an issue in this - library's source repository. Thank you! """ def __init__( self, *, host: str = "dataflow.googleapis.com", - credentials: ga_credentials.Credentials = None, - credentials_file: str = None, - scopes: Sequence[str] = None, - client_cert_source_for_mtls: Callable[[], Tuple[bytes, bytes]] = None, + credentials: Optional[ga_credentials.Credentials] = None, + credentials_file: Optional[str] = None, + scopes: Optional[Sequence[str]] = None, + client_cert_source_for_mtls: Optional[Callable[[], Tuple[bytes, bytes]]] = None, quota_project_id: Optional[str] = None, client_info: gapic_v1.client_info.ClientInfo = DEFAULT_CLIENT_INFO, always_use_jwt_access: Optional[bool] = False, @@ -277,39 +281,35 @@ def __init__( ) -> None: """Instantiate the transport. - NOTE: This REST transport functionality is currently in a beta - state (preview). We welcome your feedback via a GitHub issue in - this library's repository. Thank you! - - Args: - host (Optional[str]): - The hostname to connect to. - credentials (Optional[google.auth.credentials.Credentials]): The - authorization credentials to attach to requests. These - credentials identify the application to the service; if none - are specified, the client will attempt to ascertain the - credentials from the environment. - - credentials_file (Optional[str]): A file with credentials that can - be loaded with :func:`google.auth.load_credentials_from_file`. - This argument is ignored if ``channel`` is provided. - scopes (Optional(Sequence[str])): A list of scopes. This argument is - ignored if ``channel`` is provided. - client_cert_source_for_mtls (Callable[[], Tuple[bytes, bytes]]): Client - certificate to configure mutual TLS HTTP channel. It is ignored - if ``channel`` is provided. - quota_project_id (Optional[str]): An optional project to use for billing - and quota. - client_info (google.api_core.gapic_v1.client_info.ClientInfo): - The client info used to send a user-agent string along with - API requests. If ``None``, then default info will be used. - Generally, you only need to set this if you are developing - your own client library. - always_use_jwt_access (Optional[bool]): Whether self signed JWT should - be used for service account credentials. - url_scheme: the protocol scheme for the API endpoint. Normally - "https", but for testing or local servers, - "http" can be specified. + Args: + host (Optional[str]): + The hostname to connect to. + credentials (Optional[google.auth.credentials.Credentials]): The + authorization credentials to attach to requests. These + credentials identify the application to the service; if none + are specified, the client will attempt to ascertain the + credentials from the environment. + + credentials_file (Optional[str]): A file with credentials that can + be loaded with :func:`google.auth.load_credentials_from_file`. + This argument is ignored if ``channel`` is provided. + scopes (Optional(Sequence[str])): A list of scopes. This argument is + ignored if ``channel`` is provided. + client_cert_source_for_mtls (Callable[[], Tuple[bytes, bytes]]): Client + certificate to configure mutual TLS HTTP channel. It is ignored + if ``channel`` is provided. + quota_project_id (Optional[str]): An optional project to use for billing + and quota. + client_info (google.api_core.gapic_v1.client_info.ClientInfo): + The client info used to send a user-agent string along with + API requests. If ``None``, then default info will be used. + Generally, you only need to set this if you are developing + your own client library. + always_use_jwt_access (Optional[bool]): Whether self signed JWT should + be used for service account credentials. + url_scheme: the protocol scheme for the API endpoint. Normally + "https", but for testing or local servers, + "http" can be specified. """ # Run the base constructor # TODO(yon-mg): resolve other ctor params i.e. scopes, quota, etc. @@ -349,7 +349,7 @@ def __call__( request: jobs.ListJobsRequest, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> jobs.ListJobsResponse: r"""Call the aggregated list jobs method over HTTP. @@ -396,10 +396,12 @@ def __call__( json_format.MessageToJson( transcoded_request["query_params"], including_default_value_fields=False, - use_integers_for_enums=False, + use_integers_for_enums=True, ) ) + query_params["$alt"] = "json;enum-encoding=int" + # Send the request headers = dict(metadata) headers["Content-Type"] = "application/json" @@ -432,7 +434,7 @@ def __call__( request: jobs.CheckActiveJobsRequest, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> jobs.CheckActiveJobsResponse: raise RuntimeError( @@ -448,7 +450,7 @@ def __call__( request: jobs.CreateJobRequest, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> jobs.Job: r"""Call the create job method over HTTP. @@ -492,7 +494,7 @@ def __call__( body = json_format.MessageToJson( transcoded_request["body"], including_default_value_fields=False, - use_integers_for_enums=False, + use_integers_for_enums=True, ) uri = transcoded_request["uri"] method = transcoded_request["method"] @@ -502,10 +504,12 @@ def __call__( json_format.MessageToJson( transcoded_request["query_params"], including_default_value_fields=False, - use_integers_for_enums=False, + use_integers_for_enums=True, ) ) + query_params["$alt"] = "json;enum-encoding=int" + # Send the request headers = dict(metadata) headers["Content-Type"] = "application/json" @@ -539,7 +543,7 @@ def __call__( request: jobs.GetJobRequest, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> jobs.Job: r"""Call the get job method over HTTP. @@ -584,10 +588,12 @@ def __call__( json_format.MessageToJson( transcoded_request["query_params"], including_default_value_fields=False, - use_integers_for_enums=False, + use_integers_for_enums=True, ) ) + query_params["$alt"] = "json;enum-encoding=int" + # Send the request headers = dict(metadata) headers["Content-Type"] = "application/json" @@ -620,7 +626,7 @@ def __call__( request: jobs.ListJobsRequest, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> jobs.ListJobsResponse: r"""Call the list jobs method over HTTP. @@ -669,10 +675,12 @@ def __call__( json_format.MessageToJson( transcoded_request["query_params"], including_default_value_fields=False, - use_integers_for_enums=False, + use_integers_for_enums=True, ) ) + query_params["$alt"] = "json;enum-encoding=int" + # Send the request headers = dict(metadata) headers["Content-Type"] = "application/json" @@ -705,7 +713,7 @@ def __call__( request: jobs.SnapshotJobRequest, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> snapshots.Snapshot: r"""Call the snapshot job method over HTTP. @@ -747,7 +755,7 @@ def __call__( body = json_format.MessageToJson( transcoded_request["body"], including_default_value_fields=False, - use_integers_for_enums=False, + use_integers_for_enums=True, ) uri = transcoded_request["uri"] method = transcoded_request["method"] @@ -757,10 +765,12 @@ def __call__( json_format.MessageToJson( transcoded_request["query_params"], including_default_value_fields=False, - use_integers_for_enums=False, + use_integers_for_enums=True, ) ) + query_params["$alt"] = "json;enum-encoding=int" + # Send the request headers = dict(metadata) headers["Content-Type"] = "application/json" @@ -794,7 +804,7 @@ def __call__( request: jobs.UpdateJobRequest, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> jobs.Job: r"""Call the update job method over HTTP. @@ -838,7 +848,7 @@ def __call__( body = json_format.MessageToJson( transcoded_request["body"], including_default_value_fields=False, - use_integers_for_enums=False, + use_integers_for_enums=True, ) uri = transcoded_request["uri"] method = transcoded_request["method"] @@ -848,10 +858,12 @@ def __call__( json_format.MessageToJson( transcoded_request["query_params"], including_default_value_fields=False, - use_integers_for_enums=False, + use_integers_for_enums=True, ) ) + query_params["$alt"] = "json;enum-encoding=int" + # Send the request headers = dict(metadata) headers["Content-Type"] = "application/json" diff --git a/google/cloud/dataflow_v1beta3/services/messages_v1_beta3/async_client.py b/google/cloud/dataflow_v1beta3/services/messages_v1_beta3/async_client.py index 9709399..0157a0f 100644 --- a/google/cloud/dataflow_v1beta3/services/messages_v1_beta3/async_client.py +++ b/google/cloud/dataflow_v1beta3/services/messages_v1_beta3/async_client.py @@ -16,7 +16,17 @@ from collections import OrderedDict import functools import re -from typing import Dict, Mapping, Optional, Sequence, Tuple, Type, Union +from typing import ( + Dict, + Mapping, + MutableMapping, + MutableSequence, + Optional, + Sequence, + Tuple, + Type, + Union, +) from google.api_core import exceptions as core_exceptions from google.api_core import gapic_v1 @@ -24,7 +34,8 @@ from google.api_core.client_options import ClientOptions from google.auth import credentials as ga_credentials # type: ignore from google.oauth2 import service_account # type: ignore -import pkg_resources + +from google.cloud.dataflow_v1beta3 import gapic_version as package_version try: OptionalRetry = Union[retries.Retry, gapic_v1.method._MethodDefault] @@ -159,9 +170,9 @@ def transport(self) -> MessagesV1Beta3Transport: def __init__( self, *, - credentials: ga_credentials.Credentials = None, + credentials: Optional[ga_credentials.Credentials] = None, transport: Union[str, MessagesV1Beta3Transport] = "grpc_asyncio", - client_options: ClientOptions = None, + client_options: Optional[ClientOptions] = None, client_info: gapic_v1.client_info.ClientInfo = DEFAULT_CLIENT_INFO, ) -> None: """Instantiates the messages v1 beta3 client. @@ -205,10 +216,10 @@ def __init__( async def list_job_messages( self, - request: Union[messages.ListJobMessagesRequest, dict] = None, + request: Optional[Union[messages.ListJobMessagesRequest, dict]] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.ListJobMessagesAsyncPager: r"""Request the job status. @@ -248,7 +259,7 @@ async def sample_list_job_messages(): print(response) Args: - request (Union[google.cloud.dataflow_v1beta3.types.ListJobMessagesRequest, dict]): + request (Optional[Union[google.cloud.dataflow_v1beta3.types.ListJobMessagesRequest, dict]]): The request object. Request to list job messages. Up to max_results messages will be returned in the time range specified starting with the oldest messages first. @@ -319,14 +330,9 @@ async def __aexit__(self, exc_type, exc, tb): await self.transport.close() -try: - DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( - gapic_version=pkg_resources.get_distribution( - "google-cloud-dataflow-client", - ).version, - ) -except pkg_resources.DistributionNotFound: - DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo() +DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( + gapic_version=package_version.__version__ +) __all__ = ("MessagesV1Beta3AsyncClient",) diff --git a/google/cloud/dataflow_v1beta3/services/messages_v1_beta3/client.py b/google/cloud/dataflow_v1beta3/services/messages_v1_beta3/client.py index 435a92f..2bbd3f8 100644 --- a/google/cloud/dataflow_v1beta3/services/messages_v1_beta3/client.py +++ b/google/cloud/dataflow_v1beta3/services/messages_v1_beta3/client.py @@ -16,7 +16,18 @@ from collections import OrderedDict import os import re -from typing import Dict, Mapping, Optional, Sequence, Tuple, Type, Union +from typing import ( + Dict, + Mapping, + MutableMapping, + MutableSequence, + Optional, + Sequence, + Tuple, + Type, + Union, + cast, +) from google.api_core import client_options as client_options_lib from google.api_core import exceptions as core_exceptions @@ -27,7 +38,8 @@ from google.auth.transport import mtls # type: ignore from google.auth.transport.grpc import SslCredentials # type: ignore from google.oauth2 import service_account # type: ignore -import pkg_resources + +from google.cloud.dataflow_v1beta3 import gapic_version as package_version try: OptionalRetry = Union[retries.Retry, gapic_v1.method._MethodDefault] @@ -60,7 +72,7 @@ class MessagesV1Beta3ClientMeta(type): def get_transport_class( cls, - label: str = None, + label: Optional[str] = None, ) -> Type[MessagesV1Beta3Transport]: """Returns an appropriate transport class. @@ -315,8 +327,8 @@ def __init__( self, *, credentials: Optional[ga_credentials.Credentials] = None, - transport: Union[str, MessagesV1Beta3Transport, None] = None, - client_options: Optional[client_options_lib.ClientOptions] = None, + transport: Optional[Union[str, MessagesV1Beta3Transport]] = None, + client_options: Optional[Union[client_options_lib.ClientOptions, dict]] = None, client_info: gapic_v1.client_info.ClientInfo = DEFAULT_CLIENT_INFO, ) -> None: """Instantiates the messages v1 beta3 client. @@ -330,10 +342,7 @@ def __init__( transport (Union[str, MessagesV1Beta3Transport]): The transport to use. If set to None, a transport is chosen automatically. - NOTE: "rest" transport functionality is currently in a - beta state (preview). We welcome your feedback via an - issue in this library's source repository. - client_options (google.api_core.client_options.ClientOptions): Custom options for the + client_options (Optional[Union[google.api_core.client_options.ClientOptions, dict]]): Custom options for the client. It won't take effect if a ``transport`` instance is provided. (1) The ``api_endpoint`` property can be used to override the default endpoint provided by the client. GOOGLE_API_USE_MTLS_ENDPOINT @@ -363,6 +372,7 @@ def __init__( client_options = client_options_lib.from_dict(client_options) if client_options is None: client_options = client_options_lib.ClientOptions() + client_options = cast(client_options_lib.ClientOptions, client_options) api_endpoint, client_cert_source_func = self.get_mtls_endpoint_and_cert_source( client_options @@ -415,10 +425,10 @@ def __init__( def list_job_messages( self, - request: Union[messages.ListJobMessagesRequest, dict] = None, + request: Optional[Union[messages.ListJobMessagesRequest, dict]] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.ListJobMessagesPager: r"""Request the job status. @@ -537,14 +547,9 @@ def __exit__(self, type, value, traceback): self.transport.close() -try: - DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( - gapic_version=pkg_resources.get_distribution( - "google-cloud-dataflow-client", - ).version, - ) -except pkg_resources.DistributionNotFound: - DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo() +DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( + gapic_version=package_version.__version__ +) __all__ = ("MessagesV1Beta3Client",) diff --git a/google/cloud/dataflow_v1beta3/services/messages_v1_beta3/transports/base.py b/google/cloud/dataflow_v1beta3/services/messages_v1_beta3/transports/base.py index 32c54f8..dcc57ac 100644 --- a/google/cloud/dataflow_v1beta3/services/messages_v1_beta3/transports/base.py +++ b/google/cloud/dataflow_v1beta3/services/messages_v1_beta3/transports/base.py @@ -23,18 +23,13 @@ import google.auth # type: ignore from google.auth import credentials as ga_credentials # type: ignore from google.oauth2 import service_account # type: ignore -import pkg_resources +from google.cloud.dataflow_v1beta3 import gapic_version as package_version from google.cloud.dataflow_v1beta3.types import messages -try: - DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( - gapic_version=pkg_resources.get_distribution( - "google-cloud-dataflow-client", - ).version, - ) -except pkg_resources.DistributionNotFound: - DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo() +DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( + gapic_version=package_version.__version__ +) class MessagesV1Beta3Transport(abc.ABC): @@ -53,7 +48,7 @@ def __init__( self, *, host: str = DEFAULT_HOST, - credentials: ga_credentials.Credentials = None, + credentials: Optional[ga_credentials.Credentials] = None, credentials_file: Optional[str] = None, scopes: Optional[Sequence[str]] = None, quota_project_id: Optional[str] = None, diff --git a/google/cloud/dataflow_v1beta3/services/messages_v1_beta3/transports/grpc.py b/google/cloud/dataflow_v1beta3/services/messages_v1_beta3/transports/grpc.py index 1b6be85..c0b98ff 100644 --- a/google/cloud/dataflow_v1beta3/services/messages_v1_beta3/transports/grpc.py +++ b/google/cloud/dataflow_v1beta3/services/messages_v1_beta3/transports/grpc.py @@ -47,14 +47,14 @@ def __init__( self, *, host: str = "dataflow.googleapis.com", - credentials: ga_credentials.Credentials = None, - credentials_file: str = None, - scopes: Sequence[str] = None, - channel: grpc.Channel = None, - api_mtls_endpoint: str = None, - client_cert_source: Callable[[], Tuple[bytes, bytes]] = None, - ssl_channel_credentials: grpc.ChannelCredentials = None, - client_cert_source_for_mtls: Callable[[], Tuple[bytes, bytes]] = None, + credentials: Optional[ga_credentials.Credentials] = None, + credentials_file: Optional[str] = None, + scopes: Optional[Sequence[str]] = None, + channel: Optional[grpc.Channel] = None, + api_mtls_endpoint: Optional[str] = None, + client_cert_source: Optional[Callable[[], Tuple[bytes, bytes]]] = None, + ssl_channel_credentials: Optional[grpc.ChannelCredentials] = None, + client_cert_source_for_mtls: Optional[Callable[[], Tuple[bytes, bytes]]] = None, quota_project_id: Optional[str] = None, client_info: gapic_v1.client_info.ClientInfo = DEFAULT_CLIENT_INFO, always_use_jwt_access: Optional[bool] = False, @@ -181,8 +181,8 @@ def __init__( def create_channel( cls, host: str = "dataflow.googleapis.com", - credentials: ga_credentials.Credentials = None, - credentials_file: str = None, + credentials: Optional[ga_credentials.Credentials] = None, + credentials_file: Optional[str] = None, scopes: Optional[Sequence[str]] = None, quota_project_id: Optional[str] = None, **kwargs, diff --git a/google/cloud/dataflow_v1beta3/services/messages_v1_beta3/transports/grpc_asyncio.py b/google/cloud/dataflow_v1beta3/services/messages_v1_beta3/transports/grpc_asyncio.py index 8f67459..bc04c1d 100644 --- a/google/cloud/dataflow_v1beta3/services/messages_v1_beta3/transports/grpc_asyncio.py +++ b/google/cloud/dataflow_v1beta3/services/messages_v1_beta3/transports/grpc_asyncio.py @@ -49,7 +49,7 @@ class MessagesV1Beta3GrpcAsyncIOTransport(MessagesV1Beta3Transport): def create_channel( cls, host: str = "dataflow.googleapis.com", - credentials: ga_credentials.Credentials = None, + credentials: Optional[ga_credentials.Credentials] = None, credentials_file: Optional[str] = None, scopes: Optional[Sequence[str]] = None, quota_project_id: Optional[str] = None, @@ -92,15 +92,15 @@ def __init__( self, *, host: str = "dataflow.googleapis.com", - credentials: ga_credentials.Credentials = None, + credentials: Optional[ga_credentials.Credentials] = None, credentials_file: Optional[str] = None, scopes: Optional[Sequence[str]] = None, - channel: aio.Channel = None, - api_mtls_endpoint: str = None, - client_cert_source: Callable[[], Tuple[bytes, bytes]] = None, - ssl_channel_credentials: grpc.ChannelCredentials = None, - client_cert_source_for_mtls: Callable[[], Tuple[bytes, bytes]] = None, - quota_project_id=None, + channel: Optional[aio.Channel] = None, + api_mtls_endpoint: Optional[str] = None, + client_cert_source: Optional[Callable[[], Tuple[bytes, bytes]]] = None, + ssl_channel_credentials: Optional[grpc.ChannelCredentials] = None, + client_cert_source_for_mtls: Optional[Callable[[], Tuple[bytes, bytes]]] = None, + quota_project_id: Optional[str] = None, client_info: gapic_v1.client_info.ClientInfo = DEFAULT_CLIENT_INFO, always_use_jwt_access: Optional[bool] = False, api_audience: Optional[str] = None, diff --git a/google/cloud/dataflow_v1beta3/services/messages_v1_beta3/transports/rest.py b/google/cloud/dataflow_v1beta3/services/messages_v1_beta3/transports/rest.py index c08b7db..c6a61c7 100644 --- a/google/cloud/dataflow_v1beta3/services/messages_v1_beta3/transports/rest.py +++ b/google/cloud/dataflow_v1beta3/services/messages_v1_beta3/transports/rest.py @@ -63,12 +63,13 @@ class MessagesV1Beta3RestInterceptor: .. code-block:: python class MyCustomMessagesV1Beta3Interceptor(MessagesV1Beta3RestInterceptor): - def pre_list_job_messages(request, metadata): + def pre_list_job_messages(self, request, metadata): logging.log(f"Received request: {request}") return request, metadata - def post_list_job_messages(response): + def post_list_job_messages(self, response): logging.log(f"Received response: {response}") + return response transport = MessagesV1Beta3RestTransport(interceptor=MyCustomMessagesV1Beta3Interceptor()) client = MessagesV1Beta3Client(transport=transport) @@ -119,19 +120,16 @@ class MessagesV1Beta3RestTransport(MessagesV1Beta3Transport): It sends JSON representations of protocol buffers over HTTP/1.1 - NOTE: This REST transport functionality is currently in a beta - state (preview). We welcome your feedback via an issue in this - library's source repository. Thank you! """ def __init__( self, *, host: str = "dataflow.googleapis.com", - credentials: ga_credentials.Credentials = None, - credentials_file: str = None, - scopes: Sequence[str] = None, - client_cert_source_for_mtls: Callable[[], Tuple[bytes, bytes]] = None, + credentials: Optional[ga_credentials.Credentials] = None, + credentials_file: Optional[str] = None, + scopes: Optional[Sequence[str]] = None, + client_cert_source_for_mtls: Optional[Callable[[], Tuple[bytes, bytes]]] = None, quota_project_id: Optional[str] = None, client_info: gapic_v1.client_info.ClientInfo = DEFAULT_CLIENT_INFO, always_use_jwt_access: Optional[bool] = False, @@ -141,39 +139,35 @@ def __init__( ) -> None: """Instantiate the transport. - NOTE: This REST transport functionality is currently in a beta - state (preview). We welcome your feedback via a GitHub issue in - this library's repository. Thank you! - - Args: - host (Optional[str]): - The hostname to connect to. - credentials (Optional[google.auth.credentials.Credentials]): The - authorization credentials to attach to requests. These - credentials identify the application to the service; if none - are specified, the client will attempt to ascertain the - credentials from the environment. - - credentials_file (Optional[str]): A file with credentials that can - be loaded with :func:`google.auth.load_credentials_from_file`. - This argument is ignored if ``channel`` is provided. - scopes (Optional(Sequence[str])): A list of scopes. This argument is - ignored if ``channel`` is provided. - client_cert_source_for_mtls (Callable[[], Tuple[bytes, bytes]]): Client - certificate to configure mutual TLS HTTP channel. It is ignored - if ``channel`` is provided. - quota_project_id (Optional[str]): An optional project to use for billing - and quota. - client_info (google.api_core.gapic_v1.client_info.ClientInfo): - The client info used to send a user-agent string along with - API requests. If ``None``, then default info will be used. - Generally, you only need to set this if you are developing - your own client library. - always_use_jwt_access (Optional[bool]): Whether self signed JWT should - be used for service account credentials. - url_scheme: the protocol scheme for the API endpoint. Normally - "https", but for testing or local servers, - "http" can be specified. + Args: + host (Optional[str]): + The hostname to connect to. + credentials (Optional[google.auth.credentials.Credentials]): The + authorization credentials to attach to requests. These + credentials identify the application to the service; if none + are specified, the client will attempt to ascertain the + credentials from the environment. + + credentials_file (Optional[str]): A file with credentials that can + be loaded with :func:`google.auth.load_credentials_from_file`. + This argument is ignored if ``channel`` is provided. + scopes (Optional(Sequence[str])): A list of scopes. This argument is + ignored if ``channel`` is provided. + client_cert_source_for_mtls (Callable[[], Tuple[bytes, bytes]]): Client + certificate to configure mutual TLS HTTP channel. It is ignored + if ``channel`` is provided. + quota_project_id (Optional[str]): An optional project to use for billing + and quota. + client_info (google.api_core.gapic_v1.client_info.ClientInfo): + The client info used to send a user-agent string along with + API requests. If ``None``, then default info will be used. + Generally, you only need to set this if you are developing + your own client library. + always_use_jwt_access (Optional[bool]): Whether self signed JWT should + be used for service account credentials. + url_scheme: the protocol scheme for the API endpoint. Normally + "https", but for testing or local servers, + "http" can be specified. """ # Run the base constructor # TODO(yon-mg): resolve other ctor params i.e. scopes, quota, etc. @@ -213,7 +207,7 @@ def __call__( request: messages.ListJobMessagesRequest, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> messages.ListJobMessagesResponse: r"""Call the list job messages method over HTTP. @@ -263,10 +257,12 @@ def __call__( json_format.MessageToJson( transcoded_request["query_params"], including_default_value_fields=False, - use_integers_for_enums=False, + use_integers_for_enums=True, ) ) + query_params["$alt"] = "json;enum-encoding=int" + # Send the request headers = dict(metadata) headers["Content-Type"] = "application/json" diff --git a/google/cloud/dataflow_v1beta3/services/metrics_v1_beta3/async_client.py b/google/cloud/dataflow_v1beta3/services/metrics_v1_beta3/async_client.py index 1709efb..9af50ee 100644 --- a/google/cloud/dataflow_v1beta3/services/metrics_v1_beta3/async_client.py +++ b/google/cloud/dataflow_v1beta3/services/metrics_v1_beta3/async_client.py @@ -16,7 +16,17 @@ from collections import OrderedDict import functools import re -from typing import Dict, Mapping, Optional, Sequence, Tuple, Type, Union +from typing import ( + Dict, + Mapping, + MutableMapping, + MutableSequence, + Optional, + Sequence, + Tuple, + Type, + Union, +) from google.api_core import exceptions as core_exceptions from google.api_core import gapic_v1 @@ -24,7 +34,8 @@ from google.api_core.client_options import ClientOptions from google.auth import credentials as ga_credentials # type: ignore from google.oauth2 import service_account # type: ignore -import pkg_resources + +from google.cloud.dataflow_v1beta3 import gapic_version as package_version try: OptionalRetry = Union[retries.Retry, gapic_v1.method._MethodDefault] @@ -161,9 +172,9 @@ def transport(self) -> MetricsV1Beta3Transport: def __init__( self, *, - credentials: ga_credentials.Credentials = None, + credentials: Optional[ga_credentials.Credentials] = None, transport: Union[str, MetricsV1Beta3Transport] = "grpc_asyncio", - client_options: ClientOptions = None, + client_options: Optional[ClientOptions] = None, client_info: gapic_v1.client_info.ClientInfo = DEFAULT_CLIENT_INFO, ) -> None: """Instantiates the metrics v1 beta3 client. @@ -207,10 +218,10 @@ def __init__( async def get_job_metrics( self, - request: Union[metrics.GetJobMetricsRequest, dict] = None, + request: Optional[Union[metrics.GetJobMetricsRequest, dict]] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> metrics.JobMetrics: r"""Request the job status. @@ -249,7 +260,7 @@ async def sample_get_job_metrics(): print(response) Args: - request (Union[google.cloud.dataflow_v1beta3.types.GetJobMetricsRequest, dict]): + request (Optional[Union[google.cloud.dataflow_v1beta3.types.GetJobMetricsRequest, dict]]): The request object. Request to get job metrics. retry (google.api_core.retry.Retry): Designation of what errors, if any, should be retried. @@ -308,10 +319,10 @@ async def sample_get_job_metrics(): async def get_job_execution_details( self, - request: Union[metrics.GetJobExecutionDetailsRequest, dict] = None, + request: Optional[Union[metrics.GetJobExecutionDetailsRequest, dict]] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.GetJobExecutionDetailsAsyncPager: r"""Request detailed information about the execution @@ -346,7 +357,7 @@ async def sample_get_job_execution_details(): print(response) Args: - request (Union[google.cloud.dataflow_v1beta3.types.GetJobExecutionDetailsRequest, dict]): + request (Optional[Union[google.cloud.dataflow_v1beta3.types.GetJobExecutionDetailsRequest, dict]]): The request object. Request to get job execution details. retry (google.api_core.retry.Retry): Designation of what errors, if any, @@ -409,10 +420,10 @@ async def sample_get_job_execution_details(): async def get_stage_execution_details( self, - request: Union[metrics.GetStageExecutionDetailsRequest, dict] = None, + request: Optional[Union[metrics.GetStageExecutionDetailsRequest, dict]] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.GetStageExecutionDetailsAsyncPager: r"""Request detailed information about the execution @@ -448,7 +459,7 @@ async def sample_get_stage_execution_details(): print(response) Args: - request (Union[google.cloud.dataflow_v1beta3.types.GetStageExecutionDetailsRequest, dict]): + request (Optional[Union[google.cloud.dataflow_v1beta3.types.GetStageExecutionDetailsRequest, dict]]): The request object. Request to get information about a particular execution stage of a job. Currently only tracked for Batch jobs. @@ -518,14 +529,9 @@ async def __aexit__(self, exc_type, exc, tb): await self.transport.close() -try: - DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( - gapic_version=pkg_resources.get_distribution( - "google-cloud-dataflow-client", - ).version, - ) -except pkg_resources.DistributionNotFound: - DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo() +DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( + gapic_version=package_version.__version__ +) __all__ = ("MetricsV1Beta3AsyncClient",) diff --git a/google/cloud/dataflow_v1beta3/services/metrics_v1_beta3/client.py b/google/cloud/dataflow_v1beta3/services/metrics_v1_beta3/client.py index b032f51..faa6270 100644 --- a/google/cloud/dataflow_v1beta3/services/metrics_v1_beta3/client.py +++ b/google/cloud/dataflow_v1beta3/services/metrics_v1_beta3/client.py @@ -16,7 +16,18 @@ from collections import OrderedDict import os import re -from typing import Dict, Mapping, Optional, Sequence, Tuple, Type, Union +from typing import ( + Dict, + Mapping, + MutableMapping, + MutableSequence, + Optional, + Sequence, + Tuple, + Type, + Union, + cast, +) from google.api_core import client_options as client_options_lib from google.api_core import exceptions as core_exceptions @@ -27,7 +38,8 @@ from google.auth.transport import mtls # type: ignore from google.auth.transport.grpc import SslCredentials # type: ignore from google.oauth2 import service_account # type: ignore -import pkg_resources + +from google.cloud.dataflow_v1beta3 import gapic_version as package_version try: OptionalRetry = Union[retries.Retry, gapic_v1.method._MethodDefault] @@ -62,7 +74,7 @@ class MetricsV1Beta3ClientMeta(type): def get_transport_class( cls, - label: str = None, + label: Optional[str] = None, ) -> Type[MetricsV1Beta3Transport]: """Returns an appropriate transport class. @@ -317,8 +329,8 @@ def __init__( self, *, credentials: Optional[ga_credentials.Credentials] = None, - transport: Union[str, MetricsV1Beta3Transport, None] = None, - client_options: Optional[client_options_lib.ClientOptions] = None, + transport: Optional[Union[str, MetricsV1Beta3Transport]] = None, + client_options: Optional[Union[client_options_lib.ClientOptions, dict]] = None, client_info: gapic_v1.client_info.ClientInfo = DEFAULT_CLIENT_INFO, ) -> None: """Instantiates the metrics v1 beta3 client. @@ -332,10 +344,7 @@ def __init__( transport (Union[str, MetricsV1Beta3Transport]): The transport to use. If set to None, a transport is chosen automatically. - NOTE: "rest" transport functionality is currently in a - beta state (preview). We welcome your feedback via an - issue in this library's source repository. - client_options (google.api_core.client_options.ClientOptions): Custom options for the + client_options (Optional[Union[google.api_core.client_options.ClientOptions, dict]]): Custom options for the client. It won't take effect if a ``transport`` instance is provided. (1) The ``api_endpoint`` property can be used to override the default endpoint provided by the client. GOOGLE_API_USE_MTLS_ENDPOINT @@ -365,6 +374,7 @@ def __init__( client_options = client_options_lib.from_dict(client_options) if client_options is None: client_options = client_options_lib.ClientOptions() + client_options = cast(client_options_lib.ClientOptions, client_options) api_endpoint, client_cert_source_func = self.get_mtls_endpoint_and_cert_source( client_options @@ -417,10 +427,10 @@ def __init__( def get_job_metrics( self, - request: Union[metrics.GetJobMetricsRequest, dict] = None, + request: Optional[Union[metrics.GetJobMetricsRequest, dict]] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> metrics.JobMetrics: r"""Request the job status. @@ -519,10 +529,10 @@ def sample_get_job_metrics(): def get_job_execution_details( self, - request: Union[metrics.GetJobExecutionDetailsRequest, dict] = None, + request: Optional[Union[metrics.GetJobExecutionDetailsRequest, dict]] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.GetJobExecutionDetailsPager: r"""Request detailed information about the execution @@ -623,10 +633,10 @@ def sample_get_job_execution_details(): def get_stage_execution_details( self, - request: Union[metrics.GetStageExecutionDetailsRequest, dict] = None, + request: Optional[Union[metrics.GetStageExecutionDetailsRequest, dict]] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.GetStageExecutionDetailsPager: r"""Request detailed information about the execution @@ -742,14 +752,9 @@ def __exit__(self, type, value, traceback): self.transport.close() -try: - DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( - gapic_version=pkg_resources.get_distribution( - "google-cloud-dataflow-client", - ).version, - ) -except pkg_resources.DistributionNotFound: - DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo() +DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( + gapic_version=package_version.__version__ +) __all__ = ("MetricsV1Beta3Client",) diff --git a/google/cloud/dataflow_v1beta3/services/metrics_v1_beta3/transports/base.py b/google/cloud/dataflow_v1beta3/services/metrics_v1_beta3/transports/base.py index 3b0687a..902ce8c 100644 --- a/google/cloud/dataflow_v1beta3/services/metrics_v1_beta3/transports/base.py +++ b/google/cloud/dataflow_v1beta3/services/metrics_v1_beta3/transports/base.py @@ -23,18 +23,13 @@ import google.auth # type: ignore from google.auth import credentials as ga_credentials # type: ignore from google.oauth2 import service_account # type: ignore -import pkg_resources +from google.cloud.dataflow_v1beta3 import gapic_version as package_version from google.cloud.dataflow_v1beta3.types import metrics -try: - DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( - gapic_version=pkg_resources.get_distribution( - "google-cloud-dataflow-client", - ).version, - ) -except pkg_resources.DistributionNotFound: - DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo() +DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( + gapic_version=package_version.__version__ +) class MetricsV1Beta3Transport(abc.ABC): @@ -53,7 +48,7 @@ def __init__( self, *, host: str = DEFAULT_HOST, - credentials: ga_credentials.Credentials = None, + credentials: Optional[ga_credentials.Credentials] = None, credentials_file: Optional[str] = None, scopes: Optional[Sequence[str]] = None, quota_project_id: Optional[str] = None, diff --git a/google/cloud/dataflow_v1beta3/services/metrics_v1_beta3/transports/grpc.py b/google/cloud/dataflow_v1beta3/services/metrics_v1_beta3/transports/grpc.py index f48b555..76d8392 100644 --- a/google/cloud/dataflow_v1beta3/services/metrics_v1_beta3/transports/grpc.py +++ b/google/cloud/dataflow_v1beta3/services/metrics_v1_beta3/transports/grpc.py @@ -47,14 +47,14 @@ def __init__( self, *, host: str = "dataflow.googleapis.com", - credentials: ga_credentials.Credentials = None, - credentials_file: str = None, - scopes: Sequence[str] = None, - channel: grpc.Channel = None, - api_mtls_endpoint: str = None, - client_cert_source: Callable[[], Tuple[bytes, bytes]] = None, - ssl_channel_credentials: grpc.ChannelCredentials = None, - client_cert_source_for_mtls: Callable[[], Tuple[bytes, bytes]] = None, + credentials: Optional[ga_credentials.Credentials] = None, + credentials_file: Optional[str] = None, + scopes: Optional[Sequence[str]] = None, + channel: Optional[grpc.Channel] = None, + api_mtls_endpoint: Optional[str] = None, + client_cert_source: Optional[Callable[[], Tuple[bytes, bytes]]] = None, + ssl_channel_credentials: Optional[grpc.ChannelCredentials] = None, + client_cert_source_for_mtls: Optional[Callable[[], Tuple[bytes, bytes]]] = None, quota_project_id: Optional[str] = None, client_info: gapic_v1.client_info.ClientInfo = DEFAULT_CLIENT_INFO, always_use_jwt_access: Optional[bool] = False, @@ -181,8 +181,8 @@ def __init__( def create_channel( cls, host: str = "dataflow.googleapis.com", - credentials: ga_credentials.Credentials = None, - credentials_file: str = None, + credentials: Optional[ga_credentials.Credentials] = None, + credentials_file: Optional[str] = None, scopes: Optional[Sequence[str]] = None, quota_project_id: Optional[str] = None, **kwargs, diff --git a/google/cloud/dataflow_v1beta3/services/metrics_v1_beta3/transports/grpc_asyncio.py b/google/cloud/dataflow_v1beta3/services/metrics_v1_beta3/transports/grpc_asyncio.py index 112c320..95e2992 100644 --- a/google/cloud/dataflow_v1beta3/services/metrics_v1_beta3/transports/grpc_asyncio.py +++ b/google/cloud/dataflow_v1beta3/services/metrics_v1_beta3/transports/grpc_asyncio.py @@ -49,7 +49,7 @@ class MetricsV1Beta3GrpcAsyncIOTransport(MetricsV1Beta3Transport): def create_channel( cls, host: str = "dataflow.googleapis.com", - credentials: ga_credentials.Credentials = None, + credentials: Optional[ga_credentials.Credentials] = None, credentials_file: Optional[str] = None, scopes: Optional[Sequence[str]] = None, quota_project_id: Optional[str] = None, @@ -92,15 +92,15 @@ def __init__( self, *, host: str = "dataflow.googleapis.com", - credentials: ga_credentials.Credentials = None, + credentials: Optional[ga_credentials.Credentials] = None, credentials_file: Optional[str] = None, scopes: Optional[Sequence[str]] = None, - channel: aio.Channel = None, - api_mtls_endpoint: str = None, - client_cert_source: Callable[[], Tuple[bytes, bytes]] = None, - ssl_channel_credentials: grpc.ChannelCredentials = None, - client_cert_source_for_mtls: Callable[[], Tuple[bytes, bytes]] = None, - quota_project_id=None, + channel: Optional[aio.Channel] = None, + api_mtls_endpoint: Optional[str] = None, + client_cert_source: Optional[Callable[[], Tuple[bytes, bytes]]] = None, + ssl_channel_credentials: Optional[grpc.ChannelCredentials] = None, + client_cert_source_for_mtls: Optional[Callable[[], Tuple[bytes, bytes]]] = None, + quota_project_id: Optional[str] = None, client_info: gapic_v1.client_info.ClientInfo = DEFAULT_CLIENT_INFO, always_use_jwt_access: Optional[bool] = False, api_audience: Optional[str] = None, diff --git a/google/cloud/dataflow_v1beta3/services/metrics_v1_beta3/transports/rest.py b/google/cloud/dataflow_v1beta3/services/metrics_v1_beta3/transports/rest.py index f889276..f32c0c9 100644 --- a/google/cloud/dataflow_v1beta3/services/metrics_v1_beta3/transports/rest.py +++ b/google/cloud/dataflow_v1beta3/services/metrics_v1_beta3/transports/rest.py @@ -63,26 +63,29 @@ class MetricsV1Beta3RestInterceptor: .. code-block:: python class MyCustomMetricsV1Beta3Interceptor(MetricsV1Beta3RestInterceptor): - def pre_get_job_execution_details(request, metadata): + def pre_get_job_execution_details(self, request, metadata): logging.log(f"Received request: {request}") return request, metadata - def post_get_job_execution_details(response): + def post_get_job_execution_details(self, response): logging.log(f"Received response: {response}") + return response - def pre_get_job_metrics(request, metadata): + def pre_get_job_metrics(self, request, metadata): logging.log(f"Received request: {request}") return request, metadata - def post_get_job_metrics(response): + def post_get_job_metrics(self, response): logging.log(f"Received response: {response}") + return response - def pre_get_stage_execution_details(request, metadata): + def pre_get_stage_execution_details(self, request, metadata): logging.log(f"Received request: {request}") return request, metadata - def post_get_stage_execution_details(response): + def post_get_stage_execution_details(self, response): logging.log(f"Received response: {response}") + return response transport = MetricsV1Beta3RestTransport(interceptor=MyCustomMetricsV1Beta3Interceptor()) client = MetricsV1Beta3Client(transport=transport) @@ -175,19 +178,16 @@ class MetricsV1Beta3RestTransport(MetricsV1Beta3Transport): It sends JSON representations of protocol buffers over HTTP/1.1 - NOTE: This REST transport functionality is currently in a beta - state (preview). We welcome your feedback via an issue in this - library's source repository. Thank you! """ def __init__( self, *, host: str = "dataflow.googleapis.com", - credentials: ga_credentials.Credentials = None, - credentials_file: str = None, - scopes: Sequence[str] = None, - client_cert_source_for_mtls: Callable[[], Tuple[bytes, bytes]] = None, + credentials: Optional[ga_credentials.Credentials] = None, + credentials_file: Optional[str] = None, + scopes: Optional[Sequence[str]] = None, + client_cert_source_for_mtls: Optional[Callable[[], Tuple[bytes, bytes]]] = None, quota_project_id: Optional[str] = None, client_info: gapic_v1.client_info.ClientInfo = DEFAULT_CLIENT_INFO, always_use_jwt_access: Optional[bool] = False, @@ -197,39 +197,35 @@ def __init__( ) -> None: """Instantiate the transport. - NOTE: This REST transport functionality is currently in a beta - state (preview). We welcome your feedback via a GitHub issue in - this library's repository. Thank you! - - Args: - host (Optional[str]): - The hostname to connect to. - credentials (Optional[google.auth.credentials.Credentials]): The - authorization credentials to attach to requests. These - credentials identify the application to the service; if none - are specified, the client will attempt to ascertain the - credentials from the environment. - - credentials_file (Optional[str]): A file with credentials that can - be loaded with :func:`google.auth.load_credentials_from_file`. - This argument is ignored if ``channel`` is provided. - scopes (Optional(Sequence[str])): A list of scopes. This argument is - ignored if ``channel`` is provided. - client_cert_source_for_mtls (Callable[[], Tuple[bytes, bytes]]): Client - certificate to configure mutual TLS HTTP channel. It is ignored - if ``channel`` is provided. - quota_project_id (Optional[str]): An optional project to use for billing - and quota. - client_info (google.api_core.gapic_v1.client_info.ClientInfo): - The client info used to send a user-agent string along with - API requests. If ``None``, then default info will be used. - Generally, you only need to set this if you are developing - your own client library. - always_use_jwt_access (Optional[bool]): Whether self signed JWT should - be used for service account credentials. - url_scheme: the protocol scheme for the API endpoint. Normally - "https", but for testing or local servers, - "http" can be specified. + Args: + host (Optional[str]): + The hostname to connect to. + credentials (Optional[google.auth.credentials.Credentials]): The + authorization credentials to attach to requests. These + credentials identify the application to the service; if none + are specified, the client will attempt to ascertain the + credentials from the environment. + + credentials_file (Optional[str]): A file with credentials that can + be loaded with :func:`google.auth.load_credentials_from_file`. + This argument is ignored if ``channel`` is provided. + scopes (Optional(Sequence[str])): A list of scopes. This argument is + ignored if ``channel`` is provided. + client_cert_source_for_mtls (Callable[[], Tuple[bytes, bytes]]): Client + certificate to configure mutual TLS HTTP channel. It is ignored + if ``channel`` is provided. + quota_project_id (Optional[str]): An optional project to use for billing + and quota. + client_info (google.api_core.gapic_v1.client_info.ClientInfo): + The client info used to send a user-agent string along with + API requests. If ``None``, then default info will be used. + Generally, you only need to set this if you are developing + your own client library. + always_use_jwt_access (Optional[bool]): Whether self signed JWT should + be used for service account credentials. + url_scheme: the protocol scheme for the API endpoint. Normally + "https", but for testing or local servers, + "http" can be specified. """ # Run the base constructor # TODO(yon-mg): resolve other ctor params i.e. scopes, quota, etc. @@ -269,7 +265,7 @@ def __call__( request: metrics.GetJobExecutionDetailsRequest, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> metrics.JobExecutionDetails: r"""Call the get job execution details method over HTTP. @@ -310,10 +306,12 @@ def __call__( json_format.MessageToJson( transcoded_request["query_params"], including_default_value_fields=False, - use_integers_for_enums=False, + use_integers_for_enums=True, ) ) + query_params["$alt"] = "json;enum-encoding=int" + # Send the request headers = dict(metadata) headers["Content-Type"] = "application/json" @@ -346,7 +344,7 @@ def __call__( request: metrics.GetJobMetricsRequest, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> metrics.JobMetrics: r"""Call the get job metrics method over HTTP. @@ -398,10 +396,12 @@ def __call__( json_format.MessageToJson( transcoded_request["query_params"], including_default_value_fields=False, - use_integers_for_enums=False, + use_integers_for_enums=True, ) ) + query_params["$alt"] = "json;enum-encoding=int" + # Send the request headers = dict(metadata) headers["Content-Type"] = "application/json" @@ -434,7 +434,7 @@ def __call__( request: metrics.GetStageExecutionDetailsRequest, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> metrics.StageExecutionDetails: r"""Call the get stage execution @@ -479,10 +479,12 @@ def __call__( json_format.MessageToJson( transcoded_request["query_params"], including_default_value_fields=False, - use_integers_for_enums=False, + use_integers_for_enums=True, ) ) + query_params["$alt"] = "json;enum-encoding=int" + # Send the request headers = dict(metadata) headers["Content-Type"] = "application/json" diff --git a/google/cloud/dataflow_v1beta3/services/snapshots_v1_beta3/async_client.py b/google/cloud/dataflow_v1beta3/services/snapshots_v1_beta3/async_client.py index 5af9753..b0593b5 100644 --- a/google/cloud/dataflow_v1beta3/services/snapshots_v1_beta3/async_client.py +++ b/google/cloud/dataflow_v1beta3/services/snapshots_v1_beta3/async_client.py @@ -16,7 +16,17 @@ from collections import OrderedDict import functools import re -from typing import Dict, Mapping, Optional, Sequence, Tuple, Type, Union +from typing import ( + Dict, + Mapping, + MutableMapping, + MutableSequence, + Optional, + Sequence, + Tuple, + Type, + Union, +) from google.api_core import exceptions as core_exceptions from google.api_core import gapic_v1 @@ -24,7 +34,8 @@ from google.api_core.client_options import ClientOptions from google.auth import credentials as ga_credentials # type: ignore from google.oauth2 import service_account # type: ignore -import pkg_resources + +from google.cloud.dataflow_v1beta3 import gapic_version as package_version try: OptionalRetry = Union[retries.Retry, gapic_v1.method._MethodDefault] @@ -161,9 +172,9 @@ def transport(self) -> SnapshotsV1Beta3Transport: def __init__( self, *, - credentials: ga_credentials.Credentials = None, + credentials: Optional[ga_credentials.Credentials] = None, transport: Union[str, SnapshotsV1Beta3Transport] = "grpc_asyncio", - client_options: ClientOptions = None, + client_options: Optional[ClientOptions] = None, client_info: gapic_v1.client_info.ClientInfo = DEFAULT_CLIENT_INFO, ) -> None: """Instantiates the snapshots v1 beta3 client. @@ -207,10 +218,10 @@ def __init__( async def get_snapshot( self, - request: Union[snapshots.GetSnapshotRequest, dict] = None, + request: Optional[Union[snapshots.GetSnapshotRequest, dict]] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> snapshots.Snapshot: r"""Gets information about a snapshot. @@ -241,7 +252,7 @@ async def sample_get_snapshot(): print(response) Args: - request (Union[google.cloud.dataflow_v1beta3.types.GetSnapshotRequest, dict]): + request (Optional[Union[google.cloud.dataflow_v1beta3.types.GetSnapshotRequest, dict]]): The request object. Request to get information about a snapshot retry (google.api_core.retry.Retry): Designation of what errors, if any, @@ -290,10 +301,10 @@ async def sample_get_snapshot(): async def delete_snapshot( self, - request: Union[snapshots.DeleteSnapshotRequest, dict] = None, + request: Optional[Union[snapshots.DeleteSnapshotRequest, dict]] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> snapshots.DeleteSnapshotResponse: r"""Deletes a snapshot. @@ -324,7 +335,7 @@ async def sample_delete_snapshot(): print(response) Args: - request (Union[google.cloud.dataflow_v1beta3.types.DeleteSnapshotRequest, dict]): + request (Optional[Union[google.cloud.dataflow_v1beta3.types.DeleteSnapshotRequest, dict]]): The request object. Request to delete a snapshot. retry (google.api_core.retry.Retry): Designation of what errors, if any, should be retried. @@ -372,10 +383,10 @@ async def sample_delete_snapshot(): async def list_snapshots( self, - request: Union[snapshots.ListSnapshotsRequest, dict] = None, + request: Optional[Union[snapshots.ListSnapshotsRequest, dict]] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> snapshots.ListSnapshotsResponse: r"""Lists snapshots. @@ -406,7 +417,7 @@ async def sample_list_snapshots(): print(response) Args: - request (Union[google.cloud.dataflow_v1beta3.types.ListSnapshotsRequest, dict]): + request (Optional[Union[google.cloud.dataflow_v1beta3.types.ListSnapshotsRequest, dict]]): The request object. Request to list snapshots. retry (google.api_core.retry.Retry): Designation of what errors, if any, should be retried. @@ -459,14 +470,9 @@ async def __aexit__(self, exc_type, exc, tb): await self.transport.close() -try: - DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( - gapic_version=pkg_resources.get_distribution( - "google-cloud-dataflow-client", - ).version, - ) -except pkg_resources.DistributionNotFound: - DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo() +DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( + gapic_version=package_version.__version__ +) __all__ = ("SnapshotsV1Beta3AsyncClient",) diff --git a/google/cloud/dataflow_v1beta3/services/snapshots_v1_beta3/client.py b/google/cloud/dataflow_v1beta3/services/snapshots_v1_beta3/client.py index 015ead3..b3fc2cc 100644 --- a/google/cloud/dataflow_v1beta3/services/snapshots_v1_beta3/client.py +++ b/google/cloud/dataflow_v1beta3/services/snapshots_v1_beta3/client.py @@ -16,7 +16,18 @@ from collections import OrderedDict import os import re -from typing import Dict, Mapping, Optional, Sequence, Tuple, Type, Union +from typing import ( + Dict, + Mapping, + MutableMapping, + MutableSequence, + Optional, + Sequence, + Tuple, + Type, + Union, + cast, +) from google.api_core import client_options as client_options_lib from google.api_core import exceptions as core_exceptions @@ -27,7 +38,8 @@ from google.auth.transport import mtls # type: ignore from google.auth.transport.grpc import SslCredentials # type: ignore from google.oauth2 import service_account # type: ignore -import pkg_resources + +from google.cloud.dataflow_v1beta3 import gapic_version as package_version try: OptionalRetry = Union[retries.Retry, gapic_v1.method._MethodDefault] @@ -62,7 +74,7 @@ class SnapshotsV1Beta3ClientMeta(type): def get_transport_class( cls, - label: str = None, + label: Optional[str] = None, ) -> Type[SnapshotsV1Beta3Transport]: """Returns an appropriate transport class. @@ -317,8 +329,8 @@ def __init__( self, *, credentials: Optional[ga_credentials.Credentials] = None, - transport: Union[str, SnapshotsV1Beta3Transport, None] = None, - client_options: Optional[client_options_lib.ClientOptions] = None, + transport: Optional[Union[str, SnapshotsV1Beta3Transport]] = None, + client_options: Optional[Union[client_options_lib.ClientOptions, dict]] = None, client_info: gapic_v1.client_info.ClientInfo = DEFAULT_CLIENT_INFO, ) -> None: """Instantiates the snapshots v1 beta3 client. @@ -332,10 +344,7 @@ def __init__( transport (Union[str, SnapshotsV1Beta3Transport]): The transport to use. If set to None, a transport is chosen automatically. - NOTE: "rest" transport functionality is currently in a - beta state (preview). We welcome your feedback via an - issue in this library's source repository. - client_options (google.api_core.client_options.ClientOptions): Custom options for the + client_options (Optional[Union[google.api_core.client_options.ClientOptions, dict]]): Custom options for the client. It won't take effect if a ``transport`` instance is provided. (1) The ``api_endpoint`` property can be used to override the default endpoint provided by the client. GOOGLE_API_USE_MTLS_ENDPOINT @@ -365,6 +374,7 @@ def __init__( client_options = client_options_lib.from_dict(client_options) if client_options is None: client_options = client_options_lib.ClientOptions() + client_options = cast(client_options_lib.ClientOptions, client_options) api_endpoint, client_cert_source_func = self.get_mtls_endpoint_and_cert_source( client_options @@ -417,10 +427,10 @@ def __init__( def get_snapshot( self, - request: Union[snapshots.GetSnapshotRequest, dict] = None, + request: Optional[Union[snapshots.GetSnapshotRequest, dict]] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> snapshots.Snapshot: r"""Gets information about a snapshot. @@ -501,10 +511,10 @@ def sample_get_snapshot(): def delete_snapshot( self, - request: Union[snapshots.DeleteSnapshotRequest, dict] = None, + request: Optional[Union[snapshots.DeleteSnapshotRequest, dict]] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> snapshots.DeleteSnapshotResponse: r"""Deletes a snapshot. @@ -584,10 +594,10 @@ def sample_delete_snapshot(): def list_snapshots( self, - request: Union[snapshots.ListSnapshotsRequest, dict] = None, + request: Optional[Union[snapshots.ListSnapshotsRequest, dict]] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> snapshots.ListSnapshotsResponse: r"""Lists snapshots. @@ -679,14 +689,9 @@ def __exit__(self, type, value, traceback): self.transport.close() -try: - DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( - gapic_version=pkg_resources.get_distribution( - "google-cloud-dataflow-client", - ).version, - ) -except pkg_resources.DistributionNotFound: - DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo() +DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( + gapic_version=package_version.__version__ +) __all__ = ("SnapshotsV1Beta3Client",) diff --git a/google/cloud/dataflow_v1beta3/services/snapshots_v1_beta3/transports/base.py b/google/cloud/dataflow_v1beta3/services/snapshots_v1_beta3/transports/base.py index 1f6d366..f82725b 100644 --- a/google/cloud/dataflow_v1beta3/services/snapshots_v1_beta3/transports/base.py +++ b/google/cloud/dataflow_v1beta3/services/snapshots_v1_beta3/transports/base.py @@ -23,18 +23,13 @@ import google.auth # type: ignore from google.auth import credentials as ga_credentials # type: ignore from google.oauth2 import service_account # type: ignore -import pkg_resources +from google.cloud.dataflow_v1beta3 import gapic_version as package_version from google.cloud.dataflow_v1beta3.types import snapshots -try: - DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( - gapic_version=pkg_resources.get_distribution( - "google-cloud-dataflow-client", - ).version, - ) -except pkg_resources.DistributionNotFound: - DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo() +DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( + gapic_version=package_version.__version__ +) class SnapshotsV1Beta3Transport(abc.ABC): @@ -53,7 +48,7 @@ def __init__( self, *, host: str = DEFAULT_HOST, - credentials: ga_credentials.Credentials = None, + credentials: Optional[ga_credentials.Credentials] = None, credentials_file: Optional[str] = None, scopes: Optional[Sequence[str]] = None, quota_project_id: Optional[str] = None, diff --git a/google/cloud/dataflow_v1beta3/services/snapshots_v1_beta3/transports/grpc.py b/google/cloud/dataflow_v1beta3/services/snapshots_v1_beta3/transports/grpc.py index 3bd9e3a..b0a1fb0 100644 --- a/google/cloud/dataflow_v1beta3/services/snapshots_v1_beta3/transports/grpc.py +++ b/google/cloud/dataflow_v1beta3/services/snapshots_v1_beta3/transports/grpc.py @@ -47,14 +47,14 @@ def __init__( self, *, host: str = "dataflow.googleapis.com", - credentials: ga_credentials.Credentials = None, - credentials_file: str = None, - scopes: Sequence[str] = None, - channel: grpc.Channel = None, - api_mtls_endpoint: str = None, - client_cert_source: Callable[[], Tuple[bytes, bytes]] = None, - ssl_channel_credentials: grpc.ChannelCredentials = None, - client_cert_source_for_mtls: Callable[[], Tuple[bytes, bytes]] = None, + credentials: Optional[ga_credentials.Credentials] = None, + credentials_file: Optional[str] = None, + scopes: Optional[Sequence[str]] = None, + channel: Optional[grpc.Channel] = None, + api_mtls_endpoint: Optional[str] = None, + client_cert_source: Optional[Callable[[], Tuple[bytes, bytes]]] = None, + ssl_channel_credentials: Optional[grpc.ChannelCredentials] = None, + client_cert_source_for_mtls: Optional[Callable[[], Tuple[bytes, bytes]]] = None, quota_project_id: Optional[str] = None, client_info: gapic_v1.client_info.ClientInfo = DEFAULT_CLIENT_INFO, always_use_jwt_access: Optional[bool] = False, @@ -181,8 +181,8 @@ def __init__( def create_channel( cls, host: str = "dataflow.googleapis.com", - credentials: ga_credentials.Credentials = None, - credentials_file: str = None, + credentials: Optional[ga_credentials.Credentials] = None, + credentials_file: Optional[str] = None, scopes: Optional[Sequence[str]] = None, quota_project_id: Optional[str] = None, **kwargs, diff --git a/google/cloud/dataflow_v1beta3/services/snapshots_v1_beta3/transports/grpc_asyncio.py b/google/cloud/dataflow_v1beta3/services/snapshots_v1_beta3/transports/grpc_asyncio.py index 2137419..924ca2a 100644 --- a/google/cloud/dataflow_v1beta3/services/snapshots_v1_beta3/transports/grpc_asyncio.py +++ b/google/cloud/dataflow_v1beta3/services/snapshots_v1_beta3/transports/grpc_asyncio.py @@ -49,7 +49,7 @@ class SnapshotsV1Beta3GrpcAsyncIOTransport(SnapshotsV1Beta3Transport): def create_channel( cls, host: str = "dataflow.googleapis.com", - credentials: ga_credentials.Credentials = None, + credentials: Optional[ga_credentials.Credentials] = None, credentials_file: Optional[str] = None, scopes: Optional[Sequence[str]] = None, quota_project_id: Optional[str] = None, @@ -92,15 +92,15 @@ def __init__( self, *, host: str = "dataflow.googleapis.com", - credentials: ga_credentials.Credentials = None, + credentials: Optional[ga_credentials.Credentials] = None, credentials_file: Optional[str] = None, scopes: Optional[Sequence[str]] = None, - channel: aio.Channel = None, - api_mtls_endpoint: str = None, - client_cert_source: Callable[[], Tuple[bytes, bytes]] = None, - ssl_channel_credentials: grpc.ChannelCredentials = None, - client_cert_source_for_mtls: Callable[[], Tuple[bytes, bytes]] = None, - quota_project_id=None, + channel: Optional[aio.Channel] = None, + api_mtls_endpoint: Optional[str] = None, + client_cert_source: Optional[Callable[[], Tuple[bytes, bytes]]] = None, + ssl_channel_credentials: Optional[grpc.ChannelCredentials] = None, + client_cert_source_for_mtls: Optional[Callable[[], Tuple[bytes, bytes]]] = None, + quota_project_id: Optional[str] = None, client_info: gapic_v1.client_info.ClientInfo = DEFAULT_CLIENT_INFO, always_use_jwt_access: Optional[bool] = False, api_audience: Optional[str] = None, diff --git a/google/cloud/dataflow_v1beta3/services/snapshots_v1_beta3/transports/rest.py b/google/cloud/dataflow_v1beta3/services/snapshots_v1_beta3/transports/rest.py index e92b4df..8f32c4b 100644 --- a/google/cloud/dataflow_v1beta3/services/snapshots_v1_beta3/transports/rest.py +++ b/google/cloud/dataflow_v1beta3/services/snapshots_v1_beta3/transports/rest.py @@ -63,26 +63,29 @@ class SnapshotsV1Beta3RestInterceptor: .. code-block:: python class MyCustomSnapshotsV1Beta3Interceptor(SnapshotsV1Beta3RestInterceptor): - def pre_delete_snapshot(request, metadata): + def pre_delete_snapshot(self, request, metadata): logging.log(f"Received request: {request}") return request, metadata - def post_delete_snapshot(response): + def post_delete_snapshot(self, response): logging.log(f"Received response: {response}") + return response - def pre_get_snapshot(request, metadata): + def pre_get_snapshot(self, request, metadata): logging.log(f"Received request: {request}") return request, metadata - def post_get_snapshot(response): + def post_get_snapshot(self, response): logging.log(f"Received response: {response}") + return response - def pre_list_snapshots(request, metadata): + def pre_list_snapshots(self, request, metadata): logging.log(f"Received request: {request}") return request, metadata - def post_list_snapshots(response): + def post_list_snapshots(self, response): logging.log(f"Received response: {response}") + return response transport = SnapshotsV1Beta3RestTransport(interceptor=MyCustomSnapshotsV1Beta3Interceptor()) client = SnapshotsV1Beta3Client(transport=transport) @@ -175,19 +178,16 @@ class SnapshotsV1Beta3RestTransport(SnapshotsV1Beta3Transport): It sends JSON representations of protocol buffers over HTTP/1.1 - NOTE: This REST transport functionality is currently in a beta - state (preview). We welcome your feedback via an issue in this - library's source repository. Thank you! """ def __init__( self, *, host: str = "dataflow.googleapis.com", - credentials: ga_credentials.Credentials = None, - credentials_file: str = None, - scopes: Sequence[str] = None, - client_cert_source_for_mtls: Callable[[], Tuple[bytes, bytes]] = None, + credentials: Optional[ga_credentials.Credentials] = None, + credentials_file: Optional[str] = None, + scopes: Optional[Sequence[str]] = None, + client_cert_source_for_mtls: Optional[Callable[[], Tuple[bytes, bytes]]] = None, quota_project_id: Optional[str] = None, client_info: gapic_v1.client_info.ClientInfo = DEFAULT_CLIENT_INFO, always_use_jwt_access: Optional[bool] = False, @@ -197,39 +197,35 @@ def __init__( ) -> None: """Instantiate the transport. - NOTE: This REST transport functionality is currently in a beta - state (preview). We welcome your feedback via a GitHub issue in - this library's repository. Thank you! - - Args: - host (Optional[str]): - The hostname to connect to. - credentials (Optional[google.auth.credentials.Credentials]): The - authorization credentials to attach to requests. These - credentials identify the application to the service; if none - are specified, the client will attempt to ascertain the - credentials from the environment. - - credentials_file (Optional[str]): A file with credentials that can - be loaded with :func:`google.auth.load_credentials_from_file`. - This argument is ignored if ``channel`` is provided. - scopes (Optional(Sequence[str])): A list of scopes. This argument is - ignored if ``channel`` is provided. - client_cert_source_for_mtls (Callable[[], Tuple[bytes, bytes]]): Client - certificate to configure mutual TLS HTTP channel. It is ignored - if ``channel`` is provided. - quota_project_id (Optional[str]): An optional project to use for billing - and quota. - client_info (google.api_core.gapic_v1.client_info.ClientInfo): - The client info used to send a user-agent string along with - API requests. If ``None``, then default info will be used. - Generally, you only need to set this if you are developing - your own client library. - always_use_jwt_access (Optional[bool]): Whether self signed JWT should - be used for service account credentials. - url_scheme: the protocol scheme for the API endpoint. Normally - "https", but for testing or local servers, - "http" can be specified. + Args: + host (Optional[str]): + The hostname to connect to. + credentials (Optional[google.auth.credentials.Credentials]): The + authorization credentials to attach to requests. These + credentials identify the application to the service; if none + are specified, the client will attempt to ascertain the + credentials from the environment. + + credentials_file (Optional[str]): A file with credentials that can + be loaded with :func:`google.auth.load_credentials_from_file`. + This argument is ignored if ``channel`` is provided. + scopes (Optional(Sequence[str])): A list of scopes. This argument is + ignored if ``channel`` is provided. + client_cert_source_for_mtls (Callable[[], Tuple[bytes, bytes]]): Client + certificate to configure mutual TLS HTTP channel. It is ignored + if ``channel`` is provided. + quota_project_id (Optional[str]): An optional project to use for billing + and quota. + client_info (google.api_core.gapic_v1.client_info.ClientInfo): + The client info used to send a user-agent string along with + API requests. If ``None``, then default info will be used. + Generally, you only need to set this if you are developing + your own client library. + always_use_jwt_access (Optional[bool]): Whether self signed JWT should + be used for service account credentials. + url_scheme: the protocol scheme for the API endpoint. Normally + "https", but for testing or local servers, + "http" can be specified. """ # Run the base constructor # TODO(yon-mg): resolve other ctor params i.e. scopes, quota, etc. @@ -269,7 +265,7 @@ def __call__( request: snapshots.DeleteSnapshotRequest, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> snapshots.DeleteSnapshotResponse: r"""Call the delete snapshot method over HTTP. @@ -310,10 +306,12 @@ def __call__( json_format.MessageToJson( transcoded_request["query_params"], including_default_value_fields=False, - use_integers_for_enums=False, + use_integers_for_enums=True, ) ) + query_params["$alt"] = "json;enum-encoding=int" + # Send the request headers = dict(metadata) headers["Content-Type"] = "application/json" @@ -346,7 +344,7 @@ def __call__( request: snapshots.GetSnapshotRequest, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> snapshots.Snapshot: r"""Call the get snapshot method over HTTP. @@ -389,10 +387,12 @@ def __call__( json_format.MessageToJson( transcoded_request["query_params"], including_default_value_fields=False, - use_integers_for_enums=False, + use_integers_for_enums=True, ) ) + query_params["$alt"] = "json;enum-encoding=int" + # Send the request headers = dict(metadata) headers["Content-Type"] = "application/json" @@ -425,7 +425,7 @@ def __call__( request: snapshots.ListSnapshotsRequest, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> snapshots.ListSnapshotsResponse: r"""Call the list snapshots method over HTTP. @@ -470,10 +470,12 @@ def __call__( json_format.MessageToJson( transcoded_request["query_params"], including_default_value_fields=False, - use_integers_for_enums=False, + use_integers_for_enums=True, ) ) + query_params["$alt"] = "json;enum-encoding=int" + # Send the request headers = dict(metadata) headers["Content-Type"] = "application/json" diff --git a/google/cloud/dataflow_v1beta3/services/templates_service/async_client.py b/google/cloud/dataflow_v1beta3/services/templates_service/async_client.py index 32af536..8766d74 100644 --- a/google/cloud/dataflow_v1beta3/services/templates_service/async_client.py +++ b/google/cloud/dataflow_v1beta3/services/templates_service/async_client.py @@ -16,7 +16,17 @@ from collections import OrderedDict import functools import re -from typing import Dict, Mapping, Optional, Sequence, Tuple, Type, Union +from typing import ( + Dict, + Mapping, + MutableMapping, + MutableSequence, + Optional, + Sequence, + Tuple, + Type, + Union, +) from google.api_core import exceptions as core_exceptions from google.api_core import gapic_v1 @@ -24,7 +34,8 @@ from google.api_core.client_options import ClientOptions from google.auth import credentials as ga_credentials # type: ignore from google.oauth2 import service_account # type: ignore -import pkg_resources + +from google.cloud.dataflow_v1beta3 import gapic_version as package_version try: OptionalRetry = Union[retries.Retry, gapic_v1.method._MethodDefault] @@ -161,9 +172,9 @@ def transport(self) -> TemplatesServiceTransport: def __init__( self, *, - credentials: ga_credentials.Credentials = None, + credentials: Optional[ga_credentials.Credentials] = None, transport: Union[str, TemplatesServiceTransport] = "grpc_asyncio", - client_options: ClientOptions = None, + client_options: Optional[ClientOptions] = None, client_info: gapic_v1.client_info.ClientInfo = DEFAULT_CLIENT_INFO, ) -> None: """Instantiates the templates service client. @@ -207,10 +218,10 @@ def __init__( async def create_job_from_template( self, - request: Union[templates.CreateJobFromTemplateRequest, dict] = None, + request: Optional[Union[templates.CreateJobFromTemplateRequest, dict]] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> jobs.Job: r"""Creates a Cloud Dataflow job from a template. @@ -242,7 +253,7 @@ async def sample_create_job_from_template(): print(response) Args: - request (Union[google.cloud.dataflow_v1beta3.types.CreateJobFromTemplateRequest, dict]): + request (Optional[Union[google.cloud.dataflow_v1beta3.types.CreateJobFromTemplateRequest, dict]]): The request object. A request to create a Cloud Dataflow job from a template. retry (google.api_core.retry.Retry): Designation of what errors, if any, @@ -292,10 +303,10 @@ async def sample_create_job_from_template(): async def launch_template( self, - request: Union[templates.LaunchTemplateRequest, dict] = None, + request: Optional[Union[templates.LaunchTemplateRequest, dict]] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> templates.LaunchTemplateResponse: r"""Launch a template. @@ -327,7 +338,7 @@ async def sample_launch_template(): print(response) Args: - request (Union[google.cloud.dataflow_v1beta3.types.LaunchTemplateRequest, dict]): + request (Optional[Union[google.cloud.dataflow_v1beta3.types.LaunchTemplateRequest, dict]]): The request object. A request to launch a template. retry (google.api_core.retry.Retry): Designation of what errors, if any, should be retried. @@ -376,10 +387,10 @@ async def sample_launch_template(): async def get_template( self, - request: Union[templates.GetTemplateRequest, dict] = None, + request: Optional[Union[templates.GetTemplateRequest, dict]] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> templates.GetTemplateResponse: r"""Get the template associated with a template. @@ -411,7 +422,7 @@ async def sample_get_template(): print(response) Args: - request (Union[google.cloud.dataflow_v1beta3.types.GetTemplateRequest, dict]): + request (Optional[Union[google.cloud.dataflow_v1beta3.types.GetTemplateRequest, dict]]): The request object. A request to retrieve a Cloud Dataflow job template. retry (google.api_core.retry.Retry): Designation of what errors, if any, @@ -466,14 +477,9 @@ async def __aexit__(self, exc_type, exc, tb): await self.transport.close() -try: - DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( - gapic_version=pkg_resources.get_distribution( - "google-cloud-dataflow-client", - ).version, - ) -except pkg_resources.DistributionNotFound: - DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo() +DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( + gapic_version=package_version.__version__ +) __all__ = ("TemplatesServiceAsyncClient",) diff --git a/google/cloud/dataflow_v1beta3/services/templates_service/client.py b/google/cloud/dataflow_v1beta3/services/templates_service/client.py index ef7eefc..f6e1099 100644 --- a/google/cloud/dataflow_v1beta3/services/templates_service/client.py +++ b/google/cloud/dataflow_v1beta3/services/templates_service/client.py @@ -16,7 +16,18 @@ from collections import OrderedDict import os import re -from typing import Dict, Mapping, Optional, Sequence, Tuple, Type, Union +from typing import ( + Dict, + Mapping, + MutableMapping, + MutableSequence, + Optional, + Sequence, + Tuple, + Type, + Union, + cast, +) from google.api_core import client_options as client_options_lib from google.api_core import exceptions as core_exceptions @@ -27,7 +38,8 @@ from google.auth.transport import mtls # type: ignore from google.auth.transport.grpc import SslCredentials # type: ignore from google.oauth2 import service_account # type: ignore -import pkg_resources + +from google.cloud.dataflow_v1beta3 import gapic_version as package_version try: OptionalRetry = Union[retries.Retry, gapic_v1.method._MethodDefault] @@ -62,7 +74,7 @@ class TemplatesServiceClientMeta(type): def get_transport_class( cls, - label: str = None, + label: Optional[str] = None, ) -> Type[TemplatesServiceTransport]: """Returns an appropriate transport class. @@ -317,8 +329,8 @@ def __init__( self, *, credentials: Optional[ga_credentials.Credentials] = None, - transport: Union[str, TemplatesServiceTransport, None] = None, - client_options: Optional[client_options_lib.ClientOptions] = None, + transport: Optional[Union[str, TemplatesServiceTransport]] = None, + client_options: Optional[Union[client_options_lib.ClientOptions, dict]] = None, client_info: gapic_v1.client_info.ClientInfo = DEFAULT_CLIENT_INFO, ) -> None: """Instantiates the templates service client. @@ -332,10 +344,7 @@ def __init__( transport (Union[str, TemplatesServiceTransport]): The transport to use. If set to None, a transport is chosen automatically. - NOTE: "rest" transport functionality is currently in a - beta state (preview). We welcome your feedback via an - issue in this library's source repository. - client_options (google.api_core.client_options.ClientOptions): Custom options for the + client_options (Optional[Union[google.api_core.client_options.ClientOptions, dict]]): Custom options for the client. It won't take effect if a ``transport`` instance is provided. (1) The ``api_endpoint`` property can be used to override the default endpoint provided by the client. GOOGLE_API_USE_MTLS_ENDPOINT @@ -365,6 +374,7 @@ def __init__( client_options = client_options_lib.from_dict(client_options) if client_options is None: client_options = client_options_lib.ClientOptions() + client_options = cast(client_options_lib.ClientOptions, client_options) api_endpoint, client_cert_source_func = self.get_mtls_endpoint_and_cert_source( client_options @@ -417,10 +427,10 @@ def __init__( def create_job_from_template( self, - request: Union[templates.CreateJobFromTemplateRequest, dict] = None, + request: Optional[Union[templates.CreateJobFromTemplateRequest, dict]] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> jobs.Job: r"""Creates a Cloud Dataflow job from a template. @@ -503,10 +513,10 @@ def sample_create_job_from_template(): def launch_template( self, - request: Union[templates.LaunchTemplateRequest, dict] = None, + request: Optional[Union[templates.LaunchTemplateRequest, dict]] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> templates.LaunchTemplateResponse: r"""Launch a template. @@ -588,10 +598,10 @@ def sample_launch_template(): def get_template( self, - request: Union[templates.GetTemplateRequest, dict] = None, + request: Optional[Union[templates.GetTemplateRequest, dict]] = None, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> templates.GetTemplateResponse: r"""Get the template associated with a template. @@ -686,14 +696,9 @@ def __exit__(self, type, value, traceback): self.transport.close() -try: - DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( - gapic_version=pkg_resources.get_distribution( - "google-cloud-dataflow-client", - ).version, - ) -except pkg_resources.DistributionNotFound: - DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo() +DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( + gapic_version=package_version.__version__ +) __all__ = ("TemplatesServiceClient",) diff --git a/google/cloud/dataflow_v1beta3/services/templates_service/transports/base.py b/google/cloud/dataflow_v1beta3/services/templates_service/transports/base.py index f17f492..d1d0691 100644 --- a/google/cloud/dataflow_v1beta3/services/templates_service/transports/base.py +++ b/google/cloud/dataflow_v1beta3/services/templates_service/transports/base.py @@ -23,18 +23,13 @@ import google.auth # type: ignore from google.auth import credentials as ga_credentials # type: ignore from google.oauth2 import service_account # type: ignore -import pkg_resources +from google.cloud.dataflow_v1beta3 import gapic_version as package_version from google.cloud.dataflow_v1beta3.types import jobs, templates -try: - DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( - gapic_version=pkg_resources.get_distribution( - "google-cloud-dataflow-client", - ).version, - ) -except pkg_resources.DistributionNotFound: - DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo() +DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( + gapic_version=package_version.__version__ +) class TemplatesServiceTransport(abc.ABC): @@ -53,7 +48,7 @@ def __init__( self, *, host: str = DEFAULT_HOST, - credentials: ga_credentials.Credentials = None, + credentials: Optional[ga_credentials.Credentials] = None, credentials_file: Optional[str] = None, scopes: Optional[Sequence[str]] = None, quota_project_id: Optional[str] = None, diff --git a/google/cloud/dataflow_v1beta3/services/templates_service/transports/grpc.py b/google/cloud/dataflow_v1beta3/services/templates_service/transports/grpc.py index 9d4db08..78c210c 100644 --- a/google/cloud/dataflow_v1beta3/services/templates_service/transports/grpc.py +++ b/google/cloud/dataflow_v1beta3/services/templates_service/transports/grpc.py @@ -47,14 +47,14 @@ def __init__( self, *, host: str = "dataflow.googleapis.com", - credentials: ga_credentials.Credentials = None, - credentials_file: str = None, - scopes: Sequence[str] = None, - channel: grpc.Channel = None, - api_mtls_endpoint: str = None, - client_cert_source: Callable[[], Tuple[bytes, bytes]] = None, - ssl_channel_credentials: grpc.ChannelCredentials = None, - client_cert_source_for_mtls: Callable[[], Tuple[bytes, bytes]] = None, + credentials: Optional[ga_credentials.Credentials] = None, + credentials_file: Optional[str] = None, + scopes: Optional[Sequence[str]] = None, + channel: Optional[grpc.Channel] = None, + api_mtls_endpoint: Optional[str] = None, + client_cert_source: Optional[Callable[[], Tuple[bytes, bytes]]] = None, + ssl_channel_credentials: Optional[grpc.ChannelCredentials] = None, + client_cert_source_for_mtls: Optional[Callable[[], Tuple[bytes, bytes]]] = None, quota_project_id: Optional[str] = None, client_info: gapic_v1.client_info.ClientInfo = DEFAULT_CLIENT_INFO, always_use_jwt_access: Optional[bool] = False, @@ -181,8 +181,8 @@ def __init__( def create_channel( cls, host: str = "dataflow.googleapis.com", - credentials: ga_credentials.Credentials = None, - credentials_file: str = None, + credentials: Optional[ga_credentials.Credentials] = None, + credentials_file: Optional[str] = None, scopes: Optional[Sequence[str]] = None, quota_project_id: Optional[str] = None, **kwargs, diff --git a/google/cloud/dataflow_v1beta3/services/templates_service/transports/grpc_asyncio.py b/google/cloud/dataflow_v1beta3/services/templates_service/transports/grpc_asyncio.py index 7c249a6..ad99ceb 100644 --- a/google/cloud/dataflow_v1beta3/services/templates_service/transports/grpc_asyncio.py +++ b/google/cloud/dataflow_v1beta3/services/templates_service/transports/grpc_asyncio.py @@ -49,7 +49,7 @@ class TemplatesServiceGrpcAsyncIOTransport(TemplatesServiceTransport): def create_channel( cls, host: str = "dataflow.googleapis.com", - credentials: ga_credentials.Credentials = None, + credentials: Optional[ga_credentials.Credentials] = None, credentials_file: Optional[str] = None, scopes: Optional[Sequence[str]] = None, quota_project_id: Optional[str] = None, @@ -92,15 +92,15 @@ def __init__( self, *, host: str = "dataflow.googleapis.com", - credentials: ga_credentials.Credentials = None, + credentials: Optional[ga_credentials.Credentials] = None, credentials_file: Optional[str] = None, scopes: Optional[Sequence[str]] = None, - channel: aio.Channel = None, - api_mtls_endpoint: str = None, - client_cert_source: Callable[[], Tuple[bytes, bytes]] = None, - ssl_channel_credentials: grpc.ChannelCredentials = None, - client_cert_source_for_mtls: Callable[[], Tuple[bytes, bytes]] = None, - quota_project_id=None, + channel: Optional[aio.Channel] = None, + api_mtls_endpoint: Optional[str] = None, + client_cert_source: Optional[Callable[[], Tuple[bytes, bytes]]] = None, + ssl_channel_credentials: Optional[grpc.ChannelCredentials] = None, + client_cert_source_for_mtls: Optional[Callable[[], Tuple[bytes, bytes]]] = None, + quota_project_id: Optional[str] = None, client_info: gapic_v1.client_info.ClientInfo = DEFAULT_CLIENT_INFO, always_use_jwt_access: Optional[bool] = False, api_audience: Optional[str] = None, diff --git a/google/cloud/dataflow_v1beta3/services/templates_service/transports/rest.py b/google/cloud/dataflow_v1beta3/services/templates_service/transports/rest.py index 64b68b6..a4ed0c1 100644 --- a/google/cloud/dataflow_v1beta3/services/templates_service/transports/rest.py +++ b/google/cloud/dataflow_v1beta3/services/templates_service/transports/rest.py @@ -63,26 +63,29 @@ class TemplatesServiceRestInterceptor: .. code-block:: python class MyCustomTemplatesServiceInterceptor(TemplatesServiceRestInterceptor): - def pre_create_job_from_template(request, metadata): + def pre_create_job_from_template(self, request, metadata): logging.log(f"Received request: {request}") return request, metadata - def post_create_job_from_template(response): + def post_create_job_from_template(self, response): logging.log(f"Received response: {response}") + return response - def pre_get_template(request, metadata): + def pre_get_template(self, request, metadata): logging.log(f"Received request: {request}") return request, metadata - def post_get_template(response): + def post_get_template(self, response): logging.log(f"Received response: {response}") + return response - def pre_launch_template(request, metadata): + def pre_launch_template(self, request, metadata): logging.log(f"Received request: {request}") return request, metadata - def post_launch_template(response): + def post_launch_template(self, response): logging.log(f"Received response: {response}") + return response transport = TemplatesServiceRestTransport(interceptor=MyCustomTemplatesServiceInterceptor()) client = TemplatesServiceClient(transport=transport) @@ -175,19 +178,16 @@ class TemplatesServiceRestTransport(TemplatesServiceTransport): It sends JSON representations of protocol buffers over HTTP/1.1 - NOTE: This REST transport functionality is currently in a beta - state (preview). We welcome your feedback via an issue in this - library's source repository. Thank you! """ def __init__( self, *, host: str = "dataflow.googleapis.com", - credentials: ga_credentials.Credentials = None, - credentials_file: str = None, - scopes: Sequence[str] = None, - client_cert_source_for_mtls: Callable[[], Tuple[bytes, bytes]] = None, + credentials: Optional[ga_credentials.Credentials] = None, + credentials_file: Optional[str] = None, + scopes: Optional[Sequence[str]] = None, + client_cert_source_for_mtls: Optional[Callable[[], Tuple[bytes, bytes]]] = None, quota_project_id: Optional[str] = None, client_info: gapic_v1.client_info.ClientInfo = DEFAULT_CLIENT_INFO, always_use_jwt_access: Optional[bool] = False, @@ -197,39 +197,35 @@ def __init__( ) -> None: """Instantiate the transport. - NOTE: This REST transport functionality is currently in a beta - state (preview). We welcome your feedback via a GitHub issue in - this library's repository. Thank you! - - Args: - host (Optional[str]): - The hostname to connect to. - credentials (Optional[google.auth.credentials.Credentials]): The - authorization credentials to attach to requests. These - credentials identify the application to the service; if none - are specified, the client will attempt to ascertain the - credentials from the environment. - - credentials_file (Optional[str]): A file with credentials that can - be loaded with :func:`google.auth.load_credentials_from_file`. - This argument is ignored if ``channel`` is provided. - scopes (Optional(Sequence[str])): A list of scopes. This argument is - ignored if ``channel`` is provided. - client_cert_source_for_mtls (Callable[[], Tuple[bytes, bytes]]): Client - certificate to configure mutual TLS HTTP channel. It is ignored - if ``channel`` is provided. - quota_project_id (Optional[str]): An optional project to use for billing - and quota. - client_info (google.api_core.gapic_v1.client_info.ClientInfo): - The client info used to send a user-agent string along with - API requests. If ``None``, then default info will be used. - Generally, you only need to set this if you are developing - your own client library. - always_use_jwt_access (Optional[bool]): Whether self signed JWT should - be used for service account credentials. - url_scheme: the protocol scheme for the API endpoint. Normally - "https", but for testing or local servers, - "http" can be specified. + Args: + host (Optional[str]): + The hostname to connect to. + credentials (Optional[google.auth.credentials.Credentials]): The + authorization credentials to attach to requests. These + credentials identify the application to the service; if none + are specified, the client will attempt to ascertain the + credentials from the environment. + + credentials_file (Optional[str]): A file with credentials that can + be loaded with :func:`google.auth.load_credentials_from_file`. + This argument is ignored if ``channel`` is provided. + scopes (Optional(Sequence[str])): A list of scopes. This argument is + ignored if ``channel`` is provided. + client_cert_source_for_mtls (Callable[[], Tuple[bytes, bytes]]): Client + certificate to configure mutual TLS HTTP channel. It is ignored + if ``channel`` is provided. + quota_project_id (Optional[str]): An optional project to use for billing + and quota. + client_info (google.api_core.gapic_v1.client_info.ClientInfo): + The client info used to send a user-agent string along with + API requests. If ``None``, then default info will be used. + Generally, you only need to set this if you are developing + your own client library. + always_use_jwt_access (Optional[bool]): Whether self signed JWT should + be used for service account credentials. + url_scheme: the protocol scheme for the API endpoint. Normally + "https", but for testing or local servers, + "http" can be specified. """ # Run the base constructor # TODO(yon-mg): resolve other ctor params i.e. scopes, quota, etc. @@ -269,7 +265,7 @@ def __call__( request: templates.CreateJobFromTemplateRequest, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> jobs.Job: r"""Call the create job from template method over HTTP. @@ -315,7 +311,7 @@ def __call__( body = json_format.MessageToJson( transcoded_request["body"], including_default_value_fields=False, - use_integers_for_enums=False, + use_integers_for_enums=True, ) uri = transcoded_request["uri"] method = transcoded_request["method"] @@ -325,10 +321,12 @@ def __call__( json_format.MessageToJson( transcoded_request["query_params"], including_default_value_fields=False, - use_integers_for_enums=False, + use_integers_for_enums=True, ) ) + query_params["$alt"] = "json;enum-encoding=int" + # Send the request headers = dict(metadata) headers["Content-Type"] = "application/json" @@ -362,7 +360,7 @@ def __call__( request: templates.GetTemplateRequest, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> templates.GetTemplateResponse: r"""Call the get template method over HTTP. @@ -407,10 +405,12 @@ def __call__( json_format.MessageToJson( transcoded_request["query_params"], including_default_value_fields=False, - use_integers_for_enums=False, + use_integers_for_enums=True, ) ) + query_params["$alt"] = "json;enum-encoding=int" + # Send the request headers = dict(metadata) headers["Content-Type"] = "application/json" @@ -443,7 +443,7 @@ def __call__( request: templates.LaunchTemplateRequest, *, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> templates.LaunchTemplateResponse: r"""Call the launch template method over HTTP. @@ -485,7 +485,7 @@ def __call__( body = json_format.MessageToJson( transcoded_request["body"], including_default_value_fields=False, - use_integers_for_enums=False, + use_integers_for_enums=True, ) uri = transcoded_request["uri"] method = transcoded_request["method"] @@ -495,10 +495,12 @@ def __call__( json_format.MessageToJson( transcoded_request["query_params"], including_default_value_fields=False, - use_integers_for_enums=False, + use_integers_for_enums=True, ) ) + query_params["$alt"] = "json;enum-encoding=int" + # Send the request headers = dict(metadata) headers["Content-Type"] = "application/json" diff --git a/google/cloud/dataflow_v1beta3/types/environment.py b/google/cloud/dataflow_v1beta3/types/environment.py index 339d232..60d03cb 100644 --- a/google/cloud/dataflow_v1beta3/types/environment.py +++ b/google/cloud/dataflow_v1beta3/types/environment.py @@ -13,6 +13,8 @@ # See the License for the specific language governing permissions and # limitations under the License. # +from typing import MutableMapping, MutableSequence + from google.protobuf import any_pb2 # type: ignore from google.protobuf import struct_pb2 # type: ignore import proto # type: ignore @@ -136,12 +138,12 @@ class Environment(proto.Message): to choose a reasonable default. This should be in the form of the API service name, e.g. "compute.googleapis.com". - experiments (Sequence[str]): + experiments (MutableSequence[str]): The list of experiments to enable. This field should be used for SDK related experiments and not for service related experiments. The proper field for service related experiments is service_options. - service_options (Sequence[str]): + service_options (MutableSequence[str]): The list of service options to enable. This field should be used for service related experiments only. These experiments, when @@ -155,7 +157,7 @@ class Environment(proto.Message): Format: projects/PROJECT_ID/locations/LOCATION/keyRings/KEY_RING/cryptoKeys/KEY - worker_pools (Sequence[google.cloud.dataflow_v1beta3.types.WorkerPool]): + worker_pools (MutableSequence[google.cloud.dataflow_v1beta3.types.WorkerPool]): The worker pools. At least one "harness" worker pool must be specified in order for the job to have workers. @@ -211,78 +213,78 @@ class Environment(proto.Message): job. """ - temp_storage_prefix = proto.Field( + temp_storage_prefix: str = proto.Field( proto.STRING, number=1, ) - cluster_manager_api_service = proto.Field( + cluster_manager_api_service: str = proto.Field( proto.STRING, number=2, ) - experiments = proto.RepeatedField( + experiments: MutableSequence[str] = proto.RepeatedField( proto.STRING, number=3, ) - service_options = proto.RepeatedField( + service_options: MutableSequence[str] = proto.RepeatedField( proto.STRING, number=16, ) - service_kms_key_name = proto.Field( + service_kms_key_name: str = proto.Field( proto.STRING, number=12, ) - worker_pools = proto.RepeatedField( + worker_pools: MutableSequence["WorkerPool"] = proto.RepeatedField( proto.MESSAGE, number=4, message="WorkerPool", ) - user_agent = proto.Field( + user_agent: struct_pb2.Struct = proto.Field( proto.MESSAGE, number=5, message=struct_pb2.Struct, ) - version = proto.Field( + version: struct_pb2.Struct = proto.Field( proto.MESSAGE, number=6, message=struct_pb2.Struct, ) - dataset = proto.Field( + dataset: str = proto.Field( proto.STRING, number=7, ) - sdk_pipeline_options = proto.Field( + sdk_pipeline_options: struct_pb2.Struct = proto.Field( proto.MESSAGE, number=8, message=struct_pb2.Struct, ) - internal_experiments = proto.Field( + internal_experiments: any_pb2.Any = proto.Field( proto.MESSAGE, number=9, message=any_pb2.Any, ) - service_account_email = proto.Field( + service_account_email: str = proto.Field( proto.STRING, number=10, ) - flex_resource_scheduling_goal = proto.Field( + flex_resource_scheduling_goal: "FlexResourceSchedulingGoal" = proto.Field( proto.ENUM, number=11, enum="FlexResourceSchedulingGoal", ) - worker_region = proto.Field( + worker_region: str = proto.Field( proto.STRING, number=13, ) - worker_zone = proto.Field( + worker_zone: str = proto.Field( proto.STRING, number=14, ) - shuffle_mode = proto.Field( + shuffle_mode: "ShuffleMode" = proto.Field( proto.ENUM, number=15, enum="ShuffleMode", ) - debug_options = proto.Field( + debug_options: "DebugOptions" = proto.Field( proto.MESSAGE, number=17, message="DebugOptions", @@ -312,11 +314,11 @@ class Package(proto.Message): bucket.storage.googleapis.com/ """ - name = proto.Field( + name: str = proto.Field( proto.STRING, number=1, ) - location = proto.Field( + location: str = proto.Field( proto.STRING, number=2, ) @@ -358,15 +360,15 @@ class Disk(proto.Message): Directory in a VM where disk is mounted. """ - size_gb = proto.Field( + size_gb: int = proto.Field( proto.INT32, number=1, ) - disk_type = proto.Field( + disk_type: str = proto.Field( proto.STRING, number=2, ) - mount_point = proto.Field( + mount_point: str = proto.Field( proto.STRING, number=3, ) @@ -411,27 +413,27 @@ class WorkerSettings(proto.Message): bucket.storage.googleapis.com/{object} """ - base_url = proto.Field( + base_url: str = proto.Field( proto.STRING, number=1, ) - reporting_enabled = proto.Field( + reporting_enabled: bool = proto.Field( proto.BOOL, number=2, ) - service_path = proto.Field( + service_path: str = proto.Field( proto.STRING, number=3, ) - shuffle_service_path = proto.Field( + shuffle_service_path: str = proto.Field( proto.STRING, number=4, ) - worker_id = proto.Field( + worker_id: str = proto.Field( proto.STRING, number=5, ) - temp_storage_prefix = proto.Field( + temp_storage_prefix: str = proto.Field( proto.STRING, number=6, ) @@ -447,7 +449,7 @@ class TaskRunnerSettings(proto.Message): task_group (str): The UNIX group ID on the worker VM to use for tasks launched by taskrunner; e.g. "wheel". - oauth_scopes (Sequence[str]): + oauth_scopes (MutableSequence[str]): The OAuth2 scopes to be requested by the taskrunner in order to access the Cloud Dataflow API. @@ -514,80 +516,80 @@ class TaskRunnerSettings(proto.Message): The streaming worker main class name. """ - task_user = proto.Field( + task_user: str = proto.Field( proto.STRING, number=1, ) - task_group = proto.Field( + task_group: str = proto.Field( proto.STRING, number=2, ) - oauth_scopes = proto.RepeatedField( + oauth_scopes: MutableSequence[str] = proto.RepeatedField( proto.STRING, number=3, ) - base_url = proto.Field( + base_url: str = proto.Field( proto.STRING, number=4, ) - dataflow_api_version = proto.Field( + dataflow_api_version: str = proto.Field( proto.STRING, number=5, ) - parallel_worker_settings = proto.Field( + parallel_worker_settings: "WorkerSettings" = proto.Field( proto.MESSAGE, number=6, message="WorkerSettings", ) - base_task_dir = proto.Field( + base_task_dir: str = proto.Field( proto.STRING, number=7, ) - continue_on_exception = proto.Field( + continue_on_exception: bool = proto.Field( proto.BOOL, number=8, ) - log_to_serialconsole = proto.Field( + log_to_serialconsole: bool = proto.Field( proto.BOOL, number=9, ) - alsologtostderr = proto.Field( + alsologtostderr: bool = proto.Field( proto.BOOL, number=10, ) - log_upload_location = proto.Field( + log_upload_location: str = proto.Field( proto.STRING, number=11, ) - log_dir = proto.Field( + log_dir: str = proto.Field( proto.STRING, number=12, ) - temp_storage_prefix = proto.Field( + temp_storage_prefix: str = proto.Field( proto.STRING, number=13, ) - harness_command = proto.Field( + harness_command: str = proto.Field( proto.STRING, number=14, ) - workflow_file_name = proto.Field( + workflow_file_name: str = proto.Field( proto.STRING, number=15, ) - commandlines_file_name = proto.Field( + commandlines_file_name: str = proto.Field( proto.STRING, number=16, ) - vm_id = proto.Field( + vm_id: str = proto.Field( proto.STRING, number=17, ) - language_hint = proto.Field( + language_hint: str = proto.Field( proto.STRING, number=18, ) - streaming_worker_main_class = proto.Field( + streaming_worker_main_class: str = proto.Field( proto.STRING, number=19, ) @@ -604,12 +606,12 @@ class AutoscalingSettings(proto.Message): at. """ - algorithm = proto.Field( + algorithm: "AutoscalingAlgorithm" = proto.Field( proto.ENUM, number=1, enum="AutoscalingAlgorithm", ) - max_num_workers = proto.Field( + max_num_workers: int = proto.Field( proto.INT32, number=2, ) @@ -635,25 +637,25 @@ class SdkHarnessContainerImage(proto.Message): Environment ID for the Beam runner API proto Environment that corresponds to the current SDK Harness. - capabilities (Sequence[str]): + capabilities (MutableSequence[str]): The set of capabilities enumerated in the above Environment proto. See also https://github.com/apache/beam/blob/master/model/pipeline/src/main/proto/beam_runner_api.proto """ - container_image = proto.Field( + container_image: str = proto.Field( proto.STRING, number=1, ) - use_single_core_per_container = proto.Field( + use_single_core_per_container: bool = proto.Field( proto.BOOL, number=2, ) - environment_id = proto.Field( + environment_id: str = proto.Field( proto.STRING, number=3, ) - capabilities = proto.RepeatedField( + capabilities: MutableSequence[str] = proto.RepeatedField( proto.STRING, number=4, ) @@ -675,7 +677,7 @@ class WorkerPool(proto.Message): this pool needed to execute the job. If zero or unspecified, the service will attempt to choose a reasonable default. - packages (Sequence[google.cloud.dataflow_v1beta3.types.Package]): + packages (MutableSequence[google.cloud.dataflow_v1beta3.types.Package]): Packages to be installed on workers. default_package_set (google.cloud.dataflow_v1beta3.types.DefaultPackageSet): The default package set to install. This @@ -726,10 +728,10 @@ class WorkerPool(proto.Message): on_host_maintenance (str): The action to take on host maintenance, as defined by the Google Compute Engine API. - data_disks (Sequence[google.cloud.dataflow_v1beta3.types.Disk]): + data_disks (MutableSequence[google.cloud.dataflow_v1beta3.types.Disk]): Data disks that are used by a VM in this workflow. - metadata (Mapping[str, str]): + metadata (MutableMapping[str, str]): Metadata to set on the Google Compute Engine VMs. autoscaling_settings (google.cloud.dataflow_v1beta3.types.AutoscalingSettings): @@ -759,7 +761,7 @@ class WorkerPool(proto.Message): 1 by convention for streaming). ip_configuration (google.cloud.dataflow_v1beta3.types.WorkerIPAddressConfiguration): Configuration for VM IPs. - sdk_harness_container_images (Sequence[google.cloud.dataflow_v1beta3.types.SdkHarnessContainerImage]): + sdk_harness_container_images (MutableSequence[google.cloud.dataflow_v1beta3.types.SdkHarnessContainerImage]): Set of SDK harness containers needed to execute this pipeline. This will only be set in the Fn API path. For non-cross-language @@ -768,100 +770,102 @@ class WorkerPool(proto.Message): entries. """ - kind = proto.Field( + kind: str = proto.Field( proto.STRING, number=1, ) - num_workers = proto.Field( + num_workers: int = proto.Field( proto.INT32, number=2, ) - packages = proto.RepeatedField( + packages: MutableSequence["Package"] = proto.RepeatedField( proto.MESSAGE, number=3, message="Package", ) - default_package_set = proto.Field( + default_package_set: "DefaultPackageSet" = proto.Field( proto.ENUM, number=4, enum="DefaultPackageSet", ) - machine_type = proto.Field( + machine_type: str = proto.Field( proto.STRING, number=5, ) - teardown_policy = proto.Field( + teardown_policy: "TeardownPolicy" = proto.Field( proto.ENUM, number=6, enum="TeardownPolicy", ) - disk_size_gb = proto.Field( + disk_size_gb: int = proto.Field( proto.INT32, number=7, ) - disk_type = proto.Field( + disk_type: str = proto.Field( proto.STRING, number=16, ) - disk_source_image = proto.Field( + disk_source_image: str = proto.Field( proto.STRING, number=8, ) - zone = proto.Field( + zone: str = proto.Field( proto.STRING, number=9, ) - taskrunner_settings = proto.Field( + taskrunner_settings: "TaskRunnerSettings" = proto.Field( proto.MESSAGE, number=10, message="TaskRunnerSettings", ) - on_host_maintenance = proto.Field( + on_host_maintenance: str = proto.Field( proto.STRING, number=11, ) - data_disks = proto.RepeatedField( + data_disks: MutableSequence["Disk"] = proto.RepeatedField( proto.MESSAGE, number=12, message="Disk", ) - metadata = proto.MapField( + metadata: MutableMapping[str, str] = proto.MapField( proto.STRING, proto.STRING, number=13, ) - autoscaling_settings = proto.Field( + autoscaling_settings: "AutoscalingSettings" = proto.Field( proto.MESSAGE, number=14, message="AutoscalingSettings", ) - pool_args = proto.Field( + pool_args: any_pb2.Any = proto.Field( proto.MESSAGE, number=15, message=any_pb2.Any, ) - network = proto.Field( + network: str = proto.Field( proto.STRING, number=17, ) - subnetwork = proto.Field( + subnetwork: str = proto.Field( proto.STRING, number=19, ) - worker_harness_container_image = proto.Field( + worker_harness_container_image: str = proto.Field( proto.STRING, number=18, ) - num_threads_per_worker = proto.Field( + num_threads_per_worker: int = proto.Field( proto.INT32, number=20, ) - ip_configuration = proto.Field( + ip_configuration: "WorkerIPAddressConfiguration" = proto.Field( proto.ENUM, number=21, enum="WorkerIPAddressConfiguration", ) - sdk_harness_container_images = proto.RepeatedField( + sdk_harness_container_images: MutableSequence[ + "SdkHarnessContainerImage" + ] = proto.RepeatedField( proto.MESSAGE, number=22, message="SdkHarnessContainerImage", @@ -878,7 +882,7 @@ class DebugOptions(proto.Message): hot key to the user's Cloud Logging. """ - enable_hot_key_logging = proto.Field( + enable_hot_key_logging: bool = proto.Field( proto.BOOL, number=1, ) diff --git a/google/cloud/dataflow_v1beta3/types/jobs.py b/google/cloud/dataflow_v1beta3/types/jobs.py index 07385c3..0d2e538 100644 --- a/google/cloud/dataflow_v1beta3/types/jobs.py +++ b/google/cloud/dataflow_v1beta3/types/jobs.py @@ -13,6 +13,8 @@ # See the License for the specific language governing permissions and # limitations under the License. # +from typing import MutableMapping, MutableSequence + from google.protobuf import duration_pb2 # type: ignore from google.protobuf import struct_pb2 # type: ignore from google.protobuf import timestamp_pb2 # type: ignore @@ -124,7 +126,7 @@ class Job(proto.Message): The type of Cloud Dataflow job. environment (google.cloud.dataflow_v1beta3.types.Environment): The environment for the job. - steps (Sequence[google.cloud.dataflow_v1beta3.types.Step]): + steps (MutableSequence[google.cloud.dataflow_v1beta3.types.Step]): Exactly one of step or steps_location should be specified. The top-level steps that constitute the entire job. Only @@ -170,7 +172,7 @@ class Job(proto.Message): When sending a ``CreateJobRequest``, you can update a job by specifying it here. The job named here is stopped, and its intermediate state is transferred to this job. - transform_name_mapping (Mapping[str, str]): + transform_name_mapping (MutableMapping[str, str]): The map of transform name prefixes of the job to be replaced to the corresponding name prefixes of the new job. @@ -189,7 +191,7 @@ class Job(proto.Message): If another job is an update of this job (and thus, this job is in ``JOB_STATE_UPDATED``), this field contains the ID of that job. - temp_files (Sequence[str]): + temp_files (MutableSequence[str]): A set of files the system should be aware of that are used for temporary storage. These temporary files will be removed on job @@ -203,7 +205,7 @@ class Job(proto.Message): storage.googleapis.com/{bucket}/{object} bucket.storage.googleapis.com/{object} - labels (Mapping[str, str]): + labels (MutableMapping[str, str]): User-defined labels for this job. The labels map can contain no more than 64 entries. Entries @@ -225,7 +227,7 @@ class Job(proto.Message): time. A description of the user pipeline and stages through which it is executed. Created by Cloud Dataflow service. Only retrieved with JOB_VIEW_DESCRIPTION or JOB_VIEW_ALL. - stage_states (Sequence[google.cloud.dataflow_v1beta3.types.ExecutionStageState]): + stage_states (MutableSequence[google.cloud.dataflow_v1beta3.types.ExecutionStageState]): This field may be mutated by the Cloud Dataflow service; callers cannot mutate it. job_metadata (google.cloud.dataflow_v1beta3.types.JobMetadata): @@ -251,117 +253,117 @@ class Job(proto.Message): if it is set in any requests. """ - id = proto.Field( + id: str = proto.Field( proto.STRING, number=1, ) - project_id = proto.Field( + project_id: str = proto.Field( proto.STRING, number=2, ) - name = proto.Field( + name: str = proto.Field( proto.STRING, number=3, ) - type_ = proto.Field( + type_: gd_environment.JobType = proto.Field( proto.ENUM, number=4, enum=gd_environment.JobType, ) - environment = proto.Field( + environment: gd_environment.Environment = proto.Field( proto.MESSAGE, number=5, message=gd_environment.Environment, ) - steps = proto.RepeatedField( + steps: MutableSequence["Step"] = proto.RepeatedField( proto.MESSAGE, number=6, message="Step", ) - steps_location = proto.Field( + steps_location: str = proto.Field( proto.STRING, number=24, ) - current_state = proto.Field( + current_state: "JobState" = proto.Field( proto.ENUM, number=7, enum="JobState", ) - current_state_time = proto.Field( + current_state_time: timestamp_pb2.Timestamp = proto.Field( proto.MESSAGE, number=8, message=timestamp_pb2.Timestamp, ) - requested_state = proto.Field( + requested_state: "JobState" = proto.Field( proto.ENUM, number=9, enum="JobState", ) - execution_info = proto.Field( + execution_info: "JobExecutionInfo" = proto.Field( proto.MESSAGE, number=10, message="JobExecutionInfo", ) - create_time = proto.Field( + create_time: timestamp_pb2.Timestamp = proto.Field( proto.MESSAGE, number=11, message=timestamp_pb2.Timestamp, ) - replace_job_id = proto.Field( + replace_job_id: str = proto.Field( proto.STRING, number=12, ) - transform_name_mapping = proto.MapField( + transform_name_mapping: MutableMapping[str, str] = proto.MapField( proto.STRING, proto.STRING, number=13, ) - client_request_id = proto.Field( + client_request_id: str = proto.Field( proto.STRING, number=14, ) - replaced_by_job_id = proto.Field( + replaced_by_job_id: str = proto.Field( proto.STRING, number=15, ) - temp_files = proto.RepeatedField( + temp_files: MutableSequence[str] = proto.RepeatedField( proto.STRING, number=16, ) - labels = proto.MapField( + labels: MutableMapping[str, str] = proto.MapField( proto.STRING, proto.STRING, number=17, ) - location = proto.Field( + location: str = proto.Field( proto.STRING, number=18, ) - pipeline_description = proto.Field( + pipeline_description: "PipelineDescription" = proto.Field( proto.MESSAGE, number=19, message="PipelineDescription", ) - stage_states = proto.RepeatedField( + stage_states: MutableSequence["ExecutionStageState"] = proto.RepeatedField( proto.MESSAGE, number=20, message="ExecutionStageState", ) - job_metadata = proto.Field( + job_metadata: "JobMetadata" = proto.Field( proto.MESSAGE, number=21, message="JobMetadata", ) - start_time = proto.Field( + start_time: timestamp_pb2.Timestamp = proto.Field( proto.MESSAGE, number=22, message=timestamp_pb2.Timestamp, ) - created_from_snapshot_id = proto.Field( + created_from_snapshot_id: str = proto.Field( proto.STRING, number=23, ) - satisfies_pzs = proto.Field( + satisfies_pzs: bool = proto.Field( proto.BOOL, number=25, ) @@ -377,11 +379,11 @@ class DatastoreIODetails(proto.Message): ProjectId accessed in the connection. """ - namespace = proto.Field( + namespace: str = proto.Field( proto.STRING, number=1, ) - project_id = proto.Field( + project_id: str = proto.Field( proto.STRING, number=2, ) @@ -397,11 +399,11 @@ class PubSubIODetails(proto.Message): Subscription used in the connection. """ - topic = proto.Field( + topic: str = proto.Field( proto.STRING, number=1, ) - subscription = proto.Field( + subscription: str = proto.Field( proto.STRING, number=2, ) @@ -416,7 +418,7 @@ class FileIODetails(proto.Message): connector. """ - file_pattern = proto.Field( + file_pattern: str = proto.Field( proto.STRING, number=1, ) @@ -434,15 +436,15 @@ class BigTableIODetails(proto.Message): TableId accessed in the connection. """ - project_id = proto.Field( + project_id: str = proto.Field( proto.STRING, number=1, ) - instance_id = proto.Field( + instance_id: str = proto.Field( proto.STRING, number=2, ) - table_id = proto.Field( + table_id: str = proto.Field( proto.STRING, number=3, ) @@ -462,19 +464,19 @@ class BigQueryIODetails(proto.Message): Query used to access data in the connection. """ - table = proto.Field( + table: str = proto.Field( proto.STRING, number=1, ) - dataset = proto.Field( + dataset: str = proto.Field( proto.STRING, number=2, ) - project_id = proto.Field( + project_id: str = proto.Field( proto.STRING, number=3, ) - query = proto.Field( + query: str = proto.Field( proto.STRING, number=4, ) @@ -492,15 +494,15 @@ class SpannerIODetails(proto.Message): DatabaseId accessed in the connection. """ - project_id = proto.Field( + project_id: str = proto.Field( proto.STRING, number=1, ) - instance_id = proto.Field( + instance_id: str = proto.Field( proto.STRING, number=2, ) - database_id = proto.Field( + database_id: str = proto.Field( proto.STRING, number=3, ) @@ -527,15 +529,15 @@ class SdkSupportStatus(proto.Enum): DEPRECATED = 3 UNSUPPORTED = 4 - version = proto.Field( + version: str = proto.Field( proto.STRING, number=1, ) - version_display_name = proto.Field( + version_display_name: str = proto.Field( proto.STRING, number=2, ) - sdk_support_status = proto.Field( + sdk_support_status: SdkSupportStatus = proto.Field( proto.ENUM, number=3, enum=SdkSupportStatus, @@ -549,57 +551,57 @@ class JobMetadata(proto.Message): Attributes: sdk_version (google.cloud.dataflow_v1beta3.types.SdkVersion): The SDK version used to run the job. - spanner_details (Sequence[google.cloud.dataflow_v1beta3.types.SpannerIODetails]): + spanner_details (MutableSequence[google.cloud.dataflow_v1beta3.types.SpannerIODetails]): Identification of a Spanner source used in the Dataflow job. - bigquery_details (Sequence[google.cloud.dataflow_v1beta3.types.BigQueryIODetails]): + bigquery_details (MutableSequence[google.cloud.dataflow_v1beta3.types.BigQueryIODetails]): Identification of a BigQuery source used in the Dataflow job. - big_table_details (Sequence[google.cloud.dataflow_v1beta3.types.BigTableIODetails]): + big_table_details (MutableSequence[google.cloud.dataflow_v1beta3.types.BigTableIODetails]): Identification of a Cloud Bigtable source used in the Dataflow job. - pubsub_details (Sequence[google.cloud.dataflow_v1beta3.types.PubSubIODetails]): + pubsub_details (MutableSequence[google.cloud.dataflow_v1beta3.types.PubSubIODetails]): Identification of a Pub/Sub source used in the Dataflow job. - file_details (Sequence[google.cloud.dataflow_v1beta3.types.FileIODetails]): + file_details (MutableSequence[google.cloud.dataflow_v1beta3.types.FileIODetails]): Identification of a File source used in the Dataflow job. - datastore_details (Sequence[google.cloud.dataflow_v1beta3.types.DatastoreIODetails]): + datastore_details (MutableSequence[google.cloud.dataflow_v1beta3.types.DatastoreIODetails]): Identification of a Datastore source used in the Dataflow job. """ - sdk_version = proto.Field( + sdk_version: "SdkVersion" = proto.Field( proto.MESSAGE, number=1, message="SdkVersion", ) - spanner_details = proto.RepeatedField( + spanner_details: MutableSequence["SpannerIODetails"] = proto.RepeatedField( proto.MESSAGE, number=2, message="SpannerIODetails", ) - bigquery_details = proto.RepeatedField( + bigquery_details: MutableSequence["BigQueryIODetails"] = proto.RepeatedField( proto.MESSAGE, number=3, message="BigQueryIODetails", ) - big_table_details = proto.RepeatedField( + big_table_details: MutableSequence["BigTableIODetails"] = proto.RepeatedField( proto.MESSAGE, number=4, message="BigTableIODetails", ) - pubsub_details = proto.RepeatedField( + pubsub_details: MutableSequence["PubSubIODetails"] = proto.RepeatedField( proto.MESSAGE, number=5, message="PubSubIODetails", ) - file_details = proto.RepeatedField( + file_details: MutableSequence["FileIODetails"] = proto.RepeatedField( proto.MESSAGE, number=6, message="FileIODetails", ) - datastore_details = proto.RepeatedField( + datastore_details: MutableSequence["DatastoreIODetails"] = proto.RepeatedField( proto.MESSAGE, number=7, message="DatastoreIODetails", @@ -621,16 +623,16 @@ class ExecutionStageState(proto.Message): this state. """ - execution_stage_name = proto.Field( + execution_stage_name: str = proto.Field( proto.STRING, number=1, ) - execution_stage_state = proto.Field( + execution_stage_state: "JobState" = proto.Field( proto.ENUM, number=2, enum="JobState", ) - current_state_time = proto.Field( + current_state_time: timestamp_pb2.Timestamp = proto.Field( proto.MESSAGE, number=3, message=timestamp_pb2.Timestamp, @@ -644,27 +646,31 @@ class PipelineDescription(proto.Message): Dataflow provided metrics. Attributes: - original_pipeline_transform (Sequence[google.cloud.dataflow_v1beta3.types.TransformSummary]): + original_pipeline_transform (MutableSequence[google.cloud.dataflow_v1beta3.types.TransformSummary]): Description of each transform in the pipeline and collections between them. - execution_pipeline_stage (Sequence[google.cloud.dataflow_v1beta3.types.ExecutionStageSummary]): + execution_pipeline_stage (MutableSequence[google.cloud.dataflow_v1beta3.types.ExecutionStageSummary]): Description of each stage of execution of the pipeline. - display_data (Sequence[google.cloud.dataflow_v1beta3.types.DisplayData]): + display_data (MutableSequence[google.cloud.dataflow_v1beta3.types.DisplayData]): Pipeline level display data. """ - original_pipeline_transform = proto.RepeatedField( + original_pipeline_transform: MutableSequence[ + "TransformSummary" + ] = proto.RepeatedField( proto.MESSAGE, number=1, message="TransformSummary", ) - execution_pipeline_stage = proto.RepeatedField( + execution_pipeline_stage: MutableSequence[ + "ExecutionStageSummary" + ] = proto.RepeatedField( proto.MESSAGE, number=2, message="ExecutionStageSummary", ) - display_data = proto.RepeatedField( + display_data: MutableSequence["DisplayData"] = proto.RepeatedField( proto.MESSAGE, number=3, message="DisplayData", @@ -683,39 +689,39 @@ class TransformSummary(proto.Message): name (str): User provided name for this transform instance. - display_data (Sequence[google.cloud.dataflow_v1beta3.types.DisplayData]): + display_data (MutableSequence[google.cloud.dataflow_v1beta3.types.DisplayData]): Transform-specific display data. - output_collection_name (Sequence[str]): + output_collection_name (MutableSequence[str]): User names for all collection outputs to this transform. - input_collection_name (Sequence[str]): + input_collection_name (MutableSequence[str]): User names for all collection inputs to this transform. """ - kind = proto.Field( + kind: "KindType" = proto.Field( proto.ENUM, number=1, enum="KindType", ) - id = proto.Field( + id: str = proto.Field( proto.STRING, number=2, ) - name = proto.Field( + name: str = proto.Field( proto.STRING, number=3, ) - display_data = proto.RepeatedField( + display_data: MutableSequence["DisplayData"] = proto.RepeatedField( proto.MESSAGE, number=4, message="DisplayData", ) - output_collection_name = proto.RepeatedField( + output_collection_name: MutableSequence[str] = proto.RepeatedField( proto.STRING, number=5, ) - input_collection_name = proto.RepeatedField( + input_collection_name: MutableSequence[str] = proto.RepeatedField( proto.STRING, number=6, ) @@ -735,17 +741,17 @@ class ExecutionStageSummary(proto.Message): Dataflow service generated id for this stage. kind (google.cloud.dataflow_v1beta3.types.KindType): Type of transform this stage is executing. - input_source (Sequence[google.cloud.dataflow_v1beta3.types.ExecutionStageSummary.StageSource]): + input_source (MutableSequence[google.cloud.dataflow_v1beta3.types.ExecutionStageSummary.StageSource]): Input sources for this stage. - output_source (Sequence[google.cloud.dataflow_v1beta3.types.ExecutionStageSummary.StageSource]): + output_source (MutableSequence[google.cloud.dataflow_v1beta3.types.ExecutionStageSummary.StageSource]): Output sources for this stage. - prerequisite_stage (Sequence[str]): + prerequisite_stage (MutableSequence[str]): Other stages that must complete before this stage can run. - component_transform (Sequence[google.cloud.dataflow_v1beta3.types.ExecutionStageSummary.ComponentTransform]): + component_transform (MutableSequence[google.cloud.dataflow_v1beta3.types.ExecutionStageSummary.ComponentTransform]): Transforms that comprise this execution stage. - component_source (Sequence[google.cloud.dataflow_v1beta3.types.ExecutionStageSummary.ComponentSource]): + component_source (MutableSequence[google.cloud.dataflow_v1beta3.types.ExecutionStageSummary.ComponentSource]): Collections produced and consumed by component transforms of this stage. """ @@ -768,19 +774,19 @@ class StageSource(proto.Message): Size of the source, if measurable. """ - user_name = proto.Field( + user_name: str = proto.Field( proto.STRING, number=1, ) - name = proto.Field( + name: str = proto.Field( proto.STRING, number=2, ) - original_transform_or_collection = proto.Field( + original_transform_or_collection: str = proto.Field( proto.STRING, number=3, ) - size_bytes = proto.Field( + size_bytes: int = proto.Field( proto.INT64, number=4, ) @@ -802,15 +808,15 @@ class ComponentTransform(proto.Message): associated. """ - user_name = proto.Field( + user_name: str = proto.Field( proto.STRING, number=1, ) - name = proto.Field( + name: str = proto.Field( proto.STRING, number=2, ) - original_transform = proto.Field( + original_transform: str = proto.Field( proto.STRING, number=3, ) @@ -832,52 +838,52 @@ class ComponentSource(proto.Message): closely associated. """ - user_name = proto.Field( + user_name: str = proto.Field( proto.STRING, number=1, ) - name = proto.Field( + name: str = proto.Field( proto.STRING, number=2, ) - original_transform_or_collection = proto.Field( + original_transform_or_collection: str = proto.Field( proto.STRING, number=3, ) - name = proto.Field( + name: str = proto.Field( proto.STRING, number=1, ) - id = proto.Field( + id: str = proto.Field( proto.STRING, number=2, ) - kind = proto.Field( + kind: "KindType" = proto.Field( proto.ENUM, number=3, enum="KindType", ) - input_source = proto.RepeatedField( + input_source: MutableSequence[StageSource] = proto.RepeatedField( proto.MESSAGE, number=4, message=StageSource, ) - output_source = proto.RepeatedField( + output_source: MutableSequence[StageSource] = proto.RepeatedField( proto.MESSAGE, number=5, message=StageSource, ) - prerequisite_stage = proto.RepeatedField( + prerequisite_stage: MutableSequence[str] = proto.RepeatedField( proto.STRING, number=8, ) - component_transform = proto.RepeatedField( + component_transform: MutableSequence[ComponentTransform] = proto.RepeatedField( proto.MESSAGE, number=6, message=ComponentTransform, ) - component_source = proto.RepeatedField( + component_source: MutableSequence[ComponentSource] = proto.RepeatedField( proto.MESSAGE, number=7, message=ComponentSource, @@ -954,60 +960,60 @@ class name or programming language namespace the element. """ - key = proto.Field( + key: str = proto.Field( proto.STRING, number=1, ) - namespace = proto.Field( + namespace: str = proto.Field( proto.STRING, number=2, ) - str_value = proto.Field( + str_value: str = proto.Field( proto.STRING, number=4, oneof="Value", ) - int64_value = proto.Field( + int64_value: int = proto.Field( proto.INT64, number=5, oneof="Value", ) - float_value = proto.Field( + float_value: float = proto.Field( proto.FLOAT, number=6, oneof="Value", ) - java_class_value = proto.Field( + java_class_value: str = proto.Field( proto.STRING, number=7, oneof="Value", ) - timestamp_value = proto.Field( + timestamp_value: timestamp_pb2.Timestamp = proto.Field( proto.MESSAGE, number=8, oneof="Value", message=timestamp_pb2.Timestamp, ) - duration_value = proto.Field( + duration_value: duration_pb2.Duration = proto.Field( proto.MESSAGE, number=9, oneof="Value", message=duration_pb2.Duration, ) - bool_value = proto.Field( + bool_value: bool = proto.Field( proto.BOOL, number=10, oneof="Value", ) - short_str_value = proto.Field( + short_str_value: str = proto.Field( proto.STRING, number=11, ) - url = proto.Field( + url: str = proto.Field( proto.STRING, number=12, ) - label = proto.Field( + label: str = proto.Field( proto.STRING, number=13, ) @@ -1053,15 +1059,15 @@ class Step(proto.Message): be provided on Create. Only retrieved with JOB_VIEW_ALL. """ - kind = proto.Field( + kind: str = proto.Field( proto.STRING, number=1, ) - name = proto.Field( + name: str = proto.Field( proto.STRING, number=2, ) - properties = proto.Field( + properties: struct_pb2.Struct = proto.Field( proto.MESSAGE, number=3, message=struct_pb2.Struct, @@ -1073,12 +1079,12 @@ class JobExecutionInfo(proto.Message): executed that isn't contained in the submitted job. Attributes: - stages (Mapping[str, google.cloud.dataflow_v1beta3.types.JobExecutionStageInfo]): + stages (MutableMapping[str, google.cloud.dataflow_v1beta3.types.JobExecutionStageInfo]): A mapping from each stage to the information about that stage. """ - stages = proto.MapField( + stages: MutableMapping[str, "JobExecutionStageInfo"] = proto.MapField( proto.STRING, proto.MESSAGE, number=1, @@ -1092,14 +1098,14 @@ class JobExecutionStageInfo(proto.Message): executed. Attributes: - step_name (Sequence[str]): + step_name (MutableSequence[str]): The steps associated with the execution stage. Note that stages may have several steps, and that a given step might be run by more than one stage. """ - step_name = proto.RepeatedField( + step_name: MutableSequence[str] = proto.RepeatedField( proto.STRING, number=1, ) @@ -1126,25 +1132,25 @@ class CreateJobRequest(proto.Message): that contains this job. """ - project_id = proto.Field( + project_id: str = proto.Field( proto.STRING, number=1, ) - job = proto.Field( + job: "Job" = proto.Field( proto.MESSAGE, number=2, message="Job", ) - view = proto.Field( + view: "JobView" = proto.Field( proto.ENUM, number=3, enum="JobView", ) - replace_job_id = proto.Field( + replace_job_id: str = proto.Field( proto.STRING, number=4, ) - location = proto.Field( + location: str = proto.Field( proto.STRING, number=5, ) @@ -1168,20 +1174,20 @@ class GetJobRequest(proto.Message): that contains this job. """ - project_id = proto.Field( + project_id: str = proto.Field( proto.STRING, number=1, ) - job_id = proto.Field( + job_id: str = proto.Field( proto.STRING, number=2, ) - view = proto.Field( + view: "JobView" = proto.Field( proto.ENUM, number=3, enum="JobView", ) - location = proto.Field( + location: str = proto.Field( proto.STRING, number=4, ) @@ -1206,20 +1212,20 @@ class UpdateJobRequest(proto.Message): that contains this job. """ - project_id = proto.Field( + project_id: str = proto.Field( proto.STRING, number=1, ) - job_id = proto.Field( + job_id: str = proto.Field( proto.STRING, number=2, ) - job = proto.Field( + job: "Job" = proto.Field( proto.MESSAGE, number=3, message="Job", ) - location = proto.Field( + location: str = proto.Field( proto.STRING, number=4, ) @@ -1259,29 +1265,29 @@ class Filter(proto.Enum): TERMINATED = 2 ACTIVE = 3 - filter = proto.Field( + filter: Filter = proto.Field( proto.ENUM, number=5, enum=Filter, ) - project_id = proto.Field( + project_id: str = proto.Field( proto.STRING, number=1, ) - view = proto.Field( + view: "JobView" = proto.Field( proto.ENUM, number=2, enum="JobView", ) - page_size = proto.Field( + page_size: int = proto.Field( proto.INT32, number=3, ) - page_token = proto.Field( + page_token: str = proto.Field( proto.STRING, number=4, ) - location = proto.Field( + location: str = proto.Field( proto.STRING, number=17, ) @@ -1299,7 +1305,7 @@ class FailedLocation(proto.Message): that failed to respond. """ - name = proto.Field( + name: str = proto.Field( proto.STRING, number=1, ) @@ -1313,12 +1319,12 @@ class ListJobsResponse(proto.Message): and the requests's response body is empty {}. Attributes: - jobs (Sequence[google.cloud.dataflow_v1beta3.types.Job]): + jobs (MutableSequence[google.cloud.dataflow_v1beta3.types.Job]): A subset of the requested job information. next_page_token (str): Set if there may be more results than fit in this response. - failed_location (Sequence[google.cloud.dataflow_v1beta3.types.FailedLocation]): + failed_location (MutableSequence[google.cloud.dataflow_v1beta3.types.FailedLocation]): Zero or more messages describing the [regional endpoints] (https://cloud.google.com/dataflow/docs/concepts/regional-endpoints) that failed to respond. @@ -1328,16 +1334,16 @@ class ListJobsResponse(proto.Message): def raw_page(self): return self - jobs = proto.RepeatedField( + jobs: MutableSequence["Job"] = proto.RepeatedField( proto.MESSAGE, number=1, message="Job", ) - next_page_token = proto.Field( + next_page_token: str = proto.Field( proto.STRING, number=2, ) - failed_location = proto.RepeatedField( + failed_location: MutableSequence["FailedLocation"] = proto.RepeatedField( proto.MESSAGE, number=3, message="FailedLocation", @@ -1365,28 +1371,28 @@ class SnapshotJobRequest(proto.Message): Maybe empty. """ - project_id = proto.Field( + project_id: str = proto.Field( proto.STRING, number=1, ) - job_id = proto.Field( + job_id: str = proto.Field( proto.STRING, number=2, ) - ttl = proto.Field( + ttl: duration_pb2.Duration = proto.Field( proto.MESSAGE, number=3, message=duration_pb2.Duration, ) - location = proto.Field( + location: str = proto.Field( proto.STRING, number=4, ) - snapshot_sources = proto.Field( + snapshot_sources: bool = proto.Field( proto.BOOL, number=5, ) - description = proto.Field( + description: str = proto.Field( proto.STRING, number=6, ) @@ -1400,7 +1406,7 @@ class CheckActiveJobsRequest(proto.Message): The project which owns the jobs. """ - project_id = proto.Field( + project_id: str = proto.Field( proto.STRING, number=1, ) @@ -1415,7 +1421,7 @@ class CheckActiveJobsResponse(proto.Message): False otherwise. """ - active_jobs_exist = proto.Field( + active_jobs_exist: bool = proto.Field( proto.BOOL, number=1, ) diff --git a/google/cloud/dataflow_v1beta3/types/messages.py b/google/cloud/dataflow_v1beta3/types/messages.py index b46a5d0..e3da083 100644 --- a/google/cloud/dataflow_v1beta3/types/messages.py +++ b/google/cloud/dataflow_v1beta3/types/messages.py @@ -13,6 +13,8 @@ # See the License for the specific language governing permissions and # limitations under the License. # +from typing import MutableMapping, MutableSequence + from google.protobuf import struct_pb2 # type: ignore from google.protobuf import timestamp_pb2 # type: ignore import proto # type: ignore @@ -54,20 +56,20 @@ class JobMessage(proto.Message): Importance level of the message. """ - id = proto.Field( + id: str = proto.Field( proto.STRING, number=1, ) - time = proto.Field( + time: timestamp_pb2.Timestamp = proto.Field( proto.MESSAGE, number=2, message=timestamp_pb2.Timestamp, ) - message_text = proto.Field( + message_text: str = proto.Field( proto.STRING, number=3, ) - message_importance = proto.Field( + message_importance: "JobMessageImportance" = proto.Field( proto.ENUM, number=4, enum="JobMessageImportance", @@ -86,7 +88,7 @@ class StructuredMessage(proto.Message): Identifier for this message type. Used by external systems to internationalize or personalize message. - parameters (Sequence[google.cloud.dataflow_v1beta3.types.StructuredMessage.Parameter]): + parameters (MutableSequence[google.cloud.dataflow_v1beta3.types.StructuredMessage.Parameter]): The structured data associated with this message. """ @@ -101,25 +103,25 @@ class Parameter(proto.Message): Value for this parameter. """ - key = proto.Field( + key: str = proto.Field( proto.STRING, number=1, ) - value = proto.Field( + value: struct_pb2.Value = proto.Field( proto.MESSAGE, number=2, message=struct_pb2.Value, ) - message_text = proto.Field( + message_text: str = proto.Field( proto.STRING, number=1, ) - message_key = proto.Field( + message_key: str = proto.Field( proto.STRING, number=2, ) - parameters = proto.RepeatedField( + parameters: MutableSequence[Parameter] = proto.RepeatedField( proto.MESSAGE, number=3, message=Parameter, @@ -159,30 +161,30 @@ class AutoscalingEventType(proto.Enum): ACTUATION_FAILURE = 3 NO_CHANGE = 4 - current_num_workers = proto.Field( + current_num_workers: int = proto.Field( proto.INT64, number=1, ) - target_num_workers = proto.Field( + target_num_workers: int = proto.Field( proto.INT64, number=2, ) - event_type = proto.Field( + event_type: AutoscalingEventType = proto.Field( proto.ENUM, number=3, enum=AutoscalingEventType, ) - description = proto.Field( + description: "StructuredMessage" = proto.Field( proto.MESSAGE, number=4, message="StructuredMessage", ) - time = proto.Field( + time: timestamp_pb2.Timestamp = proto.Field( proto.MESSAGE, number=5, message=timestamp_pb2.Timestamp, ) - worker_pool = proto.Field( + worker_pool: str = proto.Field( proto.STRING, number=7, ) @@ -225,38 +227,38 @@ class ListJobMessagesRequest(proto.Message): that contains the job specified by job_id. """ - project_id = proto.Field( + project_id: str = proto.Field( proto.STRING, number=1, ) - job_id = proto.Field( + job_id: str = proto.Field( proto.STRING, number=2, ) - minimum_importance = proto.Field( + minimum_importance: "JobMessageImportance" = proto.Field( proto.ENUM, number=3, enum="JobMessageImportance", ) - page_size = proto.Field( + page_size: int = proto.Field( proto.INT32, number=4, ) - page_token = proto.Field( + page_token: str = proto.Field( proto.STRING, number=5, ) - start_time = proto.Field( + start_time: timestamp_pb2.Timestamp = proto.Field( proto.MESSAGE, number=6, message=timestamp_pb2.Timestamp, ) - end_time = proto.Field( + end_time: timestamp_pb2.Timestamp = proto.Field( proto.MESSAGE, number=7, message=timestamp_pb2.Timestamp, ) - location = proto.Field( + location: str = proto.Field( proto.STRING, number=8, ) @@ -266,12 +268,12 @@ class ListJobMessagesResponse(proto.Message): r"""Response to a request to list job messages. Attributes: - job_messages (Sequence[google.cloud.dataflow_v1beta3.types.JobMessage]): + job_messages (MutableSequence[google.cloud.dataflow_v1beta3.types.JobMessage]): Messages in ascending timestamp order. next_page_token (str): The token to obtain the next page of results if there are more. - autoscaling_events (Sequence[google.cloud.dataflow_v1beta3.types.AutoscalingEvent]): + autoscaling_events (MutableSequence[google.cloud.dataflow_v1beta3.types.AutoscalingEvent]): Autoscaling events in ascending timestamp order. """ @@ -280,16 +282,16 @@ class ListJobMessagesResponse(proto.Message): def raw_page(self): return self - job_messages = proto.RepeatedField( + job_messages: MutableSequence["JobMessage"] = proto.RepeatedField( proto.MESSAGE, number=1, message="JobMessage", ) - next_page_token = proto.Field( + next_page_token: str = proto.Field( proto.STRING, number=2, ) - autoscaling_events = proto.RepeatedField( + autoscaling_events: MutableSequence["AutoscalingEvent"] = proto.RepeatedField( proto.MESSAGE, number=3, message="AutoscalingEvent", diff --git a/google/cloud/dataflow_v1beta3/types/metrics.py b/google/cloud/dataflow_v1beta3/types/metrics.py index ef1aa20..c52f0ad 100644 --- a/google/cloud/dataflow_v1beta3/types/metrics.py +++ b/google/cloud/dataflow_v1beta3/types/metrics.py @@ -13,6 +13,8 @@ # See the License for the specific language governing permissions and # limitations under the License. # +from typing import MutableMapping, MutableSequence + from google.protobuf import struct_pb2 # type: ignore from google.protobuf import timestamp_pb2 # type: ignore import proto # type: ignore @@ -59,7 +61,7 @@ class MetricStructuredName(proto.Message): service or SDK. name (str): Worker-defined metric name. - context (Mapping[str, str]): + context (MutableMapping[str, str]): Zero or more labeled fields which identify the part of the job this metric is associated with, such as the name of a step or collection. @@ -69,15 +71,15 @@ class MetricStructuredName(proto.Message): PCollections in the SDK will have context['pcollection'] = . """ - origin = proto.Field( + origin: str = proto.Field( proto.STRING, number=1, ) - name = proto.Field( + name: str = proto.Field( proto.STRING, number=2, ) - context = proto.MapField( + context: MutableMapping[str, str] = proto.MapField( proto.STRING, proto.STRING, number=3, @@ -146,55 +148,55 @@ class MetricUpdate(proto.Message): the metrics API. """ - name = proto.Field( + name: "MetricStructuredName" = proto.Field( proto.MESSAGE, number=1, message="MetricStructuredName", ) - kind = proto.Field( + kind: str = proto.Field( proto.STRING, number=2, ) - cumulative = proto.Field( + cumulative: bool = proto.Field( proto.BOOL, number=3, ) - scalar = proto.Field( + scalar: struct_pb2.Value = proto.Field( proto.MESSAGE, number=4, message=struct_pb2.Value, ) - mean_sum = proto.Field( + mean_sum: struct_pb2.Value = proto.Field( proto.MESSAGE, number=5, message=struct_pb2.Value, ) - mean_count = proto.Field( + mean_count: struct_pb2.Value = proto.Field( proto.MESSAGE, number=6, message=struct_pb2.Value, ) - set_ = proto.Field( + set_: struct_pb2.Value = proto.Field( proto.MESSAGE, number=7, message=struct_pb2.Value, ) - distribution = proto.Field( + distribution: struct_pb2.Value = proto.Field( proto.MESSAGE, number=11, message=struct_pb2.Value, ) - gauge = proto.Field( + gauge: struct_pb2.Value = proto.Field( proto.MESSAGE, number=12, message=struct_pb2.Value, ) - internal = proto.Field( + internal: struct_pb2.Value = proto.Field( proto.MESSAGE, number=8, message=struct_pb2.Value, ) - update_time = proto.Field( + update_time: timestamp_pb2.Timestamp = proto.Field( proto.MESSAGE, number=9, message=timestamp_pb2.Timestamp, @@ -219,20 +221,20 @@ class GetJobMetricsRequest(proto.Message): that contains the job specified by job_id. """ - project_id = proto.Field( + project_id: str = proto.Field( proto.STRING, number=1, ) - job_id = proto.Field( + job_id: str = proto.Field( proto.STRING, number=2, ) - start_time = proto.Field( + start_time: timestamp_pb2.Timestamp = proto.Field( proto.MESSAGE, number=3, message=timestamp_pb2.Timestamp, ) - location = proto.Field( + location: str = proto.Field( proto.STRING, number=4, ) @@ -251,16 +253,16 @@ class JobMetrics(proto.Message): metric_time (google.protobuf.timestamp_pb2.Timestamp): Timestamp as of which metric values are current. - metrics (Sequence[google.cloud.dataflow_v1beta3.types.MetricUpdate]): + metrics (MutableSequence[google.cloud.dataflow_v1beta3.types.MetricUpdate]): All metrics for this job. """ - metric_time = proto.Field( + metric_time: timestamp_pb2.Timestamp = proto.Field( proto.MESSAGE, number=1, message=timestamp_pb2.Timestamp, ) - metrics = proto.RepeatedField( + metrics: MutableSequence["MetricUpdate"] = proto.RepeatedField( proto.MESSAGE, number=2, message="MetricUpdate", @@ -291,23 +293,23 @@ class GetJobExecutionDetailsRequest(proto.Message): of results to be returned. """ - project_id = proto.Field( + project_id: str = proto.Field( proto.STRING, number=1, ) - job_id = proto.Field( + job_id: str = proto.Field( proto.STRING, number=2, ) - location = proto.Field( + location: str = proto.Field( proto.STRING, number=3, ) - page_size = proto.Field( + page_size: int = proto.Field( proto.INT32, number=4, ) - page_token = proto.Field( + page_token: str = proto.Field( proto.STRING, number=5, ) @@ -320,7 +322,7 @@ class ProgressTimeseries(proto.Message): Attributes: current_progress (float): The current progress of the component, in the range [0,1]. - data_points (Sequence[google.cloud.dataflow_v1beta3.types.ProgressTimeseries.Point]): + data_points (MutableSequence[google.cloud.dataflow_v1beta3.types.ProgressTimeseries.Point]): History of progress for the component. Points are sorted by time. """ @@ -335,21 +337,21 @@ class Point(proto.Message): The value of the point. """ - time = proto.Field( + time: timestamp_pb2.Timestamp = proto.Field( proto.MESSAGE, number=1, message=timestamp_pb2.Timestamp, ) - value = proto.Field( + value: float = proto.Field( proto.DOUBLE, number=2, ) - current_progress = proto.Field( + current_progress: float = proto.Field( proto.DOUBLE, number=1, ) - data_points = proto.RepeatedField( + data_points: MutableSequence[Point] = proto.RepeatedField( proto.MESSAGE, number=2, message=Point, @@ -374,35 +376,35 @@ class StageSummary(proto.Message): progress (google.cloud.dataflow_v1beta3.types.ProgressTimeseries): Progress for this stage. Only applicable to Batch jobs. - metrics (Sequence[google.cloud.dataflow_v1beta3.types.MetricUpdate]): + metrics (MutableSequence[google.cloud.dataflow_v1beta3.types.MetricUpdate]): Metrics for this stage. """ - stage_id = proto.Field( + stage_id: str = proto.Field( proto.STRING, number=1, ) - state = proto.Field( + state: "ExecutionState" = proto.Field( proto.ENUM, number=2, enum="ExecutionState", ) - start_time = proto.Field( + start_time: timestamp_pb2.Timestamp = proto.Field( proto.MESSAGE, number=3, message=timestamp_pb2.Timestamp, ) - end_time = proto.Field( + end_time: timestamp_pb2.Timestamp = proto.Field( proto.MESSAGE, number=4, message=timestamp_pb2.Timestamp, ) - progress = proto.Field( + progress: "ProgressTimeseries" = proto.Field( proto.MESSAGE, number=5, message="ProgressTimeseries", ) - metrics = proto.RepeatedField( + metrics: MutableSequence["MetricUpdate"] = proto.RepeatedField( proto.MESSAGE, number=6, message="MetricUpdate", @@ -413,7 +415,7 @@ class JobExecutionDetails(proto.Message): r"""Information about the execution of a job. Attributes: - stages (Sequence[google.cloud.dataflow_v1beta3.types.StageSummary]): + stages (MutableSequence[google.cloud.dataflow_v1beta3.types.StageSummary]): The stages of the job execution. next_page_token (str): If present, this response does not contain all requested @@ -425,12 +427,12 @@ class JobExecutionDetails(proto.Message): def raw_page(self): return self - stages = proto.RepeatedField( + stages: MutableSequence["StageSummary"] = proto.RepeatedField( proto.MESSAGE, number=1, message="StageSummary", ) - next_page_token = proto.Field( + next_page_token: str = proto.Field( proto.STRING, number=2, ) @@ -469,36 +471,36 @@ class GetStageExecutionDetailsRequest(proto.Message): start time. """ - project_id = proto.Field( + project_id: str = proto.Field( proto.STRING, number=1, ) - job_id = proto.Field( + job_id: str = proto.Field( proto.STRING, number=2, ) - location = proto.Field( + location: str = proto.Field( proto.STRING, number=3, ) - stage_id = proto.Field( + stage_id: str = proto.Field( proto.STRING, number=4, ) - page_size = proto.Field( + page_size: int = proto.Field( proto.INT32, number=5, ) - page_token = proto.Field( + page_token: str = proto.Field( proto.STRING, number=6, ) - start_time = proto.Field( + start_time: timestamp_pb2.Timestamp = proto.Field( proto.MESSAGE, number=7, message=timestamp_pb2.Timestamp, ) - end_time = proto.Field( + end_time: timestamp_pb2.Timestamp = proto.Field( proto.MESSAGE, number=8, message=timestamp_pb2.Timestamp, @@ -524,39 +526,39 @@ class WorkItemDetails(proto.Message): State of this work item. progress (google.cloud.dataflow_v1beta3.types.ProgressTimeseries): Progress of this work item. - metrics (Sequence[google.cloud.dataflow_v1beta3.types.MetricUpdate]): + metrics (MutableSequence[google.cloud.dataflow_v1beta3.types.MetricUpdate]): Metrics for this work item. """ - task_id = proto.Field( + task_id: str = proto.Field( proto.STRING, number=1, ) - attempt_id = proto.Field( + attempt_id: str = proto.Field( proto.STRING, number=2, ) - start_time = proto.Field( + start_time: timestamp_pb2.Timestamp = proto.Field( proto.MESSAGE, number=3, message=timestamp_pb2.Timestamp, ) - end_time = proto.Field( + end_time: timestamp_pb2.Timestamp = proto.Field( proto.MESSAGE, number=4, message=timestamp_pb2.Timestamp, ) - state = proto.Field( + state: "ExecutionState" = proto.Field( proto.ENUM, number=5, enum="ExecutionState", ) - progress = proto.Field( + progress: "ProgressTimeseries" = proto.Field( proto.MESSAGE, number=6, message="ProgressTimeseries", ) - metrics = proto.RepeatedField( + metrics: MutableSequence["MetricUpdate"] = proto.RepeatedField( proto.MESSAGE, number=7, message="MetricUpdate", @@ -569,16 +571,16 @@ class WorkerDetails(proto.Message): Attributes: worker_name (str): Name of this worker - work_items (Sequence[google.cloud.dataflow_v1beta3.types.WorkItemDetails]): + work_items (MutableSequence[google.cloud.dataflow_v1beta3.types.WorkItemDetails]): Work items processed by this worker, sorted by time. """ - worker_name = proto.Field( + worker_name: str = proto.Field( proto.STRING, number=1, ) - work_items = proto.RepeatedField( + work_items: MutableSequence["WorkItemDetails"] = proto.RepeatedField( proto.MESSAGE, number=2, message="WorkItemDetails", @@ -589,7 +591,7 @@ class StageExecutionDetails(proto.Message): r"""Information about the workers and work items within a stage. Attributes: - workers (Sequence[google.cloud.dataflow_v1beta3.types.WorkerDetails]): + workers (MutableSequence[google.cloud.dataflow_v1beta3.types.WorkerDetails]): Workers that have done work on the stage. next_page_token (str): If present, this response does not contain all requested @@ -601,12 +603,12 @@ class StageExecutionDetails(proto.Message): def raw_page(self): return self - workers = proto.RepeatedField( + workers: MutableSequence["WorkerDetails"] = proto.RepeatedField( proto.MESSAGE, number=1, message="WorkerDetails", ) - next_page_token = proto.Field( + next_page_token: str = proto.Field( proto.STRING, number=2, ) diff --git a/google/cloud/dataflow_v1beta3/types/snapshots.py b/google/cloud/dataflow_v1beta3/types/snapshots.py index 8f01fef..926e44b 100644 --- a/google/cloud/dataflow_v1beta3/types/snapshots.py +++ b/google/cloud/dataflow_v1beta3/types/snapshots.py @@ -13,6 +13,8 @@ # See the License for the specific language governing permissions and # limitations under the License. # +from typing import MutableMapping, MutableSequence + from google.protobuf import duration_pb2 # type: ignore from google.protobuf import timestamp_pb2 # type: ignore import proto # type: ignore @@ -54,15 +56,15 @@ class PubsubSnapshotMetadata(proto.Message): The expire time of the Pubsub snapshot. """ - topic_name = proto.Field( + topic_name: str = proto.Field( proto.STRING, number=1, ) - snapshot_name = proto.Field( + snapshot_name: str = proto.Field( proto.STRING, number=2, ) - expire_time = proto.Field( + expire_time: timestamp_pb2.Timestamp = proto.Field( proto.MESSAGE, number=3, message=timestamp_pb2.Timestamp, @@ -86,7 +88,7 @@ class Snapshot(proto.Message): automatically deleted. state (google.cloud.dataflow_v1beta3.types.SnapshotState): State of the snapshot. - pubsub_metadata (Sequence[google.cloud.dataflow_v1beta3.types.PubsubSnapshotMetadata]): + pubsub_metadata (MutableSequence[google.cloud.dataflow_v1beta3.types.PubsubSnapshotMetadata]): Pub/Sub snapshot metadata. description (str): User specified description of the snapshot. @@ -99,47 +101,47 @@ class Snapshot(proto.Message): e.g., "us-central1". """ - id = proto.Field( + id: str = proto.Field( proto.STRING, number=1, ) - project_id = proto.Field( + project_id: str = proto.Field( proto.STRING, number=2, ) - source_job_id = proto.Field( + source_job_id: str = proto.Field( proto.STRING, number=3, ) - creation_time = proto.Field( + creation_time: timestamp_pb2.Timestamp = proto.Field( proto.MESSAGE, number=4, message=timestamp_pb2.Timestamp, ) - ttl = proto.Field( + ttl: duration_pb2.Duration = proto.Field( proto.MESSAGE, number=5, message=duration_pb2.Duration, ) - state = proto.Field( + state: "SnapshotState" = proto.Field( proto.ENUM, number=6, enum="SnapshotState", ) - pubsub_metadata = proto.RepeatedField( + pubsub_metadata: MutableSequence["PubsubSnapshotMetadata"] = proto.RepeatedField( proto.MESSAGE, number=7, message="PubsubSnapshotMetadata", ) - description = proto.Field( + description: str = proto.Field( proto.STRING, number=8, ) - disk_size_bytes = proto.Field( + disk_size_bytes: int = proto.Field( proto.INT64, number=9, ) - region = proto.Field( + region: str = proto.Field( proto.STRING, number=10, ) @@ -158,15 +160,15 @@ class GetSnapshotRequest(proto.Message): The location that contains this snapshot. """ - project_id = proto.Field( + project_id: str = proto.Field( proto.STRING, number=1, ) - snapshot_id = proto.Field( + snapshot_id: str = proto.Field( proto.STRING, number=2, ) - location = proto.Field( + location: str = proto.Field( proto.STRING, number=3, ) @@ -185,15 +187,15 @@ class DeleteSnapshotRequest(proto.Message): The location that contains this snapshot. """ - project_id = proto.Field( + project_id: str = proto.Field( proto.STRING, number=1, ) - snapshot_id = proto.Field( + snapshot_id: str = proto.Field( proto.STRING, number=2, ) - location = proto.Field( + location: str = proto.Field( proto.STRING, number=3, ) @@ -216,15 +218,15 @@ class ListSnapshotsRequest(proto.Message): The location to list snapshots in. """ - project_id = proto.Field( + project_id: str = proto.Field( proto.STRING, number=1, ) - job_id = proto.Field( + job_id: str = proto.Field( proto.STRING, number=3, ) - location = proto.Field( + location: str = proto.Field( proto.STRING, number=2, ) @@ -234,11 +236,11 @@ class ListSnapshotsResponse(proto.Message): r"""List of snapshots. Attributes: - snapshots (Sequence[google.cloud.dataflow_v1beta3.types.Snapshot]): + snapshots (MutableSequence[google.cloud.dataflow_v1beta3.types.Snapshot]): Returned snapshots. """ - snapshots = proto.RepeatedField( + snapshots: MutableSequence["Snapshot"] = proto.RepeatedField( proto.MESSAGE, number=1, message="Snapshot", diff --git a/google/cloud/dataflow_v1beta3/types/streaming.py b/google/cloud/dataflow_v1beta3/types/streaming.py index 8c2244b..f698d77 100644 --- a/google/cloud/dataflow_v1beta3/types/streaming.py +++ b/google/cloud/dataflow_v1beta3/types/streaming.py @@ -13,6 +13,8 @@ # See the License for the specific language governing permissions and # limitations under the License. # +from typing import MutableMapping, MutableSequence + import proto # type: ignore __protobuf__ = proto.module( @@ -41,13 +43,13 @@ class TopologyConfig(proto.Message): computations and their sharded locations. Attributes: - computations (Sequence[google.cloud.dataflow_v1beta3.types.ComputationTopology]): + computations (MutableSequence[google.cloud.dataflow_v1beta3.types.ComputationTopology]): The computations associated with a streaming Dataflow job. - data_disk_assignments (Sequence[google.cloud.dataflow_v1beta3.types.DataDiskAssignment]): + data_disk_assignments (MutableSequence[google.cloud.dataflow_v1beta3.types.DataDiskAssignment]): The disks assigned to a streaming Dataflow job. - user_stage_to_computation_name_map (Mapping[str, str]): + user_stage_to_computation_name_map (MutableMapping[str, str]): Maps user stage names to stable computation names. forwarding_key_bits (int): @@ -57,26 +59,26 @@ class TopologyConfig(proto.Message): Version number for persistent state. """ - computations = proto.RepeatedField( + computations: MutableSequence["ComputationTopology"] = proto.RepeatedField( proto.MESSAGE, number=1, message="ComputationTopology", ) - data_disk_assignments = proto.RepeatedField( + data_disk_assignments: MutableSequence["DataDiskAssignment"] = proto.RepeatedField( proto.MESSAGE, number=2, message="DataDiskAssignment", ) - user_stage_to_computation_name_map = proto.MapField( + user_stage_to_computation_name_map: MutableMapping[str, str] = proto.MapField( proto.STRING, proto.STRING, number=3, ) - forwarding_key_bits = proto.Field( + forwarding_key_bits: int = proto.Field( proto.INT32, number=4, ) - persistent_state_version = proto.Field( + persistent_state_version: int = proto.Field( proto.INT32, number=5, ) @@ -113,31 +115,31 @@ class PubsubLocation(proto.Message): pubsub attributes. """ - topic = proto.Field( + topic: str = proto.Field( proto.STRING, number=1, ) - subscription = proto.Field( + subscription: str = proto.Field( proto.STRING, number=2, ) - timestamp_label = proto.Field( + timestamp_label: str = proto.Field( proto.STRING, number=3, ) - id_label = proto.Field( + id_label: str = proto.Field( proto.STRING, number=4, ) - drop_late_data = proto.Field( + drop_late_data: bool = proto.Field( proto.BOOL, number=5, ) - tracking_subscription = proto.Field( + tracking_subscription: str = proto.Field( proto.STRING, number=6, ) - with_attributes = proto.Field( + with_attributes: bool = proto.Field( proto.BOOL, number=7, ) @@ -153,7 +155,7 @@ class StreamingStageLocation(proto.Message): streaming Dataflow job. """ - stream_id = proto.Field( + stream_id: str = proto.Field( proto.STRING, number=1, ) @@ -171,11 +173,11 @@ class StreamingSideInputLocation(proto.Message): input is stored. """ - tag = proto.Field( + tag: str = proto.Field( proto.STRING, number=1, ) - state_family = proto.Field( + state_family: str = proto.Field( proto.STRING, number=2, ) @@ -189,7 +191,7 @@ class CustomSourceLocation(proto.Message): Whether this source is stateful. """ - stateful = proto.Field( + stateful: bool = proto.Field( proto.BOOL, number=1, ) @@ -226,25 +228,25 @@ class StreamLocation(proto.Message): This field is a member of `oneof`_ ``location``. """ - streaming_stage_location = proto.Field( + streaming_stage_location: "StreamingStageLocation" = proto.Field( proto.MESSAGE, number=1, oneof="location", message="StreamingStageLocation", ) - pubsub_location = proto.Field( + pubsub_location: "PubsubLocation" = proto.Field( proto.MESSAGE, number=2, oneof="location", message="PubsubLocation", ) - side_input_location = proto.Field( + side_input_location: "StreamingSideInputLocation" = proto.Field( proto.MESSAGE, number=3, oneof="location", message="StreamingSideInputLocation", ) - custom_source_location = proto.Field( + custom_source_location: "CustomSourceLocation" = proto.Field( proto.MESSAGE, number=4, oneof="location", @@ -263,11 +265,11 @@ class StateFamilyConfig(proto.Message): operation. """ - state_family = proto.Field( + state_family: str = proto.Field( proto.STRING, number=1, ) - is_read = proto.Field( + is_read: bool = proto.Field( proto.BOOL, number=2, ) @@ -281,40 +283,40 @@ class ComputationTopology(proto.Message): The system stage name. computation_id (str): The ID of the computation. - key_ranges (Sequence[google.cloud.dataflow_v1beta3.types.KeyRangeLocation]): + key_ranges (MutableSequence[google.cloud.dataflow_v1beta3.types.KeyRangeLocation]): The key ranges processed by the computation. - inputs (Sequence[google.cloud.dataflow_v1beta3.types.StreamLocation]): + inputs (MutableSequence[google.cloud.dataflow_v1beta3.types.StreamLocation]): The inputs to the computation. - outputs (Sequence[google.cloud.dataflow_v1beta3.types.StreamLocation]): + outputs (MutableSequence[google.cloud.dataflow_v1beta3.types.StreamLocation]): The outputs from the computation. - state_families (Sequence[google.cloud.dataflow_v1beta3.types.StateFamilyConfig]): + state_families (MutableSequence[google.cloud.dataflow_v1beta3.types.StateFamilyConfig]): The state family values. """ - system_stage_name = proto.Field( + system_stage_name: str = proto.Field( proto.STRING, number=1, ) - computation_id = proto.Field( + computation_id: str = proto.Field( proto.STRING, number=5, ) - key_ranges = proto.RepeatedField( + key_ranges: MutableSequence["KeyRangeLocation"] = proto.RepeatedField( proto.MESSAGE, number=2, message="KeyRangeLocation", ) - inputs = proto.RepeatedField( + inputs: MutableSequence["StreamLocation"] = proto.RepeatedField( proto.MESSAGE, number=3, message="StreamLocation", ) - outputs = proto.RepeatedField( + outputs: MutableSequence["StreamLocation"] = proto.RepeatedField( proto.MESSAGE, number=4, message="StreamLocation", ) - state_families = proto.RepeatedField( + state_families: MutableSequence["StateFamilyConfig"] = proto.RepeatedField( proto.MESSAGE, number=7, message="StateFamilyConfig", @@ -348,23 +350,23 @@ class KeyRangeLocation(proto.Message): in the worker local filesystem. """ - start = proto.Field( + start: str = proto.Field( proto.STRING, number=1, ) - end = proto.Field( + end: str = proto.Field( proto.STRING, number=2, ) - delivery_endpoint = proto.Field( + delivery_endpoint: str = proto.Field( proto.STRING, number=3, ) - data_disk = proto.Field( + data_disk: str = proto.Field( proto.STRING, number=5, ) - deprecated_persistent_directory = proto.Field( + deprecated_persistent_directory: str = proto.Field( proto.STRING, number=4, ) @@ -382,7 +384,7 @@ class MountedDataDisk(proto.Message): "myproject-1014-104817-4c2-harness-0-disk-1". """ - data_disk = proto.Field( + data_disk: str = proto.Field( proto.STRING, number=1, ) @@ -396,7 +398,7 @@ class DataDiskAssignment(proto.Message): VM instance name the data disks mounted to, for example "myproject-1014-104817-4c2-harness-0". - data_disks (Sequence[str]): + data_disks (MutableSequence[str]): Mounted data disks. The order is important a data disk's 0-based index in this list defines which persistent directory the disk is mounted @@ -406,11 +408,11 @@ class DataDiskAssignment(proto.Message): }. """ - vm_instance = proto.Field( + vm_instance: str = proto.Field( proto.STRING, number=1, ) - data_disks = proto.RepeatedField( + data_disks: MutableSequence[str] = proto.RepeatedField( proto.STRING, number=2, ) @@ -436,15 +438,15 @@ class KeyRangeDataDiskAssignment(proto.Message): "myproject-1014-104817-4c2-harness-0-disk-1". """ - start = proto.Field( + start: str = proto.Field( proto.STRING, number=1, ) - end = proto.Field( + end: str = proto.Field( proto.STRING, number=2, ) - data_disk = proto.Field( + data_disk: str = proto.Field( proto.STRING, number=3, ) @@ -457,16 +459,18 @@ class StreamingComputationRanges(proto.Message): Attributes: computation_id (str): The ID of the computation. - range_assignments (Sequence[google.cloud.dataflow_v1beta3.types.KeyRangeDataDiskAssignment]): + range_assignments (MutableSequence[google.cloud.dataflow_v1beta3.types.KeyRangeDataDiskAssignment]): Data disk assignments for ranges from this computation. """ - computation_id = proto.Field( + computation_id: str = proto.Field( proto.STRING, number=1, ) - range_assignments = proto.RepeatedField( + range_assignments: MutableSequence[ + "KeyRangeDataDiskAssignment" + ] = proto.RepeatedField( proto.MESSAGE, number=2, message="KeyRangeDataDiskAssignment", @@ -485,11 +489,11 @@ class StreamingApplianceSnapshotConfig(proto.Message): appliance state. """ - snapshot_id = proto.Field( + snapshot_id: str = proto.Field( proto.STRING, number=1, ) - import_state_endpoint = proto.Field( + import_state_endpoint: str = proto.Field( proto.STRING, number=2, ) diff --git a/google/cloud/dataflow_v1beta3/types/templates.py b/google/cloud/dataflow_v1beta3/types/templates.py index e0ec0b8..1f01262 100644 --- a/google/cloud/dataflow_v1beta3/types/templates.py +++ b/google/cloud/dataflow_v1beta3/types/templates.py @@ -13,6 +13,8 @@ # See the License for the specific language governing permissions and # limitations under the License. # +from typing import MutableMapping, MutableSequence + from google.rpc import status_pb2 # type: ignore import proto # type: ignore @@ -71,7 +73,7 @@ class LaunchFlexTemplateResponse(proto.Message): launched. """ - job = proto.Field( + job: jobs.Job = proto.Field( proto.MESSAGE, number=1, message=jobs.Job, @@ -94,21 +96,21 @@ class ContainerSpec(proto.Message): Default runtime environment for the job. """ - image = proto.Field( + image: str = proto.Field( proto.STRING, number=1, ) - metadata = proto.Field( + metadata: "TemplateMetadata" = proto.Field( proto.MESSAGE, number=2, message="TemplateMetadata", ) - sdk_info = proto.Field( + sdk_info: "SDKInfo" = proto.Field( proto.MESSAGE, number=3, message="SDKInfo", ) - default_environment = proto.Field( + default_environment: "FlexTemplateRuntimeEnvironment" = proto.Field( proto.MESSAGE, number=4, message="FlexTemplateRuntimeEnvironment", @@ -139,9 +141,9 @@ class LaunchFlexTemplateParameter(proto.Message): serialized ContainerSpec as content. This field is a member of `oneof`_ ``template``. - parameters (Mapping[str, str]): + parameters (MutableMapping[str, str]): The parameters for FlexTemplate. Ex. {"num_workers":"5"} - launch_options (Mapping[str, str]): + launch_options (MutableMapping[str, str]): Launch options for this flex template job. This is a common set of options across languages and templates. This should not be used to pass @@ -153,46 +155,46 @@ class LaunchFlexTemplateParameter(proto.Message): Set this to true if you are sending a request to update a running streaming job. When set, the job name should be the same as the running job. - transform_name_mappings (Mapping[str, str]): + transform_name_mappings (MutableMapping[str, str]): Use this to pass transform_name_mappings for streaming update jobs. Ex:{"oldTransformName":"newTransformName",...}' """ - job_name = proto.Field( + job_name: str = proto.Field( proto.STRING, number=1, ) - container_spec = proto.Field( + container_spec: "ContainerSpec" = proto.Field( proto.MESSAGE, number=4, oneof="template", message="ContainerSpec", ) - container_spec_gcs_path = proto.Field( + container_spec_gcs_path: str = proto.Field( proto.STRING, number=5, oneof="template", ) - parameters = proto.MapField( + parameters: MutableMapping[str, str] = proto.MapField( proto.STRING, proto.STRING, number=2, ) - launch_options = proto.MapField( + launch_options: MutableMapping[str, str] = proto.MapField( proto.STRING, proto.STRING, number=6, ) - environment = proto.Field( + environment: "FlexTemplateRuntimeEnvironment" = proto.Field( proto.MESSAGE, number=7, message="FlexTemplateRuntimeEnvironment", ) - update = proto.Field( + update: bool = proto.Field( proto.BOOL, number=8, ) - transform_name_mappings = proto.MapField( + transform_name_mappings: MutableMapping[str, str] = proto.MapField( proto.STRING, proto.STRING, number=9, @@ -225,7 +227,7 @@ class FlexTemplateRuntimeEnvironment(proto.Message): machine_type (str): The machine type to use for the job. Defaults to the value from the template if not specified. - additional_experiments (Sequence[str]): + additional_experiments (MutableSequence[str]): Additional experiment flags for the job. network (str): Network to which VMs will be assigned. If @@ -239,7 +241,7 @@ class FlexTemplateRuntimeEnvironment(proto.Message): or "regions/REGION/subnetworks/SUBNETWORK". If the subnetwork is located in a Shared VPC network, you must use the complete URL. - additional_user_labels (Mapping[str, str]): + additional_user_labels (MutableMapping[str, str]): Additional user labels to be specified for the job. Keys and values must follow the restrictions specified in the `labeling @@ -302,99 +304,99 @@ class FlexTemplateRuntimeEnvironment(proto.Message): job. The default is n1-standard-1. """ - num_workers = proto.Field( + num_workers: int = proto.Field( proto.INT32, number=1, ) - max_workers = proto.Field( + max_workers: int = proto.Field( proto.INT32, number=2, ) - zone = proto.Field( + zone: str = proto.Field( proto.STRING, number=3, ) - service_account_email = proto.Field( + service_account_email: str = proto.Field( proto.STRING, number=4, ) - temp_location = proto.Field( + temp_location: str = proto.Field( proto.STRING, number=5, ) - machine_type = proto.Field( + machine_type: str = proto.Field( proto.STRING, number=6, ) - additional_experiments = proto.RepeatedField( + additional_experiments: MutableSequence[str] = proto.RepeatedField( proto.STRING, number=7, ) - network = proto.Field( + network: str = proto.Field( proto.STRING, number=8, ) - subnetwork = proto.Field( + subnetwork: str = proto.Field( proto.STRING, number=9, ) - additional_user_labels = proto.MapField( + additional_user_labels: MutableMapping[str, str] = proto.MapField( proto.STRING, proto.STRING, number=10, ) - kms_key_name = proto.Field( + kms_key_name: str = proto.Field( proto.STRING, number=11, ) - ip_configuration = proto.Field( + ip_configuration: gd_environment.WorkerIPAddressConfiguration = proto.Field( proto.ENUM, number=12, enum=gd_environment.WorkerIPAddressConfiguration, ) - worker_region = proto.Field( + worker_region: str = proto.Field( proto.STRING, number=13, ) - worker_zone = proto.Field( + worker_zone: str = proto.Field( proto.STRING, number=14, ) - enable_streaming_engine = proto.Field( + enable_streaming_engine: bool = proto.Field( proto.BOOL, number=15, ) - flexrs_goal = proto.Field( + flexrs_goal: gd_environment.FlexResourceSchedulingGoal = proto.Field( proto.ENUM, number=16, enum=gd_environment.FlexResourceSchedulingGoal, ) - staging_location = proto.Field( + staging_location: str = proto.Field( proto.STRING, number=17, ) - sdk_container_image = proto.Field( + sdk_container_image: str = proto.Field( proto.STRING, number=18, ) - disk_size_gb = proto.Field( + disk_size_gb: int = proto.Field( proto.INT32, number=20, ) - autoscaling_algorithm = proto.Field( + autoscaling_algorithm: gd_environment.AutoscalingAlgorithm = proto.Field( proto.ENUM, number=21, enum=gd_environment.AutoscalingAlgorithm, ) - dump_heap_on_oom = proto.Field( + dump_heap_on_oom: bool = proto.Field( proto.BOOL, number=22, ) - save_heap_dumps_to_gcs_path = proto.Field( + save_heap_dumps_to_gcs_path: str = proto.Field( proto.STRING, number=23, ) - launcher_machine_type = proto.Field( + launcher_machine_type: str = proto.Field( proto.STRING, number=24, ) @@ -419,20 +421,20 @@ class LaunchFlexTemplateRequest(proto.Message): actually executed. Defaults to false. """ - project_id = proto.Field( + project_id: str = proto.Field( proto.STRING, number=1, ) - launch_parameter = proto.Field( + launch_parameter: "LaunchFlexTemplateParameter" = proto.Field( proto.MESSAGE, number=2, message="LaunchFlexTemplateParameter", ) - location = proto.Field( + location: str = proto.Field( proto.STRING, number=3, ) - validate_only = proto.Field( + validate_only: bool = proto.Field( proto.BOOL, number=4, ) @@ -466,7 +468,7 @@ class RuntimeEnvironment(proto.Message): machine_type (str): The machine type to use for the job. Defaults to the value from the template if not specified. - additional_experiments (Sequence[str]): + additional_experiments (MutableSequence[str]): Additional experiment flags for the job, specified with the ``--experiments`` option. network (str): @@ -481,7 +483,7 @@ class RuntimeEnvironment(proto.Message): or "regions/REGION/subnetworks/SUBNETWORK". If the subnetwork is located in a Shared VPC network, you must use the complete URL. - additional_user_labels (Mapping[str, str]): + additional_user_labels (MutableMapping[str, str]): Additional user labels to be specified for the job. Keys and values should follow the restrictions specified in the `labeling @@ -515,69 +517,69 @@ class RuntimeEnvironment(proto.Message): job. """ - num_workers = proto.Field( + num_workers: int = proto.Field( proto.INT32, number=11, ) - max_workers = proto.Field( + max_workers: int = proto.Field( proto.INT32, number=1, ) - zone = proto.Field( + zone: str = proto.Field( proto.STRING, number=2, ) - service_account_email = proto.Field( + service_account_email: str = proto.Field( proto.STRING, number=3, ) - temp_location = proto.Field( + temp_location: str = proto.Field( proto.STRING, number=4, ) - bypass_temp_dir_validation = proto.Field( + bypass_temp_dir_validation: bool = proto.Field( proto.BOOL, number=5, ) - machine_type = proto.Field( + machine_type: str = proto.Field( proto.STRING, number=6, ) - additional_experiments = proto.RepeatedField( + additional_experiments: MutableSequence[str] = proto.RepeatedField( proto.STRING, number=7, ) - network = proto.Field( + network: str = proto.Field( proto.STRING, number=8, ) - subnetwork = proto.Field( + subnetwork: str = proto.Field( proto.STRING, number=9, ) - additional_user_labels = proto.MapField( + additional_user_labels: MutableMapping[str, str] = proto.MapField( proto.STRING, proto.STRING, number=10, ) - kms_key_name = proto.Field( + kms_key_name: str = proto.Field( proto.STRING, number=12, ) - ip_configuration = proto.Field( + ip_configuration: gd_environment.WorkerIPAddressConfiguration = proto.Field( proto.ENUM, number=14, enum=gd_environment.WorkerIPAddressConfiguration, ) - worker_region = proto.Field( + worker_region: str = proto.Field( proto.STRING, number=15, ) - worker_zone = proto.Field( + worker_zone: str = proto.Field( proto.STRING, number=16, ) - enable_streaming_engine = proto.Field( + enable_streaming_engine: bool = proto.Field( proto.BOOL, number=17, ) @@ -598,43 +600,43 @@ class ParameterMetadata(proto.Message): is_optional (bool): Optional. Whether the parameter is optional. Defaults to false. - regexes (Sequence[str]): + regexes (MutableSequence[str]): Optional. Regexes that the parameter must match. param_type (google.cloud.dataflow_v1beta3.types.ParameterType): Optional. The type of the parameter. Used for selecting input picker. - custom_metadata (Mapping[str, str]): + custom_metadata (MutableMapping[str, str]): Optional. Additional metadata for describing this parameter. """ - name = proto.Field( + name: str = proto.Field( proto.STRING, number=1, ) - label = proto.Field( + label: str = proto.Field( proto.STRING, number=2, ) - help_text = proto.Field( + help_text: str = proto.Field( proto.STRING, number=3, ) - is_optional = proto.Field( + is_optional: bool = proto.Field( proto.BOOL, number=4, ) - regexes = proto.RepeatedField( + regexes: MutableSequence[str] = proto.RepeatedField( proto.STRING, number=5, ) - param_type = proto.Field( + param_type: "ParameterType" = proto.Field( proto.ENUM, number=6, enum="ParameterType", ) - custom_metadata = proto.MapField( + custom_metadata: MutableMapping[str, str] = proto.MapField( proto.STRING, proto.STRING, number=7, @@ -649,19 +651,19 @@ class TemplateMetadata(proto.Message): Required. The name of the template. description (str): Optional. A description of the template. - parameters (Sequence[google.cloud.dataflow_v1beta3.types.ParameterMetadata]): + parameters (MutableSequence[google.cloud.dataflow_v1beta3.types.ParameterMetadata]): The parameters for the template. """ - name = proto.Field( + name: str = proto.Field( proto.STRING, number=1, ) - description = proto.Field( + description: str = proto.Field( proto.STRING, number=2, ) - parameters = proto.RepeatedField( + parameters: MutableSequence["ParameterMetadata"] = proto.RepeatedField( proto.MESSAGE, number=3, message="ParameterMetadata", @@ -684,12 +686,12 @@ class Language(proto.Enum): JAVA = 1 PYTHON = 2 - language = proto.Field( + language: Language = proto.Field( proto.ENUM, number=1, enum=Language, ) - version = proto.Field( + version: str = proto.Field( proto.STRING, number=2, ) @@ -701,16 +703,16 @@ class RuntimeMetadata(proto.Message): Attributes: sdk_info (google.cloud.dataflow_v1beta3.types.SDKInfo): SDK Info for the template. - parameters (Sequence[google.cloud.dataflow_v1beta3.types.ParameterMetadata]): + parameters (MutableSequence[google.cloud.dataflow_v1beta3.types.ParameterMetadata]): The parameters for the template. """ - sdk_info = proto.Field( + sdk_info: "SDKInfo" = proto.Field( proto.MESSAGE, number=1, message="SDKInfo", ) - parameters = proto.RepeatedField( + parameters: MutableSequence["ParameterMetadata"] = proto.RepeatedField( proto.MESSAGE, number=2, message="ParameterMetadata", @@ -735,7 +737,7 @@ class CreateJobFromTemplateRequest(proto.Message): with ``gs://``. This field is a member of `oneof`_ ``template``. - parameters (Mapping[str, str]): + parameters (MutableMapping[str, str]): The runtime parameters to pass to the job. environment (google.cloud.dataflow_v1beta3.types.RuntimeEnvironment): The runtime environment for the job. @@ -745,30 +747,30 @@ class CreateJobFromTemplateRequest(proto.Message): to which to direct the request. """ - project_id = proto.Field( + project_id: str = proto.Field( proto.STRING, number=1, ) - job_name = proto.Field( + job_name: str = proto.Field( proto.STRING, number=4, ) - gcs_path = proto.Field( + gcs_path: str = proto.Field( proto.STRING, number=2, oneof="template", ) - parameters = proto.MapField( + parameters: MutableMapping[str, str] = proto.MapField( proto.STRING, proto.STRING, number=3, ) - environment = proto.Field( + environment: "RuntimeEnvironment" = proto.Field( proto.MESSAGE, number=5, message="RuntimeEnvironment", ) - location = proto.Field( + location: str = proto.Field( proto.STRING, number=6, ) @@ -802,21 +804,21 @@ class TemplateView(proto.Enum): r"""The various views of a template that may be retrieved.""" METADATA_ONLY = 0 - project_id = proto.Field( + project_id: str = proto.Field( proto.STRING, number=1, ) - gcs_path = proto.Field( + gcs_path: str = proto.Field( proto.STRING, number=2, oneof="template", ) - view = proto.Field( + view: TemplateView = proto.Field( proto.ENUM, number=3, enum=TemplateView, ) - location = proto.Field( + location: str = proto.Field( proto.STRING, number=4, ) @@ -845,22 +847,22 @@ class TemplateType(proto.Enum): LEGACY = 1 FLEX = 2 - status = proto.Field( + status: status_pb2.Status = proto.Field( proto.MESSAGE, number=1, message=status_pb2.Status, ) - metadata = proto.Field( + metadata: "TemplateMetadata" = proto.Field( proto.MESSAGE, number=2, message="TemplateMetadata", ) - template_type = proto.Field( + template_type: TemplateType = proto.Field( proto.ENUM, number=3, enum=TemplateType, ) - runtime_metadata = proto.Field( + runtime_metadata: "RuntimeMetadata" = proto.Field( proto.MESSAGE, number=4, message="RuntimeMetadata", @@ -874,7 +876,7 @@ class LaunchTemplateParameters(proto.Message): job_name (str): Required. The job name to use for the created job. - parameters (Mapping[str, str]): + parameters (MutableMapping[str, str]): The runtime parameters to pass to the job. environment (google.cloud.dataflow_v1beta3.types.RuntimeEnvironment): The runtime environment for the job. @@ -882,32 +884,32 @@ class LaunchTemplateParameters(proto.Message): If set, replace the existing pipeline with the name specified by jobName with this pipeline, preserving state. - transform_name_mapping (Mapping[str, str]): + transform_name_mapping (MutableMapping[str, str]): Only applicable when updating a pipeline. Map of transform name prefixes of the job to be replaced to the corresponding name prefixes of the new job. """ - job_name = proto.Field( + job_name: str = proto.Field( proto.STRING, number=1, ) - parameters = proto.MapField( + parameters: MutableMapping[str, str] = proto.MapField( proto.STRING, proto.STRING, number=2, ) - environment = proto.Field( + environment: "RuntimeEnvironment" = proto.Field( proto.MESSAGE, number=3, message="RuntimeEnvironment", ) - update = proto.Field( + update: bool = proto.Field( proto.BOOL, number=4, ) - transform_name_mapping = proto.MapField( + transform_name_mapping: MutableMapping[str, str] = proto.MapField( proto.STRING, proto.STRING, number=5, @@ -952,31 +954,31 @@ class LaunchTemplateRequest(proto.Message): to which to direct the request. """ - project_id = proto.Field( + project_id: str = proto.Field( proto.STRING, number=1, ) - validate_only = proto.Field( + validate_only: bool = proto.Field( proto.BOOL, number=2, ) - gcs_path = proto.Field( + gcs_path: str = proto.Field( proto.STRING, number=3, oneof="template", ) - dynamic_template = proto.Field( + dynamic_template: "DynamicTemplateLaunchParams" = proto.Field( proto.MESSAGE, number=6, oneof="template", message="DynamicTemplateLaunchParams", ) - launch_parameters = proto.Field( + launch_parameters: "LaunchTemplateParameters" = proto.Field( proto.MESSAGE, number=4, message="LaunchTemplateParameters", ) - location = proto.Field( + location: str = proto.Field( proto.STRING, number=5, ) @@ -992,7 +994,7 @@ class LaunchTemplateResponse(proto.Message): launched. """ - job = proto.Field( + job: jobs.Job = proto.Field( proto.MESSAGE, number=1, message=jobs.Job, @@ -1004,7 +1006,7 @@ class InvalidTemplateParameters(proto.Message): indicates problems with the template parameter. Attributes: - parameter_violations (Sequence[google.cloud.dataflow_v1beta3.types.InvalidTemplateParameters.ParameterViolation]): + parameter_violations (MutableSequence[google.cloud.dataflow_v1beta3.types.InvalidTemplateParameters.ParameterViolation]): Describes all parameter violations in a template request. """ @@ -1020,16 +1022,16 @@ class ParameterViolation(proto.Message): validate. """ - parameter = proto.Field( + parameter: str = proto.Field( proto.STRING, number=1, ) - description = proto.Field( + description: str = proto.Field( proto.STRING, number=2, ) - parameter_violations = proto.RepeatedField( + parameter_violations: MutableSequence[ParameterViolation] = proto.RepeatedField( proto.MESSAGE, number=1, message=ParameterViolation, @@ -1050,11 +1052,11 @@ class DynamicTemplateLaunchParams(proto.Message): Cloud Storage URL, beginning with ``gs://``. """ - gcs_path = proto.Field( + gcs_path: str = proto.Field( proto.STRING, number=1, ) - staging_location = proto.Field( + staging_location: str = proto.Field( proto.STRING, number=2, ) diff --git a/noxfile.py b/noxfile.py index ffe9f35..d8440c0 100644 --- a/noxfile.py +++ b/noxfile.py @@ -273,12 +273,16 @@ def cover(session): session.run("coverage", "erase") -@nox.session(python=DEFAULT_PYTHON_VERSION) +@nox.session(python="3.9") def docs(session): """Build the docs for this library.""" session.install("-e", ".") - session.install("sphinx==4.0.1", "alabaster", "recommonmark") + session.install( + "sphinx==4.0.1", + "alabaster", + "recommonmark", + ) shutil.rmtree(os.path.join("docs", "_build"), ignore_errors=True) session.run( @@ -295,13 +299,16 @@ def docs(session): ) -@nox.session(python=DEFAULT_PYTHON_VERSION) +@nox.session(python="3.9") def docfx(session): """Build the docfx yaml files for this library.""" session.install("-e", ".") session.install( - "sphinx==4.0.1", "alabaster", "recommonmark", "gcp-sphinx-docfx-yaml" + "sphinx==4.0.1", + "alabaster", + "recommonmark", + "gcp-sphinx-docfx-yaml", ) shutil.rmtree(os.path.join("docs", "_build"), ignore_errors=True) diff --git a/owlbot.py b/owlbot.py index 9fedcf1..9891301 100644 --- a/owlbot.py +++ b/owlbot.py @@ -12,7 +12,9 @@ # See the License for the specific language governing permissions and # limitations under the License. +import json from pathlib import Path +import shutil import synthtool as s import synthtool.gcp as gcp @@ -22,9 +24,18 @@ # Copy the generated client from the owl-bot staging directory # ---------------------------------------------------------------------------- -default_version = "v1beta3" +clean_up_generated_samples = True + +# Load the default version defined in .repo-metadata.json. +default_version = json.load(open(".repo-metadata.json", "rt")).get( + "default_version" +) for library in s.get_staging_dirs(default_version): + if clean_up_generated_samples: + shutil.rmtree("samples/generated_samples", ignore_errors=True) + clean_up_generated_samples = False + # Work around to avoid a breaking change # Replace `set` with `set_` s.replace( @@ -35,10 +46,11 @@ s.replace( library / "google/cloud/dataflow_v1beta3/types/metrics.py", - "set = proto.Field", - "set_ = proto.Field", + "set: struct_pb2.Value = proto.Field", + "set_: struct_pb2.Value = proto.Field", ) - s.move(library, excludes=["setup.py"]) + + s.move([library], excludes=["**/gapic_version.py"]) s.remove_staging_dirs() # ---------------------------------------------------------------------------- @@ -50,11 +62,10 @@ microgenerator=True, versions=gcp.common.detect_versions(path="./google", default_first=True), ) -s.move(templated_files, excludes=[".coveragerc"]) # the microgenerator has a good coveragerc file +s.move(templated_files, excludes=[".coveragerc", ".github/release-please.yml"]) python.py_samples(skip_readmes=True) - -# run blacken session for all directories which have a noxfile +# run format session for all directories which have a noxfile for noxfile in Path(".").glob("**/noxfile.py"): s.shell.run(["nox", "-s", "format"], cwd=noxfile.parent, hide_output=False) \ No newline at end of file diff --git a/release-please-config.json b/release-please-config.json new file mode 100644 index 0000000..4248c65 --- /dev/null +++ b/release-please-config.json @@ -0,0 +1,24 @@ +{ + "$schema": "https://raw.githubusercontent.com/googleapis/release-please/main/schemas/config.json", + "packages": { + ".": { + "release-type": "python", + "extra-files": [ + "google/cloud/dataflow_v1beta3/gapic_version.py", + "google/cloud/dataflow/gapic_version.py", + { + "type": "json", + "path": "samples/generated_samples/snippet_metadata_google.dataflow.v1beta3.json", + "jsonpath": "$.clientLibrary.version" + } + ] + } + }, + "release-type": "python", + "plugins": [ + { + "type": "sentence-case" + } + ], + "initial-version": "0.1.0" +} diff --git a/samples/generated_samples/dataflow_v1beta3_generated_flex_templates_service_launch_flex_template_sync_4148a07f.py b/samples/generated_samples/dataflow_v1beta3_generated_flex_templates_service_launch_flex_template_sync.py similarity index 97% rename from samples/generated_samples/dataflow_v1beta3_generated_flex_templates_service_launch_flex_template_sync_4148a07f.py rename to samples/generated_samples/dataflow_v1beta3_generated_flex_templates_service_launch_flex_template_sync.py index 9fb332d..8f236ed 100644 --- a/samples/generated_samples/dataflow_v1beta3_generated_flex_templates_service_launch_flex_template_sync_4148a07f.py +++ b/samples/generated_samples/dataflow_v1beta3_generated_flex_templates_service_launch_flex_template_sync.py @@ -23,7 +23,7 @@ # python3 -m pip install google-cloud-dataflow-client -# [START dataflow_v1beta3_generated_FlexTemplatesService_LaunchFlexTemplate_sync_4148a07f] +# [START dataflow_v1beta3_generated_FlexTemplatesService_LaunchFlexTemplate_sync] # This snippet has been automatically generated and should be regarded as a # code template only. # It will require modifications to work: @@ -48,4 +48,4 @@ def sample_launch_flex_template(): # Handle the response print(response) -# [END dataflow_v1beta3_generated_FlexTemplatesService_LaunchFlexTemplate_sync_4148a07f] +# [END dataflow_v1beta3_generated_FlexTemplatesService_LaunchFlexTemplate_sync] diff --git a/samples/generated_samples/dataflow_v1beta3_generated_flex_templates_service_launch_flex_template_sync_a4f75f91.py b/samples/generated_samples/dataflow_v1beta3_generated_flex_templates_service_launch_flex_template_sync_a4f75f91.py deleted file mode 100644 index 2cc8a42..0000000 --- a/samples/generated_samples/dataflow_v1beta3_generated_flex_templates_service_launch_flex_template_sync_a4f75f91.py +++ /dev/null @@ -1,51 +0,0 @@ -# -*- coding: utf-8 -*- -# Copyright 2022 Google LLC -# -# Licensed under the Apache License, Version 2.0 (the "License"); -# you may not use this file except in compliance with the License. -# You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. -# -# Generated code. DO NOT EDIT! -# -# Snippet for LaunchFlexTemplate -# NOTE: This snippet has been automatically generated for illustrative purposes only. -# It may require modifications to work in your environment. - -# To install the latest published package dependency, execute the following: -# python3 -m pip install google-cloud-dataflow-client - - -# [START dataflow_v1beta3_generated_FlexTemplatesService_LaunchFlexTemplate_sync_a4f75f91] -# This snippet has been automatically generated and should be regarded as a -# code template only. -# It will require modifications to work: -# - It may require correct/in-range values for request initialization. -# - It may require specifying regional endpoints when creating the service -# client as shown in: -# https://googleapis.dev/python/google-api-core/latest/client_options.html -from google.cloud import dataflow_v1beta3 - - -def sample_launch_flex_template(): - # Create a client - client = dataflow_v1beta3.FlexTemplatesServiceClient() - - # Initialize request argument(s) - request = dataflow_v1beta3.LaunchFlexTemplateRequest( - ) - - # Make the request - response = client.launch_flex_template(request=request) - - # Handle the response - print(response) - -# [END dataflow_v1beta3_generated_FlexTemplatesService_LaunchFlexTemplate_sync_a4f75f91] diff --git a/samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_aggregated_list_jobs_sync_0d901b38.py b/samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_aggregated_list_jobs_sync.py similarity index 98% rename from samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_aggregated_list_jobs_sync_0d901b38.py rename to samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_aggregated_list_jobs_sync.py index a45077e..7b9946e 100644 --- a/samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_aggregated_list_jobs_sync_0d901b38.py +++ b/samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_aggregated_list_jobs_sync.py @@ -23,7 +23,7 @@ # python3 -m pip install google-cloud-dataflow-client -# [START dataflow_v1beta3_generated_JobsV1Beta3_AggregatedListJobs_sync_0d901b38] +# [START dataflow_v1beta3_generated_JobsV1Beta3_AggregatedListJobs_sync] # This snippet has been automatically generated and should be regarded as a # code template only. # It will require modifications to work: @@ -49,4 +49,4 @@ def sample_aggregated_list_jobs(): for response in page_result: print(response) -# [END dataflow_v1beta3_generated_JobsV1Beta3_AggregatedListJobs_sync_0d901b38] +# [END dataflow_v1beta3_generated_JobsV1Beta3_AggregatedListJobs_sync] diff --git a/samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_aggregated_list_jobs_sync_26f07383.py b/samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_aggregated_list_jobs_sync_26f07383.py deleted file mode 100644 index 778a4f1..0000000 --- a/samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_aggregated_list_jobs_sync_26f07383.py +++ /dev/null @@ -1,52 +0,0 @@ -# -*- coding: utf-8 -*- -# Copyright 2022 Google LLC -# -# Licensed under the Apache License, Version 2.0 (the "License"); -# you may not use this file except in compliance with the License. -# You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. -# -# Generated code. DO NOT EDIT! -# -# Snippet for AggregatedListJobs -# NOTE: This snippet has been automatically generated for illustrative purposes only. -# It may require modifications to work in your environment. - -# To install the latest published package dependency, execute the following: -# python3 -m pip install google-cloud-dataflow-client - - -# [START dataflow_v1beta3_generated_JobsV1Beta3_AggregatedListJobs_sync_26f07383] -# This snippet has been automatically generated and should be regarded as a -# code template only. -# It will require modifications to work: -# - It may require correct/in-range values for request initialization. -# - It may require specifying regional endpoints when creating the service -# client as shown in: -# https://googleapis.dev/python/google-api-core/latest/client_options.html -from google.cloud import dataflow_v1beta3 - - -def sample_aggregated_list_jobs(): - # Create a client - client = dataflow_v1beta3.JobsV1Beta3Client() - - # Initialize request argument(s) - request = dataflow_v1beta3.ListJobsRequest( - ) - - # Make the request - page_result = client.aggregated_list_jobs(request=request) - - # Handle the response - for response in page_result: - print(response) - -# [END dataflow_v1beta3_generated_JobsV1Beta3_AggregatedListJobs_sync_26f07383] diff --git a/samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_check_active_jobs_sync_aab49b35.py b/samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_check_active_jobs_sync.py similarity index 98% rename from samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_check_active_jobs_sync_aab49b35.py rename to samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_check_active_jobs_sync.py index d6c5c9a..591caa0 100644 --- a/samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_check_active_jobs_sync_aab49b35.py +++ b/samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_check_active_jobs_sync.py @@ -23,7 +23,7 @@ # python3 -m pip install google-cloud-dataflow-client -# [START dataflow_v1beta3_generated_JobsV1Beta3_CheckActiveJobs_sync_aab49b35] +# [START dataflow_v1beta3_generated_JobsV1Beta3_CheckActiveJobs_sync] # This snippet has been automatically generated and should be regarded as a # code template only. # It will require modifications to work: @@ -48,4 +48,4 @@ def sample_check_active_jobs(): # Handle the response print(response) -# [END dataflow_v1beta3_generated_JobsV1Beta3_CheckActiveJobs_sync_aab49b35] +# [END dataflow_v1beta3_generated_JobsV1Beta3_CheckActiveJobs_sync] diff --git a/samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_check_active_jobs_sync_baed931b.py b/samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_check_active_jobs_sync_baed931b.py deleted file mode 100644 index dc9691c..0000000 --- a/samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_check_active_jobs_sync_baed931b.py +++ /dev/null @@ -1,51 +0,0 @@ -# -*- coding: utf-8 -*- -# Copyright 2022 Google LLC -# -# Licensed under the Apache License, Version 2.0 (the "License"); -# you may not use this file except in compliance with the License. -# You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. -# -# Generated code. DO NOT EDIT! -# -# Snippet for CheckActiveJobs -# NOTE: This snippet has been automatically generated for illustrative purposes only. -# It may require modifications to work in your environment. - -# To install the latest published package dependency, execute the following: -# python3 -m pip install google-cloud-dataflow-client - - -# [START dataflow_v1beta3_generated_JobsV1Beta3_CheckActiveJobs_sync_baed931b] -# This snippet has been automatically generated and should be regarded as a -# code template only. -# It will require modifications to work: -# - It may require correct/in-range values for request initialization. -# - It may require specifying regional endpoints when creating the service -# client as shown in: -# https://googleapis.dev/python/google-api-core/latest/client_options.html -from google.cloud import dataflow_v1beta3 - - -def sample_check_active_jobs(): - # Create a client - client = dataflow_v1beta3.JobsV1Beta3Client() - - # Initialize request argument(s) - request = dataflow_v1beta3.CheckActiveJobsRequest( - ) - - # Make the request - response = client.check_active_jobs(request=request) - - # Handle the response - print(response) - -# [END dataflow_v1beta3_generated_JobsV1Beta3_CheckActiveJobs_sync_baed931b] diff --git a/samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_create_job_sync_2a1b0208.py b/samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_create_job_sync.py similarity index 92% rename from samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_create_job_sync_2a1b0208.py rename to samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_create_job_sync.py index 406ead7..b91dc91 100644 --- a/samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_create_job_sync_2a1b0208.py +++ b/samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_create_job_sync.py @@ -23,7 +23,7 @@ # python3 -m pip install google-cloud-dataflow-client -# [START dataflow_v1beta3_generated_JobsV1Beta3_CreateJob_sync_2a1b0208] +# [START dataflow_v1beta3_generated_JobsV1Beta3_CreateJob_sync] # This snippet has been automatically generated and should be regarded as a # code template only. # It will require modifications to work: @@ -48,4 +48,4 @@ def sample_create_job(): # Handle the response print(response) -# [END dataflow_v1beta3_generated_JobsV1Beta3_CreateJob_sync_2a1b0208] +# [END dataflow_v1beta3_generated_JobsV1Beta3_CreateJob_sync] diff --git a/samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_create_job_sync_ad13d605.py b/samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_create_job_sync_ad13d605.py deleted file mode 100644 index 93aadf8..0000000 --- a/samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_create_job_sync_ad13d605.py +++ /dev/null @@ -1,51 +0,0 @@ -# -*- coding: utf-8 -*- -# Copyright 2022 Google LLC -# -# Licensed under the Apache License, Version 2.0 (the "License"); -# you may not use this file except in compliance with the License. -# You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. -# -# Generated code. DO NOT EDIT! -# -# Snippet for CreateJob -# NOTE: This snippet has been automatically generated for illustrative purposes only. -# It may require modifications to work in your environment. - -# To install the latest published package dependency, execute the following: -# python3 -m pip install google-cloud-dataflow-client - - -# [START dataflow_v1beta3_generated_JobsV1Beta3_CreateJob_sync_ad13d605] -# This snippet has been automatically generated and should be regarded as a -# code template only. -# It will require modifications to work: -# - It may require correct/in-range values for request initialization. -# - It may require specifying regional endpoints when creating the service -# client as shown in: -# https://googleapis.dev/python/google-api-core/latest/client_options.html -from google.cloud import dataflow_v1beta3 - - -def sample_create_job(): - # Create a client - client = dataflow_v1beta3.JobsV1Beta3Client() - - # Initialize request argument(s) - request = dataflow_v1beta3.CreateJobRequest( - ) - - # Make the request - response = client.create_job(request=request) - - # Handle the response - print(response) - -# [END dataflow_v1beta3_generated_JobsV1Beta3_CreateJob_sync_ad13d605] diff --git a/samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_get_job_sync_3e418026.py b/samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_get_job_sync.py similarity index 92% rename from samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_get_job_sync_3e418026.py rename to samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_get_job_sync.py index 9980879..c4bef41 100644 --- a/samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_get_job_sync_3e418026.py +++ b/samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_get_job_sync.py @@ -23,7 +23,7 @@ # python3 -m pip install google-cloud-dataflow-client -# [START dataflow_v1beta3_generated_JobsV1Beta3_GetJob_sync_3e418026] +# [START dataflow_v1beta3_generated_JobsV1Beta3_GetJob_sync] # This snippet has been automatically generated and should be regarded as a # code template only. # It will require modifications to work: @@ -48,4 +48,4 @@ def sample_get_job(): # Handle the response print(response) -# [END dataflow_v1beta3_generated_JobsV1Beta3_GetJob_sync_3e418026] +# [END dataflow_v1beta3_generated_JobsV1Beta3_GetJob_sync] diff --git a/samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_get_job_sync_45212860.py b/samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_get_job_sync_45212860.py deleted file mode 100644 index 364e642..0000000 --- a/samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_get_job_sync_45212860.py +++ /dev/null @@ -1,51 +0,0 @@ -# -*- coding: utf-8 -*- -# Copyright 2022 Google LLC -# -# Licensed under the Apache License, Version 2.0 (the "License"); -# you may not use this file except in compliance with the License. -# You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. -# -# Generated code. DO NOT EDIT! -# -# Snippet for GetJob -# NOTE: This snippet has been automatically generated for illustrative purposes only. -# It may require modifications to work in your environment. - -# To install the latest published package dependency, execute the following: -# python3 -m pip install google-cloud-dataflow-client - - -# [START dataflow_v1beta3_generated_JobsV1Beta3_GetJob_sync_45212860] -# This snippet has been automatically generated and should be regarded as a -# code template only. -# It will require modifications to work: -# - It may require correct/in-range values for request initialization. -# - It may require specifying regional endpoints when creating the service -# client as shown in: -# https://googleapis.dev/python/google-api-core/latest/client_options.html -from google.cloud import dataflow_v1beta3 - - -def sample_get_job(): - # Create a client - client = dataflow_v1beta3.JobsV1Beta3Client() - - # Initialize request argument(s) - request = dataflow_v1beta3.GetJobRequest( - ) - - # Make the request - response = client.get_job(request=request) - - # Handle the response - print(response) - -# [END dataflow_v1beta3_generated_JobsV1Beta3_GetJob_sync_45212860] diff --git a/samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_list_jobs_sync_97a60855.py b/samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_list_jobs_sync.py similarity index 92% rename from samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_list_jobs_sync_97a60855.py rename to samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_list_jobs_sync.py index 8ebfabf..a975672 100644 --- a/samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_list_jobs_sync_97a60855.py +++ b/samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_list_jobs_sync.py @@ -23,7 +23,7 @@ # python3 -m pip install google-cloud-dataflow-client -# [START dataflow_v1beta3_generated_JobsV1Beta3_ListJobs_sync_97a60855] +# [START dataflow_v1beta3_generated_JobsV1Beta3_ListJobs_sync] # This snippet has been automatically generated and should be regarded as a # code template only. # It will require modifications to work: @@ -49,4 +49,4 @@ def sample_list_jobs(): for response in page_result: print(response) -# [END dataflow_v1beta3_generated_JobsV1Beta3_ListJobs_sync_97a60855] +# [END dataflow_v1beta3_generated_JobsV1Beta3_ListJobs_sync] diff --git a/samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_list_jobs_sync_bfb75b4c.py b/samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_list_jobs_sync_bfb75b4c.py deleted file mode 100644 index 633ac93..0000000 --- a/samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_list_jobs_sync_bfb75b4c.py +++ /dev/null @@ -1,52 +0,0 @@ -# -*- coding: utf-8 -*- -# Copyright 2022 Google LLC -# -# Licensed under the Apache License, Version 2.0 (the "License"); -# you may not use this file except in compliance with the License. -# You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. -# -# Generated code. DO NOT EDIT! -# -# Snippet for ListJobs -# NOTE: This snippet has been automatically generated for illustrative purposes only. -# It may require modifications to work in your environment. - -# To install the latest published package dependency, execute the following: -# python3 -m pip install google-cloud-dataflow-client - - -# [START dataflow_v1beta3_generated_JobsV1Beta3_ListJobs_sync_bfb75b4c] -# This snippet has been automatically generated and should be regarded as a -# code template only. -# It will require modifications to work: -# - It may require correct/in-range values for request initialization. -# - It may require specifying regional endpoints when creating the service -# client as shown in: -# https://googleapis.dev/python/google-api-core/latest/client_options.html -from google.cloud import dataflow_v1beta3 - - -def sample_list_jobs(): - # Create a client - client = dataflow_v1beta3.JobsV1Beta3Client() - - # Initialize request argument(s) - request = dataflow_v1beta3.ListJobsRequest( - ) - - # Make the request - page_result = client.list_jobs(request=request) - - # Handle the response - for response in page_result: - print(response) - -# [END dataflow_v1beta3_generated_JobsV1Beta3_ListJobs_sync_bfb75b4c] diff --git a/samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_snapshot_job_sync_60f4d8fe.py b/samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_snapshot_job_sync.py similarity index 95% rename from samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_snapshot_job_sync_60f4d8fe.py rename to samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_snapshot_job_sync.py index 49a3645..db5e822 100644 --- a/samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_snapshot_job_sync_60f4d8fe.py +++ b/samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_snapshot_job_sync.py @@ -23,7 +23,7 @@ # python3 -m pip install google-cloud-dataflow-client -# [START dataflow_v1beta3_generated_JobsV1Beta3_SnapshotJob_sync_60f4d8fe] +# [START dataflow_v1beta3_generated_JobsV1Beta3_SnapshotJob_sync] # This snippet has been automatically generated and should be regarded as a # code template only. # It will require modifications to work: @@ -48,4 +48,4 @@ def sample_snapshot_job(): # Handle the response print(response) -# [END dataflow_v1beta3_generated_JobsV1Beta3_SnapshotJob_sync_60f4d8fe] +# [END dataflow_v1beta3_generated_JobsV1Beta3_SnapshotJob_sync] diff --git a/samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_snapshot_job_sync_ff3cb8bd.py b/samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_snapshot_job_sync_ff3cb8bd.py deleted file mode 100644 index a767e86..0000000 --- a/samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_snapshot_job_sync_ff3cb8bd.py +++ /dev/null @@ -1,51 +0,0 @@ -# -*- coding: utf-8 -*- -# Copyright 2022 Google LLC -# -# Licensed under the Apache License, Version 2.0 (the "License"); -# you may not use this file except in compliance with the License. -# You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. -# -# Generated code. DO NOT EDIT! -# -# Snippet for SnapshotJob -# NOTE: This snippet has been automatically generated for illustrative purposes only. -# It may require modifications to work in your environment. - -# To install the latest published package dependency, execute the following: -# python3 -m pip install google-cloud-dataflow-client - - -# [START dataflow_v1beta3_generated_JobsV1Beta3_SnapshotJob_sync_ff3cb8bd] -# This snippet has been automatically generated and should be regarded as a -# code template only. -# It will require modifications to work: -# - It may require correct/in-range values for request initialization. -# - It may require specifying regional endpoints when creating the service -# client as shown in: -# https://googleapis.dev/python/google-api-core/latest/client_options.html -from google.cloud import dataflow_v1beta3 - - -def sample_snapshot_job(): - # Create a client - client = dataflow_v1beta3.JobsV1Beta3Client() - - # Initialize request argument(s) - request = dataflow_v1beta3.SnapshotJobRequest( - ) - - # Make the request - response = client.snapshot_job(request=request) - - # Handle the response - print(response) - -# [END dataflow_v1beta3_generated_JobsV1Beta3_SnapshotJob_sync_ff3cb8bd] diff --git a/samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_update_job_sync_8db1fd08.py b/samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_update_job_sync.py similarity index 92% rename from samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_update_job_sync_8db1fd08.py rename to samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_update_job_sync.py index 0eba295..9fff7ed 100644 --- a/samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_update_job_sync_8db1fd08.py +++ b/samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_update_job_sync.py @@ -23,7 +23,7 @@ # python3 -m pip install google-cloud-dataflow-client -# [START dataflow_v1beta3_generated_JobsV1Beta3_UpdateJob_sync_8db1fd08] +# [START dataflow_v1beta3_generated_JobsV1Beta3_UpdateJob_sync] # This snippet has been automatically generated and should be regarded as a # code template only. # It will require modifications to work: @@ -48,4 +48,4 @@ def sample_update_job(): # Handle the response print(response) -# [END dataflow_v1beta3_generated_JobsV1Beta3_UpdateJob_sync_8db1fd08] +# [END dataflow_v1beta3_generated_JobsV1Beta3_UpdateJob_sync] diff --git a/samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_update_job_sync_e2fa191d.py b/samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_update_job_sync_e2fa191d.py deleted file mode 100644 index c3dac71..0000000 --- a/samples/generated_samples/dataflow_v1beta3_generated_jobs_v1_beta3_update_job_sync_e2fa191d.py +++ /dev/null @@ -1,51 +0,0 @@ -# -*- coding: utf-8 -*- -# Copyright 2022 Google LLC -# -# Licensed under the Apache License, Version 2.0 (the "License"); -# you may not use this file except in compliance with the License. -# You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. -# -# Generated code. DO NOT EDIT! -# -# Snippet for UpdateJob -# NOTE: This snippet has been automatically generated for illustrative purposes only. -# It may require modifications to work in your environment. - -# To install the latest published package dependency, execute the following: -# python3 -m pip install google-cloud-dataflow-client - - -# [START dataflow_v1beta3_generated_JobsV1Beta3_UpdateJob_sync_e2fa191d] -# This snippet has been automatically generated and should be regarded as a -# code template only. -# It will require modifications to work: -# - It may require correct/in-range values for request initialization. -# - It may require specifying regional endpoints when creating the service -# client as shown in: -# https://googleapis.dev/python/google-api-core/latest/client_options.html -from google.cloud import dataflow_v1beta3 - - -def sample_update_job(): - # Create a client - client = dataflow_v1beta3.JobsV1Beta3Client() - - # Initialize request argument(s) - request = dataflow_v1beta3.UpdateJobRequest( - ) - - # Make the request - response = client.update_job(request=request) - - # Handle the response - print(response) - -# [END dataflow_v1beta3_generated_JobsV1Beta3_UpdateJob_sync_e2fa191d] diff --git a/samples/generated_samples/dataflow_v1beta3_generated_messages_v1_beta3_list_job_messages_sync_04dec136.py b/samples/generated_samples/dataflow_v1beta3_generated_messages_v1_beta3_list_job_messages_sync.py similarity index 98% rename from samples/generated_samples/dataflow_v1beta3_generated_messages_v1_beta3_list_job_messages_sync_04dec136.py rename to samples/generated_samples/dataflow_v1beta3_generated_messages_v1_beta3_list_job_messages_sync.py index 82358ed..256bde8 100644 --- a/samples/generated_samples/dataflow_v1beta3_generated_messages_v1_beta3_list_job_messages_sync_04dec136.py +++ b/samples/generated_samples/dataflow_v1beta3_generated_messages_v1_beta3_list_job_messages_sync.py @@ -23,7 +23,7 @@ # python3 -m pip install google-cloud-dataflow-client -# [START dataflow_v1beta3_generated_MessagesV1Beta3_ListJobMessages_sync_04dec136] +# [START dataflow_v1beta3_generated_MessagesV1Beta3_ListJobMessages_sync] # This snippet has been automatically generated and should be regarded as a # code template only. # It will require modifications to work: @@ -49,4 +49,4 @@ def sample_list_job_messages(): for response in page_result: print(response) -# [END dataflow_v1beta3_generated_MessagesV1Beta3_ListJobMessages_sync_04dec136] +# [END dataflow_v1beta3_generated_MessagesV1Beta3_ListJobMessages_sync] diff --git a/samples/generated_samples/dataflow_v1beta3_generated_messages_v1_beta3_list_job_messages_sync_f19ed68d.py b/samples/generated_samples/dataflow_v1beta3_generated_messages_v1_beta3_list_job_messages_sync_f19ed68d.py deleted file mode 100644 index 1aa793d..0000000 --- a/samples/generated_samples/dataflow_v1beta3_generated_messages_v1_beta3_list_job_messages_sync_f19ed68d.py +++ /dev/null @@ -1,52 +0,0 @@ -# -*- coding: utf-8 -*- -# Copyright 2022 Google LLC -# -# Licensed under the Apache License, Version 2.0 (the "License"); -# you may not use this file except in compliance with the License. -# You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. -# -# Generated code. DO NOT EDIT! -# -# Snippet for ListJobMessages -# NOTE: This snippet has been automatically generated for illustrative purposes only. -# It may require modifications to work in your environment. - -# To install the latest published package dependency, execute the following: -# python3 -m pip install google-cloud-dataflow-client - - -# [START dataflow_v1beta3_generated_MessagesV1Beta3_ListJobMessages_sync_f19ed68d] -# This snippet has been automatically generated and should be regarded as a -# code template only. -# It will require modifications to work: -# - It may require correct/in-range values for request initialization. -# - It may require specifying regional endpoints when creating the service -# client as shown in: -# https://googleapis.dev/python/google-api-core/latest/client_options.html -from google.cloud import dataflow_v1beta3 - - -def sample_list_job_messages(): - # Create a client - client = dataflow_v1beta3.MessagesV1Beta3Client() - - # Initialize request argument(s) - request = dataflow_v1beta3.ListJobMessagesRequest( - ) - - # Make the request - page_result = client.list_job_messages(request=request) - - # Handle the response - for response in page_result: - print(response) - -# [END dataflow_v1beta3_generated_MessagesV1Beta3_ListJobMessages_sync_f19ed68d] diff --git a/samples/generated_samples/dataflow_v1beta3_generated_metrics_v1_beta3_get_job_execution_details_sync_a4ff4d57.py b/samples/generated_samples/dataflow_v1beta3_generated_metrics_v1_beta3_get_job_execution_details_sync.py similarity index 97% rename from samples/generated_samples/dataflow_v1beta3_generated_metrics_v1_beta3_get_job_execution_details_sync_a4ff4d57.py rename to samples/generated_samples/dataflow_v1beta3_generated_metrics_v1_beta3_get_job_execution_details_sync.py index a1e521e..9268495 100644 --- a/samples/generated_samples/dataflow_v1beta3_generated_metrics_v1_beta3_get_job_execution_details_sync_a4ff4d57.py +++ b/samples/generated_samples/dataflow_v1beta3_generated_metrics_v1_beta3_get_job_execution_details_sync.py @@ -23,7 +23,7 @@ # python3 -m pip install google-cloud-dataflow-client -# [START dataflow_v1beta3_generated_MetricsV1Beta3_GetJobExecutionDetails_sync_a4ff4d57] +# [START dataflow_v1beta3_generated_MetricsV1Beta3_GetJobExecutionDetails_sync] # This snippet has been automatically generated and should be regarded as a # code template only. # It will require modifications to work: @@ -49,4 +49,4 @@ def sample_get_job_execution_details(): for response in page_result: print(response) -# [END dataflow_v1beta3_generated_MetricsV1Beta3_GetJobExecutionDetails_sync_a4ff4d57] +# [END dataflow_v1beta3_generated_MetricsV1Beta3_GetJobExecutionDetails_sync] diff --git a/samples/generated_samples/dataflow_v1beta3_generated_metrics_v1_beta3_get_job_execution_details_sync_b7550163.py b/samples/generated_samples/dataflow_v1beta3_generated_metrics_v1_beta3_get_job_execution_details_sync_b7550163.py deleted file mode 100644 index 211b983..0000000 --- a/samples/generated_samples/dataflow_v1beta3_generated_metrics_v1_beta3_get_job_execution_details_sync_b7550163.py +++ /dev/null @@ -1,52 +0,0 @@ -# -*- coding: utf-8 -*- -# Copyright 2022 Google LLC -# -# Licensed under the Apache License, Version 2.0 (the "License"); -# you may not use this file except in compliance with the License. -# You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. -# -# Generated code. DO NOT EDIT! -# -# Snippet for GetJobExecutionDetails -# NOTE: This snippet has been automatically generated for illustrative purposes only. -# It may require modifications to work in your environment. - -# To install the latest published package dependency, execute the following: -# python3 -m pip install google-cloud-dataflow-client - - -# [START dataflow_v1beta3_generated_MetricsV1Beta3_GetJobExecutionDetails_sync_b7550163] -# This snippet has been automatically generated and should be regarded as a -# code template only. -# It will require modifications to work: -# - It may require correct/in-range values for request initialization. -# - It may require specifying regional endpoints when creating the service -# client as shown in: -# https://googleapis.dev/python/google-api-core/latest/client_options.html -from google.cloud import dataflow_v1beta3 - - -def sample_get_job_execution_details(): - # Create a client - client = dataflow_v1beta3.MetricsV1Beta3Client() - - # Initialize request argument(s) - request = dataflow_v1beta3.GetJobExecutionDetailsRequest( - ) - - # Make the request - page_result = client.get_job_execution_details(request=request) - - # Handle the response - for response in page_result: - print(response) - -# [END dataflow_v1beta3_generated_MetricsV1Beta3_GetJobExecutionDetails_sync_b7550163] diff --git a/samples/generated_samples/dataflow_v1beta3_generated_metrics_v1_beta3_get_job_metrics_sync_02835968.py b/samples/generated_samples/dataflow_v1beta3_generated_metrics_v1_beta3_get_job_metrics_sync.py similarity index 98% rename from samples/generated_samples/dataflow_v1beta3_generated_metrics_v1_beta3_get_job_metrics_sync_02835968.py rename to samples/generated_samples/dataflow_v1beta3_generated_metrics_v1_beta3_get_job_metrics_sync.py index 54a22ec..bbe9622 100644 --- a/samples/generated_samples/dataflow_v1beta3_generated_metrics_v1_beta3_get_job_metrics_sync_02835968.py +++ b/samples/generated_samples/dataflow_v1beta3_generated_metrics_v1_beta3_get_job_metrics_sync.py @@ -23,7 +23,7 @@ # python3 -m pip install google-cloud-dataflow-client -# [START dataflow_v1beta3_generated_MetricsV1Beta3_GetJobMetrics_sync_02835968] +# [START dataflow_v1beta3_generated_MetricsV1Beta3_GetJobMetrics_sync] # This snippet has been automatically generated and should be regarded as a # code template only. # It will require modifications to work: @@ -48,4 +48,4 @@ def sample_get_job_metrics(): # Handle the response print(response) -# [END dataflow_v1beta3_generated_MetricsV1Beta3_GetJobMetrics_sync_02835968] +# [END dataflow_v1beta3_generated_MetricsV1Beta3_GetJobMetrics_sync] diff --git a/samples/generated_samples/dataflow_v1beta3_generated_metrics_v1_beta3_get_job_metrics_sync_59dce217.py b/samples/generated_samples/dataflow_v1beta3_generated_metrics_v1_beta3_get_job_metrics_sync_59dce217.py deleted file mode 100644 index ffc68c4..0000000 --- a/samples/generated_samples/dataflow_v1beta3_generated_metrics_v1_beta3_get_job_metrics_sync_59dce217.py +++ /dev/null @@ -1,51 +0,0 @@ -# -*- coding: utf-8 -*- -# Copyright 2022 Google LLC -# -# Licensed under the Apache License, Version 2.0 (the "License"); -# you may not use this file except in compliance with the License. -# You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. -# -# Generated code. DO NOT EDIT! -# -# Snippet for GetJobMetrics -# NOTE: This snippet has been automatically generated for illustrative purposes only. -# It may require modifications to work in your environment. - -# To install the latest published package dependency, execute the following: -# python3 -m pip install google-cloud-dataflow-client - - -# [START dataflow_v1beta3_generated_MetricsV1Beta3_GetJobMetrics_sync_59dce217] -# This snippet has been automatically generated and should be regarded as a -# code template only. -# It will require modifications to work: -# - It may require correct/in-range values for request initialization. -# - It may require specifying regional endpoints when creating the service -# client as shown in: -# https://googleapis.dev/python/google-api-core/latest/client_options.html -from google.cloud import dataflow_v1beta3 - - -def sample_get_job_metrics(): - # Create a client - client = dataflow_v1beta3.MetricsV1Beta3Client() - - # Initialize request argument(s) - request = dataflow_v1beta3.GetJobMetricsRequest( - ) - - # Make the request - response = client.get_job_metrics(request=request) - - # Handle the response - print(response) - -# [END dataflow_v1beta3_generated_MetricsV1Beta3_GetJobMetrics_sync_59dce217] diff --git a/samples/generated_samples/dataflow_v1beta3_generated_metrics_v1_beta3_get_stage_execution_details_sync_4acb3ded.py b/samples/generated_samples/dataflow_v1beta3_generated_metrics_v1_beta3_get_stage_execution_details_sync.py similarity index 97% rename from samples/generated_samples/dataflow_v1beta3_generated_metrics_v1_beta3_get_stage_execution_details_sync_4acb3ded.py rename to samples/generated_samples/dataflow_v1beta3_generated_metrics_v1_beta3_get_stage_execution_details_sync.py index 76af1fd..c9e9729 100644 --- a/samples/generated_samples/dataflow_v1beta3_generated_metrics_v1_beta3_get_stage_execution_details_sync_4acb3ded.py +++ b/samples/generated_samples/dataflow_v1beta3_generated_metrics_v1_beta3_get_stage_execution_details_sync.py @@ -23,7 +23,7 @@ # python3 -m pip install google-cloud-dataflow-client -# [START dataflow_v1beta3_generated_MetricsV1Beta3_GetStageExecutionDetails_sync_4acb3ded] +# [START dataflow_v1beta3_generated_MetricsV1Beta3_GetStageExecutionDetails_sync] # This snippet has been automatically generated and should be regarded as a # code template only. # It will require modifications to work: @@ -49,4 +49,4 @@ def sample_get_stage_execution_details(): for response in page_result: print(response) -# [END dataflow_v1beta3_generated_MetricsV1Beta3_GetStageExecutionDetails_sync_4acb3ded] +# [END dataflow_v1beta3_generated_MetricsV1Beta3_GetStageExecutionDetails_sync] diff --git a/samples/generated_samples/dataflow_v1beta3_generated_metrics_v1_beta3_get_stage_execution_details_sync_9490a11d.py b/samples/generated_samples/dataflow_v1beta3_generated_metrics_v1_beta3_get_stage_execution_details_sync_9490a11d.py deleted file mode 100644 index f88599e..0000000 --- a/samples/generated_samples/dataflow_v1beta3_generated_metrics_v1_beta3_get_stage_execution_details_sync_9490a11d.py +++ /dev/null @@ -1,52 +0,0 @@ -# -*- coding: utf-8 -*- -# Copyright 2022 Google LLC -# -# Licensed under the Apache License, Version 2.0 (the "License"); -# you may not use this file except in compliance with the License. -# You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. -# -# Generated code. DO NOT EDIT! -# -# Snippet for GetStageExecutionDetails -# NOTE: This snippet has been automatically generated for illustrative purposes only. -# It may require modifications to work in your environment. - -# To install the latest published package dependency, execute the following: -# python3 -m pip install google-cloud-dataflow-client - - -# [START dataflow_v1beta3_generated_MetricsV1Beta3_GetStageExecutionDetails_sync_9490a11d] -# This snippet has been automatically generated and should be regarded as a -# code template only. -# It will require modifications to work: -# - It may require correct/in-range values for request initialization. -# - It may require specifying regional endpoints when creating the service -# client as shown in: -# https://googleapis.dev/python/google-api-core/latest/client_options.html -from google.cloud import dataflow_v1beta3 - - -def sample_get_stage_execution_details(): - # Create a client - client = dataflow_v1beta3.MetricsV1Beta3Client() - - # Initialize request argument(s) - request = dataflow_v1beta3.GetStageExecutionDetailsRequest( - ) - - # Make the request - page_result = client.get_stage_execution_details(request=request) - - # Handle the response - for response in page_result: - print(response) - -# [END dataflow_v1beta3_generated_MetricsV1Beta3_GetStageExecutionDetails_sync_9490a11d] diff --git a/samples/generated_samples/dataflow_v1beta3_generated_snapshots_v1_beta3_delete_snapshot_sync_34b1dfd7.py b/samples/generated_samples/dataflow_v1beta3_generated_snapshots_v1_beta3_delete_snapshot_sync.py similarity index 98% rename from samples/generated_samples/dataflow_v1beta3_generated_snapshots_v1_beta3_delete_snapshot_sync_34b1dfd7.py rename to samples/generated_samples/dataflow_v1beta3_generated_snapshots_v1_beta3_delete_snapshot_sync.py index cc7f9e5..59a50f5 100644 --- a/samples/generated_samples/dataflow_v1beta3_generated_snapshots_v1_beta3_delete_snapshot_sync_34b1dfd7.py +++ b/samples/generated_samples/dataflow_v1beta3_generated_snapshots_v1_beta3_delete_snapshot_sync.py @@ -23,7 +23,7 @@ # python3 -m pip install google-cloud-dataflow-client -# [START dataflow_v1beta3_generated_SnapshotsV1Beta3_DeleteSnapshot_sync_34b1dfd7] +# [START dataflow_v1beta3_generated_SnapshotsV1Beta3_DeleteSnapshot_sync] # This snippet has been automatically generated and should be regarded as a # code template only. # It will require modifications to work: @@ -48,4 +48,4 @@ def sample_delete_snapshot(): # Handle the response print(response) -# [END dataflow_v1beta3_generated_SnapshotsV1Beta3_DeleteSnapshot_sync_34b1dfd7] +# [END dataflow_v1beta3_generated_SnapshotsV1Beta3_DeleteSnapshot_sync] diff --git a/samples/generated_samples/dataflow_v1beta3_generated_snapshots_v1_beta3_delete_snapshot_sync_567e01e6.py b/samples/generated_samples/dataflow_v1beta3_generated_snapshots_v1_beta3_delete_snapshot_sync_567e01e6.py deleted file mode 100644 index 938c6e9..0000000 --- a/samples/generated_samples/dataflow_v1beta3_generated_snapshots_v1_beta3_delete_snapshot_sync_567e01e6.py +++ /dev/null @@ -1,51 +0,0 @@ -# -*- coding: utf-8 -*- -# Copyright 2022 Google LLC -# -# Licensed under the Apache License, Version 2.0 (the "License"); -# you may not use this file except in compliance with the License. -# You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. -# -# Generated code. DO NOT EDIT! -# -# Snippet for DeleteSnapshot -# NOTE: This snippet has been automatically generated for illustrative purposes only. -# It may require modifications to work in your environment. - -# To install the latest published package dependency, execute the following: -# python3 -m pip install google-cloud-dataflow-client - - -# [START dataflow_v1beta3_generated_SnapshotsV1Beta3_DeleteSnapshot_sync_567e01e6] -# This snippet has been automatically generated and should be regarded as a -# code template only. -# It will require modifications to work: -# - It may require correct/in-range values for request initialization. -# - It may require specifying regional endpoints when creating the service -# client as shown in: -# https://googleapis.dev/python/google-api-core/latest/client_options.html -from google.cloud import dataflow_v1beta3 - - -def sample_delete_snapshot(): - # Create a client - client = dataflow_v1beta3.SnapshotsV1Beta3Client() - - # Initialize request argument(s) - request = dataflow_v1beta3.DeleteSnapshotRequest( - ) - - # Make the request - response = client.delete_snapshot(request=request) - - # Handle the response - print(response) - -# [END dataflow_v1beta3_generated_SnapshotsV1Beta3_DeleteSnapshot_sync_567e01e6] diff --git a/samples/generated_samples/dataflow_v1beta3_generated_snapshots_v1_beta3_get_snapshot_sync_846de01f.py b/samples/generated_samples/dataflow_v1beta3_generated_snapshots_v1_beta3_get_snapshot_sync.py similarity index 98% rename from samples/generated_samples/dataflow_v1beta3_generated_snapshots_v1_beta3_get_snapshot_sync_846de01f.py rename to samples/generated_samples/dataflow_v1beta3_generated_snapshots_v1_beta3_get_snapshot_sync.py index 9eb621a..f1861f6 100644 --- a/samples/generated_samples/dataflow_v1beta3_generated_snapshots_v1_beta3_get_snapshot_sync_846de01f.py +++ b/samples/generated_samples/dataflow_v1beta3_generated_snapshots_v1_beta3_get_snapshot_sync.py @@ -23,7 +23,7 @@ # python3 -m pip install google-cloud-dataflow-client -# [START dataflow_v1beta3_generated_SnapshotsV1Beta3_GetSnapshot_sync_846de01f] +# [START dataflow_v1beta3_generated_SnapshotsV1Beta3_GetSnapshot_sync] # This snippet has been automatically generated and should be regarded as a # code template only. # It will require modifications to work: @@ -48,4 +48,4 @@ def sample_get_snapshot(): # Handle the response print(response) -# [END dataflow_v1beta3_generated_SnapshotsV1Beta3_GetSnapshot_sync_846de01f] +# [END dataflow_v1beta3_generated_SnapshotsV1Beta3_GetSnapshot_sync] diff --git a/samples/generated_samples/dataflow_v1beta3_generated_snapshots_v1_beta3_get_snapshot_sync_ba1ca2b6.py b/samples/generated_samples/dataflow_v1beta3_generated_snapshots_v1_beta3_get_snapshot_sync_ba1ca2b6.py deleted file mode 100644 index f4ef41f..0000000 --- a/samples/generated_samples/dataflow_v1beta3_generated_snapshots_v1_beta3_get_snapshot_sync_ba1ca2b6.py +++ /dev/null @@ -1,51 +0,0 @@ -# -*- coding: utf-8 -*- -# Copyright 2022 Google LLC -# -# Licensed under the Apache License, Version 2.0 (the "License"); -# you may not use this file except in compliance with the License. -# You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. -# -# Generated code. DO NOT EDIT! -# -# Snippet for GetSnapshot -# NOTE: This snippet has been automatically generated for illustrative purposes only. -# It may require modifications to work in your environment. - -# To install the latest published package dependency, execute the following: -# python3 -m pip install google-cloud-dataflow-client - - -# [START dataflow_v1beta3_generated_SnapshotsV1Beta3_GetSnapshot_sync_ba1ca2b6] -# This snippet has been automatically generated and should be regarded as a -# code template only. -# It will require modifications to work: -# - It may require correct/in-range values for request initialization. -# - It may require specifying regional endpoints when creating the service -# client as shown in: -# https://googleapis.dev/python/google-api-core/latest/client_options.html -from google.cloud import dataflow_v1beta3 - - -def sample_get_snapshot(): - # Create a client - client = dataflow_v1beta3.SnapshotsV1Beta3Client() - - # Initialize request argument(s) - request = dataflow_v1beta3.GetSnapshotRequest( - ) - - # Make the request - response = client.get_snapshot(request=request) - - # Handle the response - print(response) - -# [END dataflow_v1beta3_generated_SnapshotsV1Beta3_GetSnapshot_sync_ba1ca2b6] diff --git a/samples/generated_samples/dataflow_v1beta3_generated_snapshots_v1_beta3_list_snapshots_sync_425e5024.py b/samples/generated_samples/dataflow_v1beta3_generated_snapshots_v1_beta3_list_snapshots_sync.py similarity index 98% rename from samples/generated_samples/dataflow_v1beta3_generated_snapshots_v1_beta3_list_snapshots_sync_425e5024.py rename to samples/generated_samples/dataflow_v1beta3_generated_snapshots_v1_beta3_list_snapshots_sync.py index d6955d2..bdd2529 100644 --- a/samples/generated_samples/dataflow_v1beta3_generated_snapshots_v1_beta3_list_snapshots_sync_425e5024.py +++ b/samples/generated_samples/dataflow_v1beta3_generated_snapshots_v1_beta3_list_snapshots_sync.py @@ -23,7 +23,7 @@ # python3 -m pip install google-cloud-dataflow-client -# [START dataflow_v1beta3_generated_SnapshotsV1Beta3_ListSnapshots_sync_425e5024] +# [START dataflow_v1beta3_generated_SnapshotsV1Beta3_ListSnapshots_sync] # This snippet has been automatically generated and should be regarded as a # code template only. # It will require modifications to work: @@ -48,4 +48,4 @@ def sample_list_snapshots(): # Handle the response print(response) -# [END dataflow_v1beta3_generated_SnapshotsV1Beta3_ListSnapshots_sync_425e5024] +# [END dataflow_v1beta3_generated_SnapshotsV1Beta3_ListSnapshots_sync] diff --git a/samples/generated_samples/dataflow_v1beta3_generated_snapshots_v1_beta3_list_snapshots_sync_6da08cb9.py b/samples/generated_samples/dataflow_v1beta3_generated_snapshots_v1_beta3_list_snapshots_sync_6da08cb9.py deleted file mode 100644 index a8fdc15..0000000 --- a/samples/generated_samples/dataflow_v1beta3_generated_snapshots_v1_beta3_list_snapshots_sync_6da08cb9.py +++ /dev/null @@ -1,51 +0,0 @@ -# -*- coding: utf-8 -*- -# Copyright 2022 Google LLC -# -# Licensed under the Apache License, Version 2.0 (the "License"); -# you may not use this file except in compliance with the License. -# You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. -# -# Generated code. DO NOT EDIT! -# -# Snippet for ListSnapshots -# NOTE: This snippet has been automatically generated for illustrative purposes only. -# It may require modifications to work in your environment. - -# To install the latest published package dependency, execute the following: -# python3 -m pip install google-cloud-dataflow-client - - -# [START dataflow_v1beta3_generated_SnapshotsV1Beta3_ListSnapshots_sync_6da08cb9] -# This snippet has been automatically generated and should be regarded as a -# code template only. -# It will require modifications to work: -# - It may require correct/in-range values for request initialization. -# - It may require specifying regional endpoints when creating the service -# client as shown in: -# https://googleapis.dev/python/google-api-core/latest/client_options.html -from google.cloud import dataflow_v1beta3 - - -def sample_list_snapshots(): - # Create a client - client = dataflow_v1beta3.SnapshotsV1Beta3Client() - - # Initialize request argument(s) - request = dataflow_v1beta3.ListSnapshotsRequest( - ) - - # Make the request - response = client.list_snapshots(request=request) - - # Handle the response - print(response) - -# [END dataflow_v1beta3_generated_SnapshotsV1Beta3_ListSnapshots_sync_6da08cb9] diff --git a/samples/generated_samples/dataflow_v1beta3_generated_templates_service_create_job_from_template_sync_9e26a4bf.py b/samples/generated_samples/dataflow_v1beta3_generated_templates_service_create_job_from_template_sync.py similarity index 97% rename from samples/generated_samples/dataflow_v1beta3_generated_templates_service_create_job_from_template_sync_9e26a4bf.py rename to samples/generated_samples/dataflow_v1beta3_generated_templates_service_create_job_from_template_sync.py index 3c0a699..175c180 100644 --- a/samples/generated_samples/dataflow_v1beta3_generated_templates_service_create_job_from_template_sync_9e26a4bf.py +++ b/samples/generated_samples/dataflow_v1beta3_generated_templates_service_create_job_from_template_sync.py @@ -23,7 +23,7 @@ # python3 -m pip install google-cloud-dataflow-client -# [START dataflow_v1beta3_generated_TemplatesService_CreateJobFromTemplate_sync_9e26a4bf] +# [START dataflow_v1beta3_generated_TemplatesService_CreateJobFromTemplate_sync] # This snippet has been automatically generated and should be regarded as a # code template only. # It will require modifications to work: @@ -49,4 +49,4 @@ def sample_create_job_from_template(): # Handle the response print(response) -# [END dataflow_v1beta3_generated_TemplatesService_CreateJobFromTemplate_sync_9e26a4bf] +# [END dataflow_v1beta3_generated_TemplatesService_CreateJobFromTemplate_sync] diff --git a/samples/generated_samples/dataflow_v1beta3_generated_templates_service_create_job_from_template_sync_e2cb482f.py b/samples/generated_samples/dataflow_v1beta3_generated_templates_service_create_job_from_template_sync_e2cb482f.py deleted file mode 100644 index 2470a87..0000000 --- a/samples/generated_samples/dataflow_v1beta3_generated_templates_service_create_job_from_template_sync_e2cb482f.py +++ /dev/null @@ -1,52 +0,0 @@ -# -*- coding: utf-8 -*- -# Copyright 2022 Google LLC -# -# Licensed under the Apache License, Version 2.0 (the "License"); -# you may not use this file except in compliance with the License. -# You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. -# -# Generated code. DO NOT EDIT! -# -# Snippet for CreateJobFromTemplate -# NOTE: This snippet has been automatically generated for illustrative purposes only. -# It may require modifications to work in your environment. - -# To install the latest published package dependency, execute the following: -# python3 -m pip install google-cloud-dataflow-client - - -# [START dataflow_v1beta3_generated_TemplatesService_CreateJobFromTemplate_sync_e2cb482f] -# This snippet has been automatically generated and should be regarded as a -# code template only. -# It will require modifications to work: -# - It may require correct/in-range values for request initialization. -# - It may require specifying regional endpoints when creating the service -# client as shown in: -# https://googleapis.dev/python/google-api-core/latest/client_options.html -from google.cloud import dataflow_v1beta3 - - -def sample_create_job_from_template(): - # Create a client - client = dataflow_v1beta3.TemplatesServiceClient() - - # Initialize request argument(s) - request = dataflow_v1beta3.CreateJobFromTemplateRequest( - gcs_path="gcs_path_value", - ) - - # Make the request - response = client.create_job_from_template(request=request) - - # Handle the response - print(response) - -# [END dataflow_v1beta3_generated_TemplatesService_CreateJobFromTemplate_sync_e2cb482f] diff --git a/samples/generated_samples/dataflow_v1beta3_generated_templates_service_get_template_sync_07cd261a.py b/samples/generated_samples/dataflow_v1beta3_generated_templates_service_get_template_sync.py similarity index 98% rename from samples/generated_samples/dataflow_v1beta3_generated_templates_service_get_template_sync_07cd261a.py rename to samples/generated_samples/dataflow_v1beta3_generated_templates_service_get_template_sync.py index 9691908..c7f9cae 100644 --- a/samples/generated_samples/dataflow_v1beta3_generated_templates_service_get_template_sync_07cd261a.py +++ b/samples/generated_samples/dataflow_v1beta3_generated_templates_service_get_template_sync.py @@ -23,7 +23,7 @@ # python3 -m pip install google-cloud-dataflow-client -# [START dataflow_v1beta3_generated_TemplatesService_GetTemplate_sync_07cd261a] +# [START dataflow_v1beta3_generated_TemplatesService_GetTemplate_sync] # This snippet has been automatically generated and should be regarded as a # code template only. # It will require modifications to work: @@ -49,4 +49,4 @@ def sample_get_template(): # Handle the response print(response) -# [END dataflow_v1beta3_generated_TemplatesService_GetTemplate_sync_07cd261a] +# [END dataflow_v1beta3_generated_TemplatesService_GetTemplate_sync] diff --git a/samples/generated_samples/dataflow_v1beta3_generated_templates_service_get_template_sync_732e6209.py b/samples/generated_samples/dataflow_v1beta3_generated_templates_service_get_template_sync_732e6209.py deleted file mode 100644 index 88cf605..0000000 --- a/samples/generated_samples/dataflow_v1beta3_generated_templates_service_get_template_sync_732e6209.py +++ /dev/null @@ -1,52 +0,0 @@ -# -*- coding: utf-8 -*- -# Copyright 2022 Google LLC -# -# Licensed under the Apache License, Version 2.0 (the "License"); -# you may not use this file except in compliance with the License. -# You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. -# -# Generated code. DO NOT EDIT! -# -# Snippet for GetTemplate -# NOTE: This snippet has been automatically generated for illustrative purposes only. -# It may require modifications to work in your environment. - -# To install the latest published package dependency, execute the following: -# python3 -m pip install google-cloud-dataflow-client - - -# [START dataflow_v1beta3_generated_TemplatesService_GetTemplate_sync_732e6209] -# This snippet has been automatically generated and should be regarded as a -# code template only. -# It will require modifications to work: -# - It may require correct/in-range values for request initialization. -# - It may require specifying regional endpoints when creating the service -# client as shown in: -# https://googleapis.dev/python/google-api-core/latest/client_options.html -from google.cloud import dataflow_v1beta3 - - -def sample_get_template(): - # Create a client - client = dataflow_v1beta3.TemplatesServiceClient() - - # Initialize request argument(s) - request = dataflow_v1beta3.GetTemplateRequest( - gcs_path="gcs_path_value", - ) - - # Make the request - response = client.get_template(request=request) - - # Handle the response - print(response) - -# [END dataflow_v1beta3_generated_TemplatesService_GetTemplate_sync_732e6209] diff --git a/samples/generated_samples/dataflow_v1beta3_generated_templates_service_launch_template_sync_140179ca.py b/samples/generated_samples/dataflow_v1beta3_generated_templates_service_launch_template_sync.py similarity index 98% rename from samples/generated_samples/dataflow_v1beta3_generated_templates_service_launch_template_sync_140179ca.py rename to samples/generated_samples/dataflow_v1beta3_generated_templates_service_launch_template_sync.py index 98b969f..7f1f81d 100644 --- a/samples/generated_samples/dataflow_v1beta3_generated_templates_service_launch_template_sync_140179ca.py +++ b/samples/generated_samples/dataflow_v1beta3_generated_templates_service_launch_template_sync.py @@ -23,7 +23,7 @@ # python3 -m pip install google-cloud-dataflow-client -# [START dataflow_v1beta3_generated_TemplatesService_LaunchTemplate_sync_140179ca] +# [START dataflow_v1beta3_generated_TemplatesService_LaunchTemplate_sync] # This snippet has been automatically generated and should be regarded as a # code template only. # It will require modifications to work: @@ -49,4 +49,4 @@ def sample_launch_template(): # Handle the response print(response) -# [END dataflow_v1beta3_generated_TemplatesService_LaunchTemplate_sync_140179ca] +# [END dataflow_v1beta3_generated_TemplatesService_LaunchTemplate_sync] diff --git a/samples/generated_samples/dataflow_v1beta3_generated_templates_service_launch_template_sync_77764eb9.py b/samples/generated_samples/dataflow_v1beta3_generated_templates_service_launch_template_sync_77764eb9.py deleted file mode 100644 index 74d62b9..0000000 --- a/samples/generated_samples/dataflow_v1beta3_generated_templates_service_launch_template_sync_77764eb9.py +++ /dev/null @@ -1,52 +0,0 @@ -# -*- coding: utf-8 -*- -# Copyright 2022 Google LLC -# -# Licensed under the Apache License, Version 2.0 (the "License"); -# you may not use this file except in compliance with the License. -# You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. -# -# Generated code. DO NOT EDIT! -# -# Snippet for LaunchTemplate -# NOTE: This snippet has been automatically generated for illustrative purposes only. -# It may require modifications to work in your environment. - -# To install the latest published package dependency, execute the following: -# python3 -m pip install google-cloud-dataflow-client - - -# [START dataflow_v1beta3_generated_TemplatesService_LaunchTemplate_sync_77764eb9] -# This snippet has been automatically generated and should be regarded as a -# code template only. -# It will require modifications to work: -# - It may require correct/in-range values for request initialization. -# - It may require specifying regional endpoints when creating the service -# client as shown in: -# https://googleapis.dev/python/google-api-core/latest/client_options.html -from google.cloud import dataflow_v1beta3 - - -def sample_launch_template(): - # Create a client - client = dataflow_v1beta3.TemplatesServiceClient() - - # Initialize request argument(s) - request = dataflow_v1beta3.LaunchTemplateRequest( - gcs_path="gcs_path_value", - ) - - # Make the request - response = client.launch_template(request=request) - - # Handle the response - print(response) - -# [END dataflow_v1beta3_generated_TemplatesService_LaunchTemplate_sync_77764eb9] diff --git a/samples/generated_samples/snippet_metadata_dataflow_v1beta3.json b/samples/generated_samples/snippet_metadata_google.dataflow.v1beta3.json similarity index 65% rename from samples/generated_samples/snippet_metadata_dataflow_v1beta3.json rename to samples/generated_samples/snippet_metadata_google.dataflow.v1beta3.json index a96a7a2..5f2c89a 100644 --- a/samples/generated_samples/snippet_metadata_dataflow_v1beta3.json +++ b/samples/generated_samples/snippet_metadata_google.dataflow.v1beta3.json @@ -7,7 +7,8 @@ } ], "language": "PYTHON", - "name": "google-cloud-dataflow-client" + "name": "google-cloud-dataflow-client", + "version": "0.7.0" }, "snippets": [ { @@ -125,7 +126,7 @@ "shortName": "launch_flex_template" }, "description": "Sample for LaunchFlexTemplate", - "file": "dataflow_v1beta3_generated_flex_templates_service_launch_flex_template_sync_a4f75f91.py", + "file": "dataflow_v1beta3_generated_flex_templates_service_launch_flex_template_sync.py", "language": "PYTHON", "origin": "API_DEFINITION", "regionTag": "dataflow_v1beta3_generated_FlexTemplatesService_LaunchFlexTemplate_sync", @@ -161,28 +162,29 @@ "type": "RESPONSE_HANDLING" } ], - "title": "dataflow_v1beta3_generated_flex_templates_service_launch_flex_template_sync_a4f75f91.py" + "title": "dataflow_v1beta3_generated_flex_templates_service_launch_flex_template_sync.py" }, { "canonical": true, "clientMethod": { + "async": true, "client": { - "fullName": "google.cloud.dataflow_v1beta3.FlexTemplatesServiceClient", - "shortName": "FlexTemplatesServiceClient" + "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3AsyncClient", + "shortName": "JobsV1Beta3AsyncClient" }, - "fullName": "google.cloud.dataflow_v1beta3.FlexTemplatesServiceClient.launch_flex_template", + "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3AsyncClient.aggregated_list_jobs", "method": { - "fullName": "google.dataflow.v1beta3.FlexTemplatesService.LaunchFlexTemplate", + "fullName": "google.dataflow.v1beta3.JobsV1Beta3.AggregatedListJobs", "service": { - "fullName": "google.dataflow.v1beta3.FlexTemplatesService", - "shortName": "FlexTemplatesService" + "fullName": "google.dataflow.v1beta3.JobsV1Beta3", + "shortName": "JobsV1Beta3" }, - "shortName": "LaunchFlexTemplate" + "shortName": "AggregatedListJobs" }, "parameters": [ { "name": "request", - "type": "google.cloud.dataflow_v1beta3.types.LaunchFlexTemplateRequest" + "type": "google.cloud.dataflow_v1beta3.types.ListJobsRequest" }, { "name": "retry", @@ -197,22 +199,22 @@ "type": "Sequence[Tuple[str, str]" } ], - "resultType": "google.cloud.dataflow_v1beta3.types.LaunchFlexTemplateResponse", - "shortName": "launch_flex_template" + "resultType": "google.cloud.dataflow_v1beta3.services.jobs_v1_beta3.pagers.AggregatedListJobsAsyncPager", + "shortName": "aggregated_list_jobs" }, - "description": "Sample for LaunchFlexTemplate", - "file": "dataflow_v1beta3_generated_flex_templates_service_launch_flex_template_sync_4148a07f.py", + "description": "Sample for AggregatedListJobs", + "file": "dataflow_v1beta3_generated_jobs_v1_beta3_aggregated_list_jobs_async.py", "language": "PYTHON", "origin": "API_DEFINITION", - "regionTag": "dataflow_v1beta3_generated_FlexTemplatesService_LaunchFlexTemplate_sync", + "regionTag": "dataflow_v1beta3_generated_JobsV1Beta3_AggregatedListJobs_async", "segments": [ { - "end": 50, + "end": 51, "start": 27, "type": "FULL" }, { - "end": 50, + "end": 51, "start": 27, "type": "SHORT" }, @@ -232,22 +234,21 @@ "type": "REQUEST_EXECUTION" }, { - "end": 51, + "end": 52, "start": 48, "type": "RESPONSE_HANDLING" } ], - "title": "dataflow_v1beta3_generated_flex_templates_service_launch_flex_template_sync_4148a07f.py" + "title": "dataflow_v1beta3_generated_jobs_v1_beta3_aggregated_list_jobs_async.py" }, { "canonical": true, "clientMethod": { - "async": true, "client": { - "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3AsyncClient", - "shortName": "JobsV1Beta3AsyncClient" + "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3Client", + "shortName": "JobsV1Beta3Client" }, - "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3AsyncClient.aggregated_list_jobs", + "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3Client.aggregated_list_jobs", "method": { "fullName": "google.dataflow.v1beta3.JobsV1Beta3.AggregatedListJobs", "service": { @@ -274,14 +275,14 @@ "type": "Sequence[Tuple[str, str]" } ], - "resultType": "google.cloud.dataflow_v1beta3.services.jobs_v1_beta3.pagers.AggregatedListJobsAsyncPager", + "resultType": "google.cloud.dataflow_v1beta3.services.jobs_v1_beta3.pagers.AggregatedListJobsPager", "shortName": "aggregated_list_jobs" }, "description": "Sample for AggregatedListJobs", - "file": "dataflow_v1beta3_generated_jobs_v1_beta3_aggregated_list_jobs_async.py", + "file": "dataflow_v1beta3_generated_jobs_v1_beta3_aggregated_list_jobs_sync.py", "language": "PYTHON", "origin": "API_DEFINITION", - "regionTag": "dataflow_v1beta3_generated_JobsV1Beta3_AggregatedListJobs_async", + "regionTag": "dataflow_v1beta3_generated_JobsV1Beta3_AggregatedListJobs_sync", "segments": [ { "end": 51, @@ -314,28 +315,29 @@ "type": "RESPONSE_HANDLING" } ], - "title": "dataflow_v1beta3_generated_jobs_v1_beta3_aggregated_list_jobs_async.py" + "title": "dataflow_v1beta3_generated_jobs_v1_beta3_aggregated_list_jobs_sync.py" }, { "canonical": true, "clientMethod": { + "async": true, "client": { - "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3Client", - "shortName": "JobsV1Beta3Client" + "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3AsyncClient", + "shortName": "JobsV1Beta3AsyncClient" }, - "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3Client.aggregated_list_jobs", + "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3AsyncClient.check_active_jobs", "method": { - "fullName": "google.dataflow.v1beta3.JobsV1Beta3.AggregatedListJobs", + "fullName": "google.dataflow.v1beta3.JobsV1Beta3.CheckActiveJobs", "service": { "fullName": "google.dataflow.v1beta3.JobsV1Beta3", "shortName": "JobsV1Beta3" }, - "shortName": "AggregatedListJobs" + "shortName": "CheckActiveJobs" }, "parameters": [ { "name": "request", - "type": "google.cloud.dataflow_v1beta3.types.ListJobsRequest" + "type": "google.cloud.dataflow_v1beta3.types.CheckActiveJobsRequest" }, { "name": "retry", @@ -350,22 +352,22 @@ "type": "Sequence[Tuple[str, str]" } ], - "resultType": "google.cloud.dataflow_v1beta3.services.jobs_v1_beta3.pagers.AggregatedListJobsPager", - "shortName": "aggregated_list_jobs" + "resultType": "google.cloud.dataflow_v1beta3.types.CheckActiveJobsResponse", + "shortName": "check_active_jobs" }, - "description": "Sample for AggregatedListJobs", - "file": "dataflow_v1beta3_generated_jobs_v1_beta3_aggregated_list_jobs_sync_0d901b38.py", + "description": "Sample for CheckActiveJobs", + "file": "dataflow_v1beta3_generated_jobs_v1_beta3_check_active_jobs_async.py", "language": "PYTHON", "origin": "API_DEFINITION", - "regionTag": "dataflow_v1beta3_generated_JobsV1Beta3_AggregatedListJobs_sync", + "regionTag": "dataflow_v1beta3_generated_JobsV1Beta3_CheckActiveJobs_async", "segments": [ { - "end": 51, + "end": 50, "start": 27, "type": "FULL" }, { - "end": 51, + "end": 50, "start": 27, "type": "SHORT" }, @@ -385,12 +387,12 @@ "type": "REQUEST_EXECUTION" }, { - "end": 52, + "end": 51, "start": 48, "type": "RESPONSE_HANDLING" } ], - "title": "dataflow_v1beta3_generated_jobs_v1_beta3_aggregated_list_jobs_sync_0d901b38.py" + "title": "dataflow_v1beta3_generated_jobs_v1_beta3_check_active_jobs_async.py" }, { "canonical": true, @@ -399,19 +401,19 @@ "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3Client", "shortName": "JobsV1Beta3Client" }, - "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3Client.aggregated_list_jobs", + "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3Client.check_active_jobs", "method": { - "fullName": "google.dataflow.v1beta3.JobsV1Beta3.AggregatedListJobs", + "fullName": "google.dataflow.v1beta3.JobsV1Beta3.CheckActiveJobs", "service": { "fullName": "google.dataflow.v1beta3.JobsV1Beta3", "shortName": "JobsV1Beta3" }, - "shortName": "AggregatedListJobs" + "shortName": "CheckActiveJobs" }, "parameters": [ { "name": "request", - "type": "google.cloud.dataflow_v1beta3.types.ListJobsRequest" + "type": "google.cloud.dataflow_v1beta3.types.CheckActiveJobsRequest" }, { "name": "retry", @@ -426,22 +428,22 @@ "type": "Sequence[Tuple[str, str]" } ], - "resultType": "google.cloud.dataflow_v1beta3.services.jobs_v1_beta3.pagers.AggregatedListJobsPager", - "shortName": "aggregated_list_jobs" + "resultType": "google.cloud.dataflow_v1beta3.types.CheckActiveJobsResponse", + "shortName": "check_active_jobs" }, - "description": "Sample for AggregatedListJobs", - "file": "dataflow_v1beta3_generated_jobs_v1_beta3_aggregated_list_jobs_sync_26f07383.py", + "description": "Sample for CheckActiveJobs", + "file": "dataflow_v1beta3_generated_jobs_v1_beta3_check_active_jobs_sync.py", "language": "PYTHON", "origin": "API_DEFINITION", - "regionTag": "dataflow_v1beta3_generated_JobsV1Beta3_AggregatedListJobs_sync", + "regionTag": "dataflow_v1beta3_generated_JobsV1Beta3_CheckActiveJobs_sync", "segments": [ { - "end": 51, + "end": 50, "start": 27, "type": "FULL" }, { - "end": 51, + "end": 50, "start": 27, "type": "SHORT" }, @@ -461,12 +463,12 @@ "type": "REQUEST_EXECUTION" }, { - "end": 52, + "end": 51, "start": 48, "type": "RESPONSE_HANDLING" } ], - "title": "dataflow_v1beta3_generated_jobs_v1_beta3_aggregated_list_jobs_sync_26f07383.py" + "title": "dataflow_v1beta3_generated_jobs_v1_beta3_check_active_jobs_sync.py" }, { "canonical": true, @@ -476,19 +478,19 @@ "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3AsyncClient", "shortName": "JobsV1Beta3AsyncClient" }, - "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3AsyncClient.check_active_jobs", + "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3AsyncClient.create_job", "method": { - "fullName": "google.dataflow.v1beta3.JobsV1Beta3.CheckActiveJobs", + "fullName": "google.dataflow.v1beta3.JobsV1Beta3.CreateJob", "service": { "fullName": "google.dataflow.v1beta3.JobsV1Beta3", "shortName": "JobsV1Beta3" }, - "shortName": "CheckActiveJobs" + "shortName": "CreateJob" }, "parameters": [ { "name": "request", - "type": "google.cloud.dataflow_v1beta3.types.CheckActiveJobsRequest" + "type": "google.cloud.dataflow_v1beta3.types.CreateJobRequest" }, { "name": "retry", @@ -503,14 +505,14 @@ "type": "Sequence[Tuple[str, str]" } ], - "resultType": "google.cloud.dataflow_v1beta3.types.CheckActiveJobsResponse", - "shortName": "check_active_jobs" + "resultType": "google.cloud.dataflow_v1beta3.types.Job", + "shortName": "create_job" }, - "description": "Sample for CheckActiveJobs", - "file": "dataflow_v1beta3_generated_jobs_v1_beta3_check_active_jobs_async.py", + "description": "Sample for CreateJob", + "file": "dataflow_v1beta3_generated_jobs_v1_beta3_create_job_async.py", "language": "PYTHON", "origin": "API_DEFINITION", - "regionTag": "dataflow_v1beta3_generated_JobsV1Beta3_CheckActiveJobs_async", + "regionTag": "dataflow_v1beta3_generated_JobsV1Beta3_CreateJob_async", "segments": [ { "end": 50, @@ -543,7 +545,7 @@ "type": "RESPONSE_HANDLING" } ], - "title": "dataflow_v1beta3_generated_jobs_v1_beta3_check_active_jobs_async.py" + "title": "dataflow_v1beta3_generated_jobs_v1_beta3_create_job_async.py" }, { "canonical": true, @@ -552,19 +554,19 @@ "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3Client", "shortName": "JobsV1Beta3Client" }, - "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3Client.check_active_jobs", + "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3Client.create_job", "method": { - "fullName": "google.dataflow.v1beta3.JobsV1Beta3.CheckActiveJobs", + "fullName": "google.dataflow.v1beta3.JobsV1Beta3.CreateJob", "service": { "fullName": "google.dataflow.v1beta3.JobsV1Beta3", "shortName": "JobsV1Beta3" }, - "shortName": "CheckActiveJobs" + "shortName": "CreateJob" }, "parameters": [ { "name": "request", - "type": "google.cloud.dataflow_v1beta3.types.CheckActiveJobsRequest" + "type": "google.cloud.dataflow_v1beta3.types.CreateJobRequest" }, { "name": "retry", @@ -579,14 +581,14 @@ "type": "Sequence[Tuple[str, str]" } ], - "resultType": "google.cloud.dataflow_v1beta3.types.CheckActiveJobsResponse", - "shortName": "check_active_jobs" + "resultType": "google.cloud.dataflow_v1beta3.types.Job", + "shortName": "create_job" }, - "description": "Sample for CheckActiveJobs", - "file": "dataflow_v1beta3_generated_jobs_v1_beta3_check_active_jobs_sync_aab49b35.py", + "description": "Sample for CreateJob", + "file": "dataflow_v1beta3_generated_jobs_v1_beta3_create_job_sync.py", "language": "PYTHON", "origin": "API_DEFINITION", - "regionTag": "dataflow_v1beta3_generated_JobsV1Beta3_CheckActiveJobs_sync", + "regionTag": "dataflow_v1beta3_generated_JobsV1Beta3_CreateJob_sync", "segments": [ { "end": 50, @@ -619,28 +621,29 @@ "type": "RESPONSE_HANDLING" } ], - "title": "dataflow_v1beta3_generated_jobs_v1_beta3_check_active_jobs_sync_aab49b35.py" + "title": "dataflow_v1beta3_generated_jobs_v1_beta3_create_job_sync.py" }, { "canonical": true, "clientMethod": { + "async": true, "client": { - "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3Client", - "shortName": "JobsV1Beta3Client" + "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3AsyncClient", + "shortName": "JobsV1Beta3AsyncClient" }, - "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3Client.check_active_jobs", + "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3AsyncClient.get_job", "method": { - "fullName": "google.dataflow.v1beta3.JobsV1Beta3.CheckActiveJobs", + "fullName": "google.dataflow.v1beta3.JobsV1Beta3.GetJob", "service": { "fullName": "google.dataflow.v1beta3.JobsV1Beta3", "shortName": "JobsV1Beta3" }, - "shortName": "CheckActiveJobs" + "shortName": "GetJob" }, "parameters": [ { "name": "request", - "type": "google.cloud.dataflow_v1beta3.types.CheckActiveJobsRequest" + "type": "google.cloud.dataflow_v1beta3.types.GetJobRequest" }, { "name": "retry", @@ -655,14 +658,14 @@ "type": "Sequence[Tuple[str, str]" } ], - "resultType": "google.cloud.dataflow_v1beta3.types.CheckActiveJobsResponse", - "shortName": "check_active_jobs" + "resultType": "google.cloud.dataflow_v1beta3.types.Job", + "shortName": "get_job" }, - "description": "Sample for CheckActiveJobs", - "file": "dataflow_v1beta3_generated_jobs_v1_beta3_check_active_jobs_sync_baed931b.py", + "description": "Sample for GetJob", + "file": "dataflow_v1beta3_generated_jobs_v1_beta3_get_job_async.py", "language": "PYTHON", "origin": "API_DEFINITION", - "regionTag": "dataflow_v1beta3_generated_JobsV1Beta3_CheckActiveJobs_sync", + "regionTag": "dataflow_v1beta3_generated_JobsV1Beta3_GetJob_async", "segments": [ { "end": 50, @@ -695,29 +698,28 @@ "type": "RESPONSE_HANDLING" } ], - "title": "dataflow_v1beta3_generated_jobs_v1_beta3_check_active_jobs_sync_baed931b.py" + "title": "dataflow_v1beta3_generated_jobs_v1_beta3_get_job_async.py" }, { "canonical": true, "clientMethod": { - "async": true, "client": { - "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3AsyncClient", - "shortName": "JobsV1Beta3AsyncClient" + "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3Client", + "shortName": "JobsV1Beta3Client" }, - "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3AsyncClient.create_job", + "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3Client.get_job", "method": { - "fullName": "google.dataflow.v1beta3.JobsV1Beta3.CreateJob", + "fullName": "google.dataflow.v1beta3.JobsV1Beta3.GetJob", "service": { "fullName": "google.dataflow.v1beta3.JobsV1Beta3", "shortName": "JobsV1Beta3" }, - "shortName": "CreateJob" + "shortName": "GetJob" }, "parameters": [ { "name": "request", - "type": "google.cloud.dataflow_v1beta3.types.CreateJobRequest" + "type": "google.cloud.dataflow_v1beta3.types.GetJobRequest" }, { "name": "retry", @@ -733,13 +735,13 @@ } ], "resultType": "google.cloud.dataflow_v1beta3.types.Job", - "shortName": "create_job" + "shortName": "get_job" }, - "description": "Sample for CreateJob", - "file": "dataflow_v1beta3_generated_jobs_v1_beta3_create_job_async.py", + "description": "Sample for GetJob", + "file": "dataflow_v1beta3_generated_jobs_v1_beta3_get_job_sync.py", "language": "PYTHON", "origin": "API_DEFINITION", - "regionTag": "dataflow_v1beta3_generated_JobsV1Beta3_CreateJob_async", + "regionTag": "dataflow_v1beta3_generated_JobsV1Beta3_GetJob_sync", "segments": [ { "end": 50, @@ -772,28 +774,29 @@ "type": "RESPONSE_HANDLING" } ], - "title": "dataflow_v1beta3_generated_jobs_v1_beta3_create_job_async.py" + "title": "dataflow_v1beta3_generated_jobs_v1_beta3_get_job_sync.py" }, { "canonical": true, "clientMethod": { + "async": true, "client": { - "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3Client", - "shortName": "JobsV1Beta3Client" + "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3AsyncClient", + "shortName": "JobsV1Beta3AsyncClient" }, - "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3Client.create_job", + "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3AsyncClient.list_jobs", "method": { - "fullName": "google.dataflow.v1beta3.JobsV1Beta3.CreateJob", + "fullName": "google.dataflow.v1beta3.JobsV1Beta3.ListJobs", "service": { "fullName": "google.dataflow.v1beta3.JobsV1Beta3", "shortName": "JobsV1Beta3" }, - "shortName": "CreateJob" + "shortName": "ListJobs" }, "parameters": [ { "name": "request", - "type": "google.cloud.dataflow_v1beta3.types.CreateJobRequest" + "type": "google.cloud.dataflow_v1beta3.types.ListJobsRequest" }, { "name": "retry", @@ -808,22 +811,22 @@ "type": "Sequence[Tuple[str, str]" } ], - "resultType": "google.cloud.dataflow_v1beta3.types.Job", - "shortName": "create_job" + "resultType": "google.cloud.dataflow_v1beta3.services.jobs_v1_beta3.pagers.ListJobsAsyncPager", + "shortName": "list_jobs" }, - "description": "Sample for CreateJob", - "file": "dataflow_v1beta3_generated_jobs_v1_beta3_create_job_sync_ad13d605.py", + "description": "Sample for ListJobs", + "file": "dataflow_v1beta3_generated_jobs_v1_beta3_list_jobs_async.py", "language": "PYTHON", "origin": "API_DEFINITION", - "regionTag": "dataflow_v1beta3_generated_JobsV1Beta3_CreateJob_sync", + "regionTag": "dataflow_v1beta3_generated_JobsV1Beta3_ListJobs_async", "segments": [ { - "end": 50, + "end": 51, "start": 27, "type": "FULL" }, { - "end": 50, + "end": 51, "start": 27, "type": "SHORT" }, @@ -843,12 +846,12 @@ "type": "REQUEST_EXECUTION" }, { - "end": 51, + "end": 52, "start": 48, "type": "RESPONSE_HANDLING" } ], - "title": "dataflow_v1beta3_generated_jobs_v1_beta3_create_job_sync_ad13d605.py" + "title": "dataflow_v1beta3_generated_jobs_v1_beta3_list_jobs_async.py" }, { "canonical": true, @@ -857,19 +860,19 @@ "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3Client", "shortName": "JobsV1Beta3Client" }, - "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3Client.create_job", + "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3Client.list_jobs", "method": { - "fullName": "google.dataflow.v1beta3.JobsV1Beta3.CreateJob", + "fullName": "google.dataflow.v1beta3.JobsV1Beta3.ListJobs", "service": { "fullName": "google.dataflow.v1beta3.JobsV1Beta3", "shortName": "JobsV1Beta3" }, - "shortName": "CreateJob" + "shortName": "ListJobs" }, "parameters": [ { "name": "request", - "type": "google.cloud.dataflow_v1beta3.types.CreateJobRequest" + "type": "google.cloud.dataflow_v1beta3.types.ListJobsRequest" }, { "name": "retry", @@ -884,22 +887,22 @@ "type": "Sequence[Tuple[str, str]" } ], - "resultType": "google.cloud.dataflow_v1beta3.types.Job", - "shortName": "create_job" + "resultType": "google.cloud.dataflow_v1beta3.services.jobs_v1_beta3.pagers.ListJobsPager", + "shortName": "list_jobs" }, - "description": "Sample for CreateJob", - "file": "dataflow_v1beta3_generated_jobs_v1_beta3_create_job_sync_2a1b0208.py", + "description": "Sample for ListJobs", + "file": "dataflow_v1beta3_generated_jobs_v1_beta3_list_jobs_sync.py", "language": "PYTHON", "origin": "API_DEFINITION", - "regionTag": "dataflow_v1beta3_generated_JobsV1Beta3_CreateJob_sync", + "regionTag": "dataflow_v1beta3_generated_JobsV1Beta3_ListJobs_sync", "segments": [ { - "end": 50, + "end": 51, "start": 27, "type": "FULL" }, { - "end": 50, + "end": 51, "start": 27, "type": "SHORT" }, @@ -919,12 +922,12 @@ "type": "REQUEST_EXECUTION" }, { - "end": 51, + "end": 52, "start": 48, "type": "RESPONSE_HANDLING" } ], - "title": "dataflow_v1beta3_generated_jobs_v1_beta3_create_job_sync_2a1b0208.py" + "title": "dataflow_v1beta3_generated_jobs_v1_beta3_list_jobs_sync.py" }, { "canonical": true, @@ -934,19 +937,19 @@ "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3AsyncClient", "shortName": "JobsV1Beta3AsyncClient" }, - "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3AsyncClient.get_job", + "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3AsyncClient.snapshot_job", "method": { - "fullName": "google.dataflow.v1beta3.JobsV1Beta3.GetJob", + "fullName": "google.dataflow.v1beta3.JobsV1Beta3.SnapshotJob", "service": { "fullName": "google.dataflow.v1beta3.JobsV1Beta3", "shortName": "JobsV1Beta3" }, - "shortName": "GetJob" + "shortName": "SnapshotJob" }, "parameters": [ { "name": "request", - "type": "google.cloud.dataflow_v1beta3.types.GetJobRequest" + "type": "google.cloud.dataflow_v1beta3.types.SnapshotJobRequest" }, { "name": "retry", @@ -961,14 +964,14 @@ "type": "Sequence[Tuple[str, str]" } ], - "resultType": "google.cloud.dataflow_v1beta3.types.Job", - "shortName": "get_job" + "resultType": "google.cloud.dataflow_v1beta3.types.Snapshot", + "shortName": "snapshot_job" }, - "description": "Sample for GetJob", - "file": "dataflow_v1beta3_generated_jobs_v1_beta3_get_job_async.py", + "description": "Sample for SnapshotJob", + "file": "dataflow_v1beta3_generated_jobs_v1_beta3_snapshot_job_async.py", "language": "PYTHON", "origin": "API_DEFINITION", - "regionTag": "dataflow_v1beta3_generated_JobsV1Beta3_GetJob_async", + "regionTag": "dataflow_v1beta3_generated_JobsV1Beta3_SnapshotJob_async", "segments": [ { "end": 50, @@ -1001,7 +1004,7 @@ "type": "RESPONSE_HANDLING" } ], - "title": "dataflow_v1beta3_generated_jobs_v1_beta3_get_job_async.py" + "title": "dataflow_v1beta3_generated_jobs_v1_beta3_snapshot_job_async.py" }, { "canonical": true, @@ -1010,19 +1013,19 @@ "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3Client", "shortName": "JobsV1Beta3Client" }, - "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3Client.get_job", + "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3Client.snapshot_job", "method": { - "fullName": "google.dataflow.v1beta3.JobsV1Beta3.GetJob", + "fullName": "google.dataflow.v1beta3.JobsV1Beta3.SnapshotJob", "service": { "fullName": "google.dataflow.v1beta3.JobsV1Beta3", "shortName": "JobsV1Beta3" }, - "shortName": "GetJob" + "shortName": "SnapshotJob" }, "parameters": [ { "name": "request", - "type": "google.cloud.dataflow_v1beta3.types.GetJobRequest" + "type": "google.cloud.dataflow_v1beta3.types.SnapshotJobRequest" }, { "name": "retry", @@ -1037,14 +1040,14 @@ "type": "Sequence[Tuple[str, str]" } ], - "resultType": "google.cloud.dataflow_v1beta3.types.Job", - "shortName": "get_job" + "resultType": "google.cloud.dataflow_v1beta3.types.Snapshot", + "shortName": "snapshot_job" }, - "description": "Sample for GetJob", - "file": "dataflow_v1beta3_generated_jobs_v1_beta3_get_job_sync_3e418026.py", + "description": "Sample for SnapshotJob", + "file": "dataflow_v1beta3_generated_jobs_v1_beta3_snapshot_job_sync.py", "language": "PYTHON", "origin": "API_DEFINITION", - "regionTag": "dataflow_v1beta3_generated_JobsV1Beta3_GetJob_sync", + "regionTag": "dataflow_v1beta3_generated_JobsV1Beta3_SnapshotJob_sync", "segments": [ { "end": 50, @@ -1077,28 +1080,29 @@ "type": "RESPONSE_HANDLING" } ], - "title": "dataflow_v1beta3_generated_jobs_v1_beta3_get_job_sync_3e418026.py" + "title": "dataflow_v1beta3_generated_jobs_v1_beta3_snapshot_job_sync.py" }, { "canonical": true, "clientMethod": { + "async": true, "client": { - "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3Client", - "shortName": "JobsV1Beta3Client" + "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3AsyncClient", + "shortName": "JobsV1Beta3AsyncClient" }, - "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3Client.get_job", + "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3AsyncClient.update_job", "method": { - "fullName": "google.dataflow.v1beta3.JobsV1Beta3.GetJob", + "fullName": "google.dataflow.v1beta3.JobsV1Beta3.UpdateJob", "service": { "fullName": "google.dataflow.v1beta3.JobsV1Beta3", "shortName": "JobsV1Beta3" }, - "shortName": "GetJob" + "shortName": "UpdateJob" }, "parameters": [ { "name": "request", - "type": "google.cloud.dataflow_v1beta3.types.GetJobRequest" + "type": "google.cloud.dataflow_v1beta3.types.UpdateJobRequest" }, { "name": "retry", @@ -1114,13 +1118,13 @@ } ], "resultType": "google.cloud.dataflow_v1beta3.types.Job", - "shortName": "get_job" + "shortName": "update_job" }, - "description": "Sample for GetJob", - "file": "dataflow_v1beta3_generated_jobs_v1_beta3_get_job_sync_45212860.py", + "description": "Sample for UpdateJob", + "file": "dataflow_v1beta3_generated_jobs_v1_beta3_update_job_async.py", "language": "PYTHON", "origin": "API_DEFINITION", - "regionTag": "dataflow_v1beta3_generated_JobsV1Beta3_GetJob_sync", + "regionTag": "dataflow_v1beta3_generated_JobsV1Beta3_UpdateJob_async", "segments": [ { "end": 50, @@ -1153,29 +1157,28 @@ "type": "RESPONSE_HANDLING" } ], - "title": "dataflow_v1beta3_generated_jobs_v1_beta3_get_job_sync_45212860.py" + "title": "dataflow_v1beta3_generated_jobs_v1_beta3_update_job_async.py" }, { "canonical": true, "clientMethod": { - "async": true, "client": { - "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3AsyncClient", - "shortName": "JobsV1Beta3AsyncClient" + "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3Client", + "shortName": "JobsV1Beta3Client" }, - "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3AsyncClient.list_jobs", + "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3Client.update_job", "method": { - "fullName": "google.dataflow.v1beta3.JobsV1Beta3.ListJobs", + "fullName": "google.dataflow.v1beta3.JobsV1Beta3.UpdateJob", "service": { "fullName": "google.dataflow.v1beta3.JobsV1Beta3", "shortName": "JobsV1Beta3" }, - "shortName": "ListJobs" + "shortName": "UpdateJob" }, "parameters": [ { "name": "request", - "type": "google.cloud.dataflow_v1beta3.types.ListJobsRequest" + "type": "google.cloud.dataflow_v1beta3.types.UpdateJobRequest" }, { "name": "retry", @@ -1190,22 +1193,22 @@ "type": "Sequence[Tuple[str, str]" } ], - "resultType": "google.cloud.dataflow_v1beta3.services.jobs_v1_beta3.pagers.ListJobsAsyncPager", - "shortName": "list_jobs" + "resultType": "google.cloud.dataflow_v1beta3.types.Job", + "shortName": "update_job" }, - "description": "Sample for ListJobs", - "file": "dataflow_v1beta3_generated_jobs_v1_beta3_list_jobs_async.py", + "description": "Sample for UpdateJob", + "file": "dataflow_v1beta3_generated_jobs_v1_beta3_update_job_sync.py", "language": "PYTHON", "origin": "API_DEFINITION", - "regionTag": "dataflow_v1beta3_generated_JobsV1Beta3_ListJobs_async", + "regionTag": "dataflow_v1beta3_generated_JobsV1Beta3_UpdateJob_sync", "segments": [ { - "end": 51, + "end": 50, "start": 27, "type": "FULL" }, { - "end": 51, + "end": 50, "start": 27, "type": "SHORT" }, @@ -1225,33 +1228,34 @@ "type": "REQUEST_EXECUTION" }, { - "end": 52, + "end": 51, "start": 48, "type": "RESPONSE_HANDLING" } ], - "title": "dataflow_v1beta3_generated_jobs_v1_beta3_list_jobs_async.py" + "title": "dataflow_v1beta3_generated_jobs_v1_beta3_update_job_sync.py" }, { "canonical": true, "clientMethod": { + "async": true, "client": { - "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3Client", - "shortName": "JobsV1Beta3Client" + "fullName": "google.cloud.dataflow_v1beta3.MessagesV1Beta3AsyncClient", + "shortName": "MessagesV1Beta3AsyncClient" }, - "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3Client.list_jobs", + "fullName": "google.cloud.dataflow_v1beta3.MessagesV1Beta3AsyncClient.list_job_messages", "method": { - "fullName": "google.dataflow.v1beta3.JobsV1Beta3.ListJobs", + "fullName": "google.dataflow.v1beta3.MessagesV1Beta3.ListJobMessages", "service": { - "fullName": "google.dataflow.v1beta3.JobsV1Beta3", - "shortName": "JobsV1Beta3" + "fullName": "google.dataflow.v1beta3.MessagesV1Beta3", + "shortName": "MessagesV1Beta3" }, - "shortName": "ListJobs" + "shortName": "ListJobMessages" }, "parameters": [ { "name": "request", - "type": "google.cloud.dataflow_v1beta3.types.ListJobsRequest" + "type": "google.cloud.dataflow_v1beta3.types.ListJobMessagesRequest" }, { "name": "retry", @@ -1266,14 +1270,14 @@ "type": "Sequence[Tuple[str, str]" } ], - "resultType": "google.cloud.dataflow_v1beta3.services.jobs_v1_beta3.pagers.ListJobsPager", - "shortName": "list_jobs" + "resultType": "google.cloud.dataflow_v1beta3.services.messages_v1_beta3.pagers.ListJobMessagesAsyncPager", + "shortName": "list_job_messages" }, - "description": "Sample for ListJobs", - "file": "dataflow_v1beta3_generated_jobs_v1_beta3_list_jobs_sync_97a60855.py", + "description": "Sample for ListJobMessages", + "file": "dataflow_v1beta3_generated_messages_v1_beta3_list_job_messages_async.py", "language": "PYTHON", "origin": "API_DEFINITION", - "regionTag": "dataflow_v1beta3_generated_JobsV1Beta3_ListJobs_sync", + "regionTag": "dataflow_v1beta3_generated_MessagesV1Beta3_ListJobMessages_async", "segments": [ { "end": 51, @@ -1306,28 +1310,28 @@ "type": "RESPONSE_HANDLING" } ], - "title": "dataflow_v1beta3_generated_jobs_v1_beta3_list_jobs_sync_97a60855.py" + "title": "dataflow_v1beta3_generated_messages_v1_beta3_list_job_messages_async.py" }, { "canonical": true, "clientMethod": { "client": { - "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3Client", - "shortName": "JobsV1Beta3Client" + "fullName": "google.cloud.dataflow_v1beta3.MessagesV1Beta3Client", + "shortName": "MessagesV1Beta3Client" }, - "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3Client.list_jobs", + "fullName": "google.cloud.dataflow_v1beta3.MessagesV1Beta3Client.list_job_messages", "method": { - "fullName": "google.dataflow.v1beta3.JobsV1Beta3.ListJobs", + "fullName": "google.dataflow.v1beta3.MessagesV1Beta3.ListJobMessages", "service": { - "fullName": "google.dataflow.v1beta3.JobsV1Beta3", - "shortName": "JobsV1Beta3" + "fullName": "google.dataflow.v1beta3.MessagesV1Beta3", + "shortName": "MessagesV1Beta3" }, - "shortName": "ListJobs" + "shortName": "ListJobMessages" }, "parameters": [ { "name": "request", - "type": "google.cloud.dataflow_v1beta3.types.ListJobsRequest" + "type": "google.cloud.dataflow_v1beta3.types.ListJobMessagesRequest" }, { "name": "retry", @@ -1342,14 +1346,14 @@ "type": "Sequence[Tuple[str, str]" } ], - "resultType": "google.cloud.dataflow_v1beta3.services.jobs_v1_beta3.pagers.ListJobsPager", - "shortName": "list_jobs" + "resultType": "google.cloud.dataflow_v1beta3.services.messages_v1_beta3.pagers.ListJobMessagesPager", + "shortName": "list_job_messages" }, - "description": "Sample for ListJobs", - "file": "dataflow_v1beta3_generated_jobs_v1_beta3_list_jobs_sync_bfb75b4c.py", + "description": "Sample for ListJobMessages", + "file": "dataflow_v1beta3_generated_messages_v1_beta3_list_job_messages_sync.py", "language": "PYTHON", "origin": "API_DEFINITION", - "regionTag": "dataflow_v1beta3_generated_JobsV1Beta3_ListJobs_sync", + "regionTag": "dataflow_v1beta3_generated_MessagesV1Beta3_ListJobMessages_sync", "segments": [ { "end": 51, @@ -1382,29 +1386,29 @@ "type": "RESPONSE_HANDLING" } ], - "title": "dataflow_v1beta3_generated_jobs_v1_beta3_list_jobs_sync_bfb75b4c.py" + "title": "dataflow_v1beta3_generated_messages_v1_beta3_list_job_messages_sync.py" }, { "canonical": true, "clientMethod": { "async": true, "client": { - "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3AsyncClient", - "shortName": "JobsV1Beta3AsyncClient" + "fullName": "google.cloud.dataflow_v1beta3.MetricsV1Beta3AsyncClient", + "shortName": "MetricsV1Beta3AsyncClient" }, - "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3AsyncClient.snapshot_job", + "fullName": "google.cloud.dataflow_v1beta3.MetricsV1Beta3AsyncClient.get_job_execution_details", "method": { - "fullName": "google.dataflow.v1beta3.JobsV1Beta3.SnapshotJob", + "fullName": "google.dataflow.v1beta3.MetricsV1Beta3.GetJobExecutionDetails", "service": { - "fullName": "google.dataflow.v1beta3.JobsV1Beta3", - "shortName": "JobsV1Beta3" + "fullName": "google.dataflow.v1beta3.MetricsV1Beta3", + "shortName": "MetricsV1Beta3" }, - "shortName": "SnapshotJob" + "shortName": "GetJobExecutionDetails" }, "parameters": [ { "name": "request", - "type": "google.cloud.dataflow_v1beta3.types.SnapshotJobRequest" + "type": "google.cloud.dataflow_v1beta3.types.GetJobExecutionDetailsRequest" }, { "name": "retry", @@ -1419,22 +1423,22 @@ "type": "Sequence[Tuple[str, str]" } ], - "resultType": "google.cloud.dataflow_v1beta3.types.Snapshot", - "shortName": "snapshot_job" + "resultType": "google.cloud.dataflow_v1beta3.services.metrics_v1_beta3.pagers.GetJobExecutionDetailsAsyncPager", + "shortName": "get_job_execution_details" }, - "description": "Sample for SnapshotJob", - "file": "dataflow_v1beta3_generated_jobs_v1_beta3_snapshot_job_async.py", + "description": "Sample for GetJobExecutionDetails", + "file": "dataflow_v1beta3_generated_metrics_v1_beta3_get_job_execution_details_async.py", "language": "PYTHON", "origin": "API_DEFINITION", - "regionTag": "dataflow_v1beta3_generated_JobsV1Beta3_SnapshotJob_async", + "regionTag": "dataflow_v1beta3_generated_MetricsV1Beta3_GetJobExecutionDetails_async", "segments": [ { - "end": 50, + "end": 51, "start": 27, "type": "FULL" }, { - "end": 50, + "end": 51, "start": 27, "type": "SHORT" }, @@ -1454,33 +1458,33 @@ "type": "REQUEST_EXECUTION" }, { - "end": 51, + "end": 52, "start": 48, "type": "RESPONSE_HANDLING" } ], - "title": "dataflow_v1beta3_generated_jobs_v1_beta3_snapshot_job_async.py" + "title": "dataflow_v1beta3_generated_metrics_v1_beta3_get_job_execution_details_async.py" }, { "canonical": true, "clientMethod": { "client": { - "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3Client", - "shortName": "JobsV1Beta3Client" + "fullName": "google.cloud.dataflow_v1beta3.MetricsV1Beta3Client", + "shortName": "MetricsV1Beta3Client" }, - "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3Client.snapshot_job", + "fullName": "google.cloud.dataflow_v1beta3.MetricsV1Beta3Client.get_job_execution_details", "method": { - "fullName": "google.dataflow.v1beta3.JobsV1Beta3.SnapshotJob", + "fullName": "google.dataflow.v1beta3.MetricsV1Beta3.GetJobExecutionDetails", "service": { - "fullName": "google.dataflow.v1beta3.JobsV1Beta3", - "shortName": "JobsV1Beta3" + "fullName": "google.dataflow.v1beta3.MetricsV1Beta3", + "shortName": "MetricsV1Beta3" }, - "shortName": "SnapshotJob" + "shortName": "GetJobExecutionDetails" }, "parameters": [ { "name": "request", - "type": "google.cloud.dataflow_v1beta3.types.SnapshotJobRequest" + "type": "google.cloud.dataflow_v1beta3.types.GetJobExecutionDetailsRequest" }, { "name": "retry", @@ -1495,22 +1499,22 @@ "type": "Sequence[Tuple[str, str]" } ], - "resultType": "google.cloud.dataflow_v1beta3.types.Snapshot", - "shortName": "snapshot_job" + "resultType": "google.cloud.dataflow_v1beta3.services.metrics_v1_beta3.pagers.GetJobExecutionDetailsPager", + "shortName": "get_job_execution_details" }, - "description": "Sample for SnapshotJob", - "file": "dataflow_v1beta3_generated_jobs_v1_beta3_snapshot_job_sync_60f4d8fe.py", + "description": "Sample for GetJobExecutionDetails", + "file": "dataflow_v1beta3_generated_metrics_v1_beta3_get_job_execution_details_sync.py", "language": "PYTHON", "origin": "API_DEFINITION", - "regionTag": "dataflow_v1beta3_generated_JobsV1Beta3_SnapshotJob_sync", + "regionTag": "dataflow_v1beta3_generated_MetricsV1Beta3_GetJobExecutionDetails_sync", "segments": [ { - "end": 50, + "end": 51, "start": 27, "type": "FULL" }, { - "end": 50, + "end": 51, "start": 27, "type": "SHORT" }, @@ -1530,33 +1534,34 @@ "type": "REQUEST_EXECUTION" }, { - "end": 51, + "end": 52, "start": 48, "type": "RESPONSE_HANDLING" } ], - "title": "dataflow_v1beta3_generated_jobs_v1_beta3_snapshot_job_sync_60f4d8fe.py" + "title": "dataflow_v1beta3_generated_metrics_v1_beta3_get_job_execution_details_sync.py" }, { "canonical": true, "clientMethod": { + "async": true, "client": { - "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3Client", - "shortName": "JobsV1Beta3Client" + "fullName": "google.cloud.dataflow_v1beta3.MetricsV1Beta3AsyncClient", + "shortName": "MetricsV1Beta3AsyncClient" }, - "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3Client.snapshot_job", + "fullName": "google.cloud.dataflow_v1beta3.MetricsV1Beta3AsyncClient.get_job_metrics", "method": { - "fullName": "google.dataflow.v1beta3.JobsV1Beta3.SnapshotJob", + "fullName": "google.dataflow.v1beta3.MetricsV1Beta3.GetJobMetrics", "service": { - "fullName": "google.dataflow.v1beta3.JobsV1Beta3", - "shortName": "JobsV1Beta3" + "fullName": "google.dataflow.v1beta3.MetricsV1Beta3", + "shortName": "MetricsV1Beta3" }, - "shortName": "SnapshotJob" + "shortName": "GetJobMetrics" }, "parameters": [ { "name": "request", - "type": "google.cloud.dataflow_v1beta3.types.SnapshotJobRequest" + "type": "google.cloud.dataflow_v1beta3.types.GetJobMetricsRequest" }, { "name": "retry", @@ -1571,14 +1576,14 @@ "type": "Sequence[Tuple[str, str]" } ], - "resultType": "google.cloud.dataflow_v1beta3.types.Snapshot", - "shortName": "snapshot_job" + "resultType": "google.cloud.dataflow_v1beta3.types.JobMetrics", + "shortName": "get_job_metrics" }, - "description": "Sample for SnapshotJob", - "file": "dataflow_v1beta3_generated_jobs_v1_beta3_snapshot_job_sync_ff3cb8bd.py", + "description": "Sample for GetJobMetrics", + "file": "dataflow_v1beta3_generated_metrics_v1_beta3_get_job_metrics_async.py", "language": "PYTHON", "origin": "API_DEFINITION", - "regionTag": "dataflow_v1beta3_generated_JobsV1Beta3_SnapshotJob_sync", + "regionTag": "dataflow_v1beta3_generated_MetricsV1Beta3_GetJobMetrics_async", "segments": [ { "end": 50, @@ -1611,29 +1616,28 @@ "type": "RESPONSE_HANDLING" } ], - "title": "dataflow_v1beta3_generated_jobs_v1_beta3_snapshot_job_sync_ff3cb8bd.py" + "title": "dataflow_v1beta3_generated_metrics_v1_beta3_get_job_metrics_async.py" }, { "canonical": true, "clientMethod": { - "async": true, "client": { - "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3AsyncClient", - "shortName": "JobsV1Beta3AsyncClient" + "fullName": "google.cloud.dataflow_v1beta3.MetricsV1Beta3Client", + "shortName": "MetricsV1Beta3Client" }, - "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3AsyncClient.update_job", + "fullName": "google.cloud.dataflow_v1beta3.MetricsV1Beta3Client.get_job_metrics", "method": { - "fullName": "google.dataflow.v1beta3.JobsV1Beta3.UpdateJob", + "fullName": "google.dataflow.v1beta3.MetricsV1Beta3.GetJobMetrics", "service": { - "fullName": "google.dataflow.v1beta3.JobsV1Beta3", - "shortName": "JobsV1Beta3" + "fullName": "google.dataflow.v1beta3.MetricsV1Beta3", + "shortName": "MetricsV1Beta3" }, - "shortName": "UpdateJob" + "shortName": "GetJobMetrics" }, "parameters": [ { "name": "request", - "type": "google.cloud.dataflow_v1beta3.types.UpdateJobRequest" + "type": "google.cloud.dataflow_v1beta3.types.GetJobMetricsRequest" }, { "name": "retry", @@ -1648,14 +1652,14 @@ "type": "Sequence[Tuple[str, str]" } ], - "resultType": "google.cloud.dataflow_v1beta3.types.Job", - "shortName": "update_job" + "resultType": "google.cloud.dataflow_v1beta3.types.JobMetrics", + "shortName": "get_job_metrics" }, - "description": "Sample for UpdateJob", - "file": "dataflow_v1beta3_generated_jobs_v1_beta3_update_job_async.py", + "description": "Sample for GetJobMetrics", + "file": "dataflow_v1beta3_generated_metrics_v1_beta3_get_job_metrics_sync.py", "language": "PYTHON", "origin": "API_DEFINITION", - "regionTag": "dataflow_v1beta3_generated_JobsV1Beta3_UpdateJob_async", + "regionTag": "dataflow_v1beta3_generated_MetricsV1Beta3_GetJobMetrics_sync", "segments": [ { "end": 50, @@ -1688,28 +1692,29 @@ "type": "RESPONSE_HANDLING" } ], - "title": "dataflow_v1beta3_generated_jobs_v1_beta3_update_job_async.py" + "title": "dataflow_v1beta3_generated_metrics_v1_beta3_get_job_metrics_sync.py" }, { "canonical": true, "clientMethod": { + "async": true, "client": { - "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3Client", - "shortName": "JobsV1Beta3Client" + "fullName": "google.cloud.dataflow_v1beta3.MetricsV1Beta3AsyncClient", + "shortName": "MetricsV1Beta3AsyncClient" }, - "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3Client.update_job", + "fullName": "google.cloud.dataflow_v1beta3.MetricsV1Beta3AsyncClient.get_stage_execution_details", "method": { - "fullName": "google.dataflow.v1beta3.JobsV1Beta3.UpdateJob", + "fullName": "google.dataflow.v1beta3.MetricsV1Beta3.GetStageExecutionDetails", "service": { - "fullName": "google.dataflow.v1beta3.JobsV1Beta3", - "shortName": "JobsV1Beta3" + "fullName": "google.dataflow.v1beta3.MetricsV1Beta3", + "shortName": "MetricsV1Beta3" }, - "shortName": "UpdateJob" + "shortName": "GetStageExecutionDetails" }, "parameters": [ { "name": "request", - "type": "google.cloud.dataflow_v1beta3.types.UpdateJobRequest" + "type": "google.cloud.dataflow_v1beta3.types.GetStageExecutionDetailsRequest" }, { "name": "retry", @@ -1724,1243 +1729,22 @@ "type": "Sequence[Tuple[str, str]" } ], - "resultType": "google.cloud.dataflow_v1beta3.types.Job", - "shortName": "update_job" + "resultType": "google.cloud.dataflow_v1beta3.services.metrics_v1_beta3.pagers.GetStageExecutionDetailsAsyncPager", + "shortName": "get_stage_execution_details" }, - "description": "Sample for UpdateJob", - "file": "dataflow_v1beta3_generated_jobs_v1_beta3_update_job_sync_8db1fd08.py", + "description": "Sample for GetStageExecutionDetails", + "file": "dataflow_v1beta3_generated_metrics_v1_beta3_get_stage_execution_details_async.py", "language": "PYTHON", "origin": "API_DEFINITION", - "regionTag": "dataflow_v1beta3_generated_JobsV1Beta3_UpdateJob_sync", + "regionTag": "dataflow_v1beta3_generated_MetricsV1Beta3_GetStageExecutionDetails_async", "segments": [ { - "end": 50, + "end": 51, "start": 27, "type": "FULL" }, - { - "end": 50, - "start": 27, - "type": "SHORT" - }, - { - "end": 40, - "start": 38, - "type": "CLIENT_INITIALIZATION" - }, - { - "end": 44, - "start": 41, - "type": "REQUEST_INITIALIZATION" - }, - { - "end": 47, - "start": 45, - "type": "REQUEST_EXECUTION" - }, - { - "end": 51, - "start": 48, - "type": "RESPONSE_HANDLING" - } - ], - "title": "dataflow_v1beta3_generated_jobs_v1_beta3_update_job_sync_8db1fd08.py" - }, - { - "canonical": true, - "clientMethod": { - "client": { - "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3Client", - "shortName": "JobsV1Beta3Client" - }, - "fullName": "google.cloud.dataflow_v1beta3.JobsV1Beta3Client.update_job", - "method": { - "fullName": "google.dataflow.v1beta3.JobsV1Beta3.UpdateJob", - "service": { - "fullName": "google.dataflow.v1beta3.JobsV1Beta3", - "shortName": "JobsV1Beta3" - }, - "shortName": "UpdateJob" - }, - "parameters": [ - { - "name": "request", - "type": "google.cloud.dataflow_v1beta3.types.UpdateJobRequest" - }, - { - "name": "retry", - "type": "google.api_core.retry.Retry" - }, - { - "name": "timeout", - "type": "float" - }, - { - "name": "metadata", - "type": "Sequence[Tuple[str, str]" - } - ], - "resultType": "google.cloud.dataflow_v1beta3.types.Job", - "shortName": "update_job" - }, - "description": "Sample for UpdateJob", - "file": "dataflow_v1beta3_generated_jobs_v1_beta3_update_job_sync_e2fa191d.py", - "language": "PYTHON", - "origin": "API_DEFINITION", - "regionTag": "dataflow_v1beta3_generated_JobsV1Beta3_UpdateJob_sync", - "segments": [ - { - "end": 50, - "start": 27, - "type": "FULL" - }, - { - "end": 50, - "start": 27, - "type": "SHORT" - }, - { - "end": 40, - "start": 38, - "type": "CLIENT_INITIALIZATION" - }, - { - "end": 44, - "start": 41, - "type": "REQUEST_INITIALIZATION" - }, - { - "end": 47, - "start": 45, - "type": "REQUEST_EXECUTION" - }, - { - "end": 51, - "start": 48, - "type": "RESPONSE_HANDLING" - } - ], - "title": "dataflow_v1beta3_generated_jobs_v1_beta3_update_job_sync_e2fa191d.py" - }, - { - "canonical": true, - "clientMethod": { - "async": true, - "client": { - "fullName": "google.cloud.dataflow_v1beta3.MessagesV1Beta3AsyncClient", - "shortName": "MessagesV1Beta3AsyncClient" - }, - "fullName": "google.cloud.dataflow_v1beta3.MessagesV1Beta3AsyncClient.list_job_messages", - "method": { - "fullName": "google.dataflow.v1beta3.MessagesV1Beta3.ListJobMessages", - "service": { - "fullName": "google.dataflow.v1beta3.MessagesV1Beta3", - "shortName": "MessagesV1Beta3" - }, - "shortName": "ListJobMessages" - }, - "parameters": [ - { - "name": "request", - "type": "google.cloud.dataflow_v1beta3.types.ListJobMessagesRequest" - }, - { - "name": "retry", - "type": "google.api_core.retry.Retry" - }, - { - "name": "timeout", - "type": "float" - }, - { - "name": "metadata", - "type": "Sequence[Tuple[str, str]" - } - ], - "resultType": "google.cloud.dataflow_v1beta3.services.messages_v1_beta3.pagers.ListJobMessagesAsyncPager", - "shortName": "list_job_messages" - }, - "description": "Sample for ListJobMessages", - "file": "dataflow_v1beta3_generated_messages_v1_beta3_list_job_messages_async.py", - "language": "PYTHON", - "origin": "API_DEFINITION", - "regionTag": "dataflow_v1beta3_generated_MessagesV1Beta3_ListJobMessages_async", - "segments": [ - { - "end": 51, - "start": 27, - "type": "FULL" - }, - { - "end": 51, - "start": 27, - "type": "SHORT" - }, - { - "end": 40, - "start": 38, - "type": "CLIENT_INITIALIZATION" - }, - { - "end": 44, - "start": 41, - "type": "REQUEST_INITIALIZATION" - }, - { - "end": 47, - "start": 45, - "type": "REQUEST_EXECUTION" - }, - { - "end": 52, - "start": 48, - "type": "RESPONSE_HANDLING" - } - ], - "title": "dataflow_v1beta3_generated_messages_v1_beta3_list_job_messages_async.py" - }, - { - "canonical": true, - "clientMethod": { - "client": { - "fullName": "google.cloud.dataflow_v1beta3.MessagesV1Beta3Client", - "shortName": "MessagesV1Beta3Client" - }, - "fullName": "google.cloud.dataflow_v1beta3.MessagesV1Beta3Client.list_job_messages", - "method": { - "fullName": "google.dataflow.v1beta3.MessagesV1Beta3.ListJobMessages", - "service": { - "fullName": "google.dataflow.v1beta3.MessagesV1Beta3", - "shortName": "MessagesV1Beta3" - }, - "shortName": "ListJobMessages" - }, - "parameters": [ - { - "name": "request", - "type": "google.cloud.dataflow_v1beta3.types.ListJobMessagesRequest" - }, - { - "name": "retry", - "type": "google.api_core.retry.Retry" - }, - { - "name": "timeout", - "type": "float" - }, - { - "name": "metadata", - "type": "Sequence[Tuple[str, str]" - } - ], - "resultType": "google.cloud.dataflow_v1beta3.services.messages_v1_beta3.pagers.ListJobMessagesPager", - "shortName": "list_job_messages" - }, - "description": "Sample for ListJobMessages", - "file": "dataflow_v1beta3_generated_messages_v1_beta3_list_job_messages_sync_f19ed68d.py", - "language": "PYTHON", - "origin": "API_DEFINITION", - "regionTag": "dataflow_v1beta3_generated_MessagesV1Beta3_ListJobMessages_sync", - "segments": [ - { - "end": 51, - "start": 27, - "type": "FULL" - }, - { - "end": 51, - "start": 27, - "type": "SHORT" - }, - { - "end": 40, - "start": 38, - "type": "CLIENT_INITIALIZATION" - }, - { - "end": 44, - "start": 41, - "type": "REQUEST_INITIALIZATION" - }, - { - "end": 47, - "start": 45, - "type": "REQUEST_EXECUTION" - }, - { - "end": 52, - "start": 48, - "type": "RESPONSE_HANDLING" - } - ], - "title": "dataflow_v1beta3_generated_messages_v1_beta3_list_job_messages_sync_f19ed68d.py" - }, - { - "canonical": true, - "clientMethod": { - "client": { - "fullName": "google.cloud.dataflow_v1beta3.MessagesV1Beta3Client", - "shortName": "MessagesV1Beta3Client" - }, - "fullName": "google.cloud.dataflow_v1beta3.MessagesV1Beta3Client.list_job_messages", - "method": { - "fullName": "google.dataflow.v1beta3.MessagesV1Beta3.ListJobMessages", - "service": { - "fullName": "google.dataflow.v1beta3.MessagesV1Beta3", - "shortName": "MessagesV1Beta3" - }, - "shortName": "ListJobMessages" - }, - "parameters": [ - { - "name": "request", - "type": "google.cloud.dataflow_v1beta3.types.ListJobMessagesRequest" - }, - { - "name": "retry", - "type": "google.api_core.retry.Retry" - }, - { - "name": "timeout", - "type": "float" - }, - { - "name": "metadata", - "type": "Sequence[Tuple[str, str]" - } - ], - "resultType": "google.cloud.dataflow_v1beta3.services.messages_v1_beta3.pagers.ListJobMessagesPager", - "shortName": "list_job_messages" - }, - "description": "Sample for ListJobMessages", - "file": "dataflow_v1beta3_generated_messages_v1_beta3_list_job_messages_sync_04dec136.py", - "language": "PYTHON", - "origin": "API_DEFINITION", - "regionTag": "dataflow_v1beta3_generated_MessagesV1Beta3_ListJobMessages_sync", - "segments": [ - { - "end": 51, - "start": 27, - "type": "FULL" - }, - { - "end": 51, - "start": 27, - "type": "SHORT" - }, - { - "end": 40, - "start": 38, - "type": "CLIENT_INITIALIZATION" - }, - { - "end": 44, - "start": 41, - "type": "REQUEST_INITIALIZATION" - }, - { - "end": 47, - "start": 45, - "type": "REQUEST_EXECUTION" - }, - { - "end": 52, - "start": 48, - "type": "RESPONSE_HANDLING" - } - ], - "title": "dataflow_v1beta3_generated_messages_v1_beta3_list_job_messages_sync_04dec136.py" - }, - { - "canonical": true, - "clientMethod": { - "async": true, - "client": { - "fullName": "google.cloud.dataflow_v1beta3.MetricsV1Beta3AsyncClient", - "shortName": "MetricsV1Beta3AsyncClient" - }, - "fullName": "google.cloud.dataflow_v1beta3.MetricsV1Beta3AsyncClient.get_job_execution_details", - "method": { - "fullName": "google.dataflow.v1beta3.MetricsV1Beta3.GetJobExecutionDetails", - "service": { - "fullName": "google.dataflow.v1beta3.MetricsV1Beta3", - "shortName": "MetricsV1Beta3" - }, - "shortName": "GetJobExecutionDetails" - }, - "parameters": [ - { - "name": "request", - "type": "google.cloud.dataflow_v1beta3.types.GetJobExecutionDetailsRequest" - }, - { - "name": "retry", - "type": "google.api_core.retry.Retry" - }, - { - "name": "timeout", - "type": "float" - }, - { - "name": "metadata", - "type": "Sequence[Tuple[str, str]" - } - ], - "resultType": "google.cloud.dataflow_v1beta3.services.metrics_v1_beta3.pagers.GetJobExecutionDetailsAsyncPager", - "shortName": "get_job_execution_details" - }, - "description": "Sample for GetJobExecutionDetails", - "file": "dataflow_v1beta3_generated_metrics_v1_beta3_get_job_execution_details_async.py", - "language": "PYTHON", - "origin": "API_DEFINITION", - "regionTag": "dataflow_v1beta3_generated_MetricsV1Beta3_GetJobExecutionDetails_async", - "segments": [ - { - "end": 51, - "start": 27, - "type": "FULL" - }, - { - "end": 51, - "start": 27, - "type": "SHORT" - }, - { - "end": 40, - "start": 38, - "type": "CLIENT_INITIALIZATION" - }, - { - "end": 44, - "start": 41, - "type": "REQUEST_INITIALIZATION" - }, - { - "end": 47, - "start": 45, - "type": "REQUEST_EXECUTION" - }, - { - "end": 52, - "start": 48, - "type": "RESPONSE_HANDLING" - } - ], - "title": "dataflow_v1beta3_generated_metrics_v1_beta3_get_job_execution_details_async.py" - }, - { - "canonical": true, - "clientMethod": { - "client": { - "fullName": "google.cloud.dataflow_v1beta3.MetricsV1Beta3Client", - "shortName": "MetricsV1Beta3Client" - }, - "fullName": "google.cloud.dataflow_v1beta3.MetricsV1Beta3Client.get_job_execution_details", - "method": { - "fullName": "google.dataflow.v1beta3.MetricsV1Beta3.GetJobExecutionDetails", - "service": { - "fullName": "google.dataflow.v1beta3.MetricsV1Beta3", - "shortName": "MetricsV1Beta3" - }, - "shortName": "GetJobExecutionDetails" - }, - "parameters": [ - { - "name": "request", - "type": "google.cloud.dataflow_v1beta3.types.GetJobExecutionDetailsRequest" - }, - { - "name": "retry", - "type": "google.api_core.retry.Retry" - }, - { - "name": "timeout", - "type": "float" - }, - { - "name": "metadata", - "type": "Sequence[Tuple[str, str]" - } - ], - "resultType": "google.cloud.dataflow_v1beta3.services.metrics_v1_beta3.pagers.GetJobExecutionDetailsPager", - "shortName": "get_job_execution_details" - }, - "description": "Sample for GetJobExecutionDetails", - "file": "dataflow_v1beta3_generated_metrics_v1_beta3_get_job_execution_details_sync_b7550163.py", - "language": "PYTHON", - "origin": "API_DEFINITION", - "regionTag": "dataflow_v1beta3_generated_MetricsV1Beta3_GetJobExecutionDetails_sync", - "segments": [ - { - "end": 51, - "start": 27, - "type": "FULL" - }, - { - "end": 51, - "start": 27, - "type": "SHORT" - }, - { - "end": 40, - "start": 38, - "type": "CLIENT_INITIALIZATION" - }, - { - "end": 44, - "start": 41, - "type": "REQUEST_INITIALIZATION" - }, - { - "end": 47, - "start": 45, - "type": "REQUEST_EXECUTION" - }, - { - "end": 52, - "start": 48, - "type": "RESPONSE_HANDLING" - } - ], - "title": "dataflow_v1beta3_generated_metrics_v1_beta3_get_job_execution_details_sync_b7550163.py" - }, - { - "canonical": true, - "clientMethod": { - "client": { - "fullName": "google.cloud.dataflow_v1beta3.MetricsV1Beta3Client", - "shortName": "MetricsV1Beta3Client" - }, - "fullName": "google.cloud.dataflow_v1beta3.MetricsV1Beta3Client.get_job_execution_details", - "method": { - "fullName": "google.dataflow.v1beta3.MetricsV1Beta3.GetJobExecutionDetails", - "service": { - "fullName": "google.dataflow.v1beta3.MetricsV1Beta3", - "shortName": "MetricsV1Beta3" - }, - "shortName": "GetJobExecutionDetails" - }, - "parameters": [ - { - "name": "request", - "type": "google.cloud.dataflow_v1beta3.types.GetJobExecutionDetailsRequest" - }, - { - "name": "retry", - "type": "google.api_core.retry.Retry" - }, - { - "name": "timeout", - "type": "float" - }, - { - "name": "metadata", - "type": "Sequence[Tuple[str, str]" - } - ], - "resultType": "google.cloud.dataflow_v1beta3.services.metrics_v1_beta3.pagers.GetJobExecutionDetailsPager", - "shortName": "get_job_execution_details" - }, - "description": "Sample for GetJobExecutionDetails", - "file": "dataflow_v1beta3_generated_metrics_v1_beta3_get_job_execution_details_sync_a4ff4d57.py", - "language": "PYTHON", - "origin": "API_DEFINITION", - "regionTag": "dataflow_v1beta3_generated_MetricsV1Beta3_GetJobExecutionDetails_sync", - "segments": [ - { - "end": 51, - "start": 27, - "type": "FULL" - }, - { - "end": 51, - "start": 27, - "type": "SHORT" - }, - { - "end": 40, - "start": 38, - "type": "CLIENT_INITIALIZATION" - }, - { - "end": 44, - "start": 41, - "type": "REQUEST_INITIALIZATION" - }, - { - "end": 47, - "start": 45, - "type": "REQUEST_EXECUTION" - }, - { - "end": 52, - "start": 48, - "type": "RESPONSE_HANDLING" - } - ], - "title": "dataflow_v1beta3_generated_metrics_v1_beta3_get_job_execution_details_sync_a4ff4d57.py" - }, - { - "canonical": true, - "clientMethod": { - "async": true, - "client": { - "fullName": "google.cloud.dataflow_v1beta3.MetricsV1Beta3AsyncClient", - "shortName": "MetricsV1Beta3AsyncClient" - }, - "fullName": "google.cloud.dataflow_v1beta3.MetricsV1Beta3AsyncClient.get_job_metrics", - "method": { - "fullName": "google.dataflow.v1beta3.MetricsV1Beta3.GetJobMetrics", - "service": { - "fullName": "google.dataflow.v1beta3.MetricsV1Beta3", - "shortName": "MetricsV1Beta3" - }, - "shortName": "GetJobMetrics" - }, - "parameters": [ - { - "name": "request", - "type": "google.cloud.dataflow_v1beta3.types.GetJobMetricsRequest" - }, - { - "name": "retry", - "type": "google.api_core.retry.Retry" - }, - { - "name": "timeout", - "type": "float" - }, - { - "name": "metadata", - "type": "Sequence[Tuple[str, str]" - } - ], - "resultType": "google.cloud.dataflow_v1beta3.types.JobMetrics", - "shortName": "get_job_metrics" - }, - "description": "Sample for GetJobMetrics", - "file": "dataflow_v1beta3_generated_metrics_v1_beta3_get_job_metrics_async.py", - "language": "PYTHON", - "origin": "API_DEFINITION", - "regionTag": "dataflow_v1beta3_generated_MetricsV1Beta3_GetJobMetrics_async", - "segments": [ - { - "end": 50, - "start": 27, - "type": "FULL" - }, - { - "end": 50, - "start": 27, - "type": "SHORT" - }, - { - "end": 40, - "start": 38, - "type": "CLIENT_INITIALIZATION" - }, - { - "end": 44, - "start": 41, - "type": "REQUEST_INITIALIZATION" - }, - { - "end": 47, - "start": 45, - "type": "REQUEST_EXECUTION" - }, - { - "end": 51, - "start": 48, - "type": "RESPONSE_HANDLING" - } - ], - "title": "dataflow_v1beta3_generated_metrics_v1_beta3_get_job_metrics_async.py" - }, - { - "canonical": true, - "clientMethod": { - "client": { - "fullName": "google.cloud.dataflow_v1beta3.MetricsV1Beta3Client", - "shortName": "MetricsV1Beta3Client" - }, - "fullName": "google.cloud.dataflow_v1beta3.MetricsV1Beta3Client.get_job_metrics", - "method": { - "fullName": "google.dataflow.v1beta3.MetricsV1Beta3.GetJobMetrics", - "service": { - "fullName": "google.dataflow.v1beta3.MetricsV1Beta3", - "shortName": "MetricsV1Beta3" - }, - "shortName": "GetJobMetrics" - }, - "parameters": [ - { - "name": "request", - "type": "google.cloud.dataflow_v1beta3.types.GetJobMetricsRequest" - }, - { - "name": "retry", - "type": "google.api_core.retry.Retry" - }, - { - "name": "timeout", - "type": "float" - }, - { - "name": "metadata", - "type": "Sequence[Tuple[str, str]" - } - ], - "resultType": "google.cloud.dataflow_v1beta3.types.JobMetrics", - "shortName": "get_job_metrics" - }, - "description": "Sample for GetJobMetrics", - "file": "dataflow_v1beta3_generated_metrics_v1_beta3_get_job_metrics_sync_02835968.py", - "language": "PYTHON", - "origin": "API_DEFINITION", - "regionTag": "dataflow_v1beta3_generated_MetricsV1Beta3_GetJobMetrics_sync", - "segments": [ - { - "end": 50, - "start": 27, - "type": "FULL" - }, - { - "end": 50, - "start": 27, - "type": "SHORT" - }, - { - "end": 40, - "start": 38, - "type": "CLIENT_INITIALIZATION" - }, - { - "end": 44, - "start": 41, - "type": "REQUEST_INITIALIZATION" - }, - { - "end": 47, - "start": 45, - "type": "REQUEST_EXECUTION" - }, - { - "end": 51, - "start": 48, - "type": "RESPONSE_HANDLING" - } - ], - "title": "dataflow_v1beta3_generated_metrics_v1_beta3_get_job_metrics_sync_02835968.py" - }, - { - "canonical": true, - "clientMethod": { - "client": { - "fullName": "google.cloud.dataflow_v1beta3.MetricsV1Beta3Client", - "shortName": "MetricsV1Beta3Client" - }, - "fullName": "google.cloud.dataflow_v1beta3.MetricsV1Beta3Client.get_job_metrics", - "method": { - "fullName": "google.dataflow.v1beta3.MetricsV1Beta3.GetJobMetrics", - "service": { - "fullName": "google.dataflow.v1beta3.MetricsV1Beta3", - "shortName": "MetricsV1Beta3" - }, - "shortName": "GetJobMetrics" - }, - "parameters": [ - { - "name": "request", - "type": "google.cloud.dataflow_v1beta3.types.GetJobMetricsRequest" - }, - { - "name": "retry", - "type": "google.api_core.retry.Retry" - }, - { - "name": "timeout", - "type": "float" - }, - { - "name": "metadata", - "type": "Sequence[Tuple[str, str]" - } - ], - "resultType": "google.cloud.dataflow_v1beta3.types.JobMetrics", - "shortName": "get_job_metrics" - }, - "description": "Sample for GetJobMetrics", - "file": "dataflow_v1beta3_generated_metrics_v1_beta3_get_job_metrics_sync_59dce217.py", - "language": "PYTHON", - "origin": "API_DEFINITION", - "regionTag": "dataflow_v1beta3_generated_MetricsV1Beta3_GetJobMetrics_sync", - "segments": [ - { - "end": 50, - "start": 27, - "type": "FULL" - }, - { - "end": 50, - "start": 27, - "type": "SHORT" - }, - { - "end": 40, - "start": 38, - "type": "CLIENT_INITIALIZATION" - }, - { - "end": 44, - "start": 41, - "type": "REQUEST_INITIALIZATION" - }, - { - "end": 47, - "start": 45, - "type": "REQUEST_EXECUTION" - }, - { - "end": 51, - "start": 48, - "type": "RESPONSE_HANDLING" - } - ], - "title": "dataflow_v1beta3_generated_metrics_v1_beta3_get_job_metrics_sync_59dce217.py" - }, - { - "canonical": true, - "clientMethod": { - "async": true, - "client": { - "fullName": "google.cloud.dataflow_v1beta3.MetricsV1Beta3AsyncClient", - "shortName": "MetricsV1Beta3AsyncClient" - }, - "fullName": "google.cloud.dataflow_v1beta3.MetricsV1Beta3AsyncClient.get_stage_execution_details", - "method": { - "fullName": "google.dataflow.v1beta3.MetricsV1Beta3.GetStageExecutionDetails", - "service": { - "fullName": "google.dataflow.v1beta3.MetricsV1Beta3", - "shortName": "MetricsV1Beta3" - }, - "shortName": "GetStageExecutionDetails" - }, - "parameters": [ - { - "name": "request", - "type": "google.cloud.dataflow_v1beta3.types.GetStageExecutionDetailsRequest" - }, - { - "name": "retry", - "type": "google.api_core.retry.Retry" - }, - { - "name": "timeout", - "type": "float" - }, - { - "name": "metadata", - "type": "Sequence[Tuple[str, str]" - } - ], - "resultType": "google.cloud.dataflow_v1beta3.services.metrics_v1_beta3.pagers.GetStageExecutionDetailsAsyncPager", - "shortName": "get_stage_execution_details" - }, - "description": "Sample for GetStageExecutionDetails", - "file": "dataflow_v1beta3_generated_metrics_v1_beta3_get_stage_execution_details_async.py", - "language": "PYTHON", - "origin": "API_DEFINITION", - "regionTag": "dataflow_v1beta3_generated_MetricsV1Beta3_GetStageExecutionDetails_async", - "segments": [ - { - "end": 51, - "start": 27, - "type": "FULL" - }, - { - "end": 51, - "start": 27, - "type": "SHORT" - }, - { - "end": 40, - "start": 38, - "type": "CLIENT_INITIALIZATION" - }, - { - "end": 44, - "start": 41, - "type": "REQUEST_INITIALIZATION" - }, - { - "end": 47, - "start": 45, - "type": "REQUEST_EXECUTION" - }, - { - "end": 52, - "start": 48, - "type": "RESPONSE_HANDLING" - } - ], - "title": "dataflow_v1beta3_generated_metrics_v1_beta3_get_stage_execution_details_async.py" - }, - { - "canonical": true, - "clientMethod": { - "client": { - "fullName": "google.cloud.dataflow_v1beta3.MetricsV1Beta3Client", - "shortName": "MetricsV1Beta3Client" - }, - "fullName": "google.cloud.dataflow_v1beta3.MetricsV1Beta3Client.get_stage_execution_details", - "method": { - "fullName": "google.dataflow.v1beta3.MetricsV1Beta3.GetStageExecutionDetails", - "service": { - "fullName": "google.dataflow.v1beta3.MetricsV1Beta3", - "shortName": "MetricsV1Beta3" - }, - "shortName": "GetStageExecutionDetails" - }, - "parameters": [ - { - "name": "request", - "type": "google.cloud.dataflow_v1beta3.types.GetStageExecutionDetailsRequest" - }, - { - "name": "retry", - "type": "google.api_core.retry.Retry" - }, - { - "name": "timeout", - "type": "float" - }, - { - "name": "metadata", - "type": "Sequence[Tuple[str, str]" - } - ], - "resultType": "google.cloud.dataflow_v1beta3.services.metrics_v1_beta3.pagers.GetStageExecutionDetailsPager", - "shortName": "get_stage_execution_details" - }, - "description": "Sample for GetStageExecutionDetails", - "file": "dataflow_v1beta3_generated_metrics_v1_beta3_get_stage_execution_details_sync_4acb3ded.py", - "language": "PYTHON", - "origin": "API_DEFINITION", - "regionTag": "dataflow_v1beta3_generated_MetricsV1Beta3_GetStageExecutionDetails_sync", - "segments": [ - { - "end": 51, - "start": 27, - "type": "FULL" - }, - { - "end": 51, - "start": 27, - "type": "SHORT" - }, - { - "end": 40, - "start": 38, - "type": "CLIENT_INITIALIZATION" - }, - { - "end": 44, - "start": 41, - "type": "REQUEST_INITIALIZATION" - }, - { - "end": 47, - "start": 45, - "type": "REQUEST_EXECUTION" - }, - { - "end": 52, - "start": 48, - "type": "RESPONSE_HANDLING" - } - ], - "title": "dataflow_v1beta3_generated_metrics_v1_beta3_get_stage_execution_details_sync_4acb3ded.py" - }, - { - "canonical": true, - "clientMethod": { - "client": { - "fullName": "google.cloud.dataflow_v1beta3.MetricsV1Beta3Client", - "shortName": "MetricsV1Beta3Client" - }, - "fullName": "google.cloud.dataflow_v1beta3.MetricsV1Beta3Client.get_stage_execution_details", - "method": { - "fullName": "google.dataflow.v1beta3.MetricsV1Beta3.GetStageExecutionDetails", - "service": { - "fullName": "google.dataflow.v1beta3.MetricsV1Beta3", - "shortName": "MetricsV1Beta3" - }, - "shortName": "GetStageExecutionDetails" - }, - "parameters": [ - { - "name": "request", - "type": "google.cloud.dataflow_v1beta3.types.GetStageExecutionDetailsRequest" - }, - { - "name": "retry", - "type": "google.api_core.retry.Retry" - }, - { - "name": "timeout", - "type": "float" - }, - { - "name": "metadata", - "type": "Sequence[Tuple[str, str]" - } - ], - "resultType": "google.cloud.dataflow_v1beta3.services.metrics_v1_beta3.pagers.GetStageExecutionDetailsPager", - "shortName": "get_stage_execution_details" - }, - "description": "Sample for GetStageExecutionDetails", - "file": "dataflow_v1beta3_generated_metrics_v1_beta3_get_stage_execution_details_sync_9490a11d.py", - "language": "PYTHON", - "origin": "API_DEFINITION", - "regionTag": "dataflow_v1beta3_generated_MetricsV1Beta3_GetStageExecutionDetails_sync", - "segments": [ - { - "end": 51, - "start": 27, - "type": "FULL" - }, - { - "end": 51, - "start": 27, - "type": "SHORT" - }, - { - "end": 40, - "start": 38, - "type": "CLIENT_INITIALIZATION" - }, - { - "end": 44, - "start": 41, - "type": "REQUEST_INITIALIZATION" - }, - { - "end": 47, - "start": 45, - "type": "REQUEST_EXECUTION" - }, - { - "end": 52, - "start": 48, - "type": "RESPONSE_HANDLING" - } - ], - "title": "dataflow_v1beta3_generated_metrics_v1_beta3_get_stage_execution_details_sync_9490a11d.py" - }, - { - "canonical": true, - "clientMethod": { - "async": true, - "client": { - "fullName": "google.cloud.dataflow_v1beta3.SnapshotsV1Beta3AsyncClient", - "shortName": "SnapshotsV1Beta3AsyncClient" - }, - "fullName": "google.cloud.dataflow_v1beta3.SnapshotsV1Beta3AsyncClient.delete_snapshot", - "method": { - "fullName": "google.dataflow.v1beta3.SnapshotsV1Beta3.DeleteSnapshot", - "service": { - "fullName": "google.dataflow.v1beta3.SnapshotsV1Beta3", - "shortName": "SnapshotsV1Beta3" - }, - "shortName": "DeleteSnapshot" - }, - "parameters": [ - { - "name": "request", - "type": "google.cloud.dataflow_v1beta3.types.DeleteSnapshotRequest" - }, - { - "name": "retry", - "type": "google.api_core.retry.Retry" - }, - { - "name": "timeout", - "type": "float" - }, - { - "name": "metadata", - "type": "Sequence[Tuple[str, str]" - } - ], - "resultType": "google.cloud.dataflow_v1beta3.types.DeleteSnapshotResponse", - "shortName": "delete_snapshot" - }, - "description": "Sample for DeleteSnapshot", - "file": "dataflow_v1beta3_generated_snapshots_v1_beta3_delete_snapshot_async.py", - "language": "PYTHON", - "origin": "API_DEFINITION", - "regionTag": "dataflow_v1beta3_generated_SnapshotsV1Beta3_DeleteSnapshot_async", - "segments": [ - { - "end": 50, - "start": 27, - "type": "FULL" - }, - { - "end": 50, - "start": 27, - "type": "SHORT" - }, - { - "end": 40, - "start": 38, - "type": "CLIENT_INITIALIZATION" - }, - { - "end": 44, - "start": 41, - "type": "REQUEST_INITIALIZATION" - }, - { - "end": 47, - "start": 45, - "type": "REQUEST_EXECUTION" - }, - { - "end": 51, - "start": 48, - "type": "RESPONSE_HANDLING" - } - ], - "title": "dataflow_v1beta3_generated_snapshots_v1_beta3_delete_snapshot_async.py" - }, - { - "canonical": true, - "clientMethod": { - "client": { - "fullName": "google.cloud.dataflow_v1beta3.SnapshotsV1Beta3Client", - "shortName": "SnapshotsV1Beta3Client" - }, - "fullName": "google.cloud.dataflow_v1beta3.SnapshotsV1Beta3Client.delete_snapshot", - "method": { - "fullName": "google.dataflow.v1beta3.SnapshotsV1Beta3.DeleteSnapshot", - "service": { - "fullName": "google.dataflow.v1beta3.SnapshotsV1Beta3", - "shortName": "SnapshotsV1Beta3" - }, - "shortName": "DeleteSnapshot" - }, - "parameters": [ - { - "name": "request", - "type": "google.cloud.dataflow_v1beta3.types.DeleteSnapshotRequest" - }, - { - "name": "retry", - "type": "google.api_core.retry.Retry" - }, - { - "name": "timeout", - "type": "float" - }, - { - "name": "metadata", - "type": "Sequence[Tuple[str, str]" - } - ], - "resultType": "google.cloud.dataflow_v1beta3.types.DeleteSnapshotResponse", - "shortName": "delete_snapshot" - }, - "description": "Sample for DeleteSnapshot", - "file": "dataflow_v1beta3_generated_snapshots_v1_beta3_delete_snapshot_sync_34b1dfd7.py", - "language": "PYTHON", - "origin": "API_DEFINITION", - "regionTag": "dataflow_v1beta3_generated_SnapshotsV1Beta3_DeleteSnapshot_sync", - "segments": [ - { - "end": 50, - "start": 27, - "type": "FULL" - }, - { - "end": 50, - "start": 27, - "type": "SHORT" - }, - { - "end": 40, - "start": 38, - "type": "CLIENT_INITIALIZATION" - }, - { - "end": 44, - "start": 41, - "type": "REQUEST_INITIALIZATION" - }, - { - "end": 47, - "start": 45, - "type": "REQUEST_EXECUTION" - }, { "end": 51, - "start": 48, - "type": "RESPONSE_HANDLING" - } - ], - "title": "dataflow_v1beta3_generated_snapshots_v1_beta3_delete_snapshot_sync_34b1dfd7.py" - }, - { - "canonical": true, - "clientMethod": { - "client": { - "fullName": "google.cloud.dataflow_v1beta3.SnapshotsV1Beta3Client", - "shortName": "SnapshotsV1Beta3Client" - }, - "fullName": "google.cloud.dataflow_v1beta3.SnapshotsV1Beta3Client.delete_snapshot", - "method": { - "fullName": "google.dataflow.v1beta3.SnapshotsV1Beta3.DeleteSnapshot", - "service": { - "fullName": "google.dataflow.v1beta3.SnapshotsV1Beta3", - "shortName": "SnapshotsV1Beta3" - }, - "shortName": "DeleteSnapshot" - }, - "parameters": [ - { - "name": "request", - "type": "google.cloud.dataflow_v1beta3.types.DeleteSnapshotRequest" - }, - { - "name": "retry", - "type": "google.api_core.retry.Retry" - }, - { - "name": "timeout", - "type": "float" - }, - { - "name": "metadata", - "type": "Sequence[Tuple[str, str]" - } - ], - "resultType": "google.cloud.dataflow_v1beta3.types.DeleteSnapshotResponse", - "shortName": "delete_snapshot" - }, - "description": "Sample for DeleteSnapshot", - "file": "dataflow_v1beta3_generated_snapshots_v1_beta3_delete_snapshot_sync_567e01e6.py", - "language": "PYTHON", - "origin": "API_DEFINITION", - "regionTag": "dataflow_v1beta3_generated_SnapshotsV1Beta3_DeleteSnapshot_sync", - "segments": [ - { - "end": 50, - "start": 27, - "type": "FULL" - }, - { - "end": 50, "start": 27, "type": "SHORT" }, @@ -2980,34 +1764,33 @@ "type": "REQUEST_EXECUTION" }, { - "end": 51, + "end": 52, "start": 48, "type": "RESPONSE_HANDLING" } ], - "title": "dataflow_v1beta3_generated_snapshots_v1_beta3_delete_snapshot_sync_567e01e6.py" + "title": "dataflow_v1beta3_generated_metrics_v1_beta3_get_stage_execution_details_async.py" }, { "canonical": true, "clientMethod": { - "async": true, "client": { - "fullName": "google.cloud.dataflow_v1beta3.SnapshotsV1Beta3AsyncClient", - "shortName": "SnapshotsV1Beta3AsyncClient" + "fullName": "google.cloud.dataflow_v1beta3.MetricsV1Beta3Client", + "shortName": "MetricsV1Beta3Client" }, - "fullName": "google.cloud.dataflow_v1beta3.SnapshotsV1Beta3AsyncClient.get_snapshot", + "fullName": "google.cloud.dataflow_v1beta3.MetricsV1Beta3Client.get_stage_execution_details", "method": { - "fullName": "google.dataflow.v1beta3.SnapshotsV1Beta3.GetSnapshot", + "fullName": "google.dataflow.v1beta3.MetricsV1Beta3.GetStageExecutionDetails", "service": { - "fullName": "google.dataflow.v1beta3.SnapshotsV1Beta3", - "shortName": "SnapshotsV1Beta3" + "fullName": "google.dataflow.v1beta3.MetricsV1Beta3", + "shortName": "MetricsV1Beta3" }, - "shortName": "GetSnapshot" + "shortName": "GetStageExecutionDetails" }, "parameters": [ { "name": "request", - "type": "google.cloud.dataflow_v1beta3.types.GetSnapshotRequest" + "type": "google.cloud.dataflow_v1beta3.types.GetStageExecutionDetailsRequest" }, { "name": "retry", @@ -3022,174 +1805,22 @@ "type": "Sequence[Tuple[str, str]" } ], - "resultType": "google.cloud.dataflow_v1beta3.types.Snapshot", - "shortName": "get_snapshot" + "resultType": "google.cloud.dataflow_v1beta3.services.metrics_v1_beta3.pagers.GetStageExecutionDetailsPager", + "shortName": "get_stage_execution_details" }, - "description": "Sample for GetSnapshot", - "file": "dataflow_v1beta3_generated_snapshots_v1_beta3_get_snapshot_async.py", + "description": "Sample for GetStageExecutionDetails", + "file": "dataflow_v1beta3_generated_metrics_v1_beta3_get_stage_execution_details_sync.py", "language": "PYTHON", "origin": "API_DEFINITION", - "regionTag": "dataflow_v1beta3_generated_SnapshotsV1Beta3_GetSnapshot_async", + "regionTag": "dataflow_v1beta3_generated_MetricsV1Beta3_GetStageExecutionDetails_sync", "segments": [ - { - "end": 50, - "start": 27, - "type": "FULL" - }, - { - "end": 50, - "start": 27, - "type": "SHORT" - }, - { - "end": 40, - "start": 38, - "type": "CLIENT_INITIALIZATION" - }, - { - "end": 44, - "start": 41, - "type": "REQUEST_INITIALIZATION" - }, - { - "end": 47, - "start": 45, - "type": "REQUEST_EXECUTION" - }, { "end": 51, - "start": 48, - "type": "RESPONSE_HANDLING" - } - ], - "title": "dataflow_v1beta3_generated_snapshots_v1_beta3_get_snapshot_async.py" - }, - { - "canonical": true, - "clientMethod": { - "client": { - "fullName": "google.cloud.dataflow_v1beta3.SnapshotsV1Beta3Client", - "shortName": "SnapshotsV1Beta3Client" - }, - "fullName": "google.cloud.dataflow_v1beta3.SnapshotsV1Beta3Client.get_snapshot", - "method": { - "fullName": "google.dataflow.v1beta3.SnapshotsV1Beta3.GetSnapshot", - "service": { - "fullName": "google.dataflow.v1beta3.SnapshotsV1Beta3", - "shortName": "SnapshotsV1Beta3" - }, - "shortName": "GetSnapshot" - }, - "parameters": [ - { - "name": "request", - "type": "google.cloud.dataflow_v1beta3.types.GetSnapshotRequest" - }, - { - "name": "retry", - "type": "google.api_core.retry.Retry" - }, - { - "name": "timeout", - "type": "float" - }, - { - "name": "metadata", - "type": "Sequence[Tuple[str, str]" - } - ], - "resultType": "google.cloud.dataflow_v1beta3.types.Snapshot", - "shortName": "get_snapshot" - }, - "description": "Sample for GetSnapshot", - "file": "dataflow_v1beta3_generated_snapshots_v1_beta3_get_snapshot_sync_846de01f.py", - "language": "PYTHON", - "origin": "API_DEFINITION", - "regionTag": "dataflow_v1beta3_generated_SnapshotsV1Beta3_GetSnapshot_sync", - "segments": [ - { - "end": 50, "start": 27, "type": "FULL" }, - { - "end": 50, - "start": 27, - "type": "SHORT" - }, - { - "end": 40, - "start": 38, - "type": "CLIENT_INITIALIZATION" - }, - { - "end": 44, - "start": 41, - "type": "REQUEST_INITIALIZATION" - }, - { - "end": 47, - "start": 45, - "type": "REQUEST_EXECUTION" - }, { "end": 51, - "start": 48, - "type": "RESPONSE_HANDLING" - } - ], - "title": "dataflow_v1beta3_generated_snapshots_v1_beta3_get_snapshot_sync_846de01f.py" - }, - { - "canonical": true, - "clientMethod": { - "client": { - "fullName": "google.cloud.dataflow_v1beta3.SnapshotsV1Beta3Client", - "shortName": "SnapshotsV1Beta3Client" - }, - "fullName": "google.cloud.dataflow_v1beta3.SnapshotsV1Beta3Client.get_snapshot", - "method": { - "fullName": "google.dataflow.v1beta3.SnapshotsV1Beta3.GetSnapshot", - "service": { - "fullName": "google.dataflow.v1beta3.SnapshotsV1Beta3", - "shortName": "SnapshotsV1Beta3" - }, - "shortName": "GetSnapshot" - }, - "parameters": [ - { - "name": "request", - "type": "google.cloud.dataflow_v1beta3.types.GetSnapshotRequest" - }, - { - "name": "retry", - "type": "google.api_core.retry.Retry" - }, - { - "name": "timeout", - "type": "float" - }, - { - "name": "metadata", - "type": "Sequence[Tuple[str, str]" - } - ], - "resultType": "google.cloud.dataflow_v1beta3.types.Snapshot", - "shortName": "get_snapshot" - }, - "description": "Sample for GetSnapshot", - "file": "dataflow_v1beta3_generated_snapshots_v1_beta3_get_snapshot_sync_ba1ca2b6.py", - "language": "PYTHON", - "origin": "API_DEFINITION", - "regionTag": "dataflow_v1beta3_generated_SnapshotsV1Beta3_GetSnapshot_sync", - "segments": [ - { - "end": 50, - "start": 27, - "type": "FULL" - }, - { - "end": 50, "start": 27, "type": "SHORT" }, @@ -3209,12 +1840,12 @@ "type": "REQUEST_EXECUTION" }, { - "end": 51, + "end": 52, "start": 48, "type": "RESPONSE_HANDLING" } ], - "title": "dataflow_v1beta3_generated_snapshots_v1_beta3_get_snapshot_sync_ba1ca2b6.py" + "title": "dataflow_v1beta3_generated_metrics_v1_beta3_get_stage_execution_details_sync.py" }, { "canonical": true, @@ -3224,19 +1855,19 @@ "fullName": "google.cloud.dataflow_v1beta3.SnapshotsV1Beta3AsyncClient", "shortName": "SnapshotsV1Beta3AsyncClient" }, - "fullName": "google.cloud.dataflow_v1beta3.SnapshotsV1Beta3AsyncClient.list_snapshots", + "fullName": "google.cloud.dataflow_v1beta3.SnapshotsV1Beta3AsyncClient.delete_snapshot", "method": { - "fullName": "google.dataflow.v1beta3.SnapshotsV1Beta3.ListSnapshots", + "fullName": "google.dataflow.v1beta3.SnapshotsV1Beta3.DeleteSnapshot", "service": { "fullName": "google.dataflow.v1beta3.SnapshotsV1Beta3", "shortName": "SnapshotsV1Beta3" }, - "shortName": "ListSnapshots" + "shortName": "DeleteSnapshot" }, "parameters": [ { "name": "request", - "type": "google.cloud.dataflow_v1beta3.types.ListSnapshotsRequest" + "type": "google.cloud.dataflow_v1beta3.types.DeleteSnapshotRequest" }, { "name": "retry", @@ -3251,14 +1882,14 @@ "type": "Sequence[Tuple[str, str]" } ], - "resultType": "google.cloud.dataflow_v1beta3.types.ListSnapshotsResponse", - "shortName": "list_snapshots" + "resultType": "google.cloud.dataflow_v1beta3.types.DeleteSnapshotResponse", + "shortName": "delete_snapshot" }, - "description": "Sample for ListSnapshots", - "file": "dataflow_v1beta3_generated_snapshots_v1_beta3_list_snapshots_async.py", + "description": "Sample for DeleteSnapshot", + "file": "dataflow_v1beta3_generated_snapshots_v1_beta3_delete_snapshot_async.py", "language": "PYTHON", "origin": "API_DEFINITION", - "regionTag": "dataflow_v1beta3_generated_SnapshotsV1Beta3_ListSnapshots_async", + "regionTag": "dataflow_v1beta3_generated_SnapshotsV1Beta3_DeleteSnapshot_async", "segments": [ { "end": 50, @@ -3291,7 +1922,7 @@ "type": "RESPONSE_HANDLING" } ], - "title": "dataflow_v1beta3_generated_snapshots_v1_beta3_list_snapshots_async.py" + "title": "dataflow_v1beta3_generated_snapshots_v1_beta3_delete_snapshot_async.py" }, { "canonical": true, @@ -3300,19 +1931,19 @@ "fullName": "google.cloud.dataflow_v1beta3.SnapshotsV1Beta3Client", "shortName": "SnapshotsV1Beta3Client" }, - "fullName": "google.cloud.dataflow_v1beta3.SnapshotsV1Beta3Client.list_snapshots", + "fullName": "google.cloud.dataflow_v1beta3.SnapshotsV1Beta3Client.delete_snapshot", "method": { - "fullName": "google.dataflow.v1beta3.SnapshotsV1Beta3.ListSnapshots", + "fullName": "google.dataflow.v1beta3.SnapshotsV1Beta3.DeleteSnapshot", "service": { "fullName": "google.dataflow.v1beta3.SnapshotsV1Beta3", "shortName": "SnapshotsV1Beta3" }, - "shortName": "ListSnapshots" + "shortName": "DeleteSnapshot" }, "parameters": [ { "name": "request", - "type": "google.cloud.dataflow_v1beta3.types.ListSnapshotsRequest" + "type": "google.cloud.dataflow_v1beta3.types.DeleteSnapshotRequest" }, { "name": "retry", @@ -3327,14 +1958,14 @@ "type": "Sequence[Tuple[str, str]" } ], - "resultType": "google.cloud.dataflow_v1beta3.types.ListSnapshotsResponse", - "shortName": "list_snapshots" + "resultType": "google.cloud.dataflow_v1beta3.types.DeleteSnapshotResponse", + "shortName": "delete_snapshot" }, - "description": "Sample for ListSnapshots", - "file": "dataflow_v1beta3_generated_snapshots_v1_beta3_list_snapshots_sync_6da08cb9.py", + "description": "Sample for DeleteSnapshot", + "file": "dataflow_v1beta3_generated_snapshots_v1_beta3_delete_snapshot_sync.py", "language": "PYTHON", "origin": "API_DEFINITION", - "regionTag": "dataflow_v1beta3_generated_SnapshotsV1Beta3_ListSnapshots_sync", + "regionTag": "dataflow_v1beta3_generated_SnapshotsV1Beta3_DeleteSnapshot_sync", "segments": [ { "end": 50, @@ -3367,28 +1998,29 @@ "type": "RESPONSE_HANDLING" } ], - "title": "dataflow_v1beta3_generated_snapshots_v1_beta3_list_snapshots_sync_6da08cb9.py" + "title": "dataflow_v1beta3_generated_snapshots_v1_beta3_delete_snapshot_sync.py" }, { "canonical": true, "clientMethod": { + "async": true, "client": { - "fullName": "google.cloud.dataflow_v1beta3.SnapshotsV1Beta3Client", - "shortName": "SnapshotsV1Beta3Client" + "fullName": "google.cloud.dataflow_v1beta3.SnapshotsV1Beta3AsyncClient", + "shortName": "SnapshotsV1Beta3AsyncClient" }, - "fullName": "google.cloud.dataflow_v1beta3.SnapshotsV1Beta3Client.list_snapshots", + "fullName": "google.cloud.dataflow_v1beta3.SnapshotsV1Beta3AsyncClient.get_snapshot", "method": { - "fullName": "google.dataflow.v1beta3.SnapshotsV1Beta3.ListSnapshots", + "fullName": "google.dataflow.v1beta3.SnapshotsV1Beta3.GetSnapshot", "service": { "fullName": "google.dataflow.v1beta3.SnapshotsV1Beta3", "shortName": "SnapshotsV1Beta3" }, - "shortName": "ListSnapshots" + "shortName": "GetSnapshot" }, "parameters": [ { "name": "request", - "type": "google.cloud.dataflow_v1beta3.types.ListSnapshotsRequest" + "type": "google.cloud.dataflow_v1beta3.types.GetSnapshotRequest" }, { "name": "retry", @@ -3403,14 +2035,14 @@ "type": "Sequence[Tuple[str, str]" } ], - "resultType": "google.cloud.dataflow_v1beta3.types.ListSnapshotsResponse", - "shortName": "list_snapshots" + "resultType": "google.cloud.dataflow_v1beta3.types.Snapshot", + "shortName": "get_snapshot" }, - "description": "Sample for ListSnapshots", - "file": "dataflow_v1beta3_generated_snapshots_v1_beta3_list_snapshots_sync_425e5024.py", + "description": "Sample for GetSnapshot", + "file": "dataflow_v1beta3_generated_snapshots_v1_beta3_get_snapshot_async.py", "language": "PYTHON", "origin": "API_DEFINITION", - "regionTag": "dataflow_v1beta3_generated_SnapshotsV1Beta3_ListSnapshots_sync", + "regionTag": "dataflow_v1beta3_generated_SnapshotsV1Beta3_GetSnapshot_async", "segments": [ { "end": 50, @@ -3443,29 +2075,28 @@ "type": "RESPONSE_HANDLING" } ], - "title": "dataflow_v1beta3_generated_snapshots_v1_beta3_list_snapshots_sync_425e5024.py" + "title": "dataflow_v1beta3_generated_snapshots_v1_beta3_get_snapshot_async.py" }, { "canonical": true, "clientMethod": { - "async": true, "client": { - "fullName": "google.cloud.dataflow_v1beta3.TemplatesServiceAsyncClient", - "shortName": "TemplatesServiceAsyncClient" + "fullName": "google.cloud.dataflow_v1beta3.SnapshotsV1Beta3Client", + "shortName": "SnapshotsV1Beta3Client" }, - "fullName": "google.cloud.dataflow_v1beta3.TemplatesServiceAsyncClient.create_job_from_template", + "fullName": "google.cloud.dataflow_v1beta3.SnapshotsV1Beta3Client.get_snapshot", "method": { - "fullName": "google.dataflow.v1beta3.TemplatesService.CreateJobFromTemplate", + "fullName": "google.dataflow.v1beta3.SnapshotsV1Beta3.GetSnapshot", "service": { - "fullName": "google.dataflow.v1beta3.TemplatesService", - "shortName": "TemplatesService" + "fullName": "google.dataflow.v1beta3.SnapshotsV1Beta3", + "shortName": "SnapshotsV1Beta3" }, - "shortName": "CreateJobFromTemplate" + "shortName": "GetSnapshot" }, "parameters": [ { "name": "request", - "type": "google.cloud.dataflow_v1beta3.types.CreateJobFromTemplateRequest" + "type": "google.cloud.dataflow_v1beta3.types.GetSnapshotRequest" }, { "name": "retry", @@ -3480,22 +2111,22 @@ "type": "Sequence[Tuple[str, str]" } ], - "resultType": "google.cloud.dataflow_v1beta3.types.Job", - "shortName": "create_job_from_template" + "resultType": "google.cloud.dataflow_v1beta3.types.Snapshot", + "shortName": "get_snapshot" }, - "description": "Sample for CreateJobFromTemplate", - "file": "dataflow_v1beta3_generated_templates_service_create_job_from_template_async.py", + "description": "Sample for GetSnapshot", + "file": "dataflow_v1beta3_generated_snapshots_v1_beta3_get_snapshot_sync.py", "language": "PYTHON", "origin": "API_DEFINITION", - "regionTag": "dataflow_v1beta3_generated_TemplatesService_CreateJobFromTemplate_async", + "regionTag": "dataflow_v1beta3_generated_SnapshotsV1Beta3_GetSnapshot_sync", "segments": [ { - "end": 51, + "end": 50, "start": 27, "type": "FULL" }, { - "end": 51, + "end": 50, "start": 27, "type": "SHORT" }, @@ -3505,43 +2136,44 @@ "type": "CLIENT_INITIALIZATION" }, { - "end": 45, + "end": 44, "start": 41, "type": "REQUEST_INITIALIZATION" }, { - "end": 48, - "start": 46, + "end": 47, + "start": 45, "type": "REQUEST_EXECUTION" }, { - "end": 52, - "start": 49, + "end": 51, + "start": 48, "type": "RESPONSE_HANDLING" } ], - "title": "dataflow_v1beta3_generated_templates_service_create_job_from_template_async.py" + "title": "dataflow_v1beta3_generated_snapshots_v1_beta3_get_snapshot_sync.py" }, { "canonical": true, "clientMethod": { + "async": true, "client": { - "fullName": "google.cloud.dataflow_v1beta3.TemplatesServiceClient", - "shortName": "TemplatesServiceClient" + "fullName": "google.cloud.dataflow_v1beta3.SnapshotsV1Beta3AsyncClient", + "shortName": "SnapshotsV1Beta3AsyncClient" }, - "fullName": "google.cloud.dataflow_v1beta3.TemplatesServiceClient.create_job_from_template", + "fullName": "google.cloud.dataflow_v1beta3.SnapshotsV1Beta3AsyncClient.list_snapshots", "method": { - "fullName": "google.dataflow.v1beta3.TemplatesService.CreateJobFromTemplate", + "fullName": "google.dataflow.v1beta3.SnapshotsV1Beta3.ListSnapshots", "service": { - "fullName": "google.dataflow.v1beta3.TemplatesService", - "shortName": "TemplatesService" + "fullName": "google.dataflow.v1beta3.SnapshotsV1Beta3", + "shortName": "SnapshotsV1Beta3" }, - "shortName": "CreateJobFromTemplate" + "shortName": "ListSnapshots" }, "parameters": [ { "name": "request", - "type": "google.cloud.dataflow_v1beta3.types.CreateJobFromTemplateRequest" + "type": "google.cloud.dataflow_v1beta3.types.ListSnapshotsRequest" }, { "name": "retry", @@ -3556,22 +2188,22 @@ "type": "Sequence[Tuple[str, str]" } ], - "resultType": "google.cloud.dataflow_v1beta3.types.Job", - "shortName": "create_job_from_template" + "resultType": "google.cloud.dataflow_v1beta3.types.ListSnapshotsResponse", + "shortName": "list_snapshots" }, - "description": "Sample for CreateJobFromTemplate", - "file": "dataflow_v1beta3_generated_templates_service_create_job_from_template_sync_9e26a4bf.py", + "description": "Sample for ListSnapshots", + "file": "dataflow_v1beta3_generated_snapshots_v1_beta3_list_snapshots_async.py", "language": "PYTHON", "origin": "API_DEFINITION", - "regionTag": "dataflow_v1beta3_generated_TemplatesService_CreateJobFromTemplate_sync", + "regionTag": "dataflow_v1beta3_generated_SnapshotsV1Beta3_ListSnapshots_async", "segments": [ { - "end": 51, + "end": 50, "start": 27, "type": "FULL" }, { - "end": 51, + "end": 50, "start": 27, "type": "SHORT" }, @@ -3581,43 +2213,43 @@ "type": "CLIENT_INITIALIZATION" }, { - "end": 45, + "end": 44, "start": 41, "type": "REQUEST_INITIALIZATION" }, { - "end": 48, - "start": 46, + "end": 47, + "start": 45, "type": "REQUEST_EXECUTION" }, { - "end": 52, - "start": 49, + "end": 51, + "start": 48, "type": "RESPONSE_HANDLING" } ], - "title": "dataflow_v1beta3_generated_templates_service_create_job_from_template_sync_9e26a4bf.py" + "title": "dataflow_v1beta3_generated_snapshots_v1_beta3_list_snapshots_async.py" }, { "canonical": true, "clientMethod": { "client": { - "fullName": "google.cloud.dataflow_v1beta3.TemplatesServiceClient", - "shortName": "TemplatesServiceClient" + "fullName": "google.cloud.dataflow_v1beta3.SnapshotsV1Beta3Client", + "shortName": "SnapshotsV1Beta3Client" }, - "fullName": "google.cloud.dataflow_v1beta3.TemplatesServiceClient.create_job_from_template", + "fullName": "google.cloud.dataflow_v1beta3.SnapshotsV1Beta3Client.list_snapshots", "method": { - "fullName": "google.dataflow.v1beta3.TemplatesService.CreateJobFromTemplate", + "fullName": "google.dataflow.v1beta3.SnapshotsV1Beta3.ListSnapshots", "service": { - "fullName": "google.dataflow.v1beta3.TemplatesService", - "shortName": "TemplatesService" + "fullName": "google.dataflow.v1beta3.SnapshotsV1Beta3", + "shortName": "SnapshotsV1Beta3" }, - "shortName": "CreateJobFromTemplate" + "shortName": "ListSnapshots" }, "parameters": [ { "name": "request", - "type": "google.cloud.dataflow_v1beta3.types.CreateJobFromTemplateRequest" + "type": "google.cloud.dataflow_v1beta3.types.ListSnapshotsRequest" }, { "name": "retry", @@ -3632,22 +2264,22 @@ "type": "Sequence[Tuple[str, str]" } ], - "resultType": "google.cloud.dataflow_v1beta3.types.Job", - "shortName": "create_job_from_template" + "resultType": "google.cloud.dataflow_v1beta3.types.ListSnapshotsResponse", + "shortName": "list_snapshots" }, - "description": "Sample for CreateJobFromTemplate", - "file": "dataflow_v1beta3_generated_templates_service_create_job_from_template_sync_e2cb482f.py", + "description": "Sample for ListSnapshots", + "file": "dataflow_v1beta3_generated_snapshots_v1_beta3_list_snapshots_sync.py", "language": "PYTHON", "origin": "API_DEFINITION", - "regionTag": "dataflow_v1beta3_generated_TemplatesService_CreateJobFromTemplate_sync", + "regionTag": "dataflow_v1beta3_generated_SnapshotsV1Beta3_ListSnapshots_sync", "segments": [ { - "end": 51, + "end": 50, "start": 27, "type": "FULL" }, { - "end": 51, + "end": 50, "start": 27, "type": "SHORT" }, @@ -3657,22 +2289,22 @@ "type": "CLIENT_INITIALIZATION" }, { - "end": 45, + "end": 44, "start": 41, "type": "REQUEST_INITIALIZATION" }, { - "end": 48, - "start": 46, + "end": 47, + "start": 45, "type": "REQUEST_EXECUTION" }, { - "end": 52, - "start": 49, + "end": 51, + "start": 48, "type": "RESPONSE_HANDLING" } ], - "title": "dataflow_v1beta3_generated_templates_service_create_job_from_template_sync_e2cb482f.py" + "title": "dataflow_v1beta3_generated_snapshots_v1_beta3_list_snapshots_sync.py" }, { "canonical": true, @@ -3682,19 +2314,19 @@ "fullName": "google.cloud.dataflow_v1beta3.TemplatesServiceAsyncClient", "shortName": "TemplatesServiceAsyncClient" }, - "fullName": "google.cloud.dataflow_v1beta3.TemplatesServiceAsyncClient.get_template", + "fullName": "google.cloud.dataflow_v1beta3.TemplatesServiceAsyncClient.create_job_from_template", "method": { - "fullName": "google.dataflow.v1beta3.TemplatesService.GetTemplate", + "fullName": "google.dataflow.v1beta3.TemplatesService.CreateJobFromTemplate", "service": { "fullName": "google.dataflow.v1beta3.TemplatesService", "shortName": "TemplatesService" }, - "shortName": "GetTemplate" + "shortName": "CreateJobFromTemplate" }, "parameters": [ { "name": "request", - "type": "google.cloud.dataflow_v1beta3.types.GetTemplateRequest" + "type": "google.cloud.dataflow_v1beta3.types.CreateJobFromTemplateRequest" }, { "name": "retry", @@ -3709,14 +2341,14 @@ "type": "Sequence[Tuple[str, str]" } ], - "resultType": "google.cloud.dataflow_v1beta3.types.GetTemplateResponse", - "shortName": "get_template" + "resultType": "google.cloud.dataflow_v1beta3.types.Job", + "shortName": "create_job_from_template" }, - "description": "Sample for GetTemplate", - "file": "dataflow_v1beta3_generated_templates_service_get_template_async.py", + "description": "Sample for CreateJobFromTemplate", + "file": "dataflow_v1beta3_generated_templates_service_create_job_from_template_async.py", "language": "PYTHON", "origin": "API_DEFINITION", - "regionTag": "dataflow_v1beta3_generated_TemplatesService_GetTemplate_async", + "regionTag": "dataflow_v1beta3_generated_TemplatesService_CreateJobFromTemplate_async", "segments": [ { "end": 51, @@ -3749,7 +2381,7 @@ "type": "RESPONSE_HANDLING" } ], - "title": "dataflow_v1beta3_generated_templates_service_get_template_async.py" + "title": "dataflow_v1beta3_generated_templates_service_create_job_from_template_async.py" }, { "canonical": true, @@ -3758,19 +2390,19 @@ "fullName": "google.cloud.dataflow_v1beta3.TemplatesServiceClient", "shortName": "TemplatesServiceClient" }, - "fullName": "google.cloud.dataflow_v1beta3.TemplatesServiceClient.get_template", + "fullName": "google.cloud.dataflow_v1beta3.TemplatesServiceClient.create_job_from_template", "method": { - "fullName": "google.dataflow.v1beta3.TemplatesService.GetTemplate", + "fullName": "google.dataflow.v1beta3.TemplatesService.CreateJobFromTemplate", "service": { "fullName": "google.dataflow.v1beta3.TemplatesService", "shortName": "TemplatesService" }, - "shortName": "GetTemplate" + "shortName": "CreateJobFromTemplate" }, "parameters": [ { "name": "request", - "type": "google.cloud.dataflow_v1beta3.types.GetTemplateRequest" + "type": "google.cloud.dataflow_v1beta3.types.CreateJobFromTemplateRequest" }, { "name": "retry", @@ -3785,14 +2417,14 @@ "type": "Sequence[Tuple[str, str]" } ], - "resultType": "google.cloud.dataflow_v1beta3.types.GetTemplateResponse", - "shortName": "get_template" + "resultType": "google.cloud.dataflow_v1beta3.types.Job", + "shortName": "create_job_from_template" }, - "description": "Sample for GetTemplate", - "file": "dataflow_v1beta3_generated_templates_service_get_template_sync_732e6209.py", + "description": "Sample for CreateJobFromTemplate", + "file": "dataflow_v1beta3_generated_templates_service_create_job_from_template_sync.py", "language": "PYTHON", "origin": "API_DEFINITION", - "regionTag": "dataflow_v1beta3_generated_TemplatesService_GetTemplate_sync", + "regionTag": "dataflow_v1beta3_generated_TemplatesService_CreateJobFromTemplate_sync", "segments": [ { "end": 51, @@ -3825,16 +2457,17 @@ "type": "RESPONSE_HANDLING" } ], - "title": "dataflow_v1beta3_generated_templates_service_get_template_sync_732e6209.py" + "title": "dataflow_v1beta3_generated_templates_service_create_job_from_template_sync.py" }, { "canonical": true, "clientMethod": { + "async": true, "client": { - "fullName": "google.cloud.dataflow_v1beta3.TemplatesServiceClient", - "shortName": "TemplatesServiceClient" + "fullName": "google.cloud.dataflow_v1beta3.TemplatesServiceAsyncClient", + "shortName": "TemplatesServiceAsyncClient" }, - "fullName": "google.cloud.dataflow_v1beta3.TemplatesServiceClient.get_template", + "fullName": "google.cloud.dataflow_v1beta3.TemplatesServiceAsyncClient.get_template", "method": { "fullName": "google.dataflow.v1beta3.TemplatesService.GetTemplate", "service": { @@ -3865,10 +2498,10 @@ "shortName": "get_template" }, "description": "Sample for GetTemplate", - "file": "dataflow_v1beta3_generated_templates_service_get_template_sync_07cd261a.py", + "file": "dataflow_v1beta3_generated_templates_service_get_template_async.py", "language": "PYTHON", "origin": "API_DEFINITION", - "regionTag": "dataflow_v1beta3_generated_TemplatesService_GetTemplate_sync", + "regionTag": "dataflow_v1beta3_generated_TemplatesService_GetTemplate_async", "segments": [ { "end": 51, @@ -3901,29 +2534,28 @@ "type": "RESPONSE_HANDLING" } ], - "title": "dataflow_v1beta3_generated_templates_service_get_template_sync_07cd261a.py" + "title": "dataflow_v1beta3_generated_templates_service_get_template_async.py" }, { "canonical": true, "clientMethod": { - "async": true, "client": { - "fullName": "google.cloud.dataflow_v1beta3.TemplatesServiceAsyncClient", - "shortName": "TemplatesServiceAsyncClient" + "fullName": "google.cloud.dataflow_v1beta3.TemplatesServiceClient", + "shortName": "TemplatesServiceClient" }, - "fullName": "google.cloud.dataflow_v1beta3.TemplatesServiceAsyncClient.launch_template", + "fullName": "google.cloud.dataflow_v1beta3.TemplatesServiceClient.get_template", "method": { - "fullName": "google.dataflow.v1beta3.TemplatesService.LaunchTemplate", + "fullName": "google.dataflow.v1beta3.TemplatesService.GetTemplate", "service": { "fullName": "google.dataflow.v1beta3.TemplatesService", "shortName": "TemplatesService" }, - "shortName": "LaunchTemplate" + "shortName": "GetTemplate" }, "parameters": [ { "name": "request", - "type": "google.cloud.dataflow_v1beta3.types.LaunchTemplateRequest" + "type": "google.cloud.dataflow_v1beta3.types.GetTemplateRequest" }, { "name": "retry", @@ -3938,14 +2570,14 @@ "type": "Sequence[Tuple[str, str]" } ], - "resultType": "google.cloud.dataflow_v1beta3.types.LaunchTemplateResponse", - "shortName": "launch_template" + "resultType": "google.cloud.dataflow_v1beta3.types.GetTemplateResponse", + "shortName": "get_template" }, - "description": "Sample for LaunchTemplate", - "file": "dataflow_v1beta3_generated_templates_service_launch_template_async.py", + "description": "Sample for GetTemplate", + "file": "dataflow_v1beta3_generated_templates_service_get_template_sync.py", "language": "PYTHON", "origin": "API_DEFINITION", - "regionTag": "dataflow_v1beta3_generated_TemplatesService_LaunchTemplate_async", + "regionTag": "dataflow_v1beta3_generated_TemplatesService_GetTemplate_sync", "segments": [ { "end": 51, @@ -3978,16 +2610,17 @@ "type": "RESPONSE_HANDLING" } ], - "title": "dataflow_v1beta3_generated_templates_service_launch_template_async.py" + "title": "dataflow_v1beta3_generated_templates_service_get_template_sync.py" }, { "canonical": true, "clientMethod": { + "async": true, "client": { - "fullName": "google.cloud.dataflow_v1beta3.TemplatesServiceClient", - "shortName": "TemplatesServiceClient" + "fullName": "google.cloud.dataflow_v1beta3.TemplatesServiceAsyncClient", + "shortName": "TemplatesServiceAsyncClient" }, - "fullName": "google.cloud.dataflow_v1beta3.TemplatesServiceClient.launch_template", + "fullName": "google.cloud.dataflow_v1beta3.TemplatesServiceAsyncClient.launch_template", "method": { "fullName": "google.dataflow.v1beta3.TemplatesService.LaunchTemplate", "service": { @@ -4018,10 +2651,10 @@ "shortName": "launch_template" }, "description": "Sample for LaunchTemplate", - "file": "dataflow_v1beta3_generated_templates_service_launch_template_sync_77764eb9.py", + "file": "dataflow_v1beta3_generated_templates_service_launch_template_async.py", "language": "PYTHON", "origin": "API_DEFINITION", - "regionTag": "dataflow_v1beta3_generated_TemplatesService_LaunchTemplate_sync", + "regionTag": "dataflow_v1beta3_generated_TemplatesService_LaunchTemplate_async", "segments": [ { "end": 51, @@ -4054,7 +2687,7 @@ "type": "RESPONSE_HANDLING" } ], - "title": "dataflow_v1beta3_generated_templates_service_launch_template_sync_77764eb9.py" + "title": "dataflow_v1beta3_generated_templates_service_launch_template_async.py" }, { "canonical": true, @@ -4094,7 +2727,7 @@ "shortName": "launch_template" }, "description": "Sample for LaunchTemplate", - "file": "dataflow_v1beta3_generated_templates_service_launch_template_sync_140179ca.py", + "file": "dataflow_v1beta3_generated_templates_service_launch_template_sync.py", "language": "PYTHON", "origin": "API_DEFINITION", "regionTag": "dataflow_v1beta3_generated_TemplatesService_LaunchTemplate_sync", @@ -4130,7 +2763,7 @@ "type": "RESPONSE_HANDLING" } ], - "title": "dataflow_v1beta3_generated_templates_service_launch_template_sync_140179ca.py" + "title": "dataflow_v1beta3_generated_templates_service_launch_template_sync.py" } ] } diff --git a/setup.py b/setup.py index ac10bd5..b1fc892 100644 --- a/setup.py +++ b/setup.py @@ -1,34 +1,46 @@ # -*- coding: utf-8 -*- -# -# Copyright 2020 Google LLC +# Copyright 2022 Google LLC # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # -# https://www.apache.org/licenses/LICENSE-2.0 +# http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. - +# import io import os -import setuptools +import setuptools # type: ignore + +package_root = os.path.abspath(os.path.dirname(__file__)) name = "google-cloud-dataflow-client" -description = "Cloud Dataflow API client library" -version = "0.6.2" -release_status = "Development Status :: 4 - Beta" -url = "https://github.com/googleapis/python-dataflow-client" + + +description = "Google Cloud Dataflow Client API client library" + +version = {} +with open(os.path.join(package_root, "google/cloud/dataflow/gapic_version.py")) as fp: + exec(fp.read(), version) +version = version["__version__"] + +if version[0] == "0": + release_status = "Development Status :: 4 - Beta" +else: + release_status = "Development Status :: 5 - Production/Stable" + dependencies = [ - "google-api-core[grpc] >= 1.33.2, <3.0.0dev,!=2.0.*,!=2.1.*,!=2.2.*,!=2.3.*,!=2.4.*,!=2.5.*,!=2.6.*,!=2.7.*", + "google-api-core[grpc] >= 1.34.0, <3.0.0dev,!=2.0.*,!=2.1.*,!=2.2.*,!=2.3.*,!=2.4.*,!=2.5.*,!=2.6.*,!=2.7.*,!=2.8.*,!=2.9.*,!=2.10.*", "proto-plus >= 1.22.0, <2.0.0dev", "protobuf>=3.19.5,<5.0.0dev,!=3.20.0,!=3.20.1,!=4.21.0,!=4.21.1,!=4.21.2,!=4.21.3,!=4.21.4,!=4.21.5", ] +url = "https://github.com/googleapis/python-dataflow-client" package_root = os.path.abspath(os.path.dirname(__file__)) @@ -46,7 +58,6 @@ if "google.cloud" in packages: namespaces.append("google.cloud") - setuptools.setup( name=name, version=version, diff --git a/testing/constraints-3.10.txt b/testing/constraints-3.10.txt index e69de29..ed7f9ae 100644 --- a/testing/constraints-3.10.txt +++ b/testing/constraints-3.10.txt @@ -0,0 +1,6 @@ +# -*- coding: utf-8 -*- +# This constraints file is required for unit tests. +# List all library dependencies and extras in this file. +google-api-core +proto-plus +protobuf diff --git a/testing/constraints-3.11.txt b/testing/constraints-3.11.txt index e69de29..ed7f9ae 100644 --- a/testing/constraints-3.11.txt +++ b/testing/constraints-3.11.txt @@ -0,0 +1,6 @@ +# -*- coding: utf-8 -*- +# This constraints file is required for unit tests. +# List all library dependencies and extras in this file. +google-api-core +proto-plus +protobuf diff --git a/testing/constraints-3.7.txt b/testing/constraints-3.7.txt index 6f3158c..6c44adf 100644 --- a/testing/constraints-3.7.txt +++ b/testing/constraints-3.7.txt @@ -4,6 +4,6 @@ # Pin the version to the lower bound. # e.g., if setup.py has "google-cloud-foo >= 1.14.0, < 2.0.0dev", # Then this file should have google-cloud-foo==1.14.0 -google-api-core==1.33.2 +google-api-core==1.34.0 proto-plus==1.22.0 protobuf==3.19.5 diff --git a/testing/constraints-3.8.txt b/testing/constraints-3.8.txt index e69de29..ed7f9ae 100644 --- a/testing/constraints-3.8.txt +++ b/testing/constraints-3.8.txt @@ -0,0 +1,6 @@ +# -*- coding: utf-8 -*- +# This constraints file is required for unit tests. +# List all library dependencies and extras in this file. +google-api-core +proto-plus +protobuf diff --git a/testing/constraints-3.9.txt b/testing/constraints-3.9.txt index e69de29..ed7f9ae 100644 --- a/testing/constraints-3.9.txt +++ b/testing/constraints-3.9.txt @@ -0,0 +1,6 @@ +# -*- coding: utf-8 -*- +# This constraints file is required for unit tests. +# List all library dependencies and extras in this file. +google-api-core +proto-plus +protobuf