From d90af54ab58a1d0afd2a38373dbe17066cfa6ce2 Mon Sep 17 00:00:00 2001 From: WhiteSource Renovate Date: Mon, 10 Oct 2022 20:16:31 +0200 Subject: [PATCH 01/13] chore(deps): update all dependencies (#336) --- samples/snippets/requirements.txt | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/samples/snippets/requirements.txt b/samples/snippets/requirements.txt index 10b47fb..172a3b2 100644 --- a/samples/snippets/requirements.txt +++ b/samples/snippets/requirements.txt @@ -1,6 +1,6 @@ google-cloud-pubsub==2.13.7 -google-cloud-containeranalysis==2.9.2 -grafeas==1.6.0 +google-cloud-containeranalysis==2.9.3 +grafeas==1.6.1 pytest==7.1.3 flaky==3.7.0 mock==4.0.3 From a92655b25e1183c7f87b1f0d9d36d487a71bdd1d Mon Sep 17 00:00:00 2001 From: WhiteSource Renovate Date: Tue, 11 Oct 2022 18:26:13 +0200 Subject: [PATCH 02/13] chore(deps): update dependency google-cloud-pubsub to v2.13.9 (#337) MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit [![Mend Renovate](https://app.renovatebot.com/images/banner.svg)](https://renovatebot.com) This PR contains the following updates: | Package | Change | Age | Adoption | Passing | Confidence | |---|---|---|---|---|---| | [google-cloud-pubsub](https://togithub.com/googleapis/python-pubsub) | `==2.13.7` -> `==2.13.9` | [![age](https://badges.renovateapi.com/packages/pypi/google-cloud-pubsub/2.13.9/age-slim)](https://docs.renovatebot.com/merge-confidence/) | [![adoption](https://badges.renovateapi.com/packages/pypi/google-cloud-pubsub/2.13.9/adoption-slim)](https://docs.renovatebot.com/merge-confidence/) | [![passing](https://badges.renovateapi.com/packages/pypi/google-cloud-pubsub/2.13.9/compatibility-slim/2.13.7)](https://docs.renovatebot.com/merge-confidence/) | [![confidence](https://badges.renovateapi.com/packages/pypi/google-cloud-pubsub/2.13.9/confidence-slim/2.13.7)](https://docs.renovatebot.com/merge-confidence/) | --- ### Release Notes
googleapis/python-pubsub ### [`v2.13.9`](https://togithub.com/googleapis/python-pubsub/releases/tag/v2.13.9) [Compare Source](https://togithub.com/googleapis/python-pubsub/compare/v2.13.7...v2.13.9) ##### Bug Fixes - **deps:** Allow protobuf 3.19.5 ([#​801](https://togithub.com/googleapis/python-pubsub/issues/801)) ([fa23503](https://togithub.com/googleapis/python-pubsub/commit/fa235033481783c2ec378b2a26b223bdff206461))
--- ### Configuration πŸ“… **Schedule**: Branch creation - At any time (no schedule defined), Automerge - At any time (no schedule defined). 🚦 **Automerge**: Disabled by config. Please merge this manually once you are satisfied. β™» **Rebasing**: Whenever PR becomes conflicted, or you tick the rebase/retry checkbox. πŸ”• **Ignore**: Close this PR and you won't be reminded about this update again. --- - [ ] If you want to rebase/retry this PR, click this checkbox. --- This PR has been generated by [Mend Renovate](https://www.mend.io/free-developer-tools/renovate/). View repository job log [here](https://app.renovatebot.com/dashboard#github/googleapis/python-containeranalysis). --- samples/snippets/requirements.txt | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/samples/snippets/requirements.txt b/samples/snippets/requirements.txt index 172a3b2..a9f96ac 100644 --- a/samples/snippets/requirements.txt +++ b/samples/snippets/requirements.txt @@ -1,4 +1,4 @@ -google-cloud-pubsub==2.13.7 +google-cloud-pubsub==2.13.9 google-cloud-containeranalysis==2.9.3 grafeas==1.6.1 pytest==7.1.3 From 5439eaf61e8ccd8ca22ac094762e2a96432c6577 Mon Sep 17 00:00:00 2001 From: WhiteSource Renovate Date: Tue, 18 Oct 2022 15:12:03 +0200 Subject: [PATCH 03/13] chore(deps): update dependency google-cloud-pubsub to v2.13.10 (#338) --- samples/snippets/requirements.txt | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/samples/snippets/requirements.txt b/samples/snippets/requirements.txt index a9f96ac..a5f9a58 100644 --- a/samples/snippets/requirements.txt +++ b/samples/snippets/requirements.txt @@ -1,4 +1,4 @@ -google-cloud-pubsub==2.13.9 +google-cloud-pubsub==2.13.10 google-cloud-containeranalysis==2.9.3 grafeas==1.6.1 pytest==7.1.3 From 02678ea64ffcbd25493edacb919718bd5bfb24ae Mon Sep 17 00:00:00 2001 From: WhiteSource Renovate Date: Wed, 26 Oct 2022 12:48:18 +0200 Subject: [PATCH 04/13] chore(deps): update dependency pytest to v7.2.0 (#339) --- samples/snippets/requirements-test.txt | 2 +- samples/snippets/requirements.txt | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/samples/snippets/requirements-test.txt b/samples/snippets/requirements-test.txt index e071685..49780e0 100644 --- a/samples/snippets/requirements-test.txt +++ b/samples/snippets/requirements-test.txt @@ -1 +1 @@ -pytest==7.1.3 +pytest==7.2.0 diff --git a/samples/snippets/requirements.txt b/samples/snippets/requirements.txt index a5f9a58..ad6c968 100644 --- a/samples/snippets/requirements.txt +++ b/samples/snippets/requirements.txt @@ -1,6 +1,6 @@ google-cloud-pubsub==2.13.10 google-cloud-containeranalysis==2.9.3 grafeas==1.6.1 -pytest==7.1.3 +pytest==7.2.0 flaky==3.7.0 mock==4.0.3 From d23e838a5e75c7c16f483049ca77c5d743d9590d Mon Sep 17 00:00:00 2001 From: "gcf-owl-bot[bot]" <78513119+gcf-owl-bot[bot]@users.noreply.github.com> Date: Tue, 8 Nov 2022 02:00:31 +0000 Subject: [PATCH 05/13] chore(python): update dependencies in .kokoro/requirements.txt [autoapprove] (#341) Source-Link: https://togithub.com/googleapis/synthtool/commit/e3a1277ac35fc88c09db1930533e24292b132ced Post-Processor: gcr.io/cloud-devrel-public-resources/owlbot-python:latest@sha256:452901c74a22f9b9a3bd02bce780b8e8805c97270d424684bff809ce5be8c2a2 --- .github/.OwlBot.lock.yaml | 2 +- .kokoro/requirements.txt | 325 +++++++++++++++++++++----------------- noxfile.py | 11 +- 3 files changed, 187 insertions(+), 151 deletions(-) diff --git a/.github/.OwlBot.lock.yaml b/.github/.OwlBot.lock.yaml index 3815c98..12edee7 100644 --- a/.github/.OwlBot.lock.yaml +++ b/.github/.OwlBot.lock.yaml @@ -13,4 +13,4 @@ # limitations under the License. docker: image: gcr.io/cloud-devrel-public-resources/owlbot-python:latest - digest: sha256:7a40313731a7cb1454eef6b33d3446ebb121836738dc3ab3d2d3ded5268c35b6 + digest: sha256:452901c74a22f9b9a3bd02bce780b8e8805c97270d424684bff809ce5be8c2a2 diff --git a/.kokoro/requirements.txt b/.kokoro/requirements.txt index d15994b..31425f1 100644 --- a/.kokoro/requirements.txt +++ b/.kokoro/requirements.txt @@ -20,9 +20,9 @@ cachetools==5.2.0 \ --hash=sha256:6a94c6402995a99c3970cc7e4884bb60b4a8639938157eeed436098bf9831757 \ --hash=sha256:f9f17d2aec496a9aa6b76f53e3b614c965223c061982d434d160f930c698a9db # via google-auth -certifi==2022.6.15 \ - --hash=sha256:84c85a9078b11105f04f3036a9482ae10e4621616db313fe045dd24743a0820d \ - --hash=sha256:fe86415d55e84719d75f8b69414f6438ac3547d2078ab91b67e779ef69378412 +certifi==2022.9.24 \ + --hash=sha256:0d9c601124e5a6ba9712dbc60d9c53c21e34f5f641fe83002317394311bdce14 \ + --hash=sha256:90c1a32f1d68f940488354e36370f6cca89f0f106db09518524c88d6ed83f382 # via requests cffi==1.15.1 \ --hash=sha256:00a9ed42e88df81ffae7a8ab6d9356b371399b91dbdf0c3cb1e84c03a13aceb5 \ @@ -110,29 +110,33 @@ commonmark==0.9.1 \ --hash=sha256:452f9dc859be7f06631ddcb328b6919c67984aca654e5fefb3914d54691aed60 \ --hash=sha256:da2f38c92590f83de410ba1a3cbceafbc74fee9def35f9251ba9a971d6d66fd9 # via rich -cryptography==37.0.4 \ - --hash=sha256:190f82f3e87033821828f60787cfa42bff98404483577b591429ed99bed39d59 \ - --hash=sha256:2be53f9f5505673eeda5f2736bea736c40f051a739bfae2f92d18aed1eb54596 \ - --hash=sha256:30788e070800fec9bbcf9faa71ea6d8068f5136f60029759fd8c3efec3c9dcb3 \ - --hash=sha256:3d41b965b3380f10e4611dbae366f6dc3cefc7c9ac4e8842a806b9672ae9add5 \ - --hash=sha256:4c590ec31550a724ef893c50f9a97a0c14e9c851c85621c5650d699a7b88f7ab \ - --hash=sha256:549153378611c0cca1042f20fd9c5030d37a72f634c9326e225c9f666d472884 \ - --hash=sha256:63f9c17c0e2474ccbebc9302ce2f07b55b3b3fcb211ded18a42d5764f5c10a82 \ - --hash=sha256:6bc95ed67b6741b2607298f9ea4932ff157e570ef456ef7ff0ef4884a134cc4b \ - --hash=sha256:7099a8d55cd49b737ffc99c17de504f2257e3787e02abe6d1a6d136574873441 \ - --hash=sha256:75976c217f10d48a8b5a8de3d70c454c249e4b91851f6838a4e48b8f41eb71aa \ - --hash=sha256:7bc997818309f56c0038a33b8da5c0bfbb3f1f067f315f9abd6fc07ad359398d \ - --hash=sha256:80f49023dd13ba35f7c34072fa17f604d2f19bf0989f292cedf7ab5770b87a0b \ - --hash=sha256:91ce48d35f4e3d3f1d83e29ef4a9267246e6a3be51864a5b7d2247d5086fa99a \ - --hash=sha256:a958c52505c8adf0d3822703078580d2c0456dd1d27fabfb6f76fe63d2971cd6 \ - --hash=sha256:b62439d7cd1222f3da897e9a9fe53bbf5c104fff4d60893ad1355d4c14a24157 \ - --hash=sha256:b7f8dd0d4c1f21759695c05a5ec8536c12f31611541f8904083f3dc582604280 \ - --hash=sha256:d204833f3c8a33bbe11eda63a54b1aad7aa7456ed769a982f21ec599ba5fa282 \ - --hash=sha256:e007f052ed10cc316df59bc90fbb7ff7950d7e2919c9757fd42a2b8ecf8a5f67 \ - --hash=sha256:f2dcb0b3b63afb6df7fd94ec6fbddac81b5492513f7b0436210d390c14d46ee8 \ - --hash=sha256:f721d1885ecae9078c3f6bbe8a88bc0786b6e749bf32ccec1ef2b18929a05046 \ - --hash=sha256:f7a6de3e98771e183645181b3627e2563dcde3ce94a9e42a3f427d2255190327 \ - --hash=sha256:f8c0a6e9e1dd3eb0414ba320f85da6b0dcbd543126e30fcc546e7372a7fbf3b9 +cryptography==38.0.3 \ + --hash=sha256:068147f32fa662c81aebab95c74679b401b12b57494872886eb5c1139250ec5d \ + --hash=sha256:06fc3cc7b6f6cca87bd56ec80a580c88f1da5306f505876a71c8cfa7050257dd \ + --hash=sha256:25c1d1f19729fb09d42e06b4bf9895212292cb27bb50229f5aa64d039ab29146 \ + --hash=sha256:402852a0aea73833d982cabb6d0c3bb582c15483d29fb7085ef2c42bfa7e38d7 \ + --hash=sha256:4e269dcd9b102c5a3d72be3c45d8ce20377b8076a43cbed6f660a1afe365e436 \ + --hash=sha256:5419a127426084933076132d317911e3c6eb77568a1ce23c3ac1e12d111e61e0 \ + --hash=sha256:554bec92ee7d1e9d10ded2f7e92a5d70c1f74ba9524947c0ba0c850c7b011828 \ + --hash=sha256:5e89468fbd2fcd733b5899333bc54d0d06c80e04cd23d8c6f3e0542358c6060b \ + --hash=sha256:65535bc550b70bd6271984d9863a37741352b4aad6fb1b3344a54e6950249b55 \ + --hash=sha256:6ab9516b85bebe7aa83f309bacc5f44a61eeb90d0b4ec125d2d003ce41932d36 \ + --hash=sha256:6addc3b6d593cd980989261dc1cce38263c76954d758c3c94de51f1e010c9a50 \ + --hash=sha256:728f2694fa743a996d7784a6194da430f197d5c58e2f4e278612b359f455e4a2 \ + --hash=sha256:785e4056b5a8b28f05a533fab69febf5004458e20dad7e2e13a3120d8ecec75a \ + --hash=sha256:78cf5eefac2b52c10398a42765bfa981ce2372cbc0457e6bf9658f41ec3c41d8 \ + --hash=sha256:7f836217000342d448e1c9a342e9163149e45d5b5eca76a30e84503a5a96cab0 \ + --hash=sha256:8d41a46251bf0634e21fac50ffd643216ccecfaf3701a063257fe0b2be1b6548 \ + --hash=sha256:984fe150f350a3c91e84de405fe49e688aa6092b3525f407a18b9646f6612320 \ + --hash=sha256:9b24bcff7853ed18a63cfb0c2b008936a9554af24af2fb146e16d8e1aed75748 \ + --hash=sha256:b1b35d9d3a65542ed2e9d90115dfd16bbc027b3f07ee3304fc83580f26e43249 \ + --hash=sha256:b1b52c9e5f8aa2b802d48bd693190341fae201ea51c7a167d69fc48b60e8a959 \ + --hash=sha256:bbf203f1a814007ce24bd4d51362991d5cb90ba0c177a9c08825f2cc304d871f \ + --hash=sha256:be243c7e2bfcf6cc4cb350c0d5cdf15ca6383bbcb2a8ef51d3c9411a9d4386f0 \ + --hash=sha256:bfbe6ee19615b07a98b1d2287d6a6073f734735b49ee45b11324d85efc4d5cbd \ + --hash=sha256:c46837ea467ed1efea562bbeb543994c2d1f6e800785bd5a2c98bc096f5cb220 \ + --hash=sha256:dfb4f4dd568de1b6af9f4cda334adf7d72cf5bc052516e1b2608b683375dd95c \ + --hash=sha256:ed7b00096790213e09eb11c97cc6e2b757f15f3d2f85833cd2d3ec3fe37c1722 # via # gcp-releasetool # secretstorage @@ -148,23 +152,23 @@ filelock==3.8.0 \ --hash=sha256:55447caa666f2198c5b6b13a26d2084d26fa5b115c00d065664b2124680c4edc \ --hash=sha256:617eb4e5eedc82fc5f47b6d61e4d11cb837c56cb4544e39081099fa17ad109d4 # via virtualenv -gcp-docuploader==0.6.3 \ - --hash=sha256:ba8c9d76b3bbac54b0311c503a373b00edc2dc02d6d54ea9507045adb8e870f7 \ - --hash=sha256:c0f5aaa82ce1854a386197e4e359b120ad6d4e57ae2c812fce42219a3288026b +gcp-docuploader==0.6.4 \ + --hash=sha256:01486419e24633af78fd0167db74a2763974765ee8078ca6eb6964d0ebd388af \ + --hash=sha256:70861190c123d907b3b067da896265ead2eeb9263969d6955c9e0bb091b5ccbf # via -r requirements.in -gcp-releasetool==1.8.7 \ - --hash=sha256:3d2a67c9db39322194afb3b427e9cb0476ce8f2a04033695f0aeb63979fc2b37 \ - --hash=sha256:5e4d28f66e90780d77f3ecf1e9155852b0c3b13cbccb08ab07e66b2357c8da8d +gcp-releasetool==1.9.1 \ + --hash=sha256:952f4055d5d986b070ae2a71c4410b250000f9cc5a1e26398fcd55a5bbc5a15f \ + --hash=sha256:d0d3c814a97c1a237517e837d8cfa668ced8df4b882452578ecef4a4e79c583b # via -r requirements.in -google-api-core==2.8.2 \ - --hash=sha256:06f7244c640322b508b125903bb5701bebabce8832f85aba9335ec00b3d02edc \ - --hash=sha256:93c6a91ccac79079ac6bbf8b74ee75db970cc899278b97d53bc012f35908cf50 +google-api-core==2.10.2 \ + --hash=sha256:10c06f7739fe57781f87523375e8e1a3a4674bf6392cd6131a3222182b971320 \ + --hash=sha256:34f24bd1d5f72a8c4519773d99ca6bf080a6c4e041b4e9f024fe230191dda62e # via # google-cloud-core # google-cloud-storage -google-auth==2.11.0 \ - --hash=sha256:be62acaae38d0049c21ca90f27a23847245c9f161ff54ede13af2cb6afecbac9 \ - --hash=sha256:ed65ecf9f681832298e29328e1ef0a3676e3732b2e56f41532d45f70a22de0fb +google-auth==2.14.0 \ + --hash=sha256:1ad5b0e6eba5f69645971abb3d2c197537d5914070a8c6d30299dfdb07c5c700 \ + --hash=sha256:cf24817855d874ede2efd071aa22125445f555de1685b739a9782fcf408c2a3d # via # gcp-releasetool # google-api-core @@ -178,72 +182,97 @@ google-cloud-storage==2.5.0 \ --hash=sha256:19a26c66c317ce542cea0830b7e787e8dac2588b6bfa4d3fd3b871ba16305ab0 \ --hash=sha256:382f34b91de2212e3c2e7b40ec079d27ee2e3dbbae99b75b1bcd8c63063ce235 # via gcp-docuploader -google-crc32c==1.3.0 \ - --hash=sha256:04e7c220798a72fd0f08242bc8d7a05986b2a08a0573396187fd32c1dcdd58b3 \ - --hash=sha256:05340b60bf05b574159e9bd940152a47d38af3fb43803ffe71f11d704b7696a6 \ - --hash=sha256:12674a4c3b56b706153a358eaa1018c4137a5a04635b92b4652440d3d7386206 \ - --hash=sha256:127f9cc3ac41b6a859bd9dc4321097b1a4f6aa7fdf71b4f9227b9e3ebffb4422 \ - --hash=sha256:13af315c3a0eec8bb8b8d80b8b128cb3fcd17d7e4edafc39647846345a3f003a \ - --hash=sha256:1926fd8de0acb9d15ee757175ce7242e235482a783cd4ec711cc999fc103c24e \ - --hash=sha256:226f2f9b8e128a6ca6a9af9b9e8384f7b53a801907425c9a292553a3a7218ce0 \ - --hash=sha256:276de6273eb074a35bc598f8efbc00c7869c5cf2e29c90748fccc8c898c244df \ - --hash=sha256:318f73f5484b5671f0c7f5f63741ab020a599504ed81d209b5c7129ee4667407 \ - --hash=sha256:3bbce1be3687bbfebe29abdb7631b83e6b25da3f4e1856a1611eb21854b689ea \ - --hash=sha256:42ae4781333e331a1743445931b08ebdad73e188fd554259e772556fc4937c48 \ - --hash=sha256:58be56ae0529c664cc04a9c76e68bb92b091e0194d6e3c50bea7e0f266f73713 \ - --hash=sha256:5da2c81575cc3ccf05d9830f9e8d3c70954819ca9a63828210498c0774fda1a3 \ - --hash=sha256:6311853aa2bba4064d0c28ca54e7b50c4d48e3de04f6770f6c60ebda1e975267 \ - --hash=sha256:650e2917660e696041ab3dcd7abac160b4121cd9a484c08406f24c5964099829 \ - --hash=sha256:6a4db36f9721fdf391646685ecffa404eb986cbe007a3289499020daf72e88a2 \ - --hash=sha256:779cbf1ce375b96111db98fca913c1f5ec11b1d870e529b1dc7354b2681a8c3a \ - --hash=sha256:7f6fe42536d9dcd3e2ffb9d3053f5d05221ae3bbcefbe472bdf2c71c793e3183 \ - --hash=sha256:891f712ce54e0d631370e1f4997b3f182f3368179198efc30d477c75d1f44942 \ - --hash=sha256:95c68a4b9b7828ba0428f8f7e3109c5d476ca44996ed9a5f8aac6269296e2d59 \ - --hash=sha256:96a8918a78d5d64e07c8ea4ed2bc44354e3f93f46a4866a40e8db934e4c0d74b \ - --hash=sha256:9c3cf890c3c0ecfe1510a452a165431b5831e24160c5fcf2071f0f85ca5a47cd \ - --hash=sha256:9f58099ad7affc0754ae42e6d87443299f15d739b0ce03c76f515153a5cda06c \ - --hash=sha256:a0b9e622c3b2b8d0ce32f77eba617ab0d6768b82836391e4f8f9e2074582bf02 \ - --hash=sha256:a7f9cbea4245ee36190f85fe1814e2d7b1e5f2186381b082f5d59f99b7f11328 \ - --hash=sha256:bab4aebd525218bab4ee615786c4581952eadc16b1ff031813a2fd51f0cc7b08 \ - --hash=sha256:c124b8c8779bf2d35d9b721e52d4adb41c9bfbde45e6a3f25f0820caa9aba73f \ - --hash=sha256:c9da0a39b53d2fab3e5467329ed50e951eb91386e9d0d5b12daf593973c3b168 \ - --hash=sha256:ca60076c388728d3b6ac3846842474f4250c91efbfe5afa872d3ffd69dd4b318 \ - --hash=sha256:cb6994fff247987c66a8a4e550ef374671c2b82e3c0d2115e689d21e511a652d \ - --hash=sha256:d1c1d6236feab51200272d79b3d3e0f12cf2cbb12b208c835b175a21efdb0a73 \ - --hash=sha256:dd7760a88a8d3d705ff562aa93f8445ead54f58fd482e4f9e2bafb7e177375d4 \ - --hash=sha256:dda4d8a3bb0b50f540f6ff4b6033f3a74e8bf0bd5320b70fab2c03e512a62812 \ - --hash=sha256:e0f1ff55dde0ebcfbef027edc21f71c205845585fffe30d4ec4979416613e9b3 \ - --hash=sha256:e7a539b9be7b9c00f11ef16b55486141bc2cdb0c54762f84e3c6fc091917436d \ - --hash=sha256:eb0b14523758e37802f27b7f8cd973f5f3d33be7613952c0df904b68c4842f0e \ - --hash=sha256:ed447680ff21c14aaceb6a9f99a5f639f583ccfe4ce1a5e1d48eb41c3d6b3217 \ - --hash=sha256:f52a4ad2568314ee713715b1e2d79ab55fab11e8b304fd1462ff5cccf4264b3e \ - --hash=sha256:fbd60c6aaa07c31d7754edbc2334aef50601b7f1ada67a96eb1eb57c7c72378f \ - --hash=sha256:fc28e0db232c62ca0c3600884933178f0825c99be4474cdd645e378a10588125 \ - --hash=sha256:fe31de3002e7b08eb20823b3735b97c86c5926dd0581c7710a680b418a8709d4 \ - --hash=sha256:fec221a051150eeddfdfcff162e6db92c65ecf46cb0f7bb1bf812a1520ec026b \ - --hash=sha256:ff71073ebf0e42258a42a0b34f2c09ec384977e7f6808999102eedd5b49920e3 +google-crc32c==1.5.0 \ + --hash=sha256:024894d9d3cfbc5943f8f230e23950cd4906b2fe004c72e29b209420a1e6b05a \ + --hash=sha256:02c65b9817512edc6a4ae7c7e987fea799d2e0ee40c53ec573a692bee24de876 \ + --hash=sha256:02ebb8bf46c13e36998aeaad1de9b48f4caf545e91d14041270d9dca767b780c \ + --hash=sha256:07eb3c611ce363c51a933bf6bd7f8e3878a51d124acfc89452a75120bc436289 \ + --hash=sha256:1034d91442ead5a95b5aaef90dbfaca8633b0247d1e41621d1e9f9db88c36298 \ + --hash=sha256:116a7c3c616dd14a3de8c64a965828b197e5f2d121fedd2f8c5585c547e87b02 \ + --hash=sha256:19e0a019d2c4dcc5e598cd4a4bc7b008546b0358bd322537c74ad47a5386884f \ + --hash=sha256:1c7abdac90433b09bad6c43a43af253e688c9cfc1c86d332aed13f9a7c7f65e2 \ + --hash=sha256:1e986b206dae4476f41bcec1faa057851f3889503a70e1bdb2378d406223994a \ + --hash=sha256:272d3892a1e1a2dbc39cc5cde96834c236d5327e2122d3aaa19f6614531bb6eb \ + --hash=sha256:278d2ed7c16cfc075c91378c4f47924c0625f5fc84b2d50d921b18b7975bd210 \ + --hash=sha256:2ad40e31093a4af319dadf503b2467ccdc8f67c72e4bcba97f8c10cb078207b5 \ + --hash=sha256:2e920d506ec85eb4ba50cd4228c2bec05642894d4c73c59b3a2fe20346bd00ee \ + --hash=sha256:3359fc442a743e870f4588fcf5dcbc1bf929df1fad8fb9905cd94e5edb02e84c \ + --hash=sha256:37933ec6e693e51a5b07505bd05de57eee12f3e8c32b07da7e73669398e6630a \ + --hash=sha256:398af5e3ba9cf768787eef45c803ff9614cc3e22a5b2f7d7ae116df8b11e3314 \ + --hash=sha256:3b747a674c20a67343cb61d43fdd9207ce5da6a99f629c6e2541aa0e89215bcd \ + --hash=sha256:461665ff58895f508e2866824a47bdee72497b091c730071f2b7575d5762ab65 \ + --hash=sha256:4c6fdd4fccbec90cc8a01fc00773fcd5fa28db683c116ee3cb35cd5da9ef6c37 \ + --hash=sha256:5829b792bf5822fd0a6f6eb34c5f81dd074f01d570ed7f36aa101d6fc7a0a6e4 \ + --hash=sha256:596d1f98fc70232fcb6590c439f43b350cb762fb5d61ce7b0e9db4539654cc13 \ + --hash=sha256:5ae44e10a8e3407dbe138984f21e536583f2bba1be9491239f942c2464ac0894 \ + --hash=sha256:635f5d4dd18758a1fbd1049a8e8d2fee4ffed124462d837d1a02a0e009c3ab31 \ + --hash=sha256:64e52e2b3970bd891309c113b54cf0e4384762c934d5ae56e283f9a0afcd953e \ + --hash=sha256:66741ef4ee08ea0b2cc3c86916ab66b6aef03768525627fd6a1b34968b4e3709 \ + --hash=sha256:67b741654b851abafb7bc625b6d1cdd520a379074e64b6a128e3b688c3c04740 \ + --hash=sha256:6ac08d24c1f16bd2bf5eca8eaf8304812f44af5cfe5062006ec676e7e1d50afc \ + --hash=sha256:6f998db4e71b645350b9ac28a2167e6632c239963ca9da411523bb439c5c514d \ + --hash=sha256:72218785ce41b9cfd2fc1d6a017dc1ff7acfc4c17d01053265c41a2c0cc39b8c \ + --hash=sha256:74dea7751d98034887dbd821b7aae3e1d36eda111d6ca36c206c44478035709c \ + --hash=sha256:759ce4851a4bb15ecabae28f4d2e18983c244eddd767f560165563bf9aefbc8d \ + --hash=sha256:77e2fd3057c9d78e225fa0a2160f96b64a824de17840351b26825b0848022906 \ + --hash=sha256:7c074fece789b5034b9b1404a1f8208fc2d4c6ce9decdd16e8220c5a793e6f61 \ + --hash=sha256:7c42c70cd1d362284289c6273adda4c6af8039a8ae12dc451dcd61cdabb8ab57 \ + --hash=sha256:7f57f14606cd1dd0f0de396e1e53824c371e9544a822648cd76c034d209b559c \ + --hash=sha256:83c681c526a3439b5cf94f7420471705bbf96262f49a6fe546a6db5f687a3d4a \ + --hash=sha256:8485b340a6a9e76c62a7dce3c98e5f102c9219f4cfbf896a00cf48caf078d438 \ + --hash=sha256:84e6e8cd997930fc66d5bb4fde61e2b62ba19d62b7abd7a69920406f9ecca946 \ + --hash=sha256:89284716bc6a5a415d4eaa11b1726d2d60a0cd12aadf5439828353662ede9dd7 \ + --hash=sha256:8b87e1a59c38f275c0e3676fc2ab6d59eccecfd460be267ac360cc31f7bcde96 \ + --hash=sha256:8f24ed114432de109aa9fd317278518a5af2d31ac2ea6b952b2f7782b43da091 \ + --hash=sha256:98cb4d057f285bd80d8778ebc4fde6b4d509ac3f331758fb1528b733215443ae \ + --hash=sha256:998679bf62b7fb599d2878aa3ed06b9ce688b8974893e7223c60db155f26bd8d \ + --hash=sha256:9ba053c5f50430a3fcfd36f75aff9caeba0440b2d076afdb79a318d6ca245f88 \ + --hash=sha256:9c99616c853bb585301df6de07ca2cadad344fd1ada6d62bb30aec05219c45d2 \ + --hash=sha256:a1fd716e7a01f8e717490fbe2e431d2905ab8aa598b9b12f8d10abebb36b04dd \ + --hash=sha256:a2355cba1f4ad8b6988a4ca3feed5bff33f6af2d7f134852cf279c2aebfde541 \ + --hash=sha256:b1f8133c9a275df5613a451e73f36c2aea4fe13c5c8997e22cf355ebd7bd0728 \ + --hash=sha256:b8667b48e7a7ef66afba2c81e1094ef526388d35b873966d8a9a447974ed9178 \ + --hash=sha256:ba1eb1843304b1e5537e1fca632fa894d6f6deca8d6389636ee5b4797affb968 \ + --hash=sha256:be82c3c8cfb15b30f36768797a640e800513793d6ae1724aaaafe5bf86f8f346 \ + --hash=sha256:c02ec1c5856179f171e032a31d6f8bf84e5a75c45c33b2e20a3de353b266ebd8 \ + --hash=sha256:c672d99a345849301784604bfeaeba4db0c7aae50b95be04dd651fd2a7310b93 \ + --hash=sha256:c6c777a480337ac14f38564ac88ae82d4cd238bf293f0a22295b66eb89ffced7 \ + --hash=sha256:cae0274952c079886567f3f4f685bcaf5708f0a23a5f5216fdab71f81a6c0273 \ + --hash=sha256:cd67cf24a553339d5062eff51013780a00d6f97a39ca062781d06b3a73b15462 \ + --hash=sha256:d3515f198eaa2f0ed49f8819d5732d70698c3fa37384146079b3799b97667a94 \ + --hash=sha256:d5280312b9af0976231f9e317c20e4a61cd2f9629b7bfea6a693d1878a264ebd \ + --hash=sha256:de06adc872bcd8c2a4e0dc51250e9e65ef2ca91be023b9d13ebd67c2ba552e1e \ + --hash=sha256:e1674e4307fa3024fc897ca774e9c7562c957af85df55efe2988ed9056dc4e57 \ + --hash=sha256:e2096eddb4e7c7bdae4bd69ad364e55e07b8316653234a56552d9c988bd2d61b \ + --hash=sha256:e560628513ed34759456a416bf86b54b2476c59144a9138165c9a1575801d0d9 \ + --hash=sha256:edfedb64740750e1a3b16152620220f51d58ff1b4abceb339ca92e934775c27a \ + --hash=sha256:f13cae8cc389a440def0c8c52057f37359014ccbc9dc1f0827936bcd367c6100 \ + --hash=sha256:f314013e7dcd5cf45ab1945d92e713eec788166262ae8deb2cfacd53def27325 \ + --hash=sha256:f583edb943cf2e09c60441b910d6a20b4d9d626c75a36c8fcac01a6c96c01183 \ + --hash=sha256:fd8536e902db7e365f49e7d9029283403974ccf29b13fc7028b97e2295b33556 \ + --hash=sha256:fe70e325aa68fa4b5edf7d1a4b6f691eb04bbccac0ace68e34820d283b5f80d4 # via google-resumable-media -google-resumable-media==2.3.3 \ - --hash=sha256:27c52620bd364d1c8116eaac4ea2afcbfb81ae9139fb3199652fcac1724bfb6c \ - --hash=sha256:5b52774ea7a829a8cdaa8bd2d4c3d4bc660c91b30857ab2668d0eb830f4ea8c5 +google-resumable-media==2.4.0 \ + --hash=sha256:2aa004c16d295c8f6c33b2b4788ba59d366677c0a25ae7382436cb30f776deaa \ + --hash=sha256:8d5518502f92b9ecc84ac46779bd4f09694ecb3ba38a3e7ca737a86d15cbca1f # via google-cloud-storage googleapis-common-protos==1.56.4 \ --hash=sha256:8eb2cbc91b69feaf23e32452a7ae60e791e09967d81d4fcc7fc388182d1bd394 \ --hash=sha256:c25873c47279387cfdcbdafa36149887901d36202cb645a0e4f29686bf6e4417 # via google-api-core -idna==3.3 \ - --hash=sha256:84d9dd047ffa80596e0f246e2eab0b391788b0503584e8945f2368256d2735ff \ - --hash=sha256:9d643ff0a55b762d5cdb124b8eaa99c66322e2157b69160bc32796e824360e6d +idna==3.4 \ + --hash=sha256:814f528e8dead7d329833b91c5faa87d60bf71824cd12a7530b5526063d02cb4 \ + --hash=sha256:90b77e79eaa3eba6de819a0c442c0b4ceefc341a7a2ab77d7562bf49f425c5c2 # via requests -importlib-metadata==4.12.0 \ - --hash=sha256:637245b8bab2b6502fcbc752cc4b7a6f6243bb02b31c5c26156ad103d3d45670 \ - --hash=sha256:7401a975809ea1fdc658c3aa4f78cc2195a0e019c5cbc4c06122884e9ae80c23 +importlib-metadata==5.0.0 \ + --hash=sha256:da31db32b304314d044d3c12c79bd59e307889b287ad12ff387b3500835fc2ab \ + --hash=sha256:ddb0e35065e8938f867ed4928d0ae5bf2a53b7773871bfe6bcc7e4fcdc7dea43 # via # -r requirements.in # twine -jaraco-classes==3.2.2 \ - --hash=sha256:6745f113b0b588239ceb49532aa09c3ebb947433ce311ef2f8e3ad64ebb74594 \ - --hash=sha256:e6ef6fd3fcf4579a7a019d87d1e56a883f4e4c35cfe925f86731abc58804e647 +jaraco-classes==3.2.3 \ + --hash=sha256:2353de3288bc6b82120752201c6b1c1a14b058267fa424ed5ce5984e3b922158 \ + --hash=sha256:89559fa5c1d3c34eff6f631ad80bb21f378dbcbb35dd161fd2c6b93f5be2f98a # via keyring jeepney==0.8.0 \ --hash=sha256:5efe48d255973902f6badc3ce55e2aa6c5c3b3bc642059ef3a91247bcfcc5806 \ @@ -255,9 +284,9 @@ jinja2==3.1.2 \ --hash=sha256:31351a702a408a9e7595a8fc6150fc3f43bb6bf7e319770cbc0db9df9437e852 \ --hash=sha256:6088930bfe239f0e6710546ab9c19c9ef35e29792895fed6e6e31a023a182a61 # via gcp-releasetool -keyring==23.9.0 \ - --hash=sha256:4c32a31174faaee48f43a7e2c7e9c3216ec5e95acf22a2bebfb4a1d05056ee44 \ - --hash=sha256:98f060ec95ada2ab910c195a2d4317be6ef87936a766b239c46aa3c7aac4f0db +keyring==23.9.3 \ + --hash=sha256:69732a15cb1433bdfbc3b980a8a36a04878a6cfd7cb99f497b573f31618001c0 \ + --hash=sha256:69b01dd83c42f590250fe7a1f503fc229b14de83857314b1933a3ddbf595c4a5 # via # gcp-releasetool # twine @@ -303,9 +332,9 @@ markupsafe==2.1.1 \ --hash=sha256:f121a1420d4e173a5d96e47e9a0c0dcff965afdf1626d28de1460815f7c4ee7a \ --hash=sha256:fc7b548b17d238737688817ab67deebb30e8073c95749d55538ed473130ec0c7 # via jinja2 -more-itertools==8.14.0 \ - --hash=sha256:1bc4f91ee5b1b31ac7ceacc17c09befe6a40a503907baf9c839c229b5095cfd2 \ - --hash=sha256:c09443cd3d5438b8dafccd867a6bc1cb0894389e90cb53d227456b0b0bccb750 +more-itertools==9.0.0 \ + --hash=sha256:250e83d7e81d0c87ca6bd942e6aeab8cc9daa6096d12c5308f3f92fa5e5c1f41 \ + --hash=sha256:5a6257e40878ef0520b1803990e3e22303a41b5714006c32a3fd8304b26ea1ab # via jaraco-classes nox==2022.8.7 \ --hash=sha256:1b894940551dc5c389f9271d197ca5d655d40bdc6ccf93ed6880e4042760a34b \ @@ -325,34 +354,34 @@ platformdirs==2.5.2 \ --hash=sha256:027d8e83a2d7de06bbac4e5ef7e023c02b863d7ea5d079477e722bb41ab25788 \ --hash=sha256:58c8abb07dcb441e6ee4b11d8df0ac856038f944ab98b7be6b27b2a3c7feef19 # via virtualenv -protobuf==3.20.2 \ - --hash=sha256:03d76b7bd42ac4a6e109742a4edf81ffe26ffd87c5993126d894fe48a120396a \ - --hash=sha256:09e25909c4297d71d97612f04f41cea8fa8510096864f2835ad2f3b3df5a5559 \ - --hash=sha256:18e34a10ae10d458b027d7638a599c964b030c1739ebd035a1dfc0e22baa3bfe \ - --hash=sha256:291fb4307094bf5ccc29f424b42268640e00d5240bf0d9b86bf3079f7576474d \ - --hash=sha256:2c0b040d0b5d5d207936ca2d02f00f765906622c07d3fa19c23a16a8ca71873f \ - --hash=sha256:384164994727f274cc34b8abd41a9e7e0562801361ee77437099ff6dfedd024b \ - --hash=sha256:3cb608e5a0eb61b8e00fe641d9f0282cd0eedb603be372f91f163cbfbca0ded0 \ - --hash=sha256:5d9402bf27d11e37801d1743eada54372f986a372ec9679673bfcc5c60441151 \ - --hash=sha256:712dca319eee507a1e7df3591e639a2b112a2f4a62d40fe7832a16fd19151750 \ - --hash=sha256:7a5037af4e76c975b88c3becdf53922b5ffa3f2cddf657574a4920a3b33b80f3 \ - --hash=sha256:8228e56a865c27163d5d1d1771d94b98194aa6917bcfb6ce139cbfa8e3c27334 \ - --hash=sha256:84a1544252a933ef07bb0b5ef13afe7c36232a774affa673fc3636f7cee1db6c \ - --hash=sha256:84fe5953b18a383fd4495d375fe16e1e55e0a3afe7b4f7b4d01a3a0649fcda9d \ - --hash=sha256:9c673c8bfdf52f903081816b9e0e612186684f4eb4c17eeb729133022d6032e3 \ - --hash=sha256:9f876a69ca55aed879b43c295a328970306e8e80a263ec91cf6e9189243c613b \ - --hash=sha256:a9e5ae5a8e8985c67e8944c23035a0dff2c26b0f5070b2f55b217a1c33bbe8b1 \ - --hash=sha256:b4fdb29c5a7406e3f7ef176b2a7079baa68b5b854f364c21abe327bbeec01cdb \ - --hash=sha256:c184485e0dfba4dfd451c3bd348c2e685d6523543a0f91b9fd4ae90eb09e8422 \ - --hash=sha256:c9cdf251c582c16fd6a9f5e95836c90828d51b0069ad22f463761d27c6c19019 \ - --hash=sha256:e39cf61bb8582bda88cdfebc0db163b774e7e03364bbf9ce1ead13863e81e359 \ - --hash=sha256:e8fbc522303e09036c752a0afcc5c0603e917222d8bedc02813fd73b4b4ed804 \ - --hash=sha256:f34464ab1207114e73bba0794d1257c150a2b89b7a9faf504e00af7c9fd58978 \ - --hash=sha256:f52dabc96ca99ebd2169dadbe018824ebda08a795c7684a0b7d203a290f3adb0 +protobuf==3.20.3 \ + --hash=sha256:03038ac1cfbc41aa21f6afcbcd357281d7521b4157926f30ebecc8d4ea59dcb7 \ + --hash=sha256:28545383d61f55b57cf4df63eebd9827754fd2dc25f80c5253f9184235db242c \ + --hash=sha256:2e3427429c9cffebf259491be0af70189607f365c2f41c7c3764af6f337105f2 \ + --hash=sha256:398a9e0c3eaceb34ec1aee71894ca3299605fa8e761544934378bbc6c97de23b \ + --hash=sha256:44246bab5dd4b7fbd3c0c80b6f16686808fab0e4aca819ade6e8d294a29c7050 \ + --hash=sha256:447d43819997825d4e71bf5769d869b968ce96848b6479397e29fc24c4a5dfe9 \ + --hash=sha256:67a3598f0a2dcbc58d02dd1928544e7d88f764b47d4a286202913f0b2801c2e7 \ + --hash=sha256:74480f79a023f90dc6e18febbf7b8bac7508420f2006fabd512013c0c238f454 \ + --hash=sha256:819559cafa1a373b7096a482b504ae8a857c89593cf3a25af743ac9ecbd23480 \ + --hash=sha256:899dc660cd599d7352d6f10d83c95df430a38b410c1b66b407a6b29265d66469 \ + --hash=sha256:8c0c984a1b8fef4086329ff8dd19ac77576b384079247c770f29cc8ce3afa06c \ + --hash=sha256:9aae4406ea63d825636cc11ffb34ad3379335803216ee3a856787bcf5ccc751e \ + --hash=sha256:a7ca6d488aa8ff7f329d4c545b2dbad8ac31464f1d8b1c87ad1346717731e4db \ + --hash=sha256:b6cc7ba72a8850621bfec987cb72623e703b7fe2b9127a161ce61e61558ad905 \ + --hash=sha256:bf01b5720be110540be4286e791db73f84a2b721072a3711efff6c324cdf074b \ + --hash=sha256:c02ce36ec760252242a33967d51c289fd0e1c0e6e5cc9397e2279177716add86 \ + --hash=sha256:d9e4432ff660d67d775c66ac42a67cf2453c27cb4d738fc22cb53b5d84c135d4 \ + --hash=sha256:daa564862dd0d39c00f8086f88700fdbe8bc717e993a21e90711acfed02f2402 \ + --hash=sha256:de78575669dddf6099a8a0f46a27e82a1783c557ccc38ee620ed8cc96d3be7d7 \ + --hash=sha256:e64857f395505ebf3d2569935506ae0dfc4a15cb80dc25261176c784662cdcc4 \ + --hash=sha256:f4bd856d702e5b0d96a00ec6b307b0f51c1982c2bf9c0052cf9019e9a544ba99 \ + --hash=sha256:f4c42102bc82a51108e449cbb32b19b180022941c727bac0cfd50170341f16ee # via # gcp-docuploader # gcp-releasetool # google-api-core + # googleapis-common-protos py==1.11.0 \ --hash=sha256:51c75c4126074b472f746a24399ad32f6053d1b34b68d2fa41e558e6f4a98719 \ --hash=sha256:607c53218732647dff4acdfcd50cb62615cedf612e72d1724fb1a0cc6405b378 @@ -377,9 +406,9 @@ pygments==2.13.0 \ # via # readme-renderer # rich -pyjwt==2.4.0 \ - --hash=sha256:72d1d253f32dbd4f5c88eaf1fdc62f3a19f676ccbadb9dbc5d07e951b2b26daf \ - --hash=sha256:d42908208c699b3b973cbeb01a969ba6a96c821eefb1c5bfe4c390c01d67abba +pyjwt==2.6.0 \ + --hash=sha256:69285c7e31fc44f68a1feb309e948e0df53259d579295e6cfe2b1792329f05fd \ + --hash=sha256:d83c3d892a77bbb74d3e1a2cfa90afaadb60945205d1095d9221f04466f64c14 # via gcp-releasetool pyparsing==3.0.9 \ --hash=sha256:2b020ecf7d21b687f219b71ecad3631f644a47f01403fa1d1036b0c6416d70fb \ @@ -392,9 +421,9 @@ python-dateutil==2.8.2 \ --hash=sha256:0123cacc1627ae19ddf3c27a5de5bd67ee4586fbdd6440d9748f8abb483d3e86 \ --hash=sha256:961d03dc3453ebbc59dbdea9e4e11c5651520a876d0f4db161e8674aae935da9 # via gcp-releasetool -readme-renderer==37.0 \ - --hash=sha256:07b7ea234e03e58f77cc222e206e6abb8f4c0435becce5104794ee591f9301c5 \ - --hash=sha256:9fa416704703e509eeb900696751c908ddeb2011319d93700d8f18baff887a69 +readme-renderer==37.3 \ + --hash=sha256:cd653186dfc73055656f090f227f5cb22a046d7f71a841dfa305f55c9a513273 \ + --hash=sha256:f67a16caedfa71eef48a31b39708637a6f4664c4394801a7b0d6432d13907343 # via twine requests==2.28.1 \ --hash=sha256:7c5599b102feddaa661c826c56ab4fee28bfd17f5abca1ebbe3e7f19d7c97983 \ @@ -405,17 +434,17 @@ requests==2.28.1 \ # google-cloud-storage # requests-toolbelt # twine -requests-toolbelt==0.9.1 \ - --hash=sha256:380606e1d10dc85c3bd47bf5a6095f815ec007be7a8b69c878507068df059e6f \ - --hash=sha256:968089d4584ad4ad7c171454f0a5c6dac23971e9472521ea3b6d49d610aa6fc0 +requests-toolbelt==0.10.1 \ + --hash=sha256:18565aa58116d9951ac39baa288d3adb5b3ff975c4f25eee78555d89e8f247f7 \ + --hash=sha256:62e09f7ff5ccbda92772a29f394a49c3ad6cb181d568b1337626b2abb628a63d # via twine rfc3986==2.0.0 \ --hash=sha256:50b1502b60e289cb37883f3dfd34532b8873c7de9f49bb546641ce9cbd256ebd \ --hash=sha256:97aacf9dbd4bfd829baad6e6309fa6573aaf1be3f6fa735c8ab05e46cecb261c # via twine -rich==12.5.1 \ - --hash=sha256:2eb4e6894cde1e017976d2975ac210ef515d7548bc595ba20e195fb9628acdeb \ - --hash=sha256:63a5c5ce3673d3d5fbbf23cd87e11ab84b6b451436f1b7f19ec54b6bc36ed7ca +rich==12.6.0 \ + --hash=sha256:a4eb26484f2c82589bd9a17c73d32a010b1e29d89f1604cd9bf3a2097b81bb5e \ + --hash=sha256:ba3a3775974105c221d31141f2c116f4fd65c5ceb0698657a11e9f295ec93fd0 # via twine rsa==4.9 \ --hash=sha256:90260d9058e514786967344d0ef75fa8727eed8a7d2e43ce9f4bcf1b536174f7 \ @@ -437,9 +466,9 @@ twine==4.0.1 \ --hash=sha256:42026c18e394eac3e06693ee52010baa5313e4811d5a11050e7d48436cf41b9e \ --hash=sha256:96b1cf12f7ae611a4a40b6ae8e9570215daff0611828f5fe1f37a16255ab24a0 # via -r requirements.in -typing-extensions==4.3.0 \ - --hash=sha256:25642c956049920a5aa49edcdd6ab1e06d7e5d467fc00e0506c44ac86fbfca02 \ - --hash=sha256:e6d2677a32f47fc7eb2795db1dd15c1f34eff616bcaf2cfb5e997f854fa1c4a6 +typing-extensions==4.4.0 \ + --hash=sha256:1511434bb92bf8dd198c12b1cc812e800d4181cfcb867674e0f8279cc93087aa \ + --hash=sha256:16fa4864408f655d35ec496218b85f79b3437c829e93320c7c9215ccfd92489e # via -r requirements.in urllib3==1.26.12 \ --hash=sha256:3fa96cf423e6987997fc326ae8df396db2a8b7c667747d47ddd8ecba91f4a74e \ @@ -447,9 +476,9 @@ urllib3==1.26.12 \ # via # requests # twine -virtualenv==20.16.4 \ - --hash=sha256:014f766e4134d0008dcaa1f95bafa0fb0f575795d07cae50b1bee514185d6782 \ - --hash=sha256:035ed57acce4ac35c82c9d8802202b0e71adac011a511ff650cbcf9635006a22 +virtualenv==20.16.6 \ + --hash=sha256:186ca84254abcbde98180fd17092f9628c5fe742273c02724972a1d8a2035108 \ + --hash=sha256:530b850b523c6449406dfba859d6345e48ef19b8439606c5d74d7d3c9e14d76e # via nox webencodings==0.5.1 \ --hash=sha256:a0af1213f3c2226497a97e2b3aa01a7e4bee4f403f95be16fc9acd2947514a78 \ @@ -459,13 +488,13 @@ wheel==0.37.1 \ --hash=sha256:4bdcd7d840138086126cd09254dc6195fb4fc6f01c050a1d7236f2630db1d22a \ --hash=sha256:e9a504e793efbca1b8e0e9cb979a249cf4a0a7b5b8c9e8b65a5e39d49529c1c4 # via -r requirements.in -zipp==3.8.1 \ - --hash=sha256:05b45f1ee8f807d0cc928485ca40a07cb491cf092ff587c0df9cb1fd154848d2 \ - --hash=sha256:47c40d7fe183a6f21403a199b3e4192cca5774656965b0a4988ad2f8feb5f009 +zipp==3.10.0 \ + --hash=sha256:4fcb6f278987a6605757302a6e40e896257570d11c51628968ccb2a47e80c6c1 \ + --hash=sha256:7a7262fd930bd3e36c50b9a64897aec3fafff3dfdeec9623ae22b40e93f99bb8 # via importlib-metadata # The following packages are considered to be unsafe in a requirements file: -setuptools==65.2.0 \ - --hash=sha256:7f4bc85450898a09f76ebf28b72fa25bc7111f6c7d665d514a60bba9c75ef2a9 \ - --hash=sha256:a3ca5857c89f82f5c9410e8508cb32f4872a3bafd4aa7ae122a24ca33bccc750 +setuptools==65.5.0 \ + --hash=sha256:512e5536220e38146176efb833d4a62aa726b7bbff82cfbc8ba9eaa3996e0b17 \ + --hash=sha256:f62ea9da9ed6289bfe868cd6845968a2c854d1427f8548d52cae02a42b4f0356 # via -r requirements.in diff --git a/noxfile.py b/noxfile.py index f764a9c..54c2285 100644 --- a/noxfile.py +++ b/noxfile.py @@ -277,7 +277,11 @@ def docs(session): """Build the docs for this library.""" session.install("-e", ".") - session.install("sphinx==4.0.1", "alabaster", "recommonmark") + session.install( + "sphinx==4.0.1", + "alabaster", + "recommonmark", + ) shutil.rmtree(os.path.join("docs", "_build"), ignore_errors=True) session.run( @@ -300,7 +304,10 @@ def docfx(session): session.install("-e", ".") session.install( - "sphinx==4.0.1", "alabaster", "recommonmark", "gcp-sphinx-docfx-yaml" + "sphinx==4.0.1", + "alabaster", + "recommonmark", + "gcp-sphinx-docfx-yaml", ) shutil.rmtree(os.path.join("docs", "_build"), ignore_errors=True) From 75fa28921beddfd3b25c4b0cce642d6106ad01af Mon Sep 17 00:00:00 2001 From: "gcf-owl-bot[bot]" <78513119+gcf-owl-bot[bot]@users.noreply.github.com> Date: Wed, 16 Nov 2022 17:14:45 +0000 Subject: [PATCH 06/13] chore(python): update release script dependencies [autoapprove] (#344) Source-Link: https://togithub.com/googleapis/synthtool/commit/25083af347468dd5f90f69627420f7d452b6c50e Post-Processor: gcr.io/cloud-devrel-public-resources/owlbot-python:latest@sha256:e6cbd61f1838d9ff6a31436dfc13717f372a7482a82fc1863ca954ec47bff8c8 --- .github/.OwlBot.lock.yaml | 2 +- .kokoro/docker/docs/Dockerfile | 12 +++---- .kokoro/requirements.in | 4 ++- .kokoro/requirements.txt | 61 ++++++++++++++++++---------------- noxfile.py | 4 +-- 5 files changed, 44 insertions(+), 39 deletions(-) diff --git a/.github/.OwlBot.lock.yaml b/.github/.OwlBot.lock.yaml index 12edee7..3f1ccc0 100644 --- a/.github/.OwlBot.lock.yaml +++ b/.github/.OwlBot.lock.yaml @@ -13,4 +13,4 @@ # limitations under the License. docker: image: gcr.io/cloud-devrel-public-resources/owlbot-python:latest - digest: sha256:452901c74a22f9b9a3bd02bce780b8e8805c97270d424684bff809ce5be8c2a2 + digest: sha256:e6cbd61f1838d9ff6a31436dfc13717f372a7482a82fc1863ca954ec47bff8c8 diff --git a/.kokoro/docker/docs/Dockerfile b/.kokoro/docker/docs/Dockerfile index 238b87b..f8137d0 100644 --- a/.kokoro/docker/docs/Dockerfile +++ b/.kokoro/docker/docs/Dockerfile @@ -60,16 +60,16 @@ RUN apt-get update \ && rm -rf /var/lib/apt/lists/* \ && rm -f /var/cache/apt/archives/*.deb -###################### Install python 3.8.11 +###################### Install python 3.9.13 -# Download python 3.8.11 -RUN wget https://www.python.org/ftp/python/3.8.11/Python-3.8.11.tgz +# Download python 3.9.13 +RUN wget https://www.python.org/ftp/python/3.9.13/Python-3.9.13.tgz # Extract files -RUN tar -xvf Python-3.8.11.tgz +RUN tar -xvf Python-3.9.13.tgz -# Install python 3.8.11 -RUN ./Python-3.8.11/configure --enable-optimizations +# Install python 3.9.13 +RUN ./Python-3.9.13/configure --enable-optimizations RUN make altinstall ###################### Install pip diff --git a/.kokoro/requirements.in b/.kokoro/requirements.in index 7718391..cbd7e77 100644 --- a/.kokoro/requirements.in +++ b/.kokoro/requirements.in @@ -5,4 +5,6 @@ typing-extensions twine wheel setuptools -nox \ No newline at end of file +nox +charset-normalizer<3 +click<8.1.0 diff --git a/.kokoro/requirements.txt b/.kokoro/requirements.txt index 31425f1..9c1b9be 100644 --- a/.kokoro/requirements.txt +++ b/.kokoro/requirements.txt @@ -93,11 +93,14 @@ cffi==1.15.1 \ charset-normalizer==2.1.1 \ --hash=sha256:5a3d016c7c547f69d6f81fb0db9449ce888b418b5b9952cc5e6e66843e9dd845 \ --hash=sha256:83e9a75d1911279afd89352c68b45348559d1fc0506b054b346651b5e7fee29f - # via requests + # via + # -r requirements.in + # requests click==8.0.4 \ --hash=sha256:6a7a62563bbfabfda3a38f3023a1db4a35978c0abd76f6c9605ecd6554d6d9b1 \ --hash=sha256:8458d7b1287c5fb128c90e23381cf99dcde74beaf6c7ff6384ce84d6fe090adb # via + # -r requirements.in # gcp-docuploader # gcp-releasetool colorlog==6.7.0 \ @@ -156,9 +159,9 @@ gcp-docuploader==0.6.4 \ --hash=sha256:01486419e24633af78fd0167db74a2763974765ee8078ca6eb6964d0ebd388af \ --hash=sha256:70861190c123d907b3b067da896265ead2eeb9263969d6955c9e0bb091b5ccbf # via -r requirements.in -gcp-releasetool==1.9.1 \ - --hash=sha256:952f4055d5d986b070ae2a71c4410b250000f9cc5a1e26398fcd55a5bbc5a15f \ - --hash=sha256:d0d3c814a97c1a237517e837d8cfa668ced8df4b882452578ecef4a4e79c583b +gcp-releasetool==1.10.0 \ + --hash=sha256:72a38ca91b59c24f7e699e9227c90cbe4dd71b789383cb0164b088abae294c83 \ + --hash=sha256:8c7c99320208383d4bb2b808c6880eb7a81424afe7cdba3c8d84b25f4f0e097d # via -r requirements.in google-api-core==2.10.2 \ --hash=sha256:10c06f7739fe57781f87523375e8e1a3a4674bf6392cd6131a3222182b971320 \ @@ -166,9 +169,9 @@ google-api-core==2.10.2 \ # via # google-cloud-core # google-cloud-storage -google-auth==2.14.0 \ - --hash=sha256:1ad5b0e6eba5f69645971abb3d2c197537d5914070a8c6d30299dfdb07c5c700 \ - --hash=sha256:cf24817855d874ede2efd071aa22125445f555de1685b739a9782fcf408c2a3d +google-auth==2.14.1 \ + --hash=sha256:ccaa901f31ad5cbb562615eb8b664b3dd0bf5404a67618e642307f00613eda4d \ + --hash=sha256:f5d8701633bebc12e0deea4df8abd8aff31c28b355360597f7f2ee60f2e4d016 # via # gcp-releasetool # google-api-core @@ -178,9 +181,9 @@ google-cloud-core==2.3.2 \ --hash=sha256:8417acf6466be2fa85123441696c4badda48db314c607cf1e5d543fa8bdc22fe \ --hash=sha256:b9529ee7047fd8d4bf4a2182de619154240df17fbe60ead399078c1ae152af9a # via google-cloud-storage -google-cloud-storage==2.5.0 \ - --hash=sha256:19a26c66c317ce542cea0830b7e787e8dac2588b6bfa4d3fd3b871ba16305ab0 \ - --hash=sha256:382f34b91de2212e3c2e7b40ec079d27ee2e3dbbae99b75b1bcd8c63063ce235 +google-cloud-storage==2.6.0 \ + --hash=sha256:104ca28ae61243b637f2f01455cc8a05e8f15a2a18ced96cb587241cdd3820f5 \ + --hash=sha256:4ad0415ff61abdd8bb2ae81c1f8f7ec7d91a1011613f2db87c614c550f97bfe9 # via gcp-docuploader google-crc32c==1.5.0 \ --hash=sha256:024894d9d3cfbc5943f8f230e23950cd4906b2fe004c72e29b209420a1e6b05a \ @@ -256,9 +259,9 @@ google-resumable-media==2.4.0 \ --hash=sha256:2aa004c16d295c8f6c33b2b4788ba59d366677c0a25ae7382436cb30f776deaa \ --hash=sha256:8d5518502f92b9ecc84ac46779bd4f09694ecb3ba38a3e7ca737a86d15cbca1f # via google-cloud-storage -googleapis-common-protos==1.56.4 \ - --hash=sha256:8eb2cbc91b69feaf23e32452a7ae60e791e09967d81d4fcc7fc388182d1bd394 \ - --hash=sha256:c25873c47279387cfdcbdafa36149887901d36202cb645a0e4f29686bf6e4417 +googleapis-common-protos==1.57.0 \ + --hash=sha256:27a849d6205838fb6cc3c1c21cb9800707a661bb21c6ce7fb13e99eb1f8a0c46 \ + --hash=sha256:a9f4a1d7f6d9809657b7f1316a1aa527f6664891531bcfcc13b6696e685f443c # via google-api-core idna==3.4 \ --hash=sha256:814f528e8dead7d329833b91c5faa87d60bf71824cd12a7530b5526063d02cb4 \ @@ -269,6 +272,7 @@ importlib-metadata==5.0.0 \ --hash=sha256:ddb0e35065e8938f867ed4928d0ae5bf2a53b7773871bfe6bcc7e4fcdc7dea43 # via # -r requirements.in + # keyring # twine jaraco-classes==3.2.3 \ --hash=sha256:2353de3288bc6b82120752201c6b1c1a14b058267fa424ed5ce5984e3b922158 \ @@ -284,9 +288,9 @@ jinja2==3.1.2 \ --hash=sha256:31351a702a408a9e7595a8fc6150fc3f43bb6bf7e319770cbc0db9df9437e852 \ --hash=sha256:6088930bfe239f0e6710546ab9c19c9ef35e29792895fed6e6e31a023a182a61 # via gcp-releasetool -keyring==23.9.3 \ - --hash=sha256:69732a15cb1433bdfbc3b980a8a36a04878a6cfd7cb99f497b573f31618001c0 \ - --hash=sha256:69b01dd83c42f590250fe7a1f503fc229b14de83857314b1933a3ddbf595c4a5 +keyring==23.11.0 \ + --hash=sha256:3dd30011d555f1345dec2c262f0153f2f0ca6bca041fb1dc4588349bb4c0ac1e \ + --hash=sha256:ad192263e2cdd5f12875dedc2da13534359a7e760e77f8d04b50968a821c2361 # via # gcp-releasetool # twine @@ -350,9 +354,9 @@ pkginfo==1.8.3 \ --hash=sha256:848865108ec99d4901b2f7e84058b6e7660aae8ae10164e015a6dcf5b242a594 \ --hash=sha256:a84da4318dd86f870a9447a8c98340aa06216bfc6f2b7bdc4b8766984ae1867c # via twine -platformdirs==2.5.2 \ - --hash=sha256:027d8e83a2d7de06bbac4e5ef7e023c02b863d7ea5d079477e722bb41ab25788 \ - --hash=sha256:58c8abb07dcb441e6ee4b11d8df0ac856038f944ab98b7be6b27b2a3c7feef19 +platformdirs==2.5.4 \ + --hash=sha256:1006647646d80f16130f052404c6b901e80ee4ed6bef6792e1f238a8969106f7 \ + --hash=sha256:af0276409f9a02373d540bf8480021a048711d572745aef4b7842dad245eba10 # via virtualenv protobuf==3.20.3 \ --hash=sha256:03038ac1cfbc41aa21f6afcbcd357281d7521b4157926f30ebecc8d4ea59dcb7 \ @@ -381,7 +385,6 @@ protobuf==3.20.3 \ # gcp-docuploader # gcp-releasetool # google-api-core - # googleapis-common-protos py==1.11.0 \ --hash=sha256:51c75c4126074b472f746a24399ad32f6053d1b34b68d2fa41e558e6f4a98719 \ --hash=sha256:607c53218732647dff4acdfcd50cb62615cedf612e72d1724fb1a0cc6405b378 @@ -476,17 +479,17 @@ urllib3==1.26.12 \ # via # requests # twine -virtualenv==20.16.6 \ - --hash=sha256:186ca84254abcbde98180fd17092f9628c5fe742273c02724972a1d8a2035108 \ - --hash=sha256:530b850b523c6449406dfba859d6345e48ef19b8439606c5d74d7d3c9e14d76e +virtualenv==20.16.7 \ + --hash=sha256:8691e3ff9387f743e00f6bb20f70121f5e4f596cae754531f2b3b3a1b1ac696e \ + --hash=sha256:efd66b00386fdb7dbe4822d172303f40cd05e50e01740b19ea42425cbe653e29 # via nox webencodings==0.5.1 \ --hash=sha256:a0af1213f3c2226497a97e2b3aa01a7e4bee4f403f95be16fc9acd2947514a78 \ --hash=sha256:b36a1c245f2d304965eb4e0a82848379241dc04b865afcc4aab16748587e1923 # via bleach -wheel==0.37.1 \ - --hash=sha256:4bdcd7d840138086126cd09254dc6195fb4fc6f01c050a1d7236f2630db1d22a \ - --hash=sha256:e9a504e793efbca1b8e0e9cb979a249cf4a0a7b5b8c9e8b65a5e39d49529c1c4 +wheel==0.38.4 \ + --hash=sha256:965f5259b566725405b05e7cf774052044b1ed30119b5d586b2703aafe8719ac \ + --hash=sha256:b60533f3f5d530e971d6737ca6d58681ee434818fab630c83a734bb10c083ce8 # via -r requirements.in zipp==3.10.0 \ --hash=sha256:4fcb6f278987a6605757302a6e40e896257570d11c51628968ccb2a47e80c6c1 \ @@ -494,7 +497,7 @@ zipp==3.10.0 \ # via importlib-metadata # The following packages are considered to be unsafe in a requirements file: -setuptools==65.5.0 \ - --hash=sha256:512e5536220e38146176efb833d4a62aa726b7bbff82cfbc8ba9eaa3996e0b17 \ - --hash=sha256:f62ea9da9ed6289bfe868cd6845968a2c854d1427f8548d52cae02a42b4f0356 +setuptools==65.5.1 \ + --hash=sha256:d0b9a8433464d5800cbe05094acf5c6d52a91bfac9b52bcfc4d41382be5d5d31 \ + --hash=sha256:e197a19aa8ec9722928f2206f8de752def0e4c9fc6953527360d1c36d94ddb2f # via -r requirements.in diff --git a/noxfile.py b/noxfile.py index 54c2285..624fe86 100644 --- a/noxfile.py +++ b/noxfile.py @@ -272,7 +272,7 @@ def cover(session): session.run("coverage", "erase") -@nox.session(python=DEFAULT_PYTHON_VERSION) +@nox.session(python="3.9") def docs(session): """Build the docs for this library.""" @@ -298,7 +298,7 @@ def docs(session): ) -@nox.session(python=DEFAULT_PYTHON_VERSION) +@nox.session(python="3.9") def docfx(session): """Build the docfx yaml files for this library.""" From 42dc02f06a2d18917b66189618c58b2e0082e704 Mon Sep 17 00:00:00 2001 From: Mend Renovate Date: Mon, 21 Nov 2022 21:23:34 +0100 Subject: [PATCH 07/13] chore(deps): update dependency google-cloud-pubsub to v2.13.11 (#342) Co-authored-by: Anthonios Partheniou --- samples/snippets/requirements.txt | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/samples/snippets/requirements.txt b/samples/snippets/requirements.txt index ad6c968..d347745 100644 --- a/samples/snippets/requirements.txt +++ b/samples/snippets/requirements.txt @@ -1,4 +1,4 @@ -google-cloud-pubsub==2.13.10 +google-cloud-pubsub==2.13.11 google-cloud-containeranalysis==2.9.3 grafeas==1.6.1 pytest==7.2.0 From 3d91a44b7152fba0508d4d3099e51f18348b0512 Mon Sep 17 00:00:00 2001 From: "gcf-owl-bot[bot]" <78513119+gcf-owl-bot[bot]@users.noreply.github.com> Date: Tue, 22 Nov 2022 11:38:55 -0500 Subject: [PATCH 08/13] chore: Update gapic-generator-python to v1.6.1 (#340) MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit * chore: update to gapic-generator-python 1.5.0 feat: add support for `google.cloud..__version__` PiperOrigin-RevId: 484665853 Source-Link: https://github.com/googleapis/googleapis/commit/8eb249a19db926c2fbc4ecf1dc09c0e521a88b22 Source-Link: https://github.com/googleapis/googleapis-gen/commit/c8aa327b5f478865fc3fd91e3c2768e54e26ad44 Copy-Tag: eyJwIjoiLmdpdGh1Yi8uT3dsQm90LnlhbWwiLCJoIjoiYzhhYTMyN2I1ZjQ3ODg2NWZjM2ZkOTFlM2MyNzY4ZTU0ZTI2YWQ0NCJ9 * πŸ¦‰ Updates from OwlBot post-processor See https://github.com/googleapis/repo-automation-bots/blob/main/packages/owl-bot/README.md * update version in gapic_version.py * add .release-please-manifest.json with correct version * set manifest to true in .github/release-please.yml * add release-please-config.json * πŸ¦‰ Updates from OwlBot post-processor See https://github.com/googleapis/repo-automation-bots/blob/main/packages/owl-bot/README.md * chore: Update to gapic-generator-python 1.6.0 feat(python): Add typing to proto.Message based class attributes feat(python): Snippetgen handling of repeated enum field PiperOrigin-RevId: 487326846 Source-Link: https://github.com/googleapis/googleapis/commit/da380c77bb87ba0f752baf07605dd1db30e1f7e1 Source-Link: https://github.com/googleapis/googleapis-gen/commit/61ef5762ee6731a0cbbfea22fd0eecee51ab1c8e Copy-Tag: eyJwIjoiLmdpdGh1Yi8uT3dsQm90LnlhbWwiLCJoIjoiNjFlZjU3NjJlZTY3MzFhMGNiYmZlYTIyZmQwZWVjZWU1MWFiMWM4ZSJ9 * πŸ¦‰ Updates from OwlBot post-processor See https://github.com/googleapis/repo-automation-bots/blob/main/packages/owl-bot/README.md * feat: new APIs added to reflect updates to the filestore service - Add ENTERPRISE Tier - Add snapshot APIs: RevertInstance, ListSnapshots, CreateSnapshot, DeleteSnapshot, UpdateSnapshot - Add multi-share APIs: ListShares, GetShare, CreateShare, DeleteShare, UpdateShare - Add ConnectMode to NetworkConfig (for Private Service Access support) - New status codes (SUSPENDED/SUSPENDING, REVERTING/RESUMING) - Add SuspensionReason (for KMS related suspension) - Add new fields to Instance information: max_capacity_gb, capacity_step_size_gb, max_share_count, capacity_gb, multi_share_enabled PiperOrigin-RevId: 487492758 Source-Link: https://github.com/googleapis/googleapis/commit/5be5981f50322cf0c7388595e0f31ac5d0693469 Source-Link: https://github.com/googleapis/googleapis-gen/commit/ab0e217f560cc2c1afc11441c2eab6b6950efd2b Copy-Tag: eyJwIjoiLmdpdGh1Yi8uT3dsQm90LnlhbWwiLCJoIjoiYWIwZTIxN2Y1NjBjYzJjMWFmYzExNDQxYzJlYWI2YjY5NTBlZmQyYiJ9 * πŸ¦‰ Updates from OwlBot post-processor See https://github.com/googleapis/repo-automation-bots/blob/main/packages/owl-bot/README.md * update path to snippet metadata json * chore: Update gapic-generator-python to v1.6.1 PiperOrigin-RevId: 488036204 Source-Link: https://github.com/googleapis/googleapis/commit/08f275f5c1c0d99056e1cb68376323414459ee19 Source-Link: https://github.com/googleapis/googleapis-gen/commit/555c0945e60649e38739ae64bc45719cdf72178f Copy-Tag: eyJwIjoiLmdpdGh1Yi8uT3dsQm90LnlhbWwiLCJoIjoiNTU1YzA5NDVlNjA2NDllMzg3MzlhZTY0YmM0NTcxOWNkZjcyMTc4ZiJ9 * πŸ¦‰ Updates from OwlBot post-processor See https://github.com/googleapis/repo-automation-bots/blob/main/packages/owl-bot/README.md * use templated owlbot.py and setup.py * exclude generated constraints file * exclude autogenerated setup.py * update replacements in owlbot.py * run nox format session Co-authored-by: Owl Bot Co-authored-by: Anthonios Partheniou --- .github/release-please.yml | 1 + .release-please-manifest.json | 3 + docs/conf.py | 2 +- docs/containeranalysis_v1/types.rst | 1 - docs/index.rst | 20 +- .../devtools/containeranalysis/__init__.py | 13 +- .../containeranalysis/gapic_version.py | 16 ++ .../cloud/devtools/containeranalysis/py.typed | 2 +- .../devtools/containeranalysis_v1/__init__.py | 15 +- .../devtools/containeranalysis_v1/py.typed | 2 +- .../services/container_analysis/__init__.py | 2 +- .../container_analysis/async_client.py | 72 ++++--- .../services/container_analysis/client.py | 70 ++++--- .../container_analysis/transports/__init__.py | 1 - .../container_analysis/transports/base.py | 12 +- .../container_analysis/transports/grpc.py | 33 ++- .../transports/grpc_asyncio.py | 29 ++- .../types/containeranalysis.py | 26 +-- noxfile.py | 3 +- owlbot.py | 71 ++++--- release-please-config.json | 23 +++ ...container_analysis_get_iam_policy_async.py | 2 +- ..._container_analysis_get_iam_policy_sync.py | 2 +- ...vulnerability_occurrences_summary_async.py | 2 +- ..._vulnerability_occurrences_summary_sync.py | 2 +- ...container_analysis_set_iam_policy_async.py | 2 +- ..._container_analysis_set_iam_policy_sync.py | 2 +- ...ner_analysis_test_iam_permissions_async.py | 2 +- ...iner_analysis_test_iam_permissions_sync.py | 2 +- ...google.devtools.containeranalysis.v1.json} | 7 +- samples/snippets/noxfile.py | 16 +- samples/snippets/samples.py | 131 +++++++----- samples/snippets/samples_test.py | 193 ++++++++---------- setup.py | 44 ++-- testing/constraints-3.10.txt | 7 + testing/constraints-3.11.txt | 7 + testing/constraints-3.7.txt | 15 +- testing/constraints-3.8.txt | 7 + testing/constraints-3.9.txt | 7 + .../test_container_analysis.py | 34 ++- tests/unit/test_get_grafeas_client.py | 2 - 41 files changed, 506 insertions(+), 397 deletions(-) create mode 100644 .release-please-manifest.json create mode 100644 google/cloud/devtools/containeranalysis/gapic_version.py create mode 100644 release-please-config.json rename samples/generated_samples/{snippet_metadata_containeranalysis_v1.json => snippet_metadata_google.devtools.containeranalysis.v1.json} (99%) diff --git a/.github/release-please.yml b/.github/release-please.yml index 29601ad..fe749ff 100644 --- a/.github/release-please.yml +++ b/.github/release-please.yml @@ -1,5 +1,6 @@ releaseType: python handleGHRelease: true +manifest: true # NOTE: this section is generated by synthtool.languages.python # See https://github.com/googleapis/synthtool/blob/master/synthtool/languages/python.py branches: diff --git a/.release-please-manifest.json b/.release-please-manifest.json new file mode 100644 index 0000000..3542679 --- /dev/null +++ b/.release-please-manifest.json @@ -0,0 +1,3 @@ +{ + ".": "2.9.3" +} diff --git a/docs/conf.py b/docs/conf.py index c6405fc..fdba699 100644 --- a/docs/conf.py +++ b/docs/conf.py @@ -24,9 +24,9 @@ # All configuration values have a default; values that are commented out # serve to show the default. -import sys import os import shlex +import sys # If extensions (or modules to document with autodoc) are in another directory, # add these directories to sys.path here. If the directory is relative to the diff --git a/docs/containeranalysis_v1/types.rst b/docs/containeranalysis_v1/types.rst index 1fa561b..c0d2aa5 100644 --- a/docs/containeranalysis_v1/types.rst +++ b/docs/containeranalysis_v1/types.rst @@ -3,5 +3,4 @@ Types for Google Cloud Devtools Containeranalysis v1 API .. automodule:: google.cloud.devtools.containeranalysis_v1.types :members: - :undoc-members: :show-inheritance: diff --git a/docs/index.rst b/docs/index.rst index cbf09a5..c84fd2e 100644 --- a/docs/index.rst +++ b/docs/index.rst @@ -2,7 +2,8 @@ .. include:: multiprocessing.rst -Api Reference + +API Reference ------------- .. toctree:: :maxdepth: 2 @@ -11,22 +12,23 @@ Api Reference containeranalysis_v1/types -Changelog ----------- +Migration Guide +--------------- -For all previous ``google-cloud-containeranalysis`` releases: +See the guide below for instructions on migrating to the latest version. .. toctree:: :maxdepth: 2 - changelog + Β UPGRADING -Migration Guide ---------------- -See the guide below for instructions on migrating to the 2.x release of this library. +Changelog +--------- + +For a list of all ``google-cloud-containeranalysis`` releases: .. toctree:: :maxdepth: 2 - UPGRADING + changelog diff --git a/google/cloud/devtools/containeranalysis/__init__.py b/google/cloud/devtools/containeranalysis/__init__.py index c0a5734..10f8cec 100644 --- a/google/cloud/devtools/containeranalysis/__init__.py +++ b/google/cloud/devtools/containeranalysis/__init__.py @@ -13,18 +13,19 @@ # See the License for the specific language governing permissions and # limitations under the License. # +from google.cloud.devtools.containeranalysis import gapic_version as package_version + +__version__ = package_version.__version__ + -from google.cloud.devtools.containeranalysis_v1.services.container_analysis.client import ( - ContainerAnalysisClient, -) from google.cloud.devtools.containeranalysis_v1.services.container_analysis.async_client import ( ContainerAnalysisAsyncClient, ) - -from google.cloud.devtools.containeranalysis_v1.types.containeranalysis import ( - GetVulnerabilityOccurrencesSummaryRequest, +from google.cloud.devtools.containeranalysis_v1.services.container_analysis.client import ( + ContainerAnalysisClient, ) from google.cloud.devtools.containeranalysis_v1.types.containeranalysis import ( + GetVulnerabilityOccurrencesSummaryRequest, VulnerabilityOccurrencesSummary, ) diff --git a/google/cloud/devtools/containeranalysis/gapic_version.py b/google/cloud/devtools/containeranalysis/gapic_version.py new file mode 100644 index 0000000..085af7e --- /dev/null +++ b/google/cloud/devtools/containeranalysis/gapic_version.py @@ -0,0 +1,16 @@ +# -*- coding: utf-8 -*- +# Copyright 2022 Google LLC +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +# +__version__ = "2.9.3" # {x-release-please-version} diff --git a/google/cloud/devtools/containeranalysis/py.typed b/google/cloud/devtools/containeranalysis/py.typed index 73accaf..4fb2a9f 100644 --- a/google/cloud/devtools/containeranalysis/py.typed +++ b/google/cloud/devtools/containeranalysis/py.typed @@ -1,2 +1,2 @@ # Marker file for PEP 561. -# The google-cloud-devtools-containeranalysis package uses inline types. +# The google-cloud-containeranalysis package uses inline types. diff --git a/google/cloud/devtools/containeranalysis_v1/__init__.py b/google/cloud/devtools/containeranalysis_v1/__init__.py index 709ca57..de0b63e 100644 --- a/google/cloud/devtools/containeranalysis_v1/__init__.py +++ b/google/cloud/devtools/containeranalysis_v1/__init__.py @@ -13,12 +13,19 @@ # See the License for the specific language governing permissions and # limitations under the License. # +from google.cloud.devtools.containeranalysis import gapic_version as package_version -from .services.container_analysis import ContainerAnalysisClient -from .services.container_analysis import ContainerAnalysisAsyncClient +__version__ = package_version.__version__ -from .types.containeranalysis import GetVulnerabilityOccurrencesSummaryRequest -from .types.containeranalysis import VulnerabilityOccurrencesSummary + +from .services.container_analysis import ( + ContainerAnalysisAsyncClient, + ContainerAnalysisClient, +) +from .types.containeranalysis import ( + GetVulnerabilityOccurrencesSummaryRequest, + VulnerabilityOccurrencesSummary, +) __all__ = ( "ContainerAnalysisAsyncClient", diff --git a/google/cloud/devtools/containeranalysis_v1/py.typed b/google/cloud/devtools/containeranalysis_v1/py.typed index 73accaf..4fb2a9f 100644 --- a/google/cloud/devtools/containeranalysis_v1/py.typed +++ b/google/cloud/devtools/containeranalysis_v1/py.typed @@ -1,2 +1,2 @@ # Marker file for PEP 561. -# The google-cloud-devtools-containeranalysis package uses inline types. +# The google-cloud-containeranalysis package uses inline types. diff --git a/google/cloud/devtools/containeranalysis_v1/services/container_analysis/__init__.py b/google/cloud/devtools/containeranalysis_v1/services/container_analysis/__init__.py index da51d0c..5525a6b 100644 --- a/google/cloud/devtools/containeranalysis_v1/services/container_analysis/__init__.py +++ b/google/cloud/devtools/containeranalysis_v1/services/container_analysis/__init__.py @@ -13,8 +13,8 @@ # See the License for the specific language governing permissions and # limitations under the License. # -from .client import ContainerAnalysisClient from .async_client import ContainerAnalysisAsyncClient +from .client import ContainerAnalysisClient __all__ = ( "ContainerAnalysisClient", diff --git a/google/cloud/devtools/containeranalysis_v1/services/container_analysis/async_client.py b/google/cloud/devtools/containeranalysis_v1/services/container_analysis/async_client.py index f51d310..69e62d5 100644 --- a/google/cloud/devtools/containeranalysis_v1/services/container_analysis/async_client.py +++ b/google/cloud/devtools/containeranalysis_v1/services/container_analysis/async_client.py @@ -16,32 +16,42 @@ from collections import OrderedDict import functools import re -from typing import Dict, Mapping, Optional, Sequence, Tuple, Type, Union -import pkg_resources +from typing import ( + Dict, + Mapping, + MutableMapping, + MutableSequence, + Optional, + Sequence, + Tuple, + Type, + Union, +) -from google.api_core.client_options import ClientOptions from google.api_core import exceptions as core_exceptions from google.api_core import gapic_v1 from google.api_core import retry as retries +from google.api_core.client_options import ClientOptions from google.auth import credentials as ga_credentials # type: ignore from google.oauth2 import service_account # type: ignore +import pkg_resources try: OptionalRetry = Union[retries.Retry, gapic_v1.method._MethodDefault] except AttributeError: # pragma: NO COVER OptionalRetry = Union[retries.Retry, object] # type: ignore -from google.cloud.devtools.containeranalysis_v1.types import containeranalysis from google.iam.v1 import iam_policy_pb2 # type: ignore from google.iam.v1 import policy_pb2 # type: ignore -from .transports.base import ContainerAnalysisTransport, DEFAULT_CLIENT_INFO -from .transports.grpc_asyncio import ContainerAnalysisGrpcAsyncIOTransport -from .client import ContainerAnalysisClient - - from grafeas import grafeas_v1 from grafeas.grafeas_v1.services.grafeas import transports +from google.cloud.devtools.containeranalysis_v1.types import containeranalysis + +from .client import ContainerAnalysisClient +from .transports.base import DEFAULT_CLIENT_INFO, ContainerAnalysisTransport +from .transports.grpc_asyncio import ContainerAnalysisGrpcAsyncIOTransport + class ContainerAnalysisAsyncClient: """Retrieves analysis results of Cloud components such as Docker @@ -176,9 +186,9 @@ def transport(self) -> ContainerAnalysisTransport: def __init__( self, *, - credentials: ga_credentials.Credentials = None, + credentials: Optional[ga_credentials.Credentials] = None, transport: Union[str, ContainerAnalysisTransport] = "grpc_asyncio", - client_options: ClientOptions = None, + client_options: Optional[ClientOptions] = None, client_info: gapic_v1.client_info.ClientInfo = DEFAULT_CLIENT_INFO, ) -> None: """Instantiates the container analysis client. @@ -236,11 +246,11 @@ def get_grafeas_client(self) -> grafeas_v1.GrafeasClient: async def set_iam_policy( self, - request: Union[iam_policy_pb2.SetIamPolicyRequest, dict] = None, + request: Optional[Union[iam_policy_pb2.SetIamPolicyRequest, dict]] = None, *, - resource: str = None, + resource: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> policy_pb2.Policy: r"""Sets the access control policy on the specified note or @@ -281,7 +291,7 @@ async def sample_set_iam_policy(): print(response) Args: - request (Union[google.iam.v1.iam_policy_pb2.SetIamPolicyRequest, dict]): + request (Optional[Union[google.iam.v1.iam_policy_pb2.SetIamPolicyRequest, dict]]): The request object. Request message for `SetIamPolicy` method. resource (:class:`str`): @@ -410,11 +420,11 @@ async def sample_set_iam_policy(): async def get_iam_policy( self, - request: Union[iam_policy_pb2.GetIamPolicyRequest, dict] = None, + request: Optional[Union[iam_policy_pb2.GetIamPolicyRequest, dict]] = None, *, - resource: str = None, + resource: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> policy_pb2.Policy: r"""Gets the access control policy for a note or an occurrence @@ -455,7 +465,7 @@ async def sample_get_iam_policy(): print(response) Args: - request (Union[google.iam.v1.iam_policy_pb2.GetIamPolicyRequest, dict]): + request (Optional[Union[google.iam.v1.iam_policy_pb2.GetIamPolicyRequest, dict]]): The request object. Request message for `GetIamPolicy` method. resource (:class:`str`): @@ -584,12 +594,12 @@ async def sample_get_iam_policy(): async def test_iam_permissions( self, - request: Union[iam_policy_pb2.TestIamPermissionsRequest, dict] = None, + request: Optional[Union[iam_policy_pb2.TestIamPermissionsRequest, dict]] = None, *, - resource: str = None, - permissions: Sequence[str] = None, + resource: Optional[str] = None, + permissions: Optional[MutableSequence[str]] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> iam_policy_pb2.TestIamPermissionsResponse: r"""Returns the permissions that a caller has on the specified note @@ -630,7 +640,7 @@ async def sample_test_iam_permissions(): print(response) Args: - request (Union[google.iam.v1.iam_policy_pb2.TestIamPermissionsRequest, dict]): + request (Optional[Union[google.iam.v1.iam_policy_pb2.TestIamPermissionsRequest, dict]]): The request object. Request message for `TestIamPermissions` method. resource (:class:`str`): @@ -642,7 +652,7 @@ async def sample_test_iam_permissions(): This corresponds to the ``resource`` field on the ``request`` instance; if ``request`` is provided, this should not be set. - permissions (:class:`Sequence[str]`): + permissions (:class:`MutableSequence[str]`): The set of permissions to check for the ``resource``. Permissions with wildcards (such as '*' or 'storage.*') are not allowed. For more information see `IAM @@ -708,14 +718,14 @@ async def sample_test_iam_permissions(): async def get_vulnerability_occurrences_summary( self, - request: Union[ - containeranalysis.GetVulnerabilityOccurrencesSummaryRequest, dict + request: Optional[ + Union[containeranalysis.GetVulnerabilityOccurrencesSummaryRequest, dict] ] = None, *, - parent: str = None, - filter: str = None, + parent: Optional[str] = None, + filter: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> containeranalysis.VulnerabilityOccurrencesSummary: r"""Gets a summary of the number and severity of @@ -748,7 +758,7 @@ async def sample_get_vulnerability_occurrences_summary(): print(response) Args: - request (Union[google.cloud.devtools.containeranalysis_v1.types.GetVulnerabilityOccurrencesSummaryRequest, dict]): + request (Optional[Union[google.cloud.devtools.containeranalysis_v1.types.GetVulnerabilityOccurrencesSummaryRequest, dict]]): The request object. Request to get a vulnerability summary for some set of occurrences. parent (:class:`str`): diff --git a/google/cloud/devtools/containeranalysis_v1/services/container_analysis/client.py b/google/cloud/devtools/containeranalysis_v1/services/container_analysis/client.py index 2be47b2..2a1073a 100644 --- a/google/cloud/devtools/containeranalysis_v1/services/container_analysis/client.py +++ b/google/cloud/devtools/containeranalysis_v1/services/container_analysis/client.py @@ -16,35 +16,46 @@ from collections import OrderedDict import os import re -from typing import Dict, Mapping, Optional, Sequence, Tuple, Type, Union -import pkg_resources +from typing import ( + Dict, + Mapping, + MutableMapping, + MutableSequence, + Optional, + Sequence, + Tuple, + Type, + Union, + cast, +) from google.api_core import client_options as client_options_lib from google.api_core import exceptions as core_exceptions from google.api_core import gapic_v1 from google.api_core import retry as retries from google.auth import credentials as ga_credentials # type: ignore +from google.auth.exceptions import MutualTLSChannelError # type: ignore from google.auth.transport import mtls # type: ignore from google.auth.transport.grpc import SslCredentials # type: ignore -from google.auth.exceptions import MutualTLSChannelError # type: ignore from google.oauth2 import service_account # type: ignore +import pkg_resources try: OptionalRetry = Union[retries.Retry, gapic_v1.method._MethodDefault] except AttributeError: # pragma: NO COVER OptionalRetry = Union[retries.Retry, object] # type: ignore -from google.cloud.devtools.containeranalysis_v1.types import containeranalysis from google.iam.v1 import iam_policy_pb2 # type: ignore from google.iam.v1 import policy_pb2 # type: ignore -from .transports.base import ContainerAnalysisTransport, DEFAULT_CLIENT_INFO -from .transports.grpc import ContainerAnalysisGrpcTransport -from .transports.grpc_asyncio import ContainerAnalysisGrpcAsyncIOTransport - - from grafeas import grafeas_v1 from grafeas.grafeas_v1.services.grafeas import transports +from google.cloud.devtools.containeranalysis_v1.types import containeranalysis + +from .transports.base import DEFAULT_CLIENT_INFO, ContainerAnalysisTransport +from .transports.grpc import ContainerAnalysisGrpcTransport +from .transports.grpc_asyncio import ContainerAnalysisGrpcAsyncIOTransport + class ContainerAnalysisClientMeta(type): """Metaclass for the ContainerAnalysis client. @@ -62,7 +73,7 @@ class ContainerAnalysisClientMeta(type): def get_transport_class( cls, - label: str = None, + label: Optional[str] = None, ) -> Type[ContainerAnalysisTransport]: """Returns an appropriate transport class. @@ -330,8 +341,8 @@ def __init__( self, *, credentials: Optional[ga_credentials.Credentials] = None, - transport: Union[str, ContainerAnalysisTransport, None] = None, - client_options: Optional[client_options_lib.ClientOptions] = None, + transport: Optional[Union[str, ContainerAnalysisTransport]] = None, + client_options: Optional[Union[client_options_lib.ClientOptions, dict]] = None, client_info: gapic_v1.client_info.ClientInfo = DEFAULT_CLIENT_INFO, ) -> None: """Instantiates the container analysis client. @@ -345,7 +356,7 @@ def __init__( transport (Union[str, ContainerAnalysisTransport]): The transport to use. If set to None, a transport is chosen automatically. - client_options (google.api_core.client_options.ClientOptions): Custom options for the + client_options (Optional[Union[google.api_core.client_options.ClientOptions, dict]]): Custom options for the client. It won't take effect if a ``transport`` instance is provided. (1) The ``api_endpoint`` property can be used to override the default endpoint provided by the client. GOOGLE_API_USE_MTLS_ENDPOINT @@ -375,6 +386,7 @@ def __init__( client_options = client_options_lib.from_dict(client_options) if client_options is None: client_options = client_options_lib.ClientOptions() + client_options = cast(client_options_lib.ClientOptions, client_options) api_endpoint, client_cert_source_func = self.get_mtls_endpoint_and_cert_source( client_options @@ -441,11 +453,11 @@ def get_grafeas_client(self) -> grafeas_v1.GrafeasClient: def set_iam_policy( self, - request: Union[iam_policy_pb2.SetIamPolicyRequest, dict] = None, + request: Optional[Union[iam_policy_pb2.SetIamPolicyRequest, dict]] = None, *, - resource: str = None, + resource: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> policy_pb2.Policy: r"""Sets the access control policy on the specified note or @@ -612,11 +624,11 @@ def sample_set_iam_policy(): def get_iam_policy( self, - request: Union[iam_policy_pb2.GetIamPolicyRequest, dict] = None, + request: Optional[Union[iam_policy_pb2.GetIamPolicyRequest, dict]] = None, *, - resource: str = None, + resource: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> policy_pb2.Policy: r"""Gets the access control policy for a note or an occurrence @@ -783,12 +795,12 @@ def sample_get_iam_policy(): def test_iam_permissions( self, - request: Union[iam_policy_pb2.TestIamPermissionsRequest, dict] = None, + request: Optional[Union[iam_policy_pb2.TestIamPermissionsRequest, dict]] = None, *, - resource: str = None, - permissions: Sequence[str] = None, + resource: Optional[str] = None, + permissions: Optional[MutableSequence[str]] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> iam_policy_pb2.TestIamPermissionsResponse: r"""Returns the permissions that a caller has on the specified note @@ -841,7 +853,7 @@ def sample_test_iam_permissions(): This corresponds to the ``resource`` field on the ``request`` instance; if ``request`` is provided, this should not be set. - permissions (Sequence[str]): + permissions (MutableSequence[str]): The set of permissions to check for the ``resource``. Permissions with wildcards (such as '*' or 'storage.*') are not allowed. For more information see `IAM @@ -905,14 +917,14 @@ def sample_test_iam_permissions(): def get_vulnerability_occurrences_summary( self, - request: Union[ - containeranalysis.GetVulnerabilityOccurrencesSummaryRequest, dict + request: Optional[ + Union[containeranalysis.GetVulnerabilityOccurrencesSummaryRequest, dict] ] = None, *, - parent: str = None, - filter: str = None, + parent: Optional[str] = None, + filter: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: float = None, + timeout: Optional[float] = None, metadata: Sequence[Tuple[str, str]] = (), ) -> containeranalysis.VulnerabilityOccurrencesSummary: r"""Gets a summary of the number and severity of diff --git a/google/cloud/devtools/containeranalysis_v1/services/container_analysis/transports/__init__.py b/google/cloud/devtools/containeranalysis_v1/services/container_analysis/transports/__init__.py index 4f6c5a2..ae59bc4 100644 --- a/google/cloud/devtools/containeranalysis_v1/services/container_analysis/transports/__init__.py +++ b/google/cloud/devtools/containeranalysis_v1/services/container_analysis/transports/__init__.py @@ -20,7 +20,6 @@ from .grpc import ContainerAnalysisGrpcTransport from .grpc_asyncio import ContainerAnalysisGrpcAsyncIOTransport - # Compile a registry of transports. _transport_registry = OrderedDict() # type: Dict[str, Type[ContainerAnalysisTransport]] _transport_registry["grpc"] = ContainerAnalysisGrpcTransport diff --git a/google/cloud/devtools/containeranalysis_v1/services/container_analysis/transports/base.py b/google/cloud/devtools/containeranalysis_v1/services/container_analysis/transports/base.py index 4349442..37d7e1c 100644 --- a/google/cloud/devtools/containeranalysis_v1/services/container_analysis/transports/base.py +++ b/google/cloud/devtools/containeranalysis_v1/services/container_analysis/transports/base.py @@ -15,24 +15,24 @@ # import abc from typing import Awaitable, Callable, Dict, Optional, Sequence, Union -import pkg_resources -import google.auth # type: ignore import google.api_core from google.api_core import exceptions as core_exceptions from google.api_core import gapic_v1 from google.api_core import retry as retries +import google.auth # type: ignore from google.auth import credentials as ga_credentials # type: ignore +from google.iam.v1 import iam_policy_pb2 # type: ignore +from google.iam.v1 import policy_pb2 # type: ignore from google.oauth2 import service_account # type: ignore +import pkg_resources from google.cloud.devtools.containeranalysis_v1.types import containeranalysis -from google.iam.v1 import iam_policy_pb2 # type: ignore -from google.iam.v1 import policy_pb2 # type: ignore try: DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( gapic_version=pkg_resources.get_distribution( - "google-cloud-devtools-containeranalysis", + "google-cloud-containeranalysis", ).version, ) except pkg_resources.DistributionNotFound: @@ -50,7 +50,7 @@ def __init__( self, *, host: str = DEFAULT_HOST, - credentials: ga_credentials.Credentials = None, + credentials: Optional[ga_credentials.Credentials] = None, credentials_file: Optional[str] = None, scopes: Optional[Sequence[str]] = None, quota_project_id: Optional[str] = None, diff --git a/google/cloud/devtools/containeranalysis_v1/services/container_analysis/transports/grpc.py b/google/cloud/devtools/containeranalysis_v1/services/container_analysis/transports/grpc.py index 969a87b..73d3255 100644 --- a/google/cloud/devtools/containeranalysis_v1/services/container_analysis/transports/grpc.py +++ b/google/cloud/devtools/containeranalysis_v1/services/container_analysis/transports/grpc.py @@ -13,21 +13,20 @@ # See the License for the specific language governing permissions and # limitations under the License. # -import warnings from typing import Callable, Dict, Optional, Sequence, Tuple, Union +import warnings -from google.api_core import grpc_helpers -from google.api_core import gapic_v1 +from google.api_core import gapic_v1, grpc_helpers import google.auth # type: ignore from google.auth import credentials as ga_credentials # type: ignore from google.auth.transport.grpc import SslCredentials # type: ignore - +from google.iam.v1 import iam_policy_pb2 # type: ignore +from google.iam.v1 import policy_pb2 # type: ignore import grpc # type: ignore from google.cloud.devtools.containeranalysis_v1.types import containeranalysis -from google.iam.v1 import iam_policy_pb2 # type: ignore -from google.iam.v1 import policy_pb2 # type: ignore -from .base import ContainerAnalysisTransport, DEFAULT_CLIENT_INFO + +from .base import DEFAULT_CLIENT_INFO, ContainerAnalysisTransport class ContainerAnalysisGrpcTransport(ContainerAnalysisTransport): @@ -63,14 +62,14 @@ def __init__( self, *, host: str = "containeranalysis.googleapis.com", - credentials: ga_credentials.Credentials = None, - credentials_file: str = None, - scopes: Sequence[str] = None, - channel: grpc.Channel = None, - api_mtls_endpoint: str = None, - client_cert_source: Callable[[], Tuple[bytes, bytes]] = None, - ssl_channel_credentials: grpc.ChannelCredentials = None, - client_cert_source_for_mtls: Callable[[], Tuple[bytes, bytes]] = None, + credentials: Optional[ga_credentials.Credentials] = None, + credentials_file: Optional[str] = None, + scopes: Optional[Sequence[str]] = None, + channel: Optional[grpc.Channel] = None, + api_mtls_endpoint: Optional[str] = None, + client_cert_source: Optional[Callable[[], Tuple[bytes, bytes]]] = None, + ssl_channel_credentials: Optional[grpc.ChannelCredentials] = None, + client_cert_source_for_mtls: Optional[Callable[[], Tuple[bytes, bytes]]] = None, quota_project_id: Optional[str] = None, client_info: gapic_v1.client_info.ClientInfo = DEFAULT_CLIENT_INFO, always_use_jwt_access: Optional[bool] = False, @@ -197,8 +196,8 @@ def __init__( def create_channel( cls, host: str = "containeranalysis.googleapis.com", - credentials: ga_credentials.Credentials = None, - credentials_file: str = None, + credentials: Optional[ga_credentials.Credentials] = None, + credentials_file: Optional[str] = None, scopes: Optional[Sequence[str]] = None, quota_project_id: Optional[str] = None, **kwargs, diff --git a/google/cloud/devtools/containeranalysis_v1/services/container_analysis/transports/grpc_asyncio.py b/google/cloud/devtools/containeranalysis_v1/services/container_analysis/transports/grpc_asyncio.py index 6a408d4..a0c2731 100644 --- a/google/cloud/devtools/containeranalysis_v1/services/container_analysis/transports/grpc_asyncio.py +++ b/google/cloud/devtools/containeranalysis_v1/services/container_analysis/transports/grpc_asyncio.py @@ -13,21 +13,20 @@ # See the License for the specific language governing permissions and # limitations under the License. # -import warnings from typing import Awaitable, Callable, Dict, Optional, Sequence, Tuple, Union +import warnings -from google.api_core import gapic_v1 -from google.api_core import grpc_helpers_async +from google.api_core import gapic_v1, grpc_helpers_async from google.auth import credentials as ga_credentials # type: ignore from google.auth.transport.grpc import SslCredentials # type: ignore - +from google.iam.v1 import iam_policy_pb2 # type: ignore +from google.iam.v1 import policy_pb2 # type: ignore import grpc # type: ignore from grpc.experimental import aio # type: ignore from google.cloud.devtools.containeranalysis_v1.types import containeranalysis -from google.iam.v1 import iam_policy_pb2 # type: ignore -from google.iam.v1 import policy_pb2 # type: ignore -from .base import ContainerAnalysisTransport, DEFAULT_CLIENT_INFO + +from .base import DEFAULT_CLIENT_INFO, ContainerAnalysisTransport from .grpc import ContainerAnalysisGrpcTransport @@ -65,7 +64,7 @@ class ContainerAnalysisGrpcAsyncIOTransport(ContainerAnalysisTransport): def create_channel( cls, host: str = "containeranalysis.googleapis.com", - credentials: ga_credentials.Credentials = None, + credentials: Optional[ga_credentials.Credentials] = None, credentials_file: Optional[str] = None, scopes: Optional[Sequence[str]] = None, quota_project_id: Optional[str] = None, @@ -108,15 +107,15 @@ def __init__( self, *, host: str = "containeranalysis.googleapis.com", - credentials: ga_credentials.Credentials = None, + credentials: Optional[ga_credentials.Credentials] = None, credentials_file: Optional[str] = None, scopes: Optional[Sequence[str]] = None, - channel: aio.Channel = None, - api_mtls_endpoint: str = None, - client_cert_source: Callable[[], Tuple[bytes, bytes]] = None, - ssl_channel_credentials: grpc.ChannelCredentials = None, - client_cert_source_for_mtls: Callable[[], Tuple[bytes, bytes]] = None, - quota_project_id=None, + channel: Optional[aio.Channel] = None, + api_mtls_endpoint: Optional[str] = None, + client_cert_source: Optional[Callable[[], Tuple[bytes, bytes]]] = None, + ssl_channel_credentials: Optional[grpc.ChannelCredentials] = None, + client_cert_source_for_mtls: Optional[Callable[[], Tuple[bytes, bytes]]] = None, + quota_project_id: Optional[str] = None, client_info: gapic_v1.client_info.ClientInfo = DEFAULT_CLIENT_INFO, always_use_jwt_access: Optional[bool] = False, api_audience: Optional[str] = None, diff --git a/google/cloud/devtools/containeranalysis_v1/types/containeranalysis.py b/google/cloud/devtools/containeranalysis_v1/types/containeranalysis.py index 4480a7a..08b4aee 100644 --- a/google/cloud/devtools/containeranalysis_v1/types/containeranalysis.py +++ b/google/cloud/devtools/containeranalysis_v1/types/containeranalysis.py @@ -13,10 +13,10 @@ # See the License for the specific language governing permissions and # limitations under the License. # -import proto # type: ignore - -from grafeas.grafeas_v1.types import severity # type: ignore +from typing import MutableMapping, MutableSequence +import grafeas.grafeas_v1 # type: ignore +import proto # type: ignore __protobuf__ = proto.module( package="google.devtools.containeranalysis.v1", @@ -39,11 +39,11 @@ class GetVulnerabilityOccurrencesSummaryRequest(proto.Message): The filter expression. """ - parent = proto.Field( + parent: str = proto.Field( proto.STRING, number=1, ) - filter = proto.Field( + filter: str = proto.Field( proto.STRING, number=2, ) @@ -54,7 +54,7 @@ class VulnerabilityOccurrencesSummary(proto.Message): resource and severity type. Attributes: - counts (Sequence[google.cloud.devtools.containeranalysis_v1.types.VulnerabilityOccurrencesSummary.FixableTotalByDigest]): + counts (MutableSequence[google.cloud.devtools.containeranalysis_v1.types.VulnerabilityOccurrencesSummary.FixableTotalByDigest]): A listing by resource of the number of fixable and total vulnerabilities. """ @@ -66,7 +66,7 @@ class FixableTotalByDigest(proto.Message): Attributes: resource_uri (str): The affected resource. - severity (grafeas.v1.severity.Severity): + severity (grafeas.v1.grafeas.grafeas_v1.Severity): The severity for this count. SEVERITY_UNSPECIFIED indicates total across all severities. fixable_count (int): @@ -77,25 +77,25 @@ class FixableTotalByDigest(proto.Message): associated with this resource. """ - resource_uri = proto.Field( + resource_uri: str = proto.Field( proto.STRING, number=1, ) - severity = proto.Field( + severity: grafeas.grafeas_v1.Severity = proto.Field( proto.ENUM, number=2, - enum=severity.Severity, + enum=grafeas.grafeas_v1.Severity, ) - fixable_count = proto.Field( + fixable_count: int = proto.Field( proto.INT64, number=3, ) - total_count = proto.Field( + total_count: int = proto.Field( proto.INT64, number=4, ) - counts = proto.RepeatedField( + counts: MutableSequence[FixableTotalByDigest] = proto.RepeatedField( proto.MESSAGE, number=1, message=FixableTotalByDigest, diff --git a/noxfile.py b/noxfile.py index 624fe86..d8440c0 100644 --- a/noxfile.py +++ b/noxfile.py @@ -17,6 +17,7 @@ # Generated by synthtool. DO NOT EDIT! from __future__ import absolute_import + import os import pathlib import re @@ -267,7 +268,7 @@ def cover(session): test runs (not system test runs), and then erases coverage data. """ session.install("coverage", "pytest-cov") - session.run("coverage", "report", "--show-missing", "--fail-under=98") + session.run("coverage", "report", "--show-missing", "--fail-under=100") session.run("coverage", "erase") diff --git a/owlbot.py b/owlbot.py index 6b2c822..f970ec7 100644 --- a/owlbot.py +++ b/owlbot.py @@ -1,4 +1,4 @@ -# Copyright 2018 Google LLC +# Copyright 2022 Google LLC # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. @@ -12,38 +12,38 @@ # See the License for the specific language governing permissions and # limitations under the License. -"""This script is used to synthesize generated parts of this library.""" +import json +from pathlib import Path +import shutil import synthtool as s import synthtool.gcp as gcp -import logging from synthtool.languages import python -logging.basicConfig(level=logging.DEBUG) +# ---------------------------------------------------------------------------- +# Copy the generated client from the owl-bot staging directory +# ---------------------------------------------------------------------------- -common = gcp.CommonTemplates() +clean_up_generated_samples = True -default_version = "v1" +# Load the default version defined in .repo-metadata.json. +default_version = json.load(open(".repo-metadata.json", "rt")).get("default_version") for library in s.get_staging_dirs(default_version): - s.replace( - library / "google/**/*client.py", - r"""google-cloud-devtools-containeranalysis""", - r"""google-cloud-containeranalysis""", - ) + if clean_up_generated_samples: + shutil.rmtree("samples/generated_samples", ignore_errors=True) + clean_up_generated_samples = False # Fix imported type from grafeas s.replace( library / "google/**/types/containeranalysis.py", "from grafeas\.v1 import severity_pb2", - "from grafeas.grafeas_v1.types import severity" + "import grafeas.grafeas_v1", ) # Fix imported type from grafeas s.replace( - library / "google/**/types/containeranalysis.py", - "severity_pb2", - "severity" + library / "google/**/types/containeranalysis.py", "severity_pb2", "grafeas.grafeas_v1" ) # Insert helper method to get grafeas client @@ -66,7 +66,7 @@ s.replace( library / "google/**/client.py", r"""(\s+)def set_iam_policy\(""", - r'''\n\g<1>def get_grafeas_client( + r"""\n\g<1>def get_grafeas_client( self ) -> grafeas_v1.GrafeasClient: grafeas_transport = grafeas_v1.services.grafeas.transports.GrafeasGrpcTransport( @@ -81,13 +81,13 @@ return grafeas_v1.GrafeasClient(transport=grafeas_transport) \g<1># Service calls - \g<1>def set_iam_policy(''', + \g<1>def set_iam_policy(""", ) s.replace( library / "google/**/async_client.py", r"""(\s+)async def set_iam_policy\(""", - r'''\n\g<1>def get_grafeas_client( + r"""\n\g<1>def get_grafeas_client( self ) -> grafeas_v1.GrafeasClient: grafeas_transport = grafeas_v1.services.grafeas.transports.GrafeasGrpcTransport( @@ -102,11 +102,12 @@ return grafeas_v1.GrafeasClient(transport=grafeas_transport) \g<1># Service calls - \g<1>async def set_iam_policy(''', + \g<1>async def set_iam_policy(""", ) # Add test to ensure that credentials propagate to client.get_grafeas_client() - num_replacements = s.replace(library / "tests/**/test_container_analysis.py", + num_replacements = s.replace( + library / "tests/**/test_container_analysis.py", """create_channel.assert_called_with\( "containeranalysis.googleapis.com:443", credentials=file_creds, @@ -142,31 +143,33 @@ # Also check client.get_grafeas_client() to make sure that the file credentials are used assert file_creds == client.get_grafeas_client().transport._credentials - """ + """, ) assert num_replacements == 1 - s.move(library, excludes=["setup.py", "README.rst", "docs/index.rst"]) - + s.move([library], excludes=["**/gapic_version.py", "setup.py", "testing/constraints-3.7.txt"]) s.remove_staging_dirs() # ---------------------------------------------------------------------------- # Add templated files # ---------------------------------------------------------------------------- -templated_files = common.py_library( - samples=False, # set to True only if there are samples + +templated_files = gcp.CommonTemplates().py_library( + cov_level=100, microgenerator=True, - cov_level=98, + versions=gcp.common.detect_versions(path="./google", default_first=True), ) -s.move(templated_files, +s.move( + templated_files, excludes=[ - ".coveragerc", # microgenerator has a good coveragerc - ".github/workflows", # exclude templated gh actions as tests require credentials - ] - ) - -python.configure_previous_major_version_branches() + ".coveragerc", + ".github/release-please.yml", + ".github/workflows", + ], +) # exclude templated gh actions as tests require credentials python.py_samples(skip_readmes=True) -s.shell.run(["nox", "-s", "blacken"], hide_output=False) \ No newline at end of file +# run format session for all directories which have a noxfile +for noxfile in Path(".").glob("**/noxfile.py"): + s.shell.run(["nox", "-s", "format"], cwd=noxfile.parent, hide_output=False) diff --git a/release-please-config.json b/release-please-config.json new file mode 100644 index 0000000..595994a --- /dev/null +++ b/release-please-config.json @@ -0,0 +1,23 @@ +{ + "$schema": "https://raw.githubusercontent.com/googleapis/release-please/main/schemas/config.json", + "packages": { + ".": { + "release-type": "python", + "extra-files": [ + "google/cloud/devtools/containeranalysis/gapic_version.py", + { + "type": "json", + "path": "samples/generated_samples/snippet_metadata_google.devtools.containeranalysis.v1.json", + "jsonpath": "$.clientLibrary.version" + } + ] + } + }, + "release-type": "python", + "plugins": [ + { + "type": "sentence-case" + } + ], + "initial-version": "0.1.0" +} diff --git a/samples/generated_samples/containeranalysis_v1_generated_container_analysis_get_iam_policy_async.py b/samples/generated_samples/containeranalysis_v1_generated_container_analysis_get_iam_policy_async.py index a23a384..c075f96 100644 --- a/samples/generated_samples/containeranalysis_v1_generated_container_analysis_get_iam_policy_async.py +++ b/samples/generated_samples/containeranalysis_v1_generated_container_analysis_get_iam_policy_async.py @@ -20,7 +20,7 @@ # It may require modifications to work in your environment. # To install the latest published package dependency, execute the following: -# python3 -m pip install google-cloud-devtools-containeranalysis +# python3 -m pip install google-cloud-containeranalysis # [START containeranalysis_v1_generated_ContainerAnalysis_GetIamPolicy_async] diff --git a/samples/generated_samples/containeranalysis_v1_generated_container_analysis_get_iam_policy_sync.py b/samples/generated_samples/containeranalysis_v1_generated_container_analysis_get_iam_policy_sync.py index 94c5c99..c5c5d1d 100644 --- a/samples/generated_samples/containeranalysis_v1_generated_container_analysis_get_iam_policy_sync.py +++ b/samples/generated_samples/containeranalysis_v1_generated_container_analysis_get_iam_policy_sync.py @@ -20,7 +20,7 @@ # It may require modifications to work in your environment. # To install the latest published package dependency, execute the following: -# python3 -m pip install google-cloud-devtools-containeranalysis +# python3 -m pip install google-cloud-containeranalysis # [START containeranalysis_v1_generated_ContainerAnalysis_GetIamPolicy_sync] diff --git a/samples/generated_samples/containeranalysis_v1_generated_container_analysis_get_vulnerability_occurrences_summary_async.py b/samples/generated_samples/containeranalysis_v1_generated_container_analysis_get_vulnerability_occurrences_summary_async.py index fe88fb5..690acda 100644 --- a/samples/generated_samples/containeranalysis_v1_generated_container_analysis_get_vulnerability_occurrences_summary_async.py +++ b/samples/generated_samples/containeranalysis_v1_generated_container_analysis_get_vulnerability_occurrences_summary_async.py @@ -20,7 +20,7 @@ # It may require modifications to work in your environment. # To install the latest published package dependency, execute the following: -# python3 -m pip install google-cloud-devtools-containeranalysis +# python3 -m pip install google-cloud-containeranalysis # [START containeranalysis_v1_generated_ContainerAnalysis_GetVulnerabilityOccurrencesSummary_async] diff --git a/samples/generated_samples/containeranalysis_v1_generated_container_analysis_get_vulnerability_occurrences_summary_sync.py b/samples/generated_samples/containeranalysis_v1_generated_container_analysis_get_vulnerability_occurrences_summary_sync.py index 97371c3..ee847bb 100644 --- a/samples/generated_samples/containeranalysis_v1_generated_container_analysis_get_vulnerability_occurrences_summary_sync.py +++ b/samples/generated_samples/containeranalysis_v1_generated_container_analysis_get_vulnerability_occurrences_summary_sync.py @@ -20,7 +20,7 @@ # It may require modifications to work in your environment. # To install the latest published package dependency, execute the following: -# python3 -m pip install google-cloud-devtools-containeranalysis +# python3 -m pip install google-cloud-containeranalysis # [START containeranalysis_v1_generated_ContainerAnalysis_GetVulnerabilityOccurrencesSummary_sync] diff --git a/samples/generated_samples/containeranalysis_v1_generated_container_analysis_set_iam_policy_async.py b/samples/generated_samples/containeranalysis_v1_generated_container_analysis_set_iam_policy_async.py index aeebc69..ceaf936 100644 --- a/samples/generated_samples/containeranalysis_v1_generated_container_analysis_set_iam_policy_async.py +++ b/samples/generated_samples/containeranalysis_v1_generated_container_analysis_set_iam_policy_async.py @@ -20,7 +20,7 @@ # It may require modifications to work in your environment. # To install the latest published package dependency, execute the following: -# python3 -m pip install google-cloud-devtools-containeranalysis +# python3 -m pip install google-cloud-containeranalysis # [START containeranalysis_v1_generated_ContainerAnalysis_SetIamPolicy_async] diff --git a/samples/generated_samples/containeranalysis_v1_generated_container_analysis_set_iam_policy_sync.py b/samples/generated_samples/containeranalysis_v1_generated_container_analysis_set_iam_policy_sync.py index d034834..35319f3 100644 --- a/samples/generated_samples/containeranalysis_v1_generated_container_analysis_set_iam_policy_sync.py +++ b/samples/generated_samples/containeranalysis_v1_generated_container_analysis_set_iam_policy_sync.py @@ -20,7 +20,7 @@ # It may require modifications to work in your environment. # To install the latest published package dependency, execute the following: -# python3 -m pip install google-cloud-devtools-containeranalysis +# python3 -m pip install google-cloud-containeranalysis # [START containeranalysis_v1_generated_ContainerAnalysis_SetIamPolicy_sync] diff --git a/samples/generated_samples/containeranalysis_v1_generated_container_analysis_test_iam_permissions_async.py b/samples/generated_samples/containeranalysis_v1_generated_container_analysis_test_iam_permissions_async.py index b5447d7..0fd4a88 100644 --- a/samples/generated_samples/containeranalysis_v1_generated_container_analysis_test_iam_permissions_async.py +++ b/samples/generated_samples/containeranalysis_v1_generated_container_analysis_test_iam_permissions_async.py @@ -20,7 +20,7 @@ # It may require modifications to work in your environment. # To install the latest published package dependency, execute the following: -# python3 -m pip install google-cloud-devtools-containeranalysis +# python3 -m pip install google-cloud-containeranalysis # [START containeranalysis_v1_generated_ContainerAnalysis_TestIamPermissions_async] diff --git a/samples/generated_samples/containeranalysis_v1_generated_container_analysis_test_iam_permissions_sync.py b/samples/generated_samples/containeranalysis_v1_generated_container_analysis_test_iam_permissions_sync.py index b1b9297..97938d0 100644 --- a/samples/generated_samples/containeranalysis_v1_generated_container_analysis_test_iam_permissions_sync.py +++ b/samples/generated_samples/containeranalysis_v1_generated_container_analysis_test_iam_permissions_sync.py @@ -20,7 +20,7 @@ # It may require modifications to work in your environment. # To install the latest published package dependency, execute the following: -# python3 -m pip install google-cloud-devtools-containeranalysis +# python3 -m pip install google-cloud-containeranalysis # [START containeranalysis_v1_generated_ContainerAnalysis_TestIamPermissions_sync] diff --git a/samples/generated_samples/snippet_metadata_containeranalysis_v1.json b/samples/generated_samples/snippet_metadata_google.devtools.containeranalysis.v1.json similarity index 99% rename from samples/generated_samples/snippet_metadata_containeranalysis_v1.json rename to samples/generated_samples/snippet_metadata_google.devtools.containeranalysis.v1.json index e7675a2..dab3e1e 100644 --- a/samples/generated_samples/snippet_metadata_containeranalysis_v1.json +++ b/samples/generated_samples/snippet_metadata_google.devtools.containeranalysis.v1.json @@ -7,7 +7,8 @@ } ], "language": "PYTHON", - "name": "google-cloud-devtools-containeranalysis" + "name": "google-cloud-containeranalysis", + "version": "0.1.0" }, "snippets": [ { @@ -529,7 +530,7 @@ }, { "name": "permissions", - "type": "Sequence[str]" + "type": "MutableSequence[str]" }, { "name": "retry", @@ -613,7 +614,7 @@ }, { "name": "permissions", - "type": "Sequence[str]" + "type": "MutableSequence[str]" }, { "name": "retry", diff --git a/samples/snippets/noxfile.py b/samples/snippets/noxfile.py index 0398d72..c171513 100644 --- a/samples/snippets/noxfile.py +++ b/samples/snippets/noxfile.py @@ -22,7 +22,6 @@ import nox - # WARNING - WARNING - WARNING - WARNING - WARNING # WARNING - WARNING - WARNING - WARNING - WARNING # DO NOT EDIT THIS FILE EVER! @@ -180,6 +179,7 @@ def blacken(session: nox.sessions.Session) -> None: # format = isort + black # + @nox.session def format(session: nox.sessions.Session) -> None: """ @@ -207,7 +207,9 @@ def _session_tests( session: nox.sessions.Session, post_install: Callable = None ) -> None: # check for presence of tests - test_list = glob.glob("**/*_test.py", recursive=True) + glob.glob("**/test_*.py", recursive=True) + test_list = glob.glob("**/*_test.py", recursive=True) + glob.glob( + "**/test_*.py", recursive=True + ) test_list.extend(glob.glob("**/tests", recursive=True)) if len(test_list) == 0: @@ -229,9 +231,7 @@ def _session_tests( if os.path.exists("requirements-test.txt"): if os.path.exists("constraints-test.txt"): - session.install( - "-r", "requirements-test.txt", "-c", "constraints-test.txt" - ) + session.install("-r", "requirements-test.txt", "-c", "constraints-test.txt") else: session.install("-r", "requirements-test.txt") with open("requirements-test.txt") as rtfile: @@ -244,9 +244,9 @@ def _session_tests( post_install(session) if "pytest-parallel" in packages: - concurrent_args.extend(['--workers', 'auto', '--tests-per-worker', 'auto']) + concurrent_args.extend(["--workers", "auto", "--tests-per-worker", "auto"]) elif "pytest-xdist" in packages: - concurrent_args.extend(['-n', 'auto']) + concurrent_args.extend(["-n", "auto"]) session.run( "pytest", @@ -276,7 +276,7 @@ def py(session: nox.sessions.Session) -> None: def _get_repo_root() -> Optional[str]: - """ Returns the root folder of the project. """ + """Returns the root folder of the project.""" # Get root of this repository. Assume we don't have directories nested deeper than 10 items. p = Path(os.getcwd()) for i in range(10): diff --git a/samples/snippets/samples.py b/samples/snippets/samples.py index ecf28c4..3f22117 100644 --- a/samples/snippets/samples.py +++ b/samples/snippets/samples.py @@ -20,30 +20,30 @@ def create_note(note_id, project_id): # note_id = 'my-note' # project_id = 'my-gcp-project' - from grafeas.grafeas_v1 import Version from google.cloud.devtools import containeranalysis_v1 + from grafeas.grafeas_v1 import Version client = containeranalysis_v1.ContainerAnalysisClient() grafeas_client = client.get_grafeas_client() project_name = f"projects/{project_id}" note = { - 'vulnerability': { - 'details': [ + "vulnerability": { + "details": [ { - 'affected_cpe_uri': 'your-uri-here', - 'affected_package': 'your-package-here', - 'affected_version_start': { - 'kind': Version.VersionKind.MINIMUM - }, - 'fixed_version': { - 'kind': Version.VersionKind.MAXIMUM - } + "affected_cpe_uri": "your-uri-here", + "affected_package": "your-package-here", + "affected_version_start": {"kind": Version.VersionKind.MINIMUM}, + "fixed_version": {"kind": Version.VersionKind.MAXIMUM}, } ] } } - response = grafeas_client.create_note(parent=project_name, note_id=note_id, note=note) + response = grafeas_client.create_note( + parent=project_name, note_id=note_id, note=note + ) return response + + # [END containeranalysis_create_note] @@ -60,20 +60,22 @@ def delete_note(note_id, project_id): note_name = f"projects/{project_id}/notes/{note_id}" grafeas_client.delete_note(name=note_name) + + # [END containeranalysis_delete_note] # [START containeranalysis_create_occurrence] def create_occurrence(resource_url, note_id, occurrence_project, note_project): - """ Creates and returns a new occurrence of a previously + """Creates and returns a new occurrence of a previously created vulnerability note.""" # resource_url = 'https://gcr.io/my-project/my-image@sha256:123' # note_id = 'my-note' # occurrence_project = 'my-gcp-project' # note_project = 'my-gcp-project' - from grafeas.grafeas_v1 import Version from google.cloud.devtools import containeranalysis_v1 + from grafeas.grafeas_v1 import Version client = containeranalysis_v1.ContainerAnalysisClient() grafeas_client = client.get_grafeas_client() @@ -81,25 +83,25 @@ def create_occurrence(resource_url, note_id, occurrence_project, note_project): formatted_project = f"projects/{occurrence_project}" occurrence = { - 'note_name': formatted_note, - 'resource_uri': resource_url, - 'vulnerability': { - 'package_issue': [ + "note_name": formatted_note, + "resource_uri": resource_url, + "vulnerability": { + "package_issue": [ { - 'affected_cpe_uri': 'your-uri-here', - 'affected_package': 'your-package-here', - 'affected_version': { - 'kind': Version.VersionKind.MINIMUM - }, - 'fixed_version': { - 'kind': Version.VersionKind.MAXIMUM - } + "affected_cpe_uri": "your-uri-here", + "affected_package": "your-package-here", + "affected_version": {"kind": Version.VersionKind.MINIMUM}, + "fixed_version": {"kind": Version.VersionKind.MAXIMUM}, } ] - } + }, } - return grafeas_client.create_occurrence(parent=formatted_project, occurrence=occurrence) + return grafeas_client.create_occurrence( + parent=formatted_project, occurrence=occurrence + ) + + # [END containeranalysis_create_occurrence] @@ -115,6 +117,8 @@ def delete_occurrence(occurrence_id, project_id): grafeas_client = client.get_grafeas_client() parent = f"projects/{project_id}/occurrences/{occurrence_id}" grafeas_client.delete_occurrence(name=parent) + + # [END containeranalysis_delete_occurrence] @@ -131,6 +135,8 @@ def get_note(note_id, project_id): note_name = f"projects/{project_id}/notes/{note_id}" response = grafeas_client.get_note(name=note_name) return response + + # [END containeranalysis_get_note] @@ -146,6 +152,8 @@ def get_occurrence(occurrence_id, project_id): grafeas_client = client.get_grafeas_client() parent = f"projects/{project_id}/occurrences/{occurrence_id}" return grafeas_client.get_occurrence(name=parent) + + # [END containeranalysis_get_occurrence] @@ -163,10 +171,11 @@ def get_discovery_info(resource_url, project_id): client = containeranalysis_v1.ContainerAnalysisClient() grafeas_client = client.get_grafeas_client() project_name = f"projects/{project_id}" - response = grafeas_client.list_occurrences(parent=project_name, - filter_=filter_str) + response = grafeas_client.list_occurrences(parent=project_name, filter_=filter_str) for occ in response: print(occ) + + # [END containeranalysis_discovery_info] @@ -190,6 +199,8 @@ def get_occurrences_for_note(note_id, project_id): # in this sample, we will simply count each one count += 1 return count + + # [END containeranalysis_occurrences_for_note] @@ -207,14 +218,15 @@ def get_occurrences_for_image(resource_url, project_id): grafeas_client = client.get_grafeas_client() project_name = f"projects/{project_id}" - response = grafeas_client.list_occurrences(parent=project_name, - filter=filter_str) + response = grafeas_client.list_occurrences(parent=project_name, filter=filter_str) count = 0 for o in response: # do something with the retrieved occurrence # in this sample, we will simply count each one count += 1 return count + + # [END containeranalysis_occurrences_for_image] @@ -226,6 +238,7 @@ def pubsub(subscription_id, timeout_seconds, project_id): # project_id = 'my-gcp-project' import time + from google.cloud.pubsub import SubscriberClient client = SubscriberClient() @@ -243,6 +256,7 @@ def pubsub(subscription_id, timeout_seconds, project_id): class MessageReceiver: """Custom class to handle incoming Pub/Sub messages.""" + def __init__(self): # initialize counter to 0 on initialization self.msg_count = 0 @@ -250,7 +264,7 @@ def __init__(self): def pubsub_callback(self, message): # every time a pubsub message comes in, print it and count it self.msg_count += 1 - print('Message {}: {}'.format(self.msg_count, message.data)) + print("Message {}: {}".format(self.msg_count, message.data)) message.ack() @@ -263,7 +277,7 @@ def create_occurrence_subscription(subscription_id, project_id): from google.api_core.exceptions import AlreadyExists from google.cloud.pubsub import SubscriberClient - topic_id = 'container-analysis-occurrences-v1' + topic_id = "container-analysis-occurrences-v1" client = SubscriberClient() topic_name = f"projects/{project_id}/topics/{topic_id}" subscription_name = client.subscription_path(project_id, subscription_id) @@ -276,6 +290,8 @@ def create_occurrence_subscription(subscription_id, project_id): else: success = False return success + + # [END containeranalysis_pubsub] @@ -288,8 +304,9 @@ def poll_discovery_finished(resource_url, timeout_seconds, project_id): # project_id = 'my-gcp-project' import time - from grafeas.grafeas_v1 import DiscoveryOccurrence + from google.cloud.devtools import containeranalysis_v1 + from grafeas.grafeas_v1 import DiscoveryOccurrence deadline = time.time() + timeout_seconds @@ -302,12 +319,13 @@ def poll_discovery_finished(resource_url, timeout_seconds, project_id): time.sleep(1) filter_str = 'resourceUrl="{}" \ AND noteProjectId="goog-analysis" \ - AND noteId="PACKAGE_VULNERABILITY"'.format(resource_url) + AND noteId="PACKAGE_VULNERABILITY"'.format( + resource_url + ) # [END containeranalysis_poll_discovery_occurrence_finished] # The above filter isn't testable, since it looks for occurrences in a # locked down project fall back to a more permissive filter for testing - filter_str = 'kind="DISCOVERY" AND resourceUrl="{}"'\ - .format(resource_url) + filter_str = 'kind="DISCOVERY" AND resourceUrl="{}"'.format(resource_url) # [START containeranalysis_poll_discovery_occurrence_finished] result = grafeas_client.list_occurrences(parent=project_name, filter=filter_str) # only one occurrence should ever be returned by ListOccurrences @@ -315,24 +333,28 @@ def poll_discovery_finished(resource_url, timeout_seconds, project_id): for item in result: discovery_occurrence = item if time.time() > deadline: - raise RuntimeError('timeout while retrieving discovery occurrence') + raise RuntimeError("timeout while retrieving discovery occurrence") status = DiscoveryOccurrence.AnalysisStatus.PENDING - while status != DiscoveryOccurrence.AnalysisStatus.FINISHED_UNSUPPORTED \ - and status != DiscoveryOccurrence.AnalysisStatus.FINISHED_FAILED \ - and status != DiscoveryOccurrence.AnalysisStatus.FINISHED_SUCCESS: + while ( + status != DiscoveryOccurrence.AnalysisStatus.FINISHED_UNSUPPORTED + and status != DiscoveryOccurrence.AnalysisStatus.FINISHED_FAILED + and status != DiscoveryOccurrence.AnalysisStatus.FINISHED_SUCCESS + ): time.sleep(1) updated = grafeas_client.get_occurrence(name=discovery_occurrence.name) status = updated.discovery.analysis_status if time.time() > deadline: - raise RuntimeError('timeout while waiting for terminal state') + raise RuntimeError("timeout while waiting for terminal state") return discovery_occurrence + + # [END containeranalysis_poll_discovery_occurrence_finished] # [START containeranalysis_vulnerability_occurrences_for_image] def find_vulnerabilities_for_image(resource_url, project_id): - """"Retrieves all vulnerability occurrences associated with a resource.""" + """ "Retrieves all vulnerability occurrences associated with a resource.""" # resource_url = 'https://gcr.io/my-project/my-image@sha256:123' # project_id = 'my-gcp-project' @@ -342,9 +364,10 @@ def find_vulnerabilities_for_image(resource_url, project_id): grafeas_client = client.get_grafeas_client() project_name = f"projects/{project_id}" - filter_str = 'kind="VULNERABILITY" AND resourceUrl="{}"'\ - .format(resource_url) + filter_str = 'kind="VULNERABILITY" AND resourceUrl="{}"'.format(resource_url) return list(grafeas_client.list_occurrences(parent=project_name, filter=filter_str)) + + # [END containeranalysis_vulnerability_occurrences_for_image] @@ -355,19 +378,25 @@ def find_high_severity_vulnerabilities_for_image(resource_url, project_id): # resource_url = 'https://gcr.io/my-project/my-image@sha256:123' # project_id = 'my-gcp-project' - from grafeas.grafeas_v1 import Severity from google.cloud.devtools import containeranalysis_v1 + from grafeas.grafeas_v1 import Severity client = containeranalysis_v1.ContainerAnalysisClient() grafeas_client = client.get_grafeas_client() project_name = f"projects/{project_id}" - filter_str = 'kind="VULNERABILITY" AND resourceUrl="{}"'\ - .format(resource_url) - vulnerabilities = grafeas_client.list_occurrences(parent=project_name, filter=filter_str) + filter_str = 'kind="VULNERABILITY" AND resourceUrl="{}"'.format(resource_url) + vulnerabilities = grafeas_client.list_occurrences( + parent=project_name, filter=filter_str + ) filtered_list = [] for v in vulnerabilities: - if v.vulnerability.effective_severity == Severity.HIGH or v.vulnerability.effective_severity == Severity.CRITICAL: + if ( + v.vulnerability.effective_severity == Severity.HIGH + or v.vulnerability.effective_severity == Severity.CRITICAL + ): filtered_list.append(v) return filtered_list + + # [END containeranalysis_filter_vulnerability_occurrences] diff --git a/samples/snippets/samples_test.py b/samples/snippets/samples_test.py index dd9bce6..8ed97fa 100644 --- a/samples/snippets/samples_test.py +++ b/samples/snippets/samples_test.py @@ -19,27 +19,22 @@ import time import uuid -from google.api_core.exceptions import AlreadyExists -from google.api_core.exceptions import InvalidArgument -from google.api_core.exceptions import NotFound +from google.api_core.exceptions import AlreadyExists, InvalidArgument, NotFound from google.cloud.devtools import containeranalysis_v1 from google.cloud.pubsub import PublisherClient, SubscriberClient - -from grafeas.grafeas_v1 import DiscoveryOccurrence -from grafeas.grafeas_v1 import NoteKind -from grafeas.grafeas_v1 import Severity -from grafeas.grafeas_v1 import Version +from grafeas.grafeas_v1 import DiscoveryOccurrence, NoteKind, Severity, Version import pytest import samples -PROJECT_ID = environ['GOOGLE_CLOUD_PROJECT'] +PROJECT_ID = environ["GOOGLE_CLOUD_PROJECT"] SLEEP_TIME = 1 TRY_LIMIT = 20 class MessageReceiver: """Custom class to handle incoming Pub/Sub messages.""" + def __init__(self, expected_msg_nums, done_event): # initialize counter to 0 on initialization self.msg_count = 0 @@ -49,22 +44,21 @@ def __init__(self, expected_msg_nums, done_event): def pubsub_callback(self, message): # every time a pubsub message comes in, print it and count it self.msg_count += 1 - print('Message {}: {}'.format(self.msg_count, message.data)) + print("Message {}: {}".format(self.msg_count, message.data)) message.ack() - if (self.msg_count == self.expected_msg_nums): + if self.msg_count == self.expected_msg_nums: self.done_event.set() class TestContainerAnalysisSamples: - def setup_method(self, test_method): - print('SETUP {}'.format(test_method.__name__)) - self.note_id = 'note-{}'.format(uuid.uuid4()) - self.image_url = '{}.{}'.format(uuid.uuid4(), test_method.__name__) + print("SETUP {}".format(test_method.__name__)) + self.note_id = "note-{}".format(uuid.uuid4()) + self.image_url = "{}.{}".format(uuid.uuid4(), test_method.__name__) self.note_obj = samples.create_note(self.note_id, PROJECT_ID) def teardown_method(self, test_method): - print('TEAR DOWN {}'.format(test_method.__name__)) + print("TEAR DOWN {}".format(test_method.__name__)) try: samples.delete_note(self.note_id, PROJECT_ID) except NotFound: @@ -82,23 +76,21 @@ def test_delete_note(self): pass else: # didn't raise exception we expected - assert (False) + assert False def test_create_occurrence(self): - created = samples.create_occurrence(self.image_url, - self.note_id, - PROJECT_ID, - PROJECT_ID) + created = samples.create_occurrence( + self.image_url, self.note_id, PROJECT_ID, PROJECT_ID + ) retrieved = samples.get_occurrence(basename(created.name), PROJECT_ID) assert created.name == retrieved.name # clean up samples.delete_occurrence(basename(created.name), PROJECT_ID) def test_delete_occurrence(self): - created = samples.create_occurrence(self.image_url, - self.note_id, - PROJECT_ID, - PROJECT_ID) + created = samples.create_occurrence( + self.image_url, self.note_id, PROJECT_ID, PROJECT_ID + ) samples.delete_occurrence(basename(created.name), PROJECT_ID) try: samples.get_occurrence(basename(created.name), PROJECT_ID) @@ -109,18 +101,15 @@ def test_delete_occurrence(self): assert False def test_occurrences_for_image(self): - orig_count = samples.get_occurrences_for_image(self.image_url, - PROJECT_ID) - occ = samples.create_occurrence(self.image_url, - self.note_id, - PROJECT_ID, - PROJECT_ID) + orig_count = samples.get_occurrences_for_image(self.image_url, PROJECT_ID) + occ = samples.create_occurrence( + self.image_url, self.note_id, PROJECT_ID, PROJECT_ID + ) new_count = 0 tries = 0 while new_count != 1 and tries < TRY_LIMIT: tries += 1 - new_count = samples.get_occurrences_for_image(self.image_url, - PROJECT_ID) + new_count = samples.get_occurrences_for_image(self.image_url, PROJECT_ID) time.sleep(SLEEP_TIME) assert new_count == 1 assert orig_count == 0 @@ -128,18 +117,15 @@ def test_occurrences_for_image(self): samples.delete_occurrence(basename(occ.name), PROJECT_ID) def test_occurrences_for_note(self): - orig_count = samples.get_occurrences_for_note(self.note_id, - PROJECT_ID) - occ = samples.create_occurrence(self.image_url, - self.note_id, - PROJECT_ID, - PROJECT_ID) + orig_count = samples.get_occurrences_for_note(self.note_id, PROJECT_ID) + occ = samples.create_occurrence( + self.image_url, self.note_id, PROJECT_ID, PROJECT_ID + ) new_count = 0 tries = 0 while new_count != 1 and tries < TRY_LIMIT: tries += 1 - new_count = samples.get_occurrences_for_note(self.note_id, - PROJECT_ID) + new_count = samples.get_occurrences_for_note(self.note_id, PROJECT_ID) time.sleep(SLEEP_TIME) assert new_count == 1 assert orig_count == 0 @@ -151,16 +137,15 @@ def test_pubsub(self): # create topic if needed client = SubscriberClient() try: - topic_id = 'container-analysis-occurrences-v1' + topic_id = "container-analysis-occurrences-v1" topic_name = {"name": f"projects/{PROJECT_ID}/topics/{topic_id}"} publisher = PublisherClient() publisher.create_topic(topic_name) except AlreadyExists: pass - subscription_id = 'container-analysis-test-{}'.format(uuid.uuid4()) - subscription_name = client.subscription_path(PROJECT_ID, - subscription_id) + subscription_id = "container-analysis-test-{}".format(uuid.uuid4()) + subscription_name = client.subscription_path(PROJECT_ID, subscription_id) samples.create_occurrence_subscription(subscription_id, PROJECT_ID) # I can not make it pass with multiple messages. My guess is @@ -173,7 +158,8 @@ def test_pubsub(self): for i in range(message_count): occ = samples.create_occurrence( - self.image_url, self.note_id, PROJECT_ID, PROJECT_ID) + self.image_url, self.note_id, PROJECT_ID, PROJECT_ID + ) time.sleep(SLEEP_TIME) samples.delete_occurrence(basename(occ.name), PROJECT_ID) time.sleep(SLEEP_TIME) @@ -181,7 +167,7 @@ def test_pubsub(self): # to 180 seconds. # See also: python-docs-samples/issues/2894 job_done.wait(timeout=180) - print('done. msg_count = {}'.format(receiver.msg_count)) + print("done. msg_count = {}".format(receiver.msg_count)) assert message_count <= receiver.msg_count finally: # clean up @@ -200,27 +186,23 @@ def test_poll_discovery_occurrence_fails(self): @pytest.mark.flaky(max_runs=3, min_passes=1) def test_poll_discovery_occurrence(self): # create discovery occurrence - note_id = 'discovery-note-{}'.format(uuid.uuid4()) + note_id = "discovery-note-{}".format(uuid.uuid4()) client = containeranalysis_v1.ContainerAnalysisClient() grafeas_client = client.get_grafeas_client() - note = { - 'discovery': { - 'analysis_kind': NoteKind.DISCOVERY - } - } - grafeas_client.\ - create_note(parent=f"projects/{PROJECT_ID}", note_id=note_id, note=note) + note = {"discovery": {"analysis_kind": NoteKind.DISCOVERY}} + grafeas_client.create_note( + parent=f"projects/{PROJECT_ID}", note_id=note_id, note=note + ) occurrence = { - 'note_name': f"projects/{PROJECT_ID}/notes/{note_id}", - 'resource_uri': self.image_url, - 'discovery': { - 'analysis_status': DiscoveryOccurrence.AnalysisStatus - .FINISHED_SUCCESS - } + "note_name": f"projects/{PROJECT_ID}/notes/{note_id}", + "resource_uri": self.image_url, + "discovery": { + "analysis_status": DiscoveryOccurrence.AnalysisStatus.FINISHED_SUCCESS + }, } - created = grafeas_client.\ - create_occurrence(parent=f"projects/{PROJECT_ID}", - occurrence=occurrence) + created = grafeas_client.create_occurrence( + parent=f"projects/{PROJECT_ID}", occurrence=occurrence + ) disc = samples.poll_discovery_finished(self.image_url, 10, PROJECT_ID) status = disc.discovery.analysis_status @@ -232,20 +214,19 @@ def test_poll_discovery_occurrence(self): samples.delete_note(note_id, PROJECT_ID) def test_find_vulnerabilities_for_image(self): - occ_list = samples.find_vulnerabilities_for_image(self.image_url, - PROJECT_ID) + occ_list = samples.find_vulnerabilities_for_image(self.image_url, PROJECT_ID) assert len(occ_list) == 0 - created = samples.create_occurrence(self.image_url, - self.note_id, - PROJECT_ID, - PROJECT_ID) + created = samples.create_occurrence( + self.image_url, self.note_id, PROJECT_ID, PROJECT_ID + ) tries = 0 count = 0 while count != 1 and tries < TRY_LIMIT: tries += 1 - occ_list = samples.find_vulnerabilities_for_image(self.image_url, - PROJECT_ID) + occ_list = samples.find_vulnerabilities_for_image( + self.image_url, PROJECT_ID + ) count = len(occ_list) time.sleep(SLEEP_TIME) assert len(occ_list) == 1 @@ -253,62 +234,56 @@ def test_find_vulnerabilities_for_image(self): def test_find_high_severity_vulnerabilities(self): occ_list = samples.find_high_severity_vulnerabilities_for_image( - self.image_url, - PROJECT_ID) + self.image_url, PROJECT_ID + ) assert len(occ_list) == 0 # create new high severity vulnerability - note_id = 'discovery-note-{}'.format(uuid.uuid4()) + note_id = "discovery-note-{}".format(uuid.uuid4()) client = containeranalysis_v1.ContainerAnalysisClient() grafeas_client = client.get_grafeas_client() note = { - 'vulnerability': { - 'severity': Severity.CRITICAL, - 'details': [ + "vulnerability": { + "severity": Severity.CRITICAL, + "details": [ { - 'affected_cpe_uri': 'your-uri-here', - 'affected_package': 'your-package-here', - 'affected_version_start': { - 'kind': Version.VersionKind.MINIMUM - }, - 'fixed_version': { - 'kind': Version.VersionKind.MAXIMUM - } + "affected_cpe_uri": "your-uri-here", + "affected_package": "your-package-here", + "affected_version_start": {"kind": Version.VersionKind.MINIMUM}, + "fixed_version": {"kind": Version.VersionKind.MAXIMUM}, } - ] + ], } } - grafeas_client.\ - create_note(parent=f"projects/{PROJECT_ID}", note_id=note_id, note=note) + grafeas_client.create_note( + parent=f"projects/{PROJECT_ID}", note_id=note_id, note=note + ) occurrence = { - 'note_name': f"projects/{PROJECT_ID}/notes/{note_id}", - 'resource_uri': self.image_url, - 'vulnerability': { - 'effective_severity': Severity.CRITICAL, - 'package_issue': [ + "note_name": f"projects/{PROJECT_ID}/notes/{note_id}", + "resource_uri": self.image_url, + "vulnerability": { + "effective_severity": Severity.CRITICAL, + "package_issue": [ { - 'affected_cpe_uri': 'your-uri-here', - 'affected_package': 'your-package-here', - 'affected_version': { - 'kind': Version.VersionKind.MINIMUM - }, - 'fixed_version': { - 'kind': Version.VersionKind.MAXIMUM - } + "affected_cpe_uri": "your-uri-here", + "affected_package": "your-package-here", + "affected_version": {"kind": Version.VersionKind.MINIMUM}, + "fixed_version": {"kind": Version.VersionKind.MAXIMUM}, } - ] - } + ], + }, } - created = grafeas_client.\ - create_occurrence(parent=f"projects/{PROJECT_ID}", - occurrence=occurrence) + created = grafeas_client.create_occurrence( + parent=f"projects/{PROJECT_ID}", occurrence=occurrence + ) # query again tries = 0 count = 0 while count != 1 and tries < TRY_LIMIT: tries += 1 - occ_list = samples.find_vulnerabilities_for_image(self.image_url, - PROJECT_ID) + occ_list = samples.find_vulnerabilities_for_image( + self.image_url, PROJECT_ID + ) count = len(occ_list) time.sleep(SLEEP_TIME) assert len(occ_list) == 1 diff --git a/setup.py b/setup.py index 69f0296..16a187f 100644 --- a/setup.py +++ b/setup.py @@ -1,38 +1,52 @@ # -*- coding: utf-8 -*- -# -# Copyright 2019 Google LLC +# Copyright 2022 Google LLC # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # -# https://www.apache.org/licenses/LICENSE-2.0 +# http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. - +# import io import os -import setuptools +import setuptools # type: ignore +package_root = os.path.abspath(os.path.dirname(__file__)) name = "google-cloud-containeranalysis" -description = "Container Analysis API API client library" -version = "2.9.3" -release_status = "Development Status :: 5 - Production/Stable" + + +description = "Google Cloud Devtools Containeranalysis API client library" + +version = {} +with open( + os.path.join( + package_root, "google/cloud/devtools/containeranalysis/gapic_version.py" + ) +) as fp: + exec(fp.read(), version) +version = version["__version__"] + +if version[0] == "0": + release_status = "Development Status :: 4 - Beta" +else: + release_status = "Development Status :: 5 - Production/Stable" + dependencies = [ - "google-api-core[grpc] >= 1.32.0, <3.0.0dev,!=2.0.*,!=2.1.*,!=2.2.*,!=2.3.*,!=2.4.*,!=2.5.*,!=2.6.*,!=2.7.*", - "grpc-google-iam-v1 >= 0.12.4, <1.0.0dev", + "google-api-core[grpc] >= 1.33.2, <3.0.0dev,!=2.0.*,!=2.1.*,!=2.2.*,!=2.3.*,!=2.4.*,!=2.5.*,!=2.6.*,!=2.7.*", "proto-plus >= 1.22.0, <2.0.0dev", "protobuf>=3.19.5,<5.0.0dev,!=3.20.0,!=3.20.1,!=4.21.0,!=4.21.1,!=4.21.2,!=4.21.3,!=4.21.4,!=4.21.5", + "grpc-google-iam-v1 >= 0.12.4, < 1.0.0dev", "grafeas >=1.4.1, <2.0dev", ] -extras = {"libcst": "libcst >= 0.2.5"} - +url = "https://github.com/googleapis/python-containeranalysis" package_root = os.path.abspath(os.path.dirname(__file__)) @@ -60,7 +74,7 @@ author="Google LLC", author_email="googleapis-packages@google.com", license="Apache 2.0", - url="https://github.com/googleapis/python-containeranalysis", + url=url, classifiers=[ release_status, "Intended Audience :: Developers", @@ -76,11 +90,9 @@ ], platforms="Posix; MacOS X; Windows", packages=packages, + python_requires=">=3.7", namespace_packages=namespaces, install_requires=dependencies, - extras_require=extras, - python_requires=">=3.7", - scripts=["scripts/fixup_containeranalysis_v1_keywords.py"], include_package_data=True, zip_safe=False, ) diff --git a/testing/constraints-3.10.txt b/testing/constraints-3.10.txt index e69de29..ad3f0fa 100644 --- a/testing/constraints-3.10.txt +++ b/testing/constraints-3.10.txt @@ -0,0 +1,7 @@ +# -*- coding: utf-8 -*- +# This constraints file is required for unit tests. +# List all library dependencies and extras in this file. +google-api-core +proto-plus +protobuf +grpc-google-iam-v1 diff --git a/testing/constraints-3.11.txt b/testing/constraints-3.11.txt index e69de29..ad3f0fa 100644 --- a/testing/constraints-3.11.txt +++ b/testing/constraints-3.11.txt @@ -0,0 +1,7 @@ +# -*- coding: utf-8 -*- +# This constraints file is required for unit tests. +# List all library dependencies and extras in this file. +google-api-core +proto-plus +protobuf +grpc-google-iam-v1 diff --git a/testing/constraints-3.7.txt b/testing/constraints-3.7.txt index f2ff391..8995c3a 100644 --- a/testing/constraints-3.7.txt +++ b/testing/constraints-3.7.txt @@ -1,14 +1,11 @@ # This constraints file is used to check that lower bounds # are correct in setup.py -# List *all* library dependencies and extras in this file. +# List all library dependencies and extras in this file. # Pin the version to the lower bound. -# -# e.g., if setup.py has "foo >= 1.14.0, < 2.0.0dev", -# Then this file should have foo==1.14.0 -google-api-core==1.32.0 -grpc-google-iam-v1==0.12.4 +# e.g., if setup.py has "google-cloud-foo >= 1.14.0, < 2.0.0dev", +# Then this file should have google-cloud-foo==1.14.0 +google-api-core==1.33.2 proto-plus==1.22.0 -libcst==0.2.5 -grafeas==1.4.1 protobuf==3.19.5 - +grpc-google-iam-v1==0.12.4 +grafeas==1.4.1 diff --git a/testing/constraints-3.8.txt b/testing/constraints-3.8.txt index e69de29..ad3f0fa 100644 --- a/testing/constraints-3.8.txt +++ b/testing/constraints-3.8.txt @@ -0,0 +1,7 @@ +# -*- coding: utf-8 -*- +# This constraints file is required for unit tests. +# List all library dependencies and extras in this file. +google-api-core +proto-plus +protobuf +grpc-google-iam-v1 diff --git a/testing/constraints-3.9.txt b/testing/constraints-3.9.txt index e69de29..ad3f0fa 100644 --- a/testing/constraints-3.9.txt +++ b/testing/constraints-3.9.txt @@ -0,0 +1,7 @@ +# -*- coding: utf-8 -*- +# This constraints file is required for unit tests. +# List all library dependencies and extras in this file. +google-api-core +proto-plus +protobuf +grpc-google-iam-v1 diff --git a/tests/unit/gapic/containeranalysis_v1/test_container_analysis.py b/tests/unit/gapic/containeranalysis_v1/test_container_analysis.py index b977996..3a4bf81 100644 --- a/tests/unit/gapic/containeranalysis_v1/test_container_analysis.py +++ b/tests/unit/gapic/containeranalysis_v1/test_container_analysis.py @@ -22,38 +22,32 @@ except ImportError: # pragma: NO COVER import mock -import grpc -from grpc.experimental import aio import math -import pytest -from proto.marshal.rules.dates import DurationRule, TimestampRule -from proto.marshal.rules import wrappers +from google.api_core import gapic_v1, grpc_helpers, grpc_helpers_async, path_template from google.api_core import client_options from google.api_core import exceptions as core_exceptions -from google.api_core import gapic_v1 -from google.api_core import grpc_helpers -from google.api_core import grpc_helpers_async -from google.api_core import path_template +import google.auth from google.auth import credentials as ga_credentials from google.auth.exceptions import MutualTLSChannelError -from google.cloud.devtools.containeranalysis_v1.services.container_analysis import ( - ContainerAnalysisAsyncClient, -) -from google.cloud.devtools.containeranalysis_v1.services.container_analysis import ( - ContainerAnalysisClient, -) -from google.cloud.devtools.containeranalysis_v1.services.container_analysis import ( - transports, -) -from google.cloud.devtools.containeranalysis_v1.types import containeranalysis from google.iam.v1 import iam_policy_pb2 # type: ignore from google.iam.v1 import options_pb2 # type: ignore from google.iam.v1 import policy_pb2 # type: ignore from google.oauth2 import service_account from google.protobuf import field_mask_pb2 # type: ignore from google.type import expr_pb2 # type: ignore -import google.auth +import grpc +from grpc.experimental import aio +from proto.marshal.rules import wrappers +from proto.marshal.rules.dates import DurationRule, TimestampRule +import pytest + +from google.cloud.devtools.containeranalysis_v1.services.container_analysis import ( + ContainerAnalysisAsyncClient, + ContainerAnalysisClient, + transports, +) +from google.cloud.devtools.containeranalysis_v1.types import containeranalysis def client_cert_source_callback(): diff --git a/tests/unit/test_get_grafeas_client.py b/tests/unit/test_get_grafeas_client.py index 66a1652..2a96216 100644 --- a/tests/unit/test_get_grafeas_client.py +++ b/tests/unit/test_get_grafeas_client.py @@ -1,7 +1,5 @@ from google.cloud.devtools.containeranalysis_v1.services.container_analysis import ( ContainerAnalysisAsyncClient, -) -from google.cloud.devtools.containeranalysis_v1.services.container_analysis import ( ContainerAnalysisClient, ) From 17a5190c53ef0523e764e2ce00546517cab84908 Mon Sep 17 00:00:00 2001 From: "gcf-owl-bot[bot]" <78513119+gcf-owl-bot[bot]@users.noreply.github.com> Date: Sat, 26 Nov 2022 19:11:54 -0500 Subject: [PATCH 09/13] chore(python): drop flake8-import-order in samples noxfile (#347) Source-Link: https://github.com/googleapis/synthtool/commit/6ed3a831cb9ff69ef8a504c353e098ec0192ad93 Post-Processor: gcr.io/cloud-devrel-public-resources/owlbot-python:latest@sha256:3abfa0f1886adaf0b83f07cb117b24a639ea1cb9cffe56d43280b977033563eb Co-authored-by: Owl Bot --- .github/.OwlBot.lock.yaml | 2 +- samples/snippets/noxfile.py | 26 +++----------------------- 2 files changed, 4 insertions(+), 24 deletions(-) diff --git a/.github/.OwlBot.lock.yaml b/.github/.OwlBot.lock.yaml index 3f1ccc0..bb21147 100644 --- a/.github/.OwlBot.lock.yaml +++ b/.github/.OwlBot.lock.yaml @@ -13,4 +13,4 @@ # limitations under the License. docker: image: gcr.io/cloud-devrel-public-resources/owlbot-python:latest - digest: sha256:e6cbd61f1838d9ff6a31436dfc13717f372a7482a82fc1863ca954ec47bff8c8 + digest: sha256:3abfa0f1886adaf0b83f07cb117b24a639ea1cb9cffe56d43280b977033563eb diff --git a/samples/snippets/noxfile.py b/samples/snippets/noxfile.py index c171513..0577084 100644 --- a/samples/snippets/noxfile.py +++ b/samples/snippets/noxfile.py @@ -18,7 +18,7 @@ import os from pathlib import Path import sys -from typing import Callable, Dict, List, Optional +from typing import Callable, Dict, Optional import nox @@ -108,22 +108,6 @@ def get_pytest_env_vars() -> Dict[str, str]: # -def _determine_local_import_names(start_dir: str) -> List[str]: - """Determines all import names that should be considered "local". - - This is used when running the linter to insure that import order is - properly checked. - """ - file_ext_pairs = [os.path.splitext(path) for path in os.listdir(start_dir)] - return [ - basename - for basename, extension in file_ext_pairs - if extension == ".py" - or os.path.isdir(os.path.join(start_dir, basename)) - and basename not in ("__pycache__") - ] - - # Linting with flake8. # # We ignore the following rules: @@ -138,7 +122,6 @@ def _determine_local_import_names(start_dir: str) -> List[str]: "--show-source", "--builtin=gettext", "--max-complexity=20", - "--import-order-style=google", "--exclude=.nox,.cache,env,lib,generated_pb2,*_pb2.py,*_pb2_grpc.py", "--ignore=E121,E123,E126,E203,E226,E24,E266,E501,E704,W503,W504,I202", "--max-line-length=88", @@ -148,14 +131,11 @@ def _determine_local_import_names(start_dir: str) -> List[str]: @nox.session def lint(session: nox.sessions.Session) -> None: if not TEST_CONFIG["enforce_type_hints"]: - session.install("flake8", "flake8-import-order") + session.install("flake8") else: - session.install("flake8", "flake8-import-order", "flake8-annotations") + session.install("flake8", "flake8-annotations") - local_names = _determine_local_import_names(".") args = FLAKE8_COMMON_ARGS + [ - "--application-import-names", - ",".join(local_names), ".", ] session.run("flake8", *args) From 8a0a4b635de8992b9f5984a337cabaf92017400a Mon Sep 17 00:00:00 2001 From: "gcf-owl-bot[bot]" <78513119+gcf-owl-bot[bot]@users.noreply.github.com> Date: Wed, 7 Dec 2022 10:33:30 -0500 Subject: [PATCH 10/13] fix(deps): Require google-api-core >=1.34.0, >=2.11.0 (#349) MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit * fix(deps): Require google-api-core >=1.34.0, >=2.11.0 fix: Drop usage of pkg_resources fix: Fix timeout default values docs(samples): Snippetgen should call await on the operation coroutine before calling result PiperOrigin-RevId: 493260409 Source-Link: https://github.com/googleapis/googleapis/commit/fea43879f83a8d0dacc9353b3f75f8f46d37162f Source-Link: https://github.com/googleapis/googleapis-gen/commit/387b7344c7529ee44be84e613b19a820508c612b Copy-Tag: eyJwIjoiLmdpdGh1Yi8uT3dsQm90LnlhbWwiLCJoIjoiMzg3YjczNDRjNzUyOWVlNDRiZTg0ZTYxM2IxOWE4MjA1MDhjNjEyYiJ9 * πŸ¦‰ Updates from OwlBot post-processor See https://github.com/googleapis/repo-automation-bots/blob/main/packages/owl-bot/README.md * add gapic_version.py Co-authored-by: Owl Bot Co-authored-by: Anthonios Partheniou --- .coveragerc | 5 ----- .../containeranalysis_v1/gapic_version.py | 16 ++++++++++++++ .../container_analysis/async_client.py | 22 ++++++++----------- .../services/container_analysis/client.py | 22 ++++++++----------- .../container_analysis/transports/base.py | 13 ++++------- release-please-config.json | 1 + 6 files changed, 39 insertions(+), 40 deletions(-) create mode 100644 google/cloud/devtools/containeranalysis_v1/gapic_version.py diff --git a/.coveragerc b/.coveragerc index f04431d..65216b3 100644 --- a/.coveragerc +++ b/.coveragerc @@ -10,8 +10,3 @@ exclude_lines = pragma: NO COVER # Ignore debug-only repr def __repr__ - # Ignore pkg_resources exceptions. - # This is added at the module level as a safeguard for if someone - # generates the code and tries to run it without pip installing. This - # makes it virtually impossible to test properly. - except pkg_resources.DistributionNotFound diff --git a/google/cloud/devtools/containeranalysis_v1/gapic_version.py b/google/cloud/devtools/containeranalysis_v1/gapic_version.py new file mode 100644 index 0000000..085af7e --- /dev/null +++ b/google/cloud/devtools/containeranalysis_v1/gapic_version.py @@ -0,0 +1,16 @@ +# -*- coding: utf-8 -*- +# Copyright 2022 Google LLC +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +# +__version__ = "2.9.3" # {x-release-please-version} diff --git a/google/cloud/devtools/containeranalysis_v1/services/container_analysis/async_client.py b/google/cloud/devtools/containeranalysis_v1/services/container_analysis/async_client.py index 69e62d5..96903fd 100644 --- a/google/cloud/devtools/containeranalysis_v1/services/container_analysis/async_client.py +++ b/google/cloud/devtools/containeranalysis_v1/services/container_analysis/async_client.py @@ -34,7 +34,8 @@ from google.api_core.client_options import ClientOptions from google.auth import credentials as ga_credentials # type: ignore from google.oauth2 import service_account # type: ignore -import pkg_resources + +from google.cloud.devtools.containeranalysis_v1 import gapic_version as package_version try: OptionalRetry = Union[retries.Retry, gapic_v1.method._MethodDefault] @@ -250,7 +251,7 @@ async def set_iam_policy( *, resource: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> policy_pb2.Policy: r"""Sets the access control policy on the specified note or @@ -424,7 +425,7 @@ async def get_iam_policy( *, resource: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> policy_pb2.Policy: r"""Gets the access control policy for a note or an occurrence @@ -599,7 +600,7 @@ async def test_iam_permissions( resource: Optional[str] = None, permissions: Optional[MutableSequence[str]] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> iam_policy_pb2.TestIamPermissionsResponse: r"""Returns the permissions that a caller has on the specified note @@ -725,7 +726,7 @@ async def get_vulnerability_occurrences_summary( parent: Optional[str] = None, filter: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> containeranalysis.VulnerabilityOccurrencesSummary: r"""Gets a summary of the number and severity of @@ -837,14 +838,9 @@ async def __aexit__(self, exc_type, exc, tb): await self.transport.close() -try: - DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( - gapic_version=pkg_resources.get_distribution( - "google-cloud-containeranalysis", - ).version, - ) -except pkg_resources.DistributionNotFound: - DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo() +DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( + gapic_version=package_version.__version__ +) __all__ = ("ContainerAnalysisAsyncClient",) diff --git a/google/cloud/devtools/containeranalysis_v1/services/container_analysis/client.py b/google/cloud/devtools/containeranalysis_v1/services/container_analysis/client.py index 2a1073a..36731a0 100644 --- a/google/cloud/devtools/containeranalysis_v1/services/container_analysis/client.py +++ b/google/cloud/devtools/containeranalysis_v1/services/container_analysis/client.py @@ -38,7 +38,8 @@ from google.auth.transport import mtls # type: ignore from google.auth.transport.grpc import SslCredentials # type: ignore from google.oauth2 import service_account # type: ignore -import pkg_resources + +from google.cloud.devtools.containeranalysis_v1 import gapic_version as package_version try: OptionalRetry = Union[retries.Retry, gapic_v1.method._MethodDefault] @@ -457,7 +458,7 @@ def set_iam_policy( *, resource: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> policy_pb2.Policy: r"""Sets the access control policy on the specified note or @@ -628,7 +629,7 @@ def get_iam_policy( *, resource: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> policy_pb2.Policy: r"""Gets the access control policy for a note or an occurrence @@ -800,7 +801,7 @@ def test_iam_permissions( resource: Optional[str] = None, permissions: Optional[MutableSequence[str]] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> iam_policy_pb2.TestIamPermissionsResponse: r"""Returns the permissions that a caller has on the specified note @@ -924,7 +925,7 @@ def get_vulnerability_occurrences_summary( parent: Optional[str] = None, filter: Optional[str] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, - timeout: Optional[float] = None, + timeout: Union[float, object] = gapic_v1.method.DEFAULT, metadata: Sequence[Tuple[str, str]] = (), ) -> containeranalysis.VulnerabilityOccurrencesSummary: r"""Gets a summary of the number and severity of @@ -1049,14 +1050,9 @@ def __exit__(self, type, value, traceback): self.transport.close() -try: - DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( - gapic_version=pkg_resources.get_distribution( - "google-cloud-containeranalysis", - ).version, - ) -except pkg_resources.DistributionNotFound: - DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo() +DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( + gapic_version=package_version.__version__ +) __all__ = ("ContainerAnalysisClient",) diff --git a/google/cloud/devtools/containeranalysis_v1/services/container_analysis/transports/base.py b/google/cloud/devtools/containeranalysis_v1/services/container_analysis/transports/base.py index 37d7e1c..1438609 100644 --- a/google/cloud/devtools/containeranalysis_v1/services/container_analysis/transports/base.py +++ b/google/cloud/devtools/containeranalysis_v1/services/container_analysis/transports/base.py @@ -25,18 +25,13 @@ from google.iam.v1 import iam_policy_pb2 # type: ignore from google.iam.v1 import policy_pb2 # type: ignore from google.oauth2 import service_account # type: ignore -import pkg_resources +from google.cloud.devtools.containeranalysis_v1 import gapic_version as package_version from google.cloud.devtools.containeranalysis_v1.types import containeranalysis -try: - DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( - gapic_version=pkg_resources.get_distribution( - "google-cloud-containeranalysis", - ).version, - ) -except pkg_resources.DistributionNotFound: - DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo() +DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( + gapic_version=package_version.__version__ +) class ContainerAnalysisTransport(abc.ABC): diff --git a/release-please-config.json b/release-please-config.json index 595994a..23cf035 100644 --- a/release-please-config.json +++ b/release-please-config.json @@ -5,6 +5,7 @@ "release-type": "python", "extra-files": [ "google/cloud/devtools/containeranalysis/gapic_version.py", + "google/cloud/devtools/containeranalysis_v1/gapic_version.py", { "type": "json", "path": "samples/generated_samples/snippet_metadata_google.devtools.containeranalysis.v1.json", From ede0f1a961b898528bb7c7181a3d7cc740b149ed Mon Sep 17 00:00:00 2001 From: Sampath Kumar Date: Wed, 7 Dec 2022 18:26:26 +0100 Subject: [PATCH 11/13] chore: Python code samples migrated to python-doc-samples (#348) MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit * Python code samples migrated to python-doc-samples For more details, please check https://github.com/GoogleCloudPlatform/python-docs-samples/pull/8529 and http://b/257074849 * πŸ¦‰ Updates from OwlBot post-processor See https://github.com/googleapis/repo-automation-bots/blob/main/packages/owl-bot/README.md Co-authored-by: Owl Bot --- samples/README.md | 4 + samples/snippets/.gitignore | 1 - samples/snippets/README.md | 54 ---- samples/snippets/noxfile.py | 292 ------------------ samples/snippets/requirements-test.txt | 1 - samples/snippets/requirements.txt | 6 - samples/snippets/samples.py | 402 ------------------------- samples/snippets/samples_test.py | 292 ------------------ 8 files changed, 4 insertions(+), 1048 deletions(-) create mode 100644 samples/README.md delete mode 100644 samples/snippets/.gitignore delete mode 100644 samples/snippets/README.md delete mode 100644 samples/snippets/noxfile.py delete mode 100644 samples/snippets/requirements-test.txt delete mode 100644 samples/snippets/requirements.txt delete mode 100644 samples/snippets/samples.py delete mode 100644 samples/snippets/samples_test.py diff --git a/samples/README.md b/samples/README.md new file mode 100644 index 0000000..e6d225b --- /dev/null +++ b/samples/README.md @@ -0,0 +1,4 @@ +Samples migrated +================ + +New location: https://github.com/GoogleCloudPlatform/python-docs-samples/tree/main/containeranalysis/snippets diff --git a/samples/snippets/.gitignore b/samples/snippets/.gitignore deleted file mode 100644 index 9e3d04c..0000000 --- a/samples/snippets/.gitignore +++ /dev/null @@ -1 +0,0 @@ -venv* diff --git a/samples/snippets/README.md b/samples/snippets/README.md deleted file mode 100644 index 73c45c3..0000000 --- a/samples/snippets/README.md +++ /dev/null @@ -1,54 +0,0 @@ -Google
-Cloud Platform logo - -# Google Cloud Container Analysis Samples - - -Container Analysis scans container images stored in Container Registry for vulnerabilities. -Continuous automated analysis of containers keep you informed about known vulnerabilities so -that you can review and address issues before deployment. - -Additionally, third-party metadata providers can use Container Analysis to store and -retrieve additional metadata for their customers' images, such as packages installed in an image. - - -## Description - -These samples show how to use the [Google Cloud Container Analysis Client Library](https://cloud.google.com/container-registry/docs/reference/libraries). - -## Build and Run -1. **Enable APIs** - - [Enable the Container Analysis API](https://console.cloud.google.com/flows/enableapi?apiid=containeranalysis.googleapis.com) - and create a new project or select an existing project. -1. **Install and Initialize Cloud SDK** - - Follow instructions from the available [quickstarts](https://cloud.google.com/sdk/docs/quickstarts) -1. **Authenticate with GCP** - - Typically, you should authenticate using a [service account key](https://cloud.google.com/docs/authentication/getting-started) -1. **Clone the repo** and cd into this directory - - ``` - git clone https://github.com/GoogleCloudPlatform/python-docs-samples - cd python-docs-samples - ``` - -1. **Set Environment Variables** - - ``` - export GCLOUD_PROJECT="YOUR_PROJECT_ID" - ``` - -1. **Run Tests** - - ``` - nox -s "py36(sample='./container_registry/container_analysis')" - ``` - -## Contributing changes - -* See [CONTRIBUTING.md](../../CONTRIBUTING.md) - -## Licensing - -* See [LICENSE](../../LICENSE) - diff --git a/samples/snippets/noxfile.py b/samples/snippets/noxfile.py deleted file mode 100644 index 0577084..0000000 --- a/samples/snippets/noxfile.py +++ /dev/null @@ -1,292 +0,0 @@ -# Copyright 2019 Google LLC -# -# Licensed under the Apache License, Version 2.0 (the "License"); -# you may not use this file except in compliance with the License. -# You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. - -from __future__ import print_function - -import glob -import os -from pathlib import Path -import sys -from typing import Callable, Dict, Optional - -import nox - -# WARNING - WARNING - WARNING - WARNING - WARNING -# WARNING - WARNING - WARNING - WARNING - WARNING -# DO NOT EDIT THIS FILE EVER! -# WARNING - WARNING - WARNING - WARNING - WARNING -# WARNING - WARNING - WARNING - WARNING - WARNING - -BLACK_VERSION = "black==22.3.0" -ISORT_VERSION = "isort==5.10.1" - -# Copy `noxfile_config.py` to your directory and modify it instead. - -# `TEST_CONFIG` dict is a configuration hook that allows users to -# modify the test configurations. The values here should be in sync -# with `noxfile_config.py`. Users will copy `noxfile_config.py` into -# their directory and modify it. - -TEST_CONFIG = { - # You can opt out from the test for specific Python versions. - "ignored_versions": [], - # Old samples are opted out of enforcing Python type hints - # All new samples should feature them - "enforce_type_hints": False, - # An envvar key for determining the project id to use. Change it - # to 'BUILD_SPECIFIC_GCLOUD_PROJECT' if you want to opt in using a - # build specific Cloud project. You can also use your own string - # to use your own Cloud project. - "gcloud_project_env": "GOOGLE_CLOUD_PROJECT", - # 'gcloud_project_env': 'BUILD_SPECIFIC_GCLOUD_PROJECT', - # If you need to use a specific version of pip, - # change pip_version_override to the string representation - # of the version number, for example, "20.2.4" - "pip_version_override": None, - # A dictionary you want to inject into your test. Don't put any - # secrets here. These values will override predefined values. - "envs": {}, -} - - -try: - # Ensure we can import noxfile_config in the project's directory. - sys.path.append(".") - from noxfile_config import TEST_CONFIG_OVERRIDE -except ImportError as e: - print("No user noxfile_config found: detail: {}".format(e)) - TEST_CONFIG_OVERRIDE = {} - -# Update the TEST_CONFIG with the user supplied values. -TEST_CONFIG.update(TEST_CONFIG_OVERRIDE) - - -def get_pytest_env_vars() -> Dict[str, str]: - """Returns a dict for pytest invocation.""" - ret = {} - - # Override the GCLOUD_PROJECT and the alias. - env_key = TEST_CONFIG["gcloud_project_env"] - # This should error out if not set. - ret["GOOGLE_CLOUD_PROJECT"] = os.environ[env_key] - - # Apply user supplied envs. - ret.update(TEST_CONFIG["envs"]) - return ret - - -# DO NOT EDIT - automatically generated. -# All versions used to test samples. -ALL_VERSIONS = ["3.7", "3.8", "3.9", "3.10"] - -# Any default versions that should be ignored. -IGNORED_VERSIONS = TEST_CONFIG["ignored_versions"] - -TESTED_VERSIONS = sorted([v for v in ALL_VERSIONS if v not in IGNORED_VERSIONS]) - -INSTALL_LIBRARY_FROM_SOURCE = os.environ.get("INSTALL_LIBRARY_FROM_SOURCE", False) in ( - "True", - "true", -) - -# Error if a python version is missing -nox.options.error_on_missing_interpreters = True - -# -# Style Checks -# - - -# Linting with flake8. -# -# We ignore the following rules: -# E203: whitespace before β€˜:’ -# E266: too many leading β€˜#’ for block comment -# E501: line too long -# I202: Additional newline in a section of imports -# -# We also need to specify the rules which are ignored by default: -# ['E226', 'W504', 'E126', 'E123', 'W503', 'E24', 'E704', 'E121'] -FLAKE8_COMMON_ARGS = [ - "--show-source", - "--builtin=gettext", - "--max-complexity=20", - "--exclude=.nox,.cache,env,lib,generated_pb2,*_pb2.py,*_pb2_grpc.py", - "--ignore=E121,E123,E126,E203,E226,E24,E266,E501,E704,W503,W504,I202", - "--max-line-length=88", -] - - -@nox.session -def lint(session: nox.sessions.Session) -> None: - if not TEST_CONFIG["enforce_type_hints"]: - session.install("flake8") - else: - session.install("flake8", "flake8-annotations") - - args = FLAKE8_COMMON_ARGS + [ - ".", - ] - session.run("flake8", *args) - - -# -# Black -# - - -@nox.session -def blacken(session: nox.sessions.Session) -> None: - """Run black. Format code to uniform standard.""" - session.install(BLACK_VERSION) - python_files = [path for path in os.listdir(".") if path.endswith(".py")] - - session.run("black", *python_files) - - -# -# format = isort + black -# - - -@nox.session -def format(session: nox.sessions.Session) -> None: - """ - Run isort to sort imports. Then run black - to format code to uniform standard. - """ - session.install(BLACK_VERSION, ISORT_VERSION) - python_files = [path for path in os.listdir(".") if path.endswith(".py")] - - # Use the --fss option to sort imports using strict alphabetical order. - # See https://pycqa.github.io/isort/docs/configuration/options.html#force-sort-within-sections - session.run("isort", "--fss", *python_files) - session.run("black", *python_files) - - -# -# Sample Tests -# - - -PYTEST_COMMON_ARGS = ["--junitxml=sponge_log.xml"] - - -def _session_tests( - session: nox.sessions.Session, post_install: Callable = None -) -> None: - # check for presence of tests - test_list = glob.glob("**/*_test.py", recursive=True) + glob.glob( - "**/test_*.py", recursive=True - ) - test_list.extend(glob.glob("**/tests", recursive=True)) - - if len(test_list) == 0: - print("No tests found, skipping directory.") - return - - if TEST_CONFIG["pip_version_override"]: - pip_version = TEST_CONFIG["pip_version_override"] - session.install(f"pip=={pip_version}") - """Runs py.test for a particular project.""" - concurrent_args = [] - if os.path.exists("requirements.txt"): - if os.path.exists("constraints.txt"): - session.install("-r", "requirements.txt", "-c", "constraints.txt") - else: - session.install("-r", "requirements.txt") - with open("requirements.txt") as rfile: - packages = rfile.read() - - if os.path.exists("requirements-test.txt"): - if os.path.exists("constraints-test.txt"): - session.install("-r", "requirements-test.txt", "-c", "constraints-test.txt") - else: - session.install("-r", "requirements-test.txt") - with open("requirements-test.txt") as rtfile: - packages += rtfile.read() - - if INSTALL_LIBRARY_FROM_SOURCE: - session.install("-e", _get_repo_root()) - - if post_install: - post_install(session) - - if "pytest-parallel" in packages: - concurrent_args.extend(["--workers", "auto", "--tests-per-worker", "auto"]) - elif "pytest-xdist" in packages: - concurrent_args.extend(["-n", "auto"]) - - session.run( - "pytest", - *(PYTEST_COMMON_ARGS + session.posargs + concurrent_args), - # Pytest will return 5 when no tests are collected. This can happen - # on travis where slow and flaky tests are excluded. - # See http://doc.pytest.org/en/latest/_modules/_pytest/main.html - success_codes=[0, 5], - env=get_pytest_env_vars(), - ) - - -@nox.session(python=ALL_VERSIONS) -def py(session: nox.sessions.Session) -> None: - """Runs py.test for a sample using the specified version of Python.""" - if session.python in TESTED_VERSIONS: - _session_tests(session) - else: - session.skip( - "SKIPPED: {} tests are disabled for this sample.".format(session.python) - ) - - -# -# Readmegen -# - - -def _get_repo_root() -> Optional[str]: - """Returns the root folder of the project.""" - # Get root of this repository. Assume we don't have directories nested deeper than 10 items. - p = Path(os.getcwd()) - for i in range(10): - if p is None: - break - if Path(p / ".git").exists(): - return str(p) - # .git is not available in repos cloned via Cloud Build - # setup.py is always in the library's root, so use that instead - # https://github.com/googleapis/synthtool/issues/792 - if Path(p / "setup.py").exists(): - return str(p) - p = p.parent - raise Exception("Unable to detect repository root.") - - -GENERATED_READMES = sorted([x for x in Path(".").rglob("*.rst.in")]) - - -@nox.session -@nox.parametrize("path", GENERATED_READMES) -def readmegen(session: nox.sessions.Session, path: str) -> None: - """(Re-)generates the readme for a sample.""" - session.install("jinja2", "pyyaml") - dir_ = os.path.dirname(path) - - if os.path.exists(os.path.join(dir_, "requirements.txt")): - session.install("-r", os.path.join(dir_, "requirements.txt")) - - in_file = os.path.join(dir_, "README.rst.in") - session.run( - "python", _get_repo_root() + "/scripts/readme-gen/readme_gen.py", in_file - ) diff --git a/samples/snippets/requirements-test.txt b/samples/snippets/requirements-test.txt deleted file mode 100644 index 49780e0..0000000 --- a/samples/snippets/requirements-test.txt +++ /dev/null @@ -1 +0,0 @@ -pytest==7.2.0 diff --git a/samples/snippets/requirements.txt b/samples/snippets/requirements.txt deleted file mode 100644 index d347745..0000000 --- a/samples/snippets/requirements.txt +++ /dev/null @@ -1,6 +0,0 @@ -google-cloud-pubsub==2.13.11 -google-cloud-containeranalysis==2.9.3 -grafeas==1.6.1 -pytest==7.2.0 -flaky==3.7.0 -mock==4.0.3 diff --git a/samples/snippets/samples.py b/samples/snippets/samples.py deleted file mode 100644 index 3f22117..0000000 --- a/samples/snippets/samples.py +++ /dev/null @@ -1,402 +0,0 @@ -#!/bin/python -# Copyright 2019 Google LLC -# -# Licensed under the Apache License, Version 2.0 (the "License"); -# you may not use this file except in compliance with the License. -# You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. - - -# [START containeranalysis_create_note] -def create_note(note_id, project_id): - """Creates and returns a new vulnerability note.""" - # note_id = 'my-note' - # project_id = 'my-gcp-project' - - from google.cloud.devtools import containeranalysis_v1 - from grafeas.grafeas_v1 import Version - - client = containeranalysis_v1.ContainerAnalysisClient() - grafeas_client = client.get_grafeas_client() - project_name = f"projects/{project_id}" - note = { - "vulnerability": { - "details": [ - { - "affected_cpe_uri": "your-uri-here", - "affected_package": "your-package-here", - "affected_version_start": {"kind": Version.VersionKind.MINIMUM}, - "fixed_version": {"kind": Version.VersionKind.MAXIMUM}, - } - ] - } - } - response = grafeas_client.create_note( - parent=project_name, note_id=note_id, note=note - ) - return response - - -# [END containeranalysis_create_note] - - -# [START containeranalysis_delete_note] -def delete_note(note_id, project_id): - """Removes an existing note from the server.""" - # note_id = 'my-note' - # project_id = 'my-gcp-project' - - from google.cloud.devtools import containeranalysis_v1 - - client = containeranalysis_v1.ContainerAnalysisClient() - grafeas_client = client.get_grafeas_client() - note_name = f"projects/{project_id}/notes/{note_id}" - - grafeas_client.delete_note(name=note_name) - - -# [END containeranalysis_delete_note] - - -# [START containeranalysis_create_occurrence] -def create_occurrence(resource_url, note_id, occurrence_project, note_project): - """Creates and returns a new occurrence of a previously - created vulnerability note.""" - # resource_url = 'https://gcr.io/my-project/my-image@sha256:123' - # note_id = 'my-note' - # occurrence_project = 'my-gcp-project' - # note_project = 'my-gcp-project' - - from google.cloud.devtools import containeranalysis_v1 - from grafeas.grafeas_v1 import Version - - client = containeranalysis_v1.ContainerAnalysisClient() - grafeas_client = client.get_grafeas_client() - formatted_note = f"projects/{note_project}/notes/{note_id}" - formatted_project = f"projects/{occurrence_project}" - - occurrence = { - "note_name": formatted_note, - "resource_uri": resource_url, - "vulnerability": { - "package_issue": [ - { - "affected_cpe_uri": "your-uri-here", - "affected_package": "your-package-here", - "affected_version": {"kind": Version.VersionKind.MINIMUM}, - "fixed_version": {"kind": Version.VersionKind.MAXIMUM}, - } - ] - }, - } - - return grafeas_client.create_occurrence( - parent=formatted_project, occurrence=occurrence - ) - - -# [END containeranalysis_create_occurrence] - - -# [START containeranalysis_delete_occurrence] -def delete_occurrence(occurrence_id, project_id): - """Removes an existing occurrence from the server.""" - # occurrence_id = basename(occurrence.name) - # project_id = 'my-gcp-project' - - from google.cloud.devtools import containeranalysis_v1 - - client = containeranalysis_v1.ContainerAnalysisClient() - grafeas_client = client.get_grafeas_client() - parent = f"projects/{project_id}/occurrences/{occurrence_id}" - grafeas_client.delete_occurrence(name=parent) - - -# [END containeranalysis_delete_occurrence] - - -# [START containeranalysis_get_note] -def get_note(note_id, project_id): - """Retrieves and prints a specified note from the server.""" - # note_id = 'my-note' - # project_id = 'my-gcp-project' - - from google.cloud.devtools import containeranalysis_v1 - - client = containeranalysis_v1.ContainerAnalysisClient() - grafeas_client = client.get_grafeas_client() - note_name = f"projects/{project_id}/notes/{note_id}" - response = grafeas_client.get_note(name=note_name) - return response - - -# [END containeranalysis_get_note] - - -# [START containeranalysis_get_occurrence] -def get_occurrence(occurrence_id, project_id): - """retrieves and prints a specified occurrence from the server.""" - # occurrence_id = basename(occurrence.name) - # project_id = 'my-gcp-project' - - from google.cloud.devtools import containeranalysis_v1 - - client = containeranalysis_v1.ContainerAnalysisClient() - grafeas_client = client.get_grafeas_client() - parent = f"projects/{project_id}/occurrences/{occurrence_id}" - return grafeas_client.get_occurrence(name=parent) - - -# [END containeranalysis_get_occurrence] - - -# [START containeranalysis_discovery_info] -def get_discovery_info(resource_url, project_id): - """Retrieves and prints the discovery occurrence created for a specified - image. The discovery occurrence contains information about the initial - scan on the image.""" - # resource_url = 'https://gcr.io/my-project/my-image@sha256:123' - # project_id = 'my-gcp-project' - - from google.cloud.devtools import containeranalysis_v1 - - filter_str = 'kind="DISCOVERY" AND resourceUrl="{}"'.format(resource_url) - client = containeranalysis_v1.ContainerAnalysisClient() - grafeas_client = client.get_grafeas_client() - project_name = f"projects/{project_id}" - response = grafeas_client.list_occurrences(parent=project_name, filter_=filter_str) - for occ in response: - print(occ) - - -# [END containeranalysis_discovery_info] - - -# [START containeranalysis_occurrences_for_note] -def get_occurrences_for_note(note_id, project_id): - """Retrieves all the occurrences associated with a specified Note. - Here, all occurrences are printed and counted.""" - # note_id = 'my-note' - # project_id = 'my-gcp-project' - - from google.cloud.devtools import containeranalysis_v1 - - client = containeranalysis_v1.ContainerAnalysisClient() - grafeas_client = client.get_grafeas_client() - note_name = f"projects/{project_id}/notes/{note_id}" - - response = grafeas_client.list_note_occurrences(name=note_name) - count = 0 - for o in response: - # do something with the retrieved occurrence - # in this sample, we will simply count each one - count += 1 - return count - - -# [END containeranalysis_occurrences_for_note] - - -# [START containeranalysis_occurrences_for_image] -def get_occurrences_for_image(resource_url, project_id): - """Retrieves all the occurrences associated with a specified image. - Here, all occurrences are simply printed and counted.""" - # resource_url = 'https://gcr.io/my-project/my-image@sha256:123' - # project_id = 'my-gcp-project' - - from google.cloud.devtools import containeranalysis_v1 - - filter_str = 'resourceUrl="{}"'.format(resource_url) - client = containeranalysis_v1.ContainerAnalysisClient() - grafeas_client = client.get_grafeas_client() - project_name = f"projects/{project_id}" - - response = grafeas_client.list_occurrences(parent=project_name, filter=filter_str) - count = 0 - for o in response: - # do something with the retrieved occurrence - # in this sample, we will simply count each one - count += 1 - return count - - -# [END containeranalysis_occurrences_for_image] - - -# [START containeranalysis_pubsub] -def pubsub(subscription_id, timeout_seconds, project_id): - """Respond to incoming occurrences using a Cloud Pub/Sub subscription.""" - # subscription_id := 'my-occurrences-subscription' - # timeout_seconds = 20 - # project_id = 'my-gcp-project' - - import time - - from google.cloud.pubsub import SubscriberClient - - client = SubscriberClient() - subscription_name = client.subscription_path(project_id, subscription_id) - receiver = MessageReceiver() - client.subscribe(subscription_name, receiver.pubsub_callback) - - # listen for 'timeout' seconds - for _ in range(timeout_seconds): - time.sleep(1) - # print and return the number of pubsub messages received - print(receiver.msg_count) - return receiver.msg_count - - -class MessageReceiver: - """Custom class to handle incoming Pub/Sub messages.""" - - def __init__(self): - # initialize counter to 0 on initialization - self.msg_count = 0 - - def pubsub_callback(self, message): - # every time a pubsub message comes in, print it and count it - self.msg_count += 1 - print("Message {}: {}".format(self.msg_count, message.data)) - message.ack() - - -def create_occurrence_subscription(subscription_id, project_id): - """Creates a new Pub/Sub subscription object listening to the - Container Analysis Occurrences topic.""" - # subscription_id := 'my-occurrences-subscription' - # project_id = 'my-gcp-project' - - from google.api_core.exceptions import AlreadyExists - from google.cloud.pubsub import SubscriberClient - - topic_id = "container-analysis-occurrences-v1" - client = SubscriberClient() - topic_name = f"projects/{project_id}/topics/{topic_id}" - subscription_name = client.subscription_path(project_id, subscription_id) - success = True - try: - client.create_subscription({"name": subscription_name, "topic": topic_name}) - except AlreadyExists: - # if subscription already exists, do nothing - pass - else: - success = False - return success - - -# [END containeranalysis_pubsub] - - -# [START containeranalysis_poll_discovery_occurrence_finished] -def poll_discovery_finished(resource_url, timeout_seconds, project_id): - """Returns the discovery occurrence for a resource once it reaches a - terminal state.""" - # resource_url = 'https://gcr.io/my-project/my-image@sha256:123' - # timeout_seconds = 20 - # project_id = 'my-gcp-project' - - import time - - from google.cloud.devtools import containeranalysis_v1 - from grafeas.grafeas_v1 import DiscoveryOccurrence - - deadline = time.time() + timeout_seconds - - client = containeranalysis_v1.ContainerAnalysisClient() - grafeas_client = client.get_grafeas_client() - project_name = f"projects/{project_id}" - - discovery_occurrence = None - while discovery_occurrence is None: - time.sleep(1) - filter_str = 'resourceUrl="{}" \ - AND noteProjectId="goog-analysis" \ - AND noteId="PACKAGE_VULNERABILITY"'.format( - resource_url - ) - # [END containeranalysis_poll_discovery_occurrence_finished] - # The above filter isn't testable, since it looks for occurrences in a - # locked down project fall back to a more permissive filter for testing - filter_str = 'kind="DISCOVERY" AND resourceUrl="{}"'.format(resource_url) - # [START containeranalysis_poll_discovery_occurrence_finished] - result = grafeas_client.list_occurrences(parent=project_name, filter=filter_str) - # only one occurrence should ever be returned by ListOccurrences - # and the given filter - for item in result: - discovery_occurrence = item - if time.time() > deadline: - raise RuntimeError("timeout while retrieving discovery occurrence") - - status = DiscoveryOccurrence.AnalysisStatus.PENDING - while ( - status != DiscoveryOccurrence.AnalysisStatus.FINISHED_UNSUPPORTED - and status != DiscoveryOccurrence.AnalysisStatus.FINISHED_FAILED - and status != DiscoveryOccurrence.AnalysisStatus.FINISHED_SUCCESS - ): - time.sleep(1) - updated = grafeas_client.get_occurrence(name=discovery_occurrence.name) - status = updated.discovery.analysis_status - if time.time() > deadline: - raise RuntimeError("timeout while waiting for terminal state") - return discovery_occurrence - - -# [END containeranalysis_poll_discovery_occurrence_finished] - - -# [START containeranalysis_vulnerability_occurrences_for_image] -def find_vulnerabilities_for_image(resource_url, project_id): - """ "Retrieves all vulnerability occurrences associated with a resource.""" - # resource_url = 'https://gcr.io/my-project/my-image@sha256:123' - # project_id = 'my-gcp-project' - - from google.cloud.devtools import containeranalysis_v1 - - client = containeranalysis_v1.ContainerAnalysisClient() - grafeas_client = client.get_grafeas_client() - project_name = f"projects/{project_id}" - - filter_str = 'kind="VULNERABILITY" AND resourceUrl="{}"'.format(resource_url) - return list(grafeas_client.list_occurrences(parent=project_name, filter=filter_str)) - - -# [END containeranalysis_vulnerability_occurrences_for_image] - - -# [START containeranalysis_filter_vulnerability_occurrences] -def find_high_severity_vulnerabilities_for_image(resource_url, project_id): - """Retrieves a list of only high vulnerability occurrences associated - with a resource.""" - # resource_url = 'https://gcr.io/my-project/my-image@sha256:123' - # project_id = 'my-gcp-project' - - from google.cloud.devtools import containeranalysis_v1 - from grafeas.grafeas_v1 import Severity - - client = containeranalysis_v1.ContainerAnalysisClient() - grafeas_client = client.get_grafeas_client() - project_name = f"projects/{project_id}" - - filter_str = 'kind="VULNERABILITY" AND resourceUrl="{}"'.format(resource_url) - vulnerabilities = grafeas_client.list_occurrences( - parent=project_name, filter=filter_str - ) - filtered_list = [] - for v in vulnerabilities: - if ( - v.vulnerability.effective_severity == Severity.HIGH - or v.vulnerability.effective_severity == Severity.CRITICAL - ): - filtered_list.append(v) - return filtered_list - - -# [END containeranalysis_filter_vulnerability_occurrences] diff --git a/samples/snippets/samples_test.py b/samples/snippets/samples_test.py deleted file mode 100644 index 8ed97fa..0000000 --- a/samples/snippets/samples_test.py +++ /dev/null @@ -1,292 +0,0 @@ -#!/bin/python -# Copyright 2019 Google LLC -# -# Licensed under the Apache License, Version 2.0 (the "License"); -# you may not use this file except in compliance with the License. -# You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. - -from os import environ -from os.path import basename -import threading -import time -import uuid - -from google.api_core.exceptions import AlreadyExists, InvalidArgument, NotFound -from google.cloud.devtools import containeranalysis_v1 -from google.cloud.pubsub import PublisherClient, SubscriberClient -from grafeas.grafeas_v1 import DiscoveryOccurrence, NoteKind, Severity, Version -import pytest - -import samples - -PROJECT_ID = environ["GOOGLE_CLOUD_PROJECT"] -SLEEP_TIME = 1 -TRY_LIMIT = 20 - - -class MessageReceiver: - """Custom class to handle incoming Pub/Sub messages.""" - - def __init__(self, expected_msg_nums, done_event): - # initialize counter to 0 on initialization - self.msg_count = 0 - self.expected_msg_nums = expected_msg_nums - self.done_event = done_event - - def pubsub_callback(self, message): - # every time a pubsub message comes in, print it and count it - self.msg_count += 1 - print("Message {}: {}".format(self.msg_count, message.data)) - message.ack() - if self.msg_count == self.expected_msg_nums: - self.done_event.set() - - -class TestContainerAnalysisSamples: - def setup_method(self, test_method): - print("SETUP {}".format(test_method.__name__)) - self.note_id = "note-{}".format(uuid.uuid4()) - self.image_url = "{}.{}".format(uuid.uuid4(), test_method.__name__) - self.note_obj = samples.create_note(self.note_id, PROJECT_ID) - - def teardown_method(self, test_method): - print("TEAR DOWN {}".format(test_method.__name__)) - try: - samples.delete_note(self.note_id, PROJECT_ID) - except NotFound: - pass - - def test_create_note(self): - new_note = samples.get_note(self.note_id, PROJECT_ID) - assert new_note.name == self.note_obj.name - - def test_delete_note(self): - samples.delete_note(self.note_id, PROJECT_ID) - try: - samples.get_note(self.note_obj, PROJECT_ID) - except InvalidArgument: - pass - else: - # didn't raise exception we expected - assert False - - def test_create_occurrence(self): - created = samples.create_occurrence( - self.image_url, self.note_id, PROJECT_ID, PROJECT_ID - ) - retrieved = samples.get_occurrence(basename(created.name), PROJECT_ID) - assert created.name == retrieved.name - # clean up - samples.delete_occurrence(basename(created.name), PROJECT_ID) - - def test_delete_occurrence(self): - created = samples.create_occurrence( - self.image_url, self.note_id, PROJECT_ID, PROJECT_ID - ) - samples.delete_occurrence(basename(created.name), PROJECT_ID) - try: - samples.get_occurrence(basename(created.name), PROJECT_ID) - except NotFound: - pass - else: - # didn't raise exception we expected - assert False - - def test_occurrences_for_image(self): - orig_count = samples.get_occurrences_for_image(self.image_url, PROJECT_ID) - occ = samples.create_occurrence( - self.image_url, self.note_id, PROJECT_ID, PROJECT_ID - ) - new_count = 0 - tries = 0 - while new_count != 1 and tries < TRY_LIMIT: - tries += 1 - new_count = samples.get_occurrences_for_image(self.image_url, PROJECT_ID) - time.sleep(SLEEP_TIME) - assert new_count == 1 - assert orig_count == 0 - # clean up - samples.delete_occurrence(basename(occ.name), PROJECT_ID) - - def test_occurrences_for_note(self): - orig_count = samples.get_occurrences_for_note(self.note_id, PROJECT_ID) - occ = samples.create_occurrence( - self.image_url, self.note_id, PROJECT_ID, PROJECT_ID - ) - new_count = 0 - tries = 0 - while new_count != 1 and tries < TRY_LIMIT: - tries += 1 - new_count = samples.get_occurrences_for_note(self.note_id, PROJECT_ID) - time.sleep(SLEEP_TIME) - assert new_count == 1 - assert orig_count == 0 - # clean up - samples.delete_occurrence(basename(occ.name), PROJECT_ID) - - @pytest.mark.flaky(max_runs=3, min_passes=1) - def test_pubsub(self): - # create topic if needed - client = SubscriberClient() - try: - topic_id = "container-analysis-occurrences-v1" - topic_name = {"name": f"projects/{PROJECT_ID}/topics/{topic_id}"} - publisher = PublisherClient() - publisher.create_topic(topic_name) - except AlreadyExists: - pass - - subscription_id = "container-analysis-test-{}".format(uuid.uuid4()) - subscription_name = client.subscription_path(PROJECT_ID, subscription_id) - samples.create_occurrence_subscription(subscription_id, PROJECT_ID) - - # I can not make it pass with multiple messages. My guess is - # the server started to dedup? - message_count = 1 - try: - job_done = threading.Event() - receiver = MessageReceiver(message_count, job_done) - client.subscribe(subscription_name, receiver.pubsub_callback) - - for i in range(message_count): - occ = samples.create_occurrence( - self.image_url, self.note_id, PROJECT_ID, PROJECT_ID - ) - time.sleep(SLEEP_TIME) - samples.delete_occurrence(basename(occ.name), PROJECT_ID) - time.sleep(SLEEP_TIME) - # We saw occational failure with 60 seconds timeout, so we bumped it - # to 180 seconds. - # See also: python-docs-samples/issues/2894 - job_done.wait(timeout=180) - print("done. msg_count = {}".format(receiver.msg_count)) - assert message_count <= receiver.msg_count - finally: - # clean up - client.delete_subscription({"subscription": subscription_name}) - - def test_poll_discovery_occurrence_fails(self): - # try with no discovery occurrence - try: - samples.poll_discovery_finished(self.image_url, 5, PROJECT_ID) - except RuntimeError: - pass - else: - # we expect timeout error - assert False - - @pytest.mark.flaky(max_runs=3, min_passes=1) - def test_poll_discovery_occurrence(self): - # create discovery occurrence - note_id = "discovery-note-{}".format(uuid.uuid4()) - client = containeranalysis_v1.ContainerAnalysisClient() - grafeas_client = client.get_grafeas_client() - note = {"discovery": {"analysis_kind": NoteKind.DISCOVERY}} - grafeas_client.create_note( - parent=f"projects/{PROJECT_ID}", note_id=note_id, note=note - ) - occurrence = { - "note_name": f"projects/{PROJECT_ID}/notes/{note_id}", - "resource_uri": self.image_url, - "discovery": { - "analysis_status": DiscoveryOccurrence.AnalysisStatus.FINISHED_SUCCESS - }, - } - created = grafeas_client.create_occurrence( - parent=f"projects/{PROJECT_ID}", occurrence=occurrence - ) - - disc = samples.poll_discovery_finished(self.image_url, 10, PROJECT_ID) - status = disc.discovery.analysis_status - assert disc is not None - assert status == DiscoveryOccurrence.AnalysisStatus.FINISHED_SUCCESS - - # clean up - samples.delete_occurrence(basename(created.name), PROJECT_ID) - samples.delete_note(note_id, PROJECT_ID) - - def test_find_vulnerabilities_for_image(self): - occ_list = samples.find_vulnerabilities_for_image(self.image_url, PROJECT_ID) - assert len(occ_list) == 0 - - created = samples.create_occurrence( - self.image_url, self.note_id, PROJECT_ID, PROJECT_ID - ) - tries = 0 - count = 0 - while count != 1 and tries < TRY_LIMIT: - tries += 1 - occ_list = samples.find_vulnerabilities_for_image( - self.image_url, PROJECT_ID - ) - count = len(occ_list) - time.sleep(SLEEP_TIME) - assert len(occ_list) == 1 - samples.delete_occurrence(basename(created.name), PROJECT_ID) - - def test_find_high_severity_vulnerabilities(self): - occ_list = samples.find_high_severity_vulnerabilities_for_image( - self.image_url, PROJECT_ID - ) - assert len(occ_list) == 0 - - # create new high severity vulnerability - note_id = "discovery-note-{}".format(uuid.uuid4()) - client = containeranalysis_v1.ContainerAnalysisClient() - grafeas_client = client.get_grafeas_client() - note = { - "vulnerability": { - "severity": Severity.CRITICAL, - "details": [ - { - "affected_cpe_uri": "your-uri-here", - "affected_package": "your-package-here", - "affected_version_start": {"kind": Version.VersionKind.MINIMUM}, - "fixed_version": {"kind": Version.VersionKind.MAXIMUM}, - } - ], - } - } - grafeas_client.create_note( - parent=f"projects/{PROJECT_ID}", note_id=note_id, note=note - ) - occurrence = { - "note_name": f"projects/{PROJECT_ID}/notes/{note_id}", - "resource_uri": self.image_url, - "vulnerability": { - "effective_severity": Severity.CRITICAL, - "package_issue": [ - { - "affected_cpe_uri": "your-uri-here", - "affected_package": "your-package-here", - "affected_version": {"kind": Version.VersionKind.MINIMUM}, - "fixed_version": {"kind": Version.VersionKind.MAXIMUM}, - } - ], - }, - } - created = grafeas_client.create_occurrence( - parent=f"projects/{PROJECT_ID}", occurrence=occurrence - ) - # query again - tries = 0 - count = 0 - while count != 1 and tries < TRY_LIMIT: - tries += 1 - occ_list = samples.find_vulnerabilities_for_image( - self.image_url, PROJECT_ID - ) - count = len(occ_list) - time.sleep(SLEEP_TIME) - assert len(occ_list) == 1 - # clean up - samples.delete_occurrence(basename(created.name), PROJECT_ID) - samples.delete_note(note_id, PROJECT_ID) From 213a530ce5e23351c4f36c7219419324565bfe02 Mon Sep 17 00:00:00 2001 From: "gcf-owl-bot[bot]" <78513119+gcf-owl-bot[bot]@users.noreply.github.com> Date: Wed, 14 Dec 2022 10:11:29 -0500 Subject: [PATCH 12/13] build(deps): bump certifi from 2022.9.24 to 2022.12.7 in /synthtool/gcp/templates/python_library/.kokoro (#351) Source-Link: https://github.com/googleapis/synthtool/commit/b4fe62efb5114b6738ad4b13d6f654f2bf4b7cc0 Post-Processor: gcr.io/cloud-devrel-public-resources/owlbot-python:latest@sha256:3bf87e47c2173d7eed42714589dc4da2c07c3268610f1e47f8e1a30decbfc7f1 Co-authored-by: Owl Bot --- .github/.OwlBot.lock.yaml | 2 +- .kokoro/requirements.txt | 6 +++--- .pre-commit-config.yaml | 2 +- 3 files changed, 5 insertions(+), 5 deletions(-) diff --git a/.github/.OwlBot.lock.yaml b/.github/.OwlBot.lock.yaml index bb21147..fccaa8e 100644 --- a/.github/.OwlBot.lock.yaml +++ b/.github/.OwlBot.lock.yaml @@ -13,4 +13,4 @@ # limitations under the License. docker: image: gcr.io/cloud-devrel-public-resources/owlbot-python:latest - digest: sha256:3abfa0f1886adaf0b83f07cb117b24a639ea1cb9cffe56d43280b977033563eb + digest: sha256:3bf87e47c2173d7eed42714589dc4da2c07c3268610f1e47f8e1a30decbfc7f1 diff --git a/.kokoro/requirements.txt b/.kokoro/requirements.txt index 9c1b9be..05dc467 100644 --- a/.kokoro/requirements.txt +++ b/.kokoro/requirements.txt @@ -20,9 +20,9 @@ cachetools==5.2.0 \ --hash=sha256:6a94c6402995a99c3970cc7e4884bb60b4a8639938157eeed436098bf9831757 \ --hash=sha256:f9f17d2aec496a9aa6b76f53e3b614c965223c061982d434d160f930c698a9db # via google-auth -certifi==2022.9.24 \ - --hash=sha256:0d9c601124e5a6ba9712dbc60d9c53c21e34f5f641fe83002317394311bdce14 \ - --hash=sha256:90c1a32f1d68f940488354e36370f6cca89f0f106db09518524c88d6ed83f382 +certifi==2022.12.7 \ + --hash=sha256:35824b4c3a97115964b408844d64aa14db1cc518f6562e8d7261699d1350a9e3 \ + --hash=sha256:4ad3232f5e926d6718ec31cfc1fcadfde020920e278684144551c91769c7bc18 # via requests cffi==1.15.1 \ --hash=sha256:00a9ed42e88df81ffae7a8ab6d9356b371399b91dbdf0c3cb1e84c03a13aceb5 \ diff --git a/.pre-commit-config.yaml b/.pre-commit-config.yaml index 46d2371..5405cc8 100644 --- a/.pre-commit-config.yaml +++ b/.pre-commit-config.yaml @@ -25,7 +25,7 @@ repos: rev: 22.3.0 hooks: - id: black -- repo: https://gitlab.com/pycqa/flake8 +- repo: https://github.com/pycqa/flake8 rev: 3.9.2 hooks: - id: flake8 From 8742acf38f10929904a2bbc3543d6c56c803f7c7 Mon Sep 17 00:00:00 2001 From: "release-please[bot]" <55107282+release-please[bot]@users.noreply.github.com> Date: Thu, 15 Dec 2022 17:32:44 -0500 Subject: [PATCH 13/13] chore(main): release 2.10.0 (#345) * chore(main): release 2.10.0 * fix(deps): require google-api-core>=1.34.0,>=2.11.0 Co-authored-by: release-please[bot] <55107282+release-please[bot]@users.noreply.github.com> Co-authored-by: Anthonios Partheniou --- .release-please-manifest.json | 2 +- CHANGELOG.md | 22 +++++++++++++++++++ .../containeranalysis/gapic_version.py | 2 +- .../containeranalysis_v1/gapic_version.py | 2 +- ..._google.devtools.containeranalysis.v1.json | 2 +- setup.py | 2 +- testing/constraints-3.7.txt | 2 +- 7 files changed, 28 insertions(+), 6 deletions(-) diff --git a/.release-please-manifest.json b/.release-please-manifest.json index 3542679..f393718 100644 --- a/.release-please-manifest.json +++ b/.release-please-manifest.json @@ -1,3 +1,3 @@ { - ".": "2.9.3" + ".": "2.10.0" } diff --git a/CHANGELOG.md b/CHANGELOG.md index cb543bd..aec77a4 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,27 @@ # Changelog +## [2.10.0](https://github.com/googleapis/python-containeranalysis/compare/v2.9.3...v2.10.0) (2022-12-15) + + +### Features + +* Add support for `google.cloud.devtools.containeranalysis.__version__` ([3d91a44](https://github.com/googleapis/python-containeranalysis/commit/3d91a44b7152fba0508d4d3099e51f18348b0512)) +* Add typing to proto.Message based class attributes ([3d91a44](https://github.com/googleapis/python-containeranalysis/commit/3d91a44b7152fba0508d4d3099e51f18348b0512)) + + +### Bug Fixes + +* Add dict typing for client_options ([3d91a44](https://github.com/googleapis/python-containeranalysis/commit/3d91a44b7152fba0508d4d3099e51f18348b0512)) +* **deps:** Require google-api-core >=1.34.0, >=2.11.0 ([8a0a4b6](https://github.com/googleapis/python-containeranalysis/commit/8a0a4b635de8992b9f5984a337cabaf92017400a)) +* Drop usage of pkg_resources ([8a0a4b6](https://github.com/googleapis/python-containeranalysis/commit/8a0a4b635de8992b9f5984a337cabaf92017400a)) +* Fix timeout default values ([8a0a4b6](https://github.com/googleapis/python-containeranalysis/commit/8a0a4b635de8992b9f5984a337cabaf92017400a)) + + +### Documentation + +* **samples:** Snippetgen handling of repeated enum field ([3d91a44](https://github.com/googleapis/python-containeranalysis/commit/3d91a44b7152fba0508d4d3099e51f18348b0512)) +* **samples:** Snippetgen should call await on the operation coroutine before calling result ([8a0a4b6](https://github.com/googleapis/python-containeranalysis/commit/8a0a4b635de8992b9f5984a337cabaf92017400a)) + ## [2.9.3](https://github.com/googleapis/python-containeranalysis/compare/v2.9.2...v2.9.3) (2022-10-07) diff --git a/google/cloud/devtools/containeranalysis/gapic_version.py b/google/cloud/devtools/containeranalysis/gapic_version.py index 085af7e..00f0a8d 100644 --- a/google/cloud/devtools/containeranalysis/gapic_version.py +++ b/google/cloud/devtools/containeranalysis/gapic_version.py @@ -13,4 +13,4 @@ # See the License for the specific language governing permissions and # limitations under the License. # -__version__ = "2.9.3" # {x-release-please-version} +__version__ = "2.10.0" # {x-release-please-version} diff --git a/google/cloud/devtools/containeranalysis_v1/gapic_version.py b/google/cloud/devtools/containeranalysis_v1/gapic_version.py index 085af7e..00f0a8d 100644 --- a/google/cloud/devtools/containeranalysis_v1/gapic_version.py +++ b/google/cloud/devtools/containeranalysis_v1/gapic_version.py @@ -13,4 +13,4 @@ # See the License for the specific language governing permissions and # limitations under the License. # -__version__ = "2.9.3" # {x-release-please-version} +__version__ = "2.10.0" # {x-release-please-version} diff --git a/samples/generated_samples/snippet_metadata_google.devtools.containeranalysis.v1.json b/samples/generated_samples/snippet_metadata_google.devtools.containeranalysis.v1.json index dab3e1e..e012702 100644 --- a/samples/generated_samples/snippet_metadata_google.devtools.containeranalysis.v1.json +++ b/samples/generated_samples/snippet_metadata_google.devtools.containeranalysis.v1.json @@ -8,7 +8,7 @@ ], "language": "PYTHON", "name": "google-cloud-containeranalysis", - "version": "0.1.0" + "version": "2.10.0" }, "snippets": [ { diff --git a/setup.py b/setup.py index 16a187f..aac4cdc 100644 --- a/setup.py +++ b/setup.py @@ -40,7 +40,7 @@ release_status = "Development Status :: 5 - Production/Stable" dependencies = [ - "google-api-core[grpc] >= 1.33.2, <3.0.0dev,!=2.0.*,!=2.1.*,!=2.2.*,!=2.3.*,!=2.4.*,!=2.5.*,!=2.6.*,!=2.7.*", + "google-api-core[grpc] >= 1.34.0, <3.0.0dev,!=2.0.*,!=2.1.*,!=2.2.*,!=2.3.*,!=2.4.*,!=2.5.*,!=2.6.*,!=2.7.*,!=2.8.*,!=2.9.*,!=2.10.*", "proto-plus >= 1.22.0, <2.0.0dev", "protobuf>=3.19.5,<5.0.0dev,!=3.20.0,!=3.20.1,!=4.21.0,!=4.21.1,!=4.21.2,!=4.21.3,!=4.21.4,!=4.21.5", "grpc-google-iam-v1 >= 0.12.4, < 1.0.0dev", diff --git a/testing/constraints-3.7.txt b/testing/constraints-3.7.txt index 8995c3a..5bdb617 100644 --- a/testing/constraints-3.7.txt +++ b/testing/constraints-3.7.txt @@ -4,7 +4,7 @@ # Pin the version to the lower bound. # e.g., if setup.py has "google-cloud-foo >= 1.14.0, < 2.0.0dev", # Then this file should have google-cloud-foo==1.14.0 -google-api-core==1.33.2 +google-api-core==1.34.0 proto-plus==1.22.0 protobuf==3.19.5 grpc-google-iam-v1==0.12.4