diff --git a/.kokoro/requirements.txt b/.kokoro/requirements.txt index d15994bac9..31425f1647 100644 --- a/.kokoro/requirements.txt +++ b/.kokoro/requirements.txt @@ -20,9 +20,9 @@ cachetools==5.2.0 \ --hash=sha256:6a94c6402995a99c3970cc7e4884bb60b4a8639938157eeed436098bf9831757 \ --hash=sha256:f9f17d2aec496a9aa6b76f53e3b614c965223c061982d434d160f930c698a9db # via google-auth -certifi==2022.6.15 \ - --hash=sha256:84c85a9078b11105f04f3036a9482ae10e4621616db313fe045dd24743a0820d \ - --hash=sha256:fe86415d55e84719d75f8b69414f6438ac3547d2078ab91b67e779ef69378412 +certifi==2022.9.24 \ + --hash=sha256:0d9c601124e5a6ba9712dbc60d9c53c21e34f5f641fe83002317394311bdce14 \ + --hash=sha256:90c1a32f1d68f940488354e36370f6cca89f0f106db09518524c88d6ed83f382 # via requests cffi==1.15.1 \ --hash=sha256:00a9ed42e88df81ffae7a8ab6d9356b371399b91dbdf0c3cb1e84c03a13aceb5 \ @@ -110,29 +110,33 @@ commonmark==0.9.1 \ --hash=sha256:452f9dc859be7f06631ddcb328b6919c67984aca654e5fefb3914d54691aed60 \ --hash=sha256:da2f38c92590f83de410ba1a3cbceafbc74fee9def35f9251ba9a971d6d66fd9 # via rich -cryptography==37.0.4 \ - --hash=sha256:190f82f3e87033821828f60787cfa42bff98404483577b591429ed99bed39d59 \ - --hash=sha256:2be53f9f5505673eeda5f2736bea736c40f051a739bfae2f92d18aed1eb54596 \ - --hash=sha256:30788e070800fec9bbcf9faa71ea6d8068f5136f60029759fd8c3efec3c9dcb3 \ - --hash=sha256:3d41b965b3380f10e4611dbae366f6dc3cefc7c9ac4e8842a806b9672ae9add5 \ - --hash=sha256:4c590ec31550a724ef893c50f9a97a0c14e9c851c85621c5650d699a7b88f7ab \ - --hash=sha256:549153378611c0cca1042f20fd9c5030d37a72f634c9326e225c9f666d472884 \ - --hash=sha256:63f9c17c0e2474ccbebc9302ce2f07b55b3b3fcb211ded18a42d5764f5c10a82 \ - --hash=sha256:6bc95ed67b6741b2607298f9ea4932ff157e570ef456ef7ff0ef4884a134cc4b \ - --hash=sha256:7099a8d55cd49b737ffc99c17de504f2257e3787e02abe6d1a6d136574873441 \ - --hash=sha256:75976c217f10d48a8b5a8de3d70c454c249e4b91851f6838a4e48b8f41eb71aa \ - --hash=sha256:7bc997818309f56c0038a33b8da5c0bfbb3f1f067f315f9abd6fc07ad359398d \ - --hash=sha256:80f49023dd13ba35f7c34072fa17f604d2f19bf0989f292cedf7ab5770b87a0b \ - --hash=sha256:91ce48d35f4e3d3f1d83e29ef4a9267246e6a3be51864a5b7d2247d5086fa99a \ - --hash=sha256:a958c52505c8adf0d3822703078580d2c0456dd1d27fabfb6f76fe63d2971cd6 \ - --hash=sha256:b62439d7cd1222f3da897e9a9fe53bbf5c104fff4d60893ad1355d4c14a24157 \ - --hash=sha256:b7f8dd0d4c1f21759695c05a5ec8536c12f31611541f8904083f3dc582604280 \ - --hash=sha256:d204833f3c8a33bbe11eda63a54b1aad7aa7456ed769a982f21ec599ba5fa282 \ - --hash=sha256:e007f052ed10cc316df59bc90fbb7ff7950d7e2919c9757fd42a2b8ecf8a5f67 \ - --hash=sha256:f2dcb0b3b63afb6df7fd94ec6fbddac81b5492513f7b0436210d390c14d46ee8 \ - --hash=sha256:f721d1885ecae9078c3f6bbe8a88bc0786b6e749bf32ccec1ef2b18929a05046 \ - --hash=sha256:f7a6de3e98771e183645181b3627e2563dcde3ce94a9e42a3f427d2255190327 \ - --hash=sha256:f8c0a6e9e1dd3eb0414ba320f85da6b0dcbd543126e30fcc546e7372a7fbf3b9 +cryptography==38.0.3 \ + --hash=sha256:068147f32fa662c81aebab95c74679b401b12b57494872886eb5c1139250ec5d \ + --hash=sha256:06fc3cc7b6f6cca87bd56ec80a580c88f1da5306f505876a71c8cfa7050257dd \ + --hash=sha256:25c1d1f19729fb09d42e06b4bf9895212292cb27bb50229f5aa64d039ab29146 \ + --hash=sha256:402852a0aea73833d982cabb6d0c3bb582c15483d29fb7085ef2c42bfa7e38d7 \ + --hash=sha256:4e269dcd9b102c5a3d72be3c45d8ce20377b8076a43cbed6f660a1afe365e436 \ + --hash=sha256:5419a127426084933076132d317911e3c6eb77568a1ce23c3ac1e12d111e61e0 \ + --hash=sha256:554bec92ee7d1e9d10ded2f7e92a5d70c1f74ba9524947c0ba0c850c7b011828 \ + --hash=sha256:5e89468fbd2fcd733b5899333bc54d0d06c80e04cd23d8c6f3e0542358c6060b \ + --hash=sha256:65535bc550b70bd6271984d9863a37741352b4aad6fb1b3344a54e6950249b55 \ + --hash=sha256:6ab9516b85bebe7aa83f309bacc5f44a61eeb90d0b4ec125d2d003ce41932d36 \ + --hash=sha256:6addc3b6d593cd980989261dc1cce38263c76954d758c3c94de51f1e010c9a50 \ + --hash=sha256:728f2694fa743a996d7784a6194da430f197d5c58e2f4e278612b359f455e4a2 \ + --hash=sha256:785e4056b5a8b28f05a533fab69febf5004458e20dad7e2e13a3120d8ecec75a \ + --hash=sha256:78cf5eefac2b52c10398a42765bfa981ce2372cbc0457e6bf9658f41ec3c41d8 \ + --hash=sha256:7f836217000342d448e1c9a342e9163149e45d5b5eca76a30e84503a5a96cab0 \ + --hash=sha256:8d41a46251bf0634e21fac50ffd643216ccecfaf3701a063257fe0b2be1b6548 \ + --hash=sha256:984fe150f350a3c91e84de405fe49e688aa6092b3525f407a18b9646f6612320 \ + --hash=sha256:9b24bcff7853ed18a63cfb0c2b008936a9554af24af2fb146e16d8e1aed75748 \ + --hash=sha256:b1b35d9d3a65542ed2e9d90115dfd16bbc027b3f07ee3304fc83580f26e43249 \ + --hash=sha256:b1b52c9e5f8aa2b802d48bd693190341fae201ea51c7a167d69fc48b60e8a959 \ + --hash=sha256:bbf203f1a814007ce24bd4d51362991d5cb90ba0c177a9c08825f2cc304d871f \ + --hash=sha256:be243c7e2bfcf6cc4cb350c0d5cdf15ca6383bbcb2a8ef51d3c9411a9d4386f0 \ + --hash=sha256:bfbe6ee19615b07a98b1d2287d6a6073f734735b49ee45b11324d85efc4d5cbd \ + --hash=sha256:c46837ea467ed1efea562bbeb543994c2d1f6e800785bd5a2c98bc096f5cb220 \ + --hash=sha256:dfb4f4dd568de1b6af9f4cda334adf7d72cf5bc052516e1b2608b683375dd95c \ + --hash=sha256:ed7b00096790213e09eb11c97cc6e2b757f15f3d2f85833cd2d3ec3fe37c1722 # via # gcp-releasetool # secretstorage @@ -148,23 +152,23 @@ filelock==3.8.0 \ --hash=sha256:55447caa666f2198c5b6b13a26d2084d26fa5b115c00d065664b2124680c4edc \ --hash=sha256:617eb4e5eedc82fc5f47b6d61e4d11cb837c56cb4544e39081099fa17ad109d4 # via virtualenv -gcp-docuploader==0.6.3 \ - --hash=sha256:ba8c9d76b3bbac54b0311c503a373b00edc2dc02d6d54ea9507045adb8e870f7 \ - --hash=sha256:c0f5aaa82ce1854a386197e4e359b120ad6d4e57ae2c812fce42219a3288026b +gcp-docuploader==0.6.4 \ + --hash=sha256:01486419e24633af78fd0167db74a2763974765ee8078ca6eb6964d0ebd388af \ + --hash=sha256:70861190c123d907b3b067da896265ead2eeb9263969d6955c9e0bb091b5ccbf # via -r requirements.in -gcp-releasetool==1.8.7 \ - --hash=sha256:3d2a67c9db39322194afb3b427e9cb0476ce8f2a04033695f0aeb63979fc2b37 \ - --hash=sha256:5e4d28f66e90780d77f3ecf1e9155852b0c3b13cbccb08ab07e66b2357c8da8d +gcp-releasetool==1.9.1 \ + --hash=sha256:952f4055d5d986b070ae2a71c4410b250000f9cc5a1e26398fcd55a5bbc5a15f \ + --hash=sha256:d0d3c814a97c1a237517e837d8cfa668ced8df4b882452578ecef4a4e79c583b # via -r requirements.in -google-api-core==2.8.2 \ - --hash=sha256:06f7244c640322b508b125903bb5701bebabce8832f85aba9335ec00b3d02edc \ - --hash=sha256:93c6a91ccac79079ac6bbf8b74ee75db970cc899278b97d53bc012f35908cf50 +google-api-core==2.10.2 \ + --hash=sha256:10c06f7739fe57781f87523375e8e1a3a4674bf6392cd6131a3222182b971320 \ + --hash=sha256:34f24bd1d5f72a8c4519773d99ca6bf080a6c4e041b4e9f024fe230191dda62e # via # google-cloud-core # google-cloud-storage -google-auth==2.11.0 \ - --hash=sha256:be62acaae38d0049c21ca90f27a23847245c9f161ff54ede13af2cb6afecbac9 \ - --hash=sha256:ed65ecf9f681832298e29328e1ef0a3676e3732b2e56f41532d45f70a22de0fb +google-auth==2.14.0 \ + --hash=sha256:1ad5b0e6eba5f69645971abb3d2c197537d5914070a8c6d30299dfdb07c5c700 \ + --hash=sha256:cf24817855d874ede2efd071aa22125445f555de1685b739a9782fcf408c2a3d # via # gcp-releasetool # google-api-core @@ -178,72 +182,97 @@ google-cloud-storage==2.5.0 \ --hash=sha256:19a26c66c317ce542cea0830b7e787e8dac2588b6bfa4d3fd3b871ba16305ab0 \ --hash=sha256:382f34b91de2212e3c2e7b40ec079d27ee2e3dbbae99b75b1bcd8c63063ce235 # via gcp-docuploader -google-crc32c==1.3.0 \ - --hash=sha256:04e7c220798a72fd0f08242bc8d7a05986b2a08a0573396187fd32c1dcdd58b3 \ - --hash=sha256:05340b60bf05b574159e9bd940152a47d38af3fb43803ffe71f11d704b7696a6 \ - --hash=sha256:12674a4c3b56b706153a358eaa1018c4137a5a04635b92b4652440d3d7386206 \ - --hash=sha256:127f9cc3ac41b6a859bd9dc4321097b1a4f6aa7fdf71b4f9227b9e3ebffb4422 \ - --hash=sha256:13af315c3a0eec8bb8b8d80b8b128cb3fcd17d7e4edafc39647846345a3f003a \ - --hash=sha256:1926fd8de0acb9d15ee757175ce7242e235482a783cd4ec711cc999fc103c24e \ - --hash=sha256:226f2f9b8e128a6ca6a9af9b9e8384f7b53a801907425c9a292553a3a7218ce0 \ - --hash=sha256:276de6273eb074a35bc598f8efbc00c7869c5cf2e29c90748fccc8c898c244df \ - --hash=sha256:318f73f5484b5671f0c7f5f63741ab020a599504ed81d209b5c7129ee4667407 \ - --hash=sha256:3bbce1be3687bbfebe29abdb7631b83e6b25da3f4e1856a1611eb21854b689ea \ - --hash=sha256:42ae4781333e331a1743445931b08ebdad73e188fd554259e772556fc4937c48 \ - --hash=sha256:58be56ae0529c664cc04a9c76e68bb92b091e0194d6e3c50bea7e0f266f73713 \ - --hash=sha256:5da2c81575cc3ccf05d9830f9e8d3c70954819ca9a63828210498c0774fda1a3 \ - --hash=sha256:6311853aa2bba4064d0c28ca54e7b50c4d48e3de04f6770f6c60ebda1e975267 \ - --hash=sha256:650e2917660e696041ab3dcd7abac160b4121cd9a484c08406f24c5964099829 \ - --hash=sha256:6a4db36f9721fdf391646685ecffa404eb986cbe007a3289499020daf72e88a2 \ - --hash=sha256:779cbf1ce375b96111db98fca913c1f5ec11b1d870e529b1dc7354b2681a8c3a \ - --hash=sha256:7f6fe42536d9dcd3e2ffb9d3053f5d05221ae3bbcefbe472bdf2c71c793e3183 \ - --hash=sha256:891f712ce54e0d631370e1f4997b3f182f3368179198efc30d477c75d1f44942 \ - --hash=sha256:95c68a4b9b7828ba0428f8f7e3109c5d476ca44996ed9a5f8aac6269296e2d59 \ - --hash=sha256:96a8918a78d5d64e07c8ea4ed2bc44354e3f93f46a4866a40e8db934e4c0d74b \ - --hash=sha256:9c3cf890c3c0ecfe1510a452a165431b5831e24160c5fcf2071f0f85ca5a47cd \ - --hash=sha256:9f58099ad7affc0754ae42e6d87443299f15d739b0ce03c76f515153a5cda06c \ - --hash=sha256:a0b9e622c3b2b8d0ce32f77eba617ab0d6768b82836391e4f8f9e2074582bf02 \ - --hash=sha256:a7f9cbea4245ee36190f85fe1814e2d7b1e5f2186381b082f5d59f99b7f11328 \ - --hash=sha256:bab4aebd525218bab4ee615786c4581952eadc16b1ff031813a2fd51f0cc7b08 \ - --hash=sha256:c124b8c8779bf2d35d9b721e52d4adb41c9bfbde45e6a3f25f0820caa9aba73f \ - --hash=sha256:c9da0a39b53d2fab3e5467329ed50e951eb91386e9d0d5b12daf593973c3b168 \ - --hash=sha256:ca60076c388728d3b6ac3846842474f4250c91efbfe5afa872d3ffd69dd4b318 \ - --hash=sha256:cb6994fff247987c66a8a4e550ef374671c2b82e3c0d2115e689d21e511a652d \ - --hash=sha256:d1c1d6236feab51200272d79b3d3e0f12cf2cbb12b208c835b175a21efdb0a73 \ - --hash=sha256:dd7760a88a8d3d705ff562aa93f8445ead54f58fd482e4f9e2bafb7e177375d4 \ - --hash=sha256:dda4d8a3bb0b50f540f6ff4b6033f3a74e8bf0bd5320b70fab2c03e512a62812 \ - --hash=sha256:e0f1ff55dde0ebcfbef027edc21f71c205845585fffe30d4ec4979416613e9b3 \ - --hash=sha256:e7a539b9be7b9c00f11ef16b55486141bc2cdb0c54762f84e3c6fc091917436d \ - --hash=sha256:eb0b14523758e37802f27b7f8cd973f5f3d33be7613952c0df904b68c4842f0e \ - --hash=sha256:ed447680ff21c14aaceb6a9f99a5f639f583ccfe4ce1a5e1d48eb41c3d6b3217 \ - --hash=sha256:f52a4ad2568314ee713715b1e2d79ab55fab11e8b304fd1462ff5cccf4264b3e \ - --hash=sha256:fbd60c6aaa07c31d7754edbc2334aef50601b7f1ada67a96eb1eb57c7c72378f \ - --hash=sha256:fc28e0db232c62ca0c3600884933178f0825c99be4474cdd645e378a10588125 \ - --hash=sha256:fe31de3002e7b08eb20823b3735b97c86c5926dd0581c7710a680b418a8709d4 \ - --hash=sha256:fec221a051150eeddfdfcff162e6db92c65ecf46cb0f7bb1bf812a1520ec026b \ - --hash=sha256:ff71073ebf0e42258a42a0b34f2c09ec384977e7f6808999102eedd5b49920e3 +google-crc32c==1.5.0 \ + --hash=sha256:024894d9d3cfbc5943f8f230e23950cd4906b2fe004c72e29b209420a1e6b05a \ + --hash=sha256:02c65b9817512edc6a4ae7c7e987fea799d2e0ee40c53ec573a692bee24de876 \ + --hash=sha256:02ebb8bf46c13e36998aeaad1de9b48f4caf545e91d14041270d9dca767b780c \ + --hash=sha256:07eb3c611ce363c51a933bf6bd7f8e3878a51d124acfc89452a75120bc436289 \ + --hash=sha256:1034d91442ead5a95b5aaef90dbfaca8633b0247d1e41621d1e9f9db88c36298 \ + --hash=sha256:116a7c3c616dd14a3de8c64a965828b197e5f2d121fedd2f8c5585c547e87b02 \ + --hash=sha256:19e0a019d2c4dcc5e598cd4a4bc7b008546b0358bd322537c74ad47a5386884f \ + --hash=sha256:1c7abdac90433b09bad6c43a43af253e688c9cfc1c86d332aed13f9a7c7f65e2 \ + --hash=sha256:1e986b206dae4476f41bcec1faa057851f3889503a70e1bdb2378d406223994a \ + --hash=sha256:272d3892a1e1a2dbc39cc5cde96834c236d5327e2122d3aaa19f6614531bb6eb \ + --hash=sha256:278d2ed7c16cfc075c91378c4f47924c0625f5fc84b2d50d921b18b7975bd210 \ + --hash=sha256:2ad40e31093a4af319dadf503b2467ccdc8f67c72e4bcba97f8c10cb078207b5 \ + --hash=sha256:2e920d506ec85eb4ba50cd4228c2bec05642894d4c73c59b3a2fe20346bd00ee \ + --hash=sha256:3359fc442a743e870f4588fcf5dcbc1bf929df1fad8fb9905cd94e5edb02e84c \ + --hash=sha256:37933ec6e693e51a5b07505bd05de57eee12f3e8c32b07da7e73669398e6630a \ + --hash=sha256:398af5e3ba9cf768787eef45c803ff9614cc3e22a5b2f7d7ae116df8b11e3314 \ + --hash=sha256:3b747a674c20a67343cb61d43fdd9207ce5da6a99f629c6e2541aa0e89215bcd \ + --hash=sha256:461665ff58895f508e2866824a47bdee72497b091c730071f2b7575d5762ab65 \ + --hash=sha256:4c6fdd4fccbec90cc8a01fc00773fcd5fa28db683c116ee3cb35cd5da9ef6c37 \ + --hash=sha256:5829b792bf5822fd0a6f6eb34c5f81dd074f01d570ed7f36aa101d6fc7a0a6e4 \ + --hash=sha256:596d1f98fc70232fcb6590c439f43b350cb762fb5d61ce7b0e9db4539654cc13 \ + --hash=sha256:5ae44e10a8e3407dbe138984f21e536583f2bba1be9491239f942c2464ac0894 \ + --hash=sha256:635f5d4dd18758a1fbd1049a8e8d2fee4ffed124462d837d1a02a0e009c3ab31 \ + --hash=sha256:64e52e2b3970bd891309c113b54cf0e4384762c934d5ae56e283f9a0afcd953e \ + --hash=sha256:66741ef4ee08ea0b2cc3c86916ab66b6aef03768525627fd6a1b34968b4e3709 \ + --hash=sha256:67b741654b851abafb7bc625b6d1cdd520a379074e64b6a128e3b688c3c04740 \ + --hash=sha256:6ac08d24c1f16bd2bf5eca8eaf8304812f44af5cfe5062006ec676e7e1d50afc \ + --hash=sha256:6f998db4e71b645350b9ac28a2167e6632c239963ca9da411523bb439c5c514d \ + --hash=sha256:72218785ce41b9cfd2fc1d6a017dc1ff7acfc4c17d01053265c41a2c0cc39b8c \ + --hash=sha256:74dea7751d98034887dbd821b7aae3e1d36eda111d6ca36c206c44478035709c \ + --hash=sha256:759ce4851a4bb15ecabae28f4d2e18983c244eddd767f560165563bf9aefbc8d \ + --hash=sha256:77e2fd3057c9d78e225fa0a2160f96b64a824de17840351b26825b0848022906 \ + --hash=sha256:7c074fece789b5034b9b1404a1f8208fc2d4c6ce9decdd16e8220c5a793e6f61 \ + --hash=sha256:7c42c70cd1d362284289c6273adda4c6af8039a8ae12dc451dcd61cdabb8ab57 \ + --hash=sha256:7f57f14606cd1dd0f0de396e1e53824c371e9544a822648cd76c034d209b559c \ + --hash=sha256:83c681c526a3439b5cf94f7420471705bbf96262f49a6fe546a6db5f687a3d4a \ + --hash=sha256:8485b340a6a9e76c62a7dce3c98e5f102c9219f4cfbf896a00cf48caf078d438 \ + --hash=sha256:84e6e8cd997930fc66d5bb4fde61e2b62ba19d62b7abd7a69920406f9ecca946 \ + --hash=sha256:89284716bc6a5a415d4eaa11b1726d2d60a0cd12aadf5439828353662ede9dd7 \ + --hash=sha256:8b87e1a59c38f275c0e3676fc2ab6d59eccecfd460be267ac360cc31f7bcde96 \ + --hash=sha256:8f24ed114432de109aa9fd317278518a5af2d31ac2ea6b952b2f7782b43da091 \ + --hash=sha256:98cb4d057f285bd80d8778ebc4fde6b4d509ac3f331758fb1528b733215443ae \ + --hash=sha256:998679bf62b7fb599d2878aa3ed06b9ce688b8974893e7223c60db155f26bd8d \ + --hash=sha256:9ba053c5f50430a3fcfd36f75aff9caeba0440b2d076afdb79a318d6ca245f88 \ + --hash=sha256:9c99616c853bb585301df6de07ca2cadad344fd1ada6d62bb30aec05219c45d2 \ + --hash=sha256:a1fd716e7a01f8e717490fbe2e431d2905ab8aa598b9b12f8d10abebb36b04dd \ + --hash=sha256:a2355cba1f4ad8b6988a4ca3feed5bff33f6af2d7f134852cf279c2aebfde541 \ + --hash=sha256:b1f8133c9a275df5613a451e73f36c2aea4fe13c5c8997e22cf355ebd7bd0728 \ + --hash=sha256:b8667b48e7a7ef66afba2c81e1094ef526388d35b873966d8a9a447974ed9178 \ + --hash=sha256:ba1eb1843304b1e5537e1fca632fa894d6f6deca8d6389636ee5b4797affb968 \ + --hash=sha256:be82c3c8cfb15b30f36768797a640e800513793d6ae1724aaaafe5bf86f8f346 \ + --hash=sha256:c02ec1c5856179f171e032a31d6f8bf84e5a75c45c33b2e20a3de353b266ebd8 \ + --hash=sha256:c672d99a345849301784604bfeaeba4db0c7aae50b95be04dd651fd2a7310b93 \ + --hash=sha256:c6c777a480337ac14f38564ac88ae82d4cd238bf293f0a22295b66eb89ffced7 \ + --hash=sha256:cae0274952c079886567f3f4f685bcaf5708f0a23a5f5216fdab71f81a6c0273 \ + --hash=sha256:cd67cf24a553339d5062eff51013780a00d6f97a39ca062781d06b3a73b15462 \ + --hash=sha256:d3515f198eaa2f0ed49f8819d5732d70698c3fa37384146079b3799b97667a94 \ + --hash=sha256:d5280312b9af0976231f9e317c20e4a61cd2f9629b7bfea6a693d1878a264ebd \ + --hash=sha256:de06adc872bcd8c2a4e0dc51250e9e65ef2ca91be023b9d13ebd67c2ba552e1e \ + --hash=sha256:e1674e4307fa3024fc897ca774e9c7562c957af85df55efe2988ed9056dc4e57 \ + --hash=sha256:e2096eddb4e7c7bdae4bd69ad364e55e07b8316653234a56552d9c988bd2d61b \ + --hash=sha256:e560628513ed34759456a416bf86b54b2476c59144a9138165c9a1575801d0d9 \ + --hash=sha256:edfedb64740750e1a3b16152620220f51d58ff1b4abceb339ca92e934775c27a \ + --hash=sha256:f13cae8cc389a440def0c8c52057f37359014ccbc9dc1f0827936bcd367c6100 \ + --hash=sha256:f314013e7dcd5cf45ab1945d92e713eec788166262ae8deb2cfacd53def27325 \ + --hash=sha256:f583edb943cf2e09c60441b910d6a20b4d9d626c75a36c8fcac01a6c96c01183 \ + --hash=sha256:fd8536e902db7e365f49e7d9029283403974ccf29b13fc7028b97e2295b33556 \ + --hash=sha256:fe70e325aa68fa4b5edf7d1a4b6f691eb04bbccac0ace68e34820d283b5f80d4 # via google-resumable-media -google-resumable-media==2.3.3 \ - --hash=sha256:27c52620bd364d1c8116eaac4ea2afcbfb81ae9139fb3199652fcac1724bfb6c \ - --hash=sha256:5b52774ea7a829a8cdaa8bd2d4c3d4bc660c91b30857ab2668d0eb830f4ea8c5 +google-resumable-media==2.4.0 \ + --hash=sha256:2aa004c16d295c8f6c33b2b4788ba59d366677c0a25ae7382436cb30f776deaa \ + --hash=sha256:8d5518502f92b9ecc84ac46779bd4f09694ecb3ba38a3e7ca737a86d15cbca1f # via google-cloud-storage googleapis-common-protos==1.56.4 \ --hash=sha256:8eb2cbc91b69feaf23e32452a7ae60e791e09967d81d4fcc7fc388182d1bd394 \ --hash=sha256:c25873c47279387cfdcbdafa36149887901d36202cb645a0e4f29686bf6e4417 # via google-api-core -idna==3.3 \ - --hash=sha256:84d9dd047ffa80596e0f246e2eab0b391788b0503584e8945f2368256d2735ff \ - --hash=sha256:9d643ff0a55b762d5cdb124b8eaa99c66322e2157b69160bc32796e824360e6d +idna==3.4 \ + --hash=sha256:814f528e8dead7d329833b91c5faa87d60bf71824cd12a7530b5526063d02cb4 \ + --hash=sha256:90b77e79eaa3eba6de819a0c442c0b4ceefc341a7a2ab77d7562bf49f425c5c2 # via requests -importlib-metadata==4.12.0 \ - --hash=sha256:637245b8bab2b6502fcbc752cc4b7a6f6243bb02b31c5c26156ad103d3d45670 \ - --hash=sha256:7401a975809ea1fdc658c3aa4f78cc2195a0e019c5cbc4c06122884e9ae80c23 +importlib-metadata==5.0.0 \ + --hash=sha256:da31db32b304314d044d3c12c79bd59e307889b287ad12ff387b3500835fc2ab \ + --hash=sha256:ddb0e35065e8938f867ed4928d0ae5bf2a53b7773871bfe6bcc7e4fcdc7dea43 # via # -r requirements.in # twine -jaraco-classes==3.2.2 \ - --hash=sha256:6745f113b0b588239ceb49532aa09c3ebb947433ce311ef2f8e3ad64ebb74594 \ - --hash=sha256:e6ef6fd3fcf4579a7a019d87d1e56a883f4e4c35cfe925f86731abc58804e647 +jaraco-classes==3.2.3 \ + --hash=sha256:2353de3288bc6b82120752201c6b1c1a14b058267fa424ed5ce5984e3b922158 \ + --hash=sha256:89559fa5c1d3c34eff6f631ad80bb21f378dbcbb35dd161fd2c6b93f5be2f98a # via keyring jeepney==0.8.0 \ --hash=sha256:5efe48d255973902f6badc3ce55e2aa6c5c3b3bc642059ef3a91247bcfcc5806 \ @@ -255,9 +284,9 @@ jinja2==3.1.2 \ --hash=sha256:31351a702a408a9e7595a8fc6150fc3f43bb6bf7e319770cbc0db9df9437e852 \ --hash=sha256:6088930bfe239f0e6710546ab9c19c9ef35e29792895fed6e6e31a023a182a61 # via gcp-releasetool -keyring==23.9.0 \ - --hash=sha256:4c32a31174faaee48f43a7e2c7e9c3216ec5e95acf22a2bebfb4a1d05056ee44 \ - --hash=sha256:98f060ec95ada2ab910c195a2d4317be6ef87936a766b239c46aa3c7aac4f0db +keyring==23.9.3 \ + --hash=sha256:69732a15cb1433bdfbc3b980a8a36a04878a6cfd7cb99f497b573f31618001c0 \ + --hash=sha256:69b01dd83c42f590250fe7a1f503fc229b14de83857314b1933a3ddbf595c4a5 # via # gcp-releasetool # twine @@ -303,9 +332,9 @@ markupsafe==2.1.1 \ --hash=sha256:f121a1420d4e173a5d96e47e9a0c0dcff965afdf1626d28de1460815f7c4ee7a \ --hash=sha256:fc7b548b17d238737688817ab67deebb30e8073c95749d55538ed473130ec0c7 # via jinja2 -more-itertools==8.14.0 \ - --hash=sha256:1bc4f91ee5b1b31ac7ceacc17c09befe6a40a503907baf9c839c229b5095cfd2 \ - --hash=sha256:c09443cd3d5438b8dafccd867a6bc1cb0894389e90cb53d227456b0b0bccb750 +more-itertools==9.0.0 \ + --hash=sha256:250e83d7e81d0c87ca6bd942e6aeab8cc9daa6096d12c5308f3f92fa5e5c1f41 \ + --hash=sha256:5a6257e40878ef0520b1803990e3e22303a41b5714006c32a3fd8304b26ea1ab # via jaraco-classes nox==2022.8.7 \ --hash=sha256:1b894940551dc5c389f9271d197ca5d655d40bdc6ccf93ed6880e4042760a34b \ @@ -325,34 +354,34 @@ platformdirs==2.5.2 \ --hash=sha256:027d8e83a2d7de06bbac4e5ef7e023c02b863d7ea5d079477e722bb41ab25788 \ --hash=sha256:58c8abb07dcb441e6ee4b11d8df0ac856038f944ab98b7be6b27b2a3c7feef19 # via virtualenv -protobuf==3.20.2 \ - --hash=sha256:03d76b7bd42ac4a6e109742a4edf81ffe26ffd87c5993126d894fe48a120396a \ - --hash=sha256:09e25909c4297d71d97612f04f41cea8fa8510096864f2835ad2f3b3df5a5559 \ - --hash=sha256:18e34a10ae10d458b027d7638a599c964b030c1739ebd035a1dfc0e22baa3bfe \ - --hash=sha256:291fb4307094bf5ccc29f424b42268640e00d5240bf0d9b86bf3079f7576474d \ - --hash=sha256:2c0b040d0b5d5d207936ca2d02f00f765906622c07d3fa19c23a16a8ca71873f \ - --hash=sha256:384164994727f274cc34b8abd41a9e7e0562801361ee77437099ff6dfedd024b \ - --hash=sha256:3cb608e5a0eb61b8e00fe641d9f0282cd0eedb603be372f91f163cbfbca0ded0 \ - --hash=sha256:5d9402bf27d11e37801d1743eada54372f986a372ec9679673bfcc5c60441151 \ - --hash=sha256:712dca319eee507a1e7df3591e639a2b112a2f4a62d40fe7832a16fd19151750 \ - --hash=sha256:7a5037af4e76c975b88c3becdf53922b5ffa3f2cddf657574a4920a3b33b80f3 \ - --hash=sha256:8228e56a865c27163d5d1d1771d94b98194aa6917bcfb6ce139cbfa8e3c27334 \ - --hash=sha256:84a1544252a933ef07bb0b5ef13afe7c36232a774affa673fc3636f7cee1db6c \ - --hash=sha256:84fe5953b18a383fd4495d375fe16e1e55e0a3afe7b4f7b4d01a3a0649fcda9d \ - --hash=sha256:9c673c8bfdf52f903081816b9e0e612186684f4eb4c17eeb729133022d6032e3 \ - --hash=sha256:9f876a69ca55aed879b43c295a328970306e8e80a263ec91cf6e9189243c613b \ - --hash=sha256:a9e5ae5a8e8985c67e8944c23035a0dff2c26b0f5070b2f55b217a1c33bbe8b1 \ - --hash=sha256:b4fdb29c5a7406e3f7ef176b2a7079baa68b5b854f364c21abe327bbeec01cdb \ - --hash=sha256:c184485e0dfba4dfd451c3bd348c2e685d6523543a0f91b9fd4ae90eb09e8422 \ - --hash=sha256:c9cdf251c582c16fd6a9f5e95836c90828d51b0069ad22f463761d27c6c19019 \ - --hash=sha256:e39cf61bb8582bda88cdfebc0db163b774e7e03364bbf9ce1ead13863e81e359 \ - --hash=sha256:e8fbc522303e09036c752a0afcc5c0603e917222d8bedc02813fd73b4b4ed804 \ - --hash=sha256:f34464ab1207114e73bba0794d1257c150a2b89b7a9faf504e00af7c9fd58978 \ - --hash=sha256:f52dabc96ca99ebd2169dadbe018824ebda08a795c7684a0b7d203a290f3adb0 +protobuf==3.20.3 \ + --hash=sha256:03038ac1cfbc41aa21f6afcbcd357281d7521b4157926f30ebecc8d4ea59dcb7 \ + --hash=sha256:28545383d61f55b57cf4df63eebd9827754fd2dc25f80c5253f9184235db242c \ + --hash=sha256:2e3427429c9cffebf259491be0af70189607f365c2f41c7c3764af6f337105f2 \ + --hash=sha256:398a9e0c3eaceb34ec1aee71894ca3299605fa8e761544934378bbc6c97de23b \ + --hash=sha256:44246bab5dd4b7fbd3c0c80b6f16686808fab0e4aca819ade6e8d294a29c7050 \ + --hash=sha256:447d43819997825d4e71bf5769d869b968ce96848b6479397e29fc24c4a5dfe9 \ + --hash=sha256:67a3598f0a2dcbc58d02dd1928544e7d88f764b47d4a286202913f0b2801c2e7 \ + --hash=sha256:74480f79a023f90dc6e18febbf7b8bac7508420f2006fabd512013c0c238f454 \ + --hash=sha256:819559cafa1a373b7096a482b504ae8a857c89593cf3a25af743ac9ecbd23480 \ + --hash=sha256:899dc660cd599d7352d6f10d83c95df430a38b410c1b66b407a6b29265d66469 \ + --hash=sha256:8c0c984a1b8fef4086329ff8dd19ac77576b384079247c770f29cc8ce3afa06c \ + --hash=sha256:9aae4406ea63d825636cc11ffb34ad3379335803216ee3a856787bcf5ccc751e \ + --hash=sha256:a7ca6d488aa8ff7f329d4c545b2dbad8ac31464f1d8b1c87ad1346717731e4db \ + --hash=sha256:b6cc7ba72a8850621bfec987cb72623e703b7fe2b9127a161ce61e61558ad905 \ + --hash=sha256:bf01b5720be110540be4286e791db73f84a2b721072a3711efff6c324cdf074b \ + --hash=sha256:c02ce36ec760252242a33967d51c289fd0e1c0e6e5cc9397e2279177716add86 \ + --hash=sha256:d9e4432ff660d67d775c66ac42a67cf2453c27cb4d738fc22cb53b5d84c135d4 \ + --hash=sha256:daa564862dd0d39c00f8086f88700fdbe8bc717e993a21e90711acfed02f2402 \ + --hash=sha256:de78575669dddf6099a8a0f46a27e82a1783c557ccc38ee620ed8cc96d3be7d7 \ + --hash=sha256:e64857f395505ebf3d2569935506ae0dfc4a15cb80dc25261176c784662cdcc4 \ + --hash=sha256:f4bd856d702e5b0d96a00ec6b307b0f51c1982c2bf9c0052cf9019e9a544ba99 \ + --hash=sha256:f4c42102bc82a51108e449cbb32b19b180022941c727bac0cfd50170341f16ee # via # gcp-docuploader # gcp-releasetool # google-api-core + # googleapis-common-protos py==1.11.0 \ --hash=sha256:51c75c4126074b472f746a24399ad32f6053d1b34b68d2fa41e558e6f4a98719 \ --hash=sha256:607c53218732647dff4acdfcd50cb62615cedf612e72d1724fb1a0cc6405b378 @@ -377,9 +406,9 @@ pygments==2.13.0 \ # via # readme-renderer # rich -pyjwt==2.4.0 \ - --hash=sha256:72d1d253f32dbd4f5c88eaf1fdc62f3a19f676ccbadb9dbc5d07e951b2b26daf \ - --hash=sha256:d42908208c699b3b973cbeb01a969ba6a96c821eefb1c5bfe4c390c01d67abba +pyjwt==2.6.0 \ + --hash=sha256:69285c7e31fc44f68a1feb309e948e0df53259d579295e6cfe2b1792329f05fd \ + --hash=sha256:d83c3d892a77bbb74d3e1a2cfa90afaadb60945205d1095d9221f04466f64c14 # via gcp-releasetool pyparsing==3.0.9 \ --hash=sha256:2b020ecf7d21b687f219b71ecad3631f644a47f01403fa1d1036b0c6416d70fb \ @@ -392,9 +421,9 @@ python-dateutil==2.8.2 \ --hash=sha256:0123cacc1627ae19ddf3c27a5de5bd67ee4586fbdd6440d9748f8abb483d3e86 \ --hash=sha256:961d03dc3453ebbc59dbdea9e4e11c5651520a876d0f4db161e8674aae935da9 # via gcp-releasetool -readme-renderer==37.0 \ - --hash=sha256:07b7ea234e03e58f77cc222e206e6abb8f4c0435becce5104794ee591f9301c5 \ - --hash=sha256:9fa416704703e509eeb900696751c908ddeb2011319d93700d8f18baff887a69 +readme-renderer==37.3 \ + --hash=sha256:cd653186dfc73055656f090f227f5cb22a046d7f71a841dfa305f55c9a513273 \ + --hash=sha256:f67a16caedfa71eef48a31b39708637a6f4664c4394801a7b0d6432d13907343 # via twine requests==2.28.1 \ --hash=sha256:7c5599b102feddaa661c826c56ab4fee28bfd17f5abca1ebbe3e7f19d7c97983 \ @@ -405,17 +434,17 @@ requests==2.28.1 \ # google-cloud-storage # requests-toolbelt # twine -requests-toolbelt==0.9.1 \ - --hash=sha256:380606e1d10dc85c3bd47bf5a6095f815ec007be7a8b69c878507068df059e6f \ - --hash=sha256:968089d4584ad4ad7c171454f0a5c6dac23971e9472521ea3b6d49d610aa6fc0 +requests-toolbelt==0.10.1 \ + --hash=sha256:18565aa58116d9951ac39baa288d3adb5b3ff975c4f25eee78555d89e8f247f7 \ + --hash=sha256:62e09f7ff5ccbda92772a29f394a49c3ad6cb181d568b1337626b2abb628a63d # via twine rfc3986==2.0.0 \ --hash=sha256:50b1502b60e289cb37883f3dfd34532b8873c7de9f49bb546641ce9cbd256ebd \ --hash=sha256:97aacf9dbd4bfd829baad6e6309fa6573aaf1be3f6fa735c8ab05e46cecb261c # via twine -rich==12.5.1 \ - --hash=sha256:2eb4e6894cde1e017976d2975ac210ef515d7548bc595ba20e195fb9628acdeb \ - --hash=sha256:63a5c5ce3673d3d5fbbf23cd87e11ab84b6b451436f1b7f19ec54b6bc36ed7ca +rich==12.6.0 \ + --hash=sha256:a4eb26484f2c82589bd9a17c73d32a010b1e29d89f1604cd9bf3a2097b81bb5e \ + --hash=sha256:ba3a3775974105c221d31141f2c116f4fd65c5ceb0698657a11e9f295ec93fd0 # via twine rsa==4.9 \ --hash=sha256:90260d9058e514786967344d0ef75fa8727eed8a7d2e43ce9f4bcf1b536174f7 \ @@ -437,9 +466,9 @@ twine==4.0.1 \ --hash=sha256:42026c18e394eac3e06693ee52010baa5313e4811d5a11050e7d48436cf41b9e \ --hash=sha256:96b1cf12f7ae611a4a40b6ae8e9570215daff0611828f5fe1f37a16255ab24a0 # via -r requirements.in -typing-extensions==4.3.0 \ - --hash=sha256:25642c956049920a5aa49edcdd6ab1e06d7e5d467fc00e0506c44ac86fbfca02 \ - --hash=sha256:e6d2677a32f47fc7eb2795db1dd15c1f34eff616bcaf2cfb5e997f854fa1c4a6 +typing-extensions==4.4.0 \ + --hash=sha256:1511434bb92bf8dd198c12b1cc812e800d4181cfcb867674e0f8279cc93087aa \ + --hash=sha256:16fa4864408f655d35ec496218b85f79b3437c829e93320c7c9215ccfd92489e # via -r requirements.in urllib3==1.26.12 \ --hash=sha256:3fa96cf423e6987997fc326ae8df396db2a8b7c667747d47ddd8ecba91f4a74e \ @@ -447,9 +476,9 @@ urllib3==1.26.12 \ # via # requests # twine -virtualenv==20.16.4 \ - --hash=sha256:014f766e4134d0008dcaa1f95bafa0fb0f575795d07cae50b1bee514185d6782 \ - --hash=sha256:035ed57acce4ac35c82c9d8802202b0e71adac011a511ff650cbcf9635006a22 +virtualenv==20.16.6 \ + --hash=sha256:186ca84254abcbde98180fd17092f9628c5fe742273c02724972a1d8a2035108 \ + --hash=sha256:530b850b523c6449406dfba859d6345e48ef19b8439606c5d74d7d3c9e14d76e # via nox webencodings==0.5.1 \ --hash=sha256:a0af1213f3c2226497a97e2b3aa01a7e4bee4f403f95be16fc9acd2947514a78 \ @@ -459,13 +488,13 @@ wheel==0.37.1 \ --hash=sha256:4bdcd7d840138086126cd09254dc6195fb4fc6f01c050a1d7236f2630db1d22a \ --hash=sha256:e9a504e793efbca1b8e0e9cb979a249cf4a0a7b5b8c9e8b65a5e39d49529c1c4 # via -r requirements.in -zipp==3.8.1 \ - --hash=sha256:05b45f1ee8f807d0cc928485ca40a07cb491cf092ff587c0df9cb1fd154848d2 \ - --hash=sha256:47c40d7fe183a6f21403a199b3e4192cca5774656965b0a4988ad2f8feb5f009 +zipp==3.10.0 \ + --hash=sha256:4fcb6f278987a6605757302a6e40e896257570d11c51628968ccb2a47e80c6c1 \ + --hash=sha256:7a7262fd930bd3e36c50b9a64897aec3fafff3dfdeec9623ae22b40e93f99bb8 # via importlib-metadata # The following packages are considered to be unsafe in a requirements file: -setuptools==65.2.0 \ - --hash=sha256:7f4bc85450898a09f76ebf28b72fa25bc7111f6c7d665d514a60bba9c75ef2a9 \ - --hash=sha256:a3ca5857c89f82f5c9410e8508cb32f4872a3bafd4aa7ae122a24ca33bccc750 +setuptools==65.5.0 \ + --hash=sha256:512e5536220e38146176efb833d4a62aa726b7bbff82cfbc8ba9eaa3996e0b17 \ + --hash=sha256:f62ea9da9ed6289bfe868cd6845968a2c854d1427f8548d52cae02a42b4f0356 # via -r requirements.in diff --git a/google/cloud/aiplatform_v1/services/featurestore_service/async_client.py b/google/cloud/aiplatform_v1/services/featurestore_service/async_client.py index f88cfcc1e4..373cb302eb 100644 --- a/google/cloud/aiplatform_v1/services/featurestore_service/async_client.py +++ b/google/cloud/aiplatform_v1/services/featurestore_service/async_client.py @@ -1593,7 +1593,7 @@ async def sample_create_feature(): become the final component of the Feature's resource name. - This value may be up to 60 characters, and valid + This value may be up to 128 characters, and valid characters are ``[a-z0-9_]``. The first character cannot be a number. diff --git a/google/cloud/aiplatform_v1/services/featurestore_service/client.py b/google/cloud/aiplatform_v1/services/featurestore_service/client.py index e485cfeda8..a4e50e56fb 100644 --- a/google/cloud/aiplatform_v1/services/featurestore_service/client.py +++ b/google/cloud/aiplatform_v1/services/featurestore_service/client.py @@ -1859,7 +1859,7 @@ def sample_create_feature(): become the final component of the Feature's resource name. - This value may be up to 60 characters, and valid + This value may be up to 128 characters, and valid characters are ``[a-z0-9_]``. The first character cannot be a number. diff --git a/google/cloud/aiplatform_v1/services/tensorboard_service/async_client.py b/google/cloud/aiplatform_v1/services/tensorboard_service/async_client.py index 4ec8f12b68..dfc219be05 100644 --- a/google/cloud/aiplatform_v1/services/tensorboard_service/async_client.py +++ b/google/cloud/aiplatform_v1/services/tensorboard_service/async_client.py @@ -1280,10 +1280,9 @@ async def sample_list_tensorboard_experiments(): The request object. Request message for [TensorboardService.ListTensorboardExperiments][google.cloud.aiplatform.v1.TensorboardService.ListTensorboardExperiments]. parent (:class:`str`): - Required. The resource name of the - Tensorboard to list + Required. The resource name of the Tensorboard to list TensorboardExperiments. Format: - 'projects/{project}/locations/{location}/tensorboards/{tensorboard}' + ``projects/{project}/locations/{location}/tensorboards/{tensorboard}`` This corresponds to the ``parent`` field on the ``request`` instance; if ``request`` is provided, this @@ -2014,10 +2013,9 @@ async def sample_list_tensorboard_runs(): The request object. Request message for [TensorboardService.ListTensorboardRuns][google.cloud.aiplatform.v1.TensorboardService.ListTensorboardRuns]. parent (:class:`str`): - Required. The resource name of the - TensorboardExperiment to list - TensorboardRuns. Format: - 'projects/{project}/locations/{location}/tensorboards/{tensorboard}/experiments/{experiment}' + Required. The resource name of the TensorboardExperiment + to list TensorboardRuns. Format: + ``projects/{project}/locations/{location}/tensorboards/{tensorboard}/experiments/{experiment}`` This corresponds to the ``parent`` field on the ``request`` instance; if ``request`` is provided, this @@ -2745,10 +2743,9 @@ async def sample_list_tensorboard_time_series(): The request object. Request message for [TensorboardService.ListTensorboardTimeSeries][google.cloud.aiplatform.v1.TensorboardService.ListTensorboardTimeSeries]. parent (:class:`str`): - Required. The resource name of the - TensorboardRun to list - TensorboardTimeSeries. Format: - 'projects/{project}/locations/{location}/tensorboards/{tensorboard}/experiments/{experiment}/runs/{run}' + Required. The resource name of the TensorboardRun to + list TensorboardTimeSeries. Format: + ``projects/{project}/locations/{location}/tensorboards/{tensorboard}/experiments/{experiment}/runs/{run}`` This corresponds to the ``parent`` field on the ``request`` instance; if ``request`` is provided, this @@ -3222,7 +3219,7 @@ async def sample_read_tensorboard_blob_data(): time_series (:class:`str`): Required. The resource name of the TensorboardTimeSeries to list Blobs. Format: - 'projects/{project}/locations/{location}/tensorboards/{tensorboard}/experiments/{experiment}/runs/{run}/timeSeries/{time_series}' + ``projects/{project}/locations/{location}/tensorboards/{tensorboard}/experiments/{experiment}/runs/{run}/timeSeries/{time_series}`` This corresponds to the ``time_series`` field on the ``request`` instance; if ``request`` is provided, this diff --git a/google/cloud/aiplatform_v1/services/tensorboard_service/client.py b/google/cloud/aiplatform_v1/services/tensorboard_service/client.py index ed3e1952fb..f9ef55f177 100644 --- a/google/cloud/aiplatform_v1/services/tensorboard_service/client.py +++ b/google/cloud/aiplatform_v1/services/tensorboard_service/client.py @@ -1566,10 +1566,9 @@ def sample_list_tensorboard_experiments(): The request object. Request message for [TensorboardService.ListTensorboardExperiments][google.cloud.aiplatform.v1.TensorboardService.ListTensorboardExperiments]. parent (str): - Required. The resource name of the - Tensorboard to list + Required. The resource name of the Tensorboard to list TensorboardExperiments. Format: - 'projects/{project}/locations/{location}/tensorboards/{tensorboard}' + ``projects/{project}/locations/{location}/tensorboards/{tensorboard}`` This corresponds to the ``parent`` field on the ``request`` instance; if ``request`` is provided, this @@ -2312,10 +2311,9 @@ def sample_list_tensorboard_runs(): The request object. Request message for [TensorboardService.ListTensorboardRuns][google.cloud.aiplatform.v1.TensorboardService.ListTensorboardRuns]. parent (str): - Required. The resource name of the - TensorboardExperiment to list - TensorboardRuns. Format: - 'projects/{project}/locations/{location}/tensorboards/{tensorboard}/experiments/{experiment}' + Required. The resource name of the TensorboardExperiment + to list TensorboardRuns. Format: + ``projects/{project}/locations/{location}/tensorboards/{tensorboard}/experiments/{experiment}`` This corresponds to the ``parent`` field on the ``request`` instance; if ``request`` is provided, this @@ -3059,10 +3057,9 @@ def sample_list_tensorboard_time_series(): The request object. Request message for [TensorboardService.ListTensorboardTimeSeries][google.cloud.aiplatform.v1.TensorboardService.ListTensorboardTimeSeries]. parent (str): - Required. The resource name of the - TensorboardRun to list - TensorboardTimeSeries. Format: - 'projects/{project}/locations/{location}/tensorboards/{tensorboard}/experiments/{experiment}/runs/{run}' + Required. The resource name of the TensorboardRun to + list TensorboardTimeSeries. Format: + ``projects/{project}/locations/{location}/tensorboards/{tensorboard}/experiments/{experiment}/runs/{run}`` This corresponds to the ``parent`` field on the ``request`` instance; if ``request`` is provided, this @@ -3554,7 +3551,7 @@ def sample_read_tensorboard_blob_data(): time_series (str): Required. The resource name of the TensorboardTimeSeries to list Blobs. Format: - 'projects/{project}/locations/{location}/tensorboards/{tensorboard}/experiments/{experiment}/runs/{run}/timeSeries/{time_series}' + ``projects/{project}/locations/{location}/tensorboards/{tensorboard}/experiments/{experiment}/runs/{run}/timeSeries/{time_series}`` This corresponds to the ``time_series`` field on the ``request`` instance; if ``request`` is provided, this diff --git a/google/cloud/aiplatform_v1/types/artifact.py b/google/cloud/aiplatform_v1/types/artifact.py index 853ae4b0cb..12d2c781d4 100644 --- a/google/cloud/aiplatform_v1/types/artifact.py +++ b/google/cloud/aiplatform_v1/types/artifact.py @@ -85,7 +85,9 @@ class Artifact(proto.Message): metadata store. metadata (google.protobuf.struct_pb2.Struct): Properties of the Artifact. - The size of this field should not exceed 200KB. + Top level metadata keys' heading and trailing + spaces will be trimmed. The size of this field + should not exceed 200KB. description (str): Description of the Artifact """ diff --git a/google/cloud/aiplatform_v1/types/batch_prediction_job.py b/google/cloud/aiplatform_v1/types/batch_prediction_job.py index 8d472beafb..a2c6075813 100644 --- a/google/cloud/aiplatform_v1/types/batch_prediction_job.py +++ b/google/cloud/aiplatform_v1/types/batch_prediction_job.py @@ -103,6 +103,16 @@ class BatchPredictionJob(proto.Message): DEDICATED_RESOURCES this config may be provided (and the job will use these resources), if the Model doesn't support AUTOMATIC_RESOURCES, this config must be provided. + service_account (str): + The service account that the DeployedModel's container runs + as. If not specified, a system generated one will be used, + which has minimal permissions and the custom container, if + used, may not have enough permission to access other GCP + resources. + + Users deploying the Model must have the + ``iam.serviceAccounts.actAs`` permission on this service + account. manual_batch_tuning_parameters (google.cloud.aiplatform_v1.types.ManualBatchTuningParameters): Immutable. Parameters configuring the batch behavior. Currently only applicable when @@ -437,6 +447,10 @@ class OutputInfo(proto.Message): number=7, message=machine_resources.BatchDedicatedResources, ) + service_account = proto.Field( + proto.STRING, + number=29, + ) manual_batch_tuning_parameters = proto.Field( proto.MESSAGE, number=8, diff --git a/google/cloud/aiplatform_v1/types/context.py b/google/cloud/aiplatform_v1/types/context.py index f65d3613a5..5740cdeed5 100644 --- a/google/cloud/aiplatform_v1/types/context.py +++ b/google/cloud/aiplatform_v1/types/context.py @@ -78,7 +78,9 @@ class Context(proto.Message): metadata store. metadata (google.protobuf.struct_pb2.Struct): Properties of the Context. - The size of this field should not exceed 200KB. + Top level metadata keys' heading and trailing + spaces will be trimmed. The size of this field + should not exceed 200KB. description (str): Description of the Context """ diff --git a/google/cloud/aiplatform_v1/types/dataset.py b/google/cloud/aiplatform_v1/types/dataset.py index 3eeaa861d2..138f869a06 100644 --- a/google/cloud/aiplatform_v1/types/dataset.py +++ b/google/cloud/aiplatform_v1/types/dataset.py @@ -166,6 +166,18 @@ class ImportDataConfig(proto.Message): labels specified inside index file referenced by [import_schema_uri][google.cloud.aiplatform.v1.ImportDataConfig.import_schema_uri], e.g. jsonl file. + annotation_labels (Mapping[str, str]): + Labels that will be applied to newly imported Annotations. + If two Annotations are identical, one of them will be + deduped. Two Annotations are considered identical if their + [payload][google.cloud.aiplatform.v1.Annotation.payload], + [payload_schema_uri][google.cloud.aiplatform.v1.Annotation.payload_schema_uri] + and all of their + [labels][google.cloud.aiplatform.v1.Annotation.labels] are + the same. These labels will be overridden by Annotation + labels specified inside index file referenced by + [import_schema_uri][google.cloud.aiplatform.v1.ImportDataConfig.import_schema_uri], + e.g. jsonl file. import_schema_uri (str): Required. Points to a YAML file stored on Google Cloud Storage describing the import format. Validation will be @@ -185,6 +197,11 @@ class ImportDataConfig(proto.Message): proto.STRING, number=2, ) + annotation_labels = proto.MapField( + proto.STRING, + proto.STRING, + number=3, + ) import_schema_uri = proto.Field( proto.STRING, number=4, diff --git a/google/cloud/aiplatform_v1/types/execution.py b/google/cloud/aiplatform_v1/types/execution.py index 9d157b8ba2..465c5d4c00 100644 --- a/google/cloud/aiplatform_v1/types/execution.py +++ b/google/cloud/aiplatform_v1/types/execution.py @@ -81,7 +81,9 @@ class Execution(proto.Message): metadata store. metadata (google.protobuf.struct_pb2.Struct): Properties of the Execution. - The size of this field should not exceed 200KB. + Top level metadata keys' heading and trailing + spaces will be trimmed. The size of this field + should not exceed 200KB. description (str): Description of the Execution """ diff --git a/google/cloud/aiplatform_v1/types/featurestore.py b/google/cloud/aiplatform_v1/types/featurestore.py index e88b5c17bc..742f03f67b 100644 --- a/google/cloud/aiplatform_v1/types/featurestore.py +++ b/google/cloud/aiplatform_v1/types/featurestore.py @@ -61,10 +61,12 @@ class Featurestore(proto.Message): System reserved label keys are prefixed with "aiplatform.googleapis.com/" and are immutable. online_serving_config (google.cloud.aiplatform_v1.types.Featurestore.OnlineServingConfig): - Optional. Config for online storage - resources. If unset, the featurestore will not - have an online store and cannot be used for - online serving. + Optional. Config for online storage resources. The field + should not co-exist with the field of + ``OnlineStoreReplicationConfig``. If both of it and + OnlineStoreReplicationConfig are unset, the feature store + will not have an online store and cannot be used for online + serving. state (google.cloud.aiplatform_v1.types.Featurestore.State): Output only. State of the featurestore. encryption_spec (google.cloud.aiplatform_v1.types.EncryptionSpec): diff --git a/google/cloud/aiplatform_v1/types/featurestore_service.py b/google/cloud/aiplatform_v1/types/featurestore_service.py index f505c7fd34..14640e33fa 100644 --- a/google/cloud/aiplatform_v1/types/featurestore_service.py +++ b/google/cloud/aiplatform_v1/types/featurestore_service.py @@ -548,6 +548,12 @@ class BatchReadFeatureValuesRequest(proto.Message): [BatchReadFeatureValuesRequest.entity_type_specs] must have a column specifying entity IDs in the EntityType in [BatchReadFeatureValuesRequest.request][] . + start_time (google.protobuf.timestamp_pb2.Timestamp): + Optional. Excludes Feature values with + feature generation timestamp before this + timestamp. If not set, retrieve oldest values + kept in Feature Store. Timestamp, if present, + must not have higher than millisecond precision. """ class PassThroughField(proto.Message): @@ -629,6 +635,11 @@ class EntityTypeSpec(proto.Message): number=7, message=EntityTypeSpec, ) + start_time = proto.Field( + proto.MESSAGE, + number=11, + message=timestamp_pb2.Timestamp, + ) class ExportFeatureValuesRequest(proto.Message): @@ -1109,7 +1120,7 @@ class CreateFeatureRequest(proto.Message): Required. The ID to use for the Feature, which will become the final component of the Feature's resource name. - This value may be up to 60 characters, and valid characters + This value may be up to 128 characters, and valid characters are ``[a-z0-9_]``. The first character cannot be a number. The value must be unique within an EntityType. diff --git a/google/cloud/aiplatform_v1/types/index_endpoint.py b/google/cloud/aiplatform_v1/types/index_endpoint.py index 6cfbc20e44..1bb2b36c9f 100644 --- a/google/cloud/aiplatform_v1/types/index_endpoint.py +++ b/google/cloud/aiplatform_v1/types/index_endpoint.py @@ -87,7 +87,7 @@ class IndexEndpoint(proto.Message): are mutually exclusive. `Format `__: - projects/{project}/global/networks/{network}. Where + ``projects/{project}/global/networks/{network}``. Where {project} is a project number, as in '12345', and {network} is network name. enable_private_service_connect (bool): diff --git a/google/cloud/aiplatform_v1/types/metadata_service.py b/google/cloud/aiplatform_v1/types/metadata_service.py index 37d42b7aed..94462113d8 100644 --- a/google/cloud/aiplatform_v1/types/metadata_service.py +++ b/google/cloud/aiplatform_v1/types/metadata_service.py @@ -359,7 +359,8 @@ class ListArtifactsRequest(proto.Message): ``in_context("projects//locations//metadataStores//contexts/")`` Each of the above supported filter types can be combined - together using logical operators (``AND`` & ``OR``). + together using logical operators (``AND`` & ``OR``). Maximum + nested expression depth allowed is 5. For example: ``display_name = "test" AND metadata.field1.bool_value = true``. @@ -667,7 +668,8 @@ class ListContextsRequest(proto.Message): "projects//locations//metadataStores//contexts/" Each of the above supported filters can be combined together - using logical operators (``AND`` & ``OR``). + using logical operators (``AND`` & ``OR``). Maximum nested + expression depth allowed is 5. For example: ``display_name = "test" AND metadata.field1.bool_value = true``. @@ -1103,7 +1105,10 @@ class ListExecutionsRequest(proto.Message): ``in_context("projects//locations//metadataStores//contexts/")`` Each of the above supported filters can be combined together - using logical operators (``AND`` & ``OR``). For example: + using logical operators (``AND`` & ``OR``). Maximum nested + expression depth allowed is 5. + + For example: ``display_name = "test" AND metadata.field1.bool_value = true``. order_by (str): How the list of messages is ordered. Specify the values to @@ -1523,7 +1528,8 @@ class QueryArtifactLineageSubgraphRequest(proto.Message): ``metadata.field_1.number_value = 10.0`` Each of the above supported filter types can be combined - together using logical operators (``AND`` & ``OR``). + together using logical operators (``AND`` & ``OR``). Maximum + nested expression depth allowed is 5. For example: ``display_name = "test" AND metadata.field1.bool_value = true``. diff --git a/google/cloud/aiplatform_v1/types/model.py b/google/cloud/aiplatform_v1/types/model.py index 611a233562..972f6e6af1 100644 --- a/google/cloud/aiplatform_v1/types/model.py +++ b/google/cloud/aiplatform_v1/types/model.py @@ -50,9 +50,9 @@ class Model(proto.Message): version_aliases (Sequence[str]): User provided version aliases so that a model version can be referenced via alias (i.e. - projects/{project}/locations/{location}/models/{model_id}@{version_alias} + ``projects/{project}/locations/{location}/models/{model_id}@{version_alias}`` instead of auto-generated version id (i.e. - projects/{project}/locations/{location}/models/{model_id}@{version_id}). + ``projects/{project}/locations/{location}/models/{model_id}@{version_id})``. The format is [a-z][a-zA-Z0-9-]{0,126}[a-z0-9] to distinguish from version_id. A default version alias will be created for the first version of the model, and there must @@ -280,6 +280,11 @@ class Model(proto.Message): be automl training pipeline, custom training pipeline, BigQuery ML, or existing Vertex AI Model. + metadata_artifact (str): + Output only. The resource name of the Artifact that was + created in MetadataStore when creating the Model. The + Artifact resource name pattern is + ``projects/{project}/locations/{location}/metadataStores/{metadata_store}/artifacts/{artifact}``. """ class DeploymentResourcesType(proto.Enum): @@ -454,6 +459,10 @@ class ExportableContent(proto.Enum): number=38, message="ModelSourceInfo", ) + metadata_artifact = proto.Field( + proto.STRING, + number=44, + ) class PredictSchemata(proto.Message): diff --git a/google/cloud/aiplatform_v1/types/pipeline_job.py b/google/cloud/aiplatform_v1/types/pipeline_job.py index 0eb9ef7c4d..0b555f6e10 100644 --- a/google/cloud/aiplatform_v1/types/pipeline_job.py +++ b/google/cloud/aiplatform_v1/types/pipeline_job.py @@ -177,8 +177,8 @@ class InputArtifact(proto.Message): Attributes: artifact_id (str): Artifact resource id from MLMD. Which is the last portion of - an artifact resource - name(projects/{project}/locations/{location}/metadataStores/default/artifacts/{artifact_id}). + an artifact resource name: + ``projects/{project}/locations/{location}/metadataStores/default/artifacts/{artifact_id}``. The artifact must stay within the same project, location and default metadatastore as the pipeline. @@ -565,6 +565,20 @@ class ContainerDetail(proto.Message): [PipelineJob.pipeline_spec][google.cloud.aiplatform.v1.PipelineJob.pipeline_spec] specifies the ``pre_caching_check`` hook in the lifecycle events. + failed_main_jobs (Sequence[str]): + Output only. The names of the previously failed + [CustomJob][google.cloud.aiplatform.v1.CustomJob] for the + main container executions. The list includes the all + attempts in chronological order. + failed_pre_caching_check_jobs (Sequence[str]): + Output only. The names of the previously failed + [CustomJob][google.cloud.aiplatform.v1.CustomJob] for the + pre-caching-check container executions. This job will be + available if the + [PipelineJob.pipeline_spec][google.cloud.aiplatform.v1.PipelineJob.pipeline_spec] + specifies the ``pre_caching_check`` hook in the lifecycle + events. The list includes the all attempts in chronological + order. """ main_job = proto.Field( @@ -575,6 +589,14 @@ class ContainerDetail(proto.Message): proto.STRING, number=2, ) + failed_main_jobs = proto.RepeatedField( + proto.STRING, + number=3, + ) + failed_pre_caching_check_jobs = proto.RepeatedField( + proto.STRING, + number=4, + ) class CustomJobDetail(proto.Message): r"""The detailed info for a custom job executor. diff --git a/google/cloud/aiplatform_v1/types/tensorboard_service.py b/google/cloud/aiplatform_v1/types/tensorboard_service.py index ccecccb39f..5d529b5690 100644 --- a/google/cloud/aiplatform_v1/types/tensorboard_service.py +++ b/google/cloud/aiplatform_v1/types/tensorboard_service.py @@ -310,10 +310,9 @@ class ListTensorboardExperimentsRequest(proto.Message): Attributes: parent (str): - Required. The resource name of the - Tensorboard to list TensorboardExperiments. - Format: - 'projects/{project}/locations/{location}/tensorboards/{tensorboard}' + Required. The resource name of the Tensorboard to list + TensorboardExperiments. Format: + ``projects/{project}/locations/{location}/tensorboards/{tensorboard}`` filter (str): Lists the TensorboardExperiments that match the filter expression. @@ -546,7 +545,7 @@ class ReadTensorboardBlobDataRequest(proto.Message): time_series (str): Required. The resource name of the TensorboardTimeSeries to list Blobs. Format: - 'projects/{project}/locations/{location}/tensorboards/{tensorboard}/experiments/{experiment}/runs/{run}/timeSeries/{time_series}' + ``projects/{project}/locations/{location}/tensorboards/{tensorboard}/experiments/{experiment}/runs/{run}/timeSeries/{time_series}`` blob_ids (Sequence[str]): IDs of the blobs to read. """ @@ -583,10 +582,9 @@ class ListTensorboardRunsRequest(proto.Message): Attributes: parent (str): - Required. The resource name of the - TensorboardExperiment to list TensorboardRuns. - Format: - 'projects/{project}/locations/{location}/tensorboards/{tensorboard}/experiments/{experiment}' + Required. The resource name of the TensorboardExperiment to + list TensorboardRuns. Format: + ``projects/{project}/locations/{location}/tensorboards/{tensorboard}/experiments/{experiment}`` filter (str): Lists the TensorboardRuns that match the filter expression. @@ -818,10 +816,9 @@ class ListTensorboardTimeSeriesRequest(proto.Message): Attributes: parent (str): - Required. The resource name of the - TensorboardRun to list TensorboardTimeSeries. - Format: - 'projects/{project}/locations/{location}/tensorboards/{tensorboard}/experiments/{experiment}/runs/{run}' + Required. The resource name of the TensorboardRun to list + TensorboardTimeSeries. Format: + ``projects/{project}/locations/{location}/tensorboards/{tensorboard}/experiments/{experiment}/runs/{run}`` filter (str): Lists the TensorboardTimeSeries that match the filter expression. diff --git a/google/cloud/aiplatform_v1/types/training_pipeline.py b/google/cloud/aiplatform_v1/types/training_pipeline.py index a2c680121c..42c117f810 100644 --- a/google/cloud/aiplatform_v1/types/training_pipeline.py +++ b/google/cloud/aiplatform_v1/types/training_pipeline.py @@ -404,6 +404,9 @@ class InputDataConfig(proto.Message): [annotation_schema_uri][google.cloud.aiplatform.v1.InputDataConfig.annotation_schema_uri] should be specified as both of them represent the same thing: problem type. + persist_ml_use_assignment (bool): + Whether to persist the ML use assignment to + data item system labels. """ fraction_split = proto.Field( @@ -464,6 +467,10 @@ class InputDataConfig(proto.Message): proto.STRING, number=7, ) + persist_ml_use_assignment = proto.Field( + proto.BOOL, + number=11, + ) class FractionSplit(proto.Message): diff --git a/google/cloud/aiplatform_v1/types/user_action_reference.py b/google/cloud/aiplatform_v1/types/user_action_reference.py index 539d958f1d..6cd433f198 100644 --- a/google/cloud/aiplatform_v1/types/user_action_reference.py +++ b/google/cloud/aiplatform_v1/types/user_action_reference.py @@ -38,16 +38,15 @@ class UserActionReference(proto.Message): Attributes: operation (str): - For API calls that return a long running - operation. Resource name of the long running - operation. Format: - 'projects/{project}/locations/{location}/operations/{operation}' + For API calls that return a long running operation. Resource + name of the long running operation. Format: + ``projects/{project}/locations/{location}/operations/{operation}`` This field is a member of `oneof`_ ``reference``. data_labeling_job (str): For API calls that start a LabelingJob. Resource name of the LabelingJob. Format: - 'projects/{project}/locations/{location}/dataLabelingJobs/{data_labeling_job}' + ``projects/{project}/locations/{location}/dataLabelingJobs/{data_labeling_job}`` This field is a member of `oneof`_ ``reference``. method (str): diff --git a/google/cloud/aiplatform_v1beta1/services/deployment_resource_pool_service/async_client.py b/google/cloud/aiplatform_v1beta1/services/deployment_resource_pool_service/async_client.py index f41517d69f..f132a7052e 100644 --- a/google/cloud/aiplatform_v1beta1/services/deployment_resource_pool_service/async_client.py +++ b/google/cloud/aiplatform_v1beta1/services/deployment_resource_pool_service/async_client.py @@ -293,11 +293,9 @@ async def sample_create_deployment_resource_pool(): The request object. Request message for CreateDeploymentResourcePool method. parent (:class:`str`): - Required. The parent location - resource where this - DeploymentResourcePool will be created. - Format: - projects/{project}/locations/{location} + Required. The parent location resource where this + DeploymentResourcePool will be created. Format: + ``projects/{project}/locations/{location}`` This corresponds to the ``parent`` field on the ``request`` instance; if ``request`` is provided, this @@ -439,7 +437,7 @@ async def sample_get_deployment_resource_pool(): name (:class:`str`): Required. The name of the DeploymentResourcePool to retrieve. Format: - projects/{project}/locations/{location}/deploymentResourcePools/{deployment_resource_pool} + ``projects/{project}/locations/{location}/deploymentResourcePools/{deployment_resource_pool}`` This corresponds to the ``name`` field on the ``request`` instance; if ``request`` is provided, this @@ -547,10 +545,9 @@ async def sample_list_deployment_resource_pools(): The request object. Request message for ListDeploymentResourcePools method. parent (:class:`str`): - Required. The parent Location which - owns this collection of - DeploymentResourcePools. Format: - projects/{project}/locations/{location} + Required. The parent Location which owns this collection + of DeploymentResourcePools. Format: + ``projects/{project}/locations/{location}`` This corresponds to the ``parent`` field on the ``request`` instance; if ``request`` is provided, this @@ -673,7 +670,7 @@ async def sample_delete_deployment_resource_pool(): name (:class:`str`): Required. The name of the DeploymentResourcePool to delete. Format: - projects/{project}/locations/{location}/deploymentResourcePools/{deployment_resource_pool} + ``projects/{project}/locations/{location}/deploymentResourcePools/{deployment_resource_pool}`` This corresponds to the ``name`` field on the ``request`` instance; if ``request`` is provided, this @@ -800,7 +797,7 @@ async def sample_query_deployed_models(): deployment_resource_pool (:class:`str`): Required. The name of the target DeploymentResourcePool to query. Format: - projects/{project}/locations/{location}/deploymentResourcePools/{deployment_resource_pool} + ``projects/{project}/locations/{location}/deploymentResourcePools/{deployment_resource_pool}`` This corresponds to the ``deployment_resource_pool`` field on the ``request`` instance; if ``request`` is provided, this diff --git a/google/cloud/aiplatform_v1beta1/services/deployment_resource_pool_service/client.py b/google/cloud/aiplatform_v1beta1/services/deployment_resource_pool_service/client.py index d03d9fc044..a3a84277ce 100644 --- a/google/cloud/aiplatform_v1beta1/services/deployment_resource_pool_service/client.py +++ b/google/cloud/aiplatform_v1beta1/services/deployment_resource_pool_service/client.py @@ -547,11 +547,9 @@ def sample_create_deployment_resource_pool(): The request object. Request message for CreateDeploymentResourcePool method. parent (str): - Required. The parent location - resource where this - DeploymentResourcePool will be created. - Format: - projects/{project}/locations/{location} + Required. The parent location resource where this + DeploymentResourcePool will be created. Format: + ``projects/{project}/locations/{location}`` This corresponds to the ``parent`` field on the ``request`` instance; if ``request`` is provided, this @@ -700,7 +698,7 @@ def sample_get_deployment_resource_pool(): name (str): Required. The name of the DeploymentResourcePool to retrieve. Format: - projects/{project}/locations/{location}/deploymentResourcePools/{deployment_resource_pool} + ``projects/{project}/locations/{location}/deploymentResourcePools/{deployment_resource_pool}`` This corresponds to the ``name`` field on the ``request`` instance; if ``request`` is provided, this @@ -812,10 +810,9 @@ def sample_list_deployment_resource_pools(): The request object. Request message for ListDeploymentResourcePools method. parent (str): - Required. The parent Location which - owns this collection of - DeploymentResourcePools. Format: - projects/{project}/locations/{location} + Required. The parent Location which owns this collection + of DeploymentResourcePools. Format: + ``projects/{project}/locations/{location}`` This corresponds to the ``parent`` field on the ``request`` instance; if ``request`` is provided, this @@ -944,7 +941,7 @@ def sample_delete_deployment_resource_pool(): name (str): Required. The name of the DeploymentResourcePool to delete. Format: - projects/{project}/locations/{location}/deploymentResourcePools/{deployment_resource_pool} + ``projects/{project}/locations/{location}/deploymentResourcePools/{deployment_resource_pool}`` This corresponds to the ``name`` field on the ``request`` instance; if ``request`` is provided, this @@ -1078,7 +1075,7 @@ def sample_query_deployed_models(): deployment_resource_pool (str): Required. The name of the target DeploymentResourcePool to query. Format: - projects/{project}/locations/{location}/deploymentResourcePools/{deployment_resource_pool} + ``projects/{project}/locations/{location}/deploymentResourcePools/{deployment_resource_pool}`` This corresponds to the ``deployment_resource_pool`` field on the ``request`` instance; if ``request`` is provided, this diff --git a/google/cloud/aiplatform_v1beta1/services/featurestore_service/async_client.py b/google/cloud/aiplatform_v1beta1/services/featurestore_service/async_client.py index df5d78ec8e..58bd8fda5a 100644 --- a/google/cloud/aiplatform_v1beta1/services/featurestore_service/async_client.py +++ b/google/cloud/aiplatform_v1beta1/services/featurestore_service/async_client.py @@ -1594,7 +1594,7 @@ async def sample_create_feature(): become the final component of the Feature's resource name. - This value may be up to 60 characters, and valid + This value may be up to 128 characters, and valid characters are ``[a-z0-9_]``. The first character cannot be a number. diff --git a/google/cloud/aiplatform_v1beta1/services/featurestore_service/client.py b/google/cloud/aiplatform_v1beta1/services/featurestore_service/client.py index f9de3ea53f..7981fdf8df 100644 --- a/google/cloud/aiplatform_v1beta1/services/featurestore_service/client.py +++ b/google/cloud/aiplatform_v1beta1/services/featurestore_service/client.py @@ -1860,7 +1860,7 @@ def sample_create_feature(): become the final component of the Feature's resource name. - This value may be up to 60 characters, and valid + This value may be up to 128 characters, and valid characters are ``[a-z0-9_]``. The first character cannot be a number. diff --git a/google/cloud/aiplatform_v1beta1/services/tensorboard_service/async_client.py b/google/cloud/aiplatform_v1beta1/services/tensorboard_service/async_client.py index 89d3d3f5f2..d02615ac25 100644 --- a/google/cloud/aiplatform_v1beta1/services/tensorboard_service/async_client.py +++ b/google/cloud/aiplatform_v1beta1/services/tensorboard_service/async_client.py @@ -1280,10 +1280,9 @@ async def sample_list_tensorboard_experiments(): The request object. Request message for [TensorboardService.ListTensorboardExperiments][google.cloud.aiplatform.v1beta1.TensorboardService.ListTensorboardExperiments]. parent (:class:`str`): - Required. The resource name of the - Tensorboard to list + Required. The resource name of the Tensorboard to list TensorboardExperiments. Format: - 'projects/{project}/locations/{location}/tensorboards/{tensorboard}' + ``projects/{project}/locations/{location}/tensorboards/{tensorboard}`` This corresponds to the ``parent`` field on the ``request`` instance; if ``request`` is provided, this @@ -2014,10 +2013,9 @@ async def sample_list_tensorboard_runs(): The request object. Request message for [TensorboardService.ListTensorboardRuns][google.cloud.aiplatform.v1beta1.TensorboardService.ListTensorboardRuns]. parent (:class:`str`): - Required. The resource name of the - TensorboardExperiment to list - TensorboardRuns. Format: - 'projects/{project}/locations/{location}/tensorboards/{tensorboard}/experiments/{experiment}' + Required. The resource name of the TensorboardExperiment + to list TensorboardRuns. Format: + ``projects/{project}/locations/{location}/tensorboards/{tensorboard}/experiments/{experiment}`` This corresponds to the ``parent`` field on the ``request`` instance; if ``request`` is provided, this @@ -2745,10 +2743,9 @@ async def sample_list_tensorboard_time_series(): The request object. Request message for [TensorboardService.ListTensorboardTimeSeries][google.cloud.aiplatform.v1beta1.TensorboardService.ListTensorboardTimeSeries]. parent (:class:`str`): - Required. The resource name of the - TensorboardRun to list - TensorboardTimeSeries. Format: - 'projects/{project}/locations/{location}/tensorboards/{tensorboard}/experiments/{experiment}/runs/{run}' + Required. The resource name of the TensorboardRun to + list TensorboardTimeSeries. Format: + ``projects/{project}/locations/{location}/tensorboards/{tensorboard}/experiments/{experiment}/runs/{run}`` This corresponds to the ``parent`` field on the ``request`` instance; if ``request`` is provided, this @@ -3222,7 +3219,7 @@ async def sample_read_tensorboard_blob_data(): time_series (:class:`str`): Required. The resource name of the TensorboardTimeSeries to list Blobs. Format: - 'projects/{project}/locations/{location}/tensorboards/{tensorboard}/experiments/{experiment}/runs/{run}/timeSeries/{time_series}' + ``projects/{project}/locations/{location}/tensorboards/{tensorboard}/experiments/{experiment}/runs/{run}/timeSeries/{time_series}`` This corresponds to the ``time_series`` field on the ``request`` instance; if ``request`` is provided, this diff --git a/google/cloud/aiplatform_v1beta1/services/tensorboard_service/client.py b/google/cloud/aiplatform_v1beta1/services/tensorboard_service/client.py index 8e9c1264a1..1afb27d40d 100644 --- a/google/cloud/aiplatform_v1beta1/services/tensorboard_service/client.py +++ b/google/cloud/aiplatform_v1beta1/services/tensorboard_service/client.py @@ -1566,10 +1566,9 @@ def sample_list_tensorboard_experiments(): The request object. Request message for [TensorboardService.ListTensorboardExperiments][google.cloud.aiplatform.v1beta1.TensorboardService.ListTensorboardExperiments]. parent (str): - Required. The resource name of the - Tensorboard to list + Required. The resource name of the Tensorboard to list TensorboardExperiments. Format: - 'projects/{project}/locations/{location}/tensorboards/{tensorboard}' + ``projects/{project}/locations/{location}/tensorboards/{tensorboard}`` This corresponds to the ``parent`` field on the ``request`` instance; if ``request`` is provided, this @@ -2312,10 +2311,9 @@ def sample_list_tensorboard_runs(): The request object. Request message for [TensorboardService.ListTensorboardRuns][google.cloud.aiplatform.v1beta1.TensorboardService.ListTensorboardRuns]. parent (str): - Required. The resource name of the - TensorboardExperiment to list - TensorboardRuns. Format: - 'projects/{project}/locations/{location}/tensorboards/{tensorboard}/experiments/{experiment}' + Required. The resource name of the TensorboardExperiment + to list TensorboardRuns. Format: + ``projects/{project}/locations/{location}/tensorboards/{tensorboard}/experiments/{experiment}`` This corresponds to the ``parent`` field on the ``request`` instance; if ``request`` is provided, this @@ -3059,10 +3057,9 @@ def sample_list_tensorboard_time_series(): The request object. Request message for [TensorboardService.ListTensorboardTimeSeries][google.cloud.aiplatform.v1beta1.TensorboardService.ListTensorboardTimeSeries]. parent (str): - Required. The resource name of the - TensorboardRun to list - TensorboardTimeSeries. Format: - 'projects/{project}/locations/{location}/tensorboards/{tensorboard}/experiments/{experiment}/runs/{run}' + Required. The resource name of the TensorboardRun to + list TensorboardTimeSeries. Format: + ``projects/{project}/locations/{location}/tensorboards/{tensorboard}/experiments/{experiment}/runs/{run}`` This corresponds to the ``parent`` field on the ``request`` instance; if ``request`` is provided, this @@ -3554,7 +3551,7 @@ def sample_read_tensorboard_blob_data(): time_series (str): Required. The resource name of the TensorboardTimeSeries to list Blobs. Format: - 'projects/{project}/locations/{location}/tensorboards/{tensorboard}/experiments/{experiment}/runs/{run}/timeSeries/{time_series}' + ``projects/{project}/locations/{location}/tensorboards/{tensorboard}/experiments/{experiment}/runs/{run}/timeSeries/{time_series}`` This corresponds to the ``time_series`` field on the ``request`` instance; if ``request`` is provided, this diff --git a/google/cloud/aiplatform_v1beta1/types/accelerator_type.py b/google/cloud/aiplatform_v1beta1/types/accelerator_type.py index 5f80e5be03..86c7c5c70f 100644 --- a/google/cloud/aiplatform_v1beta1/types/accelerator_type.py +++ b/google/cloud/aiplatform_v1beta1/types/accelerator_type.py @@ -33,6 +33,7 @@ class AcceleratorType(proto.Enum): NVIDIA_TESLA_P4 = 4 NVIDIA_TESLA_T4 = 5 NVIDIA_TESLA_A100 = 8 + NVIDIA_A100_80GB = 9 TPU_V2 = 6 TPU_V3 = 7 diff --git a/google/cloud/aiplatform_v1beta1/types/artifact.py b/google/cloud/aiplatform_v1beta1/types/artifact.py index df2e59b99e..579b62c06c 100644 --- a/google/cloud/aiplatform_v1beta1/types/artifact.py +++ b/google/cloud/aiplatform_v1beta1/types/artifact.py @@ -85,7 +85,9 @@ class Artifact(proto.Message): metadata store. metadata (google.protobuf.struct_pb2.Struct): Properties of the Artifact. - The size of this field should not exceed 200KB. + Top level metadata keys' heading and trailing + spaces will be trimmed. The size of this field + should not exceed 200KB. description (str): Description of the Artifact """ diff --git a/google/cloud/aiplatform_v1beta1/types/batch_prediction_job.py b/google/cloud/aiplatform_v1beta1/types/batch_prediction_job.py index 998a558ac8..83d696706a 100644 --- a/google/cloud/aiplatform_v1beta1/types/batch_prediction_job.py +++ b/google/cloud/aiplatform_v1beta1/types/batch_prediction_job.py @@ -85,6 +85,10 @@ class BatchPredictionJob(proto.Message): [Model's][google.cloud.aiplatform.v1beta1.BatchPredictionJob.model] [PredictSchemata's][google.cloud.aiplatform.v1beta1.Model.predict_schemata] [instance_schema_uri][google.cloud.aiplatform.v1beta1.PredictSchemata.instance_schema_uri]. + instance_config (google.cloud.aiplatform_v1beta1.types.BatchPredictionJob.InstanceConfig): + Configuration for how to convert batch + prediction input instances to the prediction + instances that are sent to the Model. model_parameters (google.protobuf.struct_pb2.Value): The parameters that govern the predictions. The schema of the parameters may be specified via the @@ -283,6 +287,130 @@ class InputConfig(proto.Message): number=1, ) + class InstanceConfig(proto.Message): + r"""Configuration defining how to transform batch prediction + input instances to the instances that the Model accepts. + + Attributes: + instance_type (str): + The format of the instance that the Model accepts. Vertex AI + will convert compatible [batch prediction input instance + formats][google.cloud.aiplatform.v1beta1.BatchPredictionJob.InputConfig.instances_format] + to the specified format. + + Supported values are: + + - ``object``: Each input is converted to JSON object + format. + + - For ``bigquery``, each row is converted to an object. + - For ``jsonl``, each line of the JSONL input must be an + object. + - Does not apply to ``csv``, ``file-list``, + ``tf-record``, or ``tf-record-gzip``. + + - ``array``: Each input is converted to JSON array format. + + - For ``bigquery``, each row is converted to an array. + The order of columns is determined by the BigQuery + column order, unless + [included_fields][google.cloud.aiplatform.v1beta1.BatchPredictionJob.InstanceConfig.included_fields] + is populated. + [included_fields][google.cloud.aiplatform.v1beta1.BatchPredictionJob.InstanceConfig.included_fields] + must be populated for specifying field orders. + - For ``jsonl``, if each line of the JSONL input is an + object, + [included_fields][google.cloud.aiplatform.v1beta1.BatchPredictionJob.InstanceConfig.included_fields] + must be populated for specifying field orders. + - Does not apply to ``csv``, ``file-list``, + ``tf-record``, or ``tf-record-gzip``. + + If not specified, Vertex AI converts the batch prediction + input as follows: + + - For ``bigquery`` and ``csv``, the behavior is the same as + ``array``. The order of columns is the same as defined in + the file or table, unless + [included_fields][google.cloud.aiplatform.v1beta1.BatchPredictionJob.InstanceConfig.included_fields] + is populated. + - For ``jsonl``, the prediction instance format is + determined by each line of the input. + - For ``tf-record``/``tf-record-gzip``, each record will be + converted to an object in the format of + ``{"b64": }``, where ```` is the + Base64-encoded string of the content of the record. + - For ``file-list``, each file in the list will be + converted to an object in the format of + ``{"b64": }``, where ```` is the + Base64-encoded string of the content of the file. + key_field (str): + The name of the field that is considered as a key. + + The values identified by the key field is not included in + the transformed instances that is sent to the Model. This is + similar to specifying this name of the field in + [excluded_fields][google.cloud.aiplatform.v1beta1.BatchPredictionJob.InstanceConfig.excluded_fields]. + In addition, the batch prediction output will not include + the instances. Instead the output will only include the + value of the key field, in a field named ``key`` in the + output: + + - For ``jsonl`` output format, the output will have a + ``key`` field instead of the ``instance`` field. + - For ``csv``/``bigquery`` output format, the output will + have have a ``key`` column instead of the instance + feature columns. + + The input must be JSONL with objects at each line, CSV, + BigQuery or TfRecord. + included_fields (Sequence[str]): + Fields that will be included in the prediction instance that + is sent to the Model. + + If + [instance_type][google.cloud.aiplatform.v1beta1.BatchPredictionJob.InstanceConfig.instance_type] + is ``array``, the order of field names in included_fields + also determines the order of the values in the array. + + When included_fields is populated, + [excluded_fields][google.cloud.aiplatform.v1beta1.BatchPredictionJob.InstanceConfig.excluded_fields] + must be empty. + + The input must be JSONL with objects at each line, CSV, + BigQuery or TfRecord. + excluded_fields (Sequence[str]): + Fields that will be excluded in the prediction instance that + is sent to the Model. + + Excluded will be attached to the batch prediction output if + [key_field][google.cloud.aiplatform.v1beta1.BatchPredictionJob.InstanceConfig.key_field] + is not specified. + + When excluded_fields is populated, + [included_fields][google.cloud.aiplatform.v1beta1.BatchPredictionJob.InstanceConfig.included_fields] + must be empty. + + The input must be JSONL with objects at each line, CSV, + BigQuery or TfRecord. + """ + + instance_type = proto.Field( + proto.STRING, + number=1, + ) + key_field = proto.Field( + proto.STRING, + number=2, + ) + included_fields = proto.RepeatedField( + proto.STRING, + number=3, + ) + excluded_fields = proto.RepeatedField( + proto.STRING, + number=4, + ) + class OutputConfig(proto.Message): r"""Configures the output of [BatchPredictionJob][google.cloud.aiplatform.v1beta1.BatchPredictionJob]. @@ -448,6 +576,11 @@ class OutputInfo(proto.Message): number=4, message=InputConfig, ) + instance_config = proto.Field( + proto.MESSAGE, + number=27, + message=InstanceConfig, + ) model_parameters = proto.Field( proto.MESSAGE, number=5, diff --git a/google/cloud/aiplatform_v1beta1/types/context.py b/google/cloud/aiplatform_v1beta1/types/context.py index 40d271bba2..8768c634c6 100644 --- a/google/cloud/aiplatform_v1beta1/types/context.py +++ b/google/cloud/aiplatform_v1beta1/types/context.py @@ -78,7 +78,9 @@ class Context(proto.Message): metadata store. metadata (google.protobuf.struct_pb2.Struct): Properties of the Context. - The size of this field should not exceed 200KB. + Top level metadata keys' heading and trailing + spaces will be trimmed. The size of this field + should not exceed 200KB. description (str): Description of the Context """ diff --git a/google/cloud/aiplatform_v1beta1/types/dataset.py b/google/cloud/aiplatform_v1beta1/types/dataset.py index 10fc9c1320..af0bb821dd 100644 --- a/google/cloud/aiplatform_v1beta1/types/dataset.py +++ b/google/cloud/aiplatform_v1beta1/types/dataset.py @@ -166,6 +166,18 @@ class ImportDataConfig(proto.Message): labels specified inside index file referenced by [import_schema_uri][google.cloud.aiplatform.v1beta1.ImportDataConfig.import_schema_uri], e.g. jsonl file. + annotation_labels (Mapping[str, str]): + Labels that will be applied to newly imported Annotations. + If two Annotations are identical, one of them will be + deduped. Two Annotations are considered identical if their + [payload][google.cloud.aiplatform.v1beta1.Annotation.payload], + [payload_schema_uri][google.cloud.aiplatform.v1beta1.Annotation.payload_schema_uri] + and all of their + [labels][google.cloud.aiplatform.v1beta1.Annotation.labels] + are the same. These labels will be overridden by Annotation + labels specified inside index file referenced by + [import_schema_uri][google.cloud.aiplatform.v1beta1.ImportDataConfig.import_schema_uri], + e.g. jsonl file. import_schema_uri (str): Required. Points to a YAML file stored on Google Cloud Storage describing the import format. Validation will be @@ -185,6 +197,11 @@ class ImportDataConfig(proto.Message): proto.STRING, number=2, ) + annotation_labels = proto.MapField( + proto.STRING, + proto.STRING, + number=3, + ) import_schema_uri = proto.Field( proto.STRING, number=4, diff --git a/google/cloud/aiplatform_v1beta1/types/deployment_resource_pool.py b/google/cloud/aiplatform_v1beta1/types/deployment_resource_pool.py index da677e6afc..4ea08fe5e8 100644 --- a/google/cloud/aiplatform_v1beta1/types/deployment_resource_pool.py +++ b/google/cloud/aiplatform_v1beta1/types/deployment_resource_pool.py @@ -36,7 +36,7 @@ class DeploymentResourcePool(proto.Message): name (str): Output only. The resource name of the DeploymentResourcePool. Format: - projects/{project}/locations/{location}/deploymentResourcePools/{deployment_resource_pool} + ``projects/{project}/locations/{location}/deploymentResourcePools/{deployment_resource_pool}`` dedicated_resources (google.cloud.aiplatform_v1beta1.types.DedicatedResources): Required. The underlying DedicatedResources that the DeploymentResourcePool uses. diff --git a/google/cloud/aiplatform_v1beta1/types/deployment_resource_pool_service.py b/google/cloud/aiplatform_v1beta1/types/deployment_resource_pool_service.py index e80cd7a6f5..e3b64208bb 100644 --- a/google/cloud/aiplatform_v1beta1/types/deployment_resource_pool_service.py +++ b/google/cloud/aiplatform_v1beta1/types/deployment_resource_pool_service.py @@ -44,9 +44,9 @@ class CreateDeploymentResourcePoolRequest(proto.Message): Attributes: parent (str): - Required. The parent location resource where - this DeploymentResourcePool will be created. - Format: projects/{project}/locations/{location} + Required. The parent location resource where this + DeploymentResourcePool will be created. Format: + ``projects/{project}/locations/{location}`` deployment_resource_pool (google.cloud.aiplatform_v1beta1.types.DeploymentResourcePool): Required. The DeploymentResourcePool to create. @@ -97,7 +97,7 @@ class GetDeploymentResourcePoolRequest(proto.Message): name (str): Required. The name of the DeploymentResourcePool to retrieve. Format: - projects/{project}/locations/{location}/deploymentResourcePools/{deployment_resource_pool} + ``projects/{project}/locations/{location}/deploymentResourcePools/{deployment_resource_pool}`` """ name = proto.Field( @@ -111,9 +111,9 @@ class ListDeploymentResourcePoolsRequest(proto.Message): Attributes: parent (str): - Required. The parent Location which owns this - collection of DeploymentResourcePools. Format: - projects/{project}/locations/{location} + Required. The parent Location which owns this collection of + DeploymentResourcePools. Format: + ``projects/{project}/locations/{location}`` page_size (int): The maximum number of DeploymentResourcePools to return. The service may return fewer than @@ -193,7 +193,7 @@ class DeleteDeploymentResourcePoolRequest(proto.Message): name (str): Required. The name of the DeploymentResourcePool to delete. Format: - projects/{project}/locations/{location}/deploymentResourcePools/{deployment_resource_pool} + ``projects/{project}/locations/{location}/deploymentResourcePools/{deployment_resource_pool}`` """ name = proto.Field( @@ -209,7 +209,7 @@ class QueryDeployedModelsRequest(proto.Message): deployment_resource_pool (str): Required. The name of the target DeploymentResourcePool to query. Format: - projects/{project}/locations/{location}/deploymentResourcePools/{deployment_resource_pool} + ``projects/{project}/locations/{location}/deploymentResourcePools/{deployment_resource_pool}`` page_size (int): The maximum number of DeployedModels to return. The service may return fewer than this @@ -251,6 +251,12 @@ class QueryDeployedModelsResponse(proto.Message): deployed_model_refs (Sequence[google.cloud.aiplatform_v1beta1.types.DeployedModelRef]): References to the DeployedModels that share the specified deploymentResourcePool. + total_deployed_model_count (int): + The total number of DeployedModels on this + DeploymentResourcePool. + total_endpoint_count (int): + The total number of Endpoints that have + DeployedModels on this DeploymentResourcePool. """ @property @@ -271,6 +277,14 @@ def raw_page(self): number=3, message=deployed_model_ref.DeployedModelRef, ) + total_deployed_model_count = proto.Field( + proto.INT32, + number=4, + ) + total_endpoint_count = proto.Field( + proto.INT32, + number=5, + ) __all__ = tuple(sorted(__protobuf__.manifest)) diff --git a/google/cloud/aiplatform_v1beta1/types/endpoint.py b/google/cloud/aiplatform_v1beta1/types/endpoint.py index 034d8e97fa..961107ad8b 100644 --- a/google/cloud/aiplatform_v1beta1/types/endpoint.py +++ b/google/cloud/aiplatform_v1beta1/types/endpoint.py @@ -219,7 +219,7 @@ class DeployedModel(proto.Message): shared_resources (str): The resource name of the shared DeploymentResourcePool to deploy on. Format: - projects/{project}/locations/{location}/deploymentResourcePools/{deployment_resource_pool} + ``projects/{project}/locations/{location}/deploymentResourcePools/{deployment_resource_pool}`` This field is a member of `oneof`_ ``prediction_resources``. id (str): diff --git a/google/cloud/aiplatform_v1beta1/types/execution.py b/google/cloud/aiplatform_v1beta1/types/execution.py index 937c7a4041..9f9a97b759 100644 --- a/google/cloud/aiplatform_v1beta1/types/execution.py +++ b/google/cloud/aiplatform_v1beta1/types/execution.py @@ -81,7 +81,9 @@ class Execution(proto.Message): metadata store. metadata (google.protobuf.struct_pb2.Struct): Properties of the Execution. - The size of this field should not exceed 200KB. + Top level metadata keys' heading and trailing + spaces will be trimmed. The size of this field + should not exceed 200KB. description (str): Description of the Execution """ diff --git a/google/cloud/aiplatform_v1beta1/types/featurestore.py b/google/cloud/aiplatform_v1beta1/types/featurestore.py index e10503041f..aa7b45e938 100644 --- a/google/cloud/aiplatform_v1beta1/types/featurestore.py +++ b/google/cloud/aiplatform_v1beta1/types/featurestore.py @@ -61,10 +61,12 @@ class Featurestore(proto.Message): System reserved label keys are prefixed with "aiplatform.googleapis.com/" and are immutable. online_serving_config (google.cloud.aiplatform_v1beta1.types.Featurestore.OnlineServingConfig): - Optional. Config for online storage - resources. If unset, the featurestore will not - have an online store and cannot be used for - online serving. + Optional. Config for online storage resources. The field + should not co-exist with the field of + ``OnlineStoreReplicationConfig``. If both of it and + OnlineStoreReplicationConfig are unset, the feature store + will not have an online store and cannot be used for online + serving. state (google.cloud.aiplatform_v1beta1.types.Featurestore.State): Output only. State of the featurestore. encryption_spec (google.cloud.aiplatform_v1beta1.types.EncryptionSpec): diff --git a/google/cloud/aiplatform_v1beta1/types/featurestore_service.py b/google/cloud/aiplatform_v1beta1/types/featurestore_service.py index ddeea2ccd1..3212de066b 100644 --- a/google/cloud/aiplatform_v1beta1/types/featurestore_service.py +++ b/google/cloud/aiplatform_v1beta1/types/featurestore_service.py @@ -555,6 +555,12 @@ class BatchReadFeatureValuesRequest(proto.Message): [BatchReadFeatureValuesRequest.entity_type_specs] must have a column specifying entity IDs in the EntityType in [BatchReadFeatureValuesRequest.request][] . + start_time (google.protobuf.timestamp_pb2.Timestamp): + Optional. Excludes Feature values with + feature generation timestamp before this + timestamp. If not set, retrieve oldest values + kept in Feature Store. Timestamp, if present, + must not have higher than millisecond precision. """ class PassThroughField(proto.Message): @@ -636,6 +642,11 @@ class EntityTypeSpec(proto.Message): number=7, message=EntityTypeSpec, ) + start_time = proto.Field( + proto.MESSAGE, + number=11, + message=timestamp_pb2.Timestamp, + ) class ExportFeatureValuesRequest(proto.Message): @@ -1116,7 +1127,7 @@ class CreateFeatureRequest(proto.Message): Required. The ID to use for the Feature, which will become the final component of the Feature's resource name. - This value may be up to 60 characters, and valid characters + This value may be up to 128 characters, and valid characters are ``[a-z0-9_]``. The first character cannot be a number. The value must be unique within an EntityType. @@ -1751,7 +1762,11 @@ class SelectEntity(proto.Message): class SelectTimeRangeAndFeature(proto.Message): r"""Message to select time range and feature. Values of the selected feature generated within an inclusive - time range will be deleted. + time range will be deleted. Using this option permanently + deletes the feature values from the specified feature IDs within + the specified time range. This might include data from the + online storage. If you want to retain any deleted historical + data in the online storage, you must re-ingest it. Attributes: time_range (google.type.interval_pb2.Interval): diff --git a/google/cloud/aiplatform_v1beta1/types/index_endpoint.py b/google/cloud/aiplatform_v1beta1/types/index_endpoint.py index 593aa01378..c7a73b5661 100644 --- a/google/cloud/aiplatform_v1beta1/types/index_endpoint.py +++ b/google/cloud/aiplatform_v1beta1/types/index_endpoint.py @@ -87,7 +87,7 @@ class IndexEndpoint(proto.Message): are mutually exclusive. `Format `__: - projects/{project}/global/networks/{network}. Where + ``projects/{project}/global/networks/{network}``. Where {project} is a project number, as in '12345', and {network} is network name. enable_private_service_connect (bool): diff --git a/google/cloud/aiplatform_v1beta1/types/metadata_service.py b/google/cloud/aiplatform_v1beta1/types/metadata_service.py index 828e180ddc..9b9dfaabb1 100644 --- a/google/cloud/aiplatform_v1beta1/types/metadata_service.py +++ b/google/cloud/aiplatform_v1beta1/types/metadata_service.py @@ -359,7 +359,8 @@ class ListArtifactsRequest(proto.Message): ``in_context("projects//locations//metadataStores//contexts/")`` Each of the above supported filter types can be combined - together using logical operators (``AND`` & ``OR``). + together using logical operators (``AND`` & ``OR``). Maximum + nested expression depth allowed is 5. For example: ``display_name = "test" AND metadata.field1.bool_value = true``. @@ -668,7 +669,8 @@ class ListContextsRequest(proto.Message): "projects//locations//metadataStores//contexts/" Each of the above supported filters can be combined together - using logical operators (``AND`` & ``OR``). + using logical operators (``AND`` & ``OR``). Maximum nested + expression depth allowed is 5. For example: ``display_name = "test" AND metadata.field1.bool_value = true``. @@ -1105,7 +1107,10 @@ class ListExecutionsRequest(proto.Message): ``in_context("projects//locations//metadataStores//contexts/")`` Each of the above supported filters can be combined together - using logical operators (``AND`` & ``OR``). For example: + using logical operators (``AND`` & ``OR``). Maximum nested + expression depth allowed is 5. + + For example: ``display_name = "test" AND metadata.field1.bool_value = true``. order_by (str): How the list of messages is ordered. Specify the values to @@ -1525,7 +1530,8 @@ class QueryArtifactLineageSubgraphRequest(proto.Message): ``metadata.field_1.number_value = 10.0`` Each of the above supported filter types can be combined - together using logical operators (``AND`` & ``OR``). + together using logical operators (``AND`` & ``OR``). Maximum + nested expression depth allowed is 5. For example: ``display_name = "test" AND metadata.field1.bool_value = true``. diff --git a/google/cloud/aiplatform_v1beta1/types/model.py b/google/cloud/aiplatform_v1beta1/types/model.py index 1e2a735031..d25dac0006 100644 --- a/google/cloud/aiplatform_v1beta1/types/model.py +++ b/google/cloud/aiplatform_v1beta1/types/model.py @@ -50,9 +50,9 @@ class Model(proto.Message): version_aliases (Sequence[str]): User provided version aliases so that a model version can be referenced via alias (i.e. - projects/{project}/locations/{location}/models/{model_id}@{version_alias} + ``projects/{project}/locations/{location}/models/{model_id}@{version_alias}`` instead of auto-generated version id (i.e. - projects/{project}/locations/{location}/models/{model_id}@{version_id}). + ``projects/{project}/locations/{location}/models/{model_id}@{version_id})``. The format is [a-z][a-zA-Z0-9-]{0,126}[a-z0-9] to distinguish from version_id. A default version alias will be created for the first version of the model, and there must @@ -280,6 +280,11 @@ class Model(proto.Message): be automl training pipeline, custom training pipeline, BigQuery ML, or existing Vertex AI Model. + metadata_artifact (str): + Output only. The resource name of the Artifact that was + created in MetadataStore when creating the Model. The + Artifact resource name pattern is + ``projects/{project}/locations/{location}/metadataStores/{metadata_store}/artifacts/{artifact}``. """ class DeploymentResourcesType(proto.Enum): @@ -454,6 +459,10 @@ class ExportableContent(proto.Enum): number=38, message="ModelSourceInfo", ) + metadata_artifact = proto.Field( + proto.STRING, + number=44, + ) class PredictSchemata(proto.Message): diff --git a/google/cloud/aiplatform_v1beta1/types/pipeline_job.py b/google/cloud/aiplatform_v1beta1/types/pipeline_job.py index 42fd5821f6..400b8b4ab0 100644 --- a/google/cloud/aiplatform_v1beta1/types/pipeline_job.py +++ b/google/cloud/aiplatform_v1beta1/types/pipeline_job.py @@ -177,8 +177,8 @@ class InputArtifact(proto.Message): Attributes: artifact_id (str): Artifact resource id from MLMD. Which is the last portion of - an artifact resource - name(projects/{project}/locations/{location}/metadataStores/default/artifacts/{artifact_id}). + an artifact resource name: + ``projects/{project}/locations/{location}/metadataStores/default/artifacts/{artifact_id}``. The artifact must stay within the same project, location and default metadatastore as the pipeline. @@ -565,6 +565,20 @@ class ContainerDetail(proto.Message): [PipelineJob.pipeline_spec][google.cloud.aiplatform.v1beta1.PipelineJob.pipeline_spec] specifies the ``pre_caching_check`` hook in the lifecycle events. + failed_main_jobs (Sequence[str]): + Output only. The names of the previously failed + [CustomJob][google.cloud.aiplatform.v1beta1.CustomJob] for + the main container executions. The list includes the all + attempts in chronological order. + failed_pre_caching_check_jobs (Sequence[str]): + Output only. The names of the previously failed + [CustomJob][google.cloud.aiplatform.v1beta1.CustomJob] for + the pre-caching-check container executions. This job will be + available if the + [PipelineJob.pipeline_spec][google.cloud.aiplatform.v1beta1.PipelineJob.pipeline_spec] + specifies the ``pre_caching_check`` hook in the lifecycle + events. The list includes the all attempts in chronological + order. """ main_job = proto.Field( @@ -575,6 +589,14 @@ class ContainerDetail(proto.Message): proto.STRING, number=2, ) + failed_main_jobs = proto.RepeatedField( + proto.STRING, + number=3, + ) + failed_pre_caching_check_jobs = proto.RepeatedField( + proto.STRING, + number=4, + ) class CustomJobDetail(proto.Message): r"""The detailed info for a custom job executor. diff --git a/google/cloud/aiplatform_v1beta1/types/tensorboard_service.py b/google/cloud/aiplatform_v1beta1/types/tensorboard_service.py index 04103dddfe..46e9939b83 100644 --- a/google/cloud/aiplatform_v1beta1/types/tensorboard_service.py +++ b/google/cloud/aiplatform_v1beta1/types/tensorboard_service.py @@ -310,10 +310,9 @@ class ListTensorboardExperimentsRequest(proto.Message): Attributes: parent (str): - Required. The resource name of the - Tensorboard to list TensorboardExperiments. - Format: - 'projects/{project}/locations/{location}/tensorboards/{tensorboard}' + Required. The resource name of the Tensorboard to list + TensorboardExperiments. Format: + ``projects/{project}/locations/{location}/tensorboards/{tensorboard}`` filter (str): Lists the TensorboardExperiments that match the filter expression. @@ -546,7 +545,7 @@ class ReadTensorboardBlobDataRequest(proto.Message): time_series (str): Required. The resource name of the TensorboardTimeSeries to list Blobs. Format: - 'projects/{project}/locations/{location}/tensorboards/{tensorboard}/experiments/{experiment}/runs/{run}/timeSeries/{time_series}' + ``projects/{project}/locations/{location}/tensorboards/{tensorboard}/experiments/{experiment}/runs/{run}/timeSeries/{time_series}`` blob_ids (Sequence[str]): IDs of the blobs to read. """ @@ -583,10 +582,9 @@ class ListTensorboardRunsRequest(proto.Message): Attributes: parent (str): - Required. The resource name of the - TensorboardExperiment to list TensorboardRuns. - Format: - 'projects/{project}/locations/{location}/tensorboards/{tensorboard}/experiments/{experiment}' + Required. The resource name of the TensorboardExperiment to + list TensorboardRuns. Format: + ``projects/{project}/locations/{location}/tensorboards/{tensorboard}/experiments/{experiment}`` filter (str): Lists the TensorboardRuns that match the filter expression. @@ -818,10 +816,9 @@ class ListTensorboardTimeSeriesRequest(proto.Message): Attributes: parent (str): - Required. The resource name of the - TensorboardRun to list TensorboardTimeSeries. - Format: - 'projects/{project}/locations/{location}/tensorboards/{tensorboard}/experiments/{experiment}/runs/{run}' + Required. The resource name of the TensorboardRun to list + TensorboardTimeSeries. Format: + ``projects/{project}/locations/{location}/tensorboards/{tensorboard}/experiments/{experiment}/runs/{run}`` filter (str): Lists the TensorboardTimeSeries that match the filter expression. diff --git a/google/cloud/aiplatform_v1beta1/types/training_pipeline.py b/google/cloud/aiplatform_v1beta1/types/training_pipeline.py index a32b4257be..5cd7cacab7 100644 --- a/google/cloud/aiplatform_v1beta1/types/training_pipeline.py +++ b/google/cloud/aiplatform_v1beta1/types/training_pipeline.py @@ -405,6 +405,9 @@ class InputDataConfig(proto.Message): [annotation_schema_uri][google.cloud.aiplatform.v1beta1.InputDataConfig.annotation_schema_uri] should be specified as both of them represent the same thing: problem type. + persist_ml_use_assignment (bool): + Whether to persist the ML use assignment to + data item system labels. """ fraction_split = proto.Field( @@ -465,6 +468,10 @@ class InputDataConfig(proto.Message): proto.STRING, number=7, ) + persist_ml_use_assignment = proto.Field( + proto.BOOL, + number=11, + ) class FractionSplit(proto.Message): diff --git a/google/cloud/aiplatform_v1beta1/types/user_action_reference.py b/google/cloud/aiplatform_v1beta1/types/user_action_reference.py index 0b1676026c..d26405340f 100644 --- a/google/cloud/aiplatform_v1beta1/types/user_action_reference.py +++ b/google/cloud/aiplatform_v1beta1/types/user_action_reference.py @@ -38,16 +38,15 @@ class UserActionReference(proto.Message): Attributes: operation (str): - For API calls that return a long running - operation. Resource name of the long running - operation. Format: - 'projects/{project}/locations/{location}/operations/{operation}' + For API calls that return a long running operation. Resource + name of the long running operation. Format: + ``projects/{project}/locations/{location}/operations/{operation}`` This field is a member of `oneof`_ ``reference``. data_labeling_job (str): For API calls that start a LabelingJob. Resource name of the LabelingJob. Format: - 'projects/{project}/locations/{location}/dataLabelingJobs/{data_labeling_job}' + ``projects/{project}/locations/{location}/dataLabelingJobs/{data_labeling_job}`` This field is a member of `oneof`_ ``reference``. method (str): diff --git a/owlbot.py b/owlbot.py index 5d3313eada..4c1a23515e 100644 --- a/owlbot.py +++ b/owlbot.py @@ -127,13 +127,13 @@ s.replace( ".kokoro/samples/python3.*/common.cfg", """env_vars: \{ - key: "BUILD_SPECIFIC_GCLOUD_PROJECT" - value: "python-docs-samples-tests-.*?" - \}""", + key: "BUILD_SPECIFIC_GCLOUD_PROJECT" + value: "python-docs-samples-tests-.*?" +\}""", """env_vars: { - key: "BUILD_SPECIFIC_GCLOUD_PROJECT" - value: "ucaip-sample-tests" - }""", + key: "BUILD_SPECIFIC_GCLOUD_PROJECT" + value: "ucaip-sample-tests" +}""", ) s.replace( diff --git a/tests/unit/gapic/aiplatform_v1/test_job_service.py b/tests/unit/gapic/aiplatform_v1/test_job_service.py index d97749d184..6aa9a746be 100644 --- a/tests/unit/gapic/aiplatform_v1/test_job_service.py +++ b/tests/unit/gapic/aiplatform_v1/test_job_service.py @@ -5096,6 +5096,7 @@ def test_create_batch_prediction_job(request_type, transport: str = "grpc"): display_name="display_name_value", model="model_value", model_version_id="model_version_id_value", + service_account="service_account_value", generate_explanation=True, state=job_state.JobState.JOB_STATE_QUEUED, ) @@ -5112,6 +5113,7 @@ def test_create_batch_prediction_job(request_type, transport: str = "grpc"): assert response.display_name == "display_name_value" assert response.model == "model_value" assert response.model_version_id == "model_version_id_value" + assert response.service_account == "service_account_value" assert response.generate_explanation is True assert response.state == job_state.JobState.JOB_STATE_QUEUED @@ -5159,6 +5161,7 @@ async def test_create_batch_prediction_job_async( display_name="display_name_value", model="model_value", model_version_id="model_version_id_value", + service_account="service_account_value", generate_explanation=True, state=job_state.JobState.JOB_STATE_QUEUED, ) @@ -5176,6 +5179,7 @@ async def test_create_batch_prediction_job_async( assert response.display_name == "display_name_value" assert response.model == "model_value" assert response.model_version_id == "model_version_id_value" + assert response.service_account == "service_account_value" assert response.generate_explanation is True assert response.state == job_state.JobState.JOB_STATE_QUEUED @@ -5381,6 +5385,7 @@ def test_get_batch_prediction_job(request_type, transport: str = "grpc"): display_name="display_name_value", model="model_value", model_version_id="model_version_id_value", + service_account="service_account_value", generate_explanation=True, state=job_state.JobState.JOB_STATE_QUEUED, ) @@ -5397,6 +5402,7 @@ def test_get_batch_prediction_job(request_type, transport: str = "grpc"): assert response.display_name == "display_name_value" assert response.model == "model_value" assert response.model_version_id == "model_version_id_value" + assert response.service_account == "service_account_value" assert response.generate_explanation is True assert response.state == job_state.JobState.JOB_STATE_QUEUED @@ -5444,6 +5450,7 @@ async def test_get_batch_prediction_job_async( display_name="display_name_value", model="model_value", model_version_id="model_version_id_value", + service_account="service_account_value", generate_explanation=True, state=job_state.JobState.JOB_STATE_QUEUED, ) @@ -5461,6 +5468,7 @@ async def test_get_batch_prediction_job_async( assert response.display_name == "display_name_value" assert response.model == "model_value" assert response.model_version_id == "model_version_id_value" + assert response.service_account == "service_account_value" assert response.generate_explanation is True assert response.state == job_state.JobState.JOB_STATE_QUEUED diff --git a/tests/unit/gapic/aiplatform_v1/test_model_service.py b/tests/unit/gapic/aiplatform_v1/test_model_service.py index f24a884b32..fa1969d90a 100644 --- a/tests/unit/gapic/aiplatform_v1/test_model_service.py +++ b/tests/unit/gapic/aiplatform_v1/test_model_service.py @@ -963,6 +963,7 @@ def test_get_model(request_type, transport: str = "grpc"): supported_input_storage_formats=["supported_input_storage_formats_value"], supported_output_storage_formats=["supported_output_storage_formats_value"], etag="etag_value", + metadata_artifact="metadata_artifact_value", ) response = client.get_model(request) @@ -992,6 +993,7 @@ def test_get_model(request_type, transport: str = "grpc"): "supported_output_storage_formats_value" ] assert response.etag == "etag_value" + assert response.metadata_artifact == "metadata_artifact_value" def test_get_model_empty_call(): @@ -1047,6 +1049,7 @@ async def test_get_model_async( "supported_output_storage_formats_value" ], etag="etag_value", + metadata_artifact="metadata_artifact_value", ) ) response = await client.get_model(request) @@ -1077,6 +1080,7 @@ async def test_get_model_async( "supported_output_storage_formats_value" ] assert response.etag == "etag_value" + assert response.metadata_artifact == "metadata_artifact_value" @pytest.mark.asyncio @@ -2121,6 +2125,7 @@ def test_update_model(request_type, transport: str = "grpc"): supported_input_storage_formats=["supported_input_storage_formats_value"], supported_output_storage_formats=["supported_output_storage_formats_value"], etag="etag_value", + metadata_artifact="metadata_artifact_value", ) response = client.update_model(request) @@ -2150,6 +2155,7 @@ def test_update_model(request_type, transport: str = "grpc"): "supported_output_storage_formats_value" ] assert response.etag == "etag_value" + assert response.metadata_artifact == "metadata_artifact_value" def test_update_model_empty_call(): @@ -2205,6 +2211,7 @@ async def test_update_model_async( "supported_output_storage_formats_value" ], etag="etag_value", + metadata_artifact="metadata_artifact_value", ) ) response = await client.update_model(request) @@ -2235,6 +2242,7 @@ async def test_update_model_async( "supported_output_storage_formats_value" ] assert response.etag == "etag_value" + assert response.metadata_artifact == "metadata_artifact_value" @pytest.mark.asyncio @@ -2896,6 +2904,7 @@ def test_merge_version_aliases(request_type, transport: str = "grpc"): supported_input_storage_formats=["supported_input_storage_formats_value"], supported_output_storage_formats=["supported_output_storage_formats_value"], etag="etag_value", + metadata_artifact="metadata_artifact_value", ) response = client.merge_version_aliases(request) @@ -2925,6 +2934,7 @@ def test_merge_version_aliases(request_type, transport: str = "grpc"): "supported_output_storage_formats_value" ] assert response.etag == "etag_value" + assert response.metadata_artifact == "metadata_artifact_value" def test_merge_version_aliases_empty_call(): @@ -2985,6 +2995,7 @@ async def test_merge_version_aliases_async( "supported_output_storage_formats_value" ], etag="etag_value", + metadata_artifact="metadata_artifact_value", ) ) response = await client.merge_version_aliases(request) @@ -3015,6 +3026,7 @@ async def test_merge_version_aliases_async( "supported_output_storage_formats_value" ] assert response.etag == "etag_value" + assert response.metadata_artifact == "metadata_artifact_value" @pytest.mark.asyncio diff --git a/tests/unit/gapic/aiplatform_v1beta1/test_deployment_resource_pool_service.py b/tests/unit/gapic/aiplatform_v1beta1/test_deployment_resource_pool_service.py index 01de157360..6c9d070c91 100644 --- a/tests/unit/gapic/aiplatform_v1beta1/test_deployment_resource_pool_service.py +++ b/tests/unit/gapic/aiplatform_v1beta1/test_deployment_resource_pool_service.py @@ -2019,6 +2019,8 @@ def test_query_deployed_models(request_type, transport: str = "grpc"): call.return_value = ( deployment_resource_pool_service.QueryDeployedModelsResponse( next_page_token="next_page_token_value", + total_deployed_model_count=2769, + total_endpoint_count=2156, ) ) response = client.query_deployed_models(request) @@ -2031,6 +2033,8 @@ def test_query_deployed_models(request_type, transport: str = "grpc"): # Establish that the response is the type that we expect. assert isinstance(response, pagers.QueryDeployedModelsPager) assert response.next_page_token == "next_page_token_value" + assert response.total_deployed_model_count == 2769 + assert response.total_endpoint_count == 2156 def test_query_deployed_models_empty_call(): @@ -2073,6 +2077,8 @@ async def test_query_deployed_models_async( call.return_value = grpc_helpers_async.FakeUnaryUnaryCall( deployment_resource_pool_service.QueryDeployedModelsResponse( next_page_token="next_page_token_value", + total_deployed_model_count=2769, + total_endpoint_count=2156, ) ) response = await client.query_deployed_models(request) @@ -2085,6 +2091,8 @@ async def test_query_deployed_models_async( # Establish that the response is the type that we expect. assert isinstance(response, pagers.QueryDeployedModelsAsyncPager) assert response.next_page_token == "next_page_token_value" + assert response.total_deployed_model_count == 2769 + assert response.total_endpoint_count == 2156 @pytest.mark.asyncio diff --git a/tests/unit/gapic/aiplatform_v1beta1/test_model_service.py b/tests/unit/gapic/aiplatform_v1beta1/test_model_service.py index 54700bad74..8ba27c2f57 100644 --- a/tests/unit/gapic/aiplatform_v1beta1/test_model_service.py +++ b/tests/unit/gapic/aiplatform_v1beta1/test_model_service.py @@ -967,6 +967,7 @@ def test_get_model(request_type, transport: str = "grpc"): supported_input_storage_formats=["supported_input_storage_formats_value"], supported_output_storage_formats=["supported_output_storage_formats_value"], etag="etag_value", + metadata_artifact="metadata_artifact_value", ) response = client.get_model(request) @@ -996,6 +997,7 @@ def test_get_model(request_type, transport: str = "grpc"): "supported_output_storage_formats_value" ] assert response.etag == "etag_value" + assert response.metadata_artifact == "metadata_artifact_value" def test_get_model_empty_call(): @@ -1051,6 +1053,7 @@ async def test_get_model_async( "supported_output_storage_formats_value" ], etag="etag_value", + metadata_artifact="metadata_artifact_value", ) ) response = await client.get_model(request) @@ -1081,6 +1084,7 @@ async def test_get_model_async( "supported_output_storage_formats_value" ] assert response.etag == "etag_value" + assert response.metadata_artifact == "metadata_artifact_value" @pytest.mark.asyncio @@ -2125,6 +2129,7 @@ def test_update_model(request_type, transport: str = "grpc"): supported_input_storage_formats=["supported_input_storage_formats_value"], supported_output_storage_formats=["supported_output_storage_formats_value"], etag="etag_value", + metadata_artifact="metadata_artifact_value", ) response = client.update_model(request) @@ -2154,6 +2159,7 @@ def test_update_model(request_type, transport: str = "grpc"): "supported_output_storage_formats_value" ] assert response.etag == "etag_value" + assert response.metadata_artifact == "metadata_artifact_value" def test_update_model_empty_call(): @@ -2209,6 +2215,7 @@ async def test_update_model_async( "supported_output_storage_formats_value" ], etag="etag_value", + metadata_artifact="metadata_artifact_value", ) ) response = await client.update_model(request) @@ -2239,6 +2246,7 @@ async def test_update_model_async( "supported_output_storage_formats_value" ] assert response.etag == "etag_value" + assert response.metadata_artifact == "metadata_artifact_value" @pytest.mark.asyncio @@ -3141,6 +3149,7 @@ def test_merge_version_aliases(request_type, transport: str = "grpc"): supported_input_storage_formats=["supported_input_storage_formats_value"], supported_output_storage_formats=["supported_output_storage_formats_value"], etag="etag_value", + metadata_artifact="metadata_artifact_value", ) response = client.merge_version_aliases(request) @@ -3170,6 +3179,7 @@ def test_merge_version_aliases(request_type, transport: str = "grpc"): "supported_output_storage_formats_value" ] assert response.etag == "etag_value" + assert response.metadata_artifact == "metadata_artifact_value" def test_merge_version_aliases_empty_call(): @@ -3230,6 +3240,7 @@ async def test_merge_version_aliases_async( "supported_output_storage_formats_value" ], etag="etag_value", + metadata_artifact="metadata_artifact_value", ) ) response = await client.merge_version_aliases(request) @@ -3260,6 +3271,7 @@ async def test_merge_version_aliases_async( "supported_output_storage_formats_value" ] assert response.etag == "etag_value" + assert response.metadata_artifact == "metadata_artifact_value" @pytest.mark.asyncio