diff --git "a/perf-df-bnb-1xA10.csv" "b/perf-df-bnb-1xA10.csv" --- "a/perf-df-bnb-1xA10.csv" +++ "b/perf-df-bnb-1xA10.csv" @@ -128,7 +128,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66777601-00df468c29c4b6341780e42b;f501422e-c1b4-4d94-bf37-cb1c7180a1e8) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbbbb-1306026a4013a6230f8d5ac8;d2918c57-4242-42d3-af7c-165bf5a4e605) Repository Not Found for url: https://huggingface.co/m/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -200,7 +200,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667779e6-2e111d104bf7fac824875d84;775b51a7-fd49-42ec-b0e5-2c8bd2137beb) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbfa1-40dd0346718423ca15919988;9e79c82a-5f20-4ebb-b71c-bf30408dc0b1) Repository Not Found for url: https://huggingface.co/v/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -284,7 +284,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66777a12-642be1ce02bc0bbd03fce16c;b5e45589-8c13-45f3-8aef-298edc128ebc) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbfcc-69c8044b4beaa9ca1889890b;effd42c7-23f1-4db8-93f5-952e9c0eb08d) Repository Not Found for url: https://huggingface.co/0/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -363,7 +363,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6677795a-55acc65f1892f1156b533bd7;08d736f5-e0c0-41ee-8af1-255b1d9fa9d0) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbf22-5127ed0256a0d72335e7e191;d7fd4fae-916b-4b5e-9c82-21bbfd4e7f7f) Repository Not Found for url: https://huggingface.co/2/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -434,7 +434,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6677788a-6d800ad32ac58ec412565f1b;1499a3aa-60c2-4ff4-a55d-93e3feffb6a6) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbe5b-786557e16169adb705f96fec;4f4fd21d-6d06-4d63-b594-086753dbf3a8) Repository Not Found for url: https://huggingface.co/l/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -531,7 +531,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66777651-04ea3d6d604c84b25f370235;e8c78f62-0adf-4c6b-ae6b-a927fec889d5) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbc1c-61e41abe43f0aa5017a29607;31e1034a-342f-4e07-a246-f829f78fd30e) Repository Not Found for url: https://huggingface.co/s/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -601,7 +601,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66777986-26cc18791498b57c3131ee49;c3e1f7fe-4c25-4f4f-bc9c-76286390c141) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbf4b-5f7cf95c28faf7aa312bd6bd;a7f39f3d-2d80-4d7d-8e4f-b3829e515ee4) Repository Not Found for url: https://huggingface.co/B/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -675,7 +675,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66777863-13cd83556024abab623e63c4;e0367870-7736-4fe6-aafa-dfc69bf54925) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbe32-4b056bcf54da9748243fcd7f;b3cef86a-a707-4013-a7c1-69a9ca60f26c) Repository Not Found for url: https://huggingface.co/a/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -755,7 +755,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66777a61-5dcf491e4aa9e44f6fb7f1da;15f9e7ec-7d28-43ac-8eb9-54392c468a58) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cc025-2edd63691d25912742a8894c;a29ac1ad-c5d0-4bc7-bbd4-7d76c44ddf3e) Repository Not Found for url: https://huggingface.co/1/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -826,7 +826,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66777811-2e5d44c1513c1f52048e7f96;c0540a25-6837-4e4d-b82c-53a12497c829) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbde3-19f945d22ac1133420af9fc9;908c608f-14af-4cf9-b4db-af5119dfc7be) Repository Not Found for url: https://huggingface.co/t/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -916,7 +916,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667a1466-3396c2d6560cd53d6164fae3;15599486-4ec7-418a-9325-a99750665c3f) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667cb69a-00d7acd47d7233e72938c9c1;613dd07d-8e7a-448a-8da8-8092a93ed7a5) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/recurrentgemma-2b/resolve/main/config.json. @@ -1051,7 +1051,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667a1493-3dfb6bba548ab4767a9a2060;90bbf638-28bc-45cc-bf08-ab41b6d66825) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cb6c2-7c1363bc254a0beb1469acfc;ea8a38d8-2be3-4304-9712-c1a6caef8058) Repository Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -1112,7 +1112,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667a1409-5801cd1e6298084a7046c9b6;1ff2c613-51fc-42e6-969d-5e7d0ba16abe) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667cb648-11345f3956cca7e51a245863;121082cc-2793-404d-bc35-3c492d9e4504) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-2b/resolve/main/config.json. @@ -1260,7 +1260,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667a143a-025fd49f3074eab656635e11;e2ce13f9-abdb-4cd7-818f-9bd3ffd4d681) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667cb672-2106321024fd56af6d27e847;ddae412c-785f-42bf-adcd-a196891f4168) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-7b/resolve/main/config.json. @@ -1345,7 +1345,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667777c2-1d9bb857223e3e92772d05ab;40d9f9e4-8b5b-4fb1-8b21-696e173a8c40) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbd8a-44a034e01d1838f17a1eeffc;84706f3e-b8ae-4efa-aa5f-1ab9ff391927) Repository Not Found for url: https://huggingface.co/i/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -1426,7 +1426,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6677779a-5876ff635d9cbee11032ed4f;f16b3ee9-e7eb-4049-a656-d40a2e3f9881) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbd62-7dd4c1942791e8bd55afffc1;19e98039-7703-4f1a-b878-d1501f8b06b2) Repository Not Found for url: https://huggingface.co/M/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -1504,7 +1504,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66777905-7b7899082dd00a511087fbd7;ee6acb33-4324-4434-a684-ffbff4b83cfc) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbed2-62483b2d239e85021a3fb032;0c6072ef-6d65-4d4b-9880-6083ab50efc6) Repository Not Found for url: https://huggingface.co/x/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -1575,7 +1575,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66777b71-27abb4161a058a0514ce0a72;656ba79c-e84b-4131-aba5-4d0eedad084b) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667cc13b-5e7e595d21e34c7a757dcbab;e47396cf-017a-46e8-95da-dc33ad15e4ac) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json. @@ -1665,7 +1665,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667778da-055825bc12705ea02c3b813f;fc02811d-ff1b-41c2-8c99-3f3c3f1591d4) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbeaa-51731132442c9c0d53795b61;4a64d874-7c92-4486-98d1-eee42937b25f) Repository Not Found for url: https://huggingface.co/8/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -1754,7 +1754,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667a11d2-69b57fdc45f086621971e625;cc096f9d-1047-4f5f-a1b9-80008a9453e5) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667cb40f-0a70152a2b770a177780b20a;36f56417-b741-4b05-872e-5149aea6b372) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/databricks/dbrx-base/resolve/main/config.json. @@ -1861,7 +1861,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6677783b-71b71b8152836f9017f0fd0e;3e7ea32f-7abb-4ebf-95da-b60cc4383000) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbe0a-66538fbe2bc40ffc7c9990b5;6d2e2bb8-fbbf-45a7-9e97-8af92acd4c5a) Repository Not Found for url: https://huggingface.co/r/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -2152,7 +2152,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66777608-4cda6d9f20a596bc6203b598;084126f8-9565-42ae-b40e-9bda214c5e06) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbbc2-1333ea13403019660baf7b3c;804aed62-ca91-4726-8ce6-4bf6e71b22ae) Repository Not Found for url: https://huggingface.co/m/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -2224,7 +2224,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667779ed-795730510367f364317aa7c0;652fc65a-ff0f-47db-bd2b-baf955abb898) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbfa7-5a1148c6647145085f88b752;4917e9e2-1fbc-407f-9f4e-df12254ed09c) Repository Not Found for url: https://huggingface.co/v/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -2367,7 +2367,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66777a19-066cbcce2965da80269933ac;75ce4e0a-86b4-45cd-8e3f-11c3b0df2fe3) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbfd4-77727c4901355e026acbe11e;d6851fd1-e4f7-44be-9ab5-a4151fee6da4) Repository Not Found for url: https://huggingface.co/0/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -2478,7 +2478,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66777962-38c2a4e644f499415d071bb3;a9eb5625-34e9-4f75-ab01-8d0bcd864c28) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbf28-0be2a84e32a310b078df2925;415453de-80db-4675-989e-a44f482fc103) Repository Not Found for url: https://huggingface.co/2/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -2577,7 +2577,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66777891-5afe10283df8098a565b1465;2afb1bd8-e7f1-4c01-bd43-82600d319149) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbe62-6575548132bb7015033406ce;49c5245e-c6d9-4e46-adb6-560c2eae9497) Repository Not Found for url: https://huggingface.co/l/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -2730,7 +2730,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66777658-2dbbfac345026d5a390935c3;5dafaa97-89e3-42f9-b9fb-df10c324e691) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbc24-08514fc658cffae20e664d46;36dc63cf-3b1e-4cc7-8e90-7fd45d5484b2) Repository Not Found for url: https://huggingface.co/s/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -2800,7 +2800,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6677798e-49050cb736ec364733e8aacc;88f70ca6-59a9-448d-ac00-b0531f4148f1) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbf53-382a78fc04237a997f2b6339;45d21400-c9dd-49a3-80de-3f9dfb6c7abe) Repository Not Found for url: https://huggingface.co/B/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -2930,7 +2930,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66777869-1114b67b45ed0e785d6fdd48;bf348b30-ac31-493e-b0b2-4d70328f767c) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbe39-27426b0542cc30fc548e6c94;3ad61f53-ed64-4a16-871f-a8d6241eee45) Repository Not Found for url: https://huggingface.co/a/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -3058,7 +3058,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66777a68-087ae56b0661afbc6a46f011;897a35ee-06b5-4c1a-895b-9714818cb255) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cc02b-1ccbf5e666346cc3597fb6ee;f79643ec-ab1f-4917-89d0-eddeb7aa8eff) Repository Not Found for url: https://huggingface.co/1/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -3157,7 +3157,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66777818-6a9526167f1e306608b25804;8701590f-e31a-4f38-b30a-84195dae0782) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbde9-4aa93b16428063ad75db6595;f53f87c7-facf-4477-8a4a-42c7c9e7e638) Repository Not Found for url: https://huggingface.co/t/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -3275,7 +3275,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667a146d-63f1c7e766e0f1705a5c8082;d842abff-2b26-4188-af83-930ebd69caff) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667cb6a1-549bb83d698e962d274fbbd8;bfde7f63-3731-4dcf-8720-410863b8f81a) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/recurrentgemma-2b/resolve/main/config.json. @@ -3470,7 +3470,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667a1499-35da756b34d796bf5e9469ad;76b1c40f-2061-45f3-8a60-2aafc401c3d1) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cb6c8-274ac487101d7c2364bb24a2;89d2fa01-0c17-4060-aebf-ac3468f57815) Repository Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -3531,7 +3531,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667a1414-4f951c9b7acd78830fcb93bf;59855539-7278-4cf5-bbec-5df6376cb2db) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667cb64f-400a82f60dbc5cd261aac22f;f5fae84d-ae5a-4bb0-ae18-569d8a73a2d1) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-2b/resolve/main/config.json. @@ -3765,7 +3765,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667a1441-65c1eae23f86af274d3ae018;044a4f7d-b0e8-403b-8281-7c8338f51bf4) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667cb679-715c22e62c35572f13e18ca8;ee15508a-3711-4b87-b8af-d1a368b8584c) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-7b/resolve/main/config.json. @@ -3850,7 +3850,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667777c8-529fdba23c65a4f07507b821;3f5038b9-68b7-4a62-a718-08514a0dfd5b) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbd91-3e8c9c6c208cd70d4bed9ed2;21a0f6fe-ee32-4744-a31c-b677cad287b8) Repository Not Found for url: https://huggingface.co/i/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -3979,7 +3979,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667777a1-2a85b69e4c5e549106fcd4da;025c2f56-2c61-41da-a45f-6a9bf743b50c) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbd69-5b80c9e2198610cb147407f8;a730971a-4d4c-4990-aa2f-0f51833bcc4e) Repository Not Found for url: https://huggingface.co/M/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -4257,7 +4257,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6677790c-5aa04b3250d6f6e21f18aa29;7d1b7e26-c415-4412-8867-40bdea33cd82) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbed9-6cab8a245cb7edae7929bb72;7f7a652d-a57e-4304-9720-064e56025e33) Repository Not Found for url: https://huggingface.co/x/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -4348,7 +4348,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66777b78-05a1dae01917ee8079249a47;2a7cdf7b-cdb5-441e-b3eb-e16487284f75) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667cc141-201c131a1e1b9a5e35f82a09;2a2ffeb9-7d3e-4c03-8d0d-32d3e26dc7e6) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json. @@ -4521,7 +4521,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667778e1-7a7cd89b64187e586c2fec66;a641da2b-77fe-4920-90b8-367ca04d07b3) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbeb1-3c8103f428093dbe7757820c;6872358e-c927-4c79-90c9-64e46b3de192) Repository Not Found for url: https://huggingface.co/8/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -4610,7 +4610,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667a11d9-4ce4c44e196916db21729cdb;c2b56de1-3d0c-4d31-a2bc-32ba2ae9b50d) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667cb41b-6f5c59b1403c726e7e77b7ab;47939dbb-c740-42d3-8dbe-7c71aa6726b4) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/databricks/dbrx-base/resolve/main/config.json. @@ -4825,7 +4825,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66777842-39318a4d49b20ba52a0134f2;1214f229-e2cc-44f8-a62b-841d64371428) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbe11-16de9d183c5369576a4fe2b2;4cf928c1-335e-404d-9296-7dfb280a41c8) Repository Not Found for url: https://huggingface.co/r/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -5032,7 +5032,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpnx72l9l8/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpssiwn2y0/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 8bit-bnb-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-7b,tiiuae/falcon-7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.218-208.862.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): @@ -5061,7 +5061,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpu35oly06/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp92flzuco/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 8bit-bnb-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/polyglot-ko-12.8b,,cuda,0,42,,,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,MB,1280.970752,15055.978496,0.0,14401.142784,14091.235328,s,10,0.911755844116211,0.0911755844116211,0.003503716309260247,0.0899611701965332,0.09135012588500976,0.09651647071838378,0.10064954658508302,"[0.10168281555175782, 0.08995196533203124, 0.09020204925537109, 0.09018704223632812, 0.08998716735839844, 0.08992880249023437, 0.0899645767211914, 0.089957763671875, 0.08994207763671876, 0.08995158386230469]",tokens/s,2807.7692251936983,kWh,1.0672762825921015e-06,5.848164793172366e-07,4.081284546306311e-06,5.733377308215649e-06,tokens/kWh,44650820.31722638,MB,1280.970752,15055.978496,0.0,14401.142784,14373.645824,s,10,56.190349121093746,5.619034912109375,0.02299042134398943,5.61815380859375,5.6374814453125,5.655444091796875,5.669814208984375,"[5.67340673828125, 5.63348974609375, 5.6078583984375, 5.6256748046875, 5.6210537109375, 5.61525390625, 5.589076171875, 5.62949365234375, 5.59795361328125, 5.59708837890625]",tokens/s,11.211889761394973,kWh,6.606488198027001e-05,3.6207935308025915e-05,0.00015840742877789291,0.0002606802460661888,tokens/kWh,241675.3894884839,,s,629,56.90303385925292,0.09046587259022723,0.010788190315583574,0.0888248291015625,0.09056419982910156,0.09153167572021484,0.17906814758300782,"[0.09563340759277343, 0.09580134582519531, 0.09466777801513672, 0.0926033935546875, 0.09181593322753906, 0.08967375946044921, 0.08890982055664062, 0.08843772888183593, 0.08906861114501953, 0.08885446166992188, 0.08926207733154297, 0.09002393341064453, 0.08977101135253907, 0.08987750244140626, 0.0895436782836914, 0.08805382537841797, 0.09028601837158202, 0.09159270477294922, 0.09364582061767578, 0.09197977447509766, 0.08911974334716796, 0.08923442840576172, 0.08883609771728515, 0.08881664276123047, 0.0892200927734375, 0.0888084487915039, 0.08843065643310546, 0.08809056091308594, 0.08846131134033203, 0.08865792083740234, 0.0889169921875, 0.08865484619140625, 0.09008537292480469, 0.09012326049804688, 0.08916275024414062, 0.08890367889404296, 0.09002803039550782, 0.08879312133789062, 0.09043247985839843, 0.08949247741699219, 0.09023078155517578, 0.08963174438476562, 0.0891494369506836, 0.09034239959716797, 0.08892729949951172, 0.08855443572998047, 0.08864972686767578, 0.09032704162597656, 0.08862105560302734, 0.08865382385253906, 0.08996044921875, 0.08907161712646484, 0.09264537811279297, 0.0897259521484375, 0.09017139434814453, 0.09024205017089844, 0.08983245086669922, 0.08925081634521484, 0.08993177795410157, 0.08881356811523437, 0.08985600280761719, 0.09134591674804687, 0.1800427551269531, 0.08937779235839843, 0.09021952056884766, 0.08930713653564454, 0.08967987060546875, 0.08884432220458985, 0.09044477081298828, 0.08937062072753907, 0.09019187164306641, 0.0900843505859375, 0.09056153869628907, 0.09008230590820313, 0.08906752014160156, 0.08896717071533203, 0.0893460464477539, 0.08846540832519531, 0.0888309783935547, 0.08988569641113281, 0.08939212799072266, 0.08937983703613281, 0.08940544128417968, 0.09017549133300781, 0.08905625915527343, 0.08908697509765626, 0.08904806518554688, 0.09027174377441406, 0.08899584197998046, 0.08811827087402344, 0.08806092834472656, 0.08937165069580078, 0.08884019470214843, 0.08860569763183594, 0.08889036560058594, 0.0890777587890625, 0.08860272216796874, 0.08943606567382813, 0.0903720932006836, 0.08918118286132813, 0.08861081695556641, 0.08817254638671874, 0.08963791656494141, 0.0888043212890625, 0.08874188995361328, 0.08855859375, 0.08917401885986329, 0.08873779296875, 0.08895590209960938, 0.09063014221191407, 0.0934103012084961, 0.09070285034179687, 0.09282150268554687, 0.08926105499267578, 0.08840704345703125, 0.08865692901611329, 0.08886678314208984, 0.08939417266845703, 0.08903577423095703, 0.08914841461181641, 0.09056050872802734, 0.08847666931152344, 0.08911360168457032, 0.08892623901367187, 0.09052464294433593, 0.18026290893554686, 0.09159884643554687, 0.09052671813964844, 0.09308262634277344, 0.09123123168945313, 0.08937881469726562, 0.08895692443847657, 0.08857190704345703, 0.08876953887939452, 0.0886446075439453, 0.08878598022460937, 0.08888620758056641, 0.08876338958740235, 0.08851353454589844, 0.08919660949707031, 0.09101920318603515, 0.08886380767822266, 0.0884755859375, 0.08855142211914062, 0.08854528045654297, 0.08779468536376953, 0.08868556976318359, 0.08835481262207032, 0.08848489379882812, 0.08814588928222657, 0.08789094543457031, 0.0888084487915039, 0.08872345733642578, 0.08870912170410156, 0.08860678100585938, 0.08863225555419922, 0.08824320220947265, 0.08874086761474609, 0.08859954833984375, 0.08831488037109375, 0.0887562255859375, 0.08898252868652344, 0.09153536224365234, 0.09118105316162109, 0.08888422393798828, 0.08850534057617188, 0.08875929260253906, 0.08930099487304688, 0.08871116638183593, 0.0886456298828125, 0.08868966674804687, 0.08873779296875, 0.08862105560302734, 0.08871424102783203, 0.08837939453125, 0.08884838104248047, 0.0890245132446289, 0.08877875518798828, 0.08846540832519531, 0.08852377319335937, 0.08870502471923829, 0.08910550689697265, 0.08923126220703125, 0.08878284454345703, 0.08849305725097656, 0.08854220581054688, 0.08906956481933594, 0.08879513549804688, 0.179198974609375, 0.089059326171875, 0.08898047637939453, 0.08859954833984375, 0.08874188995361328, 0.08884633636474609, 0.08884333038330078, 0.08872544097900391, 0.08922828674316406, 0.08896614074707031, 0.08875520324707031, 0.08898560333251954, 0.08993689727783204, 0.09102134704589844, 0.08864150238037109, 0.08862310028076172, 0.08874905395507812, 0.08847257232666016, 0.08898457336425782, 0.08916889953613281, 0.08887296295166015, 0.08859136199951172, 0.08856575775146484, 0.08861695861816406, 0.08863641357421875, 0.08884121704101562, 0.08859136199951172, 0.08876032257080078, 0.08867737579345703, 0.08836812591552734, 0.08881561279296875, 0.08884941101074219, 0.09208319854736328, 0.08874598693847656, 0.08885350036621094, 0.09429401397705078, 0.08980582427978516, 0.08934706878662109, 0.089818115234375, 0.08979052734375, 0.08960403442382812, 0.08980992126464844, 0.08962662506103515, 0.08860163116455078, 0.08847663879394531, 0.08892825317382813, 0.09077555084228515, 0.09088409423828125, 0.090029052734375, 0.09089433288574218, 0.08916582489013672, 0.0916684799194336, 0.09086156463623046, 0.09048371124267578, 0.08888626861572266, 0.08887296295166015, 0.08889651489257812, 0.08875417327880859, 0.08879206085205078, 0.0886302719116211, 0.08851046752929688, 0.0885063705444336, 0.0887715835571289, 0.1790873565673828, 0.0886138916015625, 0.08829952239990234, 0.08850943756103516, 0.08878598022460937, 0.0886025619506836, 0.08858316802978515, 0.08816947174072266, 0.08859442901611328, 0.08869068908691406, 0.08875724792480469, 0.0879636459350586, 0.0906219482421875, 0.08922112274169922, 0.08890573120117187, 0.08819097900390625, 0.08866815948486328, 0.08924467468261718, 0.0883978271484375, 0.08852787017822265, 0.08843775939941406, 0.09123532867431641, 0.08988569641113281, 0.08953043365478516, 0.08961222076416016, 0.08972185516357421, 0.08925286102294921, 0.08868351745605468, 0.08843571472167969, 0.08855654144287109, 0.08849100494384765, 0.08853094482421875, 0.08851148986816407, 0.08863846588134766, 0.08949555206298829, 0.08989286041259766, 0.08865280151367187, 0.0888248291015625, 0.09010688018798828, 0.08883814239501953, 0.08964198303222656, 0.0890224609375, 0.08917810821533204, 0.09040589141845704, 0.0900874252319336, 0.09133776092529297, 0.09127318572998047, 0.088953857421875, 0.088953857421875, 0.08924979400634765, 0.08956723022460937, 0.08911360168457032, 0.08889344024658204, 0.08959283447265624, 0.0890613784790039, 0.08980992126464844, 0.09030553436279297, 0.0905881576538086, 0.08932044982910156, 0.0902451171875, 0.0890091552734375, 0.0905011215209961, 0.08919859313964844, 0.1797570495605469, 0.08919152069091797, 0.08989276885986328, 0.09019699096679687, 0.09020518493652344, 0.08884633636474609, 0.08860262298583985, 0.08851769256591797, 0.0897484130859375, 0.09059123229980469, 0.09004032135009765, 0.08961331176757813, 0.08879414367675781, 0.09001980590820312, 0.08915865325927734, 0.0908953628540039, 0.09152614593505859, 0.08884326171875, 0.08911974334716796, 0.09173299407958985, 0.09045094299316406, 0.08962969970703125, 0.08883916473388671, 0.08865078735351563, 0.08867222595214844, 0.08906547546386719, 0.08866304016113281, 0.08861695861816406, 0.08890470123291015, 0.08905625915527343, 0.08878284454345703, 0.08887910461425781, 0.0888309783935547, 0.08851967620849609, 0.0885524444580078, 0.09017446136474609, 0.08971571350097657, 0.08914022064208985, 0.08856678771972656, 0.08845516967773437, 0.08873267364501954, 0.08884838104248047, 0.08881254577636719, 0.088342529296875, 0.088331298828125, 0.08850121307373048, 0.08843981170654297, 0.08860569763183594, 0.08851148986816407, 0.08871321868896484, 0.0890808334350586, 0.08856269073486328, 0.08867021179199219, 0.08865280151367187, 0.08879411315917969, 0.0888279037475586, 0.08886483001708985, 0.08875001525878906, 0.08844287872314453, 0.08874188995361328, 0.08904192352294922, 0.08877875518798828, 0.0887224349975586, 0.17901875305175782, 0.08858214569091796, 0.08852377319335937, 0.08847052764892578, 0.08856371307373047, 0.09038848114013671, 0.08959283447265624, 0.08846745300292969, 0.08873884582519531, 0.08847459411621093, 0.08998502349853515, 0.088806396484375, 0.08873270416259765, 0.08851350402832031, 0.08923442840576172, 0.0883763198852539, 0.08847666931152344, 0.08833433532714843, 0.08853913879394532, 0.08896102142333985, 0.08875110626220703, 0.08846546936035156, 0.08872032165527344, 0.08880435180664062, 0.08854732513427735, 0.08834969329833985, 0.08838349151611329, 0.08835379028320313, 0.0886476821899414, 0.08830770874023437, 0.08847052764892578, 0.08830464172363281, 0.08960205078125, 0.08854732513427735, 0.08877362823486327, 0.08805375671386718, 0.08761344146728516, 0.08860057830810547, 0.08845926666259765, 0.08865074920654296, 0.08887602996826172, 0.08902758026123046, 0.08902349090576171, 0.08860160064697266, 0.0883927001953125, 0.08863436889648438, 0.08900505828857422, 0.090029052734375, 0.08875929260253906, 0.08860060882568359, 0.08884323120117188, 0.08851455688476563, 0.08909926605224609, 0.0889722900390625, 0.08867737579345703, 0.08832717132568359, 0.08853708648681641, 0.08906034851074218, 0.08870297241210938, 0.08868863677978515, 0.08841318511962891, 0.08852582550048828, 0.08858726501464843, 0.1811179504394531, 0.09168281555175781, 0.0910387191772461, 0.08872652435302734, 0.09157427215576172, 0.08882994842529297, 0.0912332763671875, 0.08960409545898437, 0.08876441955566407, 0.08879417419433594, 0.08938182067871094, 0.08908595275878907, 0.08931839752197265, 0.08953036499023438, 0.08840089416503906, 0.08855757141113281, 0.08861087799072266, 0.08875718688964844, 0.0889886703491211, 0.08883507537841796, 0.09120870208740234, 0.0889354248046875, 0.0889200668334961, 0.09017139434814453, 0.0885401611328125, 0.09057484436035156, 0.09145037078857422, 0.0887357406616211, 0.09021440124511719, 0.08873474884033203, 0.08950985717773438, 0.09025638580322265, 0.08948941040039063, 0.09198592376708985, 0.09077350616455078, 0.08855654144287109, 0.09060044860839844, 0.08890879821777344, 0.0886118392944336, 0.08994815826416015, 0.08982835388183594, 0.09078479766845703, 0.08890262603759766, 0.08843878173828125, 0.08852684783935547, 0.08863948822021485, 0.08868966674804687, 0.08841318511962891, 0.08846438598632812, 0.08843366241455078, 0.08817049407958985, 0.08823910522460937, 0.08853094482421875, 0.0891361312866211, 0.08953446197509765, 0.08867635345458984, 0.08880435180664062, 0.08883712005615234, 0.08851660919189454, 0.08844287872314453, 0.08944947052001953, 0.08840908813476563, 0.08865280151367187, 0.17881907653808593, 0.08859442901611328, 0.08859033966064453, 0.08999321746826172, 0.08891085052490234, 0.08854425811767579, 0.08847360229492188, 0.08889036560058594, 0.08820531463623046, 0.08935424041748047, 0.08840294647216797, 0.0885186538696289, 0.08860467529296875, 0.08960511779785156, 0.08912281799316406, 0.08925491333007812, 0.0885524444580078, 0.08846438598632812, 0.08842034912109376, 0.08860774230957032, 0.08824217224121093, 0.08945356750488281, 0.08896409606933593, 0.08848896026611328, 0.08946278381347657, 0.08854937744140624, 0.08882073974609375, 0.08870604705810547, 0.08820531463623046, 0.0887910385131836, 0.08850841522216797, 0.08841728210449219, 0.0884142074584961, 0.08876032257080078, 0.08838963317871094, 0.08873267364501954, 0.08856575775146484, 0.08869478607177735, 0.08855039978027343, 0.0884142074584961, 0.08823193359375, 0.08837120056152344, 0.08842034912109376, 0.08867737579345703, 0.08846438598632812, 0.08844697570800782, 0.08890777587890625, 0.08893856048583984, 0.08988256072998047, 0.09074073791503906, 0.08874188995361328, 0.0885032958984375, 0.09115955352783203, 0.09228288269042968, 0.08916070556640625, 0.08868045043945312, 0.08863948822021485, 0.08855347442626953, 0.08844390106201172, 0.08854220581054688, 0.08874188995361328, 0.08864665222167968, 0.08968601226806641, 0.179198974609375, 0.08783257293701172, 0.08828518676757813, 0.08802512359619141, 0.09111650848388672, 0.08991539001464843, 0.08878387451171875, 0.08888934326171875, 0.08867737579345703, 0.08844697570800782, 0.0892600326538086, 0.08864256286621094, 0.08862515258789062, 0.08854732513427735, 0.08844902038574219, 0.08837017822265625, 0.08905625915527343, 0.08858316802978515, 0.08906342315673828, 0.0887715835571289, 0.08865382385253906, 0.08851763153076173, 0.08859340667724609, 0.08834969329833985, 0.08993075561523438, 0.08878182220458984, 0.08858828735351562, 0.0884510726928711, 0.08838349151611329, 0.0891361312866211, 0.08937267303466796, 0.09036902618408203, 0.08898457336425782, 0.08876137542724609, 0.08834249877929687, 0.08855142211914062, 0.08849308776855469, 0.08859235382080079, 0.09291366577148437, 0.09064755249023437, 0.0885432357788086, 0.08920780944824219, 0.08876338958740235, 0.08868351745605468, 0.08851148986816407, 0.08855039978027343, 0.08836710357666015, 0.08864972686767578, 0.08863539123535157, 0.08854937744140624, 0.08885862731933594, 0.08887296295166015, 0.0889886703491211, 0.08871116638183593, 0.088880126953125, 0.08814591979980468, 0.08878387451171875, 0.08865484619140625, 0.08849919891357422, 0.08836505889892578, 0.08813977813720703, 0.08861491394042968, 0.0881080322265625]",tokens/s,11.053892162512863,,,main,False,False, @@ -5161,7 +5161,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6677761c-6943da0a301f08063b7ae39f;e3351b57-5caa-4ec6-81be-1d704cfa3466) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbbdc-5385222e0862f4b92578f785;41076877-8d90-42af-8b74-084b1a2213e0) Repository Not Found for url: https://huggingface.co/m/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -5233,7 +5233,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66777a05-75a936ba209aecd97a2f2e3b;a362f9d0-5c9a-4bf2-86fb-0c69563bdafc) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbfbd-6ad48a41268ea1db5cee02dd;944336b6-2c08-4be9-8178-239413e9d04b) Repository Not Found for url: https://huggingface.co/v/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -5317,7 +5317,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66777a2d-1e95ce0b321f2e0a1d31e8e1;65b9040f-178a-4351-92fd-54707b1e4c6b) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbfee-348ff90a549c805d39914783;a54f6e3d-8780-4f6b-b63d-9b050faed7a3) Repository Not Found for url: https://huggingface.co/0/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -5396,7 +5396,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66777978-56fbdda10864d37d4082343c;858841f9-171f-4c5b-a57e-195c69dce924) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbf3d-5eec8f105ae9cf181155d0da;b70b4891-079a-43eb-95f6-5db154d84b59) Repository Not Found for url: https://huggingface.co/2/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -5467,7 +5467,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667778a5-2eff531127f98b472c3a6281;022b7d63-68dd-4d0b-9ae4-2c3ea7c6be06) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbe76-722a95bc4286dae062dc85ad;9e40a538-f554-411b-8642-0619888e5c9c) Repository Not Found for url: https://huggingface.co/l/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -5564,7 +5564,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6677766d-2efdf1843744b09c31070b4b;eedbfeac-d332-4948-bbf2-542f4a8b583a) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbc38-798e43e63db6c78251aeff2c;913a269f-976d-4c12-8aaf-ec1542c22743) Repository Not Found for url: https://huggingface.co/s/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -5634,7 +5634,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667779a5-3721340445e63d02444aed8c;ee7cfee4-b935-412b-81b3-b66fb615e2b2) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbf6a-63667fcd0358cb075f144cc7;7047020b-bda3-4612-becd-34051dbb9f80) Repository Not Found for url: https://huggingface.co/B/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -5693,7 +5693,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: CodeGenForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp7a4kx4lh/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: CodeGenForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpz5xhaz7c/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 8bit-bnb-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-2.7b,,cuda,0,42,,,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,MB,1753.206784,4370.989056,0.0,3716.153344,3502.931968,s,10,0.7158005905151366,0.07158005905151367,0.0026151097512319512,0.07056062698364257,0.07317087783813477,0.0761810390472412,0.07858916801452637,"[0.07919120025634765, 0.07085072326660156, 0.07094694519042968, 0.07031388854980469, 0.07023776245117187, 0.07029948425292969, 0.07049616241455078, 0.072501953125, 0.07033737945556641, 0.07062509155273437]",tokens/s,3576.4150434098656,kWh,8.388341447797885e-07,4.5958770674104095e-07,1.8877674676594226e-06,3.1861893191802524e-06,tokens/kWh,80346763.5959134,MB,1753.206784,4370.989056,0.0,3716.153344,3632.817664,s,10,44.667627929687505,4.46676279296875,0.018600767003160216,4.460619384765625,4.485654638671876,4.498310229492188,4.5084347021484374,"[4.48284228515625, 4.4655009765625, 4.45241162109375, 4.45341796875, 4.44769091796875, 4.4522490234375, 4.481310546875, 4.4602392578125, 4.46099951171875, 4.5109658203125]",tokens/s,14.104174078634749,kWh,5.310124745159473e-05,2.9102769588021433e-05,9.395025364714023e-05,0.0001761542706867564,tokens/kWh,357641.05947807967,,s,629,45.231303550720156,0.07190986256076345,0.008457534828436213,0.07048806762695313,0.07254323120117187,0.07291353302001953,0.14091268249511718,"[0.07149763488769531, 0.07236300659179687, 0.07048397064208985, 0.07265382385253906, 0.07177318572998047, 0.07225138854980469, 0.07384473419189454, 0.07193087768554687, 0.07062732696533203, 0.07043583679199218, 0.070434814453125, 0.07033241271972657, 0.07061817932128907, 0.07013779449462891, 0.07195852661132812, 0.07246438598632812, 0.07255142211914062, 0.07255961608886718, 0.07078912353515625, 0.07081983947753906, 0.07044915008544922, 0.07034265899658203, 0.07060787200927734, 0.07021568298339843, 0.07060275268554687, 0.07041433715820312, 0.07039282989501953, 0.07039385223388672, 0.07066316986083984, 0.07061299133300782, 0.07034162902832031, 0.07017164611816407, 0.07033654022216797, 0.07016553497314452, 0.0703917465209961, 0.07013990020751953, 0.07160626983642578, 0.07050240325927734, 0.07033548736572266, 0.07311052703857422, 0.07278694152832031, 0.07141990661621093, 0.07023411560058594, 0.07154585266113281, 0.07023104095458985, 0.07205171203613281, 0.07074508666992188, 0.07263231658935547, 0.07006617736816406, 0.07114649963378906, 0.07070105743408203, 0.07260057830810547, 0.07075430297851562, 0.07085670471191406, 0.07270502471923829, 0.0717649917602539, 0.070434814453125, 0.07098060607910156, 0.07048089599609375, 0.07204761505126953, 0.07027814483642578, 0.07079833221435547, 0.14196018981933595, 0.07093965148925781, 0.07223808288574218, 0.0704194564819336, 0.07073382568359375, 0.0712837142944336, 0.07221247863769531, 0.07236198425292968, 0.0719656982421875, 0.07015936279296875, 0.07168716430664063, 0.07038979339599609, 0.07106658935546875, 0.0719288330078125, 0.07191756439208985, 0.07084031677246094, 0.07375462341308593, 0.07192269134521484, 0.07429017639160156, 0.07091712188720703, 0.07006105804443359, 0.0702371826171875, 0.07047277069091797, 0.0703354263305664, 0.07038566589355469, 0.07033856201171874, 0.07036927795410156, 0.07030681610107421, 0.07035903930664063, 0.07051980590820313, 0.0704532470703125, 0.07031910705566406, 0.07042253112792969, 0.0697681884765625, 0.07008972930908203, 0.07012351989746093, 0.06991155242919922, 0.07039590454101563, 0.0702730255126953, 0.0704368667602539, 0.0709191665649414, 0.07057107543945312, 0.07038355255126953, 0.06974566650390625, 0.07057408142089844, 0.07198207855224609, 0.07142912292480469, 0.07038668823242188, 0.07020134735107422, 0.07028018951416015, 0.0721817626953125, 0.07068876647949218, 0.0739788818359375, 0.07211827087402344, 0.07048703765869141, 0.07026073455810547, 0.07033856201171874, 0.06981843566894531, 0.07036307525634766, 0.07035903930664063, 0.07015936279296875, 0.0702525405883789, 0.07032115173339844, 0.14248550415039063, 0.0704532470703125, 0.07049625396728515, 0.07205888366699219, 0.07040105438232422, 0.07024227142333984, 0.0702730255126953, 0.0700549087524414, 0.07016038513183594, 0.07020751953125, 0.07015318298339844, 0.07036006164550782, 0.07208345794677734, 0.07120384216308594, 0.07030886077880859, 0.07037747192382812, 0.07025459289550781, 0.07048806762695313, 0.07146086120605469, 0.07031705474853515, 0.07024639892578124, 0.07025357055664062, 0.07015628814697265, 0.07031603240966797, 0.07018701171875, 0.07035289764404297, 0.07062220764160156, 0.07034982299804687, 0.07380480194091797, 0.07419801330566406, 0.07048499298095703, 0.07023104095458985, 0.07022489929199219, 0.07049625396728515, 0.07031910705566406, 0.06968934631347656, 0.07044915008544922, 0.07046144104003907, 0.07019007873535156, 0.0704931869506836, 0.07031504058837891, 0.07061500549316406, 0.07025459289550781, 0.06980095672607421, 0.07020543670654297, 0.07057920074462891, 0.07047270202636718, 0.07049215698242188, 0.07036313629150391, 0.0703969955444336, 0.07029548645019532, 0.07073689270019531, 0.07023206329345703, 0.07331123352050781, 0.07239373016357421, 0.07279411315917969, 0.07025049591064453, 0.07048601531982422, 0.07033139038085938, 0.07038771057128906, 0.07031398773193359, 0.07084646606445312, 0.07017683410644532, 0.1412003173828125, 0.07280127716064454, 0.07291295623779297, 0.07079011535644532, 0.07045833587646484, 0.06992281341552735, 0.07036006164550782, 0.07027712249755859, 0.07034368133544922, 0.070329345703125, 0.07038361358642578, 0.07045938873291016, 0.07043276977539062, 0.07054847717285156, 0.07040716552734375, 0.07084953308105468, 0.07044198608398437, 0.07054541015625, 0.07037337493896484, 0.07072358703613281, 0.07076454162597656, 0.07148953247070312, 0.07052082824707032, 0.0703795166015625, 0.0707430419921875, 0.07142809295654297, 0.07075020599365234, 0.07038259124755859, 0.07049215698242188, 0.07050035095214843, 0.07211110687255859, 0.07184690856933594, 0.07365119934082032, 0.0705771484375, 0.06972825622558594, 0.0703672332763672, 0.07038566589355469, 0.07066214752197265, 0.07053619384765625, 0.07096217346191407, 0.07095398712158203, 0.0704368667602539, 0.07037849426269531, 0.07100518035888671, 0.07086080169677735, 0.07033446502685547, 0.07027609252929687, 0.07033446502685547, 0.07015017700195313, 0.07036412811279297, 0.07049420928955077, 0.07031603240966797, 0.07026585388183594, 0.07072870635986328, 0.07130521392822266, 0.07034572601318359, 0.07048703765869141, 0.07038976287841797, 0.07045529937744141, 0.07031807708740234, 0.07038566589355469, 0.07027097320556641, 0.07006105804443359, 0.14092185974121094, 0.0704901123046875, 0.07320371246337891, 0.07042150115966797, 0.07148339080810547, 0.07050240325927734, 0.06999040222167968, 0.07028530883789062, 0.0701317138671875, 0.07035596466064453, 0.07015526580810547, 0.07016550445556641, 0.07022592163085938, 0.07046144104003907, 0.07004672241210938, 0.07037337493896484, 0.07001395416259766, 0.07048703765869141, 0.07069593811035156, 0.0704676513671875, 0.07074195098876954, 0.0702485122680664, 0.0703180160522461, 0.07263129425048828, 0.07168307495117188, 0.07106047821044922, 0.07031193542480468, 0.0709939193725586, 0.07001087951660157, 0.0700231704711914, 0.06932275390625, 0.0694999008178711, 0.07037133026123046, 0.07136972808837891, 0.06993408203125, 0.06980403137207031, 0.07046963500976562, 0.07041843414306641, 0.07022796630859375, 0.07048397064208985, 0.07091302490234375, 0.07066422271728516, 0.07354160308837891, 0.07238349151611329, 0.07298252868652344, 0.07026687622070313, 0.07172300720214844, 0.07047782135009766, 0.07016754913330078, 0.07050137329101562, 0.07046348571777344, 0.0704194564819336, 0.07021363067626953, 0.07042662048339844, 0.07023513793945313, 0.07038361358642578, 0.07011328125, 0.07033245086669922, 0.07018902587890626, 0.07036927795410156, 0.06995763397216796, 0.07021363067626953, 0.07022182464599609, 0.14130278015136719, 0.07017369842529297, 0.07012454223632812, 0.07298764801025391, 0.07130521392822266, 0.07037849426269531, 0.070181884765625, 0.07114342498779297, 0.07037644958496093, 0.07040921783447265, 0.07054438018798828, 0.07054847717285156, 0.07034060668945312, 0.0706529312133789, 0.07021977233886718, 0.0706529312133789, 0.07132982635498047, 0.07049520111083984, 0.0703672332763672, 0.07049113464355469, 0.07051776123046875, 0.07055667114257813, 0.07032627105712891, 0.07019929504394531, 0.07023411560058594, 0.0705638427734375, 0.0705955810546875, 0.07053311920166015, 0.07051264190673828, 0.07064784240722656, 0.07040406036376953, 0.07043993377685547, 0.07029350280761719, 0.07095097351074219, 0.07063545227050781, 0.07042969512939454, 0.0700794906616211, 0.07004876708984376, 0.07055564880371094, 0.07064268493652344, 0.0701470718383789, 0.07053823852539062, 0.07053209686279296, 0.07260671997070313, 0.071804931640625, 0.07340850830078124, 0.07225446319580078, 0.07109529876708984, 0.07053926086425781, 0.0705269775390625, 0.07047885131835938, 0.07037849426269531, 0.07057817840576172, 0.07064985656738282, 0.07054950714111329, 0.07037337493896484, 0.0700426254272461, 0.07076044464111328, 0.07062937927246093, 0.07047270202636718, 0.07053215789794921, 0.07040300750732421, 0.07008665466308593, 0.1408890838623047, 0.0700579833984375, 0.07040306854248046, 0.07095603179931641, 0.07067135620117188, 0.07052902221679687, 0.07032217407226563, 0.07087411499023437, 0.07056179046630859, 0.07040306854248046, 0.07119155120849609, 0.07045222473144531, 0.07153561401367188, 0.07102361297607422, 0.07295692443847657, 0.07263641357421875, 0.07663104248046874, 0.0731514892578125, 0.07261190032958985, 0.07177107238769531, 0.07096627044677735, 0.07251455688476563, 0.07052288055419922, 0.07034982299804687, 0.07089356994628906, 0.07131443023681641, 0.07168000030517578, 0.07194931030273438, 0.07047577667236328, 0.07285247802734375, 0.07253094482421875, 0.07324364471435547, 0.07126322937011718, 0.0699504623413086, 0.0695050277709961, 0.06952652740478515, 0.0703641586303711, 0.07017676544189454, 0.06994944000244141, 0.06934425354003906, 0.0704686050415039, 0.0703969955444336, 0.07023200225830079, 0.07009894561767578, 0.0696596450805664, 0.07089459228515625, 0.07034674835205078, 0.0717158432006836, 0.07167692565917969, 0.07193702697753906, 0.07162572479248047, 0.07148748779296875, 0.07270706939697266, 0.07029759979248047, 0.0700979232788086, 0.07177833557128906, 0.07124784088134765, 0.07156326293945313, 0.0718397445678711, 0.07128985595703125, 0.07057920074462891, 0.07054438018798828, 0.07034880065917969, 0.14348287963867187, 0.07049830627441406, 0.07024639892578124, 0.0706170883178711, 0.06970982360839843, 0.07032524871826172, 0.07015833282470703, 0.07006208038330078, 0.07034371185302735, 0.07042147064208984, 0.0702003173828125, 0.07272038269042969, 0.07259852600097656, 0.07238041687011719, 0.07269478607177735, 0.07254118347167969, 0.07344640350341797, 0.07283404541015626, 0.0721981430053711, 0.07240908813476563, 0.07258624267578125, 0.07021670532226562, 0.07021670532226562, 0.07076249694824219, 0.07023308563232422, 0.07016960144042969, 0.0699504623413086, 0.07017472076416016, 0.07204351806640626, 0.07038873291015625, 0.07032012939453125, 0.07037439727783203, 0.07026687622070313, 0.07033139038085938, 0.0702525405883789, 0.07038259124755859, 0.07035801696777344, 0.0702894058227539, 0.07015731048583984, 0.07010304260253906, 0.0701470718383789, 0.07023312377929687, 0.07039177703857422, 0.07029657745361328, 0.0702371826171875, 0.07127552032470703, 0.07026278686523438, 0.07011532592773438, 0.07021670532226562, 0.0706324462890625, 0.07047885131835938, 0.07055052947998047, 0.07046451568603515, 0.07022284698486328, 0.07026892852783204, 0.07008460998535156, 0.07233433532714843, 0.07216537475585938, 0.07277977752685547, 0.07074610900878907, 0.06977126312255859, 0.06972006225585937, 0.06992998504638671, 0.14085635375976563, 0.06998320007324219, 0.070329345703125, 0.07024127960205079, 0.07040921783447265, 0.06970572662353515, 0.07047065734863281, 0.07047577667236328, 0.07034880065917969, 0.07042047882080078, 0.07067340850830078, 0.07035801696777344, 0.07065087890625, 0.07035494232177734, 0.07040921783447265, 0.07037542724609375, 0.07243059539794922, 0.07194419097900391, 0.07237017822265625, 0.070329345703125, 0.07091404724121093, 0.07100927734375, 0.07039794921875, 0.07029964447021485, 0.07054438018798828, 0.07038976287841797, 0.07050137329101562, 0.07172505950927735, 0.073301025390625, 0.07376380920410157, 0.0728616943359375, 0.07057202911376953, 0.07004160308837891, 0.06957158660888672, 0.07051570892333985, 0.07044198608398437, 0.07060582733154297, 0.0704716796875, 0.07048806762695313, 0.07044403076171875, 0.07057510375976563, 0.07046041870117188, 0.07126732635498047, 0.07257907104492188, 0.07077581024169922, 0.07048601531982422, 0.07047987365722656, 0.0702730255126953, 0.06994739532470703, 0.07239577484130859, 0.07052902221679687, 0.07033036804199219, 0.07229440307617188, 0.07042969512939454, 0.07047782135009766, 0.07042156982421875, 0.07127855682373047, 0.07062627410888672, 0.07017164611816407, 0.0716042251586914, 0.07137177276611328, 0.07010201263427734, 0.07049420928955077, 0.14187315368652345, 0.07062322998046874, 0.07053721618652344, 0.07057612609863281, 0.07067750549316407, 0.070793212890625, 0.07038566589355469, 0.07079936218261719, 0.0702730255126953, 0.070614013671875, 0.07046553802490234, 0.0705116195678711, 0.07048397064208985, 0.07051673889160157, 0.0705423355102539, 0.07029043579101563, 0.07032319641113281, 0.0703078384399414, 0.07160012817382813, 0.07250431823730469, 0.07043788909912109, 0.07060384368896484, 0.07045728302001954, 0.07036825561523438, 0.07142400360107422, 0.07265177917480468, 0.07246028900146484, 0.07253606414794922, 0.07204659271240234, 0.07295078277587891, 0.0729139175415039, 0.07195340728759765, 0.072416259765625, 0.07247154998779297, 0.07184690856933594, 0.07263231658935547, 0.07274086761474609, 0.0717096939086914, 0.07236402893066406, 0.07166976165771484, 0.07194931030273438, 0.07173426818847656, 0.07251148986816407, 0.07147315216064454, 0.07262515258789062, 0.0719482879638672, 0.07177523040771484, 0.07152435302734375, 0.0726824951171875, 0.07250125122070313, 0.07197081756591797, 0.07249203491210937, 0.07244595336914063, 0.07192063903808593, 0.07164313507080078, 0.07215821075439453, 0.07262723541259766, 0.071847900390625, 0.07213568115234376, 0.07099801635742188, 0.07239577484130859, 0.07256371307373047, 0.07234559631347656]",tokens/s,13.906298307203762,,,main,False,False, @@ -5736,7 +5736,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6677787d-5723ea7031cd0eff43c9edea;e1c07994-1004-4ad8-95f4-60b8eb25dd0b) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbe4e-0258505163d30a930c386998;4e55a9a0-96e3-455d-8e75-6e688176022f) Repository Not Found for url: https://huggingface.co/a/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -5794,7 +5794,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp2bm46unq/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpnp9_wh0z/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 8bit-bnb-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,1,1,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.218-208.862.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): @@ -5836,7 +5836,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66777a7c-103b14e37db6259e07498614;112a5332-4dc4-477b-870e-795a40acdd7a) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cc03f-625df1fe70cf2bda3a14e750;1482927f-81f9-49ef-b6d7-cfcc4d3f5089) Repository Not Found for url: https://huggingface.co/1/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -5907,7 +5907,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6677782b-6ead48736e2ab23a3a5269fe;d75bde6f-18fb-442b-8407-af013204ba2c) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbdfd-3c6f3c30526b4d7861a62c16;194fb7a7-c762-46c7-a697-6790ec51aa93) Repository Not Found for url: https://huggingface.co/t/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -5997,7 +5997,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667a1485-5bda482b1becebf26e06eed6;daecc226-2804-4d1c-99eb-728b569575b9) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667cb6b4-3f1ac7515b749e5b77c30e15;c1efea31-15ee-4907-adeb-549e35ef98da) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/recurrentgemma-2b/resolve/main/config.json. @@ -6118,7 +6118,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: CodeGenForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpf_wrkr98/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: CodeGenForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp6zp9p81k/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 8bit-bnb-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/recurrentgemma-7b,google/recurrentgemma-7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.218-208.862.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): @@ -6160,7 +6160,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667a14ad-7764cd5e3083b8573945a4bf;46c746c8-d734-4eae-98d9-50e76407ba6b) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cb6dd-5c0c7b2024d8fa7d2c4c944e;038a0034-4dbf-46dd-984e-13544fde3f47) Repository Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -6221,7 +6221,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667a142a-73bb2d150e2cb01f2afc904f;86dcacd6-b3e6-4392-96fe-1ac878567c6a) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667cb664-63e44b7f66f5c7761d722bd6;2d32fe8e-7bb7-48a5-9bbb-493a413268ef) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-2b/resolve/main/config.json. @@ -6369,7 +6369,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667a1456-675ebc755000167e43d8c147;60601b45-fe61-473b-b09d-c1ebb3c21986) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667cb68d-2f6c04033808365e79ea999f;202c0216-f480-431f-bc6b-8e54f95107aa) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-7b/resolve/main/config.json. @@ -6454,7 +6454,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667777dc-6ec03f50740a8e2a7f86cdf0;e563a7f9-aeed-4d0b-b540-4123ea0f8b18) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbda5-2ddf35467914dd21669158a2;18b3d9d3-0dc2-4a64-bc34-9c0a9cf14b05) Repository Not Found for url: https://huggingface.co/i/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -6535,7 +6535,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667777b5-147644de7ca0fe4f4313d7cc;643a6bb0-7764-4be1-89ae-373b9080d497) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbd7d-3d4c1a015ad0837b5ee0de02;15bc3a0c-3faf-422d-8720-b4321683391d) Repository Not Found for url: https://huggingface.co/M/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -6597,7 +6597,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp8k5g4k21/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp2rons8m3/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 8bit-bnb-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-125m,,cuda,0,42,,,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,MB,1218.818048,908.591104,0.0,253.755392,210.714624,s,10,0.3346063995361328,0.03346063995361328,0.001386687306170187,0.032872127532958983,0.03444729194641113,0.03575335865020752,0.03679821201324463,"[0.03705942535400391, 0.032388286590576174, 0.033940639495849606, 0.03399001693725586, 0.03415705490112304, 0.032402080535888673, 0.0324686393737793, 0.03245711898803711, 0.032467521667480466, 0.033275615692138674]",tokens/s,7650.780151093781,kWh,3.894642774255304e-07,2.134087789511939e-07,6.046422965788539e-07,1.207515352955578e-06,tokens/kWh,212005585.99391797,MB,1218.818048,908.591104,0.0,253.755392,219.642368,s,10,20.79267517089844,2.079267517089844,0.04302757885741849,2.079251953125,2.125889990234375,2.1366427490234376,2.1452449560546873,"[2.1473955078125, 2.12350048828125, 2.106216796875, 2.108764404296875, 2.116618408203125, 2.052287109375, 2.036617431640625, 2.044056640625, 2.030193115234375, 2.0270252685546875]",tokens/s,30.29913153655918,kWh,2.4556693654855963e-05,1.3457733793142036e-05,3.6514748510021464e-05,7.452917595801947e-05,tokens/kWh,845306.541903622,,s,629,21.054208013534534,0.03347250876555572,0.0040572122385470755,0.03302195358276367,0.033839513397216796,0.03419217910766602,0.06512308166503906,"[0.03506790542602539, 0.03536383819580078, 0.03539251327514648, 0.03406131362915039, 0.03338956832885742, 0.033960960388183595, 0.03545395278930664, 0.03595468902587891, 0.035366912841796876, 0.034855934143066404, 0.035699710845947266, 0.03513753509521484, 0.033291263580322264, 0.03509862518310547, 0.038406143188476564, 0.034492416381835936, 0.03402035140991211, 0.033710079193115236, 0.0336445426940918, 0.033949695587158206, 0.033759231567382815, 0.033693695068359376, 0.033726463317871096, 0.03363430404663086, 0.03370905685424805, 0.033718273162841796, 0.03363840103149414, 0.03382271957397461, 0.03367628860473633, 0.03388927841186523, 0.03364863967895508, 0.033555454254150394, 0.033323009490966796, 0.0337448959350586, 0.033667072296142575, 0.03362713623046875, 0.035326976776123044, 0.03380633544921875, 0.033600574493408204, 0.03373056030273437, 0.03366393661499024, 0.03363430404663086, 0.03360358428955078, 0.03364352035522461, 0.03383500671386719, 0.0331253776550293, 0.03432243347167969, 0.03396915054321289, 0.03426406478881836, 0.03361075210571289, 0.033581119537353515, 0.03367724609375, 0.03362508773803711, 0.03364761734008789, 0.03364659118652344, 0.03356467056274414, 0.033538047790527346, 0.03366809463500976, 0.033442817687988284, 0.033157119750976564, 0.03295846557617187, 0.03325132751464844, 0.06499839782714843, 0.03318374252319336, 0.0341473274230957, 0.03378073501586914, 0.03378278350830078, 0.03416883087158203, 0.033841152191162106, 0.03394867324829102, 0.03400908660888672, 0.03304140853881836, 0.03325235366821289, 0.033445888519287106, 0.033825790405273434, 0.033960960388183595, 0.03381452941894531, 0.03439718246459961, 0.03385958480834961, 0.033844223022460936, 0.03383910369873047, 0.03377151870727539, 0.03375519943237305, 0.03394655990600586, 0.03372748947143555, 0.033788928985595705, 0.033960960388183595, 0.0338155517578125, 0.03378278350830078, 0.03379302215576172, 0.03368243026733399, 0.03369472122192383, 0.033753089904785157, 0.03369779205322266, 0.033800193786621094, 0.033838081359863284, 0.03368038558959961, 0.03366809463500976, 0.033649665832519535, 0.033723392486572266, 0.03341721725463867, 0.033089534759521484, 0.03323699188232422, 0.033683456420898435, 0.03367731094360352, 0.03377459335327149, 0.03365273666381836, 0.034062335968017575, 0.03369676971435547, 0.033754112243652344, 0.03363123321533203, 0.03370393753051758, 0.03369779205322266, 0.03355852890014648, 0.033649665832519535, 0.033614849090576174, 0.03380121612548828, 0.03374899291992187, 0.03370086288452148, 0.03362406539916992, 0.03363942337036133, 0.03373567962646484, 0.03367833709716797, 0.033685504913330076, 0.03373158264160156, 0.06784614562988281, 0.03375001525878906, 0.033716224670410154, 0.03369881439208984, 0.03363840103149414, 0.033797119140625, 0.033685504913330076, 0.03370393753051758, 0.03382886505126953, 0.03364761734008789, 0.033691646575927735, 0.0342108154296875, 0.033650688171386715, 0.03352883148193359, 0.03361075210571289, 0.033031169891357424, 0.03316841506958008, 0.0337817268371582, 0.03371417617797851, 0.0337336311340332, 0.03317353439331055, 0.03308438491821289, 0.03307929611206055, 0.03295846557617187, 0.03303219223022461, 0.03358617782592774, 0.03301068878173828, 0.032979969024658204, 0.03301068878173828, 0.03297382354736328, 0.03310086441040039, 0.03358406448364258, 0.03310899353027344, 0.03299123382568359, 0.03307827377319336, 0.032965633392333986, 0.033068031311035154, 0.03363942337036133, 0.03364044952392578, 0.03363020706176758, 0.03377664184570312, 0.03362815856933594, 0.03361587142944336, 0.03311513519287109, 0.03304550552368164, 0.03344384002685547, 0.03360153579711914, 0.03362406539916992, 0.03359027099609375, 0.03359334564208984, 0.033691646575927735, 0.033686527252197264, 0.03362406539916992, 0.033683456420898435, 0.03362508773803711, 0.03307724761962891, 0.032894977569580076, 0.03303833770751953, 0.03294617462158203, 0.03296768188476563, 0.03366400146484375, 0.03360768127441406, 0.033532928466796875, 0.06804991912841797, 0.03370393753051758, 0.03369267272949219, 0.03355852890014648, 0.03361075210571289, 0.032906238555908206, 0.03340083312988281, 0.03333222579956055, 0.0335994873046875, 0.03355955123901367, 0.03355955123901367, 0.033546241760253906, 0.03303014373779297, 0.03306703948974609, 0.032993247985839844, 0.033097728729248044, 0.03316940689086914, 0.03303628921508789, 0.032917503356933595, 0.033020927429199216, 0.033121280670166016, 0.03305267333984375, 0.0336517105102539, 0.03363942337036133, 0.03374387359619141, 0.03376230239868164, 0.033704959869384765, 0.03378790283203125, 0.033710079193115236, 0.03364863967895508, 0.03372544097900391, 0.033119232177734374, 0.03297792053222656, 0.03309673690795899, 0.033073184967041015, 0.03301267242431641, 0.03359641647338867, 0.03416883087158203, 0.033704959869384765, 0.03363225555419922, 0.033770496368408204, 0.033683456420898435, 0.0337151985168457, 0.033713153839111325, 0.0336978874206543, 0.03346422576904297, 0.03374796676635742, 0.03367424011230469, 0.0335994873046875, 0.03465318298339844, 0.03302707290649414, 0.03370598220825195, 0.03378585433959961, 0.033844223022460936, 0.03379916763305664, 0.03383193588256836, 0.033807361602783206, 0.0331253776550293, 0.03276287841796875, 0.03292160034179688, 0.032947200775146485, 0.03291648101806641, 0.03377052688598633, 0.06837859344482422, 0.033849342346191406, 0.03383705520629883, 0.03409920120239258, 0.03617484664916992, 0.034237438201904294, 0.033957889556884766, 0.03362918472290039, 0.034103294372558594, 0.0338678092956543, 0.03370902252197266, 0.03375001525878906, 0.03376540756225586, 0.03382780838012695, 0.03376025772094727, 0.03385139083862305, 0.03415859222412109, 0.033865726470947266, 0.0337151985168457, 0.03377151870727539, 0.03380428695678711, 0.033737728118896484, 0.03367935943603516, 0.03375820922851563, 0.033808383941650394, 0.03374694442749023, 0.033811454772949216, 0.03399782562255859, 0.032228351593017575, 0.03217510223388672, 0.03217203140258789, 0.03261542510986328, 0.0338155517578125, 0.03369574356079102, 0.03379097747802735, 0.033710079193115236, 0.0337174072265625, 0.03359830474853515, 0.03377766418457031, 0.033759231567382815, 0.03384524917602539, 0.03381248092651367, 0.03373056030273437, 0.03383091354370117, 0.03386880111694336, 0.03381760025024414, 0.033770496368408204, 0.033770496368408204, 0.0337786865234375, 0.03320032119750976, 0.03335353469848633, 0.03627008056640625, 0.033067008972167966, 0.033360897064208986, 0.03245363235473633, 0.03248025512695313, 0.0323768310546875, 0.03333631896972656, 0.03375513458251953, 0.03243622589111328, 0.03228979110717774, 0.03227340698242188, 0.0322529296875, 0.06581247711181641, 0.032353279113769534, 0.032876544952392575, 0.03321343994140625, 0.03256934356689453, 0.03236249542236328, 0.03248332977294922, 0.03272499084472656, 0.0323583984375, 0.03230822372436523, 0.03232153701782227, 0.03267583847045898, 0.03309260940551758, 0.03223654556274414, 0.032405502319335935, 0.03415039825439453, 0.03420774459838867, 0.03242086410522461, 0.03232767868041992, 0.032674816131591795, 0.03317964935302734, 0.033137664794921876, 0.03230822372436523, 0.03260006332397461, 0.033100833892822264, 0.03465212631225586, 0.033567745208740236, 0.032487422943115234, 0.03227033615112305, 0.033037311553955076, 0.03228160095214844, 0.03231948852539063, 0.032287742614746096, 0.032456703186035156, 0.03239321517944336, 0.03227852630615234, 0.03222323226928711, 0.032198654174804685, 0.03280998229980469, 0.03214556884765625, 0.03213910293579102, 0.03215359878540039, 0.03229695892333984, 0.03275161743164062, 0.03222118377685547, 0.03222630310058594, 0.03265740966796875, 0.03288371276855469, 0.03292160034179688, 0.03300454330444336, 0.03221196746826172, 0.032215038299560544, 0.03282227325439453, 0.03225702285766602, 0.03218227386474609, 0.03181056022644043, 0.031680511474609374, 0.032132095336914065, 0.03218022537231445, 0.03221913528442383, 0.03215359878540039, 0.0321638412475586, 0.0321341438293457, 0.0651171875, 0.03283456039428711, 0.032178176879882815, 0.032178176879882815, 0.03220377731323242, 0.03217510223388672, 0.03203379058837891, 0.03250995254516602, 0.03220889663696289, 0.03211673736572265, 0.032210945129394535, 0.03215568161010742, 0.0324351692199707, 0.03214438247680664, 0.03280588912963867, 0.03239731216430664, 0.03262054443359375, 0.03189043235778809, 0.03181056022644043, 0.03175628852844238, 0.03160166358947754, 0.03174297523498535, 0.032231422424316404, 0.032301055908203126, 0.032216064453125, 0.03225395202636719, 0.032709632873535156, 0.03238809585571289, 0.032331775665283204, 0.03216793441772461, 0.03280691146850586, 0.032674816131591795, 0.032330753326416016, 0.03217715072631836, 0.03217919921875, 0.03218534469604492, 0.03220684814453125, 0.032290817260742184, 0.03227852630615234, 0.032440319061279296, 0.032244735717773435, 0.03235635375976562, 0.032200702667236326, 0.0325478401184082, 0.03278438568115234, 0.03226009750366211, 0.032894977569580076, 0.03207372665405273, 0.03267891311645508, 0.03228057479858398, 0.03221299362182617, 0.032263168334960936, 0.032284671783447266, 0.03237580871582031, 0.03304755020141602, 0.032368640899658206, 0.032658432006835936, 0.03284384155273438, 0.032491455078125, 0.03231846237182617, 0.03193343925476074, 0.032467967987060545, 0.03225190353393555, 0.06513868713378906, 0.03216998291015625, 0.03215462493896484, 0.032166976928710934, 0.03221907043457031, 0.032320510864257815, 0.03224172973632813, 0.03257952117919922, 0.03226828765869141, 0.03223654556274414, 0.03213926315307617, 0.032196670532226565, 0.032206783294677736, 0.032247806549072264, 0.032113662719726564, 0.032718849182128903, 0.032312320709228515, 0.03223865509033203, 0.032221118927001954, 0.03225190353393555, 0.03228364944458008, 0.032292865753173826, 0.03234406280517578, 0.032233470916748046, 0.03294105529785156, 0.03326566314697266, 0.03223551940917969, 0.03181977653503418, 0.03170918464660644, 0.031768575668334964, 0.03173075294494629, 0.0317573127746582, 0.0318668155670166, 0.03224371337890625, 0.032266239166259765, 0.03222630310058594, 0.032309249877929686, 0.03230515289306641, 0.032696319580078126, 0.032307201385498044, 0.032249855041503905, 0.032140350341796874, 0.03224057769775391, 0.03225395202636719, 0.03226115036010742, 0.03228771209716797, 0.032851966857910156, 0.03227756881713867, 0.03244230270385742, 0.033306625366210936, 0.033107967376708985, 0.03316940689086914, 0.03331379318237305, 0.03323801422119141, 0.03233894348144531, 0.03316121673583984, 0.03319705581665039, 0.0333496322631836, 0.03302195358276367, 0.03323187255859375, 0.033337345123291014, 0.032889854431152346, 0.03215052795410156, 0.06512537384033203, 0.03225190353393555, 0.03228672027587891, 0.032553985595703126, 0.03173990440368652, 0.03209011077880859, 0.032361473083496094, 0.03228057479858398, 0.0323583984375, 0.032345088958740234, 0.0332677116394043, 0.03224576187133789, 0.03219251251220703, 0.031730688095092774, 0.03205836868286133, 0.03226521682739258, 0.03231027221679687, 0.032193695068359375, 0.03185443115234375, 0.03187302398681641, 0.03193343925476074, 0.032295936584472655, 0.03230310440063477, 0.03223251342773437, 0.03222419357299805, 0.03230003356933594, 0.03231027221679687, 0.03231027221679687, 0.03224063873291016, 0.032254974365234376, 0.032244735717773435, 0.03167334365844727, 0.0318474235534668, 0.03177574348449707, 0.03181363105773926, 0.031751167297363284, 0.03181465530395508, 0.03225088119506836, 0.032266239166259765, 0.03229183959960937, 0.03220889663696289, 0.032279552459716795, 0.0323133430480957, 0.03229183959960937, 0.032107521057128906, 0.03160985565185547, 0.031704063415527346, 0.03168767929077149, 0.032517120361328124, 0.0325591049194336, 0.032330753326416016, 0.032302078247070314, 0.03224576187133789, 0.03226419067382812, 0.03259392166137695, 0.03251609420776367, 0.03244646453857422, 0.032249855041503905, 0.032200702667236326, 0.03279564666748047, 0.03288576126098633, 0.03276595306396484, 0.03228364944458008, 0.06702489471435547, 0.03328307342529297, 0.03296051025390625, 0.03329951858520508, 0.0322303352355957, 0.032233470916748046, 0.032287742614746096, 0.03219558334350586, 0.03220377731323242, 0.032091136932373046, 0.031753215789794925, 0.03159040069580078, 0.03159449577331543, 0.03172454452514648, 0.03171225547790527, 0.03160780715942383, 0.031768575668334964, 0.03189760017395019, 0.032264225006103514, 0.03219657516479492, 0.032178176879882815, 0.032189441680908204, 0.03222016143798828, 0.03198464012145996, 0.031703039169311525, 0.03156787109375, 0.03197235107421875, 0.03173785591125488, 0.03216998291015625, 0.032118785858154295, 0.03217407989501953, 0.03215769577026367, 0.03223859024047852, 0.03223551940917969, 0.03220787048339844, 0.03213824081420898, 0.032210945129394535, 0.031893503189086916, 0.0316682243347168, 0.03148287963867188, 0.033770496368408204, 0.03371929550170898, 0.03225088119506836, 0.03219251251220703, 0.03217820739746094, 0.032163806915283207, 0.032277503967285154, 0.03223961639404297, 0.032307201385498044, 0.0322242546081543, 0.032200702667236326, 0.03181363105773926, 0.031602752685546874, 0.03169478416442871, 0.031775808334350585, 0.031826879501342775, 0.03226009750366211, 0.03222937774658203, 0.03223756790161133, 0.032263168334960936, 0.032247806549072264, 0.032302078247070314, 0.032385025024414066]",tokens/s,29.87526292110593,,,main,False,False, @@ -6627,7 +6627,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpx99s_9ee/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpd3ubq6as/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 8bit-bnb-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,x,x,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.218-208.862.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): @@ -6669,7 +6669,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66777921-5c7d0b2854b65f2a6457327b;9e23f0b7-ba8f-40dd-8914-d6a9f4e623d3) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbeed-043b183001d3fcce597f6ca0;591c13f1-c8bf-4c03-9a00-1fabe7811129) Repository Not Found for url: https://huggingface.co/x/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -6740,7 +6740,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66777b8c-70b891c96beb479b2d8b6703;92a41cc6-44f2-4dcb-88b8-bafa24c91caf) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667cc156-033d1314151297d119a92bc0;1808aaaa-a5b9-49c2-a267-7688d09f575f) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json. @@ -6830,7 +6830,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667778f6-00825c7e788453d1770782f7;e0bd8de0-c4d0-40ec-8917-aedd331f7755) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbec4-2f4029490b3434e716d16d44;ca4dcb3d-1144-4684-a967-d5dd03a0a53a) Repository Not Found for url: https://huggingface.co/8/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -6919,7 +6919,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667a11ed-52a1b57c5328906a351e4ea2;57edafe3-571d-4df7-9e06-b4d9a3f0bbfd) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667cb432-0af3c5d340f088a57e2469fa;100f2c2e-a25a-43c7-8406-8dfa04037475) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/databricks/dbrx-base/resolve/main/config.json. @@ -7026,7 +7026,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66777856-350f54ff5a204d68226b91ce;0ad0f24a-52f8-4477-8125-40e0e94bfc48) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbe24-60767ac15030cf9f370e65f0;bb7356b7-37b8-4d42-8eec-108f74aaf117) Repository Not Found for url: https://huggingface.co/r/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -7125,7 +7125,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpkmxovvz8/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpeqieo8h3/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 8bit-bnb-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,TencentARC/Mistral_Pro_8B_v0.1,,cuda,0,42,,,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,MB,5660.131328,10962.337792,0.0,10307.50208,10029.210624,s,10,1.4342690734863281,0.1434269073486328,0.0032239528002932865,0.14276050567626952,0.14646627197265624,0.14882899169921876,0.15071916748046876,"[0.15119171142578125, 0.14055494689941406, 0.14345762634277343, 0.14594122314453126, 0.14206338500976562, 0.1407220458984375, 0.14052482604980468, 0.14054188537597656, 0.1442948760986328, 0.14497654724121095]",tokens/s,1784.8812662308324,kWh,1.7037357295316364e-06,9.333370224703686e-07,4.414690039685912e-06,7.051762791687916e-06,tokens/kWh,36302979.490710236,MB,5660.131328,10962.337792,0.0,10307.50208,10056.90112,s,10,91.03206542968749,9.10320654296875,0.08025953626004148,9.07454150390625,9.2241923828125,9.259001464843749,9.28684873046875,"[9.293810546875, 9.0487578125, 9.21645703125, 9.09006640625, 9.034521484375, 9.0707138671875, 9.078369140625, 9.0585595703125, 9.098578125, 9.0422314453125]",tokens/s,6.920638316029503,kWh,0.00010834198840080747,5.937985369364469e-05,0.00020726067374491383,0.000374982515839366,tokens/kWh,168007.83326918574,,s,629,92.16478802490228,0.1465258951111325,0.017154203671883243,0.14422015380859374,0.14755471496582032,0.14824775695800782,0.2843094567871094,"[0.1534976043701172, 0.15138217163085937, 0.14815318298339844, 0.14723379516601562, 0.14589552307128906, 0.14708316040039063, 0.14719488525390625, 0.1486684112548828, 0.14777650451660157, 0.14928997802734376, 0.14852505493164062, 0.14738124084472656, 0.14825267028808595, 0.14617190551757814, 0.14766905212402343, 0.14603974914550782, 0.14723788452148437, 0.14720204162597655, 0.14758604431152345, 0.14771302795410157, 0.14740480041503906, 0.14766592407226561, 0.1476065216064453, 0.14713037109375, 0.14739353942871095, 0.14730035400390626, 0.1475030975341797, 0.1471815643310547, 0.1472184295654297, 0.14708837890625, 0.147631103515625, 0.14701676940917968, 0.1473258819580078, 0.14693174743652343, 0.14660508728027344, 0.147368896484375, 0.14727372741699218, 0.14718360900878907, 0.14627839660644532, 0.14710476684570312, 0.14769258117675782, 0.14700233459472656, 0.1472604217529297, 0.14715289306640625, 0.14720204162597655, 0.14623641967773438, 0.14750617980957031, 0.14715493774414062, 0.1470310363769531, 0.147125244140625, 0.1474846649169922, 0.14677708435058595, 0.1466603546142578, 0.14490419006347657, 0.14707200622558594, 0.14715187072753907, 0.14748672485351563, 0.14915379333496093, 0.14982861328125, 0.14755635070800782, 0.14706585693359375, 0.14647807312011718, 0.28261068725585936, 0.14193667602539062, 0.14197244262695313, 0.1418618927001953, 0.14242816162109376, 0.14205644226074218, 0.1417164764404297, 0.14202674865722656, 0.1422878723144531, 0.14224485778808593, 0.14182809448242187, 0.1425244140625, 0.144990234375, 0.14621385192871095, 0.14743449401855468, 0.14801100158691408, 0.14670028686523437, 0.14654258728027345, 0.14298725891113281, 0.14222029113769533, 0.14218547058105468, 0.14629273986816407, 0.14658047485351564, 0.14634701538085937, 0.14226022338867186, 0.14339993286132813, 0.14181272888183594, 0.14232882690429688, 0.14208409118652343, 0.14561077880859374, 0.14698086547851563, 0.1468139190673828, 0.1429698486328125, 0.14215065002441407, 0.14218547058105468, 0.141549560546875, 0.14222950744628907, 0.14219468688964843, 0.14192127990722656, 0.1424701385498047, 0.14214451599121095, 0.14214247131347657, 0.142055419921875, 0.14200729370117188, 0.1421486053466797, 0.14182298278808594, 0.14230117797851563, 0.14190899658203124, 0.14127513122558594, 0.14705255126953126, 0.14671974182128905, 0.1458841552734375, 0.1467658233642578, 0.14730137634277343, 0.14554623413085938, 0.14709759521484375, 0.14683135986328125, 0.14590156555175782, 0.1421158447265625, 0.14215884399414064, 0.14210560607910155, 0.1419694061279297, 0.14311935424804687, 0.2923089904785156, 0.14832540893554688, 0.14229808044433595, 0.14185369873046874, 0.14219161987304688, 0.14211276245117188, 0.1411440887451172, 0.14231138610839844, 0.14515507507324218, 0.14639616394042967, 0.14236569213867187, 0.14185267639160157, 0.14324223327636718, 0.14543565368652345, 0.1462415313720703, 0.14700749206542968, 0.14749183654785156, 0.14913433837890624, 0.14708224487304689, 0.14785842895507811, 0.14836531066894532, 0.14659686279296874, 0.14642585754394533, 0.14595379638671874, 0.1468037109375, 0.14741299438476563, 0.1471866912841797, 0.14695018005371094, 0.14615753173828125, 0.14739865112304687, 0.14753074645996095, 0.14726451110839844, 0.14712832641601561, 0.14760858154296874, 0.14934323120117188, 0.14819839477539062, 0.14772531127929686, 0.14653030395507813, 0.14631321716308593, 0.1442170867919922, 0.1412689971923828, 0.14300569152832032, 0.1467852783203125, 0.144468994140625, 0.1477908477783203, 0.14747853088378907, 0.14710578918457032, 0.14732902526855468, 0.1479761962890625, 0.14744166564941405, 0.14737202453613282, 0.14715904235839844, 0.14731980895996094, 0.14735565185546876, 0.14734335327148437, 0.14759117126464844, 0.14715699768066406, 0.14793318176269532, 0.1475758056640625, 0.1477191619873047, 0.14755430603027345, 0.14633676147460936, 0.1478974151611328, 0.293482421875, 0.14644940185546876, 0.1467545623779297, 0.14730239868164063, 0.14824038696289063, 0.14709564208984374, 0.1482085418701172, 0.14197247314453126, 0.14193356323242187, 0.14270669555664062, 0.14446080017089843, 0.1439488067626953, 0.14200114440917969, 0.141955078125, 0.14212515258789063, 0.14191094970703125, 0.14166220092773438, 0.14212095642089845, 0.1453096923828125, 0.14467481994628906, 0.141955078125, 0.14581964111328125, 0.1451612091064453, 0.14201548767089844, 0.14746214294433593, 0.14255410766601562, 0.14240460205078126, 0.1456005096435547, 0.14647500610351563, 0.14677606201171875, 0.14548991394042968, 0.1466112060546875, 0.14662661743164063, 0.14529632568359374, 0.14660301208496093, 0.14532095336914064, 0.14720204162597655, 0.14515609741210939, 0.14182818603515626, 0.14359234619140626, 0.14211993408203125, 0.14597222900390625, 0.14219981384277344, 0.14482534790039062, 0.1466050567626953, 0.1439139862060547, 0.14232984924316405, 0.14218138122558593, 0.14511410522460938, 0.1470054473876953, 0.14371839904785155, 0.14676480102539063, 0.14239027404785157, 0.1433210906982422, 0.14554112243652345, 0.14293913269042968, 0.14217727661132812, 0.1420052490234375, 0.1420738525390625, 0.14203392028808592, 0.1424547882080078, 0.14198886108398437, 0.1421096954345703, 0.28633599853515623, 0.1421793212890625, 0.1419683837890625, 0.14203494262695313, 0.14164070129394532, 0.142129150390625, 0.1420369873046875, 0.1409792022705078, 0.14163763427734374, 0.14192434692382813, 0.14748876953125, 0.1480816650390625, 0.14249778747558595, 0.14207693481445313, 0.1418946533203125, 0.14191513061523436, 0.14302822875976562, 0.1462671356201172, 0.14379315185546876, 0.1420042266845703, 0.14189056396484376, 0.142166015625, 0.14207283020019532, 0.14194586181640625, 0.1419632568359375, 0.1417574462890625, 0.14206259155273437, 0.14469427490234374, 0.14457958984375, 0.14223770141601563, 0.14819123840332032, 0.14243942260742187, 0.14197760009765625, 0.14243327331542968, 0.14391500854492187, 0.14455091857910157, 0.1463655090332031, 0.1469173126220703, 0.14727372741699218, 0.14673715209960939, 0.14708428955078126, 0.14422015380859374, 0.14662864685058594, 0.14663778686523438, 0.14674227905273438, 0.14578790283203125, 0.14431129455566405, 0.14236671447753907, 0.14213632202148438, 0.14202879333496093, 0.14202983093261717, 0.14231961059570314, 0.144500732421875, 0.1421271057128906, 0.1421967315673828, 0.14232269287109375, 0.1416243133544922, 0.14191513061523436, 0.14222438049316405, 0.14256434631347656, 0.14218240356445314, 0.14231039428710937, 0.14231858825683594, 0.28481637573242186, 0.14673817443847656, 0.1422530517578125, 0.1425244140625, 0.1424435272216797, 0.14243942260742187, 0.141549560546875, 0.14195712280273437, 0.14179942321777345, 0.14219264221191405, 0.14188543701171874, 0.14629682922363282, 0.14243536376953125, 0.14196937561035156, 0.1419141082763672, 0.14319308471679687, 0.14707200622558594, 0.14835813903808595, 0.146914306640625, 0.14225202941894532, 0.1429432373046875, 0.1466439666748047, 0.1463582763671875, 0.14676173400878906, 0.142097412109375, 0.14210456848144531, 0.142328857421875, 0.14207894897460938, 0.14204415893554687, 0.14187519836425783, 0.14248448181152343, 0.1420748748779297, 0.14453657531738281, 0.1497241668701172, 0.14483045959472657, 0.14850559997558593, 0.1470443572998047, 0.1467535400390625, 0.14210560607910155, 0.14510182189941406, 0.1472430114746094, 0.14219366455078125, 0.14655795288085938, 0.14226329040527344, 0.14511616516113282, 0.1430425567626953, 0.14588108825683593, 0.14455807495117187, 0.14212095642089845, 0.14284185791015624, 0.1465180206298828, 0.14465434265136717, 0.14633779907226563, 0.14715391540527345, 0.14545919799804688, 0.14233804321289062, 0.14329347229003905, 0.14232981872558595, 0.14286131286621093, 0.14224076843261718, 0.14246502685546875, 0.14232473754882813, 0.14228172302246095, 0.28571749877929686, 0.14685081481933593, 0.1466060791015625, 0.14657945251464843, 0.14648013305664062, 0.1426606140136719, 0.142166015625, 0.1418383331298828, 0.14211891174316407, 0.14176870727539062, 0.1424547882080078, 0.14432870483398438, 0.1421844482421875, 0.14212300109863282, 0.14295245361328124, 0.14228480529785156, 0.14206668090820312, 0.14567730712890625, 0.14654360961914062, 0.1468538818359375, 0.14368153381347656, 0.14674943542480468, 0.14677920532226563, 0.14910252380371095, 0.1466337890625, 0.14349510192871093, 0.1429053497314453, 0.1412884521484375, 0.14192025756835938, 0.14216397094726563, 0.1419171905517578, 0.14165298461914064, 0.14216294860839843, 0.14221107482910156, 0.1421271057128906, 0.14187930297851561, 0.14223359680175782, 0.14176153564453126, 0.14771302795410157, 0.1455615997314453, 0.1417820129394531, 0.1478830108642578, 0.146044921875, 0.14544895935058594, 0.14200320434570313, 0.1432238006591797, 0.14634495544433593, 0.14659584045410157, 0.14687437438964843, 0.1460561981201172, 0.14661734008789062, 0.14204722595214844, 0.1441239013671875, 0.14674227905273438, 0.14656819152832032, 0.1423207092285156, 0.14195193481445312, 0.1423093719482422, 0.14223155212402344, 0.14204928588867188, 0.14666957092285157, 0.14673715209960939, 0.1421107177734375, 0.2830059509277344, 0.14187110900878908, 0.14194586181640625, 0.1410232391357422, 0.1422387237548828, 0.142055419921875, 0.14226739501953126, 0.14214656066894532, 0.14211482238769532, 0.14605413818359375, 0.1423134765625, 0.14269541931152344, 0.14528614807128906, 0.14245785522460938, 0.14232167053222655, 0.1421558074951172, 0.14252642822265624, 0.14217010498046875, 0.1435924530029297, 0.14213938903808593, 0.14155264282226562, 0.14155264282226562, 0.14195097351074218, 0.14192947387695312, 0.14090956115722655, 0.14435122680664061, 0.1428019561767578, 0.14582473754882813, 0.14670745849609376, 0.14971084594726564, 0.14662042236328124, 0.14527590942382812, 0.14401536560058595, 0.14318284606933593, 0.14638182067871094, 0.14626200866699218, 0.14467277526855468, 0.1420185546875, 0.14180870056152345, 0.14723167419433594, 0.14519500732421875, 0.14663987731933595, 0.1462353973388672, 0.14677606201171875, 0.14902784729003907, 0.1424701385498047, 0.1466265869140625, 0.1469142761230469, 0.144500732421875, 0.14280601501464843, 0.1422776336669922, 0.14262168884277343, 0.14193458557128907, 0.1422878723144531, 0.1456906280517578, 0.14675865173339844, 0.14654360961914062, 0.14573875427246094, 0.14238514709472655, 0.14148812866210939, 0.14229196166992186, 0.1426042938232422, 0.14217625427246094, 0.2911856689453125, 0.14181683349609375, 0.14218751525878906, 0.14203904724121094, 0.14148403930664064, 0.14213529968261718, 0.14200729370117188, 0.14192250061035155, 0.14223033142089844, 0.14233088684082032, 0.1445386199951172, 0.14171449279785156, 0.1443634490966797, 0.1467105255126953, 0.14244557189941406, 0.143383544921875, 0.1467310028076172, 0.1423687744140625, 0.1447403564453125, 0.14611251831054686, 0.1460643768310547, 0.14269541931152344, 0.14153523254394532, 0.1422387237548828, 0.14567526245117188, 0.14662451171875, 0.14528204345703125, 0.14394776916503907, 0.14590975952148438, 0.1455422058105469, 0.14265849304199218, 0.1465630645751953, 0.15092428588867188, 0.14452120971679688, 0.147051513671875, 0.14275379943847658, 0.14206771850585936, 0.14204620361328124, 0.14201449584960937, 0.14219261169433595, 0.14223974609375, 0.14218853759765626, 0.142202880859375, 0.1419990997314453, 0.14330368041992186, 0.14642994689941408, 0.1418434600830078, 0.1423319091796875, 0.1424486389160156, 0.1481861114501953, 0.145069091796875, 0.14216802978515625, 0.1488527374267578, 0.14834072875976562, 0.14806016540527345, 0.1470699462890625, 0.14825881958007814, 0.1478973388671875, 0.1470586853027344, 0.1465149383544922, 0.1468651580810547, 0.14660096740722656, 0.14620057678222656, 0.28747674560546876, 0.14185987854003906, 0.14227145385742188, 0.14205952453613283, 0.14222848510742186, 0.14239846801757813, 0.14227865600585937, 0.14170317077636718, 0.14202879333496093, 0.14202163696289063, 0.14184652709960938, 0.14195814514160157, 0.14193463134765624, 0.1420113983154297, 0.14294326782226563, 0.14217619323730468, 0.14186393737792968, 0.14214041137695313, 0.1420779571533203, 0.14214247131347657, 0.14200831604003905, 0.14226943969726563, 0.14271385192871094, 0.14213020324707032, 0.14204412841796876, 0.14196633911132814, 0.141591552734375, 0.1420789794921875, 0.14242713928222656, 0.14199501037597656, 0.1419878387451172, 0.1420779571533203, 0.14217625427246094, 0.14183628845214843, 0.14210867309570313, 0.1430630340576172, 0.1449400329589844, 0.14324327087402344, 0.1468098602294922, 0.14377165222167967, 0.1458401336669922, 0.14713446044921874, 0.14582681274414064, 0.14665216064453124, 0.1467658233642578, 0.14667059326171875, 0.14654669189453126, 0.14684979248046875, 0.14678425598144532, 0.14661325073242187, 0.14607052612304688, 0.14693376159667967, 0.1446860809326172, 0.14237184143066406, 0.1426483154296875, 0.14714572143554688, 0.14626815795898437, 0.14899200439453125, 0.14655795288085938, 0.14257254028320313, 0.14196223449707032, 0.14228688049316407, 0.14241789245605468]",tokens/s,6.824732237544431,,,main,False,False, @@ -7189,7 +7189,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpr40amp3b/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpyco_21kz/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,True 4bit-bnb-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-7b,tiiuae/falcon-7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.218-208.862.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): @@ -7218,7 +7218,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp_def28qr/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp1mruqgln/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,True 4bit-bnb-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/polyglot-ko-12.8b,,cuda,0,42,,,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,MB,1321.848832,9760.669696,0.0,9114.222592,8556.99456,s,10,5.6361691284179685,0.5636169128417968,0.0002636735538146785,0.5635266418457031,0.5638524169921876,0.5640790100097657,0.5642602844238281,"[0.5643056030273438, 0.563530029296875, 0.563430908203125, 0.5634299926757812, 0.5633757934570313, 0.5636502685546875, 0.5634279174804687, 0.5638020629882813, 0.5635232543945312, 0.5636932983398437]",tokens/s,454.20922290857044,kWh,6.662574842756177e-06,3.6507952148895126e-06,4.353401630866677e-05,5.3847386366312467e-05,tokens/kWh,4754176.8927926365,MB,1321.848832,9760.669696,0.0,9114.222592,8556.99712,s,10,37.283349609375,3.7283349609375,0.025639681867448014,3.739591796875,3.7490041259765623,3.7491127807617186,3.7491997045898438,"[3.74897998046875, 3.749221435546875, 3.744506103515625, 3.72951806640625, 3.7369013671875, 3.745256591796875, 3.7422822265625, 3.691360107421875, 3.668978271484375, 3.726345458984375]",tokens/s,16.897623378817464,kWh,4.2741739681841415e-05,2.3425537683815484e-05,0.00015242902935053365,0.00021859630671619052,tokens/kWh,288202.49045558943,,s,629,42.29440714645385,0.06724071088466432,0.06693317443297017,0.059417598724365236,0.05987716979980469,0.060254003143310544,0.6220912817382813,"[0.05945753479003906, 0.057845760345458984, 0.059858943939208986, 0.05932444763183594, 0.05945135879516601, 0.05909299087524414, 0.05979750442504883, 0.059681793212890626, 0.059268096923828124, 0.05947084808349609, 0.05948416137695312, 0.059458560943603515, 0.05942988967895508, 0.059456512451171874, 0.05929983901977539, 0.05915238571166992, 0.05947395324707031, 0.0594114875793457, 0.05908575820922852, 0.059481086730957033, 0.05948723220825195, 0.05952102279663086, 0.05949747085571289, 0.05976063919067383, 0.06056447982788086, 0.06034636688232422, 0.059627521514892576, 0.05948211288452149, 0.05956095886230469, 0.0595230712890625, 0.05960908889770508, 0.05933363342285156, 0.05907251358032226, 0.05955379104614258, 0.05959372711181641, 0.05938278579711914, 0.05954969787597656, 0.05912371063232422, 0.05917900848388672, 0.05992963027954101, 0.06006371307373047, 0.05890150451660156, 0.05959372711181641, 0.059415550231933595, 0.059619327545166016, 0.05942169570922851, 0.05960908889770508, 0.05949747085571289, 0.05950566482543945, 0.05810995101928711, 0.05940019226074219, 0.059557918548583985, 0.05974319839477539, 0.05965107345581055, 0.05963264083862305, 0.059486209869384764, 0.060214271545410154, 0.059660289764404295, 0.05956403350830078, 0.05943807983398437, 0.059551742553710936, 0.05952716827392578, 0.622244873046875, 0.057376766204833986, 0.05801062393188477, 0.059979774475097655, 0.06020710372924805, 0.05947187042236328, 0.05951795196533203, 0.05940121459960938, 0.059622398376464845, 0.059319297790527345, 0.05938790512084961, 0.05936742401123047, 0.05957017517089844, 0.05947903823852539, 0.059474945068359375, 0.05883391952514649, 0.05939507293701172, 0.059243518829345705, 0.060115966796875, 0.059658241271972653, 0.059834369659423826, 0.05920358276367187, 0.059540481567382814, 0.05932953643798828, 0.05965727996826172, 0.05942060852050781, 0.059510784149169924, 0.05921791839599609, 0.05966745758056641, 0.06004633712768555, 0.059769889831542966, 0.05953123092651367, 0.05963776016235352, 0.05931622314453125, 0.059772926330566405, 0.06038425445556641, 0.05985587310791016, 0.05941964721679688, 0.05961318588256836, 0.0594442253112793, 0.059595775604248044, 0.059504638671875, 0.05969414520263672, 0.0592608642578125, 0.05973299026489258, 0.05941964721679688, 0.059935745239257814, 0.05940224075317383, 0.05969100952148437, 0.059261951446533206, 0.05987635040283203, 0.05943500900268555, 0.06044979095458984, 0.059512832641601565, 0.05959987258911133, 0.059388927459716793, 0.0595230712890625, 0.0594442253112793, 0.05965414428710938, 0.05927833557128906, 0.0599736328125, 0.05959987258911133, 0.05961318588256836, 0.6222254028320312, 0.05769318389892578, 0.058367008209228514, 0.05979849624633789, 0.059407360076904295, 0.05951692962646484, 0.057761791229248044, 0.059431934356689455, 0.05828505706787109, 0.058851329803466794, 0.05913087844848633, 0.059101184844970706, 0.05941350555419922, 0.06017638397216797, 0.06029721450805664, 0.061257728576660155, 0.05956915283203125, 0.05950259017944336, 0.05960396957397461, 0.06097100830078125, 0.059834369659423826, 0.058797054290771485, 0.059374591827392575, 0.06008524703979492, 0.05975043106079102, 0.059678688049316406, 0.05931008148193359, 0.059501598358154294, 0.059232223510742185, 0.059617279052734375, 0.05955583953857422, 0.05956403350830078, 0.059240447998046876, 0.05943296051025391, 0.059202560424804686, 0.05903257751464844, 0.05897420883178711, 0.05957120132446289, 0.059259902954101565, 0.058821632385253904, 0.05923328018188476, 0.05961830520629883, 0.05937561416625976, 0.05960396957397461, 0.059319297790527345, 0.05965926361083984, 0.05953023910522461, 0.06018048095703125, 0.060985343933105465, 0.060464126586914066, 0.058726398468017575, 0.059504638671875, 0.05922918319702149, 0.059646976470947265, 0.05935001754760742, 0.05964086532592774, 0.05936943817138672, 0.0589854736328125, 0.05943603134155274, 0.059597824096679686, 0.05940838241577148, 0.05960908889770508, 0.05936844635009766, 0.6219386596679688, 0.057728000640869144, 0.05745663833618164, 0.057534465789794924, 0.05752115249633789, 0.05768703842163086, 0.057353214263916014, 0.05928345489501953, 0.059881473541259764, 0.05962035369873047, 0.05921279907226563, 0.05917695999145508, 0.059200511932373044, 0.059302913665771485, 0.05935923385620117, 0.05943603134155274, 0.0592988166809082, 0.05914316940307617, 0.05944729614257813, 0.059791358947753906, 0.059291744232177736, 0.05929257583618164, 0.059224063873291016, 0.059433982849121096, 0.05916672134399414, 0.059954177856445315, 0.060200958251953124, 0.06023987197875977, 0.059481086730957033, 0.05916159820556641, 0.05934694290161133, 0.05926911926269531, 0.059322368621826174, 0.059284481048583984, 0.05936640167236328, 0.05954252624511719, 0.05922099304199219, 0.059450366973876956, 0.059302913665771485, 0.059494400024414064, 0.05925888061523438, 0.05938687896728516, 0.05971558380126953, 0.05938585662841797, 0.05918310546875, 0.05924863815307617, 0.05934796905517578, 0.059474945068359375, 0.059407360076904295, 0.05938790512084961, 0.05923737716674805, 0.05932748794555664, 0.059284481048583984, 0.059388927459716793, 0.058929153442382816, 0.05934592056274414, 0.05936947250366211, 0.05904076766967774, 0.05913907241821289, 0.059834369659423826, 0.05922304153442383, 0.059483135223388675, 0.05919334411621094, 0.6216222534179687, 0.05771980667114258, 0.05739929580688476, 0.05780387115478516, 0.05821430587768555, 0.05933260726928711, 0.05934080123901367, 0.05915340805053711, 0.05948211288452149, 0.05930188751220703, 0.05944019317626953, 0.058399681091308595, 0.05925888061523438, 0.05914316940307617, 0.05931932830810547, 0.05937763214111328, 0.05931827163696289, 0.05941145706176758, 0.059440128326416014, 0.059355136871337894, 0.059734016418457034, 0.05896089553833008, 0.059300865173339844, 0.05940633773803711, 0.059292671203613284, 0.05947391891479492, 0.059004993438720704, 0.059514816284179685, 0.05890150451660156, 0.05942988967895508, 0.059061248779296874, 0.05944934463500977, 0.05938585662841797, 0.059491329193115235, 0.05931008148193359, 0.05937670516967773, 0.05942367935180664, 0.05972172927856445, 0.0593070068359375, 0.060731391906738284, 0.05981798553466797, 0.05953126525878906, 0.060475391387939455, 0.060034046173095705, 0.060170238494873046, 0.060184574127197264, 0.059582462310791014, 0.05998284912109375, 0.05955583953857422, 0.05952716827392578, 0.0593807373046875, 0.05965926361083984, 0.06001152038574219, 0.06018764877319336, 0.0603351058959961, 0.05938790512084961, 0.0593704948425293, 0.0595333137512207, 0.0594442253112793, 0.059763713836669924, 0.058894336700439455, 0.05771263885498047, 0.05765427017211914, 0.6244096069335937, 0.058859519958496094, 0.05931520080566406, 0.05961011123657227, 0.05942272186279297, 0.05950566482543945, 0.059504638671875, 0.05947391891479492, 0.05931315231323242, 0.05940019226074219, 0.05950054550170898, 0.059633663177490234, 0.05913190460205078, 0.05934080123901367, 0.05938380813598633, 0.06031155014038086, 0.059829246520996096, 0.059445247650146485, 0.05959990310668945, 0.05962646484375, 0.059412479400634766, 0.05943091201782227, 0.05933670425415039, 0.05953945541381836, 0.059205631256103515, 0.05936537551879883, 0.0593930549621582, 0.059442142486572265, 0.05954252624511719, 0.05954764938354492, 0.05942272186279297, 0.059486209869384764, 0.05980057525634765, 0.05878169631958008, 0.05949030303955078, 0.059417598724365236, 0.05961011123657227, 0.05941657638549805, 0.05928243255615234, 0.05957017517089844, 0.05931520080566406, 0.05900595092773438, 0.05867520141601563, 0.05951385498046875, 0.05943807983398437, 0.05924556732177735, 0.05943296051025391, 0.059109375, 0.059747329711914064, 0.059643905639648435, 0.06056755065917969, 0.05980166244506836, 0.059280319213867186, 0.05948723220825195, 0.05943091201782227, 0.05953945541381836, 0.05943603134155274, 0.05937356948852539, 0.0597391357421875, 0.059597824096679686, 0.058409984588623044, 0.05871104049682617, 0.05933465576171875, 0.6235218505859375, 0.05865363311767578, 0.05926399993896484, 0.059676673889160155, 0.05959481430053711, 0.05947385787963867, 0.05966438293457031, 0.05968076705932617, 0.059650047302246094, 0.05991219329833984, 0.05946060943603516, 0.059510784149169924, 0.05825024032592773, 0.059481086730957033, 0.059510784149169924, 0.059448318481445314, 0.05944319915771484, 0.05946060943603516, 0.059428863525390625, 0.05969715118408203, 0.05980672073364258, 0.059582462310791014, 0.05940127944946289, 0.05918304061889648, 0.05903769683837891, 0.05903564834594727, 0.05914419174194336, 0.060007423400878904, 0.059786239624023435, 0.05904182434082031, 0.05932028961181641, 0.0596234245300293, 0.05931315231323242, 0.05959372711181641, 0.059417598724365236, 0.059566078186035154, 0.05921177673339844, 0.058913791656494144, 0.059474945068359375, 0.05947596740722656, 0.05943091201782227, 0.0596234245300293, 0.059415550231933595, 0.05990092849731445, 0.059663360595703124, 0.05955072021484375, 0.059412479400634766, 0.05959065628051758, 0.05926604843139648, 0.05959372711181641, 0.059493377685546876, 0.059458656311035155, 0.059190174102783204, 0.05972275161743164, 0.059493438720703125, 0.05991110229492187, 0.05935923385620117, 0.05942995071411133, 0.05948921585083008, 0.059535358428955076, 0.05799935913085937, 0.05778227233886719, 0.05850726318359375, 0.6226585693359376, 0.05902953720092773, 0.05935305786132813, 0.05927731323242187, 0.06000537490844727, 0.05953843307495117, 0.05951590347290039, 0.05741670227050781, 0.05765119934082031, 0.05912575912475586, 0.057673728942871094, 0.05749964904785156, 0.057409534454345705, 0.05713612747192383, 0.056941566467285154, 0.05738393783569336, 0.057409534454345705, 0.057504768371582034, 0.057264129638671876, 0.05749555206298828, 0.057332736968994144, 0.05736959838867187, 0.0576808967590332, 0.05938995361328125, 0.06057574462890625, 0.05938278579711914, 0.05959065628051758, 0.05842534255981445, 0.05766451263427735, 0.05919846343994141, 0.057627647399902344, 0.05816320037841797, 0.05820211029052735, 0.059109375, 0.057766910552978515, 0.057388031005859375, 0.05870694351196289, 0.05923535919189453, 0.05923123168945312, 0.05992652893066406, 0.059867103576660155, 0.05920870590209961, 0.05745151901245117, 0.05859430313110352, 0.059474945068359375, 0.05781094360351562, 0.058805248260498044, 0.05747814559936523, 0.058598400115966794, 0.05952819061279297, 0.059400257110595704, 0.05760095977783203, 0.059671550750732424, 0.05966745758056641, 0.05945446395874023, 0.059390975952148435, 0.059635711669921876, 0.05962956619262695, 0.05981695938110351, 0.05914931106567383, 0.058600448608398435, 0.05954969787597656, 0.057766910552978515, 0.6239180908203125, 0.06053068923950195, 0.06112870407104492, 0.05972275161743164, 0.05924454498291016, 0.0597022705078125, 0.05951692962646484, 0.059579391479492184, 0.059328510284423826, 0.0595333137512207, 0.05918003082275391, 0.05955788803100586, 0.05931622314453125, 0.0594411506652832, 0.059420673370361325, 0.059412479400634766, 0.05771366500854492, 0.05833318328857422, 0.057640960693359375, 0.0577525749206543, 0.05741567993164062, 0.05754880142211914, 0.05771059036254883, 0.057745407104492184, 0.05755904006958008, 0.05749657440185547, 0.05753241729736328, 0.05803212738037109, 0.05803417587280273, 0.057763839721679686, 0.05743513488769531, 0.057568321228027346, 0.05752313613891601, 0.05771366500854492, 0.05804646301269531, 0.058156032562255856, 0.059630592346191405, 0.05753343963623047, 0.05736959838867187, 0.057608192443847656, 0.05749248123168945, 0.05760313415527344, 0.057407424926757815, 0.057605121612548826, 0.057823230743408206, 0.05905408096313477, 0.05778227233886719, 0.057629695892333986, 0.057609249114990234, 0.05772796630859375, 0.057450496673583984, 0.058041343688964846, 0.05773823928833008, 0.057662464141845705, 0.05769625473022461, 0.057491455078125, 0.05757747268676758, 0.05770240020751953, 0.05764812850952149, 0.057678848266601565, 0.05749964904785156, 0.057599998474121096, 0.057591808319091796, 0.622150634765625, 0.05772083282470703, 0.05746585464477539, 0.05763379287719726, 0.05734195327758789, 0.05864243316650391, 0.05824512100219727, 0.0575467529296875, 0.05751193618774414, 0.05752934265136719, 0.05732761764526367, 0.05749657440185547, 0.05744025421142578, 0.057524223327636716, 0.0591196174621582, 0.05774335861206055, 0.059643905639648435, 0.060537857055664064, 0.0596049919128418, 0.060581886291503906, 0.060014591217041016, 0.05957427215576172, 0.05956198501586914, 0.05975859069824219, 0.05988044738769531, 0.05833014297485352, 0.0590868148803711, 0.058676223754882816, 0.0594595832824707, 0.05944319915771484, 0.05970534515380859, 0.05982617568969727, 0.05938585662841797, 0.05920870590209961, 0.059415550231933595, 0.06034022521972656, 0.059717632293701174, 0.05967871856689453, 0.05945241546630859, 0.059540481567382814, 0.05938790512084961, 0.059786239624023435, 0.05946060943603516, 0.05948825454711914, 0.059216896057128904, 0.05953638458251953, 0.05947699356079102, 0.05993369674682617, 0.06045695877075195, 0.06139289474487305, 0.05952716827392578, 0.059509761810302736, 0.059442176818847656, 0.05957017517089844, 0.05947084808349609, 0.059481086730957033, 0.05932032012939453, 0.06026342391967773, 0.05931008148193359, 0.05923737716674805, 0.05933260726928711, 0.059504638671875, 0.059112449645996094]",tokens/s,14.871942709161203,,,main,False,False,True @@ -7318,7 +7318,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66777615-293a012649ce8b1152668997;3db1c489-43dd-4da5-9ebf-8d5389f5ac39) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbbd5-14f2127342f2aaee0873dd02;23e400ce-1a70-41ff-8178-4e8dcbf11b85) Repository Not Found for url: https://huggingface.co/m/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -7390,7 +7390,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667779fe-3e3ea3216577c0954e509adf;d53d6014-9e1b-4407-bf9d-f0c21fe55845) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbfb6-1c18536058a072242a18467d;763883cf-b473-4160-98c6-23b8308e1455) Repository Not Found for url: https://huggingface.co/v/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -7506,7 +7506,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66777a26-3488fcbf767a4823312d733b;c917f626-53c0-4a52-b2a3-800205a92282) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbfe6-1a4aac743dc4022d647fa314;fcf76d84-b655-4524-a39c-183aa7b4790e) Repository Not Found for url: https://huggingface.co/0/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -7617,7 +7617,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66777970-0b1bd110288952076685285f;b02ad0df-ab8c-4a9f-afc3-21610c91a747) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbf36-0985ae893696f8312a4ab264;2b25d2e1-485a-4041-9b35-5282f5dc8d4a) Repository Not Found for url: https://huggingface.co/2/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -7688,7 +7688,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6677789e-693e3f4832ab2a3a3e56bbfc;ebf29ef3-12d5-4dea-a431-f0ae383dcd16) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbe70-333de4cc43a8b91d1714d64a;cd70e35b-14fd-4df9-9e22-47b70fdc9dbf) Repository Not Found for url: https://huggingface.co/l/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -7841,7 +7841,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66777666-6274423c184b9c796ecaff36;3f17890a-1b81-4469-b318-d3eadd67f470) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbc32-0c07c08d024eabcc6e699a97;32a6a3af-a93d-401d-819d-4fbc20fa4d36) Repository Not Found for url: https://huggingface.co/s/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -7911,7 +7911,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6677799d-2d523bcd3ad8cebd6fcc9787;730e1bcc-f566-4d1c-aaee-cf413c004858) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbf62-4590dba816d2aeaa48c1a7bf;e129e90c-8e78-40ae-ad68-506f4c4e29ab) Repository Not Found for url: https://huggingface.co/B/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -7970,7 +7970,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: CodeGenForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmptfhqxa5h/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: CodeGenForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpc5107qhu/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,True 4bit-bnb-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-2.7b,,cuda,0,42,,,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,MB,1278.603264,3276.275712,0.0,2629.828608,2358.099968,s,10,1.363180648803711,0.1363180648803711,0.00026217317487357416,0.1361774597167969,0.13673182220458985,0.1368010154724121,0.13685637008666993,"[0.13687020874023437, 0.1363632354736328, 0.13641459655761717, 0.13614486694335937, 0.1360360565185547, 0.1361745300292969, 0.13614434814453125, 0.13671644592285156, 0.1361803894042969, 0.13613597106933595]",tokens/s,1877.9609307442738,kWh,1.6118631997444966e-06,8.832266388659214e-07,9.71787188840562e-06,1.2212961727016039e-05,tokens/kWh,20961336.465478946,MB,1278.603264,3276.275712,0.0,2629.828608,2358.102528,s,10,29.777363037109374,2.9777363037109374,0.02588726324207309,2.9713291015625,3.018982763671875,3.0193787109375,3.0196954687499997,"[2.9692578125, 2.954369873046875, 2.95029638671875, 2.94651025390625, 2.959674560546875, 3.018894775390625, 2.99993896484375, 2.973400390625, 3.019774658203125, 2.985245361328125]",tokens/s,21.157011089762264,kWh,3.511612323535098e-05,1.924535570923163e-05,7.12104706319921e-05,0.00012557194957657467,tokens/kWh,501704.4030329572,,s,629,30.95603612136841,0.04921468381775582,0.016312179544641713,0.046707710266113284,0.04863692855834961,0.04895887298583985,0.1833139434814453,"[0.04888371276855469, 0.04917145538330078, 0.048249855041503906, 0.04791910552978516, 0.04700057601928711, 0.04656435012817383, 0.04804095840454101, 0.048145408630371096, 0.05266739273071289, 0.052951038360595705, 0.04698521423339844, 0.04662681579589844, 0.04637593460083008, 0.04695449447631836, 0.046437374114990236, 0.04646912002563477, 0.046312446594238284, 0.046415870666503906, 0.046268417358398435, 0.04664115142822266, 0.04859187316894531, 0.048347137451171876, 0.048274433135986325, 0.048266239166259765, 0.04822531127929688, 0.046888927459716796, 0.04660531234741211, 0.046668800354003906, 0.0465428466796875, 0.04667289733886719, 0.04662374496459961, 0.046584831237792966, 0.046617633819580076, 0.04658787155151367, 0.04661043167114258, 0.0464793586730957, 0.04622335815429687, 0.046317569732666014, 0.04641593551635742, 0.046725055694580075, 0.04673023986816406, 0.04653977584838867, 0.046630912780761716, 0.046773246765136715, 0.04663603210449219, 0.046601215362548826, 0.04661043167114258, 0.04664422225952149, 0.0466063346862793, 0.04671692657470703, 0.04665651321411133, 0.04674764633178711, 0.046693374633789066, 0.04663808059692383, 0.04658892822265625, 0.0465797119140625, 0.04658278274536133, 0.046614528656005856, 0.04653158569335938, 0.0465797119140625, 0.04661964797973633, 0.046682113647460936, 0.18365440368652344, 0.046698497772216796, 0.046683135986328124, 0.04674764633178711, 0.04665350341796875, 0.046623680114746095, 0.04651724624633789, 0.04652339172363281, 0.0466165771484375, 0.046535678863525394, 0.04666572952270508, 0.046543872833251954, 0.04715827178955078, 0.049702945709228515, 0.04864508819580078, 0.04698214340209961, 0.04639231872558594, 0.04759756851196289, 0.04860313415527344, 0.04817203140258789, 0.04827340698242188, 0.04671078491210937, 0.04645171356201172, 0.04654489517211914, 0.04678246307373047, 0.046707710266113284, 0.04657766342163086, 0.04651827239990235, 0.04720640182495117, 0.04652032089233398, 0.04651007843017578, 0.04648038482666016, 0.04660019302368164, 0.04666572952270508, 0.04597760009765625, 0.045873153686523435, 0.046465023040771485, 0.04595916748046875, 0.04644659042358398, 0.04680294418334961, 0.0465428466796875, 0.04660639953613281, 0.04682745742797852, 0.04664115142822266, 0.046595073699951174, 0.04654694366455078, 0.04647731018066406, 0.04648857498168945, 0.046524417877197265, 0.046465023040771485, 0.049186817169189455, 0.04894003295898437, 0.047738910675048825, 0.04649980926513672, 0.04653366470336914, 0.046507999420166014, 0.048075775146484374, 0.04660326385498047, 0.04653158569335938, 0.04651827239990235, 0.04651212692260742, 0.046837760925292966, 0.04677119827270508, 0.1835816955566406, 0.04658380889892578, 0.047800319671630856, 0.046117889404296876, 0.046486526489257815, 0.046491649627685545, 0.046429183959960936, 0.04612607955932617, 0.046516223907470705, 0.04642201614379883, 0.0464640007019043, 0.04640972900390625, 0.047303680419921876, 0.04697292709350586, 0.046301185607910154, 0.04644454574584961, 0.04625612640380859, 0.046429183959960936, 0.046516223907470705, 0.04658790588378906, 0.046541854858398436, 0.046540767669677734, 0.04672512054443359, 0.0474337272644043, 0.0465428466796875, 0.04739686584472656, 0.0466165771484375, 0.0465797119140625, 0.046473217010498044, 0.046413822174072264, 0.04655001449584961, 0.04657766342163086, 0.046514175415039063, 0.04658796691894531, 0.04752275085449219, 0.04734975814819336, 0.04658892822265625, 0.04655411148071289, 0.04851200103759765, 0.04827852630615234, 0.0474224624633789, 0.04655001449584961, 0.046489601135253904, 0.046475265502929686, 0.048115711212158206, 0.04731289672851562, 0.046516223907470705, 0.04650188827514649, 0.049426433563232425, 0.04711423873901367, 0.04655104064941406, 0.046489601135253904, 0.04655615997314453, 0.04638924789428711, 0.046486526489257815, 0.04640563201904297, 0.04651212692260742, 0.04775321578979492, 0.04838092803955078, 0.04723814392089844, 0.04639539337158203, 0.04654694366455078, 0.04658790588378906, 0.18326527404785156, 0.04662681579589844, 0.04649369430541992, 0.046584831237792966, 0.04655513763427734, 0.04655923080444336, 0.04653875350952148, 0.04658995056152344, 0.04652851104736328, 0.04656639862060547, 0.04646297454833984, 0.04653363037109375, 0.04658995056152344, 0.046565376281738284, 0.0465530891418457, 0.04670361709594727, 0.046663681030273435, 0.04671897506713867, 0.04660019302368164, 0.04658073425292969, 0.046650367736816405, 0.04658687973022461, 0.04647731018066406, 0.046604286193847655, 0.04630326461791992, 0.046703582763671876, 0.05004185485839844, 0.046838783264160154, 0.04659609603881836, 0.04656435012817383, 0.046630912780761716, 0.046698558807373045, 0.04662172698974609, 0.04658371353149414, 0.0465797119140625, 0.04662579345703125, 0.047288318634033204, 0.048198654174804685, 0.04846694564819336, 0.04801331329345703, 0.04657664108276367, 0.046884864807128904, 0.048734302520751956, 0.046059425354003904, 0.04659616088867188, 0.04658988952636719, 0.04660636901855469, 0.04648137664794922, 0.04628275299072265, 0.046652416229248046, 0.046268417358398435, 0.046714881896972656, 0.04660838317871094, 0.04664934539794922, 0.04652544021606445, 0.04658585739135742, 0.04658995056152344, 0.04665651321411133, 0.0466063346862793, 0.046688255310058595, 0.046693374633789066, 0.046680065155029295, 0.04660224151611328, 0.18300210571289063, 0.04653670501708984, 0.046473217010498044, 0.0461578254699707, 0.046496768951416016, 0.04647423934936523, 0.04640256118774414, 0.04640665435791016, 0.04638003158569336, 0.046358528137207033, 0.04634316635131836, 0.046322689056396485, 0.04616806411743164, 0.046317569732666014, 0.046516223907470705, 0.04638412857055664, 0.04650495910644531, 0.04644454574584961, 0.046458881378173826, 0.04602982330322265, 0.04636876678466797, 0.046475265502929686, 0.046412799835205076, 0.046519294738769534, 0.046527488708496094, 0.04621311950683594, 0.046467071533203126, 0.04649369430541992, 0.04691763305664062, 0.04690124893188476, 0.04652544021606445, 0.046150657653808595, 0.04600428771972656, 0.04650182342529297, 0.04653055953979492, 0.04657766342163086, 0.04649267196655273, 0.04682963180541992, 0.046623680114746095, 0.04650495910644531, 0.04646809768676758, 0.046461952209472655, 0.047339519500732424, 0.048694271087646485, 0.04841471862792969, 0.04851200103759765, 0.04836249542236328, 0.047900672912597655, 0.04835948944091797, 0.048065471649169925, 0.04837171173095703, 0.04834201431274414, 0.04732108688354492, 0.047524864196777344, 0.0466063346862793, 0.04696985626220703, 0.047857662200927735, 0.048399360656738284, 0.048811008453369144, 0.047870975494384765, 0.04825497436523438, 0.048231422424316404, 0.04836454391479492, 0.18703053283691407, 0.04863692855834961, 0.048642047882080076, 0.04825190353393555, 0.048363521575927736, 0.04759347152709961, 0.048571392059326174, 0.048143360137939455, 0.04746547317504883, 0.04823654556274414, 0.048881664276123046, 0.048694271087646485, 0.04863692855834961, 0.04817715072631836, 0.04658790588378906, 0.04653055953979492, 0.04757913589477539, 0.0470302734375, 0.04621004867553711, 0.04739891052246094, 0.04717567825317383, 0.047252479553222655, 0.04823040008544922, 0.04765497589111328, 0.04656838226318359, 0.04729753494262695, 0.04845977783203125, 0.04829695892333984, 0.04857958221435547, 0.048396289825439455, 0.048353279113769534, 0.04774604797363281, 0.04651724624633789, 0.0465428466796875, 0.04867177581787109, 0.04863484954833985, 0.046604286193847655, 0.047421440124511716, 0.047592449188232425, 0.046993408203125, 0.04728115081787109, 0.0483164176940918, 0.04786175918579102, 0.04827238464355469, 0.04739788818359375, 0.04658892822265625, 0.047234046936035154, 0.04739891052246094, 0.046611457824707034, 0.04801126480102539, 0.048333824157714846, 0.04657664108276367, 0.04751564788818359, 0.0487014389038086, 0.048292865753173826, 0.04835635375976562, 0.04862259292602539, 0.048143360137939455, 0.049186817169189455, 0.04864614486694336, 0.04837887954711914, 0.04873011016845703, 0.05090409469604492, 0.1834936065673828, 0.04677836990356445, 0.046693374633789066, 0.046698497772216796, 0.046824447631835936, 0.04659302520751953, 0.04657561492919922, 0.046635009765625, 0.04619468688964844, 0.04654694366455078, 0.048099327087402347, 0.04747673416137695, 0.04760268783569336, 0.04588032150268555, 0.04759347152709961, 0.048570369720458986, 0.048827392578125, 0.047435840606689456, 0.047363006591796875, 0.047933441162109375, 0.04711936187744141, 0.04856524658203125, 0.047963134765625, 0.04738457489013672, 0.04847923278808594, 0.0484771842956543, 0.04849356842041016, 0.04822220611572266, 0.047645694732666014, 0.04873932647705078, 0.04708966445922851, 0.04839321517944336, 0.048563201904296874, 0.04717977523803711, 0.046927871704101565, 0.046620670318603515, 0.046639102935791016, 0.04680089569091797, 0.047659008026123044, 0.047280128479003904, 0.05217792129516602, 0.04737740707397461, 0.0483133430480957, 0.04905779266357422, 0.04962508773803711, 0.04874444961547852, 0.048589824676513675, 0.04659814453125, 0.04643532943725586, 0.04657664108276367, 0.04866457748413086, 0.04680908966064453, 0.04842291259765625, 0.047442943572998046, 0.04648448181152344, 0.04641689682006836, 0.047323135375976565, 0.048643070220947264, 0.04859187316894531, 0.0487823371887207, 0.047777793884277345, 0.04702310562133789, 0.04630220794677734, 0.1877821502685547, 0.04894412612915039, 0.04664115142822266, 0.04641177749633789, 0.04643430328369141, 0.04662579345703125, 0.046532608032226565, 0.04658790588378906, 0.04613836669921875, 0.046475265502929686, 0.046382080078125, 0.046450687408447267, 0.04670873641967774, 0.04652134323120117, 0.046268417358398435, 0.045902847290039066, 0.046025726318359376, 0.04748185729980469, 0.04684902572631836, 0.046461952209472655, 0.047372287750244144, 0.04662169647216797, 0.04873625564575195, 0.04875161743164062, 0.04831545639038086, 0.04846995162963867, 0.046698497772216796, 0.04655411148071289, 0.04661248016357422, 0.04654079818725586, 0.04655001449584961, 0.046561279296875, 0.04655411148071289, 0.04647423934936523, 0.04652339172363281, 0.0464824333190918, 0.04624998474121094, 0.04660940933227539, 0.04664524841308594, 0.04669747161865234, 0.04654489517211914, 0.04839321517944336, 0.04714393615722656, 0.046601215362548826, 0.048126976013183595, 0.048192512512207034, 0.046617599487304685, 0.04764876937866211, 0.04862464141845703, 0.048363521575927736, 0.04870355224609375, 0.04846380615234375, 0.048282623291015625, 0.04703539276123047, 0.047862785339355465, 0.047541248321533204, 0.04650393676757812, 0.046617599487304685, 0.046676990509033206, 0.048495616912841794, 0.04905779266357422, 0.04838915252685547, 0.046614494323730465, 0.18333287048339844, 0.04656025695800781, 0.046601215362548826, 0.04660838317871094, 0.04676095962524414, 0.04911718368530273, 0.04858879852294922, 0.04851923370361328, 0.049417152404785156, 0.04852633666992188, 0.048494590759277346, 0.04827033615112305, 0.050447425842285155, 0.04946220779418945, 0.04804710388183594, 0.0470384635925293, 0.048968704223632815, 0.0485928955078125, 0.047710208892822265, 0.04654079818725586, 0.04716032028198242, 0.04854988861083984, 0.048535552978515625, 0.04889395141601562, 0.048568321228027345, 0.04860927963256836, 0.048935935974121096, 0.04853964614868164, 0.04840755081176758, 0.04842905426025391, 0.04652134323120117, 0.04660326385498047, 0.047483905792236325, 0.04876697540283203, 0.04867993545532227, 0.04855295944213867, 0.048557056427001956, 0.04691046524047852, 0.04793552017211914, 0.04787401580810547, 0.04848332977294922, 0.04837273788452148, 0.048336894989013675, 0.048418815612792966, 0.04837376022338867, 0.047164417266845705, 0.04696780776977539, 0.04809318542480469, 0.04653875350952148, 0.046362686157226565, 0.04724319839477539, 0.04828672027587891, 0.04692582321166992, 0.04737945556640625, 0.048061439514160156, 0.04743270492553711, 0.047702014923095705, 0.048449569702148435, 0.04834406280517578, 0.048519134521484375, 0.04654182434082031, 0.04656639862060547, 0.047287296295166016, 0.18399539184570313, 0.046808063507080076, 0.04854272079467774, 0.048263168334960936, 0.04669852828979492, 0.04928508758544922, 0.046868480682373044, 0.046581760406494144, 0.046752769470214846, 0.05028966522216797, 0.048942081451416014, 0.0466247673034668, 0.04674150466918945, 0.046622718811035156, 0.046601215362548826, 0.04654489517211914, 0.04664012908935547, 0.04781158447265625, 0.04701696014404297, 0.04669235229492188, 0.04635340881347656, 0.04624076843261719, 0.0465797119140625, 0.04681216049194336, 0.04678041458129883, 0.04680908966064453, 0.048568321228027345, 0.04837887954711914, 0.048671745300292966, 0.04843622589111328, 0.04868096160888672, 0.04844339370727539, 0.04698316955566406, 0.04852121734619141, 0.046637054443359374, 0.04786380767822265, 0.04649683380126953, 0.04812998580932617, 0.0482979850769043, 0.046993408203125, 0.04710297775268555, 0.04897587203979492, 0.04673535919189453, 0.046611457824707034, 0.046875648498535157, 0.046709758758544925, 0.04660940933227539, 0.046548992156982424, 0.04666777420043945, 0.046676990509033206, 0.046685184478759766, 0.04659609603881836, 0.046706687927246096, 0.04922675323486328, 0.04920729446411133, 0.04852019119262695, 0.04854272079467774, 0.048086017608642576, 0.046620670318603515, 0.0466063346862793, 0.046553150177001956, 0.046623680114746095, 0.04688281631469727]",tokens/s,20.319138972893636,,,main,False,False,True @@ -8013,7 +8013,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66777877-1f7297001acd91355365df2b;e2cd045c-791d-4fa0-8640-4810d119145e) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbe47-7b45ded012f1b6b160d20ce5;284231d0-0283-40a4-86d9-6cb0c39065dc) Repository Not Found for url: https://huggingface.co/a/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -8071,7 +8071,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpnipn3c64/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpzpn48b98/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,True 4bit-bnb-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,1,1,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.218-208.862.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): @@ -8113,7 +8113,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66777a75-34f9ef515e24a1470d4472e8;5c23ac40-770e-4df2-9cd0-48e2888b8b6b) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cc039-0a368cd75d1bff0c427918cc;fb2e908e-1f5f-44e2-9b3e-c609c6bdf4a4) Repository Not Found for url: https://huggingface.co/1/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -8184,7 +8184,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66777825-224c908621fbdcb240d00c6c;8c9e4e39-fd92-4643-96ab-bc33042f3b31) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbdf6-5482ab123e5b89a10dcfe29b;20b8f2aa-4970-44d2-8bfd-795f776d0110) Repository Not Found for url: https://huggingface.co/t/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -8274,7 +8274,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667a147c-28f2f8281101ce6307d72082;71753a12-e6db-43fd-ac89-baa4f2857d40) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667cb6ad-08792b92568804fe0fcf14b2;d4a16552-9fed-4991-9199-a81866613272) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/recurrentgemma-2b/resolve/main/config.json. @@ -8427,7 +8427,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: CodeGenForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpizgvshnd/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: CodeGenForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpyvl5tbmk/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,True 4bit-bnb-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/recurrentgemma-7b,google/recurrentgemma-7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.218-208.862.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): @@ -8469,7 +8469,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667a14a7-3ed4d3b372956296714dcfbd;0917ebf3-68c7-4b75-823a-b360ecb627d4) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cb6d7-55ac437b35e69f616db246c5;749e0184-9be1-4243-be77-4be7ba8c6bd8) Repository Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -8530,7 +8530,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667a1423-3a05a6ed01fd4fa737d4fe18;0a2e8f1f-9481-4297-9efd-689ab50f0e80) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667cb65c-2556a27d4f6b12794e083c26;2a62206a-62c5-4291-9a0c-3a6244f119b8) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-2b/resolve/main/config.json. @@ -8708,7 +8708,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667a144f-39c6357d71d932434951e621;7b2878e1-fe10-430a-b2ab-4e18482f18c1) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667cb686-54f753e967009a8341c6d7ff;05047eb7-e46c-498f-b6a7-276ca45a1e97) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-7b/resolve/main/config.json. @@ -8793,7 +8793,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667777d5-4ad127f24907e11e7709b2d7;a71c989a-d8ab-4175-b423-fb9e6190336b) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbd9e-29c8ee553905192359c80de7;1cae282e-8fee-4ff7-a382-4f4745229959) Repository Not Found for url: https://huggingface.co/i/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -8866,7 +8866,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667777ae-1151b07214fa8162237187c9;34cc2e65-5c2c-48cf-abdd-194b071403de) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbd76-5c36c73451e7e8111d63e4b6;d21c30cf-94de-40ba-ac99-0443aab063e0) Repository Not Found for url: https://huggingface.co/M/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -8928,7 +8928,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpq_ykfflt/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpksvvawfn/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,True 4bit-bnb-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-125m,,cuda,0,42,,,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,MB,1160.658944,858.259456,0.0,211.812352,169.494528,s,10,0.24654441452026363,0.024654441452026367,0.0009328759589081531,0.024561103820800782,0.025379053497314452,0.02620999069213867,0.026874740447998046,"[0.02704092788696289, 0.025194400787353514, 0.02377552032470703, 0.024137279510498048, 0.0248092155456543, 0.023764095306396484, 0.024857343673706053, 0.02481926345825195, 0.023833375930786133, 0.024312992095947266]",tokens/s,10383.52462772825,kWh,2.878979259357159e-07,1.5775497047891068e-07,6.606060975572359e-07,1.1062589939718624e-06,tokens/kWh,231410547.97743985,MB,1160.658944,858.259456,0.0,211.812352,178.502656,s,10,12.388275268554686,1.2388275268554687,0.015951853779150135,1.239268798828125,1.2552052124023438,1.2602007507324218,1.2641971813964843,"[1.253686279296875, 1.2651962890625, 1.2181383056640624, 1.2391370849609376, 1.2394005126953125, 1.2504305419921875, 1.21899365234375, 1.2204739990234375, 1.2287235107421874, 1.2540950927734376]",tokens/s,50.85453675695573,kWh,1.443376470801314e-05,7.909398606806443e-06,2.189574528084272e-05,4.4238908595662296e-05,tokens/kWh,1424085.7652210991,,s,629,12.585495553970341,0.02000873696974616,0.002962854768789734,0.019876863479614256,0.02025267219543457,0.020336845016479494,0.04331901016235351,"[0.02230784034729004, 0.019478591918945312, 0.01914771270751953, 0.019123199462890626, 0.01940377616882324, 0.019408895492553712, 0.01923788833618164, 0.019212287902832033, 0.019133440017700197, 0.019200000762939453, 0.019137535095214844, 0.019135488510131835, 0.019117055892944337, 0.019139583587646485, 0.019149824142456053, 0.019310592651367187, 0.019128320693969726, 0.019135488510131835, 0.019120128631591796, 0.019164159774780275, 0.019017728805541992, 0.018712575912475587, 0.018928640365600585, 0.01920614433288574, 0.019151872634887695, 0.019176448822021484, 0.01921331214904785, 0.019141632080078123, 0.01943244743347168, 0.021131263732910157, 0.02114566421508789, 0.02101958465576172, 0.020304895401000975, 0.02029267120361328, 0.020274112701416016, 0.020288511276245116, 0.020207616806030275, 0.020247552871704103, 0.020207616806030275, 0.02024345588684082, 0.020297727584838866, 0.02024550437927246, 0.02020966339111328, 0.02007859230041504, 0.02031001663208008, 0.02024345588684082, 0.02025267219543457, 0.020264959335327147, 0.02025062370300293, 0.020370431900024414, 0.020256767272949217, 0.020808704376220705, 0.020332544326782227, 0.020361215591430663, 0.02026803207397461, 0.020328447341918944, 0.02031724739074707, 0.020308927536010744, 0.02030182456970215, 0.02025984001159668, 0.02025881576538086, 0.020239360809326173, 0.04594483184814453, 0.020180992126464844, 0.02021785545349121, 0.020272192001342774, 0.020241344451904297, 0.02023321533203125, 0.02025984001159668, 0.020207616806030275, 0.020271104812622072, 0.020212736129760742, 0.020199424743652345, 0.02041753578186035, 0.020349952697753908, 0.020256767272949217, 0.020339712142944336, 0.02027724838256836, 0.02022502326965332, 0.020174848556518556, 0.020189184188842774, 0.02025267219543457, 0.020328447341918944, 0.020279296875, 0.020265983581542968, 0.02007859230041504, 0.02024345588684082, 0.0202926082611084, 0.02022809600830078, 0.02025164794921875, 0.02026803207397461, 0.02002841567993164, 0.020358144760131838, 0.020280319213867186, 0.020341760635375978, 0.020381696701049806, 0.02031001663208008, 0.01963929557800293, 0.01957683181762695, 0.01986355209350586, 0.019989503860473632, 0.019963903427124022, 0.019986431121826173, 0.019973119735717772, 0.01988403129577637, 0.01990656089782715, 0.019887104034423828, 0.019951616287231445, 0.019901439666748046, 0.019933183670043944, 0.01989836883544922, 0.019877887725830077, 0.01988812828063965, 0.019353599548339845, 0.019693567276000978, 0.019942399978637695, 0.019877887725830077, 0.019886079788208007, 0.01990553665161133, 0.019942399978637695, 0.019899391174316407, 0.0194467830657959, 0.019901439666748046, 0.019946495056152345, 0.01985024070739746, 0.04329983901977539, 0.018983936309814452, 0.01904128074645996, 0.018948095321655273, 0.01903411293029785, 0.018976768493652343, 0.018988031387329102, 0.01901055908203125, 0.01905971145629883, 0.018968704223632813, 0.018982784271240234, 0.019094528198242186, 0.01905971145629883, 0.01900032043457031, 0.01907302474975586, 0.019054592132568358, 0.019062784194946288, 0.01904742431640625, 0.019396608352661132, 0.01900748825073242, 0.01903718376159668, 0.019099647521972657, 0.01900339126586914, 0.019122175216674805, 0.01904435157775879, 0.019050559997558593, 0.019016639709472656, 0.01904435157775879, 0.019466239929199217, 0.0190699520111084, 0.01906892776489258, 0.01906175994873047, 0.018949119567871094, 0.01899622344970703, 0.01903615951538086, 0.019058687210083008, 0.01903001594543457, 0.01965977668762207, 0.019956735610961913, 0.02103603172302246, 0.020271104812622072, 0.02001203155517578, 0.019939327239990236, 0.019982336044311523, 0.019927040100097656, 0.020059135437011717, 0.01990553665161133, 0.019886079788208007, 0.019917856216430663, 0.019955680847167968, 0.0198604793548584, 0.01985638427734375, 0.019731456756591798, 0.019915775299072267, 0.019812351226806642, 0.01899519920349121, 0.019110912322998046, 0.01908940887451172, 0.01901158332824707, 0.01906483268737793, 0.01903104019165039, 0.019042303085327148, 0.019615743637084963, 0.04370739364624023, 0.01906790351867676, 0.01904435157775879, 0.019017728805541992, 0.019418111801147463, 0.01903104019165039, 0.01908121681213379, 0.01903615951538086, 0.019094528198242186, 0.019175424575805664, 0.019152959823608397, 0.01912416076660156, 0.0191016960144043, 0.01900851249694824, 0.018962432861328125, 0.019132415771484376, 0.019056640625, 0.0190382080078125, 0.019104768753051758, 0.019113983154296875, 0.01903001594543457, 0.019042303085327148, 0.019107839584350587, 0.019056640625, 0.019095552444458007, 0.019088384628295898, 0.019715072631835938, 0.019935232162475586, 0.019949567794799804, 0.019964927673339843, 0.019920896530151368, 0.01989433670043945, 0.019905471801757814, 0.02003763198852539, 0.020720640182495118, 0.020930559158325195, 0.019993663787841797, 0.01995462417602539, 0.020130815505981444, 0.019916799545288084, 0.01992095947265625, 0.019911615371704102, 0.019927040100097656, 0.019916799545288084, 0.019957759857177734, 0.019934207916259765, 0.01992192077636719, 0.019904512405395508, 0.01987174415588379, 0.019899391174316407, 0.020517887115478514, 0.02150912094116211, 0.02027827262878418, 0.02008166313171387, 0.02004275131225586, 0.020414464950561522, 0.019876863479614256, 0.019938304901123048, 0.019895456314086915, 0.019840864181518553, 0.019991552352905274, 0.019999744415283204, 0.019950592041015625, 0.045183998107910156, 0.019936256408691407, 0.019985408782958985, 0.019911680221557617, 0.019895296096801757, 0.019914751052856446, 0.019899391174316407, 0.01985945510864258, 0.019899391174316407, 0.019896320343017578, 0.019955743789672853, 0.019968032836914062, 0.02001299285888672, 0.020059135437011717, 0.02003660774230957, 0.019944448471069336, 0.019941375732421874, 0.02008166313171387, 0.020024320602416993, 0.019936256408691407, 0.01993427276611328, 0.019801023483276368, 0.01990553665161133, 0.019941375732421874, 0.020172800064086914, 0.019942399978637695, 0.01999667167663574, 0.020011007308959963, 0.020002815246582033, 0.019954687118530275, 0.019938304901123048, 0.019895296096801757, 0.019937280654907227, 0.019929088592529298, 0.019131391525268555, 0.019140607833862306, 0.01909760093688965, 0.019090431213378906, 0.019100671768188478, 0.01905561637878418, 0.0190248966217041, 0.0194969596862793, 0.01904640007019043, 0.019160064697265625, 0.01893791961669922, 0.01905971145629883, 0.019122112274169923, 0.019056640625, 0.01908121681213379, 0.019087360382080077, 0.01908940887451172, 0.019878911972045898, 0.019887104034423828, 0.019908607482910155, 0.01992192077636719, 0.01992192077636719, 0.019883007049560548, 0.019914751052856446, 0.019934207916259765, 0.019113983154296875, 0.01921023941040039, 0.019126272201538085, 0.019074047088623047, 0.043225086212158204, 0.019141632080078123, 0.01898716735839844, 0.018990943908691407, 0.0190515193939209, 0.019430400848388672, 0.019955711364746095, 0.01999667167663574, 0.019954687118530275, 0.01993011283874512, 0.019891199111938478, 0.019885055541992186, 0.019919872283935547, 0.019927040100097656, 0.019869695663452147, 0.019809280395507813, 0.019934207916259765, 0.019854335784912108, 0.019826688766479493, 0.019891199111938478, 0.019903488159179687, 0.01989017677307129, 0.019886079788208007, 0.019942399978637695, 0.019780607223510743, 0.019745792388916016, 0.019920896530151368, 0.0198973445892334, 0.019977216720581056, 0.020221952438354493, 0.019899391174316407, 0.019877887725830077, 0.019928064346313477, 0.01991891288757324, 0.01990649604797363, 0.01987993621826172, 0.019963903427124022, 0.02006425666809082, 0.019954687118530275, 0.019834880828857423, 0.01989836883544922, 0.01989836883544922, 0.019955711364746095, 0.01995267105102539, 0.019988447189331054, 0.019922943115234376, 0.019895296096801757, 0.019478527069091797, 0.019919872283935547, 0.0198656005859375, 0.01991372871398926, 0.019918848037719726, 0.019928064346313477, 0.01993011283874512, 0.02042367935180664, 0.019944448471069336, 0.019966976165771484, 0.019916799545288084, 0.019975168228149414, 0.019900415420532228, 0.019902463912963866, 0.019899391174316407, 0.019944448471069336, 0.044477439880371096, 0.019088384628295898, 0.01927475166320801, 0.019088384628295898, 0.018994176864624023, 0.01900748825073242, 0.019083263397216797, 0.01906380844116211, 0.01906790351867676, 0.019154943466186524, 0.019079168319702147, 0.01902694320678711, 0.0190515193939209, 0.019102720260620116, 0.019095552444458007, 0.019500032424926757, 0.019173376083374022, 0.019125247955322267, 0.01908121681213379, 0.01906892776489258, 0.019113983154296875, 0.019098623275756836, 0.019082239151000976, 0.01903923225402832, 0.019092479705810548, 0.018961439132690428, 0.018838495254516602, 0.018892799377441406, 0.018918399810791017, 0.019075071334838867, 0.019138559341430664, 0.019129344940185547, 0.019136512756347656, 0.019009536743164062, 0.018994176864624023, 0.01909760093688965, 0.01906073570251465, 0.019086336135864256, 0.019072000503540038, 0.019883039474487305, 0.01924399948120117, 0.01904435157775879, 0.019086336135864256, 0.0190515193939209, 0.01923174476623535, 0.019760128021240234, 0.019893247604370116, 0.01988915252685547, 0.01990553665161133, 0.019934207916259765, 0.019931135177612306, 0.019982336044311523, 0.01993011283874512, 0.019923967361450197, 0.02025164794921875, 0.019977216720581056, 0.019780607223510743, 0.019938304901123048, 0.019915775299072267, 0.019942399978637695, 0.02001408004760742, 0.019959808349609375, 0.019970048904418947, 0.04435353469848633, 0.019103744506835937, 0.019500032424926757, 0.019130367279052735, 0.019078144073486326, 0.01903718376159668, 0.019074047088623047, 0.019132415771484376, 0.01904947280883789, 0.019098623275756836, 0.019118080139160155, 0.01922662353515625, 0.019959871292114257, 0.020028352737426758, 0.01977446365356445, 0.019495935440063478, 0.018988031387329102, 0.019141632080078123, 0.01902899169921875, 0.019115007400512696, 0.019086368560791017, 0.019036127090454102, 0.019074047088623047, 0.019042303085327148, 0.01907711982727051, 0.0190248966217041, 0.019113983154296875, 0.01904640007019043, 0.019133440017700197, 0.01902899169921875, 0.019168256759643554, 0.01904025650024414, 0.01938230323791504, 0.019840991973876954, 0.019902463912963866, 0.019878944396972655, 0.01993724822998047, 0.019933183670043944, 0.019915775299072267, 0.01982771110534668, 0.01997929573059082, 0.02008675193786621, 0.019927040100097656, 0.019876863479614256, 0.019920896530151368, 0.019903520584106445, 0.019916767120361327, 0.019933183670043944, 0.019861503601074217, 0.019966976165771484, 0.01908531188964844, 0.01904640007019043, 0.019021823883056642, 0.01946316719055176, 0.01900748825073242, 0.01904128074645996, 0.01905971145629883, 0.01905766487121582, 0.019096576690673828, 0.018969600677490234, 0.019095552444458007, 0.019084287643432618, 0.01900748825073242, 0.04332646560668945, 0.019127328872680663, 0.01910371208190918, 0.01902899169921875, 0.01903001594543457, 0.01907302474975586, 0.01906483268737793, 0.019078208923339845, 0.019072959899902344, 0.01908940887451172, 0.01903513526916504, 0.019743743896484374, 0.019964927673339843, 0.020405248641967775, 0.02020249557495117, 0.019948543548583983, 0.019940351486206053, 0.01985536003112793, 0.01968435287475586, 0.019915775299072267, 0.019928064346313477, 0.01985024070739746, 0.019883007049560548, 0.019982400894165038, 0.019937215805053712, 0.019919872283935547, 0.01985843276977539, 0.01987583923339844, 0.020171775817871093, 0.019935232162475586, 0.019904512405395508, 0.019891199111938478, 0.01984819221496582, 0.0198973445892334, 0.019929088592529298, 0.01988198471069336, 0.019894271850585937, 0.019891199111938478, 0.019147775650024415, 0.019072000503540038, 0.018897920608520507, 0.0194201602935791, 0.01904435157775879, 0.01902694320678711, 0.018981887817382814, 0.019070976257324217, 0.01901260757446289, 0.018988031387329102, 0.01902592086791992, 0.01902284812927246, 0.01903513526916504, 0.018988031387329102, 0.01903104019165039, 0.019021823883056642, 0.019142656326293944, 0.019705856323242187, 0.01904947280883789, 0.019132415771484376, 0.01902284812927246, 0.019725311279296876, 0.02004172706604004, 0.01986867141723633, 0.01993011283874512, 0.04468019104003906, 0.01988915252685547, 0.01985740852355957, 0.019842048645019532, 0.019811328887939454, 0.019804159164428712, 0.020006912231445313, 0.0198604793548584, 0.019883007049560548, 0.0198154239654541, 0.019903488159179687, 0.019721216201782226, 0.019908607482910155, 0.01989017677307129, 0.019943424224853516, 0.020171775817871093, 0.019951616287231445, 0.019893247604370116, 0.01985740852355957, 0.01985945510864258, 0.01985638427734375, 0.01983590316772461, 0.01990553665161133, 0.019809280395507813, 0.019894271850585937, 0.01985228729248047, 0.019735647201538087, 0.02017475128173828, 0.019817472457885742, 0.020883455276489257, 0.020572160720825194, 0.02007961654663086, 0.02004684829711914, 0.019984384536743165, 0.01986355209350586, 0.01982156753540039, 0.019912704467773438, 0.01985024070739746, 0.019910655975341796, 0.019795967102050782, 0.01982464027404785, 0.019935232162475586, 0.01987174415588379, 0.01988812828063965, 0.019911680221557617, 0.019833856582641602, 0.01992192077636719, 0.01986662483215332, 0.019749984741210938, 0.019742624282836914, 0.01989836883544922, 0.01985024070739746, 0.019860511779785157, 0.019882976531982424, 0.019827743530273438, 0.019787744522094728, 0.019870719909667968, 0.019802112579345704, 0.01979903984069824, 0.01982361602783203, 0.019938304901123048, 0.019529727935791014, 0.019878911972045898]",tokens/s,49.97816711329813,,,main,False,False,True @@ -8958,7 +8958,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpltr74um5/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp1himpwow/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,True 4bit-bnb-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,x,x,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.218-208.862.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): @@ -9000,7 +9000,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6677791b-663fb8571915a6db62512f03;66def062-3377-47ac-8061-48f7949483d9) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbee6-404ae96a1cd99130060c6107;c2877d08-bc46-4e5b-8fd8-ac2b443160a5) Repository Not Found for url: https://huggingface.co/x/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -9101,7 +9101,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66777b86-2b1977636475fe7313b434a8;2dc2541a-afda-4b43-9278-7d7238848d10) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667cc14f-66505b203296419778068821;e2393fd6-249a-4df3-858a-06258982d887) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json. @@ -9191,7 +9191,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667778ef-5d9d637d0ff7bdda07792249;22ddd190-7ce6-4916-931a-b849be446b77) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbebe-7926799c495b8d5f0bdfdda2;47eab1c8-435a-470d-93d7-4c5cb9655807) Repository Not Found for url: https://huggingface.co/8/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -9280,7 +9280,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667a11e6-6aa9e69e39179e873a99ce8b;a4db5d84-1913-4eba-bec1-6011eef99989) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667cb42b-5e10655f62941b711cfb797f;ced86591-6fa4-4e9a-86ee-9dbb40a2579c) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/databricks/dbrx-base/resolve/main/config.json. @@ -9411,7 +9411,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6677784f-2b56627308de2d0e70b4baf4;d14a0e8b-b75d-4159-b501-e48bbce59cf0) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbe1e-5fced4ae4da82d1852c162ea;ac3572fd-8f70-4251-b19c-8d102bbbc3b3) Repository Not Found for url: https://huggingface.co/r/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -9470,7 +9470,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpemm2s4ow/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmppo54evrs/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,True 4bit-bnb-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,TencentARC/Mistral_Pro_8B_v0.1,,cuda,0,42,,,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,MB,2088.595456,7149.715456,0.0,6503.268352,6130.0736,s,10,4.354886688232422,0.43548866882324216,0.0009500705262691482,0.43506301879882814,0.4360664276123047,0.43712642974853516,0.4379744314575195,"[0.43818643188476564, 0.4357357177734375, 0.43495596313476564, 0.43503363037109377, 0.43495343017578125, 0.4350522766113281, 0.43583087158203127, 0.43494912719726564, 0.4350737609863281, 0.435115478515625]",tokens/s,587.8453753842819,kWh,5.1466148040720816e-06,2.82011353963595e-06,3.132101539495727e-05,3.92877437386653e-05,tokens/kWh,6516027.026211126,MB,2088.595456,7149.715456,0.0,6503.268352,6130.07616,s,10,52.50473486328125,5.250473486328125,0.09201662696971703,5.295069091796875,5.3397248046875,5.347057958984375,5.352924482421875,"[5.33809521484375, 5.3243525390625, 5.32192578125, 5.2103232421875, 5.2842451171875, 5.35439111328125, 5.30589306640625, 5.11702490234375, 5.11619873046875, 5.13228515625]",tokens/s,11.998917843133139,kWh,6.21844898056293e-05,3.408169959179007e-05,0.00014644368237224062,0.00024270987176965998,tokens/kWh,259569.1701398498,,s,629,56.34009081268308,0.08957089159409079,0.05164629418121464,0.08460291290283203,0.08540467224121094,0.08579256439208983,0.5164155053710938,"[0.0808826904296875, 0.08587264251708984, 0.08512409973144532, 0.08462745666503907, 0.083346435546875, 0.08475443267822266, 0.08069427490234375, 0.08058879852294921, 0.08235929870605468, 0.08576921844482421, 0.08510054779052735, 0.0855224609375, 0.08536675262451172, 0.08540262603759766, 0.085212158203125, 0.08498175811767578, 0.08550911712646485, 0.08472882843017578, 0.0831272964477539, 0.08487731170654297, 0.08553369903564453, 0.08621772766113281, 0.08491315460205077, 0.08398233795166016, 0.08506777954101563, 0.08603961944580078, 0.08499603271484375, 0.0849438705444336, 0.08485478210449218, 0.08516095733642579, 0.08507596588134765, 0.08500531005859376, 0.08534015655517578, 0.08566169738769532, 0.0850708465576172, 0.08518041229248047, 0.08493363189697266, 0.08562995147705078, 0.08498790740966797, 0.08498585510253906, 0.08526950073242187, 0.08520499420166015, 0.08501145935058593, 0.08523878479003906, 0.082661376953125, 0.0845854721069336, 0.08474726104736328, 0.0852326431274414, 0.08525414276123047, 0.08547122955322266, 0.08509030151367188, 0.08530841827392578, 0.08540364837646484, 0.08494489288330079, 0.08497561645507813, 0.08481587219238282, 0.08475341033935548, 0.08502579498291016, 0.08509542083740235, 0.08502067565917969, 0.08512102508544922, 0.08473705291748047, 0.5177200317382813, 0.08092364501953125, 0.08514765167236328, 0.0842577896118164, 0.08333106994628907, 0.08667545318603516, 0.08635289764404297, 0.08515583801269531, 0.08509849548339844, 0.0846909408569336, 0.08494908905029297, 0.08473385620117188, 0.08456703948974609, 0.08469811248779296, 0.08491417694091796, 0.08470425415039062, 0.080932861328125, 0.08101580810546875, 0.08081203460693359, 0.08047824096679687, 0.08428540802001953, 0.0845998077392578, 0.08174591827392579, 0.08479027557373046, 0.08491417694091796, 0.08499625396728516, 0.08498262023925782, 0.08467250823974609, 0.08499609375, 0.0854128646850586, 0.08483328247070313, 0.08498381042480468, 0.08468889617919922, 0.08487423706054688, 0.08468582153320313, 0.0848015365600586, 0.0852674560546875, 0.08585830688476563, 0.08500326538085938, 0.08540876770019531, 0.08543743896484375, 0.08685164642333984, 0.08631084442138671, 0.08498073577880859, 0.08495001220703124, 0.08483430480957031, 0.08489369964599609, 0.08494694519042968, 0.08446463775634766, 0.08407654571533203, 0.08470425415039062, 0.08453324890136718, 0.08203366088867188, 0.08531148529052734, 0.08482201385498046, 0.08499097442626953, 0.08538829040527343, 0.08473804473876953, 0.08488960266113281, 0.08508415985107422, 0.08515789031982422, 0.08494287872314453, 0.08479126739501953, 0.5212528686523438, 0.084890625, 0.08508108520507812, 0.08611840057373046, 0.08516403198242188, 0.08458956909179688, 0.08463667297363281, 0.08557878112792969, 0.08558902740478516, 0.08462841796875, 0.08482508850097656, 0.08524390411376953, 0.0847267837524414, 0.08469811248779296, 0.08470425415039062, 0.08479539489746094, 0.08464895629882813, 0.08452607727050782, 0.08485379028320313, 0.08477590179443359, 0.08533299255371093, 0.08490598297119141, 0.0849090576171875, 0.08484352111816407, 0.08475955200195312, 0.080753662109375, 0.08075059509277344, 0.08288256072998047, 0.08464486694335938, 0.08470015716552734, 0.08492134094238281, 0.08503501129150391, 0.08485887908935547, 0.08471449279785156, 0.084569091796875, 0.08482099151611328, 0.08460594940185547, 0.08432537841796875, 0.08461209869384766, 0.0847984619140625, 0.0846192626953125, 0.08486605072021484, 0.08466534423828125, 0.0845301742553711, 0.08466636657714843, 0.08366899108886719, 0.08472576141357421, 0.08461824035644532, 0.08455577850341797, 0.08461619567871094, 0.08433561706542969, 0.0844554214477539, 0.08434893035888671, 0.08446463775634766, 0.08080691528320312, 0.08115507507324218, 0.08446771240234376, 0.08471449279785156, 0.084748291015625, 0.08502272033691406, 0.08421692657470703, 0.08310678100585937, 0.08451372528076172, 0.5164462280273437, 0.08081817626953125, 0.08075878143310547, 0.08023859405517578, 0.08031231689453125, 0.0796651840209961, 0.08062461090087891, 0.08074342346191406, 0.08057036590576172, 0.08049049377441406, 0.08030207824707031, 0.08085913848876954, 0.08074854278564453, 0.08060313415527344, 0.0807567367553711, 0.08064617919921875, 0.08057341003417968, 0.0805775375366211, 0.08061542510986328, 0.08045158386230469, 0.08406835174560547, 0.08073932647705079, 0.08241868591308593, 0.08514765167236328, 0.08534220886230469, 0.08399565124511718, 0.08037580871582031, 0.08123699188232422, 0.08072294616699219, 0.08092572784423828, 0.08314777374267578, 0.08090415954589844, 0.08274329376220703, 0.08470527648925781, 0.0807874526977539, 0.08084786987304687, 0.08369868469238281, 0.08458444976806641, 0.08475750732421874, 0.08480563354492188, 0.08099635314941406, 0.08079974365234376, 0.08065535736083984, 0.08275865936279297, 0.08473603057861329, 0.08466326141357422, 0.0848384017944336, 0.08517120361328125, 0.08535552215576171, 0.08433663940429688, 0.08464588928222656, 0.08481587219238282, 0.08481484985351563, 0.08510975646972656, 0.08468275451660157, 0.0861143035888672, 0.08548454284667968, 0.08472064208984376, 0.08493055725097656, 0.08681574249267578, 0.08619725036621094, 0.08521932983398438, 0.08474726104736328, 0.5163919067382813, 0.08359117126464843, 0.08464691162109375, 0.08525004577636719, 0.08071167755126953, 0.0807608642578125, 0.08082121276855468, 0.08076799774169922, 0.08060006713867188, 0.08175001525878907, 0.08488243103027343, 0.08472166442871094, 0.08394445037841797, 0.08496435546875, 0.0847267837524414, 0.08517120361328125, 0.08457526397705079, 0.08479126739501953, 0.08574156951904296, 0.08485990142822265, 0.08481382751464844, 0.08460291290283203, 0.08464790344238281, 0.08110489654541016, 0.08444927978515625, 0.08513843536376953, 0.08490290832519531, 0.08127078247070313, 0.08087654113769531, 0.08389324951171875, 0.0847083511352539, 0.08498381042480468, 0.08485068511962891, 0.08423526763916016, 0.08490918731689454, 0.08473484802246094, 0.08484044647216797, 0.08504319763183593, 0.0810618896484375, 0.0811878433227539, 0.08054271697998047, 0.0806123504638672, 0.0807034912109375, 0.08408370971679688, 0.08534937286376953, 0.08435711669921875, 0.08488448333740234, 0.08481177520751954, 0.08471552276611329, 0.08489574432373047, 0.08478822326660156, 0.0851937255859375, 0.08481692504882812, 0.08490595245361328, 0.08488140869140624, 0.08486809539794922, 0.08466534423828125, 0.08551526641845703, 0.0852490234375, 0.08489266967773437, 0.08475852966308593, 0.08525721740722657, 0.0845660171508789, 0.5165946655273438, 0.08096256256103515, 0.08067686462402343, 0.08383283233642579, 0.08545587158203125, 0.08592691040039062, 0.08542105865478515, 0.08580812835693359, 0.08653107452392578, 0.08479027557373046, 0.08508006286621093, 0.08506777954101563, 0.08509951782226563, 0.08516403198242188, 0.08509235382080078, 0.08480870056152344, 0.08497872161865234, 0.0852776641845703, 0.08520703887939453, 0.08543846130371094, 0.08479027557373046, 0.08502579498291016, 0.08532685089111328, 0.08513433837890624, 0.08513228607177735, 0.0849991683959961, 0.08475341033935548, 0.08500121307373047, 0.08525004577636719, 0.08603238677978516, 0.0857149429321289, 0.08504524993896484, 0.08503193664550782, 0.08063385772705078, 0.08257945251464843, 0.08591462707519532, 0.08676454162597656, 0.08414208221435547, 0.08625151824951172, 0.0863109130859375, 0.085501953125, 0.08678092956542968, 0.08597503662109375, 0.0852490234375, 0.0853780517578125, 0.08530944061279297, 0.08511385345458984, 0.08488550567626953, 0.0850882568359375, 0.08629043579101563, 0.08513536071777343, 0.08496640014648438, 0.08540364837646484, 0.08508108520507812, 0.0853729248046875, 0.08491827392578125, 0.08530226898193359, 0.08532994842529297, 0.08504521942138672, 0.0850882568359375, 0.08540160369873047, 0.08498381042480468, 0.08480668640136718, 0.5231185302734375, 0.08555007934570312, 0.085212158203125, 0.0849991683959961, 0.08495820617675781, 0.08518758392333985, 0.08489881896972656, 0.08519577789306641, 0.08499404907226563, 0.0853544921875, 0.08507087707519531, 0.08468988800048828, 0.0851773452758789, 0.08512409973144532, 0.08484146881103516, 0.08553472137451172, 0.08454454040527344, 0.08496841430664062, 0.08502374267578125, 0.08565964508056641, 0.08568627166748047, 0.0853903350830078, 0.0851230697631836, 0.0851251220703125, 0.0851600341796875, 0.08550185394287109, 0.08520089721679687, 0.08487014770507813, 0.08488038635253906, 0.08486502075195312, 0.08505958557128906, 0.08553062438964844, 0.08553369903564453, 0.08480973052978516, 0.08504524993896484, 0.08505241394042969, 0.08521318054199219, 0.08505753326416016, 0.08548044586181641, 0.08534835052490235, 0.08536576080322265, 0.08545996856689453, 0.08545996856689453, 0.08539647674560547, 0.08540057373046875, 0.08492031860351562, 0.08551526641845703, 0.08133529663085938, 0.08056217956542969, 0.08036351776123046, 0.08066969299316407, 0.08025395202636719, 0.08039628601074218, 0.08287641906738281, 0.08132608032226563, 0.08082329559326172, 0.08066867065429688, 0.0822999038696289, 0.08375296020507812, 0.08239103698730468, 0.08026112365722657, 0.08034611511230469, 0.08190873718261718, 0.5161922607421875, 0.08048332977294922, 0.08032358551025391, 0.08309555053710938, 0.08317644500732421, 0.0805580825805664, 0.08210636901855468, 0.08081407928466797, 0.08045260620117188, 0.08023040008544922, 0.08245350646972656, 0.08249139404296875, 0.08060006713867188, 0.08049459075927734, 0.07992115020751953, 0.08049977874755859, 0.08055289459228515, 0.08052326202392578, 0.08077516937255859, 0.08043520355224609, 0.0833034210205078, 0.08294092559814453, 0.08160768127441406, 0.08223337554931641, 0.08221385955810546, 0.08159744262695312, 0.08116838073730469, 0.08049049377441406, 0.08033586883544921, 0.08034918212890625, 0.08060214233398437, 0.08028873443603515, 0.08058675384521484, 0.08057036590576172, 0.08030413055419922, 0.08308633422851562, 0.08312934112548828, 0.08227123260498047, 0.08059187316894531, 0.08236550140380859, 0.08107414245605468, 0.08041264343261718, 0.08030416107177735, 0.08019350433349609, 0.0805038070678711, 0.08295935821533203, 0.08059903717041016, 0.0819609603881836, 0.08271769714355469, 0.08190156555175782, 0.08018841552734375, 0.08054988861083984, 0.08026624298095703, 0.08369356536865234, 0.08052531433105468, 0.08051507568359376, 0.08030719757080078, 0.08242793273925782, 0.08046998596191406, 0.08084377288818359, 0.08229580688476562, 0.08067276763916016, 0.08043110656738281, 0.5164246826171875, 0.08162815856933593, 0.08057965087890626, 0.08042387390136718, 0.08143666839599609, 0.08144281768798828, 0.08223232269287109, 0.08032972717285156, 0.0812779541015625, 0.08238285064697265, 0.08033280181884765, 0.08039218902587891, 0.0803594207763672, 0.08050176239013672, 0.0800552978515625, 0.08041165161132813, 0.080468994140625, 0.08084172821044922, 0.08128614044189453, 0.08042393493652343, 0.08206028747558594, 0.08061952209472656, 0.08051507568359376, 0.0803420181274414, 0.08279654693603515, 0.08056729888916016, 0.08061548614501954, 0.080133056640625, 0.08056626892089844, 0.08235724639892578, 0.08163430023193359, 0.08038706970214844, 0.08184012603759766, 0.08037785339355469, 0.0805027847290039, 0.08062770843505859, 0.0814172134399414, 0.08300748443603516, 0.08400902557373047, 0.08350918579101563, 0.08451481628417969, 0.0817254409790039, 0.08179814147949219, 0.08327993774414062, 0.08242066955566406, 0.08332598114013672, 0.08047100830078124, 0.08046387481689453, 0.08042803192138671, 0.08188108825683593, 0.08300032043457031, 0.08036966705322265, 0.08046489715576172, 0.08080178833007813, 0.08087039947509765, 0.08029388427734375, 0.08035430145263672, 0.08065331268310547, 0.08069631958007813, 0.08063385772705078, 0.08074342346191406, 0.08058367919921874, 0.08041881561279297, 0.5169356689453125, 0.080574462890625, 0.08031948852539063, 0.08043929290771484, 0.08171826934814454, 0.0848189468383789, 0.08166194915771484, 0.08131072235107421, 0.08058777618408203, 0.08282316589355469, 0.08184524536132813, 0.08173056030273437, 0.08203263854980469, 0.08320511627197266, 0.08302694702148437, 0.08298188781738282, 0.08223846435546875, 0.08015769958496094, 0.08293170928955078, 0.08339968109130859, 0.08318873596191406, 0.08285183715820313, 0.08314470672607421, 0.0827658233642578, 0.08057651519775391, 0.08049561309814453, 0.08271158599853516, 0.08238076782226562, 0.08045980834960938, 0.08164553833007812, 0.08335871887207032, 0.08263782501220703, 0.08058477020263671, 0.08046176147460937, 0.08078028869628906, 0.08137318420410156, 0.08054271697998047, 0.08029901123046874, 0.08063180541992188, 0.08216371154785156, 0.08048230743408204, 0.0804874267578125, 0.08042700958251953, 0.08028467559814453, 0.08053555297851563, 0.08054886627197265, 0.08183500671386719, 0.08184320068359376, 0.0805406723022461, 0.08033177947998046, 0.08072908782958985, 0.08007577514648437, 0.08035020446777344, 0.08032051086425782, 0.08077721405029296, 0.08069427490234375, 0.08040857696533203, 0.07999078369140625, 0.08025190734863281, 0.08065843200683594, 0.08293888092041016, 0.08307711791992188, 0.08205107116699219]",tokens/s,11.164341251974012,,,main,False,False,True @@ -9690,7 +9690,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6677760e-5bd78e692cc14150065caa69;041df985-37ef-4bf8-9917-2c604dec154e) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbbca-206b1f2f515d8d0f58673a1e;e0894f48-b08d-44cb-922a-fd1d422cdd7f) Repository Not Found for url: https://huggingface.co/m/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -9762,7 +9762,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667779f5-0a2008bc24d408843fe5394c;2ad7b2c0-5b78-45a1-bc5a-010e1d0f42b1) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbfaf-063edf285843a9d6791e3ae1;e90a7e86-d2ac-4f81-b36d-2d5f842855cc) Repository Not Found for url: https://huggingface.co/v/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -9873,7 +9873,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66777a20-76441506109344d25a51a8c4;c3f24e00-81b8-4bf4-a3c5-b3120c174d7a) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbfdd-2a8743ac29789e2e24d884f3;7be4e013-b8dd-46e2-bb59-38d726d3b8e0) Repository Not Found for url: https://huggingface.co/0/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -9952,7 +9952,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66777969-1054ac2f235ce61a55ebb342;81f11f49-c984-4d59-a089-bf681433ca39) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbf2f-1d12c33113ed929c44272f70;ae6844e4-574f-4400-adf8-2096af0e0187) Repository Not Found for url: https://huggingface.co/2/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -10051,7 +10051,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66777898-326ed25722a086321bc8bd77;ef1d32e1-62f2-4a32-b133-756b1304dbf6) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbe68-1b3320af5d0b714e793aa135;b4673cfb-167d-4a97-8702-3f679e290642) Repository Not Found for url: https://huggingface.co/l/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -10148,7 +10148,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6677765f-74cb5a317d8ccfd442afded1;c67f1e1b-daf1-4334-9f93-abd02e460331) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbc2b-2a99ed9e4826c9b11c5bd29e;fc34d879-b3a5-4ac5-aad3-11ff4ba3f272) Repository Not Found for url: https://huggingface.co/s/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -10218,7 +10218,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66777995-3f7576bc651edc573bfd57f2;03c6b011-f165-48ba-8237-790b779817e7) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbf5a-3307b8405aa580a64719759d;4e43fccb-0bc1-4929-af03-5b74797bc5d7) Repository Not Found for url: https://huggingface.co/B/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -10348,7 +10348,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66777870-42d8f97a42153d284036299c;c4e868b9-f666-40d9-8e17-5f15b0f92266) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbe41-62b7b9ff4d1c6f0a5d457663;ec6d79f8-20b2-4e4d-a928-cd455b29042b) Repository Not Found for url: https://huggingface.co/a/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -10476,7 +10476,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66777a6e-41bfa99f2e2aec2808cb395c;f35d02e5-4b5f-4a14-8479-edff7236b3af) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cc032-29a452b04103547a451a7b8a;bdc803cc-0014-467f-bb90-a4ab2f5de0c0) Repository Not Found for url: https://huggingface.co/1/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -10575,7 +10575,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6677781e-141a04dd1720d8bb40db5edf;9f5ffb0f-cf2d-45ff-bfed-f41f427af974) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbdf0-7de6b3f36674fab13e9d3de9;1b8c9922-f4fa-4a41-88f3-7c36f9bea179) Repository Not Found for url: https://huggingface.co/t/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -10693,7 +10693,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667a1475-18ffd84467cf43317125a57b;29e873f8-643e-4ccb-acd7-e588bc21af16) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667cb6a7-3cb43d9f17d5a8e4043c99c4;e161569f-c5e9-4921-a105-fd31a0a69876) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/recurrentgemma-2b/resolve/main/config.json. @@ -10856,7 +10856,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667a14a0-497418973ae7848a2eaf9ad2;e9c0db56-7055-491c-9459-5ca0a67c31e1) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cb6d0-39b177434c36ee2615d878a6;80481869-443b-45e1-b4b8-c6a189bf7297) Repository Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -10917,7 +10917,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667a141c-3f58184b3a2a11f6226d5c2c;b2ec6eec-00bd-43d1-b81e-70e76e68fcf0) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667cb655-432997295c1bceec391f33ca;10122fcb-27ff-4ac1-b89f-a8ca3b3296ac) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-2b/resolve/main/config.json. @@ -11121,7 +11121,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667a1448-58ebdded098043b83cc679be;19f17ad1-4d92-48f0-8b3a-a99310e580a8) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667cb680-23155a052b6e7ece1981c568;ae35891c-1ef1-4a4f-970b-4542de5c80e3) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-7b/resolve/main/config.json. @@ -11206,7 +11206,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667777cf-784c3f3111fee1b3273bfec8;705c2eac-a441-4bd4-b742-57c672d9e520) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbd97-098442804b4841445ddbfeb4;d8fe99e0-7b91-4d5c-a526-96c2cf341764) Repository Not Found for url: https://huggingface.co/i/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -11343,7 +11343,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667777a7-6337412502d53375033fd232;02190067-fde5-4e9f-a1c3-5ff9913d67e2) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbd6f-2e8a70e93c1ab0a632bb10ae;e7d4e242-cd72-4283-a19e-ac24f07c306c) Repository Not Found for url: https://huggingface.co/M/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -11621,7 +11621,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66777914-279367cf336c0d4a3091b885;c6fd923e-b1a3-4fa5-882f-15d6dac2e967) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbedf-4925d0674887280d6fa6018b;d9fd2361-2308-4e72-bc32-eb2e4b2afcd1) Repository Not Found for url: https://huggingface.co/x/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -11712,7 +11712,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66777b7f-6c40db516857205f054fed6a;d199b3bd-4058-481a-aa23-2c3a5c95cf96) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667cc148-06d7faed3cbfa33b5d954477;031fb9e8-3bf1-49e6-af61-a6378d0e9d2c) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json. @@ -11885,7 +11885,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667778e8-584a375068a79bf740d0b3be;4e682836-c4c9-42d9-946d-589e1fe0c358) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbeb7-6ce446ea048504727c508506;61d17381-9bea-4a87-953d-268532ced21f) Repository Not Found for url: https://huggingface.co/8/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -11974,7 +11974,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667a11e0-560eb506098661ba79a9fba9;802e6971-5407-4421-a4ab-5781759096ed) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667cb423-62ed65fd4b1da3bd438f27a5;ec23739c-ceea-4167-8c5a-18f4cf5956d8) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/databricks/dbrx-base/resolve/main/config.json. @@ -12165,7 +12165,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66777848-116deadb0df7801a62ffdd9b;5197d0c2-fc01-4f06-aff5-ccc0abb54b48) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbe17-1369f758387aade40cda21fb;56a531b9-dea7-4857-9a4e-21e942a6dcf8) Repository Not Found for url: https://huggingface.co/r/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -12445,7 +12445,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667775f9-4be5600e11468f930f517ef4;a88f7a91-5685-41c6-87ad-96a4ac6868a7) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbbb4-54c551775ab9076d3fafbbe4;60e3c45b-6a66-46f8-998a-e209290252bb) Repository Not Found for url: https://huggingface.co/m/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -12517,7 +12517,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667779df-130ccd1d0062a1950c2c96cf;30e4db5e-3ab9-4358-88a4-c500b5dbd722) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbf9a-1904dd3d6eb19e50099ee27f;25119bcf-57b1-4671-a13f-590c11a2d05d) Repository Not Found for url: https://huggingface.co/v/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -12633,7 +12633,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66777a0b-281fd0c2219727346d232b95;ed4cc185-70ee-4605-a586-3871ea7233c9) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbfc4-04f52ff6512fe28c0af14833;450155b3-f6ce-4a2f-9e52-793b95623e5e) Repository Not Found for url: https://huggingface.co/0/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -12744,7 +12744,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66777953-5395a415228cd326253ebace;cb2bbd69-4c84-4637-aecd-a8160b91ff46) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbf1b-440fe7762058c84e34d32022;742e56a7-9532-4841-b821-01a7c4b705c6) Repository Not Found for url: https://huggingface.co/2/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -12815,7 +12815,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66777884-03022a466385b3f564eb1af4;a7524f23-9a3d-4661-96b0-0f3b47c3d6bd) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbe55-1e33c5943e50c505429f6d26;dacec756-b622-4458-bb14-259431b3671d) Repository Not Found for url: https://huggingface.co/l/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -12968,7 +12968,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6677764a-13e04dcd1ae0e373608e1c20;59335f51-dcfb-4a24-a9c0-98e2658087fa) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbc13-70e3dcf56f0485e55f3c12b9;40e8b81c-9d88-47b8-b4b6-8b905c326793) Repository Not Found for url: https://huggingface.co/s/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -13038,7 +13038,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6677797f-21f5df441907e206246fe922;2b10423a-4256-4f3a-a4ce-bff11c96d4bb) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbf45-4e90cacd4c620f56226a2586;7704e568-4f43-4e26-8fd7-da8f38d5f576) Repository Not Found for url: https://huggingface.co/B/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -13112,7 +13112,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6677785c-3007e8750a4fbbe061db70f0;ad6caefa-63af-46c3-b25c-8b24161cae47) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbe2b-2cb75a775e6f7ecf1350503b;cb1cb0e2-c49e-4efc-b36d-0fad6fc5325a) Repository Not Found for url: https://huggingface.co/a/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -13256,7 +13256,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66777a5b-0204701c557bb1187a1e2928;c46df4d3-eb68-4628-8b8c-70c12cd45702) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cc01e-05220bf4592d99d178ff7585;2bcc85aa-cc5b-40ac-ab81-47a271307524) Repository Not Found for url: https://huggingface.co/1/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -13327,7 +13327,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6677780a-49af3d8a47ca5c344c595dc6;ea345797-5a1e-47a9-b756-d0a1e541af78) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbddc-39977e5534bf7d583805c0ee;0e43264d-54bc-4eef-b686-6c2299dac9e5) Repository Not Found for url: https://huggingface.co/t/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -13417,7 +13417,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667a145e-6e0895e16a95a5b86c691146;45713311-6360-46eb-a5c8-113be785f60a) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667cb693-4ffc052e42df88ba28480736;d8d36e67-9acc-4c4b-bfaa-e8a823030a12) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/recurrentgemma-2b/resolve/main/config.json. @@ -13584,7 +13584,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667a148c-17d78e836c69adb17f340647;a6738cb0-6eb0-4a12-82bf-33614206c1f6) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cb6bb-4af5af333c26b8e03737ff02;2f1b4d9d-79e5-478e-9f83-0f3e71749659) Repository Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -13645,7 +13645,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667a1402-6578d57a437420da73de70f3;605aa99c-5115-4acb-b8bd-e39a69ddfb2c) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667cb641-22297db77fc501273959c0fd;df3471f4-49dd-4395-8321-13f5af9b6eff) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-2b/resolve/main/config.json. @@ -13823,7 +13823,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667a1433-660599cc454b416e16345607;140a84dc-93f4-4a1e-b024-d76f74de4cc7) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667cb66b-43e8c04866b611d345a2330a;7755e545-7732-454e-a8fb-e86c434329a9) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-7b/resolve/main/config.json. @@ -13908,7 +13908,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667777bb-6db2c2660cfef40c5b391f6c;ff6622c2-e7b0-4a56-8d8b-dcf72c6913da) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbd83-5e21f4541ac9524216f8a4f2;c4dbf62d-4661-47bc-bc05-e373991d14f6) Repository Not Found for url: https://huggingface.co/i/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -13981,7 +13981,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66777793-79c10f5d261088ca1bc7954f;1c957696-dc49-460d-b63e-b85d4e1d3996) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbd5b-650e08ef1ee6cbe96e72e259;f1f172cb-4bdb-4830-a64b-5aec91b4e220) Repository Not Found for url: https://huggingface.co/M/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -14059,7 +14059,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667778fd-7c5b78ac5172600767b51bad;c59417ca-9b78-497d-9a03-3e1f2db4edb6) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbecb-66ea6c2a05365c2a22c9934d;b035ed03-a419-4180-84f8-eb4e760e0c9e) Repository Not Found for url: https://huggingface.co/x/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -14160,7 +14160,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66777b6a-2b57974c6dee18da07d58f64;aa540fe2-347c-4a7a-a2b8-3dd195365240) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667cc134-76c6cd093d1b945444fd6135;f1209b2d-f4fa-46dc-a3aa-a6a850b6270f) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json. @@ -14250,7 +14250,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667778d3-7c53e0ab41928a0f49f99c4b;3ea49d3a-e2e8-482d-a11e-e4907bd8e40b) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbea4-40f289965d3522135e95da0b;8fd80350-a867-428f-878d-5d164eee5e1c) Repository Not Found for url: https://huggingface.co/8/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -14339,7 +14339,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667a11cb-30f88d2752cba3822f98d14f;09265190-11c9-4daf-a54d-f959f5356f85) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667cb406-40b0ea9740cfce90126a33bd;6cbb945a-b4d8-490c-8476-060866ab1466) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/databricks/dbrx-base/resolve/main/config.json. @@ -14470,7 +14470,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66777832-6ecd02ae40501fbe14fefe64;d513be79-6cd5-4fa1-b7f1-eaf493870ee7) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667cbe04-53206457091f62731959732b;71dc136e-26e6-4d56-a4a9-4c37c1334f92) Repository Not Found for url: https://huggingface.co/r/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`.