diff --git "a/perf-df-unquantized-1xA10.csv" "b/perf-df-unquantized-1xA10.csv" --- "a/perf-df-unquantized-1xA10.csv" +++ "b/perf-df-unquantized-1xA10.csv" @@ -151,7 +151,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66931ded-504dd3c136e5df0f1d0680f0;fe0d88e1-fb51-49cb-9ef6-f6be639c07a0) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6694700a-788e108f6c2da2ac63fc6563;fd1c63e7-e1c8-42d6-b86a-28c921d41a90) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/recurrentgemma-2b/resolve/main/config.json. @@ -420,7 +420,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66931dba-14f0806f425c15bc35b7dcca;5e8197cc-dc25-4d75-b6f4-46b04fa13c91) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66946fcf-15b76fcc6ea47d993dc3c74a;73a695c0-9fb7-483f-9439-f3b28a847fc5) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-7b/resolve/main/config.json. @@ -505,7 +505,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6693239f-26d87c9358dd294f7d0d3651;6a670ea7-5873-415f-91af-848b78d31922) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6694748b-31bbd64524e54ab57045f92d;75ceb8d1-167e-4921-bef0-545c959d72a9) Repository Not Found for url: https://huggingface.co/a/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -946,7 +946,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669323d2-790fa4a6159804977c60917a;b495935d-e181-4cef-bdab-f99d306a43c4) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66947455-584228bf50d38f264d1608e5;64741d5e-945a-417c-b062-5d1bd8c01159) Repository Not Found for url: https://huggingface.co/l/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -1016,7 +1016,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669322d0-1ccd5c7c411dfbb208e39d7f;a2adf87e-7386-4912-9052-3b048023c6fa) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669474c0-47815cb04b15ea3c0e3d0c0b;0eb05746-6b32-43b7-b397-5b6f8c4eb946) Repository Not Found for url: https://huggingface.co/i/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -1087,7 +1087,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66931e20-2dbc6fe636162125612ed57f;3a2edbf6-509b-4a8c-b4be-f062f0ea4951) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66947042-458157e14426494755b52e7a;ab95f370-378f-46fb-ac48-4911724784c0) Repository Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -1184,7 +1184,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66932096-020d59fa64d982236603cdb6;b2eb6557-4361-486a-a0be-2b3975cfc332) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669472e3-1863effa6690288a682bf23f;4faff99d-4442-4995-ba9b-1f62773e0070) Repository Not Found for url: https://huggingface.co/m/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -1619,7 +1619,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66932338-646d13785d0af4bb421c9f3a;3ecad2f5-9ffc-422b-be21-1074914697b6) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669473a6-2ca4942377d64462366409e8;cb17bce6-021f-4019-babe-483053c85f61) Repository Not Found for url: https://huggingface.co/t/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -1989,7 +1989,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6693236c-6e7f178b07be29c103d77b63;22c39c0c-ac75-469d-a47b-b5b71da15693) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669473e3-33adb42f7affe0995bf78662;77caad29-b50e-4056-911c-3ab5fbe58e49) Repository Not Found for url: https://huggingface.co/r/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -2181,7 +2181,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66931d86-2b960d4300a5e5b07250a36c;07b9d637-6d5f-4df5-9758-6ca526842fb6) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66946f94-17f2584c0e69a45a728d1c29;3a63f7d7-0859-4fa0-ae06-a04f5e250b83) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-2b/resolve/main/config.json. @@ -2638,7 +2638,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669320fc-41e525eb7a0a2b7d610605e8;a391a967-0abc-49c2-9aca-b655daf90491) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66947363-08e5e89f68e1b8012d1b76bd;662573c1-9152-46c5-a4d9-5143f0981568) Repository Not Found for url: https://huggingface.co/s/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -2698,7 +2698,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66931bc9-6c25be3f6c661bb82a0d31c3;b30ffd90-ee3e-4eff-84ee-df4de65c298e) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66946da3-1a365249711d95f80bca86b0;d7e5d613-c82b-4a20-a410-f7f89bcee801) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/databricks/dbrx-base/resolve/main/config.json. @@ -3079,7 +3079,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66931dd3-58f19bf1044861914e5e58d6;340e844f-0a4f-46e5-9b3e-222bb10054c1) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66946fed-7fa8aae9235c8d004a54d174;8b1b7557-6412-4537-b382-841d3ab2912a) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/recurrentgemma-2b/resolve/main/config.json. @@ -3416,7 +3416,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66931da0-28376ac2561cadd279900bc8;b9224984-295b-4e4a-8125-285de07972c8) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66946fb1-260892c700b4aa0d3ff5bb63;2d784bb5-4593-4084-983a-96705289117c) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-7b/resolve/main/config.json. @@ -3501,7 +3501,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66932385-3f2c50ae4c24bb105f1725be;12fd1659-2328-4ac4-9fdf-4ecf0a68e5ca) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66947470-4fb3414119f7c5871a2b7787;20e4f0a6-29a7-4b33-9bee-9b2268573547) Repository Not Found for url: https://huggingface.co/a/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -4042,7 +4042,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669323b9-14434f161cdf6441423b27b1;f3c5007a-4df3-4a7e-9f97-70421d6fe20b) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6694743a-0bc35fda54c4d66a36e78186;6440c2b8-0774-4472-a7ba-faa47f03beae) Repository Not Found for url: https://huggingface.co/l/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -4112,7 +4112,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669322b6-48a12a5e5a7400583beb3f69;bb98a15b-44e2-4bac-9373-b4fa6ab78f2a) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669474a6-738578c128f9da5b7eb7db0b;32cc0dc6-06ee-4ed9-aba2-9c4a58dac5cb) Repository Not Found for url: https://huggingface.co/i/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -4212,7 +4212,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66931e07-7366784a0a52438751625647;ef5561d8-9247-42f5-85f9-ff9daae5c010) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66947027-61f3eed65ebd874e31134d16;89bd6c95-d9f8-4e68-ab55-5a0e7c96ded9) Repository Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -4283,7 +4283,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6693207c-29a71aeb3b58b7c06d43a738;7bac3392-e156-41f3-b337-785ed28ddc9c) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669472c6-2b77031e786b9f78671a3548;252aa2e9-2958-42e9-b788-bff74573d83f) Repository Not Found for url: https://huggingface.co/m/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -4720,7 +4720,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6693231d-2554abf13627e3c4588b22cd;38533e08-9c1d-405f-b265-ecddf9f791ef) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66947383-60fe704b1dc0830f382c28c2;e9956a5c-60a2-4d48-838d-3e80c4033e7b) Repository Not Found for url: https://huggingface.co/t/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -5084,7 +5084,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66932352-2f53c37533e7d0b635f84a68;39c52563-ad8c-41db-8579-3a1292df4e81) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669473c4-470cf07d6836099f2ae90965;34e8fe25-e865-4606-bc46-314e988603d6) Repository Not Found for url: https://huggingface.co/r/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -5295,7 +5295,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66931d6c-4f8fbb8148ad37a019f48a67;45d49c73-d0f1-4072-8142-68924e2b022c) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66946f77-1339df357c176fb738e1a3ee;fba5fd68-579b-4733-9efb-c6db3de9637d) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-2b/resolve/main/config.json. @@ -5828,7 +5828,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669320e2-129d3d954edeadb179a1058e;b4fe790c-3001-47e8-9400-2efe59145c23) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66947346-6367fb272e7841bf3d3b39bb;ff678302-bfd1-4828-99da-81b9bf86e423) Repository Not Found for url: https://huggingface.co/s/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -5888,7 +5888,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66931baf-5f293ac7304976db2ff0ccbe;68796226-cef9-41ae-866d-aee9033b4b41) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66946d88-255ca90a27a63bf44dc6569d;2e447ced-5d52-416e-a92f-90a7f6927efc) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/databricks/dbrx-base/resolve/main/config.json. @@ -6270,7 +6270,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66931de7-564bcd1056669c451e9144f8;19f43a92-6a8c-4d98-be2f-5119fef11ab7) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66947003-0ddbc3310bedeec571198da3;3e01d067-17d3-4807-854a-aaaeea7dd82d) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/recurrentgemma-2b/resolve/main/config.json. @@ -6611,7 +6611,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66931db3-22d619af0caffc3a7e5fac3f;7b0a306e-de98-424b-b8a8-ed7d5d3da7da) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66946fc8-3b18dfb720fc6aba1ca03916;31af7f7d-50ce-4dbb-9639-732dab195573) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-7b/resolve/main/config.json. @@ -6696,7 +6696,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66932398-74125d6624dcb0e33f977228;5f5d8d4e-b9e4-4275-bbcb-9ad3fad040b0) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66947484-653353db0d98d04951dd1a65;9f97ec32-b6a0-4739-9cde-33681cb7411f) Repository Not Found for url: https://huggingface.co/a/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -7209,7 +7209,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669323cc-69bcb0b825b61a5c10d88590;b17634c3-dc03-4140-9391-a6184e6d9084) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6694744e-712a93c4783b51d04ae3344d;d5c632eb-8c05-40f0-b2f5-49e14531a52f) Repository Not Found for url: https://huggingface.co/l/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -7279,7 +7279,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669322c9-50d80e886a39ac2e47ff6d2b;9fbe2baa-eb14-44a9-a331-e01f577e1b5d) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669474ba-41ace83e737fa42b3701f939;8a02723a-b3e0-4fbe-935e-7171972b67d6) Repository Not Found for url: https://huggingface.co/i/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -7378,7 +7378,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66931e1a-7a812deb659831f545bfc2f4;389ed2db-7e06-461e-8332-a35e64dd4b5b) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6694703b-12263a711dc1333c03b211f2;90ddfe47-8e92-4ecc-8500-1ad11442ab70) Repository Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -7475,7 +7475,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6693208f-20b4f2e173c15ab55de439d1;1b165986-4b60-47ac-a424-1fbd523a8f01) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669472da-18aa21bc6b2b69bc663ff33e;4d074193-0d32-4dd1-998d-2c6235f4b10d) Repository Not Found for url: https://huggingface.co/m/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -7910,7 +7910,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66932331-79685a76601d1a5040dd391e;07e00d8f-5b46-40a0-a44a-e043f2ae6a40) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6694739f-7626b30c1b22fa1d2e4e8109;f9a19fca-4468-40ac-91d9-363e5d71246c) Repository Not Found for url: https://huggingface.co/t/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -8278,7 +8278,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66932366-1d89012a6a7314bc710c93a7;357cdae1-2a08-4728-a748-a8057b6ef8e1) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669473d9-6dd19c917c34caca28c99768;a6028978-f4ab-42fe-ac38-5e0457847fff) Repository Not Found for url: https://huggingface.co/r/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -8506,7 +8506,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66931d80-413969fa7d582d1d64af44b7;9272b443-abe6-4512-9259-8159f189035b) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66946f8c-5353a2262475f65665dbc9e8;3b2b81fb-3f0c-4497-b9ae-8a642535ef48) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-2b/resolve/main/config.json. @@ -9073,7 +9073,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669320f6-63023c7d713589d43ce795b5;a07e408a-0656-4f2d-a609-e675f5fd1f7a) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6694735c-34752f973fd1e62729838657;8cd1b69a-5f46-4541-a0cf-0a08b36db210) Repository Not Found for url: https://huggingface.co/s/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -9133,7 +9133,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66931bc3-13d40b1420e4309e649fe490;db4c8a28-8109-4e1f-bfc1-cdb8ffbdd6cf) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66946d9c-6186a8070b9b77472e8adf90;fbb62cfb-55df-45a1-b4fb-4795ec27dae5) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/databricks/dbrx-base/resolve/main/config.json. @@ -9478,7 +9478,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66931dda-6ca00ede4bd11555166bd9d3;995466b8-a44c-4e53-a3f5-91bf0f50d7f1) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66946ff5-5496777e0c627d34710fe173;d7b3ecc2-e47f-4faa-a257-6a3479f63558) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/recurrentgemma-2b/resolve/main/config.json. @@ -9669,7 +9669,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66931da7-343730ba4fa69c175f2a39bd;a1d758de-a45e-4e5c-af4e-7df778d2856c) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66946fb8-2275110a729971ad60f8fa90;7227e037-592b-4d4a-b08a-4721108ecec0) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-7b/resolve/main/config.json. @@ -9754,7 +9754,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6693238c-0d5b5a10172f8980288ae4a4;5f6733b8-d3a1-459d-b033-529709a2bc34) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66947477-5fb528cb577bb69b158aa93e;28cdf4a1-8e4b-4f04-b3a2-4b548c956072) Repository Not Found for url: https://huggingface.co/a/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -10188,7 +10188,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669323bf-2fe8388b353b510752e41c10;d39aebaf-9252-4e4f-bfcd-9aa007055afb) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66947441-700c2fa25d9a47d060d6c891;67fb0dff-fbf7-4f10-b011-a8c17af2499c) Repository Not Found for url: https://huggingface.co/l/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -10258,7 +10258,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669322bc-7d7d6b70276d2e8e4d505b63;b9c99b09-40fe-496f-9363-c6c5df8f45b4) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669474ad-2664787d2fdd5bcc60a24be0;bf576584-f03a-4bf0-a2ee-f49d9441bbd7) Repository Not Found for url: https://huggingface.co/i/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -10330,7 +10330,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66931e0d-2266a4ae22101758226cfa4c;4394806d-3de8-4254-afd5-e893f72e6eca) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6694702e-45e94dfb0d2424ba19619b8f;6d73ef94-1ef0-486a-ba48-b597765df1b9) Repository Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -10401,7 +10401,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66932082-6d3668ac22dc4371364eea12;349b5b29-0885-4647-ab3a-e443c683ae34) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669472cd-6b7deacc2cc3bb367942f8aa;caf3e1fe-dcba-46d1-8b33-c6e5baf854a7) Repository Not Found for url: https://huggingface.co/m/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -10838,7 +10838,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66932324-4bc17d7534ef3e7e354f041f;8be7daa4-71ac-485a-a62f-7a0fae8771ab) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6694738f-1ba9805e2122e4be3754abe1;36f6c1f2-5559-41d0-8e4d-3f0d28d3b50a) Repository Not Found for url: https://huggingface.co/t/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -11204,7 +11204,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66932359-76202bff7b650e4239c73a49;7060e145-ef05-40d4-b2f4-a7ec5b123830) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669473cb-28a2688f5f24ed381eeb55e5;a69867d9-f9ac-45b0-8f60-bb1458245830) Repository Not Found for url: https://huggingface.co/r/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -11345,7 +11345,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66931d73-3a87844177c45ef50009983a;341b36b5-4cab-4a5e-823c-774d4a84a09d) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66946f7d-0593968f648c994038747d06;0d0ca0ee-467e-43cb-974c-70ac7c4b1b95) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-2b/resolve/main/config.json. @@ -11771,7 +11771,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669320e9-6a1227392fb48c9244831c68;969ec0ed-066b-47af-b9c1-07b1e40c00b8) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6694734e-239c6f3a377a159a27825dff;f03a233b-84fe-4c0d-88a5-1579097ed82b) Repository Not Found for url: https://huggingface.co/s/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -11831,7 +11831,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66931bb6-2ce23ea33770d3e529c94082;8e485786-79d3-42c4-958d-c69240955144) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66946d8f-1a3926c6183677a319d34e4d;189315c0-7507-4604-b1a0-0bf55eec2324) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/databricks/dbrx-base/resolve/main/config.json. @@ -12177,7 +12177,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66931dfa-634507e476d8ea9730b6e78a;a3746df2-ca91-453a-8147-42e5dcad7528) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6694701a-2453f8e502e19a2d6dcc7928;0f4daa73-67e4-42c1-8b5e-e6c08ad152b6) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/recurrentgemma-2b/resolve/main/config.json. @@ -12394,7 +12394,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66931dc7-5a2b43f43514809915cac473;0408fae8-994f-47e3-8354-acd987158cbe) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66946fde-051bef132ab73fc34e4a610f;c0a865e0-9f4b-4e7a-83a3-e9f96e1935c8) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-7b/resolve/main/config.json. @@ -12479,7 +12479,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669323ac-55f23d122e5ea1927e313dac;64236865-d8e0-41d2-8f38-459a18cb7b53) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66947498-58a4713947ab72d54b3de1b8;267061e9-cc25-442b-aa61-3d3134f8a64d) Repository Not Found for url: https://huggingface.co/a/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -12754,7 +12754,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1572, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: CodeGenForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpo6jh9mxe/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: CodeGenForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp9z0qyri6/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,mistralai/Mistral-7B-v0.1,mistralai/Mistral-7B-v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.1,,,MB,1270.10816,15987.113984,0.0,15340.66688,15114.865664,s,10,0.8908641891479493,0.08908641891479493,0.002248653393122016,0.08837084579467774,0.08936671295166015,0.09259126205444335,0.09517090133666993,"[0.09581581115722657, 0.08845929718017578, 0.08830262756347657, 0.08810169219970704, 0.088650146484375, 0.0884567642211914, 0.08838969421386719, 0.08835199737548828, 0.08812274932861328, 0.08821340942382813]",tokens/s,2873.614217727693,kWh,1.0468745888041418e-06,5.736383472775847e-07,6.696745770371453e-06,8.31725870645318e-06,tokens/kWh,30779372.030519523,MB,1270.10816,15987.113984,0.0,15340.66688,15134.114304,s,10,31.351924072265625,3.1351924072265627,0.01343642581392157,3.135545288085938,3.1540085205078126,3.1547807495117186,3.1553985327148437,"[3.155552978515625, 3.132075439453125, 3.1538369140625, 3.13901513671875, 3.146458984375, 3.139015869140625, 3.125260986328125, 3.11593408203125, 3.116068603515625, 3.128705078125]",tokens/s,20.09446050417388,kWh,3.680663284386857e-05,2.0171711897392094e-05,0.00013959245459422792,0.00019657079933548858,tokens/kWh,320495.21196928906,,s,629,32.095909866332995,0.05102688373025915,0.010553122864185334,0.04955648040771484,0.05033717651367187,0.050655847167968754,0.13825425720214846,"[0.05150822448730469, 0.05064396667480469, 0.05129523086547851, 0.05078326416015625, 0.05084156799316406, 0.05203763198852539, 0.051097599029541016, 0.05099622344970703, 0.05120716857910156, 0.05065523147583008, 0.050206718444824217, 0.05030400085449219, 0.04976435089111328, 0.049514495849609375, 0.049509376525878904, 0.051076095581054685, 0.05063884735107422, 0.05023641586303711, 0.05009100723266602, 0.050151424407958986, 0.050012161254882816, 0.05011663818359375, 0.04939878463745117, 0.04968854522705078, 0.04967833709716797, 0.049827838897705076, 0.05010227203369141, 0.049858558654785154, 0.05008486557006836, 0.05005926513671875, 0.04980633544921875, 0.04946432113647461, 0.04954111862182617, 0.049481727600097655, 0.04949913787841797, 0.04947558212280274, 0.04950630569458008, 0.049498111724853515, 0.049516544342041016, 0.04955340957641602, 0.04962611389160156, 0.049509376525878904, 0.04947558212280274, 0.04943462371826172, 0.04942540740966797, 0.04960870361328125, 0.04969472122192383, 0.050229248046875, 0.05003366470336914, 0.05021798324584961, 0.050135040283203126, 0.050260990142822266, 0.050267135620117184, 0.05017910385131836, 0.05020975875854492, 0.050331649780273435, 0.05035212707519531, 0.05030092620849609, 0.04970700836181641, 0.04964556884765625, 0.04923494338989258, 0.04957183837890625, 0.13839155578613282, 0.04947251129150391, 0.04955648040771484, 0.04964044952392578, 0.050106433868408205, 0.050107326507568356, 0.05006748962402344, 0.049955806732177734, 0.04954214477539062, 0.04988313674926758, 0.05001728057861328, 0.04995993423461914, 0.05014220809936523, 0.05013401412963867, 0.049616897583007816, 0.04947763061523437, 0.04960768127441406, 0.05022719955444336, 0.04988415908813477, 0.04981248092651367, 0.050051071166992187, 0.049926143646240234, 0.05006335830688476, 0.049353729248046874, 0.04936294555664063, 0.049729534149169925, 0.049554431915283206, 0.04951039886474609, 0.04948787307739258, 0.049462272644042966, 0.04944588851928711, 0.04953804779052735, 0.04948179244995117, 0.049499073028564454, 0.04950527954101563, 0.04995993423461914, 0.05023027038574219, 0.05015039825439453, 0.04978278350830078, 0.050282497406005856, 0.05033369445800781, 0.049585151672363284, 0.0496629753112793, 0.04955750274658203, 0.049522689819335934, 0.04967833709716797, 0.04951039886474609, 0.04962713623046875, 0.04952166366577149, 0.04959231948852539, 0.04946124649047851, 0.04948582458496094, 0.049503231048583986, 0.049539070129394534, 0.04958924865722656, 0.04955852890014648, 0.049509376525878904, 0.049473567962646486, 0.04956361770629883, 0.049535999298095705, 0.04949401473999023, 0.0496445426940918, 0.04965785598754883, 0.13929983520507813, 0.05099520111083984, 0.05012582397460937, 0.05029580688476563, 0.05021084976196289, 0.04981449508666992, 0.05031942367553711, 0.05046572875976563, 0.05136383819580078, 0.050411518096923826, 0.050383872985839843, 0.050495487213134765, 0.0498524169921875, 0.049560577392578124, 0.04931071853637695, 0.04940390396118164, 0.04941823959350586, 0.04949708938598633, 0.04954111862182617, 0.049980415344238284, 0.05037875366210937, 0.05026508712768555, 0.05039718246459961, 0.05020876693725586, 0.05030297470092773, 0.0502026252746582, 0.050229248046875, 0.050078720092773435, 0.05004390335083008, 0.04976025772094726, 0.05017804718017578, 0.05026924896240234, 0.05031315231323242, 0.05025484848022461, 0.05010227203369141, 0.05037875366210937, 0.05018316650390625, 0.05012070465087891, 0.049895423889160156, 0.05019340896606445, 0.04981350326538086, 0.049498111724853515, 0.04962303924560547, 0.04949606323242187, 0.04951347351074219, 0.049584129333496096, 0.04954111862182617, 0.04955750274658203, 0.049563648223876954, 0.0496517105102539, 0.050270206451416014, 0.05033062362670898, 0.05029683303833008, 0.0502026252746582, 0.0500316162109375, 0.050307071685791016, 0.05026201629638672, 0.04968038558959961, 0.05005721664428711, 0.050129920959472656, 0.04984115219116211, 0.05033267211914062, 0.050070526123046875, 0.13817240905761718, 0.05058867263793945, 0.051151905059814456, 0.05071356964111328, 0.05061529541015625, 0.05039820861816406, 0.05039923095703125, 0.05047808074951172, 0.05027635192871094, 0.05030297470092773, 0.050301952362060545, 0.050146305084228515, 0.05001728057861328, 0.05039718246459961, 0.05006950378417969, 0.05029785537719727, 0.050298881530761716, 0.04991078567504883, 0.050282497406005856, 0.0502026252746582, 0.0501288948059082, 0.04972544097900391, 0.0512542724609375, 0.0502108154296875, 0.04959539031982422, 0.049509376525878904, 0.04960665512084961, 0.04942950439453125, 0.04935168075561523, 0.049452030181884765, 0.049486846923828126, 0.04944076919555664, 0.04976537704467773, 0.04984524917602539, 0.04969574356079102, 0.04945619201660156, 0.04941817474365234, 0.04943360137939453, 0.04935475158691406, 0.04957183837890625, 0.04941823959350586, 0.049462272644042966, 0.049500160217285157, 0.04948889541625977, 0.049519615173339845, 0.049509376525878904, 0.04964966583251953, 0.04968243026733398, 0.04969267272949219, 0.04940902328491211, 0.04943462371826172, 0.04943462371826172, 0.04942233657836914, 0.049460224151611325, 0.049363967895507815, 0.04937318420410156, 0.04945407867431641, 0.04951859283447266, 0.04944793701171875, 0.049460224151611325, 0.04945510482788086, 0.04940288162231445, 0.049838081359863284, 0.13938482666015625, 0.05005721664428711, 0.050495487213134765, 0.050092033386230465, 0.04980838394165039, 0.05002342224121094, 0.05002035140991211, 0.050165760040283204, 0.04978278350830078, 0.04995072174072265, 0.05005926513671875, 0.05003776168823242, 0.049748992919921874, 0.049593345642089844, 0.04950531387329102, 0.04947657775878906, 0.049430526733398435, 0.05007257461547852, 0.05000396728515625, 0.050071617126464844, 0.04999980926513672, 0.04993228912353516, 0.05018931198120117, 0.04961996841430664, 0.04951551818847656, 0.04973875045776367, 0.04973161697387695, 0.04977967834472656, 0.05037875366210937, 0.05026201629638672, 0.05004185485839844, 0.050200576782226565, 0.05030297470092773, 0.049873920440673826, 0.05007462310791016, 0.04976128005981445, 0.05007462310791016, 0.049683456420898435, 0.04982271957397461, 0.0502927360534668, 0.049718273162841796, 0.04939059066772461, 0.04944076919555664, 0.04933017730712891, 0.0497520637512207, 0.04955033493041992, 0.05018624114990235, 0.049748992919921874, 0.05006950378417969, 0.05020774459838867, 0.04993228912353516, 0.049979393005371096, 0.050407424926757816, 0.049964031219482424, 0.04987596893310547, 0.04954828643798828, 0.05025689697265625, 0.05018828964233398, 0.04990259170532227, 0.04986777496337891, 0.05033574295043945, 0.049865726470947266, 0.05046169662475586, 0.13854208374023438, 0.05013708877563477, 0.05030809783935547, 0.05033267211914062, 0.050239486694335936, 0.05022719955444336, 0.05025791931152344, 0.05054054260253906, 0.050285568237304686, 0.050388992309570314, 0.05001113510131836, 0.050293758392333986, 0.05029171371459961, 0.05048012924194336, 0.05126860809326172, 0.05061427307128906, 0.04989440155029297, 0.05028044891357422, 0.04963840103149414, 0.05015654373168945, 0.04982988739013672, 0.049616897583007816, 0.049783809661865235, 0.04948582458496094, 0.05065625762939453, 0.05040435028076172, 0.05024051284790039, 0.050310142517089845, 0.05030809783935547, 0.05027328109741211, 0.05018316650390625, 0.050342910766601565, 0.050165760040283204, 0.0504535026550293, 0.05025894546508789, 0.04985139083862305, 0.049384513854980466, 0.0494376335144043, 0.0493568000793457, 0.04944179153442383, 0.049446910858154294, 0.04942748641967774, 0.049387489318847656, 0.04940083312988281, 0.04912025451660156, 0.04925337600708008, 0.049223678588867184, 0.049276927947998046, 0.049165313720703124, 0.049355777740478515, 0.04916428756713867, 0.049255424499511716, 0.049159168243408206, 0.04920217514038086, 0.049099777221679686, 0.04922880172729492, 0.049261566162109374, 0.049363967895507815, 0.04929024124145508, 0.04944076919555664, 0.0493568000793457, 0.04930460739135742, 0.04947452926635742, 0.13843763732910155, 0.049470462799072266, 0.04944076919555664, 0.049476608276367184, 0.049452030181884765, 0.04938751983642578, 0.04941312026977539, 0.04939980697631836, 0.05066649627685547, 0.05098086547851562, 0.05060095977783203, 0.050267135620117184, 0.05009100723266602, 0.049555454254150394, 0.04936703872680664, 0.04934860610961914, 0.04946432113647461, 0.04939980697631836, 0.049463294982910154, 0.049554431915283206, 0.04955750274658203, 0.04935168075561523, 0.0494284782409668, 0.04946636962890625, 0.04936601638793945, 0.049503231048583986, 0.049317886352539066, 0.049459201812744144, 0.04959027099609375, 0.050514942169189454, 0.04999174499511719, 0.049602497100830076, 0.04951551818847656, 0.049549312591552735, 0.04944486236572266, 0.049397792816162106, 0.049336288452148436, 0.049463294982910154, 0.049481727600097655, 0.04997836685180664, 0.05016371154785156, 0.049331199645996096, 0.04944998550415039, 0.04951859283447266, 0.04943360137939453, 0.049517566680908204, 0.049484798431396484, 0.04974796676635742, 0.04963328170776367, 0.04942745590209961, 0.04949401473999023, 0.049438751220703125, 0.04965270233154297, 0.04947251129150391, 0.04961177444458008, 0.04948787307739258, 0.04948992156982422, 0.049426433563232425, 0.04952166366577149, 0.049484798431396484, 0.04940697479248047, 0.049462272644042966, 0.049463294982910154, 0.13833319091796875, 0.049393665313720705, 0.04953702545166016, 0.049364990234375, 0.04948787307739258, 0.049535999298095705, 0.049337345123291014, 0.04940492630004883, 0.04935475158691406, 0.049446910858154294, 0.049465343475341796, 0.04941823959350586, 0.049301502227783206, 0.04938956832885742, 0.049307647705078124, 0.04935065460205078, 0.049430526733398435, 0.04942745590209961, 0.049324031829833984, 0.04948992156982422, 0.04935782241821289, 0.049377281188964846, 0.04944486236572266, 0.049430526733398435, 0.04940492630004883, 0.04945305633544922, 0.04943462371826172, 0.04951551818847656, 0.04935475158691406, 0.04977664184570312, 0.04969369506835938, 0.049484798431396484, 0.04934963226318359, 0.04940806579589844, 0.049469375610351564, 0.05001523208618164, 0.049430526733398435, 0.04937420654296875, 0.04934348678588867, 0.04939263916015625, 0.04958617782592773, 0.04946944046020508, 0.049391616821289064, 0.04938137435913086, 0.049375232696533204, 0.049344512939453126, 0.04950527954101563, 0.049585151672363284, 0.049463294982910154, 0.04938547134399414, 0.04948992156982422, 0.04953497695922852, 0.049648670196533205, 0.04945711898803711, 0.04945305633544922, 0.04946739196777344, 0.04941516876220703, 0.04949401473999023, 0.04941823959350586, 0.04942335891723633, 0.04950732803344727, 0.04944076919555664, 0.049481727600097655, 0.13801370239257812, 0.049347583770751956, 0.04943769454956055, 0.04936908721923828, 0.04947455978393555, 0.04968038558959961, 0.0495206413269043, 0.04937420654296875, 0.04921446228027344, 0.04945612716674805, 0.04941619110107422, 0.04937011337280273, 0.04930252838134765, 0.049442817687988284, 0.04929024124145508, 0.049400863647460935, 0.049390560150146486, 0.049442817687988284, 0.049364990234375, 0.04983603286743164, 0.04946944046020508, 0.05073920059204102, 0.05070848083496094, 0.04989132690429687, 0.049511425018310545, 0.04998758316040039, 0.0494919662475586, 0.049397758483886715, 0.049452030181884765, 0.04941209411621094, 0.04955852890014648, 0.04943155288696289, 0.04941209411621094, 0.049552383422851565, 0.049271808624267575, 0.04940083312988281, 0.04940492630004883, 0.04987395095825195, 0.04926153564453125, 0.04942134475708008, 0.04945612716674805, 0.049363937377929684, 0.049228832244873046, 0.04926358413696289, 0.049160190582275394, 0.04935987091064453, 0.049206272125244144, 0.049239040374755856, 0.049203201293945314, 0.04919705581665039, 0.049102848052978515, 0.049135616302490234, 0.04922470474243164, 0.049353729248046874, 0.04939468765258789, 0.04946124649047851, 0.04940083312988281, 0.0494284782409668, 0.04942335891723633, 0.0494571533203125, 0.049468414306640625, 0.049468414306640625, 0.04942950439453125, 0.13828608703613282, 0.04951551818847656, 0.04944588851928711, 0.04945407867431641, 0.04940902328491211, 0.04939263916015625, 0.04941519927978515, 0.04929840087890625, 0.049401920318603514, 0.04938643264770508, 0.04948992156982422, 0.04942233657836914, 0.0493568000793457, 0.04933324813842774, 0.04933017730712891, 0.04942335891723633, 0.05143040084838867, 0.050685951232910156, 0.05024870300292969, 0.04951555252075195, 0.04936291122436524, 0.049509376525878904, 0.049495040893554686, 0.04945407867431641, 0.0494284782409668, 0.049500160217285157, 0.04971212768554688, 0.050369537353515625, 0.04944588851928711, 0.04949401473999023, 0.04957593536376953, 0.04948992156982422, 0.04967119979858398, 0.04950729751586914, 0.049436702728271484, 0.0494601936340332, 0.05009100723266602, 0.050715648651123046, 0.04955750274658203, 0.049914878845214845, 0.049552383422851565, 0.0499947509765625, 0.049535999298095705, 0.049833984375, 0.049620990753173826, 0.04941516876220703, 0.04950732803344727, 0.04948889541625977, 0.049704959869384766, 0.04957593536376953, 0.049479679107666014, 0.049522689819335934, 0.04950835037231445, 0.04954009628295898, 0.050326526641845705, 0.049444961547851565, 0.049495967864990234, 0.04948793411254883, 0.04950214385986328, 0.05026406478881836, 0.049793022155761715, 0.05005926513671875, 0.049838081359863284]",tokens/s,19.597512661879367,,,,, @@ -12941,7 +12941,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669323df-1a0f6abe72cd8b177c0b419d;b1c11457-cf18-491f-81b1-6c9501f367b8) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66947463-5b622f330cd55e823d157cd1;8333c871-1397-4b11-8f2f-3dd0d8e1335f) Repository Not Found for url: https://huggingface.co/l/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -13011,7 +13011,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669322dd-6a0771e07f854e676922ba47;e64af932-f7ff-41a4-a585-1f164439947c) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669474cd-2309aaf353ce01945d8df075;b31070f1-0b41-4df9-b71b-3a1b815aec8c) Repository Not Found for url: https://huggingface.co/i/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -13083,7 +13083,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66931e2d-2498a334276d60e54997e44c;35bd801b-0093-4f72-93af-8475d01780ad) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6694704f-2baf8c5f7c21869a7a6e8b6b;5daef34a-af20-4316-b5f7-49f6f89eecf3) Repository Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -13138,7 +13138,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1572, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp23za11cj/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmplim5b1yn/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,m,m,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): @@ -13180,7 +13180,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669320a3-638c43c771b1a0d9269a48ad;8b31e4d3-c3d2-4185-a152-99a20ac5b48e) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669472f0-36e202fb7a9b46df3c949191;8f2a5909-b11d-4c1d-9ad4-de6828372d32) Repository Not Found for url: https://huggingface.co/m/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -13619,7 +13619,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66932345-4a51e602330ed1ed0bf3cb90;8162f2da-7398-427e-9a2d-45ca2d258e18) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669473b5-7f93ef0e47e0c47c21f285f3;03ef5cd4-75af-4c13-8d4a-d3a8905c7b6f) Repository Not Found for url: https://huggingface.co/t/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -13985,7 +13985,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66932379-15ae90982db55f4c2cdfb924;c68ca9f7-8e2b-4138-8ed0-eca70425505a) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669473f2-4ba3570e1c0301a66d936d4d;81272ae6-f6a0-494f-8cbf-babca6f900ef) Repository Not Found for url: https://huggingface.co/r/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -14041,7 +14041,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1572, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpjiww47gg/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpc5p2vkl2/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): @@ -14117,7 +14117,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1572, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpf80lgx5q/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp348xovw9/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,togethercomputer/RedPajama-INCITE-Base-3B-v1,togethercomputer/RedPajama-INCITE-Base-3B-v1,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,MB,1290.743808,6833.045504,0.0,6186.5984,6020.355072,s,10,0.3598095359802245,0.03598095359802246,0.000327616359819926,0.03585665512084961,0.03615181999206543,0.03654773464202881,0.03686446636199951,"[0.03694364929199219, 0.03587587356567383, 0.03584486389160156, 0.036063838958740234, 0.035855487823486326, 0.03585529708862305, 0.03582822418212891, 0.03581388854980469, 0.03587059020996094, 0.03585782241821289]",tokens/s,7114.875354889703,kWh,4.261281898172258e-07,2.3349698723394658e-07,2.2525587564819797e-06,2.9121839335331525e-06,tokens/kWh,87906535.38473883,MB,1290.743808,6833.045504,0.0,6186.5984,6150.076416,s,10,14.46971630859375,1.446971630859375,0.025161149540616375,1.4332265014648438,1.4859286865234376,1.488804248046875,1.491104697265625,"[1.4852896728515625, 1.4916798095703125, 1.4388095703125, 1.4768404541015625, 1.426799072265625, 1.4347098388671875, 1.4234737548828125, 1.429978515625, 1.4303924560546875, 1.4317431640625]",tokens/s,43.53920882511255,kWh,1.7033960242017947e-05,9.332253111065811e-06,5.951512954651741e-05,8.588134289960118e-05,tokens/kWh,733570.271178102,,s,629,14.768255001068102,0.023478942767993824,0.004322459082862367,0.022701055526733398,0.02381783103942871,0.023979827880859374,0.05897809753417969,"[0.02389708709716797, 0.023614463806152345, 0.02374143981933594, 0.023562240600585937, 0.023763967514038087, 0.02350284767150879, 0.02431283187866211, 0.024577024459838868, 0.02473472023010254, 0.02489753532409668, 0.024993791580200195, 0.024582143783569335, 0.024782880783081055, 0.024524768829345702, 0.024769535064697267, 0.023834623336791993, 0.024599552154541016, 0.023565311431884766, 0.023340032577514647, 0.023840768814086914, 0.02469068717956543, 0.02433433532714844, 0.023974912643432617, 0.023983104705810547, 0.023916543960571288, 0.023834623336791993, 0.02413670349121094, 0.024687616348266602, 0.024155136108398437, 0.02381520080566406, 0.023843807220458986, 0.023802879333496094, 0.023795711517333985, 0.023799808502197265, 0.022724607467651366, 0.02269593620300293, 0.02265395164489746, 0.022740991592407226, 0.022673408508300782, 0.022766592025756836, 0.02269696044921875, 0.022793216705322264, 0.0226693115234375, 0.022740991592407226, 0.022781951904296875, 0.022732831954956054, 0.02254742431640625, 0.022795263290405272, 0.022791168212890626, 0.02272870445251465, 0.022777856826782225, 0.022759424209594727, 0.023467071533203127, 0.023888864517211915, 0.024092639923095703, 0.022802463531494142, 0.022794240951538085, 0.022647775650024415, 0.02269491195678711, 0.022641664505004884, 0.022601728439331056, 0.023748607635498048, 0.058893310546875, 0.0224901123046875, 0.022862848281860353, 0.023766016006469725, 0.02369024085998535, 0.023801855087280274, 0.02391551971435547, 0.023760896682739258, 0.023767040252685546, 0.0239554557800293, 0.023773183822631837, 0.023768064498901367, 0.023706623077392578, 0.023770143508911133, 0.02378646469116211, 0.023682048797607422, 0.023773183822631837, 0.023766016006469725, 0.023764991760253908, 0.023569408416748046, 0.023801855087280274, 0.023739391326904297, 0.023744512557983398, 0.023160831451416015, 0.02384588813781738, 0.023434240341186522, 0.02306662368774414, 0.023752704620361328, 0.023441408157348635, 0.023702527999877928, 0.023707679748535156, 0.02371887969970703, 0.0237076473236084, 0.02368511962890625, 0.02373222351074219, 0.02369740867614746, 0.023764991760253908, 0.023750656127929686, 0.023526432037353516, 0.02338915252685547, 0.023546880722045898, 0.02371379280090332, 0.023748672485351563, 0.023665599822998047, 0.023837696075439452, 0.023817216873168946, 0.023798784255981444, 0.023771135330200196, 0.023792640686035156, 0.023777280807495117, 0.023834623336791993, 0.023790592193603514, 0.023829504013061522, 0.023880704879760743, 0.02392780876159668, 0.023774208068847655, 0.023773183822631837, 0.02374553680419922, 0.023805952072143553, 0.023706623077392578, 0.023802879333496094, 0.023775232315063476, 0.023824384689331055, 0.05904383850097656, 0.02270310401916504, 0.022656000137329102, 0.022617151260375976, 0.022664127349853517, 0.022697984695434572, 0.02263859176635742, 0.02266726493835449, 0.022665216445922853, 0.022642688751220705, 0.022649856567382814, 0.022648832321166993, 0.022595584869384764, 0.022683679580688478, 0.022656991958618165, 0.02267852783203125, 0.02262015914916992, 0.02268671989440918, 0.02274508857727051, 0.02271129608154297, 0.02264371109008789, 0.022701055526733398, 0.022739967346191405, 0.02273587226867676, 0.02270515251159668, 0.02265292739868164, 0.02266316795349121, 0.02272051239013672, 0.02266828727722168, 0.022601728439331056, 0.022647808074951172, 0.02268057632446289, 0.022675455093383787, 0.022626304626464845, 0.022738943099975584, 0.022664192199707032, 0.022730752944946288, 0.02270515251159668, 0.022700031280517577, 0.02314035224914551, 0.02368307113647461, 0.02370150375366211, 0.023736320495605468, 0.023764991760253908, 0.023828479766845705, 0.023770111083984375, 0.023796735763549806, 0.023827455520629884, 0.024102912902832032, 0.023197696685791015, 0.022767616271972657, 0.022635520935058592, 0.022733823776245117, 0.022602752685546876, 0.022701055526733398, 0.022577152252197266, 0.02265088081359863, 0.02264371109008789, 0.022401023864746093, 0.022349855422973634, 0.022381536483764647, 0.022433792114257813, 0.022598655700683593, 0.059652095794677736, 0.022591487884521484, 0.022658048629760744, 0.022368255615234374, 0.02244915199279785, 0.022794240951538085, 0.022759424209594727, 0.02241433525085449, 0.022368255615234374, 0.022437887191772463, 0.022311935424804686, 0.022552576065063477, 0.02365951919555664, 0.02346188735961914, 0.02350387191772461, 0.023454816818237304, 0.023383968353271483, 0.02352639961242676, 0.02348236846923828, 0.023430143356323242, 0.023377920150756838, 0.023352319717407227, 0.02334623908996582, 0.02340345573425293, 0.023405567169189453, 0.023669759750366212, 0.02351411247253418, 0.02389504051208496, 0.023983104705810547, 0.023820287704467775, 0.02390220832824707, 0.024028160095214843, 0.023669759750366212, 0.024034303665161134, 0.023782400131225585, 0.023734272003173826, 0.023827455520629884, 0.023744543075561522, 0.023795679092407228, 0.02386534309387207, 0.023770111083984375, 0.023706623077392578, 0.02387046432495117, 0.023718912124633788, 0.023916543960571288, 0.023763967514038087, 0.023718912124633788, 0.023747583389282227, 0.023932928085327147, 0.023440383911132814, 0.02370150375366211, 0.023833599090576172, 0.023767040252685546, 0.023734272003173826, 0.02394726371765137, 0.023848960876464844, 0.023773183822631837, 0.023867391586303712, 0.02371788787841797, 0.023813119888305666, 0.02262835121154785, 0.022700031280517577, 0.022624256134033204, 0.059025409698486325, 0.02266214370727539, 0.022606847763061523, 0.02266111946105957, 0.0227061767578125, 0.022656000137329102, 0.022632448196411133, 0.02262937545776367, 0.022580223083496095, 0.02266726493835449, 0.022640640258789063, 0.022567935943603516, 0.022656000137329102, 0.022571008682250978, 0.022601728439331056, 0.022615039825439453, 0.022595584869384764, 0.022593536376953126, 0.02268876838684082, 0.022600704193115235, 0.02264371109008789, 0.022650943756103516, 0.022594495773315428, 0.022609920501708985, 0.02267852783203125, 0.022647808074951172, 0.022588415145874022, 0.022768640518188478, 0.022725631713867187, 0.02265497589111328, 0.02271334457397461, 0.02264678382873535, 0.02265193557739258, 0.0225382080078125, 0.022714368820190428, 0.022563840866088865, 0.022764543533325195, 0.022610944747924806, 0.02263654327392578, 0.02263654327392578, 0.02268671989440918, 0.02262015914916992, 0.02272870445251465, 0.022617088317871094, 0.02267136001586914, 0.022723583221435546, 0.02251571273803711, 0.022607872009277344, 0.022706239700317384, 0.022386751174926757, 0.02251968002319336, 0.0223242244720459, 0.022586368560791017, 0.022722560882568358, 0.022736000061035155, 0.022624128341674803, 0.0227061767578125, 0.022631423950195313, 0.02263654327392578, 0.02267238426208496, 0.02270412826538086, 0.02261203193664551, 0.022956991195678712, 0.05908889770507812, 0.02271129608154297, 0.022689792633056642, 0.02323967933654785, 0.022937599182128905, 0.022674495697021485, 0.022642623901367186, 0.022681600570678712, 0.022618112564086915, 0.022579200744628908, 0.02243891143798828, 0.02248806381225586, 0.024610815048217775, 0.023311359405517578, 0.02262835121154785, 0.02267136001586914, 0.022648832321166993, 0.022658048629760744, 0.022709247589111328, 0.022777856826782225, 0.022740991592407226, 0.02262937545776367, 0.022700031280517577, 0.02269388771057129, 0.02284339141845703, 0.022977535247802734, 0.022939647674560547, 0.023004159927368165, 0.022811647415161132, 0.023335935592651368, 0.022734848022460938, 0.02264473533630371, 0.022912000656127928, 0.022658111572265625, 0.022656959533691408, 0.02270412826538086, 0.022838272094726563, 0.022748159408569335, 0.0227194881439209, 0.022610944747924806, 0.022732799530029296, 0.022749183654785156, 0.022897663116455077, 0.022675455093383787, 0.022579200744628908, 0.022684671401977538, 0.02269388771057129, 0.0226693115234375, 0.022631423950195313, 0.022616064071655274, 0.022631423950195313, 0.022664192199707032, 0.022658048629760744, 0.022689792633056642, 0.022734848022460938, 0.0227061767578125, 0.022734848022460938, 0.02266316795349121, 0.02274508857727051, 0.022595584869384764, 0.022632448196411133, 0.022787071228027343, 0.0227194881439209, 0.05901107025146484, 0.022639616012573242, 0.022617088317871094, 0.022618112564086915, 0.02367180824279785, 0.022838272094726563, 0.022619136810302733, 0.022578176498413087, 0.022633472442626954, 0.022692863464355468, 0.02262118339538574, 0.02262937545776367, 0.02267238426208496, 0.022578176498413087, 0.022632448196411133, 0.02262015914916992, 0.022603776931762694, 0.022607872009277344, 0.02271232032775879, 0.02264575958251953, 0.02264473533630371, 0.022657024383544923, 0.022573055267333983, 0.02264678382873535, 0.02260479927062988, 0.022573055267333983, 0.02268262481689453, 0.02272051239013672, 0.02268569564819336, 0.022625280380249024, 0.022618112564086915, 0.02271334457397461, 0.022665216445922853, 0.022563840866088865, 0.022607872009277344, 0.022573055267333983, 0.022673408508300782, 0.022676479339599608, 0.02264678382873535, 0.02263859176635742, 0.022674432754516603, 0.02265190315246582, 0.02247065544128418, 0.02254243278503418, 0.02230672073364258, 0.022425600051879883, 0.022452224731445314, 0.022297632217407225, 0.02271331214904785, 0.022339584350585938, 0.02259459114074707, 0.022308832168579103, 0.022419456481933595, 0.022495231628417968, 0.022432767868041992, 0.022329343795776366, 0.02243071937561035, 0.02230067253112793, 0.02249932861328125, 0.022351871490478514, 0.0224399356842041, 0.022395904541015626, 0.02240716743469238, 0.05876838302612305, 0.0224768009185791, 0.022375423431396483, 0.02249830436706543, 0.022388832092285156, 0.02228316879272461, 0.022397951126098634, 0.02243395233154297, 0.0223702392578125, 0.02237225532531738, 0.02307174491882324, 0.02408448028564453, 0.0230645751953125, 0.022642688751220705, 0.022345727920532226, 0.02264678382873535, 0.022676479339599608, 0.022535167694091796, 0.022792192459106447, 0.022700128555297853, 0.02259343910217285, 0.022675455093383787, 0.022549663543701172, 0.02256368064880371, 0.022648832321166993, 0.022687744140625, 0.02266726493835449, 0.02271232032775879, 0.022973440170288087, 0.022657024383544923, 0.022641664505004884, 0.022684671401977538, 0.02267852783203125, 0.02265292739868164, 0.022749183654785156, 0.02265497589111328, 0.022788095474243163, 0.0226375675201416, 0.022625312805175782, 0.022774784088134766, 0.02274505615234375, 0.022564863204956053, 0.022730752944946288, 0.02287718391418457, 0.02272768020629883, 0.022845439910888672, 0.022715391159057616, 0.02269900894165039, 0.022658048629760744, 0.022701055526733398, 0.022770687103271483, 0.022738943099975584, 0.022764543533325195, 0.02266111946105957, 0.022725631713867187, 0.0231014404296875, 0.02266111946105957, 0.022709247589111328, 0.022790143966674805, 0.022674432754516603, 0.022958080291748048, 0.022675455093383787, 0.022807552337646485, 0.05912063980102539, 0.0227061767578125, 0.022684671401977538, 0.02268671989440918, 0.022595584869384764, 0.022754304885864256, 0.022627328872680662, 0.022742015838623047, 0.022815807342529297, 0.022767551422119142, 0.022714368820190428, 0.022640640258789063, 0.022809600830078124, 0.02269081687927246, 0.022593536376953126, 0.0226693115234375, 0.02267750358581543, 0.022600704193115235, 0.022769664764404295, 0.022675455093383787, 0.02265497589111328, 0.023618560791015625, 0.022724607467651366, 0.022915071487426757, 0.022746112823486327, 0.022640640258789063, 0.02264271926879883, 0.022668256759643554, 0.022647872924804687, 0.022655935287475587, 0.0227194881439209, 0.02266316795349121, 0.022640640258789063, 0.02266111946105957, 0.022855680465698244, 0.02263654327392578, 0.02271027183532715, 0.022592512130737305, 0.022724607467651366, 0.022576160430908203, 0.02271023941040039, 0.02265292739868164, 0.02262937545776367, 0.022642688751220705, 0.0226375675201416, 0.022594560623168947, 0.022689792633056642, 0.022674432754516603, 0.02265395164489746, 0.0226375675201416, 0.022697984695434572, 0.022625280380249024, 0.022797311782836914, 0.02282598304748535, 0.02264371109008789, 0.02265497589111328, 0.0227194881439209, 0.022623231887817383, 0.022658048629760744, 0.022565887451171874, 0.022601728439331056, 0.022665216445922853, 0.02266316795349121, 0.05907763290405273, 0.02271334457397461, 0.022686752319335937, 0.022666208267211913, 0.02265907287597656, 0.022639616012573242, 0.022631423950195313, 0.022611967086791994, 0.02266009521484375, 0.022798336029052735, 0.022822912216186524, 0.02272870445251465, 0.022731775283813475, 0.02269081687927246, 0.022754304885864256, 0.022575103759765625, 0.022715391159057616, 0.022684671401977538, 0.022606847763061523, 0.022540288925170897, 0.022549503326416014, 0.0225864315032959, 0.022452159881591795, 0.022485055923461915, 0.02272863960266113, 0.022777856826782225, 0.02266422462463379, 0.022675424575805663, 0.02269900894165039, 0.02260479927062988, 0.022625280380249024, 0.02270412826538086, 0.0226693115234375, 0.022657024383544923, 0.022723583221435546, 0.022800416946411134, 0.022975456237792968, 0.02269593620300293, 0.022673408508300782, 0.022716415405273437, 0.02271129608154297, 0.02272972869873047, 0.022573055267333983, 0.022741024017333984, 0.022423519134521484, 0.022541311264038084, 0.022618112564086915, 0.02265088081359863, 0.02248192024230957, 0.02287615966796875, 0.023621631622314454, 0.023591936111450194, 0.02302463912963867, 0.023005184173583985, 0.023147520065307618, 0.022754304885864256, 0.02265497589111328, 0.022691839218139647, 0.02275225639343262, 0.022615039825439453, 0.022665216445922853, 0.02267238426208496, 0.022602783203125]",tokens/s,42.59135557684422,,,,, @@ -14178,7 +14178,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66931d93-797250734b7038a266cbb491;65829534-bf53-46f0-8be7-94a66af82588) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66946fa2-280a54a14385067b776ea7c3;56bac209-eb2a-4b05-b9e3-bf9170762b6b) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-2b/resolve/main/config.json. @@ -14429,7 +14429,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1572, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: CodeGenForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpdqs1ehtq/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: CodeGenForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpkytt17jj/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Llama-2-70b-hf,meta-llama/Llama-2-70b-hf,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): @@ -14610,7 +14610,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66932109-534b7d450074a7474663b847;0be26737-a2a3-49e4-8096-abd00ff5f34b) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66947374-34b2c09d1124fa2d569010da;302b1a4b-f9e8-43a9-9a89-fc444943510f) Repository Not Found for url: https://huggingface.co/s/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -14670,7 +14670,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66931bd6-2fd5a95f4d0a0c1062890e10;272eb332-69e1-4715-9a31-0e1e134e775d) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66946db1-4078e7360175ba7c3e0300be;e58d49b4-87b8-4fb4-9f76-2eefab62ea2e) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/databricks/dbrx-base/resolve/main/config.json. @@ -15016,7 +15016,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66931de0-00dd07bb2a3b5071533746f5;6f246b87-402f-40ea-a5b0-78253f34582d) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66946ffd-1784a49515cc8fb44525bb2e;f8b477c1-a2db-4a6b-9c10-6b9158aa8701) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/recurrentgemma-2b/resolve/main/config.json. @@ -15207,7 +15207,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66931dad-0f2644982f231aa679eab5fa;fb579367-7449-4f37-bffa-140cbf2daa2d) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66946fbf-302ea3be27ac573e549f4c76;c3eae5c7-082a-4c70-b61f-b8b7ede122d8) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-7b/resolve/main/config.json. @@ -15292,7 +15292,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66932392-79de8ce255522092666a2efa;340635b8-e468-4241-9fd6-d970dc347cb6) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6694747d-2be61c275c0d8282412a95e5;a32edf69-6e44-4e64-915b-6ac371637313) Repository Not Found for url: https://huggingface.co/a/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -15726,7 +15726,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669323c6-556d32000a88680243028d49;2da2dd55-9469-41b7-a5b9-504d1db4534c) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66947448-6846d2463d5b138d5d38a1c4;e8ae6f08-ac6a-41e6-bdbb-f4b78086be8e) Repository Not Found for url: https://huggingface.co/l/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -15796,7 +15796,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669322c3-7aa3fd0031f9ee944bb14a1e;c7dcd742-7a30-45e1-8e1b-49d1f9c3fe97) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669474b3-5456d6ca6c7db421315c927d;bfe57325-826e-4cd1-8410-27a5335b8ba7) Repository Not Found for url: https://huggingface.co/i/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -15868,7 +15868,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66931e13-7025d92a39b01af9271f7687;893384b5-213f-4228-b546-db3013c092e0) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66947035-354592131c07fd051b497d90;b0375409-8cc0-44ff-a8a8-cb575bf186e5) Repository Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -15939,7 +15939,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66932089-0a84704430ae09e165666b52;2672c119-69ef-4192-8549-2ce125455da3) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669472d3-0e58b1f67aa709c95a8d7426;c59bea95-5ca1-4c17-81b1-955f74ee5309) Repository Not Found for url: https://huggingface.co/m/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -16376,7 +16376,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6693232b-7c6415720001cbfe1cbb9983;489b1841-7f8f-4a05-88b6-3731f7afd4cd) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66947396-6460313908a75e12626d5686;103f2a9c-028e-4560-bb48-c706cba0c653) Repository Not Found for url: https://huggingface.co/t/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -16742,7 +16742,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6693235f-1d29aed7610cec3509106504;c98080ed-8167-40c7-ae4a-597b34616069) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669473d3-6fa063ad2591412516235075;bbea7084-0185-418e-bf46-73a8f0e8155d) Repository Not Found for url: https://huggingface.co/r/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -16883,7 +16883,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66931d79-310d219e4a19273034e712f4;d02e08af-38a5-4bd3-b299-477a18c3eee4) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66946f85-37bb91ec12e2b80071576a22;4c2ca101-d491-4e15-a2ee-6a4ce0aa55a2) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-2b/resolve/main/config.json. @@ -17309,7 +17309,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669320ef-2b19640143e219d340bacb9b;118f3b58-b851-4be9-a42c-387b8274f52f) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66947355-76598b0f7f7d19526254e134;d40b0c79-a852-4447-8f2e-112429e3a992) Repository Not Found for url: https://huggingface.co/s/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -17369,7 +17369,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66931bbc-00fd76f114658ebd141b61bd;52d8754c-e63b-4550-b099-f482c480eb9e) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66946d96-08ba88c8625dc7686aa984b7;fea3252f-105d-45af-83c6-05583cd0b846) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/databricks/dbrx-base/resolve/main/config.json. @@ -17715,7 +17715,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66931e00-7a33ea7a2ba1aef337dcd8b8;b4229588-5cb9-470a-9abc-7072ed50f1eb) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66947020-3cd7363d7c8981b76fd0f90a;164d2675-2e02-470c-914c-69d93ad15188) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/recurrentgemma-2b/resolve/main/config.json. @@ -17932,7 +17932,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66931dcd-28a266710b1827ba40222a4d;d3e8a7ac-ae9b-4dd2-a8e5-86ad74a2f852) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66946fe5-3ba5f6542d37aa0740c08e0b;b5836c50-9f40-4d52-9946-bf4f0a86db48) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-7b/resolve/main/config.json. @@ -18017,7 +18017,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669323b2-178841560257f7b562863cca;479f8975-fd83-43d7-864f-b4551396eb95) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6694749f-490bdf8678dde2fd46980025;1caa4945-183e-4ba8-ab68-cbe5091d9404) Repository Not Found for url: https://huggingface.co/a/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -18292,7 +18292,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1572, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: CodeGenForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp8_ofwy55/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: CodeGenForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp3y554aig/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,mistralai/Mistral-7B-v0.1,mistralai/Mistral-7B-v0.1,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.1,,,MB,1289.433088,15987.113984,0.0,15340.66688,15114.865664,s,10,0.8963494186401368,0.08963494186401369,0.002654900259774695,0.08876291275024414,0.09013825759887695,0.09383840827941893,0.09679852882385254,"[0.09753855895996094, 0.08931600189208984, 0.08908214569091796, 0.08915328216552734, 0.08856617736816406, 0.0882488021850586, 0.08849651336669923, 0.08842211151123047, 0.08864672088623046, 0.08887910461425781]",tokens/s,2856.0290738892973,kWh,1.0485136945573514e-06,5.745374322362142e-07,6.507330672920453e-06,8.130381799714018e-06,tokens/kWh,31486836.20355992,MB,1289.433088,15987.113984,0.0,15340.66688,15134.114304,s,10,31.359611328125002,3.1359611328125,0.006538257602936493,3.13266650390625,3.14416083984375,3.144967016601562,3.1456119580078123,"[3.143981689453125, 3.133189697265625, 3.141480224609375, 3.14351806640625, 3.145773193359375, 3.1281259765625, 3.132143310546875, 3.13075732421875, 3.131896484375, 3.128745361328125]",tokens/s,20.089534701438783,kWh,3.688938523933013e-05,2.021716879712013e-05,0.00013762656241187918,0.0001947331164483294,tokens/kWh,323519.70301218104,,s,629,32.10659848785398,0.05104387676924325,0.010630762351609589,0.04962918472290039,0.05035558319091797,0.05054604721069336,0.1386133514404297,"[0.05176627349853516, 0.05060812759399414, 0.05173350524902344, 0.0510300178527832, 0.05065523147583008, 0.05116825485229492, 0.05026201629638672, 0.05043302536010742, 0.05133107376098633, 0.05058560180664062, 0.05035212707519531, 0.05014323043823242, 0.04956979370117188, 0.04954521560668945, 0.04950630569458008, 0.04941619110107422, 0.049560577392578124, 0.049912830352783204, 0.0500398063659668, 0.04959027099609375, 0.04955340957641602, 0.04945612716674805, 0.0495728645324707, 0.04961180877685547, 0.049627105712890626, 0.04956159973144531, 0.04952985763549805, 0.04955340957641602, 0.04956159973144531, 0.04952473449707031, 0.049410049438476565, 0.049593345642089844, 0.049713153839111325, 0.049598464965820314, 0.049565696716308595, 0.04955955123901367, 0.04956265640258789, 0.04957283020019531, 0.049576961517333984, 0.04948787307739258, 0.04947353744506836, 0.04943974304199219, 0.04945100784301758, 0.0495206413269043, 0.049522689819335934, 0.049530879974365234, 0.04960563278198242, 0.049448959350585936, 0.04958720016479492, 0.04955750274658203, 0.049565696716308595, 0.049546241760253906, 0.049514495849609375, 0.04951039886474609, 0.04955955123901367, 0.04949708938598633, 0.049448959350585936, 0.05102796936035156, 0.05060300827026367, 0.050165760040283204, 0.049696769714355465, 0.050162689208984375, 0.13955279541015625, 0.04955337524414063, 0.04958924865722656, 0.04955033493041992, 0.05021388626098633, 0.05041459274291992, 0.05030809783935547, 0.04988927841186523, 0.049460224151611325, 0.04950630569458008, 0.04966912078857422, 0.04952985763549805, 0.049442817687988284, 0.05044224166870117, 0.05031219100952149, 0.049585151672363284, 0.049604606628417966, 0.04972748947143555, 0.0500469741821289, 0.049775615692138675, 0.049584129333496096, 0.049601535797119144, 0.04978176116943359, 0.049568767547607424, 0.04956159973144531, 0.049565696716308595, 0.04966912078857422, 0.049770496368408204, 0.049337345123291014, 0.0499989128112793, 0.04930553436279297, 0.049509376525878904, 0.049588222503662106, 0.049691646575927735, 0.04963942337036133, 0.04956979370117188, 0.049642494201660156, 0.04973567962646484, 0.04995276641845703, 0.050601982116699216, 0.05022617721557617, 0.049620990753173826, 0.04945107269287109, 0.04966598510742187, 0.04956671905517578, 0.04962406539916992, 0.049442817687988284, 0.04979507064819336, 0.04965273666381836, 0.0496363525390625, 0.049665088653564456, 0.049749950408935546, 0.049751041412353515, 0.049588222503662106, 0.049549312591552735, 0.04959231948852539, 0.04971110534667969, 0.04952883148193359, 0.049757183074951174, 0.049860607147216796, 0.04971724700927734, 0.04971212768554688, 0.04980633544921875, 0.13930291748046875, 0.04950732803344727, 0.05038489532470703, 0.05042892837524414, 0.050418689727783204, 0.050427902221679685, 0.05011763381958008, 0.05026816177368164, 0.0503818244934082, 0.05037055969238281, 0.05037055969238281, 0.050536449432373044, 0.050016254425048826, 0.05027635192871094, 0.050121726989746096, 0.04980326461791992, 0.049552383422851565, 0.04955340957641602, 0.049514495849609375, 0.04953702545166016, 0.04959641647338867, 0.04962918472290039, 0.050098175048828124, 0.04995379257202148, 0.04961177444458008, 0.04957593536376953, 0.04957081604003906, 0.049604606628417966, 0.04962713623046875, 0.04953497695922852, 0.049478656768798826, 0.04953497695922852, 0.049568767547607424, 0.04961280059814453, 0.0497786865234375, 0.05025075149536133, 0.049772544860839846, 0.04972851181030274, 0.04970086288452148, 0.049598464965820314, 0.04959743881225586, 0.04958924865722656, 0.049721343994140625, 0.04957081604003906, 0.04943974304199219, 0.04937625503540039, 0.04947455978393555, 0.049552383422851565, 0.04962201690673828, 0.04952883148193359, 0.049576961517333984, 0.049667072296142575, 0.04962815856933594, 0.0496445426940918, 0.049788928985595705, 0.050032638549804685, 0.05043609619140625, 0.04997017669677734, 0.04985958480834961, 0.050411518096923826, 0.05037772750854492, 0.05033574295043945, 0.05064089584350586, 0.14009855651855468, 0.05035724639892578, 0.050334720611572265, 0.05036544036865234, 0.05063372802734375, 0.05133414459228516, 0.05056204986572266, 0.05053235244750977, 0.05031731033325195, 0.050743297576904295, 0.050492416381835936, 0.05053440093994141, 0.050252799987792966, 0.04953804779052735, 0.04966502380371094, 0.049478656768798826, 0.04935987091064453, 0.049667072296142575, 0.04947561645507813, 0.04948579025268555, 0.04950630569458008, 0.04948582458496094, 0.04941516876220703, 0.049667072296142575, 0.0494837760925293, 0.04958720016479492, 0.049704959869384766, 0.04980326461791992, 0.05031321716308594, 0.050321407318115234, 0.04961587142944336, 0.04953804779052735, 0.049560577392578124, 0.04953500747680664, 0.049557472229003904, 0.04969881439208984, 0.04961996841430664, 0.04956979370117188, 0.05037062454223633, 0.0496956787109375, 0.04966195297241211, 0.0498073616027832, 0.04941107177734375, 0.049598464965820314, 0.05002547073364258, 0.04978585433959961, 0.04953702545166016, 0.049865726470947266, 0.049718273162841796, 0.049838081359863284, 0.050405376434326174, 0.05039718246459961, 0.04984524917602539, 0.049503231048583986, 0.0495728645324707, 0.0494837760925293, 0.05007462310791016, 0.050405376434326174, 0.04990771102905273, 0.04962713623046875, 0.04973056030273437, 0.04966195297241211, 0.049620990753173826, 0.14034739685058595, 0.05061228942871094, 0.05054355239868164, 0.04983910369873047, 0.04953190231323242, 0.051095550537109374, 0.05050777435302734, 0.050344959259033206, 0.0504268798828125, 0.050392063140869144, 0.05050470352172851, 0.05045043182373047, 0.05006028747558594, 0.04996505737304688, 0.04969779205322265, 0.050267135620117184, 0.049759231567382815, 0.05034086227416992, 0.04962201690673828, 0.05132287979125977, 0.05077811050415039, 0.05013913726806641, 0.049635326385498044, 0.04962406539916992, 0.049805313110351565, 0.05001728057861328, 0.049670143127441405, 0.04955750274658203, 0.04946739196777344, 0.04950425720214844, 0.04961996841430664, 0.04953395080566406, 0.049547264099121094, 0.04956774520874024, 0.04963840103149414, 0.049955841064453124, 0.049634304046630856, 0.04952678298950195, 0.04958924865722656, 0.049562625885009766, 0.050016254425048826, 0.049585151672363284, 0.04969062423706055, 0.04964556884765625, 0.04955648040771484, 0.04984627151489258, 0.04954111862182617, 0.050195457458496094, 0.04973158264160156, 0.049552417755126955, 0.050172897338867185, 0.04964556884765625, 0.049617919921875, 0.05019647979736328, 0.049944576263427735, 0.05037055969238281, 0.049742847442626956, 0.04965785598754883, 0.04961075210571289, 0.04952371215820312, 0.049530879974365234, 0.04948070526123047, 0.049516544342041016, 0.1384806365966797, 0.04946636962890625, 0.0495022087097168, 0.049582080841064455, 0.049495040893554686, 0.04942745590209961, 0.04957900619506836, 0.04948889541625977, 0.04944179153442383, 0.049519615173339845, 0.05005209732055664, 0.05029171371459961, 0.049696769714355465, 0.049667072296142575, 0.04953804779052735, 0.049821697235107425, 0.05025689697265625, 0.05009920120239258, 0.0495728645324707, 0.049347583770751956, 0.049903617858886716, 0.050032638549804685, 0.04965785598754883, 0.049495040893554686, 0.04966912078857422, 0.04957388687133789, 0.04955855941772461, 0.04934140777587891, 0.04946124649047851, 0.04945612716674805, 0.04948070526123047, 0.04962713623046875, 0.049468448638916016, 0.04953699111938477, 0.049516544342041016, 0.049486846923828126, 0.04951244735717773, 0.04953497695922852, 0.049522689819335934, 0.04966502380371094, 0.04976537704467773, 0.04975513458251953, 0.0496097297668457, 0.04966604614257813, 0.04972851181030274, 0.04957900619506836, 0.04949606323242187, 0.04954111862182617, 0.049495040893554686, 0.04967116928100586, 0.04949606323242187, 0.04952883148193359, 0.04943667221069336, 0.049527809143066405, 0.049549312591552735, 0.04962918472290039, 0.049616897583007816, 0.04955136108398438, 0.049598464965820314, 0.04965683364868164, 0.049914878845214845, 0.050547710418701174, 0.0501923828125, 0.1386649627685547, 0.04951859283447266, 0.04958720016479492, 0.04959641647338867, 0.049683456420898435, 0.04953395080566406, 0.04952678298950195, 0.04991692733764649, 0.04969369506835938, 0.049598464965820314, 0.0496629753112793, 0.05059379196166992, 0.05047808074951172, 0.0496363525390625, 0.04962303924560547, 0.049696769714355465, 0.0500316162109375, 0.0495728645324707, 0.04954316711425781, 0.04957900619506836, 0.049479679107666014, 0.04951859283447266, 0.04937830352783203, 0.04956979370117188, 0.049631233215332034, 0.049719295501708984, 0.05034700775146484, 0.05031935882568359, 0.04985036849975586, 0.0501473274230957, 0.05020364761352539, 0.04962918472290039, 0.049634304046630856, 0.04963328170776367, 0.049527809143066405, 0.049734657287597656, 0.049904640197753904, 0.050233345031738284, 0.049977344512939455, 0.04960665512084961, 0.04992102432250976, 0.04967116928100586, 0.049598464965820314, 0.04958720016479492, 0.04961280059814453, 0.04984729766845703, 0.04958720016479492, 0.04956467056274414, 0.04951244735717773, 0.04949094390869141, 0.0495994873046875, 0.04959743881225586, 0.049516544342041016, 0.049604606628417966, 0.04959539031982422, 0.049547264099121094, 0.04954214477539062, 0.0495810546875, 0.04963020706176758, 0.04962406539916992, 0.04962201690673828, 0.04954316711425781, 0.04952985763549805, 0.1384315185546875, 0.04950729751586914, 0.04957081604003906, 0.049511425018310545, 0.049547264099121094, 0.04958924865722656, 0.049955841064453124, 0.050337791442871094, 0.04959539031982422, 0.04947251129150391, 0.04956671905517578, 0.049658878326416016, 0.05014835357666016, 0.05004185485839844, 0.04967424011230469, 0.049650688171386716, 0.04959641647338867, 0.05004496002197266, 0.049627105712890626, 0.04964966583251953, 0.0497367057800293, 0.04956467056274414, 0.04955852890014648, 0.04960870361328125, 0.049448959350585936, 0.04947455978393555, 0.05081497573852539, 0.05029785537719727, 0.04954118347167969, 0.04969875335693359, 0.049568767547607424, 0.04948992156982422, 0.05003878402709961, 0.04977971267700195, 0.04958617782592773, 0.04957388687133789, 0.04964556884765625, 0.049634304046630856, 0.04957798385620117, 0.04962406539916992, 0.04977766418457031, 0.04961894226074219, 0.0495994873046875, 0.049576961517333984, 0.04956159973144531, 0.04971212768554688, 0.04956774520874024, 0.04968960189819336, 0.049600513458251956, 0.049658878326416016, 0.049527809143066405, 0.04960768127441406, 0.04967222213745117, 0.04973871994018555, 0.0495810546875, 0.04955033493041992, 0.04959027099609375, 0.049616897583007816, 0.049565696716308595, 0.04960259246826172, 0.04988924789428711, 0.04965785598754883, 0.04977664184570312, 0.1387509765625, 0.04978278350830078, 0.05017497634887695, 0.04994047927856445, 0.049650688171386716, 0.049732608795166014, 0.04998758316040039, 0.04980428695678711, 0.04976947021484375, 0.04964863967895508, 0.04976332855224609, 0.04972851181030274, 0.04952883148193359, 0.04971212768554688, 0.04970700836181641, 0.04964044952392578, 0.049704959869384766, 0.04968447875976562, 0.04958003234863281, 0.049535999298095705, 0.049616897583007816, 0.04967935943603516, 0.04959641647338867, 0.04958617782592773, 0.049379390716552736, 0.049415103912353514, 0.0496517105102539, 0.049740798950195314, 0.049788928985595705, 0.04975513458251953, 0.04962815856933594, 0.04978176116943359, 0.04962508773803711, 0.0497367057800293, 0.04992716979980469, 0.04976537704467773, 0.04985760116577148, 0.049624000549316406, 0.049503231048583986, 0.04947558212280274, 0.04990054321289063, 0.049651744842529294, 0.04961180877685547, 0.04998854446411133, 0.04954729461669922, 0.04976022338867187, 0.04952985763549805, 0.04959743881225586, 0.05013401412963867, 0.04964044952392578, 0.0498790397644043, 0.049600513458251956, 0.049788928985595705, 0.04956467056274414, 0.04955750274658203, 0.049718273162841796, 0.049775615692138675, 0.04999270248413086, 0.04965580749511719, 0.049584129333496096, 0.04969472122192383, 0.04979916763305664, 0.049675262451171875, 0.13942681884765626, 0.049716224670410154, 0.04972032165527344, 0.04984012985229492, 0.0497039680480957, 0.04961891174316406, 0.04955852890014648, 0.04950835037231445, 0.04962406539916992, 0.0496517105102539, 0.0496363525390625, 0.049708030700683595, 0.049582080841064455, 0.049721343994140625, 0.04949094390869141, 0.04954214477539062, 0.049514495849609375, 0.04958924865722656, 0.05028252792358399, 0.05035516738891602, 0.04951859283447266, 0.04948582458496094, 0.049492992401123044, 0.04997119903564453, 0.049775615692138675, 0.04958003234863281, 0.04955750274658203, 0.049500160217285157, 0.049498111724853515, 0.04959027099609375, 0.04976128005981445, 0.049588222503662106, 0.04953702545166016, 0.04955340957641602, 0.049452030181884765, 0.04951039886474609, 0.049511425018310545, 0.04958003234863281, 0.04952473449707031, 0.049734657287597656, 0.049635326385498044, 0.049601535797119144, 0.049842174530029294, 0.04967116928100586, 0.04973567962646484, 0.049563648223876954, 0.049653793334960936, 0.04958819198608398, 0.04954009628295898, 0.04952473449707031, 0.04955446243286133, 0.04974998474121094, 0.04962508773803711, 0.04956979370117188, 0.04967731094360352, 0.0496855354309082, 0.05006742477416992, 0.049721343994140625, 0.04965990447998047, 0.049650688171386716, 0.049675262451171875, 0.04963328170776367, 0.049506336212158206]",tokens/s,19.5909884455045,,,,, @@ -18479,7 +18479,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669323e5-5cc8ae443e1783eb19fbcfb9;56b8238c-591a-404f-af1d-4e5e85c166a2) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6694746a-0c240cfa4fe5ad0030b27285;5d7d01f6-486a-4d78-9f4c-46cb35a86805) Repository Not Found for url: https://huggingface.co/l/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -18549,7 +18549,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669322e3-378aff2a4f78f5804684476b;f929e051-7d8d-4166-9352-def27f14a976) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669474d4-2c191d8d1c3d1dae66c79b4c;505bb106-6a63-453a-8b3b-0d0f5037a99c) Repository Not Found for url: https://huggingface.co/i/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -18621,7 +18621,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66931e33-4c2da0b352a6c7f50b816b2d;f3f2f539-0d27-4e85-8a51-f10fbe144669) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66947056-5ef7b8911e1bd81e04545160;5e092540-4d2d-4cf8-9c51-39abf7a25706) Repository Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -18676,7 +18676,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1572, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmplu6fv6n_/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp21_mk5l8/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,m,m,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): @@ -18718,7 +18718,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669320a9-3d8bbcfc1c3cbd1b238d1422;fede179f-8419-4673-81b0-0665494f5a66) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669472f8-26458458117a660f6c201551;55c0804a-6026-44d1-86aa-b7cc0d729371) Repository Not Found for url: https://huggingface.co/m/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -19157,7 +19157,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6693234c-1e7c5f971d12f9b463eaa781;5e1eaf1c-7bee-4eee-8019-684690f60ad0) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669473bc-67c1246513079b5054186974;55b27377-a1cd-4583-83fa-050e3ebec93d) Repository Not Found for url: https://huggingface.co/t/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -19523,7 +19523,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6693237f-7f7bee023564416140a87afc;2b666767-52c7-4581-be28-01e7d61221eb) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669473fa-683404ae216356980b7abc0c;adfc29e0-9ecf-4f29-be56-b9bea6981d19) Repository Not Found for url: https://huggingface.co/r/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -19579,7 +19579,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1572, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpessy1mfg/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpewdb52cw/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): @@ -19655,7 +19655,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1572, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp4ddln4lo/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpxbgiwzzz/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,togethercomputer/RedPajama-INCITE-Base-3B-v1,togethercomputer/RedPajama-INCITE-Base-3B-v1,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,MB,1270.267904,6833.045504,0.0,6186.5984,6020.355072,s,10,0.3213976917266846,0.03213976917266846,0.00036806584513412007,0.03201321601867676,0.0321806095123291,0.0327108320236206,0.03313501003265381,"[0.03324105453491211, 0.03205043029785156, 0.03199142456054688, 0.032021217346191407, 0.031999135971069334, 0.031981311798095706, 0.03200521469116211, 0.03199555206298828, 0.032049568176269534, 0.03206278228759766]",tokens/s,7965.20966359962,kWh,3.804204537639366e-07,2.0845340897599228e-07,2.0093245799486035e-06,2.5981984426885324e-06,tokens/kWh,98529810.42321749,MB,1270.267904,6833.045504,0.0,6186.5984,6150.076416,s,10,14.384355590820311,1.4384355590820312,0.02662642684824122,1.4396050415039063,1.4704298217773437,1.4776764587402342,1.4834737683105468,"[1.484923095703125, 1.4688194580078124, 1.4525489501953126, 1.4121748046875, 1.4084876708984375, 1.4079842529296875, 1.4099942626953126, 1.44141064453125, 1.4602130126953126, 1.4377994384765624]",tokens/s,43.79758245145498,kWh,1.6617015781584444e-05,9.106055178637102e-06,5.852164051145294e-05,8.424471147167449e-05,tokens/kWh,747821.4228460195,,s,629,14.64821042633056,0.02328809288764796,0.0038780692554575094,0.022871999740600585,0.02342133674621582,0.024104550552368164,0.054898360290527345,"[0.024374271392822267, 0.024207359313964845, 0.024366079330444337, 0.024183807373046876, 0.024327167510986326, 0.024180736541748047, 0.024491008758544923, 0.02410700798034668, 0.024161279678344725, 0.02447871971130371, 0.024222719192504884, 0.023268384933471678, 0.024062944412231446, 0.02410086441040039, 0.024192031860351564, 0.024439775466918945, 0.024195135116577147, 0.02318022346496582, 0.024639488220214844, 0.023451648712158202, 0.024154111862182616, 0.02433228874206543, 0.024045568466186523, 0.023796735763549806, 0.023248895645141602, 0.02353561592102051, 0.023292959213256834, 0.02330006408691406, 0.023423999786376954, 0.023216127395629883, 0.02349772834777832, 0.023311359405517578, 0.023134208679199218, 0.02311065673828125, 0.023119871139526366, 0.0231844482421875, 0.02304915237426758, 0.023045120239257814, 0.023418880462646483, 0.02312294387817383, 0.022880319595336915, 0.022871999740600585, 0.023043071746826172, 0.023772159576416017, 0.02446335983276367, 0.0236298885345459, 0.023316415786743164, 0.023207935333251953, 0.023149568557739256, 0.02309017562866211, 0.022320127487182616, 0.022221824645996095, 0.023546880722045898, 0.023133184432983397, 0.022952959060668944, 0.02292531204223633, 0.02289254379272461, 0.023043071746826172, 0.0230830078125, 0.023371776580810546, 0.02326937675476074, 0.023151615142822265, 0.05610393524169922, 0.023630847930908205, 0.023274496078491212, 0.023163904190063478, 0.023244800567626952, 0.023198720932006835, 0.023384063720703126, 0.02323865509033203, 0.023167999267578124, 0.023385087966918947, 0.023422975540161133, 0.023143423080444335, 0.02308710479736328, 0.02326835250854492, 0.023241727828979493, 0.02330624008178711, 0.02330112075805664, 0.02352230453491211, 0.023266304016113282, 0.023143423080444335, 0.023262239456176757, 0.023076831817626955, 0.02328883171081543, 0.023189504623413085, 0.02326425552368164, 0.023202816009521485, 0.023218175888061524, 0.023237632751464843, 0.02327961540222168, 0.023237632751464843, 0.023258111953735353, 0.023307264328002928, 0.02328985595703125, 0.023129087448120117, 0.023185407638549805, 0.023200767517089844, 0.023386112213134767, 0.023215103149414062, 0.02325503921508789, 0.023348224639892577, 0.023408639907836915, 0.023215103149414062, 0.023198720932006835, 0.02325196838378906, 0.023225343704223633, 0.023380992889404296, 0.023137279510498047, 0.023146495819091797, 0.023323648452758788, 0.023992319107055664, 0.02430771255493164, 0.023427072525024413, 0.023596031188964844, 0.023347200393676756, 0.023249919891357423, 0.023185407638549805, 0.02345062446594238, 0.02323865509033203, 0.023203840255737306, 0.023153663635253906, 0.02369740867614746, 0.023154687881469727, 0.02329804801940918, 0.055006206512451174, 0.02348748779296875, 0.023326719284057617, 0.02326016044616699, 0.023147520065307618, 0.022994943618774414, 0.02306559944152832, 0.023174144744873046, 0.023447551727294923, 0.023389184951782226, 0.023565311431884766, 0.023165952682495116, 0.023085056304931642, 0.023145471572875977, 0.023149568557739256, 0.023155712127685548, 0.023034879684448242, 0.023188480377197264, 0.02331340789794922, 0.023370752334594725, 0.023353343963623048, 0.023044095993041993, 0.022890495300292968, 0.023018495559692383, 0.02308403205871582, 0.02289664077758789, 0.023003135681152344, 0.02308710479736328, 0.02304921531677246, 0.022914047241210937, 0.023018495559692383, 0.023031808853149413, 0.022821887969970703, 0.02298067283630371, 0.022905855178833007, 0.02285152053833008, 0.022922367095947267, 0.022827903747558595, 0.023295135498046876, 0.023107423782348632, 0.022951936721801756, 0.022935552597045897, 0.022916095733642578, 0.022816768646240236, 0.02286796760559082, 0.02301241683959961, 0.022929344177246094, 0.022952959060668944, 0.022854656219482423, 0.022815744400024415, 0.02323148727416992, 0.022874111175537108, 0.022898687362670898, 0.022929407119750975, 0.023088319778442383, 0.022900543212890624, 0.022822912216186524, 0.022861824035644532, 0.02287820816040039, 0.022920192718505858, 0.022947839736938477, 0.023216127395629883, 0.023380992889404296, 0.054978591918945316, 0.022420448303222658, 0.022261760711669923, 0.022445056915283205, 0.02230681610107422, 0.02225971221923828, 0.022347776412963868, 0.02240716743469238, 0.022294559478759766, 0.022413280487060545, 0.02229350471496582, 0.02228326416015625, 0.02229964828491211, 0.022344703674316405, 0.022305791854858398, 0.02227609634399414, 0.02230271911621094, 0.022305791854858398, 0.022409215927124023, 0.02269900894165039, 0.02227609634399414, 0.02230886459350586, 0.022331392288208008, 0.022313983917236328, 0.022390783309936522, 0.022305791854858398, 0.022381568908691408, 0.022410240173339844, 0.022396928787231447, 0.022297599792480468, 0.022150144577026368, 0.02231603240966797, 0.022344703674316405, 0.022328319549560546, 0.022337535858154296, 0.022336511611938475, 0.022351903915405272, 0.02230473518371582, 0.022351871490478514, 0.02229452705383301, 0.022379520416259766, 0.022337535858154296, 0.02231808090209961, 0.022429695129394533, 0.02230784034729004, 0.02240716743469238, 0.02230988883972168, 0.02225564765930176, 0.022439903259277343, 0.022329343795776366, 0.02230681610107422, 0.022268928527832032, 0.022403072357177735, 0.0223242244720459, 0.022365184783935548, 0.022380544662475587, 0.0222740478515625, 0.02226585578918457, 0.02309529685974121, 0.023193599700927735, 0.023163904190063478, 0.023185407638549805, 0.02330521583557129, 0.05482905578613281, 0.02245427131652832, 0.02236115264892578, 0.022313919067382813, 0.02228223991394043, 0.022371328353881836, 0.0223191032409668, 0.022286336898803712, 0.022301759719848633, 0.023443391799926758, 0.02293350410461426, 0.022311935424804686, 0.022399999618530272, 0.02228428840637207, 0.02236832046508789, 0.022343616485595703, 0.022319168090820313, 0.022335424423217773, 0.022378496170043945, 0.022296575546264647, 0.022331392288208008, 0.022312959671020507, 0.02233241653442383, 0.022323232650756836, 0.023557088851928712, 0.023173120498657225, 0.022648895263671875, 0.02297235107421875, 0.022961151123046874, 0.022391807556152343, 0.02242355155944824, 0.022566911697387695, 0.02236419105529785, 0.02232009506225586, 0.022374399185180666, 0.021836799621582033, 0.022156288146972656, 0.02208665657043457, 0.022064128875732423, 0.022032384872436524, 0.02206003189086914, 0.02205286407470703, 0.022112255096435548, 0.02201190376281738, 0.02205183982849121, 0.022923263549804687, 0.02270310401916504, 0.02267852783203125, 0.02229964828491211, 0.022033407211303712, 0.022090751647949217, 0.02209791946411133, 0.022072351455688477, 0.022026208877563475, 0.022060127258300782, 0.02209270477294922, 0.022041599273681642, 0.022029312133789062, 0.022191104888916017, 0.02206208038330078, 0.022204416275024414, 0.022189056396484375, 0.02229350471496582, 0.05492531204223633, 0.022343679428100584, 0.02228428840637207, 0.02227916717529297, 0.022185983657836913, 0.022288383483886717, 0.02228223991394043, 0.022210559844970702, 0.02231603240966797, 0.022432767868041992, 0.022215679168701173, 0.022350847244262697, 0.022230016708374024, 0.022297599792480468, 0.02233241653442383, 0.022194175720214843, 0.022393856048583984, 0.02233344078063965, 0.022311935424804686, 0.022337535858154296, 0.02229862403869629, 0.022334463119506837, 0.02227712059020996, 0.022296575546264647, 0.022363136291503907, 0.02230271911621094, 0.022361087799072265, 0.022368255615234374, 0.02226483154296875, 0.02281062316894531, 0.02223411178588867, 0.022297599792480468, 0.022492160797119142, 0.02226688003540039, 0.022296575546264647, 0.02225663948059082, 0.022658048629760744, 0.022839296340942384, 0.022271999359130858, 0.023152639389038086, 0.022804479598999023, 0.022354944229125977, 0.02231500816345215, 0.022258752822875975, 0.022264768600463867, 0.022363136291503907, 0.022227008819580077, 0.02221766471862793, 0.022206464767456056, 0.02224947166442871, 0.022252544403076172, 0.02220134353637695, 0.022303743362426756, 0.02230681610107422, 0.022221824645996095, 0.022328319549560546, 0.0224849910736084, 0.02227302360534668, 0.02227609634399414, 0.02224844741821289, 0.0223242244720459, 0.022181983947753905, 0.022335391998291015, 0.05477478408813476, 0.022334463119506837, 0.022175743103027345, 0.022268928527832032, 0.022262847900390625, 0.02219411277770996, 0.022338560104370117, 0.02213580894470215, 0.022252544403076172, 0.022255615234375, 0.022323200225830078, 0.022255615234375, 0.022323200225830078, 0.022238208770751954, 0.02227712059020996, 0.022320127487182616, 0.022331392288208008, 0.022271999359130858, 0.022347776412963868, 0.022512704849243163, 0.022311872482299804, 0.02227302360534668, 0.022271999359130858, 0.02229452705383301, 0.022231039047241212, 0.02229555130004883, 0.022358015060424806, 0.02229145622253418, 0.022254592895507814, 0.02225049591064453, 0.022368255615234374, 0.022252544403076172, 0.022336511611938475, 0.022330368041992187, 0.02228223991394043, 0.02227302360534668, 0.022361087799072265, 0.022373376846313478, 0.02244095993041992, 0.02249830436706543, 0.022293535232543946, 0.022239200592041014, 0.02234880065917969, 0.02230067253112793, 0.022412288665771486, 0.022352895736694335, 0.022262815475463868, 0.02237539291381836, 0.02224844741821289, 0.022271999359130858, 0.02245631980895996, 0.02468556785583496, 0.023274496078491212, 0.02234880065917969, 0.02233241653442383, 0.02228326416015625, 0.022328319549560546, 0.022331392288208008, 0.022355968475341798, 0.022785024642944338, 0.022534143447875975, 0.022313983917236328, 0.022358015060424806, 0.05495500946044922, 0.022394880294799805, 0.022388736724853517, 0.022261760711669923, 0.02227712059020996, 0.02220953559875488, 0.02226585578918457, 0.022288383483886717, 0.022326271057128907, 0.022501375198364256, 0.02230681610107422, 0.022296575546264647, 0.02229145622253418, 0.022257696151733397, 0.023346143722534178, 0.022845439910888672, 0.02240716743469238, 0.02243174362182617, 0.022399999618530272, 0.022125568389892578, 0.022190080642700196, 0.02239897537231445, 0.022405120849609376, 0.022321151733398437, 0.022328319549560546, 0.022372352600097657, 0.022429695129394533, 0.02243071937561035, 0.022985727310180663, 0.023193599700927735, 0.02323148727416992, 0.023157760620117186, 0.022347776412963868, 0.022377471923828125, 0.022624256134033204, 0.024639488220214844, 0.02385411262512207, 0.023233503341674806, 0.023357440948486328, 0.02325503921508789, 0.02332467269897461, 0.023221248626708983, 0.023212032318115236, 0.023362560272216795, 0.02342092704772949, 0.023341056823730468, 0.02329497528076172, 0.023375871658325196, 0.023243839263916016, 0.023259071350097655, 0.023259136199951173, 0.02325196838378906, 0.023240703582763672, 0.023149568557739256, 0.02332876777648926, 0.02327859115600586, 0.023368703842163087, 0.023267328262329103, 0.02326527976989746, 0.02323967933654785, 0.023372800827026367, 0.023365631103515624, 0.02326937675476074, 0.05528575897216797, 0.023348224639892577, 0.023189504623413085, 0.023409664154052736, 0.02348646354675293, 0.023257087707519532, 0.023149568557739256, 0.02330009651184082, 0.022226943969726562, 0.022347776412963868, 0.023008384704589845, 0.02307468795776367, 0.023234560012817384, 0.023200767517089844, 0.02330112075805664, 0.02325503921508789, 0.023205888748168944, 0.023311359405517578, 0.023451648712158202, 0.02327552032470703, 0.023208959579467774, 0.022408191680908202, 0.02262118339538574, 0.02304614448547363, 0.023221248626708983, 0.023258111953735353, 0.023235584259033205, 0.023236608505249022, 0.023235584259033205, 0.022334463119506837, 0.022381568908691408, 0.022387712478637696, 0.02251263999938965, 0.023274496078491212, 0.02308710479736328, 0.023072959899902344, 0.022961088180541992, 0.022991743087768555, 0.02309529685974121, 0.02432307243347168, 0.02454630470275879, 0.023562240600585937, 0.02330112075805664, 0.023259136199951173, 0.023229440689086913, 0.02329804801940918, 0.02327859115600586, 0.02330316734313965, 0.02322329521179199, 0.023451648712158202, 0.023347200393676756, 0.02330316734313965, 0.023218175888061524, 0.023214080810546874, 0.02334617614746094, 0.023282688140869142, 0.023197696685791015, 0.023258111953735353, 0.023222272872924804, 0.023178239822387696, 0.023195648193359376, 0.023611455917358398, 0.02323961639404297, 0.05541686248779297, 0.02276041603088379, 0.022387712478637696, 0.02230988883972168, 0.022323200225830078, 0.022363136291503907, 0.022377471923828125, 0.02245734405517578, 0.022372352600097657, 0.022353919982910156, 0.02231603240966797, 0.022331392288208008, 0.02234060859680176, 0.022339584350585938, 0.02230067253112793, 0.022322175979614257, 0.022346752166748047, 0.022386688232421875, 0.02245631980895996, 0.023244800567626952, 0.023359487533569336, 0.02327654457092285, 0.023224319458007812, 0.023212032318115236, 0.023193599700927735, 0.023214080810546874, 0.02327859115600586, 0.02330112075805664, 0.02330112075805664, 0.023209983825683594, 0.023391231536865235, 0.023365631103515624, 0.023362560272216795, 0.023244800567626952, 0.02329804801940918, 0.023152639389038086, 0.023310335159301757, 0.023232511520385742, 0.02327654457092285, 0.02311680030822754, 0.023367679595947266, 0.022374399185180666, 0.022413312911987306, 0.022374399185180666, 0.022373376846313478, 0.02270412826538086, 0.02244607925415039, 0.022389759063720704, 0.022420480728149415, 0.022487039566040038, 0.022367231369018553, 0.022314048767089843, 0.022409151077270508, 0.02242252731323242, 0.022409215927124023, 0.02255564880371094, 0.023267328262329103, 0.023315456390380858, 0.02327244758605957, 0.02329702377319336, 0.023222272872924804, 0.02346188735961914, 0.023400447845458985]",tokens/s,42.940398976611846,,,,, @@ -19716,7 +19716,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66931d9a-4a10412c2f3fbac707538685;d60793a6-a2a9-41eb-8de8-a94a30d66a3a) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66946fa9-026215554a425cfb01d11f96;6b1ec6fd-4bd1-4188-9035-64e3089894f3) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-2b/resolve/main/config.json. @@ -19967,7 +19967,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1572, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: CodeGenForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp3a4zoivb/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: CodeGenForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp4v6my2hw/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Llama-2-70b-hf,meta-llama/Llama-2-70b-hf,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): @@ -20148,7 +20148,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6693210f-3c67c55c43ce096e2e41fa28;4f224867-7eea-443d-8e85-6510af419f55) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6694737c-1b7f88db07f5900351a51119;c5600467-201d-45eb-9b2e-f9b2bd21e23b) Repository Not Found for url: https://huggingface.co/s/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -20208,7 +20208,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66931bdd-2689beb24a7069cc7c71b75d;50d4fbd1-9b11-4c72-ae6f-70f424e53c20) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66946db8-4a7b4a2234902636529b08b7;b14ad759-a194-41e9-91e4-062811fca6a8) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/databricks/dbrx-base/resolve/main/config.json. @@ -20554,7 +20554,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66931df3-4e7354b01f003f7047171b1b;3cf40d7e-a98b-4fb4-a9d6-725dcd4bfe8f) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66947011-3806455a29b75d166f60149f;1601182f-1e87-498a-a153-f1dfb352699e) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/recurrentgemma-2b/resolve/main/config.json. @@ -20823,7 +20823,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66931dc0-6a9b190232cdc96d664d9090;e2b07218-7c10-4d3b-9ca3-36937ecf48e1) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66946fd6-2e680eec526a339b43fb7594;45388a4f-ea9d-46dd-ab44-46648e8bc22e) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-7b/resolve/main/config.json. @@ -20908,7 +20908,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669323a5-6e81966c7644871b2091611f;68859509-da75-4a53-8f5c-7e4b70a2ab67) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66947492-789c207f7cc768ab1fc98ebb;7d259cac-2af5-4fbc-b861-9b7e93c14ae3) Repository Not Found for url: https://huggingface.co/a/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -21349,7 +21349,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669323d9-37d809721af8009c3d430736;239c8e0c-516a-4e54-b058-b5c582ca8774) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6694745d-06c2c7f1543af17b66cd03f0;62749f78-948a-495f-b1a6-96a29d5cd358) Repository Not Found for url: https://huggingface.co/l/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -21419,7 +21419,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669322d6-6d54ca6f1da10ed17bf06055;0ef82abe-4916-421c-82f2-720acd953f5a) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669474c7-11c53e2b0a32eabf7141f0c3;61e2f5e0-389a-4406-8dc9-85d494e86ef4) Repository Not Found for url: https://huggingface.co/i/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -21490,7 +21490,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66931e26-57226a056fb6a1eb217967f9;2e5a7379-37fa-474a-af8a-753f33701948) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66947048-4f3ad67a68bafb75790312ff;cc8ec58f-e063-46b7-a33f-30aa26a60e77) Repository Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -21587,7 +21587,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6693209c-57754e1820cf7698199963c9;19a0bd94-da84-4b56-9586-56e0612ec2eb) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669472ea-2818c28145d4623108d141f4;ae60b401-f432-4183-a446-e9a82bb9bb78) Repository Not Found for url: https://huggingface.co/m/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -22023,7 +22023,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6693233f-35ceebb154dcdd584d783ee0;fcc4144e-1237-4447-b4e5-97f7dceb4530) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669473ad-586e3d0b72f94d346722e174;8c42337d-2e5a-4918-8fd9-d51a00703850) Repository Not Found for url: https://huggingface.co/t/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -22393,7 +22393,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66932372-2b98a38f47902c0126a3c653;29b7d104-bd55-4049-9cfb-51b444c0f2ee) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-669473ea-69adecae312f9779412c9183;8023eb00-40b5-46cb-860d-fea063a86411) Repository Not Found for url: https://huggingface.co/r/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -22585,7 +22585,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66931d8d-5bfd626118c8246409350ddf;b9f30a07-09e7-4092-9cbc-6b5038de68f3) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66946f9b-4207b3b4684543400c2c4b8f;43e3c556-2667-45a8-9c72-a24e6d498d93) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-2b/resolve/main/config.json. @@ -23042,7 +23042,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66932103-1ebafb9d22da89d0218bfbcc;14bff0b3-85a7-4e2b-bb4f-6edf2c6bcf42) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6694736c-7dc7f9183d347ac031da4ff3;9fc07c5c-19d5-4eb5-922d-d33cf3318a42) Repository Not Found for url: https://huggingface.co/s/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -23102,7 +23102,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66931bd0-5525e2fb7ddcf82f7f8264d8;3a76ee64-7920-4f23-9196-a19c2c519ee6) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66946daa-2d104ac65586cd7e02d5b16f;af254e79-a901-460b-ae8c-164d93557f9c) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/databricks/dbrx-base/resolve/main/config.json.