sharpenb commited on
Commit
b3f5061
1 Parent(s): 0a36c3b

Upload folder using huggingface_hub (#6)

Browse files

- 03a02bba037bf2e5439fac7b6e106c21f4e2ea32796fdba92b39e88cc351b846 (feb21add89996b2ba767c0f1b63b287142904296)

Files changed (5) hide show
  1. config.json +1 -1
  2. model.safetensors +1 -1
  3. plots.png +0 -0
  4. results.json +13 -13
  5. smash_config.json +1 -1
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "/tmp/tmpu6eey9a0",
3
  "_remove_final_layer_norm": false,
4
  "activation_dropout": 0.0,
5
  "activation_function": "relu",
 
1
  {
2
+ "_name_or_path": "/tmp/tmpzu_5c0ni",
3
  "_remove_final_layer_norm": false,
4
  "activation_dropout": 0.0,
5
  "activation_function": "relu",
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1206bd904e52ca38ea1bb908811eafb9b85b9b6b69d10ee75cb2d51ccae977a3
3
  size 201991864
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c72c8ed6eea9a8b4a088882bcef6168a2e7ac773572c98dfc6140f31e5e5e484
3
  size 201991864
plots.png CHANGED
results.json CHANGED
@@ -3,20 +3,20 @@
3
  "base_current_gpu_total_memory": 40339.3125,
4
  "base_memory_inference_first": 690.0,
5
  "base_memory_inference": 570.0,
6
- "base_token_generation_latency_sync": 26.18392219543457,
7
- "base_token_generation_latency_async": 26.8399927765131,
8
- "base_token_generation_throughput_sync": 0.038191375323226404,
9
- "base_token_generation_throughput_async": 0.03725783417032329,
10
- "base_token_generation_CO2_emissions": 7.27957891391666e-06,
11
- "base_token_generation_energy_consumption": 0.002083295566052838,
12
  "smashed_current_gpu_type": "NVIDIA A100-PCIE-40GB",
13
  "smashed_current_gpu_total_memory": 40339.3125,
14
- "smashed_memory_inference_first": 184.0,
15
  "smashed_memory_inference": 206.0,
16
- "smashed_token_generation_latency_sync": 20.504255294799805,
17
- "smashed_token_generation_latency_async": 20.576557517051697,
18
- "smashed_token_generation_throughput_sync": 0.048770364279146265,
19
- "smashed_token_generation_throughput_async": 0.04859899422783936,
20
- "smashed_token_generation_CO2_emissions": 6.411004057764176e-06,
21
- "smashed_token_generation_energy_consumption": 0.001626662612168566
22
  }
 
3
  "base_current_gpu_total_memory": 40339.3125,
4
  "base_memory_inference_first": 690.0,
5
  "base_memory_inference": 570.0,
6
+ "base_token_generation_latency_sync": 25.73595085144043,
7
+ "base_token_generation_latency_async": 25.555139780044556,
8
+ "base_token_generation_throughput_sync": 0.03885615129483473,
9
+ "base_token_generation_throughput_async": 0.03913107142465634,
10
+ "base_token_generation_CO2_emissions": 7.04025152217974e-06,
11
+ "base_token_generation_energy_consumption": 0.00201063437593726,
12
  "smashed_current_gpu_type": "NVIDIA A100-PCIE-40GB",
13
  "smashed_current_gpu_total_memory": 40339.3125,
14
+ "smashed_memory_inference_first": 164.0,
15
  "smashed_memory_inference": 206.0,
16
+ "smashed_token_generation_latency_sync": 20.38736572265625,
17
+ "smashed_token_generation_latency_async": 21.09651416540146,
18
+ "smashed_token_generation_throughput_sync": 0.04904998583945111,
19
+ "smashed_token_generation_throughput_async": 0.047401195863912546,
20
+ "smashed_token_generation_CO2_emissions": 6.962162215496025e-06,
21
+ "smashed_token_generation_energy_consumption": 0.001568448312001272
22
  }
smash_config.json CHANGED
@@ -14,7 +14,7 @@
14
  "controlnet": "None",
15
  "unet_dim": 4,
16
  "device": "cuda",
17
- "cache_dir": "/ceph/hdd/staff/charpent/.cache/models5ojho2i3",
18
  "batch_size": 1,
19
  "tokenizer": "GPT2TokenizerFast(name_or_path='facebook/opt-125m', vocab_size=50265, model_max_length=1000000000000000019884624838656, is_fast=True, padding_side='right', truncation_side='right', special_tokens={'bos_token': '</s>', 'eos_token': '</s>', 'unk_token': '</s>', 'pad_token': '<pad>'}, clean_up_tokenization_spaces=True), added_tokens_decoder={\n\t1: AddedToken(\"<pad>\", rstrip=False, lstrip=False, single_word=False, normalized=True, special=True),\n\t2: AddedToken(\"</s>\", rstrip=False, lstrip=False, single_word=False, normalized=True, special=True),\n}",
20
  "task": "text_text_generation",
 
14
  "controlnet": "None",
15
  "unet_dim": 4,
16
  "device": "cuda",
17
+ "cache_dir": "/ceph/hdd/staff/charpent/.cache/modelsmqfybacy",
18
  "batch_size": 1,
19
  "tokenizer": "GPT2TokenizerFast(name_or_path='facebook/opt-125m', vocab_size=50265, model_max_length=1000000000000000019884624838656, is_fast=True, padding_side='right', truncation_side='right', special_tokens={'bos_token': '</s>', 'eos_token': '</s>', 'unk_token': '</s>', 'pad_token': '<pad>'}, clean_up_tokenization_spaces=True), added_tokens_decoder={\n\t1: AddedToken(\"<pad>\", rstrip=False, lstrip=False, single_word=False, normalized=True, special=True),\n\t2: AddedToken(\"</s>\", rstrip=False, lstrip=False, single_word=False, normalized=True, special=True),\n}",
20
  "task": "text_text_generation",