mgladden commited on
Commit
78c9ad0
1 Parent(s): 9ed4f75

Upload 7 files

Browse files
Files changed (6) hide show
  1. ManaGPT_logo_01.png +0 -0
  2. README.md +36 -0
  3. config.json +41 -0
  4. generation_config.json +9 -0
  5. tf_model.h5 +3 -0
  6. tokenizer.json +0 -0
ManaGPT_logo_01.png ADDED
README.md CHANGED
@@ -1,3 +1,39 @@
1
  ---
2
  license: mit
 
 
 
 
 
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: mit
3
+ tags:
4
+ - management
5
+ - text generation
6
+ model-index:
7
+ - name: ManaGPT-1010
8
+ results: []
9
+ language:
10
+ - en
11
+ pipeline_tag: text-generation
12
  ---
13
+
14
+ # ManaGPT-1020
15
+ <img style="float:right; margin:10px; margin-right:30px" src="https://huggingface.co/NeuraXenetica/ManaGPT-1010/resolve/main/ManaGPT_logo_01.png" width="150" height="150"></img>
16
+ **ManaGPT-1020** is an experimental open-source text-generating AI designed to offer insights on the role of emerging technologies in organizational management.
17
+
18
+ _(Please note that ManaGPT-1020 has superseded by **[ManaGPT-1010](https://huggingface.co/NeuraXenetica/ManaGPT-1010)**: the newer model has been fine-tuned on a dataset roughly 6.45 times the size of that used to fine-tune ManaGPT-1010.)_
19
+
20
+ ## Model description
21
+
22
+ The model is a fine-tuned version of GPT-2 that has been trained on a custom corpus of scholarly and popular texts from the field of organizational management that relate to ongoing effects of posthumanizing technologies (e.g., relating to advanced artificial intelligence, social robotics, virtual reality, neuroprosthetics, and cyber-physical systems) on the structure of organizations and human beings’ experience of organizational life.
23
+
24
+ ## Intended uses & limitations
25
+
26
+ This model has been designed for experimental research purposes; it isn’t intended for use in a production setting or in any sensitive or potentially hazardous contexts.
27
+
28
+ ## Training procedure and hyperparameters
29
+
30
+ The model was trained using a Tesla T4 with 16GB of GPU memory. The following hyperparameters were used during training:
31
+ - optimizer: {'name': 'AdamWeightDecay', 'learning_rate': {'class_name': 'ExponentialDecay', 'config': {'initial_learning_rate': 0.0005, 'decay_steps': 500, 'decay_rate': 0.95, 'staircase': False, 'name': None}}, 'decay': 0.0, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-07, 'amsgrad': False, 'weight_decay_rate': 0.01}
32
+ - training_precision: float32
33
+
34
+ ### Framework versions
35
+
36
+ - Transformers 4.27.1
37
+ - TensorFlow 2.11.0
38
+ - Datasets 2.10.1
39
+ - Tokenizers 0.13.2
config.json ADDED
@@ -0,0 +1,41 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "gpt2",
3
+ "activation_function": "gelu_new",
4
+ "architectures": [
5
+ "GPT2LMHeadModel"
6
+ ],
7
+ "attn_pdrop": 0.1,
8
+ "bos_token_id": 50256,
9
+ "do_sample": true,
10
+ "embd_pdrop": 0.1,
11
+ "eos_token_id": 50256,
12
+ "initializer_range": 0.02,
13
+ "layer_norm_epsilon": 1e-05,
14
+ "max_length": 50,
15
+ "model_type": "gpt2",
16
+ "n_ctx": 1024,
17
+ "n_embd": 768,
18
+ "n_head": 12,
19
+ "n_inner": null,
20
+ "n_layer": 12,
21
+ "n_positions": 1024,
22
+ "pad_token_id": 50256,
23
+ "reorder_and_upcast_attn": false,
24
+ "resid_pdrop": 0.1,
25
+ "scale_attn_by_inverse_layer_idx": false,
26
+ "scale_attn_weights": true,
27
+ "summary_activation": null,
28
+ "summary_first_dropout": 0.1,
29
+ "summary_proj_to_labels": true,
30
+ "summary_type": "cls_index",
31
+ "summary_use_proj": true,
32
+ "task_specific_params": {
33
+ "text-generation": {
34
+ "do_sample": true,
35
+ "max_length": 50
36
+ }
37
+ },
38
+ "transformers_version": "4.27.2",
39
+ "use_cache": true,
40
+ "vocab_size": 50257
41
+ }
generation_config.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 50256,
4
+ "do_sample": true,
5
+ "eos_token_id": 50256,
6
+ "max_length": 50,
7
+ "pad_token_id": 50256,
8
+ "transformers_version": "4.27.2"
9
+ }
tf_model.h5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5b4c1d25cec73df540aefddb6441e4efb11b1adf598cea7cd0827936cf8d2894
3
+ size 497935440
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff