jtatman commited on
Commit
2d22813
1 Parent(s): a91149e

Upload MixtralForCausalLM

Browse files
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "jtatman/TinyDolphin-3x-MoE",
3
  "architectures": [
4
  "MixtralForCausalLM"
5
  ],
 
1
  {
2
+ "_name_or_path": "jtatman/tinydolphin-3x-MoE-samantha-chatml-therabot",
3
  "architectures": [
4
  "MixtralForCausalLM"
5
  ],
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:10cbdaae08487c0d585b367887075212bf1b15257d7f913af43adc1ef1e9ef56
3
  size 4999308976
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3b75330bf84734e250e2a539552859481757716ae30dda38155cdd4d063f1757
3
  size 4999308976
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:96e557ee08229f3d00261502c209afadf44bfd353d7032973b475dfabf1604cd
3
  size 246690872
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bb44febaf2db904bb9baf5e9018c365bdc7ed411f90eb2544a7dbc02dfdd29cf
3
  size 246690872