modelId
stringlengths 5
122
| author
stringlengths 2
42
| last_modified
unknown | downloads
int64 0
738M
| likes
int64 0
11k
| library_name
stringclasses 245
values | tags
sequencelengths 1
4.05k
| pipeline_tag
stringclasses 48
values | createdAt
unknown | card
stringlengths 1
901k
|
---|---|---|---|---|---|---|---|---|---|
yepyepyo/Minji_NewJeans | yepyepyo | "2024-06-22T03:34:44Z" | 0 | 0 | null | [
"region:us"
] | null | "2024-06-22T03:24:58Z" | Entry not found |
mattzhang/idefics-9b-doodles | mattzhang | "2024-06-22T05:41:07Z" | 0 | 0 | transformers | [
"transformers",
"safetensors",
"idefics",
"pretraining",
"arxiv:1910.09700",
"endpoints_compatible",
"text-generation-inference",
"4-bit",
"bitsandbytes",
"region:us"
] | null | "2024-06-22T03:26:13Z" | ---
library_name: transformers
tags: []
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed] |
mttgermano/q-FrozenLake-v1-4x4-noSlippery | mttgermano | "2024-06-22T03:27:27Z" | 0 | 0 | null | [
"FrozenLake-v1-4x4-no_slippery",
"q-learning",
"reinforcement-learning",
"custom-implementation",
"model-index",
"region:us"
] | reinforcement-learning | "2024-06-22T03:27:23Z" | ---
tags:
- FrozenLake-v1-4x4-no_slippery
- q-learning
- reinforcement-learning
- custom-implementation
model-index:
- name: q-FrozenLake-v1-4x4-noSlippery
results:
- task:
type: reinforcement-learning
name: reinforcement-learning
dataset:
name: FrozenLake-v1-4x4-no_slippery
type: FrozenLake-v1-4x4-no_slippery
metrics:
- type: mean_reward
value: 1.00 +/- 0.00
name: mean_reward
verified: false
---
# **Q-Learning** Agent playing **FrozenLake-v1**
This is a trained model of a **Q-Learning** agent playing **FrozenLake-v1** .
## Usage
```python
model = load_from_hub(repo_id="mttgermano/q-FrozenLake-v1-4x4-noSlippery", filename="q-learning.pkl")
# Don't forget to check if you need to add additional attributes (is_slippery=False etc)
env = gym.make(model["env_id"])
```
|
mttgermano/q-Taxi-v3 | mttgermano | "2024-06-22T03:32:06Z" | 0 | 0 | null | [
"Taxi-v3",
"q-learning",
"reinforcement-learning",
"custom-implementation",
"model-index",
"region:us"
] | reinforcement-learning | "2024-06-22T03:32:02Z" | ---
tags:
- Taxi-v3
- q-learning
- reinforcement-learning
- custom-implementation
model-index:
- name: q-Taxi-v3
results:
- task:
type: reinforcement-learning
name: reinforcement-learning
dataset:
name: Taxi-v3
type: Taxi-v3
metrics:
- type: mean_reward
value: 7.54 +/- 2.74
name: mean_reward
verified: false
---
# **Q-Learning** Agent playing **Taxi-v3**
This is a trained model of a **Q-Learning** agent playing **Taxi-v3** .
## Usage
```python
model = load_from_hub(repo_id="mttgermano/q-Taxi-v3", filename="q-learning.pkl")
# Don't forget to check if you need to add additional attributes (is_slippery=False etc)
env = gym.make(model["env_id"])
```
|
vakev/TEST | vakev | "2024-06-22T03:33:30Z" | 0 | 0 | null | [
"license:apache-2.0",
"region:us"
] | null | "2024-06-22T03:33:30Z" | ---
license: apache-2.0
---
|
Andy1124233/Capstone_Forecaster | Andy1124233 | "2024-06-22T03:38:52Z" | 0 | 0 | null | [
"safetensors",
"license:mit",
"region:us"
] | null | "2024-06-22T03:37:48Z" | ---
license: mit
---
|
wanghx47/sdxl_perturbed_attention_guidance | wanghx47 | "2024-06-22T03:44:17Z" | 0 | 0 | null | [
"Diffusion Models",
"Stable Diffusion",
"Perturbed-Attention Guidance",
"PAG",
"unconditional-image-generation",
"en",
"arxiv:2403.17377",
"region:us"
] | unconditional-image-generation | "2024-06-22T03:43:23Z" | ---
language:
- en
pipeline_tag: unconditional-image-generation
tags:
- Diffusion Models
- Stable Diffusion
- Perturbed-Attention Guidance
- PAG
---
# Perturbed-Attention Guidance for SDXL
<div style="display:flex">
<video width=50% autoplay loop controls>
<source src="https://huggingface.co/multimodalart/sdxl_perturbed_attention_guidance/resolve/main/pag_sdxl.mp4" type="video/mp4">
</video>
<video width=50% autoplay loop controls>
<source src="https://huggingface.co/multimodalart/sdxl_perturbed_attention_guidance/resolve/main/pag_uncond.mp4" type="video/mp4">
</video>
</div>
The original Perturbed-Attention Guidance for unconditional models and SD1.5 by [Hyoungwon Cho](https://huggingface.co/hyoungwoncho) is availiable at [hyoungwoncho/sd_perturbed_attention_guidance](https://huggingface.co/hyoungwoncho/sd_perturbed_attention_guidance)
[Project](https://ku-cvlab.github.io/Perturbed-Attention-Guidance/) / [arXiv](https://arxiv.org/abs/2403.17377) / [GitHub](https://github.com/KU-CVLAB/Perturbed-Attention-Guidance)
This repository is just a simple SDXL implementation of the Perturbed-Attention Guidance (PAG) on Stable Diffusion XL (SDXL) for the 🧨 diffusers library.
## Quickstart
Loading Custom Pipeline:
```py
from diffusers import StableDiffusionXLPipeline
pipe = StableDiffusionXLPipeline.from_pretrained(
"stabilityai/stable-diffusion-xl-base-1.0",
custom_pipeline="multimodalart/sdxl_perturbed_attention_guidance",
torch_dtype=torch.float16
)
device="cuda"
pipe = pipe.to(device)
```
Unconditional sampling with PAG:
![image/jpeg](uncond_generation_pag.jpg)
```py
output = pipe(
"",
num_inference_steps=50,
guidance_scale=0.0,
pag_scale=5.0,
pag_applied_layers=['mid']
).images
```
Sampling with PAG and CFG:
![image/jpeg](cfgpag.jpg)
```py
output = pipe(
"the spirit of a tamagotchi wandering in the city of Vienna",
num_inference_steps=25,
guidance_scale=4.0,
pag_scale=3.0,
pag_applied_layers=['mid']
).images
```
## Parameters
`guidance_scale` : guidance scale of CFG (ex: `7.5`)
`pag_scale` : guidance scale of PAG (ex: `4.0`)
`pag_applied_layers`: layer to apply perturbation (ex: ['mid'])
`pag_applied_layers_index` : index of the layers to apply perturbation (ex: ['m0', 'm1'])
## Stable Diffusion XL Demo
[Try it here](https://huggingface.co/spaces/multimodalart/perturbed-attention-guidance-sdxl) |
Abosteet/openai-whisper-large-v3-Arabic | Abosteet | "2024-06-24T10:59:39Z" | 0 | 0 | transformers | [
"transformers",
"safetensors",
"automatic-speech-recognition",
"ar",
"dataset:mozilla-foundation/common_voice_17_0",
"arxiv:1910.09700",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] | automatic-speech-recognition | "2024-06-22T03:50:45Z" | ---
library_name: transformers
datasets:
- mozilla-foundation/common_voice_17_0
language:
- ar
pipeline_tag: automatic-speech-recognition
license: apache-2.0
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed] |
bmehrba/Meta-Llama-3-8B-Instruct-fine-tuned-adapters_Llama3_8b_rephrasetesting2_50epochs | bmehrba | "2024-06-22T03:52:49Z" | 0 | 0 | null | [
"region:us"
] | null | "2024-06-22T03:52:49Z" | Entry not found |
bmehrba/Meta-Llama-3-8B-Instruct-fine-tuned-adapters_Llama3_8b_rephrasetesting2_100epochs | bmehrba | "2024-06-22T03:57:13Z" | 0 | 0 | null | [
"region:us"
] | null | "2024-06-22T03:57:13Z" | Entry not found |
nihil117/QJab_v.03 | nihil117 | "2024-06-22T04:07:19Z" | 0 | 0 | transformers | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"text-generation-inference",
"unsloth",
"trl",
"en",
"base_model:unsloth/mistral-7b-v0.3-bnb-4bit",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-generation | "2024-06-22T04:01:33Z" | ---
base_model: unsloth/mistral-7b-v0.3-bnb-4bit
language:
- en
license: apache-2.0
tags:
- text-generation-inference
- transformers
- unsloth
- mistral
- trl
---
# Uploaded model
- **Developed by:** nihil117
- **License:** apache-2.0
- **Finetuned from model :** unsloth/mistral-7b-v0.3-bnb-4bit
This mistral model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
[<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
|
vdavidr/CodeLlama-13b-Instruct-hf_En__translations_size_104_epochs_10_2024-06-22_07-00-27_3558000 | vdavidr | "2024-06-22T09:10:43Z" | 0 | 0 | null | [
"tensorboard",
"safetensors",
"generated_from_trainer",
"base_model:codellama/CodeLlama-13b-Instruct-hf",
"license:llama2",
"region:us"
] | null | "2024-06-22T04:01:43Z" | ---
license: llama2
base_model: codellama/CodeLlama-13b-Instruct-hf
tags:
- generated_from_trainer
metrics:
- accuracy
- bleu
- sacrebleu
- rouge
model-index:
- name: CodeLlama-13b-Instruct-hf_En__translations_size_104_epochs_10_2024-06-22_07-00-27_3558000
results: []
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# CodeLlama-13b-Instruct-hf_En__translations_size_104_epochs_10_2024-06-22_07-00-27_3558000
This model is a fine-tuned version of [codellama/CodeLlama-13b-Instruct-hf](https://huggingface.co/codellama/CodeLlama-13b-Instruct-hf) on an unknown dataset.
It achieves the following results on the evaluation set:
- Loss: 1.7291
- Accuracy: 0.053
- Chrf: 0.654
- Bleu: 0.554
- Sacrebleu: 0.6
- Rouge1: 0.61
- Rouge2: 0.367
- Rougel: 0.556
- Rougelsum: 0.603
- Meteor: 0.542
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.001
- train_batch_size: 1
- eval_batch_size: 1
- seed: 3407
- distributed_type: multi-GPU
- num_devices: 4
- total_train_batch_size: 4
- total_eval_batch_size: 4
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-06
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 104
- training_steps: 1040
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy | Chrf | Bleu | Sacrebleu | Rouge1 | Rouge2 | Rougel | Rougelsum | Meteor |
|:-------------:|:-----:|:----:|:---------------:|:--------:|:-----:|:-----:|:---------:|:------:|:------:|:------:|:---------:|:------:|
| 0.2903 | 4.0 | 104 | 2.0336 | 0.052 | 0.541 | 0.417 | 0.4 | 0.524 | 0.268 | 0.48 | 0.519 | 0.477 |
| 0.2305 | 8.0 | 208 | 2.2999 | 0.05 | 0.513 | 0.412 | 0.4 | 0.495 | 0.26 | 0.454 | 0.489 | 0.453 |
| 0.1581 | 12.0 | 312 | 2.0944 | 0.054 | 0.566 | 0.475 | 0.5 | 0.548 | 0.313 | 0.502 | 0.543 | 0.52 |
| 0.6402 | 16.0 | 416 | 1.9048 | 0.05 | 0.601 | 0.482 | 0.5 | 0.568 | 0.32 | 0.521 | 0.562 | 0.544 |
| 0.1487 | 20.0 | 520 | 2.0162 | 0.054 | 0.591 | 0.487 | 0.5 | 0.567 | 0.324 | 0.516 | 0.556 | 0.472 |
| 0.1857 | 24.0 | 624 | 2.0241 | 0.055 | 0.572 | 0.481 | 0.5 | 0.545 | 0.304 | 0.499 | 0.542 | 0.509 |
| 0.718 | 28.0 | 728 | 1.7555 | 0.054 | 0.63 | 0.529 | 0.5 | 0.595 | 0.345 | 0.549 | 0.588 | 0.532 |
| 0.1329 | 32.0 | 832 | 1.7785 | 0.053 | 0.651 | 0.554 | 0.6 | 0.619 | 0.379 | 0.567 | 0.614 | 0.536 |
| 0.2488 | 36.0 | 936 | 1.7203 | 0.052 | 0.657 | 0.556 | 0.6 | 0.608 | 0.371 | 0.558 | 0.599 | 0.549 |
| 0.1337 | 40.0 | 1040 | 1.7291 | 0.053 | 0.654 | 0.554 | 0.6 | 0.61 | 0.367 | 0.556 | 0.603 | 0.542 |
### Framework versions
- Transformers 4.37.0
- Pytorch 2.2.1+cu121
- Datasets 2.20.0
- Tokenizers 0.15.2
|
mltang2/my-test-llm-michael | mltang2 | "2024-06-22T04:01:54Z" | 0 | 0 | null | [
"license:llama3",
"region:us"
] | null | "2024-06-22T04:01:54Z" | ---
license: llama3
---
|
Fudan-FUXI/llm-conditioned-diffusion-v1.0 | Fudan-FUXI | "2024-06-22T04:31:04Z" | 0 | 0 | null | [
"text-to-image",
"en",
"zh",
"license:apache-2.0",
"region:us"
] | text-to-image | "2024-06-22T04:03:44Z" | ---
license: apache-2.0
language:
- en
- zh
pipeline_tag: text-to-image
--- |
magnifi/parser_user_v8-0621-epoch7-0.002_systempromptv3_trainonly | magnifi | "2024-06-22T04:38:11Z" | 0 | 0 | transformers | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"text-generation-inference",
"unsloth",
"trl",
"conversational",
"en",
"base_model:unsloth/Phi-3-mini-4k-instruct-bnb-4bit",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-generation | "2024-06-22T04:05:09Z" | ---
language:
- en
license: apache-2.0
tags:
- text-generation-inference
- transformers
- unsloth
- mistral
- trl
base_model: unsloth/Phi-3-mini-4k-instruct-bnb-4bit
---
# Uploaded model
- **Developed by:** magnifi
- **License:** apache-2.0
- **Finetuned from model :** unsloth/Phi-3-mini-4k-instruct-bnb-4bit
This mistral model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
[<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
|
hchcsuim/batch-size16_Celeb-DF-v2_opencv-1FPS_faces-expand10-aligned_unaugmentation | hchcsuim | "2024-06-22T04:26:25Z" | 0 | 0 | transformers | [
"transformers",
"tensorboard",
"safetensors",
"swin",
"image-classification",
"generated_from_trainer",
"dataset:imagefolder",
"base_model:microsoft/swin-tiny-patch4-window7-224",
"license:apache-2.0",
"model-index",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | image-classification | "2024-06-22T04:05:14Z" | ---
license: apache-2.0
base_model: microsoft/swin-tiny-patch4-window7-224
tags:
- generated_from_trainer
datasets:
- imagefolder
metrics:
- accuracy
- precision
- recall
- f1
model-index:
- name: batch-size16_Celeb-DF-v2_opencv-1FPS_faces-expand10-aligned_unaugmentation
results:
- task:
name: Image Classification
type: image-classification
dataset:
name: imagefolder
type: imagefolder
config: default
split: test
args: default
metrics:
- name: Accuracy
type: accuracy
value: 0.9927594429960584
- name: Precision
type: precision
value: 0.9939030199104506
- name: Recall
type: recall
value: 0.9981139553636103
- name: F1
type: f1
value: 0.9960040368774208
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# batch-size16_Celeb-DF-v2_opencv-1FPS_faces-expand10-aligned_unaugmentation
This model is a fine-tuned version of [microsoft/swin-tiny-patch4-window7-224](https://huggingface.co/microsoft/swin-tiny-patch4-window7-224) on the imagefolder dataset.
It achieves the following results on the evaluation set:
- Loss: 0.0211
- Accuracy: 0.9928
- Precision: 0.9939
- Recall: 0.9981
- F1: 0.9960
- Roc Auc: 0.9990
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- gradient_accumulation_steps: 4
- total_train_batch_size: 64
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_ratio: 0.1
- num_epochs: 1
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy | Precision | Recall | F1 | Roc Auc |
|:-------------:|:------:|:----:|:---------------:|:--------:|:---------:|:------:|:------:|:-------:|
| 0.0526 | 0.9994 | 1264 | 0.0211 | 0.9928 | 0.9939 | 0.9981 | 0.9960 | 0.9990 |
### Framework versions
- Transformers 4.41.2
- Pytorch 2.3.1
- Datasets 2.20.0
- Tokenizers 0.19.1
|
fudan-generative-ai/hallo | fudan-generative-ai | "2024-06-22T05:11:24Z" | 0 | 41 | diffusers | [
"diffusers",
"onnx",
"safetensors",
"arxiv:2406.08801",
"license:mit",
"region:us"
] | null | "2024-06-22T04:05:34Z" | ---
license: mit
---
<h1 align='Center'>Hallo: Hierarchical Audio-Driven Visual Synthesis for Portrait Image Animation</h1>
<div align='Center'>
<a href='https://github.com/xumingw' target='_blank'>Mingwang Xu</a><sup>1*</sup> 
<a href='https://github.com/crystallee-ai' target='_blank'>Hui Li</a><sup>1*</sup> 
<a href='https://github.com/subazinga' target='_blank'>Qingkun Su</a><sup>1*</sup> 
<a href='https://github.com/NinoNeumann' target='_blank'>Hanlin Shang</a><sup>1</sup> 
<a href='https://github.com/AricGamma' target='_blank'>Liwei Zhang</a><sup>1</sup> 
<a href='https://github.com/cnexah' target='_blank'>Ce Liu</a><sup>3</sup> 
</div>
<div align='center'>
<a href='https://jingdongwang2017.github.io/' target='_blank'>Jingdong Wang</a><sup>2</sup> 
<a href='https://yoyo000.github.io/' target='_blank'>Yao Yao</a><sup>4</sup> 
<a href='https://sites.google.com/site/zhusiyucs/home' target='_blank'>Siyu Zhu</a><sup>1</sup> 
</div>
<div align='Center'>
<sup>1</sup>Fudan University  <sup>2</sup>Baidu Inc  <sup>3</sup>ETH Zurich  <sup>4</sup>Nanjing University
</div>
<br>
<div align='center'>
<a href='https://github.com/fudan-generative-vision/hallo'><img src='https://img.shields.io/github/stars/fudan-generative-vision/hallo?style=social'></a>
<a href='https://fudan-generative-vision.github.io/hallo/#/'><img src='https://img.shields.io/badge/Project-HomePage-Green'></a>
<a href='https://arxiv.org/pdf/2406.08801'><img src='https://img.shields.io/badge/Paper-Arxiv-red'></a>
<a href='https://huggingface.co/fudan-generative-ai/hallo'><img src='https://img.shields.io/badge/%F0%9F%A4%97%20HuggingFace-Model-yellow'></a>
<a href='assets/wechat.jpeg'><img src='https://badges.aleen42.com/src/wechat.svg'></a>
</div>
<br>
# Social Risks and Mitigations
The development of portrait image animation technologies driven by audio inputs poses social risks, such as the ethical implications of creating realistic portraits that could be misused for deepfakes. To mitigate these risks, it is crucial to establish ethical guidelines and responsible use practices. Privacy and consent concerns also arise from using individuals' images and voices. Addressing these involves transparent data usage policies, informed consent, and safeguarding privacy rights. By addressing these risks and implementing mitigations, the research aims to ensure the responsible and ethical development of this technology.
|
Coolwowsocoolwow/Morshu_11 | Coolwowsocoolwow | "2024-06-22T04:09:43Z" | 0 | 0 | null | [
"license:openrail",
"region:us"
] | null | "2024-06-22T04:07:41Z" | ---
license: openrail
---
|
bmehrba/Meta-Llama-3-8B-Instruct-fine-tuned-adapters_Llama3_8b_rephrasetesting2_20epochs | bmehrba | "2024-06-22T04:10:14Z" | 0 | 0 | null | [
"region:us"
] | null | "2024-06-22T04:10:14Z" | Entry not found |
davidyu2023/Qwen-Qwen1.5-7B-1719029444 | davidyu2023 | "2024-06-22T04:10:50Z" | 0 | 0 | peft | [
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:Qwen/Qwen1.5-7B",
"region:us"
] | null | "2024-06-22T04:10:45Z" | ---
base_model: Qwen/Qwen1.5-7B
library_name: peft
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
### Framework versions
- PEFT 0.11.1 |
davidyu2023/google-gemma-2b-1719029522 | davidyu2023 | "2024-06-22T04:12:11Z" | 0 | 0 | peft | [
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:google/gemma-2b",
"region:us"
] | null | "2024-06-22T04:12:02Z" | ---
base_model: google/gemma-2b
library_name: peft
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
### Framework versions
- PEFT 0.11.1 |
hasininawoda/3d-icon-sdxl-lora | hasininawoda | "2024-06-22T04:12:11Z" | 0 | 0 | null | [
"region:us"
] | null | "2024-06-22T04:12:11Z" | Entry not found |
bmehrba/Meta-Llama-3-8B-Instruct-fine-tuned-adapters_Llama3_8b_rephrasetesting2_30epochs | bmehrba | "2024-06-22T04:12:42Z" | 0 | 0 | null | [
"region:us"
] | null | "2024-06-22T04:12:42Z" | Entry not found |
Wawaworker/exsswimsuit | Wawaworker | "2024-06-22T04:33:22Z" | 0 | 0 | null | [
"region:us"
] | null | "2024-06-22T04:15:00Z" | Entry not found |
AdamSayyed/my_awesome_model | AdamSayyed | "2024-06-22T04:16:40Z" | 0 | 0 | null | [
"region:us"
] | null | "2024-06-22T04:16:40Z" | Entry not found |
xummer/adversarial_qa_dbert_based_on | xummer | "2024-06-22T04:37:29Z" | 0 | 0 | transformers | [
"transformers",
"safetensors",
"t5",
"text2text-generation",
"generated_from_trainer",
"base_model:google-t5/t5-large",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | text2text-generation | "2024-06-22T04:20:37Z" | ---
license: apache-2.0
base_model: google-t5/t5-large
tags:
- generated_from_trainer
metrics:
- bleu
model-index:
- name: fft-t5-large/adversarial_qa_dbert_based_on
results: []
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# fft-t5-large/adversarial_qa_dbert_based_on
This model is a fine-tuned version of [google-t5/t5-large](https://huggingface.co/google-t5/t5-large) on an unknown dataset.
It achieves the following results on the evaluation set:
- Loss: 1.1381
- Exact Match: 0.3467
- Bleu: 0.3083
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0002
- train_batch_size: 2
- eval_batch_size: 16
- seed: 42
- distributed_type: multi-GPU
- num_devices: 4
- total_train_batch_size: 8
- total_eval_batch_size: 64
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 5.0
### Training results
| Training Loss | Epoch | Step | Validation Loss | Exact Match | Bleu |
|:-------------:|:-----:|:----:|:---------------:|:-----------:|:------:|
| 1.0162 | 1.0 | 63 | 0.7607 | 0.2754 | 0.2749 |
| 0.3929 | 2.0 | 126 | 0.7943 | 0.2959 | 0.2412 |
| 0.1542 | 3.0 | 189 | 1.0053 | 0.3018 | 0.2720 |
| 0.0544 | 4.0 | 252 | 1.1005 | 0.3457 | 0.3185 |
| 0.0239 | 5.0 | 315 | 1.1381 | 0.3467 | 0.3083 |
### Framework versions
- Transformers 4.41.2
- Pytorch 2.2.1+cu121
- Datasets 2.20.0
- Tokenizers 0.19.1
|
RajuEEE/GPT2_FineTunedModel_GPTData_ThreeLabel | RajuEEE | "2024-06-22T04:22:16Z" | 0 | 0 | transformers | [
"transformers",
"safetensors",
"arxiv:1910.09700",
"endpoints_compatible",
"region:us"
] | null | "2024-06-22T04:22:09Z" | ---
library_name: transformers
tags: []
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed] |
synap-du/kocamel-13b-inst | synap-du | "2024-06-22T04:22:15Z" | 0 | 0 | null | [
"region:us"
] | null | "2024-06-22T04:22:15Z" | Entry not found |
arabic-translation-prompt-engineering/atpe-notebooks | arabic-translation-prompt-engineering | "2024-06-22T12:14:42Z" | 0 | 0 | null | [
"license:openrail",
"region:us"
] | null | "2024-06-22T04:24:56Z" | ---
license: openrail
---
|
nishantup/example-model | nishantup | "2024-06-22T04:31:34Z" | 0 | 0 | null | [
"license:mit",
"region:us"
] | null | "2024-06-22T04:31:34Z" | ---
license: mit
---
|
geraldabrhm/llama-3-8b-regulardataset-simplecontext-32lora | geraldabrhm | "2024-06-22T05:26:04Z" | 0 | 0 | null | [
"safetensors",
"region:us"
] | null | "2024-06-22T04:33:28Z" | Entry not found |
01N02/StudioVip | 01N02 | "2024-06-22T04:34:27Z" | 0 | 0 | null | [
"license:unknown",
"region:us"
] | null | "2024-06-22T04:34:27Z" | ---
license: unknown
---
|
samrocksc/learngod | samrocksc | "2024-06-22T04:53:25Z" | 0 | 0 | null | [
"token-classification",
"en",
"dataset:OpenGVLab/ShareGPT-4o",
"license:mit",
"region:us"
] | token-classification | "2024-06-22T04:35:32Z" | ---
license: mit
datasets:
- OpenGVLab/ShareGPT-4o
language:
- en
metrics:
- accuracy
- character
pipeline_tag: token-classification
--- |
ParZiVal04/lora_model_1 | ParZiVal04 | "2024-06-22T04:39:13Z" | 0 | 0 | transformers | [
"transformers",
"safetensors",
"text-generation-inference",
"unsloth",
"llama",
"trl",
"en",
"base_model:unsloth/llama-3-8b-bnb-4bit",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] | null | "2024-06-22T04:38:53Z" | ---
base_model: unsloth/llama-3-8b-bnb-4bit
language:
- en
license: apache-2.0
tags:
- text-generation-inference
- transformers
- unsloth
- llama
- trl
---
# Uploaded model
- **Developed by:** ParZiVal04
- **License:** apache-2.0
- **Finetuned from model :** unsloth/llama-3-8b-bnb-4bit
This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
[<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
|
klandtech/kland_name3 | klandtech | "2024-06-22T04:39:26Z" | 0 | 0 | null | [
"region:us"
] | null | "2024-06-22T04:39:26Z" | Entry not found |
cxfajar197/bert-base-multilingual-cased-finetuned-imdb | cxfajar197 | "2024-06-22T14:42:54Z" | 0 | 0 | transformers | [
"transformers",
"pytorch",
"tensorboard",
"bert",
"fill-mask",
"generated_from_trainer",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | fill-mask | "2024-06-22T04:40:49Z" | ---
license: apache-2.0
tags:
- generated_from_trainer
model-index:
- name: bert-base-multilingual-cased-finetuned-imdb
results: []
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# bert-base-multilingual-cased-finetuned-imdb
This model is a fine-tuned version of [bert-base-multilingual-cased](https://huggingface.co/bert-base-multilingual-cased) on an unknown dataset.
It achieves the following results on the evaluation set:
- Loss: 2.2852
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3.0
- mixed_precision_training: Native AMP
### Training results
| Training Loss | Epoch | Step | Validation Loss |
|:-------------:|:-----:|:----:|:---------------:|
| 1.3958 | 1.0 | 50 | 2.2363 |
| 1.5939 | 2.0 | 100 | 1.9750 |
| 1.8532 | 3.0 | 150 | 2.1309 |
### Framework versions
- Transformers 4.20.1
- Pytorch 1.11.0
- Datasets 2.1.0
- Tokenizers 0.12.1
|
shiromiya/program | shiromiya | "2024-06-22T04:51:08Z" | 0 | 0 | null | [
"region:us"
] | null | "2024-06-22T04:47:56Z" | Entry not found |
xummer/adversarial_qa_dbert_answer_the_following_q | xummer | "2024-06-22T05:23:34Z" | 0 | 0 | transformers | [
"transformers",
"safetensors",
"t5",
"text2text-generation",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | text2text-generation | "2024-06-22T04:49:01Z" | Entry not found |
Dahepix/models | Dahepix | "2024-06-22T05:00:34Z" | 0 | 0 | null | [
"region:us"
] | null | "2024-06-22T04:50:16Z" | Entry not found |
habberrih/olive-branch | habberrih | "2024-06-22T04:50:57Z" | 0 | 0 | null | [
"license:apache-2.0",
"region:us"
] | null | "2024-06-22T04:50:57Z" | ---
license: apache-2.0
---
|
ankunjin/test | ankunjin | "2024-06-24T02:20:01Z" | 0 | 0 | peft | [
"peft",
"arxiv:1910.09700",
"base_model:beomi/KoAlpaca-Polyglot-5.8B",
"region:us"
] | null | "2024-06-22T04:52:19Z" | ---
base_model: beomi/KoAlpaca-Polyglot-5.8B
library_name: peft
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
### Framework versions
- PEFT 0.11.1 |
hchcsuim/batch-size16_FFPP-raw_opencv-1FPS_unaugmentation | hchcsuim | "2024-06-22T06:11:40Z" | 0 | 0 | transformers | [
"transformers",
"tensorboard",
"safetensors",
"swin",
"image-classification",
"generated_from_trainer",
"dataset:imagefolder",
"base_model:microsoft/swin-tiny-patch4-window7-224",
"license:apache-2.0",
"model-index",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | image-classification | "2024-06-22T04:53:24Z" | ---
license: apache-2.0
base_model: microsoft/swin-tiny-patch4-window7-224
tags:
- generated_from_trainer
datasets:
- imagefolder
metrics:
- accuracy
- precision
- recall
- f1
model-index:
- name: batch-size16_FFPP-raw_opencv-1FPS_unaugmentation
results:
- task:
name: Image Classification
type: image-classification
dataset:
name: imagefolder
type: imagefolder
config: default
split: test
args: default
metrics:
- name: Accuracy
type: accuracy
value: 0.9033487951125693
- name: Precision
type: precision
value: 0.8971202577231833
- name: Recall
type: recall
value: 0.990071106486299
- name: F1
type: f1
value: 0.9413066030930314
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# batch-size16_FFPP-raw_opencv-1FPS_unaugmentation
This model is a fine-tuned version of [microsoft/swin-tiny-patch4-window7-224](https://huggingface.co/microsoft/swin-tiny-patch4-window7-224) on the imagefolder dataset.
It achieves the following results on the evaluation set:
- Loss: 0.2388
- Accuracy: 0.9033
- Precision: 0.8971
- Recall: 0.9901
- F1: 0.9413
- Roc Auc: 0.9623
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- gradient_accumulation_steps: 4
- total_train_batch_size: 64
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_ratio: 0.1
- num_epochs: 1
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy | Precision | Recall | F1 | Roc Auc |
|:-------------:|:------:|:----:|:---------------:|:--------:|:---------:|:------:|:------:|:-------:|
| 0.2365 | 0.9998 | 1381 | 0.2388 | 0.9033 | 0.8971 | 0.9901 | 0.9413 | 0.9623 |
### Framework versions
- Transformers 4.41.2
- Pytorch 2.3.1
- Datasets 2.20.0
- Tokenizers 0.19.1
|
Hanbitchan/CoreMLModels | Hanbitchan | "2024-06-22T04:56:48Z" | 0 | 0 | null | [
"region:us"
] | null | "2024-06-22T04:54:09Z" | Entry not found |
Ketansomewhere/sd-naruto-model | Ketansomewhere | "2024-06-22T04:57:23Z" | 0 | 0 | null | [
"region:us"
] | null | "2024-06-22T04:57:23Z" | Entry not found |
HyperdustProtocol/HyperAuto-cog-llama2-7b-4555 | HyperdustProtocol | "2024-06-22T04:59:52Z" | 0 | 0 | transformers | [
"transformers",
"safetensors",
"text-generation-inference",
"unsloth",
"llama",
"trl",
"en",
"base_model:unsloth/llama-2-7b-bnb-4bit",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] | null | "2024-06-22T04:59:38Z" | ---
base_model: unsloth/llama-2-7b-bnb-4bit
language:
- en
license: apache-2.0
tags:
- text-generation-inference
- transformers
- unsloth
- llama
- trl
---
# Uploaded model
- **Developed by:** HyperdustProtocol
- **License:** apache-2.0
- **Finetuned from model :** unsloth/llama-2-7b-bnb-4bit
This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
[<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
|
bmehrba/Meta-Llama-3-8B-Instruct-fine-tuned-adapters_Llama3_8b_rephrasetesting2_200epochs | bmehrba | "2024-06-22T05:05:59Z" | 0 | 0 | null | [
"region:us"
] | null | "2024-06-22T05:05:59Z" | Entry not found |
frankperdomo1947/Videocool | frankperdomo1947 | "2024-06-22T05:14:46Z" | 0 | 0 | null | [
"license:apache-2.0",
"region:us"
] | null | "2024-06-22T05:14:46Z" | ---
license: apache-2.0
---
|
AliGhiasvand86/last_checkpoint | AliGhiasvand86 | "2024-06-22T05:19:40Z" | 0 | 0 | transformers | [
"transformers",
"safetensors",
"longt5",
"text2text-generation",
"license:mit",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text2text-generation | "2024-06-22T05:19:06Z" | ---
license: mit
---
|
sosnaji/mohammd | sosnaji | "2024-06-22T05:20:42Z" | 0 | 0 | null | [
"region:us"
] | null | "2024-06-22T05:20:42Z" | Entry not found |
wallaceblaia/wisper-icm-17v | wallaceblaia | "2024-06-22T05:25:04Z" | 0 | 0 | null | [
"region:us"
] | null | "2024-06-22T05:25:04Z" | Entry not found |
hiudy/Stella | hiudy | "2024-06-22T05:26:03Z" | 0 | 0 | null | [
"license:other",
"region:us"
] | null | "2024-06-22T05:26:03Z" | ---
license: other
license_name: nzn
license_link: https://painel.nodzhost.com.br/license
---
|
Imran1/embadding | Imran1 | "2024-06-22T05:47:47Z" | 0 | 0 | transformers | [
"transformers",
"safetensors",
"bert",
"license:mit",
"endpoints_compatible",
"region:us"
] | null | "2024-06-22T05:26:59Z" | ---
license: mit
---
# Model using
```python
from transformers import AutoConfig, AutoTokenizer
from torch import nn
import torch.nn.functional as F
import torch
# First, define your custom model class again
class HFCustomBertModel(nn.Module):
def __init__(self, config):
super().__init__()
self.bert = BertModel(config)
self.pooler = nn.Sequential(
nn.Linear(config.hidden_size, config.hidden_size),
nn.Tanh()
)
def forward(self, input_ids, attention_mask=None, token_type_ids=None):
outputs = self.bert(input_ids, attention_mask=attention_mask, token_type_ids=token_type_ids)
pooled_output = self.pooler(outputs.pooler_output)
return pooled_output
def load_custom_model_and_tokenizer(model_path):
# Load the config
config = AutoConfig.from_pretrained(model_path)
# Initialize the custom model with the config
model = HFCustomBertModel(config)
# Load the tokenizer
tokenizer = AutoTokenizer.from_pretrained(model_path)
return model, tokenizer
# Usage
model_path = "Imran1/embadding"
model, tokenizer = load_custom_model_and_tokenizer(model_path)
queries = ["how much protein should a female eat"]
documents = ["As a general guideline, the CDC's average requirement of protein for women ages 19 to 70 is 46 grams per day. But, as you can see from this chart, you'll need to increase that if you're expecting or training for a marathon. Check out the chart below to see how much protein you should be eating each day."]
model.eval() # Set the model to evaluation mode
with torch.no_grad():
# Tokenize and encode the queries and documents
query_inputs = tokenizer(queries, padding=True, truncation=True, return_tensors="pt")
document_inputs = tokenizer(documents, padding=True, truncation=True, return_tensors="pt")
# Get embeddings
query_embeddings = model(**query_inputs)
document_embeddings = model(**document_inputs)
# Normalize embeddings
query_embeddings = F.normalize(query_embeddings, p=2, dim=1)
document_embeddings = F.normalize(document_embeddings, p=2, dim=1)
# Calculate cosine similarity
scores = torch.matmul(query_embeddings, document_embeddings.transpose(0, 1))
print(f"Similarity score: {scores.item():.4f}")
Similarity score: 0.9605
``` |
Zzainy/Z.comic | Zzainy | "2024-06-22T05:27:51Z" | 0 | 0 | null | [
"license:apache-2.0",
"region:us"
] | null | "2024-06-22T05:27:50Z" | ---
license: apache-2.0
---
|
bmehrba/Llama-2-7b-chat-hf-fine-tuned-adapters_Llama3_8b_rephrasetesting2_150epochs | bmehrba | "2024-06-22T05:29:06Z" | 0 | 0 | null | [
"region:us"
] | null | "2024-06-22T05:29:06Z" | Entry not found |
v0dkapapi/falcon-7b-sharded-bf16-finetuned-mental-health-conversational | v0dkapapi | "2024-06-26T09:43:08Z" | 0 | 0 | null | [
"generated_from_trainer",
"base_model:ybelkada/falcon-7b-sharded-bf16",
"region:us"
] | null | "2024-06-22T05:32:12Z" | ---
base_model: ybelkada/falcon-7b-sharded-bf16
tags:
- generated_from_trainer
model-index:
- name: falcon-7b-sharded-bf16-finetuned-mental-health-conversational
results: []
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# falcon-7b-sharded-bf16-finetuned-mental-health-conversational
This model is a fine-tuned version of [ybelkada/falcon-7b-sharded-bf16](https://huggingface.co/ybelkada/falcon-7b-sharded-bf16) on an unknown dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0002
- train_batch_size: 16
- eval_batch_size: 8
- seed: 42
- gradient_accumulation_steps: 4
- total_train_batch_size: 64
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: cosine
- lr_scheduler_warmup_ratio: 0.03
- training_steps: 50
### Training results
### Framework versions
- Transformers 4.32.0
- Pytorch 2.3.0+cu121
- Datasets 2.13.1
- Tokenizers 0.13.3
|
RajuEEE/LLama_FineTunedModel_GPTEDataThreeLabel | RajuEEE | "2024-06-22T05:44:02Z" | 0 | 0 | transformers | [
"transformers",
"safetensors",
"arxiv:1910.09700",
"endpoints_compatible",
"region:us"
] | null | "2024-06-22T05:43:56Z" | ---
library_name: transformers
tags: []
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed] |
vdavidr/CodeLlama-7b-Instruct-hf_En__translations_size_104_epochs_10_2024-06-22_08-44-43_3558001 | vdavidr | "2024-06-22T08:34:18Z" | 0 | 0 | null | [
"tensorboard",
"safetensors",
"region:us"
] | null | "2024-06-22T05:45:29Z" | Entry not found |
VKapseln475/Nexalyn59 | VKapseln475 | "2024-06-22T05:48:09Z" | 0 | 0 | null | [
"region:us"
] | null | "2024-06-22T05:45:38Z" | # Nexalyn Danmark Anmeldelser - Nexalyn Dose & Intake Works Officiel hjemmeside Køb nu
Nexalyn Anmeldelser Dosis og Virker At forbrænde fedt i besværlige områder er en udfordring for mange mennesker på deres vægttabsrejse. Dette stædige kropsfedt kan være frustrerende og svært at målrette mod med kost og motion alene. Nexaslim-tillægget kan dog give den løsning, du har ledt efter.
## **[Klik her for at købe nu fra Nexalyns officielle hjemmeside](https://capsules24x7.com/nexalyn-danmark)**
## Fordele du får ved at bruge Nexalyn Me Supplement Piller - Forventede resultater
### Øget libido
En øget libido er resultatet af Nexalyns forbedring af ophidselse og lyst. Der kan opstå en mærkbar stigning i brugernes ønske om nærhed, hvilket fører til større iver og entusiasme for at have samleje med deres partnere.
### Forbedret erektil funktion
Med Nexalyn kan brugerne forvente erektioner, der er stærkere, mere komplekse og varer længere. I soveværelset fører denne forbedring af erektil funktion til mere tilfredsstillende og behagelige romantiske møder, som øger selvtilliden og selvværdet.
### Forbedret virilitet
Brugere af Nexalyn føler sig mere magtfulde og maskuline, da stoffet tilskynder til større virilitet. Nexalyn hjælper mænd med at projicere selvtillid og styrke under intime interaktioner ved at fremme hormonbalancen og forbedre den romantiske præstation.
### Mere intense orgasmer
Nexalyn kan få brugere til at få mere potente og intense orgasmer. Tilskuddet arbejder for at øge følelsen af følsomhed og nydelse under samleje, hvilket resulterer i forbedrede oplevelser og øget klimakstilfredshed.
### Øget energiniveau
Nexalyn giver kunderne øget energi, så de kan blive sent oppe. Forbedret udholdenhed og udholdenhed gør det muligt for brugerne at deltage i længerevarende romantiske aktiviteter uden at opleve træthed eller udmattelse, hvilket letter mere behagelige og givende møder.
## Hvem er Nexalyn for? Kan kvinder også bruge Nexalyn?
Den primære målgruppe for Nexalyn er mænd, der ønsker at forbedre deres præstationer, sundhed og præstationer. Mænd i alle aldre, der kan have problemer såsom erektil dysfunktion, dårlig libido eller nedsat romantisk tilfredshed, bør tage dette vitamin. Nexalyn er en sikker og effektiv måde at forbedre mænds fysiske sundhed på, uanset dine mål – forbedre kvaliteten af din erektion, øge din lyst eller øge din generelle romantiske vitalitet.
Mænd, der ønsker at føle sig mere tilfredse og selvsikre under samleje, kan opleve, at Nexalyn hjælper. Nexalyn kan få dig til at føle dig mere selvsikker og tilfreds i dine intime forhold, uanset om du er langtidsdating med nogen eller lige er startet. Mænd kan opleve mere tilfredsstillende og glædelige romantiske møder med Nexalyn, hvilket øger nærhed og lykke i forhold ved at fremme mange facetter af mænds intime sundhed.
## **[Klik her for at købe nu fra Nexalyns officielle hjemmeside](https://capsules24x7.com/nexalyn-danmark)** |
Krish778/gpt-neo-2.7B-custom | Krish778 | "2024-06-22T05:46:17Z" | 0 | 0 | null | [
"region:us"
] | null | "2024-06-22T05:46:17Z" | Entry not found |
Seikaijyu/RWKV-x060-World-3B-v2.1-Chat-Anthropomorphic | Seikaijyu | "2024-06-22T07:23:24Z" | 0 | 1 | null | [
"zh",
"license:mit",
"region:us"
] | null | "2024-06-22T05:51:58Z" | ---
license: mit
language:
- zh
---
### 模型说明
#### 基于RWKV6-v2.1-3B基底模型微调的进行pissa微调的chat模型,此模型开箱即用,无需过多准备
#### 使用ChatGLM4基于[优美的中国话](https://huggingface.co/datasets/Seikaijyu/Beautiful-Chinese)语料进行异构,使单轮语料成为多轮语料,并进行了大量清洗后得到了2632条多轮对话数据集后训练的模型
#### 此模型仅微调了Assistant对话,无任何额外数据微调,但是经过测试,语料对其它角色有一定的泛化作用
#### 效果如下:
![image/png](https://cdn-uploads.huggingface.co/production/uploads/6417b108b03817ada6444bb8/KyPxh5yjQszaEgDEc-DXO.png)
![image/png](https://cdn-uploads.huggingface.co/production/uploads/6417b108b03817ada6444bb8/eoLDGEX6xqyWzU2oswLkp.png)
![image/png](https://cdn-uploads.huggingface.co/production/uploads/6417b108b03817ada6444bb8/Q5FmKi3iLK5vZF7ykFZZF.png)
推荐参数如下:
##### Temperature=1-3之间
##### Top_P=0.55-0.65之间
##### Presence Penalty=0.4-0之间
##### Frequency Penalty=0.4-1.2之间 |
hussnainahmedsaqib/xlnet_ner_res_pipeline | hussnainahmedsaqib | "2024-06-22T05:53:13Z" | 0 | 0 | null | [
"region:us"
] | null | "2024-06-22T05:53:13Z" | Entry not found |
Minutor/new-dummy-model | Minutor | "2024-06-22T06:02:14Z" | 0 | 0 | null | [
"region:us"
] | null | "2024-06-22T05:53:45Z" | Entry not found |
ymoslem/whisper-medium-ga2en-v6.3.2-15k-r | ymoslem | "2024-06-22T14:09:06Z" | 0 | 1 | transformers | [
"transformers",
"tensorboard",
"safetensors",
"whisper",
"automatic-speech-recognition",
"generated_from_trainer",
"ga",
"en",
"dataset:ymoslem/IWSLT2023-GA-EN",
"dataset:ymoslem/FLEURS-GA-EN",
"dataset:ymoslem/BitesizeIrish-GA-EN",
"dataset:ymoslem/SpokenWords-GA-EN-MTed",
"dataset:ymoslem/Tatoeba-Speech-Irish",
"dataset:ymoslem/Wikimedia-Speech-Irish",
"dataset:ymoslem/EUbookshop-Speech-Irish",
"base_model:openai/whisper-medium",
"license:apache-2.0",
"model-index",
"endpoints_compatible",
"region:us"
] | automatic-speech-recognition | "2024-06-22T05:54:18Z" | ---
language:
- ga
- en
license: apache-2.0
base_model: openai/whisper-medium
tags:
- generated_from_trainer
datasets:
- ymoslem/IWSLT2023-GA-EN
- ymoslem/FLEURS-GA-EN
- ymoslem/BitesizeIrish-GA-EN
- ymoslem/SpokenWords-GA-EN-MTed
- ymoslem/Tatoeba-Speech-Irish
- ymoslem/Wikimedia-Speech-Irish
- ymoslem/EUbookshop-Speech-Irish
metrics:
- bleu
- wer
model-index:
- name: Whisper Medium GA-EN Speech Translation
results:
- task:
name: Automatic Speech Recognition
type: automatic-speech-recognition
dataset:
name: IWSLT-2023, FLEURS, BiteSize, SpokenWords, Tatoeba, Wikimedia, and EUbookshop
type: ymoslem/IWSLT2023-GA-EN
metrics:
- name: Bleu
type: bleu
value: 34.85
- name: Wer
type: wer
value: 60.91850517784781
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# Whisper Medium GA-EN Speech Translation
This model is a fine-tuned version of [openai/whisper-medium](https://huggingface.co/openai/whisper-medium) on the IWSLT-2023, FLEURS, BiteSize, SpokenWords, Tatoeba, Wikimedia, and EUbookshop dataset.
It achieves the following results on the evaluation set:
- Loss: 1.2038
- Bleu: 34.85
- Chrf: 54.43
- Wer: 60.9185
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0001
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_ratio: 0.03
- training_steps: 15000
- mixed_precision_training: Native AMP
### Training results
| Training Loss | Epoch | Step | Bleu | Chrf | Validation Loss | Wer |
|:-------------:|:------:|:-----:|:-----:|:-----:|:---------------:|:--------:|
| 2.5219 | 0.0138 | 100 | 0.44 | 10.48 | 2.1106 | 107.2490 |
| 2.4608 | 0.0276 | 200 | 3.3 | 20.43 | 2.1816 | 179.1535 |
| 2.3008 | 0.0414 | 300 | 3.66 | 21.59 | 2.0587 | 206.4836 |
| 2.2095 | 0.0552 | 400 | 8.79 | 27.66 | 1.9459 | 100.3602 |
| 2.0454 | 0.0690 | 500 | 8.14 | 27.36 | 1.8681 | 122.1522 |
| 1.9937 | 0.0828 | 600 | 11.05 | 30.26 | 1.8717 | 97.2535 |
| 1.868 | 0.0966 | 700 | 9.14 | 29.03 | 1.7917 | 129.0410 |
| 1.9924 | 0.1103 | 800 | 12.62 | 33.2 | 1.7170 | 89.6443 |
| 1.8646 | 0.1241 | 900 | 11.98 | 30.77 | 1.7252 | 97.8838 |
| 1.7644 | 0.1379 | 1000 | 10.87 | 31.0 | 1.6832 | 109.1851 |
| 1.692 | 0.1517 | 1100 | 13.05 | 34.46 | 1.6837 | 93.3814 |
| 1.7044 | 0.1655 | 1200 | 20.95 | 37.42 | 1.5527 | 75.2364 |
| 1.6824 | 0.1793 | 1300 | 14.91 | 35.56 | 1.5611 | 92.6159 |
| 1.6557 | 0.1931 | 1400 | 14.0 | 36.54 | 1.5554 | 99.8199 |
| 1.5456 | 0.2069 | 1500 | 19.72 | 39.81 | 1.5058 | 83.5660 |
| 1.3755 | 0.2207 | 1600 | 18.04 | 37.95 | 1.5039 | 82.9806 |
| 1.3959 | 0.2345 | 1700 | 17.01 | 39.5 | 1.4374 | 85.2319 |
| 1.5012 | 0.2483 | 1800 | 14.93 | 39.24 | 1.4242 | 114.4079 |
| 1.4278 | 0.2621 | 1900 | 23.85 | 42.69 | 1.3904 | 73.0302 |
| 1.3285 | 0.2759 | 2000 | 17.7 | 37.23 | 1.4493 | 83.8811 |
| 1.2655 | 0.2897 | 2100 | 20.1 | 40.32 | 1.3661 | 79.7839 |
| 1.2074 | 0.3034 | 2200 | 24.45 | 43.79 | 1.3387 | 72.9851 |
| 1.1893 | 0.3172 | 2300 | 21.45 | 42.61 | 1.3308 | 82.3953 |
| 1.1236 | 0.3310 | 2400 | 22.77 | 44.17 | 1.3050 | 77.3075 |
| 1.0934 | 0.3448 | 2500 | 25.54 | 46.32 | 1.2793 | 72.2647 |
| 1.06 | 0.3586 | 2600 | 28.27 | 47.32 | 1.2396 | 65.6911 |
| 1.0327 | 0.3724 | 2700 | 28.45 | 47.01 | 1.2577 | 67.3570 |
| 1.1623 | 0.3862 | 2800 | 24.54 | 47.43 | 1.2194 | 73.6155 |
| 1.0215 | 0.4 | 2900 | 27.4 | 49.6 | 1.2039 | 69.2481 |
| 0.9185 | 0.4138 | 3000 | 27.04 | 49.24 | 1.1724 | 67.8973 |
| 0.9003 | 0.4276 | 3100 | 31.08 | 50.11 | 1.1674 | 63.8001 |
| 0.9839 | 0.4414 | 3200 | 30.24 | 50.63 | 1.1580 | 64.5655 |
| 0.9396 | 0.4552 | 3300 | 30.79 | 51.72 | 1.1202 | 64.9257 |
| 0.9051 | 0.4690 | 3400 | 30.34 | 53.08 | 1.1180 | 66.4566 |
| 0.8621 | 0.4828 | 3500 | 33.3 | 53.86 | 1.1042 | 60.7834 |
| 0.8236 | 0.4966 | 3600 | 32.77 | 53.21 | 1.1070 | 62.0441 |
| 0.829 | 0.5103 | 3700 | 32.49 | 54.21 | 1.0771 | 62.5844 |
| 0.8375 | 0.5241 | 3800 | 32.27 | 53.98 | 1.0780 | 63.0797 |
| 0.8206 | 0.5379 | 3900 | 33.26 | 55.07 | 1.0615 | 61.6389 |
| 0.8059 | 0.5517 | 4000 | 33.24 | 55.16 | 1.0552 | 61.5038 |
| 0.9133 | 0.5655 | 4100 | 29.38 | 49.22 | 1.2218 | 66.0964 |
| 1.051 | 0.5793 | 4200 | 25.12 | 46.01 | 1.2304 | 71.8145 |
| 0.954 | 0.5931 | 4300 | 25.47 | 45.88 | 1.2501 | 75.3715 |
| 0.939 | 0.6069 | 4400 | 29.19 | 47.63 | 1.2204 | 66.9068 |
| 0.9887 | 0.6207 | 4500 | 27.99 | 47.01 | 1.2099 | 67.7172 |
| 1.0044 | 0.6345 | 4600 | 23.77 | 45.33 | 1.2080 | 73.3904 |
| 0.9881 | 0.6483 | 4700 | 26.46 | 47.36 | 1.2188 | 68.5277 |
| 0.9674 | 0.6621 | 4800 | 26.11 | 45.92 | 1.2296 | 68.3026 |
| 0.8845 | 0.6759 | 4900 | 27.3 | 46.08 | 1.2347 | 68.0324 |
| 0.8297 | 0.6897 | 5000 | 29.48 | 48.96 | 1.2108 | 64.6105 |
| 0.9065 | 0.7034 | 5100 | 29.81 | 49.94 | 1.1873 | 64.2503 |
| 0.8096 | 0.7172 | 5200 | 28.5 | 46.93 | 1.2122 | 66.2314 |
| 0.8077 | 0.7310 | 5300 | 29.26 | 48.21 | 1.1945 | 64.4755 |
| 0.8227 | 0.7448 | 5400 | 26.82 | 48.43 | 1.2310 | 71.4093 |
| 0.7587 | 0.7586 | 5500 | 29.45 | 49.03 | 1.2067 | 65.3309 |
| 0.7206 | 0.7724 | 5600 | 29.89 | 49.33 | 1.2114 | 65.5561 |
| 0.8088 | 0.7862 | 5700 | 31.88 | 51.4 | 1.1689 | 64.2954 |
| 0.693 | 0.8 | 5800 | 27.23 | 48.11 | 1.1644 | 68.7078 |
| 0.7099 | 0.8138 | 5900 | 31.01 | 49.42 | 1.1852 | 63.3949 |
| 0.7564 | 0.8276 | 6000 | 28.3 | 50.34 | 1.1554 | 71.0941 |
| 0.584 | 0.8414 | 6100 | 34.79 | 51.69 | 1.1566 | 59.0725 |
| 0.6817 | 0.8552 | 6200 | 34.08 | 51.95 | 1.1245 | 59.8829 |
| 0.5968 | 0.8690 | 6300 | 32.4 | 51.59 | 1.1475 | 62.9896 |
| 0.6092 | 0.8828 | 6400 | 32.83 | 52.82 | 1.1250 | 62.5844 |
| 0.6325 | 0.8966 | 6500 | 29.29 | 51.68 | 1.1108 | 69.1130 |
| 0.6002 | 0.9103 | 6600 | 27.64 | 52.7 | 1.0993 | 71.0941 |
| 0.6247 | 0.9241 | 6700 | 28.39 | 52.4 | 1.0898 | 68.3026 |
| 0.6257 | 0.9379 | 6800 | 28.54 | 52.33 | 1.0863 | 70.9140 |
| 0.6719 | 0.9517 | 6900 | 31.43 | 53.53 | 1.0891 | 66.1414 |
| 0.4994 | 0.9655 | 7000 | 33.81 | 52.77 | 1.1066 | 61.0986 |
| 0.5469 | 0.9793 | 7100 | 30.52 | 53.13 | 1.0891 | 67.3570 |
| 0.6031 | 0.9931 | 7200 | 33.16 | 54.03 | 1.0933 | 62.1792 |
| 0.2469 | 1.0069 | 7300 | 33.76 | 52.38 | 1.1426 | 62.8546 |
| 0.2572 | 1.0207 | 7400 | 33.16 | 51.71 | 1.1292 | 64.8807 |
| 0.2762 | 1.0345 | 7500 | 34.76 | 54.28 | 1.1090 | 60.7384 |
| 0.2332 | 1.0483 | 7600 | 30.95 | 52.28 | 1.1073 | 66.1864 |
| 0.2069 | 1.0621 | 7700 | 32.39 | 53.08 | 1.0999 | 65.5561 |
| 0.2417 | 1.0759 | 7800 | 31.3 | 53.87 | 1.1008 | 65.1058 |
| 0.2403 | 1.0897 | 7900 | 32.18 | 53.3 | 1.1053 | 66.4566 |
| 0.208 | 1.1034 | 8000 | 32.0 | 52.48 | 1.1067 | 66.7717 |
| 0.3328 | 1.1172 | 8100 | 28.92 | 49.12 | 1.2137 | 68.4376 |
| 0.4045 | 1.1310 | 8200 | 28.47 | 51.53 | 1.2165 | 68.3926 |
| 0.4175 | 1.1448 | 8300 | 26.88 | 47.57 | 1.2790 | 74.5160 |
| 0.3976 | 1.1586 | 8400 | 21.56 | 44.64 | 1.3060 | 84.1513 |
| 0.4026 | 1.1724 | 8500 | 25.22 | 47.73 | 1.2476 | 73.1202 |
| 0.4088 | 1.1862 | 8600 | 26.03 | 48.08 | 1.2387 | 72.8050 |
| 0.4245 | 1.2 | 8700 | 29.8 | 49.69 | 1.2136 | 67.4021 |
| 0.4083 | 1.2138 | 8800 | 26.26 | 48.23 | 1.2784 | 73.4804 |
| 0.3832 | 1.2276 | 8900 | 29.06 | 49.36 | 1.2527 | 66.4115 |
| 0.4335 | 1.2414 | 9000 | 30.11 | 49.24 | 1.2772 | 67.2670 |
| 0.4056 | 1.2552 | 9100 | 32.51 | 50.18 | 1.3013 | 63.3048 |
| 0.3877 | 1.2690 | 9200 | 26.91 | 47.47 | 1.2897 | 71.5894 |
| 0.3787 | 1.2828 | 9300 | 1.2430| 30.16 | 50.61 | 65.1058 |
| 0.3947 | 1.2966 | 9400 | 1.2318| 29.9 | 50.77 | 66.0964 |
| 0.3908 | 1.3103 | 9500 | 1.1927| 30.7 | 51.62 | 64.6105 |
| 0.405 | 1.3241 | 9600 | 1.2249| 26.56 | 49.05 | 71.7695 |
| 0.3847 | 1.3379 | 9700 | 1.2105| 33.22 | 51.98 | 61.8640 |
| 0.3674 | 1.3517 | 9800 | 1.2545| 30.93 | 50.34 | 65.6011 |
| 0.3642 | 1.3655 | 9900 | 1.2443| 25.23 | 47.97 | 77.9379 |
| 0.3636 | 1.3793 | 10000 | 1.2796| 26.78 | 48.07 | 73.6155 |
| 0.329 | 1.3931 | 10100 | 1.2373| 29.06 | 49.55 | 66.4566 |
| 0.4195 | 1.4069 | 10200 | 1.2187| 29.11 | 50.65 | 66.2314 |
| 0.4244 | 1.4207 | 10300 | 1.2346| 27.97 | 49.86 | 69.0680 |
| 0.3338 | 1.4345 | 10400 | 1.2239| 29.96 | 50.45 | 66.0063 |
| 0.3401 | 1.4483 | 10500 | 1.2501| 29.84 | 51.0 | 65.6911 |
| 0.3792 | 1.4621 | 10600 | 1.2353| 28.38 | 49.19 | 69.1130 |
| 0.3549 | 1.4759 | 10700 | 1.2178| 28.63 | 49.73 | 68.5727 |
| 0.3326 | 1.4897 | 10800 | 1.1936| 29.57 | 51.1 | 64.4755 |
| 0.3418 | 1.5034 | 10900 | 1.1741| 33.06 | 52.86 | 60.9185 |
| 0.3143 | 1.5172 | 11000 | 1.2046| 31.49 | 50.4 | 63.5750 |
| 0.3245 | 1.5310 | 11100 | 1.2145| 30.9 | 50.17 | 64.6105 |
| 0.3268 | 1.5448 | 11200 | 1.2119| 33.5 | 53.0 | 60.2431 |
| 0.2894 | 1.5586 | 11300 | 1.2126| 32.01 | 52.17 | 61.0986 |
| 0.2702 | 1.5724 | 11400 | 1.2213| 31.33 | 50.89 | 63.7551 |
| 0.2876 | 1.5862 | 11500 | 1.2126| 31.44 | 51.28 | 63.1697 |
| 0.2759 | 1.6 | 11600 | 1.2283| 30.49 | 51.02 | 64.7456 |
| 0.2902 | 1.6138 | 11700 | 1.2205| 32.33 | 50.53 | 63.2148 |
| 0.2638 | 1.6276 | 11800 | 1.2097| 31.89 | 51.14 | 62.6745 |
| 0.2605 | 1.6414 | 11900 | 1.2129| 31.35 | 50.63 | 63.3048 |
| 0.2374 | 1.6552 | 12000 | 1.2319| 31.48 | 51.73 | 63.4849 |
| 0.2436 | 1.6690 | 12100 | 1.2219| 30.43 | 50.92 | 65.5110 |
| 0.2366 | 1.6828 | 12200 | 1.2367| 31.64 | 51.14 | 64.7006 |
| 0.218 | 1.6966 | 12300 | 1.2142| 30.8 | 51.63 | 64.1153 |
| 0.2313 | 1.7103 | 12400 | 1.1877| 30.8 | 50.63 | 64.5655 |
| 0.2307 | 1.7241 | 12500 | 1.1817| 32.22 | 51.41 | 63.3498 |
| 0.2638 | 1.7379 | 12600 | 1.1514| 33.74 | 52.11 | 60.6033 |
| 0.2211 | 1.7517 | 12700 | 1.1563| 30.71 | 52.07 | 64.5655 |
| 0.197 | 1.7655 | 12800 | 1.1941| 32.22 | 52.9 | 62.8546 |
| 0.2307 | 1.7793 | 12900 | 1.1771| 32.83 | 52.96 | 62.7645 |
| 0.198 | 1.7931 | 13000 | 1.1908| 32.16 | 51.85 | 63.9352 |
| 0.1716 | 1.8069 | 13100 | 1.2065| 31.91 | 51.37 | 62.6294 |
| 0.2031 | 1.8207 | 13200 | 1.1745| 31.83 | 51.86 | 64.0252 |
| 0.1785 | 1.8345 | 13300 | 1.1607| 31.33 | 52.57 | 64.7006 |
| 0.2013 | 1.8483 | 13400 | 1.1785| 33.29 | 53.34 | 62.6745 |
| 0.1842 | 1.8621 | 13500 | 1.1723| 34.41 | 54.31 | 60.0630 |
| 0.2015 | 1.8759 | 13600 | 1.1859| 32.88 | 53.07 | 62.2692 |
| 0.1848 | 1.8897 | 13700 | 1.1668| 33.62 | 53.75 | 62.8095 |
| 0.1394 | 1.9034 | 13800 | 1.1734| 34.33 | 54.03 | 61.2787 |
| 0.1774 | 1.9172 | 13900 | 1.1735| 32.63 | 53.37 | 62.8996 |
| 0.1506 | 1.9310 | 14000 | 1.1768| 35.17 | 54.34 | 59.4327 |
| 0.1399 | 1.9448 | 14100 | 1.1827| 33.68 | 53.8 | 62.1792 |
| 0.1434 | 1.9586 | 14200 | 1.1721| 34.62 | 54.24 | 60.9185 |
| 0.1203 | 1.9724 | 14300 | 1.1733| 34.08 | 53.75 | 61.8190 |
| 0.1417 | 1.9862 | 14400 | 1.1615| 33.98 | 54.19 | 62.1792 |
| 0.1458 | 2.0 | 14500 | 1.1739| 33.65 | 53.31 | 62.9896 |
| 0.07 | 2.0138 | 14600 | 1.1916| 33.98 | 53.96 | 61.9090 |
| 0.051 | 2.0276 | 14700 | 1.1967| 34.13 | 54.36 | 61.1887 |
| 0.0481 | 2.0414 | 14800 | 1.2024| 34.06 | 54.38 | 61.4588 |
| 0.0574 | 2.0552 | 14900 | 1.2038| 34.23 | 54.08 | 61.2787 |
| 0.0621 | 2.0690 | 15000 | 1.2038| 34.85 | 54.43 | 60.9185 |
### Framework versions
- Transformers 4.41.2
- Pytorch 2.2.0+cu121
- Datasets 2.20.0
- Tokenizers 0.19.1
|
ShapeKapseln33/BPZone34 | ShapeKapseln33 | "2024-06-22T05:57:05Z" | 0 | 0 | null | [
"region:us"
] | null | "2024-06-22T05:55:26Z" | BP Zone USA Reviews BP Zone is a dietary supplement to help keep blood pressure healthy by lowering oxidative stress using various organic, non-harmful ingredients. The official website says that the BP Zone supplement makes it possible for the body to fully absorb the powerful components and start immediately to speed up the process of making nitric oxide.
**[Click here to buy now from official website of BP Zone](https://capsules24x7.com/bp-zone)**
##BP Zone Ingredient?
BP zone ingredients are the goodwill ambassador of this natural supplement. Naturally enriched herbs help blood pressure patients to feel comfortable and enjoy boosted energy, particularly in old age. Dr. Ryan Shylton at the Zenith Labs did a great effort to formulate this advanced formula. Following are the ingredients that have been playing a vital role in controlling high and low blood pressure.
##Ingredients
##Saffron:
BP zone ingredients are the goodwill ambassador of this natural supplement. Naturally enriched herbs help blood pressure patients to feel comfortable and enjoy boosted energy, particularly in old age. Following are the ingredients that have been playing a vital role in controlling high and low blood pressure.
##Garlic:
Garlic is a source of all, which is anti-oxidative stress. Allin has fiber that has the potential to reduce oxidative stress. Researches show great impact in flexing the tightened platelets. Its various advantages for your body are evident, particularly for cardiac rehabilitation.
##Hawthorn:
Inflammatory has adverse impacts on arteries and veins, maintaining blood supply—increased inflammation in arteries weakening heart.
Hawthorn has unique anti-inflammatory properties. Hence, Hawthorn not only reduces inflammation from veins and arteries but increases the efficiency of oxygen in the blood.
##Ginger:
A powerful medicinal property that has been using in a series of medicine formulas. BP zone supplement has gingerol properties, which are useful to reduce inflammation and oxidants from your body. It is helpful in smooth blood flow. Its antiseptic properties eliminate harmful chemicals and bacteria inside of your body.
##Danshen:
Known as a red sage in medical history is a useful herb in treating heart and blood vessels. It plays a vital role in opening arteries and veins. Unblocking blood supply to the heart boosts blood supply and maintains the flow of blood in your body. It also empowers the body to combat oxidative stress and maintain blood pressure.
**[Click here to buy now from official website of BP Zone](https://capsules24x7.com/bp-zone)**
##Calcium:
The antagonist’s effect of calcium helps to maintain blood pressure. Calcium plays an imperative role in removing blockades to the cardiovascular. Its natural properties hit enzymes and reduce inflammation in your body. It treats diastolic blood pressure and resists hypertension.
##Arjuna:
Known as cardiotonic in heart failure. Incredible herb with a series of benefits for human health has a long medical history. Arjuna helps to treat cardiomyopathy and stimulates blood circulation in your body. It helps arteries and vessels supply blood to the heart, where the heart pumps blood to normalize blood pressure.
##Magnesium:
A wonderful natural remedy to regulate the blood pressure in your body. It influences the reaction of the enzyme in your body and supporting the immune system to protect your body from disease. Pumpkins, cashew spinach, and almonds are rich foods having magnesium properties.
##CoQ10:
Coenzyme Q10 has been supporting heart treatments for a long time. It is a particular ingredient that deals with congestive heart failure and blood pressure. Studies show that CoQ10 has an important role in lowering systolic blood pressure. Systolic is a top reading. Coenzyme Q10 also controls the diastolic pressure by 10mm Hg without significant side effects.
##Berberine HCL:
A series of plans include barberry, Oregon grape, and tree t4urmeric provide berberine HCL. The extracts of berberine are useful to control high levels of cholesterol and high blood pressure. A vital extract of plants is also anti-inflammatory.
**[Click here to buy now from official website of BP Zone](https://capsules24x7.com/bp-zone)**
|
geraldabrhm/llama-2-13b-regulardataset-simplecontext-16lora | geraldabrhm | "2024-06-22T07:29:57Z" | 0 | 0 | null | [
"safetensors",
"region:us"
] | null | "2024-06-22T06:04:19Z" | Entry not found |
jddllwqa/Qwen-Qwen1.5-1.8B-1719036380 | jddllwqa | "2024-06-22T06:06:25Z" | 0 | 0 | peft | [
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:Qwen/Qwen1.5-1.8B",
"region:us"
] | null | "2024-06-22T06:06:21Z" | ---
base_model: Qwen/Qwen1.5-1.8B
library_name: peft
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
### Framework versions
- PEFT 0.11.1 |
jddllwqa/Qwen-Qwen1.5-7B-1719036460 | jddllwqa | "2024-06-22T06:07:45Z" | 0 | 0 | peft | [
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:Qwen/Qwen1.5-7B",
"region:us"
] | null | "2024-06-22T06:07:40Z" | ---
base_model: Qwen/Qwen1.5-7B
library_name: peft
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
### Framework versions
- PEFT 0.11.1 |
shiromiya/public-models | shiromiya | "2024-06-22T21:30:33Z" | 0 | 0 | null | [
"region:us"
] | null | "2024-06-22T06:07:42Z" | Entry not found |
jddllwqa/google-gemma-2b-1719036503 | jddllwqa | "2024-06-22T06:08:29Z" | 0 | 0 | peft | [
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:google/gemma-2b",
"region:us"
] | null | "2024-06-22T06:08:24Z" | ---
base_model: google/gemma-2b
library_name: peft
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
### Framework versions
- PEFT 0.11.1 |
Jada1/distilbert-base-uncased-finetuned-cola | Jada1 | "2024-06-22T06:08:39Z" | 0 | 0 | null | [
"region:us"
] | null | "2024-06-22T06:08:39Z" | Entry not found |
jddllwqa/Qwen-Qwen1.5-0.5B-1719036629 | jddllwqa | "2024-06-22T06:10:34Z" | 0 | 0 | peft | [
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:Qwen/Qwen1.5-0.5B",
"region:us"
] | null | "2024-06-22T06:10:29Z" | ---
base_model: Qwen/Qwen1.5-0.5B
library_name: peft
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
### Framework versions
- PEFT 0.11.1 |
jddllwqa/Qwen-Qwen1.5-1.8B-1719036687 | jddllwqa | "2024-06-22T06:11:30Z" | 0 | 0 | peft | [
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:Qwen/Qwen1.5-1.8B",
"region:us"
] | null | "2024-06-22T06:11:27Z" | ---
base_model: Qwen/Qwen1.5-1.8B
library_name: peft
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
### Framework versions
- PEFT 0.11.1 |
jddllwqa/Qwen-Qwen1.5-7B-1719036766 | jddllwqa | "2024-06-22T06:12:46Z" | 0 | 0 | null | [
"region:us"
] | null | "2024-06-22T06:12:46Z" | Entry not found |
jddllwqa/google-gemma-2b-1719036806 | jddllwqa | "2024-06-22T06:13:32Z" | 0 | 0 | peft | [
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:google/gemma-2b",
"region:us"
] | null | "2024-06-22T06:13:26Z" | ---
base_model: google/gemma-2b
library_name: peft
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
### Framework versions
- PEFT 0.11.1 |
El-chapoo/segmind_tiny-stable_diffusion_q8.ov | El-chapoo | "2024-06-22T06:15:08Z" | 0 | 0 | null | [
"region:us"
] | null | "2024-06-22T06:14:00Z" | Entry not found |
zandercoffman/SatTutor | zandercoffman | "2024-06-22T06:16:18Z" | 0 | 0 | null | [
"license:mit",
"region:us"
] | null | "2024-06-22T06:16:18Z" | ---
license: mit
---
|
happyneishon/git | happyneishon | "2024-06-30T16:09:01Z" | 0 | 0 | null | [
"region:us"
] | null | "2024-06-22T06:17:50Z" | Entry not found |
Eka-Korn/git-base-ru | Eka-Korn | "2024-06-22T06:21:29Z" | 0 | 0 | null | [
"region:us"
] | null | "2024-06-22T06:21:28Z" | Entry not found |
slelab/AES3 | slelab | "2024-06-22T08:07:09Z" | 0 | 0 | null | [
"region:us"
] | null | "2024-06-22T06:22:58Z" | Entry not found |
north/scandinavian_linguistic_classifier_bert | north | "2024-06-23T09:49:29Z" | 0 | 0 | transformers | [
"transformers",
"safetensors",
"bert",
"text-classification",
"license:mit",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-classification | "2024-06-22T06:24:37Z" | ---
license: mit
---
# Scandinavian Linguistic Classifier NB-BERT
Trained using code from [ComsmoPedia[https://github.com/huggingface/cosmopedia/tree/main/classification], but with the [nb-bert-base](https://huggingface.co/NbAiLab/nb-bert-base) as starting point. The [data](https://huggingface.co/datasets/north/scandinavian-llama3-annotations) used in classification is from [GlotCC](https://huggingface.co/datasets/cis-lmu/GlotCC-V1) and have been annotated using Gemini 1.5 Flash.
The following command where used for training. Please note that `train_regressor_bert.py` has a few minor changes to the original `train_edu_bert.py`:
```
python train_regressor_bert.py --base_model_name="NbAiLab/nb-bert-base" --dataset_name="north/scandinavian-linguistic-annotations" --target_column="score" --checkpoint_dir="/home/pere/checkpoints/scandinavian_bert/"
```
We also provide an evaluation script. This is targetted toward the same dataset.
```
python eval_regressor_bert.py --checkpoint_dir="/home/pere/checkpoints/scandinavian_bert/final/" --dataset_name="north/scandinavian-linguistic-annotations"
```
For convenience we also provide the `run_regressor_bert.py` script. This is also based on `run_edu_bert.py` from Cosmopedia. You can modify this script to annotate HuggingFace datasets directly. Cosmopedia also provides slurm-scripts here. We have not included these since we have had the opportunity to test them.
## Classification Report
| Class | Precision | Recall | F1-score | Support |
|-------|-----------|--------|----------|---------|
| 0 | 0.84 | 0.60 | 0.70 | 12209 |
| 1 | 0.70 | 0.72 | 0.71 | 24316 |
| 2 | 0.41 | 0.49 | 0.44 | 10499 |
| 3 | 0.38 | 0.51 | 0.43 | 5833 |
| 4 | 0.10 | 0.24 | 0.14 | 1342 |
| 5 | 0.87 | 0.39 | 0.54 | 5656 |
### Overall Metrics
| Metric | Value |
|---------------|-------|
| Accuracy | 0.59 |
| Macro Avg | |
| - Precision | 0.55 |
| - Recall | 0.49 |
| - F1-score | 0.50 |
| Weighted Avg | |
| - Precision | 0.65 |
| - Recall | 0.59 |
| - F1-score | 0.61 |
| Support | 59855 |
## Confusion Matrix
| | Predicted 0 | Predicted 1 | Predicted 2 | Predicted 3 | Predicted 4 | Predicted 5 |
|-------|--------------|--------------|--------------|--------------|--------------|--------------|
| Actual 0 | 7318 | 4278 | 529 | 63 | 19 | 2 |
| Actual 1 | 1364 | 17602 | 4414 | 785 | 135 | 16 |
| Actual 2 | 38 | 2615 | 5130 | 2289 | 369 | 58 |
| Actual 3 | 10 | 333 | 1726 | 2952 | 664 | 148 |
| Actual 4 | 3 | 83 | 350 | 476 | 324 | 106 |
| Actual 5 | 6 | 98 | 479 | 1205 | 1639 | 2229 |
## Evaluation Metrics
| Metric | Value |
|--------------------------|-----------------------|
| Eval Loss | 0.673861563205719 |
| Eval Precision | 0.5502142676492386 |
| Eval Recall | 0.49225148166352145 |
| Eval F1 Macro | 0.49616318856882935 |
| Eval Accuracy | 0.5940188789574806 |
| Eval Runtime | 285.9726 |
| Eval Samples per Second | 209.303 |
| Eval Steps per Second | 3.273 |
| Epoch | 19.96 |
## Training Runtime
| Metric | Value |
|--------------------------|-----------------------|
| Train Runtime | 105056.8322 |
| Train Samples per Second | 102.552 |
| Train Steps per Second | 1.603 |
| Train Loss | 0.6785072675819606 |
| Epoch | 20.0 |
### Run Summary
| Metric | Value |
|----------------------------|-----------------------|
| Eval Accuracy | 0.59402 |
| Eval F1 Macro | 0.49616 |
| Eval Loss | 0.67386 |
| Eval Precision | 0.55021 |
| Eval Recall | 0.49225 |
| Eval Runtime | 285.9726 |
| Eval Samples per Second | 209.303 |
| Eval Steps per Second | 3.273 |
| Total FLOPs | 2.8346790572921083e+18|
| Train Epoch | 20.0 |
| Train Global Step | 168360 |
| Train Grad Norm | 2.77268 |
| Train Learning Rate | 0.0 |
| Train Loss | 0.6201 |
| Train Loss (Final) | 0.67851 |
| Train Runtime | 105056.8322 |
| Train Samples per Second | 102.552 |
| Train Steps per Second | 1.603 |
|
Lululaudu/Cv | Lululaudu | "2024-06-22T06:25:33Z" | 0 | 0 | null | [
"license:apache-2.0",
"region:us"
] | null | "2024-06-22T06:25:33Z" | ---
license: apache-2.0
---
|
whoisotna/imas | whoisotna | "2024-06-22T06:26:56Z" | 0 | 0 | null | [
"license:openrail",
"region:us"
] | null | "2024-06-22T06:26:43Z" | ---
license: openrail
---
|
BLAGODEL/Hub | BLAGODEL | "2024-06-22T06:30:46Z" | 0 | 0 | null | [
"license:apache-2.0",
"region:us"
] | null | "2024-06-22T06:30:46Z" | ---
license: apache-2.0
---
|
vorstcavry/onclite | vorstcavry | "2024-06-22T06:41:42Z" | 0 | 0 | null | [
"region:us"
] | null | "2024-06-22T06:30:57Z" | Entry not found |
jddllwqa/Qwen-Qwen1.5-0.5B-1719037966 | jddllwqa | "2024-06-22T06:32:46Z" | 0 | 0 | null | [
"region:us"
] | null | "2024-06-22T06:32:46Z" | Entry not found |
acd23/new-model | acd23 | "2024-06-22T07:53:21Z" | 0 | 0 | null | [
"region:us"
] | null | "2024-06-22T06:38:21Z" | First time |
jddllwqa/Qwen-Qwen1.5-0.5B-1719038330 | jddllwqa | "2024-06-22T06:39:05Z" | 0 | 0 | peft | [
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:Qwen/Qwen1.5-0.5B",
"region:us"
] | null | "2024-06-22T06:38:50Z" | ---
base_model: Qwen/Qwen1.5-0.5B
library_name: peft
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
### Framework versions
- PEFT 0.11.1 |
teowu/longvideodb | teowu | "2024-06-23T03:13:20Z" | 0 | 0 | null | [
"region:us"
] | null | "2024-06-22T06:39:40Z" | Entry not found |
jddllwqa/Qwen-Qwen1.5-1.8B-1719038406 | jddllwqa | "2024-06-22T06:40:13Z" | 0 | 0 | peft | [
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:Qwen/Qwen1.5-1.8B",
"region:us"
] | null | "2024-06-22T06:40:06Z" | ---
base_model: Qwen/Qwen1.5-1.8B
library_name: peft
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
### Framework versions
- PEFT 0.11.1 |
sh9/face-mask-cls | sh9 | "2024-06-22T06:49:59Z" | 0 | 0 | null | [
"license:gpl-3.0",
"region:us"
] | null | "2024-06-22T06:40:19Z" | ---
license: gpl-3.0
---
|
jddllwqa/google-gemma-2b-1719038456 | jddllwqa | "2024-06-22T06:42:02Z" | 0 | 0 | peft | [
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:google/gemma-2b",
"region:us"
] | null | "2024-06-22T06:40:56Z" | ---
base_model: google/gemma-2b
library_name: peft
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
### Framework versions
- PEFT 0.11.1 |
jddllwqa/Qwen-Qwen1.5-0.5B-1719038712 | jddllwqa | "2024-06-22T06:45:19Z" | 0 | 0 | peft | [
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:Qwen/Qwen1.5-0.5B",
"region:us"
] | null | "2024-06-22T06:45:12Z" | ---
base_model: Qwen/Qwen1.5-0.5B
library_name: peft
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
### Framework versions
- PEFT 0.11.1 |
jddllwqa/Qwen-Qwen1.5-1.8B-1719038781 | jddllwqa | "2024-06-22T06:46:26Z" | 0 | 0 | peft | [
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:Qwen/Qwen1.5-1.8B",
"region:us"
] | null | "2024-06-22T06:46:22Z" | ---
base_model: Qwen/Qwen1.5-1.8B
library_name: peft
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
### Framework versions
- PEFT 0.11.1 |
Alefsouza72/Vortix | Alefsouza72 | "2024-06-22T06:47:20Z" | 0 | 0 | null | [
"region:us"
] | null | "2024-06-22T06:46:29Z" | Introdução à Aliança Sombra
Apresentação do grupo de vilões e seus planos contra Vortix.
Investigação de Vortix
Daniel detecta atividades suspeitas e investiga como Vortix.
Confronto no Armazém
Vortix confronta a Aliança Sombra e enfrenta seus membros.
Intervenção de Dr. Nova
Dr. Nova entra em cena e luta contra Vortix.
Batalha Intensa
A batalha entre Vortix e Dr. Nova se intensifica.
Retirada e Promessa de Vingança
Dr. Nova e a Aliança Sombra recuam, prometendo voltar. |
jddllwqa/google-gemma-2b-1719038828 | jddllwqa | "2024-06-22T06:47:14Z" | 0 | 0 | peft | [
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:google/gemma-2b",
"region:us"
] | null | "2024-06-22T06:47:08Z" | ---
base_model: google/gemma-2b
library_name: peft
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
### Framework versions
- PEFT 0.11.1 |
Ohrfeige/test | Ohrfeige | "2024-06-22T06:52:19Z" | 0 | 0 | mlx | [
"mlx",
"de",
"en",
"dataset:google/MusicCaps",
"license:apache-2.0",
"region:us"
] | null | "2024-06-22T06:51:15Z" | ---
license: apache-2.0
datasets:
- google/MusicCaps
language:
- de
- en
library_name: mlx
--- |
AI-Wheelz/AriesV2 | AI-Wheelz | "2024-06-22T06:55:34Z" | 0 | 0 | null | [
"license:openrail",
"region:us"
] | null | "2024-06-22T06:54:01Z" | ---
license: openrail
---
|
dalietng/yolov10l-visdrone | dalietng | "2024-06-22T06:54:55Z" | 0 | 0 | null | [
"region:us"
] | null | "2024-06-22T06:54:55Z" | Entry not found |
nerdthingz/rl_course_vizdoom_health_gathering_supreme | nerdthingz | "2024-06-22T06:55:15Z" | 0 | 0 | sample-factory | [
"sample-factory",
"tensorboard",
"deep-reinforcement-learning",
"reinforcement-learning",
"model-index",
"region:us"
] | reinforcement-learning | "2024-06-22T06:55:06Z" | ---
library_name: sample-factory
tags:
- deep-reinforcement-learning
- reinforcement-learning
- sample-factory
model-index:
- name: APPO
results:
- task:
type: reinforcement-learning
name: reinforcement-learning
dataset:
name: doom_health_gathering_supreme
type: doom_health_gathering_supreme
metrics:
- type: mean_reward
value: 10.20 +/- 5.61
name: mean_reward
verified: false
---
A(n) **APPO** model trained on the **doom_health_gathering_supreme** environment.
This model was trained using Sample-Factory 2.0: https://github.com/alex-petrenko/sample-factory.
Documentation for how to use Sample-Factory can be found at https://www.samplefactory.dev/
## Downloading the model
After installing Sample-Factory, download the model with:
```
python -m sample_factory.huggingface.load_from_hub -r nerdthingz/rl_course_vizdoom_health_gathering_supreme
```
## Using the model
To run the model after download, use the `enjoy` script corresponding to this environment:
```
python -m .usr.local.lib.python3.10.dist-packages.colab_kernel_launcher --algo=APPO --env=doom_health_gathering_supreme --train_dir=./train_dir --experiment=rl_course_vizdoom_health_gathering_supreme
```
You can also upload models to the Hugging Face Hub using the same script with the `--push_to_hub` flag.
See https://www.samplefactory.dev/10-huggingface/huggingface/ for more details
## Training with this model
To continue training with this model, use the `train` script corresponding to this environment:
```
python -m .usr.local.lib.python3.10.dist-packages.colab_kernel_launcher --algo=APPO --env=doom_health_gathering_supreme --train_dir=./train_dir --experiment=rl_course_vizdoom_health_gathering_supreme --restart_behavior=resume --train_for_env_steps=10000000000
```
Note, you may have to adjust `--train_for_env_steps` to a suitably high number as the experiment will resume at the number of steps it concluded at.
|
Hafees/emotion-llm | Hafees | "2024-06-22T06:55:38Z" | 0 | 0 | null | [
"region:us"
] | null | "2024-06-22T06:55:38Z" | Entry not found |