com3dian commited on
Commit
d3a1351
1 Parent(s): 502b69a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +73 -0
README.md CHANGED
@@ -6,3 +6,76 @@ tags:
6
  license:
7
  - mit
8
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
6
  license:
7
  - mit
8
  ---
9
+ # Bart-Large Summarization Model
10
+
11
+ ![Bart Logo](https://huggingface.co/front/assets/huggingface_logo.svg)
12
+
13
+ This repository contains the **Bart-Large-paper2slides-summarizer Model**, which has been fine-tuned on the [Automatic Slide Generation from Scientific Papers dataset](https://www.kaggle.com/datasets/andrewmvd/automatic-slide-generation-from-scientific-papers) using unsupervised learning techniques.
14
+ The model is made available for usage and experimentation through the Hugging Face Model Hub.
15
+
16
+ ## Model Details
17
+
18
+ - **Model Architecture**: Bart-Large
19
+ - **Fine-tuning Dataset**: [Automatic Slide Generation from Scientific Papers](https://www.kaggle.com/datasets/andrewmvd/automatic-slide-generation-from-scientific-papers)
20
+ - **Fine-tuning Method**: Unsupervised Learning
21
+
22
+ [Bart](https://huggingface.co/transformers/model_doc/bart.html) (Bidirectional and Auto-Regressive Transformers) is a sequence-to-sequence (seq2seq) model developed by Facebook AI Research. It has shown exceptional performance in various natural language processing (NLP) tasks such as text summarization, text generation, and machine translation.
23
+
24
+ This particular model, Bart-Large, is the larger version of the Bart model. It consists of 12 encoder and decoder layers and has a total of 400 million parameters.
25
+
26
+ ## Dataset
27
+
28
+ The model has been fine-tuned on a slide generation dataset. Slide generation involves automatically generating visually appealing and informative slide decks based on input data. The dataset used for fine-tuning this model contains examples of input data along with corresponding slide deck outputs. The fine-tuning process enables the model to learn to generate high-quality slides based on the given input.
29
+
30
+ ## Usage
31
+
32
+ To use this model, you can leverage the Hugging Face [Transformers](https://huggingface.co/transformers/) library. Here's an example of how to use it in Python:
33
+
34
+ ```python
35
+ from transformers import BartTokenizer, BartForConditionalGeneration
36
+
37
+ # Load the model and tokenizer
38
+ model_name = "your-username/bart-large-slide-generation"
39
+ tokenizer = BartTokenizer.from_pretrained(model_name)
40
+ model = BartForConditionalGeneration.from_pretrained(model_name)
41
+
42
+ # Generate slides from input text
43
+ input_text = "Your input text here..."
44
+ input_ids = tokenizer.encode(input_text, return_tensors="pt")
45
+ output = model.generate(input_ids)
46
+
47
+ # Decode generated slides
48
+ slides = tokenizer.decode(output[0], skip_special_tokens=True)
49
+ print(slides)
50
+ ```
51
+
52
+ Ensure you have the `transformers` library installed before running the code. You can install it using `pip`:
53
+
54
+ ```
55
+ pip install transformers
56
+ ```
57
+
58
+ ## Model Fine-tuning Details
59
+
60
+ The fine-tuning process for this model involved training on the slide generation dataset using unsupervised learning techniques. Unsupervised learning refers to training a model without explicit human-labeled targets. Instead, the model learns to generate slides by maximizing the likelihood of generating the correct output given the input data.
61
+
62
+ The specific hyperparameters and training details used for fine-tuning this model are as follows:
63
+
64
+ - Batch Size: 32
65
+ - Learning Rate: 1e-4
66
+ - Training Steps: 100,000
67
+ - Optimizer: AdamW
68
+
69
+ ## Model Performance
70
+
71
+ The performance of the Bart-Large Slide Generation Model has been evaluated on various metrics, including slide quality, coherence, and relevance. While the model has achieved promising results during evaluation, it is essential to note that no model is perfect, and its performance may vary depending on the input data and specific use cases.
72
+
73
+ ## Acknowledgments
74
+
75
+ We would like to acknowledge the authors of the Bart model and the creators of the slide generation dataset for their valuable contributions, which have enabled the development of this fine-tuned model.
76
+
77
+ If you use this model or find it helpful in your work, please consider citing the original Bart model and the slide generation dataset to provide proper credit to the respective authors.
78
+
79
+ ## License
80
+
81
+ This model and the associated code are released under the [Apache 2.0 license](https://www.apache.org/licenses/LICENSE-2.0).