mstatt commited on
Commit
874dece
1 Parent(s): 0960708

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +98 -163
README.md CHANGED
@@ -1,199 +1,134 @@
1
  ---
2
- library_name: transformers
3
- tags: []
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
4
  ---
5
-
6
- # Model Card for Model ID
7
-
8
- <!-- Provide a quick summary of what the model is/does. -->
9
-
10
-
11
-
12
  ## Model Details
13
 
14
- ### Model Description
15
-
16
- <!-- Provide a longer summary of what this model is. -->
17
-
18
- This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
19
-
20
- - **Developed by:** [More Information Needed]
21
- - **Funded by [optional]:** [More Information Needed]
22
- - **Shared by [optional]:** [More Information Needed]
23
- - **Model type:** [More Information Needed]
24
- - **Language(s) (NLP):** [More Information Needed]
25
- - **License:** [More Information Needed]
26
- - **Finetuned from model [optional]:** [More Information Needed]
27
-
28
- ### Model Sources [optional]
29
-
30
- <!-- Provide the basic links for the model. -->
31
-
32
- - **Repository:** [More Information Needed]
33
- - **Paper [optional]:** [More Information Needed]
34
- - **Demo [optional]:** [More Information Needed]
35
-
36
- ## Uses
37
-
38
- <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
39
-
40
- ### Direct Use
41
-
42
- <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
43
-
44
- [More Information Needed]
45
 
46
- ### Downstream Use [optional]
47
 
48
- <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
49
 
50
- [More Information Needed]
51
 
52
- ### Out-of-Scope Use
53
 
54
- <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
55
 
56
- [More Information Needed]
57
 
58
- ## Bias, Risks, and Limitations
59
 
60
- <!-- This section is meant to convey both technical and sociotechnical limitations. -->
61
 
62
- [More Information Needed]
63
 
64
- ### Recommendations
65
 
66
- <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
67
 
68
- Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
69
 
70
- ## How to Get Started with the Model
 
71
 
72
- Use the code below to get started with the model.
 
73
 
74
- [More Information Needed]
 
 
75
 
76
- ## Training Details
77
 
78
- ### Training Data
79
 
80
- <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
81
-
82
- [More Information Needed]
83
-
84
- ### Training Procedure
85
-
86
- <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
87
-
88
- #### Preprocessing [optional]
89
-
90
- [More Information Needed]
91
-
92
-
93
- #### Training Hyperparameters
94
-
95
- - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
96
-
97
- #### Speeds, Sizes, Times [optional]
98
-
99
- <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
100
-
101
- [More Information Needed]
102
 
103
  ## Evaluation
104
 
105
- <!-- This section describes the evaluation protocols and provides the results. -->
106
-
107
- ### Testing Data, Factors & Metrics
108
-
109
- #### Testing Data
110
-
111
- <!-- This should link to a Dataset Card if possible. -->
112
-
113
- [More Information Needed]
114
-
115
- #### Factors
116
-
117
- <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
118
-
119
- [More Information Needed]
120
-
121
- #### Metrics
122
-
123
- <!-- These are the evaluation metrics being used, ideally with a description of why. -->
124
-
125
- [More Information Needed]
126
-
127
- ### Results
128
-
129
- [More Information Needed]
130
-
131
- #### Summary
132
-
133
-
134
-
135
- ## Model Examination [optional]
136
-
137
- <!-- Relevant interpretability work for the model goes here -->
138
-
139
- [More Information Needed]
140
-
141
- ## Environmental Impact
142
-
143
- <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
144
-
145
- Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
146
-
147
- - **Hardware Type:** [More Information Needed]
148
- - **Hours used:** [More Information Needed]
149
- - **Cloud Provider:** [More Information Needed]
150
- - **Compute Region:** [More Information Needed]
151
- - **Carbon Emitted:** [More Information Needed]
152
-
153
- ## Technical Specifications [optional]
154
-
155
- ### Model Architecture and Objective
156
-
157
- [More Information Needed]
158
-
159
- ### Compute Infrastructure
160
-
161
- [More Information Needed]
162
-
163
- #### Hardware
164
-
165
- [More Information Needed]
166
-
167
- #### Software
168
-
169
- [More Information Needed]
170
-
171
- ## Citation [optional]
172
-
173
- <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
174
-
175
- **BibTeX:**
176
-
177
- [More Information Needed]
178
 
179
- **APA:**
180
 
181
- [More Information Needed]
 
182
 
183
- ## Glossary [optional]
184
 
185
- <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
186
 
187
- [More Information Needed]
 
188
 
189
- ## More Information [optional]
190
 
191
- [More Information Needed]
192
 
193
- ## Model Card Authors [optional]
194
 
195
- [More Information Needed]
196
 
197
- ## Model Card Contact
 
 
 
 
 
 
 
198
 
199
- [More Information Needed]
 
1
  ---
2
+ license: apache-2.0
3
+ language:
4
+ - en
5
+ pipeline_tag: summarization
6
+ widget:
7
+ - text: >-
8
+ Maria: Alright, let's start dinner. How did everyone's day go?
9
+ Steven: It wasn't too bad. Had a couple of meetings but I did manage to finalize the design for the new project I've been working on.
10
+ Lilly: That's great, Dad! Finally!
11
+ Ryan: Cool, does that mean we get to see it now?
12
+ Steven: Well, I'll have to keep it under wraps for a bit longer, but soon.
13
+ Maria: Congrats, Steven. Lilly, how about you? Anything interesting at school today?
14
+ Lilly: Yeah, we had an astronaut from NASA come in and talk about the space missions he's been on. It was quite cool.
15
+ Steven: That sounds really interesting. Anything in particular that stood out?
16
+ Lilly: Well, he said he was about my age when he first started training. It’s really amazing how much they have to learn and experience.
17
+ Ryan: Did he talk about aliens?
18
+ Lilly: He did Ryan, but mostly about how we're not sure if they exist yet. It was thought-provoking.
19
+ Maria: Those experiences are definitely something to think about. Ryan, what about your day? Anything exciting in school?
20
+ Ryan: We did a geography quiz today. I got the highest marks!
21
+ Steven: That's my boy! Did you enjoy it?
22
+ Ryan: Yeah, it was fun! I want to try more at home.
23
+ Maria: Well, we can definitely help you with that. What do you think about setting up some geography puzzles this weekend?
24
+ Ryan: Yes, let's do it! Lilly, you're helping me.
25
+ Lilly: Absolutely, I'd love to help, Ryan. We can make it a fun game.
26
+ Steven: Sounds like a plan. Maria, remember we need to go grocery shopping this weekend.
27
+ Maria: Yes, thanks for reminding me. We’ll do that on Saturday afternoon.
28
+ Lilly: Can I go to the library after that? I need to pick up some books for my project.
29
+ Maria: Sure, just make sure to finish your homework before that.
30
+ Lilly: Deal.
31
+ Ryan: Can we go to the park on Saturday?
32
+ Steven: If it's not raining, sure. We'll all go together.
33
+ Maria: Sounds like we've got a fun weekend planned. Let's make sure to finish all the chores tomorrow so we have time.
34
+ Lilly: Agreed.
35
+ Ryan: Can I get extra dessert tonight?
36
+ Maria: Only if you finish your vegetables, Ryan.
37
+ Ryan: Deal, but only because you make the best roasted broccoli.
38
+ Steven: Good job, Ryan. And Maria, thanks for the dinner. It was perfect.
39
+ Maria: You're welcome, everyone. Enjoy your meal.
40
+ Lilly: Is it time for some dessert now?
41
+ Ryan: Yes, dessert time!
42
+ Maria: Alright, dessert it is. But remember, tomorrow is chore day, okay?
43
+ Everyone: Deal!
44
+
45
+ example_title: Conversation Arc Example 1
46
+ tags:
47
+ - NLP
48
  ---
49
+ # Model Card for Conversation Arc PredictorArc of the Conversation Model
 
 
 
 
 
 
50
  ## Model Details
51
 
52
+ - **Model Name:** arc_of_conversation
53
+ - **Model Type:** Fine-tuned `google/t5-small`
54
+ - **Language:** English
55
+ - **License:** MIT
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
56
 
57
+ ## Overview
58
 
59
+ The Conversation Arc Predictor model is designed to predict the arc of a conversation given its text. It is based on the `google/t5-small` model, fine-tuned on a custom dataset of conversations and their corresponding arcs. This model can be used to analyze and categorize conversation texts into predefined arcs.
60
 
61
+ ## Model Description
62
 
63
+ ### Model Architecture
64
 
65
+ The base model architecture is T5 (Text-To-Text Transfer Transformer), which treats every NLP problem as a text-to-text problem. The specific version used here is `google/t5-small`, which has been fine-tuned to understand and predict conversation arcs.
66
 
67
+ ### Fine-Tuning Data
68
 
69
+ The model was fine-tuned on a dataset consisting of conversation texts and their corresponding arcs. The dataset should be formatted in a CSV file with two columns: `conversation` and `arc`.
70
 
71
+ ### Intended Use
72
 
73
+ The model is intended for categorizing the arc of conversation texts. It can be useful for applications in customer service, chatbots, conversational analysis, and other areas where understanding the flow of a conversation is important.
74
 
75
+ ## How to Use
76
 
77
+ ### Inference
78
 
79
+ To use this model for inference, you need to load the fine-tuned model and tokenizer. Here is an example of how to do this using the `transformers` library:
80
 
81
+ ```python
82
+ from transformers import pipeline
83
 
84
+ # Generate the Arc of the conversation
85
+ convo2 = """ Your conversation text Here"""
86
 
87
+ summarizer = pipeline("summarization", model='Falconsai/arc_of_conversation', tokenizer='Falconsai/arc_of_conversation')
88
+ result = summarizer(convo2, max_length=2048, min_length=1024, do_sample=False)
89
+ ```
90
 
91
+ ## Training
92
 
93
+ The training process involves the following steps:
94
 
95
+ 1. **Load and Explore Data:** Load the dataset and perform initial exploration to understand the data distribution.
96
+ 2. **Preprocess Data:** Tokenize the conversations and prepare them for the T5 model.
97
+ 3. **Fine-Tune Model:** Fine-tune the `google/t5-small` model using the preprocessed data.
98
+ 4. **Evaluate Model:** Evaluate the model's performance on a validation set to ensure it's learning correctly.
99
+ 5. **Save Model:** Save the fine-tuned model for future use.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
100
 
101
  ## Evaluation
102
 
103
+ The model's performance should be evaluated on a separate validation set to ensure it accurately predicts the conversation arcs. Metrics such as accuracy, precision, recall, and F1 score can be used to assess its performance.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
104
 
105
+ ## Limitations
106
 
107
+ - **Data Dependency:** The model's performance is highly dependent on the quality and representativeness of the training data.
108
+ - **Generalization:** The model may not generalize well to conversation texts that are significantly different from the training data.
109
 
110
+ ## Ethical Considerations
111
 
112
+ When deploying the model, be mindful of the ethical implications, including but not limited to:
113
 
114
+ - **Privacy:** Ensure that conversation data used for training and inference does not contain sensitive or personally identifiable information.
115
+ - **Bias:** Be aware of potential biases in the training data that could affect the model's predictions.
116
 
117
+ ## License
118
 
119
+ This project is licensed under the MIT License. See the [LICENSE](LICENSE) file for details.
120
 
121
+ ## Citation
122
 
123
+ If you use this model in your research, please cite it as follows:
124
 
125
+ ```
126
+ @misc{conversation_arc_predictor,
127
+ author = {Michael Stattelman},
128
+ title = {Conversation Arc Predictor},
129
+ year = {2024},
130
+ publisher = {Falcons.ai},
131
+ }
132
+ ```
133
 
134
+ ---