Spaces:
Runtime error
Runtime error
Kevin Xu
commited on
Commit
•
4994dd4
1
Parent(s):
df6985d
try all pegasus models
Browse files- app.py +49 -4
- requirements.txt +1 -0
app.py
CHANGED
@@ -1,7 +1,52 @@
|
|
1 |
import gradio as gr
|
2 |
|
3 |
-
|
4 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
5 |
|
6 |
-
iface
|
7 |
-
iface.launch()
|
|
|
1 |
import gradio as gr
|
2 |
|
3 |
+
from transformers import pipeline
|
4 |
+
import gradio as gr
|
5 |
+
from gradio.mix import Parallel, Series
|
6 |
+
|
7 |
+
io1 = gr.Interface.load('huggingface/google/pegasus-large')
|
8 |
+
io2 = gr.Interface.load("huggingface/google/pegasus-cnn_dailymail")
|
9 |
+
io3 = gr.Interface.load("huggingface/google/pegasus-xsum")
|
10 |
+
io4 = gr.Interface.load('huggingface/google/pegasus-newsroom')
|
11 |
+
io5 = gr.Interface.load("huggingface/google/pegasus-multi_news")
|
12 |
+
io6 = gr.Interface.load("huggingface/google/pegasus-reddit_tifu")
|
13 |
+
io7 = gr.Interface.load('huggingface/google/pegasus-arxiv')
|
14 |
+
io8 = gr.Interface.load("huggingface/google/pegasus-pubmed")
|
15 |
+
io9 = gr.Interface.load("huggingface/google/pegasus-wikihow")
|
16 |
+
io10 = gr.Interface.load('huggingface/google/pegasus-gigaword')
|
17 |
+
io11 = gr.Interface.load("huggingface/google/pegasus-billsum")
|
18 |
+
io12 = gr.Interface.load("huggingface/google/pegasus-big_patent")
|
19 |
+
io13 = gr.Interface.load("huggingface/google/pegasus-aeslc")
|
20 |
+
|
21 |
+
desc = "Let Hugging Face models summarize texts for you. Note: Shorter articles generate faster summaries. This summarizer uses bart-large-cnn model by Facebook, pegasus by Google and distilbart-cnn-12-6 by Sshleifer. You can compare these models against each other on their performances. Sample Text input is provided!"
|
22 |
+
|
23 |
+
x = """ What's A Lawyer Now? Simply put… there is a tremendous manifest and latent need for just about ALL legal services. There are solid interrelated sociological and structural reasons for this including considerable societal divisiveness, meaningful changes in laws and regulations, and fast-paced disruptive technological innovations. At the same time, there are psychological factors that strongly prompt the need for various legal services such as hubris, arrogance, and Machiavellianism. The opportunities, across a wide spectrum of law firm practice areas, have probably never been greater. Although there is a tremendous amount of untapped potential for legal services, there is one major obstacle to opening the spigot – lawyers. From solo practices to mega-international law firms, many lawyers because of their inherent inclinations (e.g., risk aversion) reinforced by their education and firm experience are not going to take advantage of the incredible latent demand for legal services. As commoditization is rampant in the legal profession, the path to success is not just having “excellent knowledge of the law.” Being technical proficient is table stakes. Unfortunately, a large percentage of lawyers equate legal competence with the success of their practice, and the great majority is proven wrong. What is also required of lawyers at all levels, in order to truly excel in today’s legal environment, is a touch of entrepreneurialism coupled with some business savvy. The opportunities for lawyers are most everywhere from inside their own book of business to the clients of other lawyers in their firms to the many other types of professionals they know or can fairly easily get to know. The complication is that when it comes to the business development side of legal work, few lawyers have the expertise to create a steady stream of new work for their practices or their firms. Unless lawyers adopt these best practices, it is unlikely that they will be able to greatly benefit from all the tremendous pent up demand that exists for legal services. Conversely, for those lawyers who take a proactive and systemic approach to business development, their practices could easily grow exponentially.
|
24 |
+
"""
|
25 |
+
|
26 |
+
y = '''What is Text Summarization?
|
27 |
+
Text summarization is an important NLP task, which has several applications. The two broad categories of approaches to text summarization are extraction and abstraction. Extractive methods select a subset of existing words, phrases, or sentences in the original text to form a summary. In contrast, abstractive methods first build an internal semantic representation and then use natural language generation techniques to create a summary. Such a summary might contain words that are not explicitly present in the original document. Most text summarization systems are based on some form of extractive summarization.
|
28 |
+
In general, topic identification, interpretation, summary generation, and evaluation of the generated summary are the key challenges in text summarization. The critical tasks in extraction-based summarization are identifying key phrases in the document and using them to select sentences in the document for inclusion in the summary. In contrast, abstraction-based methods paraphrase sections of the source document.
|
29 |
+
All extraction-based summarizers perform the following three relatively independent tasks (Nenkova and McKeown, 2011, 2012): (a) capturing key aspects of text and storing as an intermediate representation, (b) scoring sentences in the text based on that representation, (c) and composing a summary by selecting several sentences.'''
|
30 |
+
|
31 |
+
z = '''Machine Learning Technology Trends To Impact Business in 2022
|
32 |
+
In this article, we will discuss the latest innovations in machine learning technology in 2021 from our perspective as a machine learning software development company. We’ll go over 9 trends and explain how the latest innovations in machine learning technologies can benefit you and your business in 2022.
|
33 |
+
1. No-Code Machine Learning
|
34 |
+
2. TinyML
|
35 |
+
3. AutoML
|
36 |
+
4. Machine Learning Operationalization Management
|
37 |
+
5. Full-stack Deep Learning
|
38 |
+
6. Generative Adversarial Networks
|
39 |
+
7. Unsupervised ML
|
40 |
+
8. Reinforcement Learning
|
41 |
+
'''
|
42 |
+
|
43 |
+
sample = [[y],[x],[z]]
|
44 |
+
|
45 |
+
iface = Parallel(io1, io2, io3, io4, io5, io6, io7, io8, io9, io10, io11, io12, io13,
|
46 |
+
theme='huggingface',
|
47 |
+
title= 'Hugging Face Text Summarizer',
|
48 |
+
description = desc,
|
49 |
+
examples=sample, #replace "sample" with directory to let gradio scan through those files and give you the text
|
50 |
+
inputs = gr.inputs.Textbox(lines = 10, label="Text"))
|
51 |
|
52 |
+
iface.launch(inline = False)
|
|
requirements.txt
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
transformers
|