entslscheia commited on
Commit
9439737
1 Parent(s): d3f7c56

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -10,7 +10,7 @@ size_categories:
10
 
11
  **Introduction**
12
 
13
- In traditional knowledge base question answering (KBQA) methods, semantic parsing plays a crucial role. It requires requires a semantic parser to be extensively trained on a vast dataset of labeled examples, typically consisting of question-answer or question-program pairs. This necessity arises primarily because smaller models before the advent of large language models (LLMs) were data-hungry, needing extensive data to effectively master tasks. Additionally, these methods often relied on the assumption that data is independent and identically distributed (i.i.d.), meaning the questions a model could answer needed to match the distribution of the training data. This demanded training data to cover a broad spectrum of the KB's entities and relationships for the model to understand it adequately.
14
 
15
  However, the rise of LLMs has shifted this paradigm. LLMs excel in learning from few (or even zero) in-context examples. They utilize natural language as a general vehicle of thought, enabling them to actively navigate and interact with KBs using auxiliary tools, without the need for training on comprehensive datasets. This advance suggests LLMs can sidestep the earlier limitations and eliminate the dependency on extensive, high-coverage training data.
16
 
 
10
 
11
  **Introduction**
12
 
13
+ In traditional knowledge base question answering (KBQA) methods, semantic parsing plays a crucial role. It requires a semantic parser to be extensively trained on a vast dataset of labeled examples, typically consisting of question-answer or question-program pairs.
14
 
15
  However, the rise of LLMs has shifted this paradigm. LLMs excel in learning from few (or even zero) in-context examples. They utilize natural language as a general vehicle of thought, enabling them to actively navigate and interact with KBs using auxiliary tools, without the need for training on comprehensive datasets. This advance suggests LLMs can sidestep the earlier limitations and eliminate the dependency on extensive, high-coverage training data.
16