claudios commited on
Commit
9e655ab
1 Parent(s): 8c3eacb

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +12 -1
README.md CHANGED
@@ -1,6 +1,17 @@
 
 
 
 
 
 
1
 
 
2
 
3
- This is an *unofficial* reupload of [microsoft/graphcodebert-base](https://huggingface.co/microsoft/graphcodebert-base) in the `SafeTensors` format using `transformers` `4.40.1`. The goal of this reupload is to prevent older models that are still relevant baselines from becoming stale as a result of changes in HuggingFace. Additionally, I may include minor corrections, such as model max length configuration.## GraphCodeBERT model
 
 
 
 
4
 
5
  GraphCodeBERT is a graph-based pre-trained model based on the Transformer architecture for programming language, which also considers data-flow information along with code sequences. GraphCodeBERT consists of 12 layers, 768 dimensional hidden states, and 12 attention heads. The maximum sequence length for the model is 512. The model is trained on the CodeSearchNet dataset, which includes 2.3M functions with document pairs for six programming languages.
6
 
 
1
+ ---
2
+ arxiv: 2009.08366
3
+ language:
4
+ - code
5
+ license: mit
6
+ ---
7
 
8
+ This is an *unofficial* reupload of [microsoft/graphcodebert-base](https://huggingface.co/microsoft/graphcodebert-base) in the `SafeTensors` format using `transformers` `4.40.1`. The goal of this reupload is to prevent older models that are still relevant baselines from becoming stale as a result of changes in HuggingFace. Additionally, I may include minor corrections, such as model max length configuration.
9
 
10
+ Original model card below:
11
+
12
+ ---
13
+
14
+ ## GraphCodeBERT model
15
 
16
  GraphCodeBERT is a graph-based pre-trained model based on the Transformer architecture for programming language, which also considers data-flow information along with code sequences. GraphCodeBERT consists of 12 layers, 768 dimensional hidden states, and 12 attention heads. The maximum sequence length for the model is 512. The model is trained on the CodeSearchNet dataset, which includes 2.3M functions with document pairs for six programming languages.
17