File size: 1,681 Bytes
a9979f5
 
9f5e4f0
 
 
 
 
 
 
 
 
 
 
 
1930b02
 
4116153
b9434c2
4116153
4d6cc9b
 
eab57b7
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
4d6cc9b
eab57b7
 
 
 
4d6cc9b
eab57b7
4d6cc9b
b9434c2
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
---
license: wtfpl
datasets:
- cakiki/rosetta-code
language:
- en
metrics:
- accuracy
library_name: transformers
pipeline_tag: text-classification
tags:
- code
- programming-language
- code-classification
base_model: huggingface/CodeBERTa-small-v1
---
This Model is a fine-tuned version of *huggingface/CodeBERTa-small-v1* on *cakiki/rosetta-code* Dataset for 25 Programming Languages as mentioned below.
## Training Details:
Model is trained for 25 epochs on Azure for nearly 26000 Datapoints for above Mentioned 25 Programming Languages<br> extracted from Dataset having 1006 of total Programming Language.
### Programming Languages this model is able to detect vs Examples used for training
<ol>
  <li>'ARM Assembly'</li>
 <li>'AppleScript'</li>
 <li>'C'</li>
 <li>'C#'</li>
 <li>'C++'</li>
 <li>'COBOL'</li>
 <li>'Erlang'</li>
 <li>'Fortran'</li>
 <li>'Go'</li>
 <li>'Java'</li>
 <li>'JavaScript'</li>
 <li>'Kotlin'</li>
 <li>'Lua</li>
 <li>'Mathematica/Wolfram Language'</li>
 <li>'PHP'</li>
 <li>'Pascal'</li>
 <li>'Perl'</li>
 <li>'PowerShell'</li>
 <li>'Python'</li>
 <li>'R</li>
 <li>'Ruby'</li>
 <li>'Rust'</li>
 <li>'Scala'</li>
 <li>'Swift'</li>
 <li>'Visual Basic .NET'</li>
 <li>'jq'</li>
</ol>
<br>

## Below is the Training Result for 25 epochs.
Training Computer Configuration: GPU:1xNvidia Tesla T4, VRam: 16GB, Ram:112GB,Cores:6 Cores
Training Time taken: exactly 7 hours for 25 epochs
Training Hyper-parameters:

![image/png](https://cdn-uploads.huggingface.co/production/uploads/645c859ad90782b1a6a3e957/yRqjKVFKZIT_zXjcA3yFW.png)

![training detail.png](https://cdn-uploads.huggingface.co/production/uploads/645c859ad90782b1a6a3e957/Oi9TuJ8nEjtt6Z_W56myn.png)