File size: 1,681 Bytes
a9979f5 9f5e4f0 1930b02 4116153 b9434c2 4116153 4d6cc9b eab57b7 4d6cc9b eab57b7 4d6cc9b eab57b7 4d6cc9b b9434c2 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 |
---
license: wtfpl
datasets:
- cakiki/rosetta-code
language:
- en
metrics:
- accuracy
library_name: transformers
pipeline_tag: text-classification
tags:
- code
- programming-language
- code-classification
base_model: huggingface/CodeBERTa-small-v1
---
This Model is a fine-tuned version of *huggingface/CodeBERTa-small-v1* on *cakiki/rosetta-code* Dataset for 25 Programming Languages as mentioned below.
## Training Details:
Model is trained for 25 epochs on Azure for nearly 26000 Datapoints for above Mentioned 25 Programming Languages<br> extracted from Dataset having 1006 of total Programming Language.
### Programming Languages this model is able to detect vs Examples used for training
<ol>
<li>'ARM Assembly'</li>
<li>'AppleScript'</li>
<li>'C'</li>
<li>'C#'</li>
<li>'C++'</li>
<li>'COBOL'</li>
<li>'Erlang'</li>
<li>'Fortran'</li>
<li>'Go'</li>
<li>'Java'</li>
<li>'JavaScript'</li>
<li>'Kotlin'</li>
<li>'Lua</li>
<li>'Mathematica/Wolfram Language'</li>
<li>'PHP'</li>
<li>'Pascal'</li>
<li>'Perl'</li>
<li>'PowerShell'</li>
<li>'Python'</li>
<li>'R</li>
<li>'Ruby'</li>
<li>'Rust'</li>
<li>'Scala'</li>
<li>'Swift'</li>
<li>'Visual Basic .NET'</li>
<li>'jq'</li>
</ol>
<br>
## Below is the Training Result for 25 epochs.
Training Computer Configuration: GPU:1xNvidia Tesla T4, VRam: 16GB, Ram:112GB,Cores:6 Cores
Training Time taken: exactly 7 hours for 25 epochs
Training Hyper-parameters:
![image/png](https://cdn-uploads.huggingface.co/production/uploads/645c859ad90782b1a6a3e957/yRqjKVFKZIT_zXjcA3yFW.png)
![training detail.png](https://cdn-uploads.huggingface.co/production/uploads/645c859ad90782b1a6a3e957/Oi9TuJ8nEjtt6Z_W56myn.png) |