--- license: apache-2.0 datasets: - Novora/CodeClassifier_v1 pipeline_tag: text-classification --- # Introduction Novora Code Classifier v1 Tiny, is a tiny `Text Classification` model, which classifies given code text input under 1 of `31` different classes (programming languages). This model is designed to be able to run on CPU, but optimally runs on GPUs. # Info - 1 of 31 classes output - 512 token input dimension - 64 hidden dimensions - 2 linear layers - The `snowflake-arctic-embed-xs` model is used as the embeddings model. - Dataset split into 80% training set, 20% testing set. - The combined test and training data is 100 chunks per programming language, the data is 3,100 chunks (entries) as 512 tokens per chunk, being a snippet of the code. # Architecture The `CodeClassifier-v1-Tiny` model employs a neural network architecture optimized for text classification tasks, specifically for classifying programming languages from code snippets. This model includes: - **Bidirectional LSTM Feature Extractor**: This bidirectional LSTM layer processes input embeddings, effectively capturing contextual relationships in both forward and reverse directions within the code snippets. - **Adaptive Pooling**: Following the LSTM, adaptive average pooling reduces the feature dimension to a fixed size, accommodating variable-length inputs. - **Fully Connected Layers**: The network includes two linear layers. The first projects the pooled features into a hidden feature space, and the second linear layer maps these to the output classes, which correspond to different programming languages. A dropout layer with a rate of 0.5 between these layers helps mitigate overfitting. The model's bidirectional nature and architectural components make it adept at understanding the syntax and structure crucial for code classification.