|
--- |
|
license: apache-2.0 |
|
datasets: |
|
- Novora/CodeClassifier_v1 |
|
pipeline_tag: text-classification |
|
--- |
|
|
|
# Introduction |
|
|
|
Novora Code Classifier v1 Tiny, is a tiny `Text Classification` model, which classifies given code text input under 1 of `31` different classes (programming languages). |
|
|
|
This model is designed to be able to run on CPU, but optimally runs on GPUs. |
|
|
|
# Info |
|
- 1 of 31 classes output |
|
- 512 token input dimension |
|
- 64 hidden dimensions |
|
- 2 linear layers |
|
- The `snowflake-arctic-embed-xs` model is used as the embeddings model. |
|
- Dataset split into 80% training set, 20% testing set. |
|
- The combined test and training data is 100 chunks per programming language, the data is 3,100 chunks (entries) as 512 tokens per chunk, being a snippet of the code. |
|
|
|
# Architecture |
|
|
|
The `CodeClassifier-v1-Tiny` model employs a neural network architecture optimized for text classification tasks, specifically for classifying programming languages from code snippets. This model includes: |
|
|
|
- **Bidirectional LSTM Feature Extractor**: This bidirectional LSTM layer processes input embeddings, effectively capturing contextual relationships in both forward and reverse directions within the code snippets. |
|
|
|
- **Adaptive Pooling**: Following the LSTM, adaptive average pooling reduces the feature dimension to a fixed size, accommodating variable-length inputs. |
|
|
|
- **Fully Connected Layers**: The network includes two linear layers. The first projects the pooled features into a hidden feature space, and the second linear layer maps these to the output classes, which correspond to different programming languages. A dropout layer with a rate of 0.5 between these layers helps mitigate overfitting. |
|
|
|
The model's bidirectional nature and architectural components make it adept at understanding the syntax and structure crucial for code classification. |
|
|
|
|