TinyLlama-1.1bee 🐝

image/png

As we feverishly hit the refresh button on hf.co's homepage, on the hunt for the newest waifu chatbot to grace the AI stage, an epiphany struck us like a bee sting. What could we offer to the hive-mind of the community? The answer was as clear as honeyβ€”beekeeping, naturally. And thus, this un-bee-lievable model was born.

Details

This model is a fine-tuned version of PY007/TinyLlama-1.1B-intermediate-step-240k-503b on the BEE-spoke-data/bees-internal dataset. It achieves the following results on the evaluation set:

  • Loss: 2.4285
  • Accuracy: 0.4969
***** eval metrics *****
  eval_accuracy           =     0.4972                                              
  eval_loss               =     2.4283
  eval_runtime            = 0:00:53.12
  eval_samples            =        239
  eval_samples_per_second =      4.499
  eval_steps_per_second   =      1.129
  perplexity              =    11.3391

πŸ“œ Intended Uses & Limitations πŸ“œ

Intended Uses:

  1. Educational Engagement: Whether you're a novice beekeeper, an enthusiast, or someone just looking to understand the buzz around bees, this model aims to serve as an informative and entertaining resource.
  2. General Queries: Have questions about hive management, bee species, or honey extraction? Feel free to consult the model for general insights.
  3. Academic & Research Inspiration: If you're diving into the world of apiculture studies or environmental science, our model could offer some preliminary insights and ideas.

Limitations:

  1. Not a Beekeeping Expert: As much as we admire bees and their hard work, this model is not a certified apiculturist. Please consult professional beekeeping resources or experts for serious decisions related to hive management, bee health, and honey production.
  2. Licensing: Apache-2.0, following TinyLlama
  3. Infallibility: Our model can err, just like any other piece of technology (or bee). Always double-check the information before applying it to your own hive or research.
  4. Ethical Constraints: This model may not be used for any illegal or unethical activities, including but not limited to: bioterrorism & standard terrorism, harassment, or spreading disinformation.

Training and evaluation data

While the full dataset is not yet complete and therefore not yet released for "safety reasons", you can check out a preliminary sample at: bees-v0

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 0.0001
  • train_batch_size: 4
  • eval_batch_size: 4
  • seed: 80085
  • gradient_accumulation_steps: 8
  • total_train_batch_size: 32
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: cosine
  • lr_scheduler_warmup_ratio: 0.03
  • num_epochs: 2.0

Open LLM Leaderboard Evaluation Results

Detailed results can be found here

Metric Value
Avg. 29.15
ARC (25-shot) 30.55
HellaSwag (10-shot) 51.8
MMLU (5-shot) 24.25
TruthfulQA (0-shot) 39.01
Winogrande (5-shot) 54.46
GSM8K (5-shot) 0.23
DROP (3-shot) 3.74
Downloads last month
34
Safetensors
Model size
1.1B params
Tensor type
F32
Β·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for BEE-spoke-data/TinyLlama-1.1bee

Finetuned
(1)
this model
Merges
3 models

Dataset used to train BEE-spoke-data/TinyLlama-1.1bee

Collection including BEE-spoke-data/TinyLlama-1.1bee