File size: 3,574 Bytes
afd30ec
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
---
license: apache-2.0
datasets:
- josedamico/sugarcane
language:
- en
base_model:
- TinyLlama/TinyLlama-1.1B-Chat-v1.0
tags:
- sugarcane
---
# 🌱 TinyLLaMA-Sugarcane

Welcome to the *first open-source LLM fine-tuned for sugarcane production*! 🧠🌾

This model is a fine-tuned version of [`TinyLLaMA`](https://huggingface.co/czi/TinyLlama-1.1B-Chat-v1.0), trained specifically on sugarcane-focused data. Developed by [SciCrop](https://scicrop.com) as part of its commitment to open innovation in agriculture, this is one of the first domain-specific small language models (SLMs) created for the agribusiness sector.

---

## 🚜 Why Sugarcane?

Sugarcane is one of the most important crops in Brazil and globally β€” but most LLMs know very little about its specific production cycle, challenges, and terminology.

By fine-tuning TinyLLaMA on 2,000+ question/answer pairs from real-world sugarcane use cases, we aim to deliver:

- βœ… Better accuracy  
- βœ… Clearer answers  
- βœ… Local deployment capabilities for agricultural experts, cooperatives, and researchers

---

## πŸ” Model Details

- **Base model**: `TinyLLaMA-1.1B-Chat`
- **Fine-tuned on**: Domain-specific QA pairs related to sugarcane
- **Architecture**: Causal LM with LoRA + QLoRA
- **Tokenizer**: `LLaMATokenizer`
- **Model size**: ~1.1B parameters
- **Format**: Available in both HF standard and `GGUF` for local/Ollama use

---

## πŸ§ͺ Try it locally with Ollama

We believe local models are the future for privacy-sensitive, domain-specific AI.

You can run this model locally using [Ollama](https://ollama.com):

```bash
ollama run infinitestack/tinyllama-sugarcane
```

πŸ‘‰ Or explore the model directly:  
https://ollama.com/infinitestack/tinyllama-sugarcane

---

## 🌐 About InfiniteStack

This model is part of **InfiniteStack**, a platform by [SciCrop](https://scicrop.com) that helps companies in the agri-food-energy-environment chain create, train, and deploy their own AI and analytics solutions β€” securely and at scale.

### πŸ“¦ InfiniteStack offers:

- A containerized platform that runs on-prem or in private cloud  
- Full support for **SLMs and LLMs** using your **real and private data**  
- No/Low-code interfaces to *Collect*, *Automate*, *Leverage*, *Catalog*, *Observe*, and *Track* data pipelines and AI assets

🌐 Learn more: https://infinitestack.ai

---

## 🧠 Why Small Language Models (SLMs)?

SLMs are great when:

- You need local inference (offline, on-device, or private)
- Your domain is narrow and specific
- You want full control over fine-tuning and usage
- You care about speed, size, and cost-efficiency

Big isn’t always better. Sometimes, smart and focused beats giant and generic. πŸ’‘

---

## 🀝 Community & Open Innovation

This work reflects SciCrop’s ongoing commitment to the open-source ecosystem, and to creating useful, usable AI for real-world agribusiness.

Feel free to fork, contribute, fine-tune further, or use it in your own ag project.  
We’d love to hear how you're using it!

---

## πŸ“‚ Files included

This repo includes:

- `config.json`  
- `tokenizer.model`  
- `tokenizer.json`  
- `model.safetensors`  
- `special_tokens_map.json`  
- `generation_config.json`  
- `tokenizer_config.json`  
- `README.md`

A merged and converted `.gguf` version is also available at **Ollama Hub**.

---

## πŸ“¬ Questions or Contributions?

Ping us at:  
πŸ“§ [email protected]  
🌐 https://scicrop.com  
🌱 https://infinitestack.ai

Made with β˜•, 🌾 and ❀️ in Brazil  
by @josedamico and the InfiniteStack team