Edit model card

Data Is Everything.

To try other models(involving commercial-available model), please check out our Demo Page(🔨constructing)

This model is made by Ados based on upstage/SOLAR-10.7B-Instruct-v1.0.

Train Dataset

Dataset used for training is collected primarily from huggingface and utilized using our own translation model.

  • Language
    • KR 73%
    • EN 24%
    • Others 3%
  • Type
    • single turn QA (alpaca style) 29%
    • multi turn QA (vicuna style) 21%
    • instructed QA 26%
    • summary 12%
    • translation 12%

After collecting data, we removed low quality rows. We chose 30% high quality from raw data manually and using deduplication methods.

We also refined problematic data such as code blocks, listing, repetition and other common issues we found.

Prompt template

### System:
You are an AI assistant, please behave and help the user. Your name is OLLM(오름) by Ados(주식회사아도스), OLLM stands for On-premise LLM.

### User: On-premise LLM이 뭔가요?

### Assistant:

For more informations, please contact us.

To try other models(involving commercial-available model), please check out our Demo Page(🔨constructing)

License

@misc{kim2023solar,
      title={SOLAR 10.7B: Scaling Large Language Models with Simple yet Effective Depth Up-Scaling}, 
      author={Dahyun Kim and Chanjun Park and Sanghoon Kim and Wonsung Lee and Wonho Song and Yunsu Kim and Hyeonwoo Kim and Yungi Kim and Hyeonju Lee and Jihoo Kim and Changbae Ahn and Seonghoon Yang and Sukyung Lee and Hyunbyung Park and Gyoungjin Gim and Mikyoung Cha and Hwalsuk Lee and Sunghun Kim},
      year={2023},
      eprint={2312.15166},
      archivePrefix={arXiv},
      primaryClass={cs.CL}
}
Downloads last month
1,971
Safetensors
Model size
10.7B params
Tensor type
FP16
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.