Text Generation
Transformers
PyTorch
English
llama
finance
llms
text-generation-inference
Edit model card

You need to agree to share your contact information to access this model

This repository is publicly accessible, but you have to accept the conditions to access its files and content.

Log in or Sign Up to review the conditions and access this model content.

FinMA-7B-full

FinMA-7B-full is a comprehensive financial large language model (LLM) developed as part of the PIXIU project. It is designed to understand complex financial language and concepts, and is fine-tuned to follow natural language instructions, enhancing its performance in downstream financial tasks. Specifically, FinMA-7B-full is trained with the full instruction data from the PIXIU dataset, covering both NLP and prediction tasks. This makes it a more comprehensive model capable of handling a wider range of financial tasks.

Other Models in the PIXIU Project

In addition to FinMA-7B-full, the PIXIU project includes two other models: FinMA-7B-NLP and FinMA-30B.

  • FinMA-7B-NLP: This model is trained only on the NLP tasks of the PIXIU dataset, making it specialized for tasks such as sentiment analysis, news headline classification, named entity recognition, and question answering.

  • FinMA-30B: This model is a larger version of FinMA, fine-tuned on the LLaMA-30B model. Like FinMA-7B-NLP, it is trained with the NLP instruction data.

Usage

You can use the FinMA-7B-full model in your Python project with the Hugging Face Transformers library. Here is a simple example of how to load the model:

from transformers import LlamaTokenizer, LlamaForCausalLM

tokenizer = LlamaTokenizer.from_pretrained('ChanceFocus/finma-7b-full')
model = LlamaForCausalLM.from_pretrained('ChanceFocus/finma-7b-full', device_map='auto')

In this example, LlamaTokenizer is used to load the tokenizer, and LlamaForCausalLM is used to load the model. The device_map='auto' argument is used to automatically use the GPU if it's available.

Hosted Inference API

You can also use the model through the Hugging Face Inference API. This allows you to generate text without having to set up your own inference environment. The model can be loaded on the Inference API on-demand.

License

FinMA-7B-NLP is licensed under MIT. For more details, please see the MIT file.

About

This model is part of the PIXIU project, an open-source resource featuring the first financial large language models (LLMs), instruction tuning data, and evaluation benchmarks to holistically assess financial LLMs. The goal is to continually push forward the open-source development of financial artificial intelligence (AI).

For more information, you can visit the PIXIU project on GitHub.

Citation

If you use FinMA-7B-NLP in your work, please cite the PIXIU paper:

@misc{xie2023pixiu,
 title={PIXIU: A Large Language Model, Instruction Data and Evaluation Benchmark for Finance}, 
 author={Qianqian Xie and Weiguang Han and Xiao Zhang and Yanzhao Lai and Min Peng and Alejandro Lopez-Lira and Jimin Huang},
 year={2023},
 eprint={2306.05443},
 archivePrefix={arXiv},
 primaryClass={cs.CL}
}
Downloads last month
128
Inference Examples
Inference API (serverless) has been turned off for this model.