--- license: mit datasets: - hyokwan/common language: - ko metrics: - accuracy base_model: - google/gemma-7b-it pipeline_tag: text-generation library_name: transformers --- Model Details model model is continued pretrained language model based on google/gemma-7b-it This model is trained with specific department of university (Korea Plytechnics Fintech) . Intended Use TBD How to use TBD Responsibility & Safety We believe that an open approach to AI leads to better, safer products, faster innovation, and a bigger overall market. We are committed to Responsible AI development and took a series of steps to limit misuse and harm and support the open source community. Foundation models are widely capable technologies that are built to be used for a diverse range of applications. They are not designed to meet every developer preference on safety levels for all use cases, out-of-the-box, as those by their nature will differ across different applications. Rather, responsible LLM-application deployment is achieved by implementing a series of safety best practices throughout the development of such applications, from the model pre-training, fine-tuning and the deployment of systems composed of safeguards to tailor the safety needs specifically to the use case and audience. As part of the Llama 3 release, we updated our Responsible Use Guide to outline the steps and best practices for developers to implement model and system level safety for their application. We also provide a set of resources including Meta Llama Guard 2 and Code Shield safeguards. These tools have proven to drastically reduce residual risks of LLM Systems, while maintaining a high level of helpfulness. We encourage developers to tune and deploy these safeguards according to their needs and we provide a reference implementation to get you started. Responsible release In addition to responsible use considerations outlined above, we followed a rigorous process that requires us to take extra measures against misuse and critical risks before we make our release decision. Misuse --- license: mit datasets: - hyokwan/common language: - ko metrics: - accuracy base_model: - google/gemma-7b-it pipeline_tag: text-generation library_name: transformers tags: - finance - education ---