File size: 1,856 Bytes
0771155 bafd612 0771155 271db01 0771155 271db01 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 |
---
license: llama3
language:
- en
base_model:
- m3rg-iitd/llamat-3
tags:
- crystal structure generation
- CIF understanding
---
# Model Card for LLaMat-3-CIF
**LLaMat-3-CIF** is a specialized large language model designed to generate and extract information from Crystallographic Information Files.
The model is developed after continued pretraining of **LLaMat-3** on 7M instruction-output pairs obtained using CIFs from Materials Project, Google GNoME, and AMCSD
---
## Overview
- **Model Type:** Large Language Model (LLM)
- **Base Model:** LLaMat-3 (continued pretraining of LLaMat-3 on CIFs)
- **Language:** English
- **License:** LLaMA-3 License
- **Tags:** Material Science, Domain Adaptation, Crystal Structure Generation
---
## Model Details
### Key Features
- **Instruction Following Abilities:** Answers questions based on CIF files.
- **Applications:** Crystal structure generation
### Development and Support
- **Developed by:** [M3RG, IIT Delhi](https://github.com/M3RG-IITD/) & [DAIR, IIT Delhi](https://github.com/dair-iitd)
- **Compute Support:**
- **Edinburgh International Data Facility (EIDF):** Provided access to Cerebras CS2 clusters for pretraining.
- **IIT Delhi High-Performance Computing Cluster:** Supported fine-tuning and inference stages.
---
## Technical Specifications
### Hardware Infrastructure
- **Pretraining:** 2 Cerebras CS-2 Wafer-Scale Engines (WSE-2)
- **Finetuning:** 2 Cerebras CS-2 Wafer-Scale Engines (WSE-2)
- **Inferencing:** 1 NVIDIA A100 80GB GPU
### Software Stack
- **Frameworks:** PyTorch, Hugging Face Transformers
---
## Model Sources
- **Repository:** [LLaMat on GitHub](https://github.com/M3RG-IITD/llamat)
- **Compute Resources:** [EIDF Cerebras CS Clusters](https://edinburgh-international-data-facility.ed.ac.uk/services/computing/cerebras-cs) |