bmah-dmx commited on
Commit
8f39c12
·
verified ·
1 Parent(s): 69c46a4

Delete README.md

Browse files
Files changed (1) hide show
  1. README.md +0 -58
README.md DELETED
@@ -1,58 +0,0 @@
1
- ---
2
- model-index:
3
- - name: CodeLlama-7b
4
- results:
5
- - task:
6
- type: code-generation
7
- dataset:
8
- name: Humaneval
9
- type: humaneval
10
- metrics:
11
- - type: pass@1 (BASELINE)
12
- value: 0.3048780487804878
13
- - type: pass@1 (BASIC)
14
- value: 0.3170731707317073
15
- ---
16
- This is a d-Matrix functional reference of the CODELLAMA-7B model.
17
- The reference provides the following functional *configurations*:
18
- Configuration | Explanation
19
- :-- | :--
20
- **`BASELINE`** | a reference functionally equivalent to the original model
21
- **`BASIC`** | all linear algebraic operands quantized to `BFP16-64`, and all other operations transformed to approximated kernel simulations
22
-
23
-
24
- ### Usage
25
-
26
- Install d-Matrix [Dmx_Compressor](https://github.com/d-matrix-ai/dmx-compressor) first.
27
- ```sh
28
- pip install dmx_compressor
29
-
30
- git clone https://github.com/bigcode-project/bigcode-evaluation-harness.git
31
- cd bigcode-evaluation-harness
32
- pip install .
33
- ```
34
-
35
- The following is an example model and its evaluation.
36
-
37
- ```python
38
- from dmx.compressor import config_rules
39
- from dmx.compressor.modeling import DmxModel
40
- from bigcode_eval.evaluator import Evaluator
41
-
42
- pipe = pipeline(
43
- task="text-generation",
44
- model="d-matrix/d-matrix/CodeLlama-7b",
45
- trust_remote_code=True,
46
- device=0,
47
- )
48
-
49
- # Transform the model with DMX
50
- model = DmxModel.from_torch(pipe.model).to("cuda")
51
- tensor = torch.randint(1, 100, (1, max_length)).to("cuda") # Assign desired max length of generation to max_length
52
- model.transform(model.dmx_config, *eval(f"config_rules.{dmx_config}"))
53
- setup = model(tensor)
54
-
55
- evaluator = Evaluator(accelerator, model, tokenizer, eval_args)
56
-
57
- result = evaluator.evaluate(task)
58
- ```