Spaces:
Running
Running
Upload 7646 files
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- .gitattributes +103 -0
- llama.cpp/.DS_Store +0 -0
- llama.cpp/.cache/huggingface/.gitignore +1 -0
- llama.cpp/.cache/huggingface/download/.gitattributes.lock +0 -0
- llama.cpp/.cache/huggingface/download/.gitattributes.metadata +3 -0
- llama.cpp/.cache/huggingface/download/README.md.lock +0 -0
- llama.cpp/.cache/huggingface/download/README.md.metadata +3 -0
- llama.cpp/.cache/huggingface/download/index.html.lock +0 -0
- llama.cpp/.cache/huggingface/download/index.html.metadata +3 -0
- llama.cpp/.cache/huggingface/download/llama.cpp/.DS_Store.lock +0 -0
- llama.cpp/.cache/huggingface/download/llama.cpp/.DS_Store.metadata +3 -0
- llama.cpp/.cache/huggingface/download/llama.cpp/.clang-format.lock +0 -0
- llama.cpp/.cache/huggingface/download/llama.cpp/.clang-format.metadata +3 -0
- llama.cpp/.cache/huggingface/download/llama.cpp/.clang-tidy.lock +0 -0
- llama.cpp/.cache/huggingface/download/llama.cpp/.clang-tidy.metadata +3 -0
- llama.cpp/.cache/huggingface/download/llama.cpp/.devops/cloud-v-pipeline.lock +0 -0
- llama.cpp/.cache/huggingface/download/llama.cpp/.devops/cloud-v-pipeline.metadata +3 -0
- llama.cpp/.cache/huggingface/download/llama.cpp/.devops/cpu.Dockerfile.lock +0 -0
- llama.cpp/.cache/huggingface/download/llama.cpp/.devops/cpu.Dockerfile.metadata +3 -0
- llama.cpp/.cache/huggingface/download/llama.cpp/.devops/cuda.Dockerfile.lock +0 -0
- llama.cpp/.cache/huggingface/download/llama.cpp/.devops/cuda.Dockerfile.metadata +3 -0
- llama.cpp/.cache/huggingface/download/llama.cpp/.devops/intel.Dockerfile.lock +0 -0
- llama.cpp/.cache/huggingface/download/llama.cpp/.devops/intel.Dockerfile.metadata +3 -0
- llama.cpp/.cache/huggingface/download/llama.cpp/.devops/llama-cli-cann.Dockerfile.lock +0 -0
- llama.cpp/.cache/huggingface/download/llama.cpp/.devops/llama-cli-cann.Dockerfile.metadata +3 -0
- llama.cpp/.cache/huggingface/download/llama.cpp/.devops/llama-cpp-cuda.srpm.spec.lock +0 -0
- llama.cpp/.cache/huggingface/download/llama.cpp/.devops/llama-cpp-cuda.srpm.spec.metadata +3 -0
- llama.cpp/.cache/huggingface/download/llama.cpp/.devops/llama-cpp.srpm.spec.lock +0 -0
- llama.cpp/.cache/huggingface/download/llama.cpp/.devops/llama-cpp.srpm.spec.metadata +3 -0
- llama.cpp/.cache/huggingface/download/llama.cpp/.devops/musa.Dockerfile.lock +0 -0
- llama.cpp/.cache/huggingface/download/llama.cpp/.devops/musa.Dockerfile.metadata +3 -0
- llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/apps.nix.lock +0 -0
- llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/apps.nix.metadata +3 -0
- llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/devshells.nix.lock +0 -0
- llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/devshells.nix.metadata +3 -0
- llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/docker.nix.lock +0 -0
- llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/docker.nix.metadata +3 -0
- llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/jetson-support.nix.lock +0 -0
- llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/jetson-support.nix.metadata +3 -0
- llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/nixpkgs-instances.nix.lock +0 -0
- llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/nixpkgs-instances.nix.metadata +3 -0
- llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/package-gguf-py.nix.lock +0 -0
- llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/package-gguf-py.nix.metadata +3 -0
- llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/package.nix.lock +0 -0
- llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/package.nix.metadata +3 -0
- llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/python-scripts.nix.lock +0 -0
- llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/python-scripts.nix.metadata +3 -0
- llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/scope.nix.lock +0 -0
- llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/scope.nix.metadata +3 -0
- llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/sif.nix.lock +0 -0
.gitattributes
CHANGED
@@ -33,3 +33,106 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
36 |
+
llama.cpp/build/bin/libggml-base.dylib filter=lfs diff=lfs merge=lfs -text
|
37 |
+
llama.cpp/build/bin/libggml-cpu.dylib filter=lfs diff=lfs merge=lfs -text
|
38 |
+
llama.cpp/build/bin/libggml-metal.dylib filter=lfs diff=lfs merge=lfs -text
|
39 |
+
llama.cpp/build/bin/libllama.dylib filter=lfs diff=lfs merge=lfs -text
|
40 |
+
llama.cpp/build/bin/libllava_shared.dylib filter=lfs diff=lfs merge=lfs -text
|
41 |
+
llama.cpp/build/bin/llama-batched filter=lfs diff=lfs merge=lfs -text
|
42 |
+
llama.cpp/build/bin/llama-batched-bench filter=lfs diff=lfs merge=lfs -text
|
43 |
+
llama.cpp/build/bin/llama-bench filter=lfs diff=lfs merge=lfs -text
|
44 |
+
llama.cpp/build/bin/llama-cli filter=lfs diff=lfs merge=lfs -text
|
45 |
+
llama.cpp/build/bin/llama-convert-llama2c-to-ggml filter=lfs diff=lfs merge=lfs -text
|
46 |
+
llama.cpp/build/bin/llama-cvector-generator filter=lfs diff=lfs merge=lfs -text
|
47 |
+
llama.cpp/build/bin/llama-embedding filter=lfs diff=lfs merge=lfs -text
|
48 |
+
llama.cpp/build/bin/llama-eval-callback filter=lfs diff=lfs merge=lfs -text
|
49 |
+
llama.cpp/build/bin/llama-export-lora filter=lfs diff=lfs merge=lfs -text
|
50 |
+
llama.cpp/build/bin/llama-gen-docs filter=lfs diff=lfs merge=lfs -text
|
51 |
+
llama.cpp/build/bin/llama-gritlm filter=lfs diff=lfs merge=lfs -text
|
52 |
+
llama.cpp/build/bin/llama-imatrix filter=lfs diff=lfs merge=lfs -text
|
53 |
+
llama.cpp/build/bin/llama-infill filter=lfs diff=lfs merge=lfs -text
|
54 |
+
llama.cpp/build/bin/llama-llava-cli filter=lfs diff=lfs merge=lfs -text
|
55 |
+
llama.cpp/build/bin/llama-llava-clip-quantize-cli filter=lfs diff=lfs merge=lfs -text
|
56 |
+
llama.cpp/build/bin/llama-lookahead filter=lfs diff=lfs merge=lfs -text
|
57 |
+
llama.cpp/build/bin/llama-lookup filter=lfs diff=lfs merge=lfs -text
|
58 |
+
llama.cpp/build/bin/llama-lookup-create filter=lfs diff=lfs merge=lfs -text
|
59 |
+
llama.cpp/build/bin/llama-lookup-stats filter=lfs diff=lfs merge=lfs -text
|
60 |
+
llama.cpp/build/bin/llama-minicpmv-cli filter=lfs diff=lfs merge=lfs -text
|
61 |
+
llama.cpp/build/bin/llama-parallel filter=lfs diff=lfs merge=lfs -text
|
62 |
+
llama.cpp/build/bin/llama-passkey filter=lfs diff=lfs merge=lfs -text
|
63 |
+
llama.cpp/build/bin/llama-perplexity filter=lfs diff=lfs merge=lfs -text
|
64 |
+
llama.cpp/build/bin/llama-quantize filter=lfs diff=lfs merge=lfs -text
|
65 |
+
llama.cpp/build/bin/llama-quantize-stats filter=lfs diff=lfs merge=lfs -text
|
66 |
+
llama.cpp/build/bin/llama-qwen2vl-cli filter=lfs diff=lfs merge=lfs -text
|
67 |
+
llama.cpp/build/bin/llama-retrieval filter=lfs diff=lfs merge=lfs -text
|
68 |
+
llama.cpp/build/bin/llama-run filter=lfs diff=lfs merge=lfs -text
|
69 |
+
llama.cpp/build/bin/llama-save-load-state filter=lfs diff=lfs merge=lfs -text
|
70 |
+
llama.cpp/build/bin/llama-server filter=lfs diff=lfs merge=lfs -text
|
71 |
+
llama.cpp/build/bin/llama-speculative filter=lfs diff=lfs merge=lfs -text
|
72 |
+
llama.cpp/build/bin/llama-speculative-simple filter=lfs diff=lfs merge=lfs -text
|
73 |
+
llama.cpp/build/bin/llama-tokenize filter=lfs diff=lfs merge=lfs -text
|
74 |
+
llama.cpp/build/bin/llama-tts filter=lfs diff=lfs merge=lfs -text
|
75 |
+
llama.cpp/build/bin/test-arg-parser filter=lfs diff=lfs merge=lfs -text
|
76 |
+
llama.cpp/build/bin/test-backend-ops filter=lfs diff=lfs merge=lfs -text
|
77 |
+
llama.cpp/build/bin/test-chat filter=lfs diff=lfs merge=lfs -text
|
78 |
+
llama.cpp/build/bin/test-chat-template filter=lfs diff=lfs merge=lfs -text
|
79 |
+
llama.cpp/build/bin/test-grammar-integration filter=lfs diff=lfs merge=lfs -text
|
80 |
+
llama.cpp/build/bin/test-json-schema-to-grammar filter=lfs diff=lfs merge=lfs -text
|
81 |
+
llama.cpp/build/bin/test-tokenizer-0 filter=lfs diff=lfs merge=lfs -text
|
82 |
+
llama.cpp/build/bin/test-tokenizer-1-bpe filter=lfs diff=lfs merge=lfs -text
|
83 |
+
llama.cpp/build/bin/test-tokenizer-1-spm filter=lfs diff=lfs merge=lfs -text
|
84 |
+
llama.cpp/build/common/CMakeFiles/common.dir/arg.cpp.o filter=lfs diff=lfs merge=lfs -text
|
85 |
+
llama.cpp/build/common/CMakeFiles/common.dir/chat.cpp.o filter=lfs diff=lfs merge=lfs -text
|
86 |
+
llama.cpp/build/common/CMakeFiles/common.dir/common.cpp.o filter=lfs diff=lfs merge=lfs -text
|
87 |
+
llama.cpp/build/common/CMakeFiles/common.dir/json-schema-to-grammar.cpp.o filter=lfs diff=lfs merge=lfs -text
|
88 |
+
llama.cpp/build/common/libcommon.a filter=lfs diff=lfs merge=lfs -text
|
89 |
+
llama.cpp/build/examples/llama-bench/CMakeFiles/llama-bench.dir/llama-bench.cpp.o filter=lfs diff=lfs merge=lfs -text
|
90 |
+
llama.cpp/build/examples/llava/CMakeFiles/llava.dir/clip.cpp.o filter=lfs diff=lfs merge=lfs -text
|
91 |
+
llama.cpp/build/examples/llava/libllava_static.a filter=lfs diff=lfs merge=lfs -text
|
92 |
+
llama.cpp/build/examples/perplexity/CMakeFiles/llama-perplexity.dir/perplexity.cpp.o filter=lfs diff=lfs merge=lfs -text
|
93 |
+
llama.cpp/build/examples/quantize-stats/CMakeFiles/llama-quantize-stats.dir/quantize-stats.cpp.o filter=lfs diff=lfs merge=lfs -text
|
94 |
+
llama.cpp/build/examples/run/CMakeFiles/llama-run.dir/run.cpp.o filter=lfs diff=lfs merge=lfs -text
|
95 |
+
llama.cpp/build/examples/server/CMakeFiles/llama-server.dir/server.cpp.o filter=lfs diff=lfs merge=lfs -text
|
96 |
+
llama.cpp/build/examples/tts/CMakeFiles/llama-tts.dir/tts.cpp.o filter=lfs diff=lfs merge=lfs -text
|
97 |
+
llama.cpp/build/ggml/src/CMakeFiles/ggml-base.dir/ggml-quants.c.o filter=lfs diff=lfs merge=lfs -text
|
98 |
+
llama.cpp/build/ggml/src/CMakeFiles/ggml-base.dir/ggml.c.o filter=lfs diff=lfs merge=lfs -text
|
99 |
+
llama.cpp/build/ggml/src/CMakeFiles/ggml-base.dir/gguf.cpp.o filter=lfs diff=lfs merge=lfs -text
|
100 |
+
llama.cpp/build/ggml/src/CMakeFiles/ggml-cpu.dir/ggml-cpu/ggml-cpu.c.o filter=lfs diff=lfs merge=lfs -text
|
101 |
+
llama.cpp/build/ggml/src/CMakeFiles/ggml-cpu.dir/ggml-cpu/llamafile/sgemm.cpp.o filter=lfs diff=lfs merge=lfs -text
|
102 |
+
llama.cpp/build/ggml/src/ggml-metal/CMakeFiles/ggml-metal.dir/__/__/__/autogenerated/ggml-metal-embed.s.o filter=lfs diff=lfs merge=lfs -text
|
103 |
+
llama.cpp/build/ggml/src/ggml-metal/CMakeFiles/ggml-metal.dir/ggml-metal.m.o filter=lfs diff=lfs merge=lfs -text
|
104 |
+
llama.cpp/build/src/CMakeFiles/llama.dir/llama-model-loader.cpp.o filter=lfs diff=lfs merge=lfs -text
|
105 |
+
llama.cpp/build/src/CMakeFiles/llama.dir/llama-model.cpp.o filter=lfs diff=lfs merge=lfs -text
|
106 |
+
llama.cpp/build/src/CMakeFiles/llama.dir/llama-sampling.cpp.o filter=lfs diff=lfs merge=lfs -text
|
107 |
+
llama.cpp/build/src/CMakeFiles/llama.dir/llama-vocab.cpp.o filter=lfs diff=lfs merge=lfs -text
|
108 |
+
llama.cpp/build/src/CMakeFiles/llama.dir/llama.cpp.o filter=lfs diff=lfs merge=lfs -text
|
109 |
+
llama.cpp/build/src/CMakeFiles/llama.dir/unicode.cpp.o filter=lfs diff=lfs merge=lfs -text
|
110 |
+
llama.cpp/build/tests/CMakeFiles/test-backend-ops.dir/test-backend-ops.cpp.o filter=lfs diff=lfs merge=lfs -text
|
111 |
+
llama.cpp/build/tests/CMakeFiles/test-chat-template.dir/test-chat-template.cpp.o filter=lfs diff=lfs merge=lfs -text
|
112 |
+
llama.cpp/build/tests/CMakeFiles/test-chat.dir/test-chat.cpp.o filter=lfs diff=lfs merge=lfs -text
|
113 |
+
llama.cpp/build/tests/CMakeFiles/test-grammar-integration.dir/test-grammar-integration.cpp.o filter=lfs diff=lfs merge=lfs -text
|
114 |
+
llama.cpp/build/tests/CMakeFiles/test-json-schema-to-grammar.dir/test-json-schema-to-grammar.cpp.o filter=lfs diff=lfs merge=lfs -text
|
115 |
+
llama.cpp/docs/development/llama-star/idea-arch.key filter=lfs diff=lfs merge=lfs -text
|
116 |
+
llama.cpp/examples/server/themes/buttons-top/buttons_top.png filter=lfs diff=lfs merge=lfs -text
|
117 |
+
llama.cpp/examples/server/themes/wild/llamapattern.png filter=lfs diff=lfs merge=lfs -text
|
118 |
+
llama.cpp/examples/server/themes/wild/wild.png filter=lfs diff=lfs merge=lfs -text
|
119 |
+
llama.cpp/gguf-py/gguf/__pycache__/quants.cpython-311.pyc filter=lfs diff=lfs merge=lfs -text
|
120 |
+
llama.cpp/media/llama0-banner.png filter=lfs diff=lfs merge=lfs -text
|
121 |
+
llama.cpp/media/llama0-logo.png filter=lfs diff=lfs merge=lfs -text
|
122 |
+
llama.cpp/media/matmul.png filter=lfs diff=lfs merge=lfs -text
|
123 |
+
llama.cpp/models/ggml-vocab-aquila.gguf filter=lfs diff=lfs merge=lfs -text
|
124 |
+
llama.cpp/models/ggml-vocab-baichuan.gguf filter=lfs diff=lfs merge=lfs -text
|
125 |
+
llama.cpp/models/ggml-vocab-bert-bge.gguf filter=lfs diff=lfs merge=lfs -text
|
126 |
+
llama.cpp/models/ggml-vocab-command-r.gguf filter=lfs diff=lfs merge=lfs -text
|
127 |
+
llama.cpp/models/ggml-vocab-deepseek-coder.gguf filter=lfs diff=lfs merge=lfs -text
|
128 |
+
llama.cpp/models/ggml-vocab-deepseek-llm.gguf filter=lfs diff=lfs merge=lfs -text
|
129 |
+
llama.cpp/models/ggml-vocab-falcon.gguf filter=lfs diff=lfs merge=lfs -text
|
130 |
+
llama.cpp/models/ggml-vocab-gpt-2.gguf filter=lfs diff=lfs merge=lfs -text
|
131 |
+
llama.cpp/models/ggml-vocab-gpt-neox.gguf filter=lfs diff=lfs merge=lfs -text
|
132 |
+
llama.cpp/models/ggml-vocab-llama-bpe.gguf filter=lfs diff=lfs merge=lfs -text
|
133 |
+
llama.cpp/models/ggml-vocab-llama-spm.gguf filter=lfs diff=lfs merge=lfs -text
|
134 |
+
llama.cpp/models/ggml-vocab-mpt.gguf filter=lfs diff=lfs merge=lfs -text
|
135 |
+
llama.cpp/models/ggml-vocab-phi-3.gguf filter=lfs diff=lfs merge=lfs -text
|
136 |
+
llama.cpp/models/ggml-vocab-qwen2.gguf filter=lfs diff=lfs merge=lfs -text
|
137 |
+
llama.cpp/models/ggml-vocab-refact.gguf filter=lfs diff=lfs merge=lfs -text
|
138 |
+
llama.cpp/models/ggml-vocab-starcoder.gguf filter=lfs diff=lfs merge=lfs -text
|
llama.cpp/.DS_Store
ADDED
Binary file (10.2 kB). View file
|
|
llama.cpp/.cache/huggingface/.gitignore
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
*
|
llama.cpp/.cache/huggingface/download/.gitattributes.lock
ADDED
File without changes
|
llama.cpp/.cache/huggingface/download/.gitattributes.metadata
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
2a8cee5d00b0b7cdb12bf10b2ec30ef7eec53134
|
2 |
+
7a3e23c37facac317a5fbc446d28a2167ace09f1
|
3 |
+
1738945643.956943
|
llama.cpp/.cache/huggingface/download/README.md.lock
ADDED
File without changes
|
llama.cpp/.cache/huggingface/download/README.md.metadata
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
2a8cee5d00b0b7cdb12bf10b2ec30ef7eec53134
|
2 |
+
11d013d8e5248d73e5b0732cd43d6657c84488c1
|
3 |
+
1738945643.734035
|
llama.cpp/.cache/huggingface/download/index.html.lock
ADDED
File without changes
|
llama.cpp/.cache/huggingface/download/index.html.metadata
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
2a8cee5d00b0b7cdb12bf10b2ec30ef7eec53134
|
2 |
+
b0c4b3666032a737f3903db53e6a8a9272483e28
|
3 |
+
1738945643.5221171
|
llama.cpp/.cache/huggingface/download/llama.cpp/.DS_Store.lock
ADDED
File without changes
|
llama.cpp/.cache/huggingface/download/llama.cpp/.DS_Store.metadata
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
2a8cee5d00b0b7cdb12bf10b2ec30ef7eec53134
|
2 |
+
b384d4d95f90147eaf6f703455c11609eeceb4be
|
3 |
+
1738945644.0605388
|
llama.cpp/.cache/huggingface/download/llama.cpp/.clang-format.lock
ADDED
File without changes
|
llama.cpp/.cache/huggingface/download/llama.cpp/.clang-format.metadata
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
2a8cee5d00b0b7cdb12bf10b2ec30ef7eec53134
|
2 |
+
45232b80ed8cd1c7ce3892a9d181d38b131240d7
|
3 |
+
1738945643.594333
|
llama.cpp/.cache/huggingface/download/llama.cpp/.clang-tidy.lock
ADDED
File without changes
|
llama.cpp/.cache/huggingface/download/llama.cpp/.clang-tidy.metadata
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
2a8cee5d00b0b7cdb12bf10b2ec30ef7eec53134
|
2 |
+
310c3d182c8f2ed2c3bc0ddfa9f6fb74f86e39ae
|
3 |
+
1738945643.5322912
|
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/cloud-v-pipeline.lock
ADDED
File without changes
|
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/cloud-v-pipeline.metadata
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
2a8cee5d00b0b7cdb12bf10b2ec30ef7eec53134
|
2 |
+
af8c0cea6155c71c0c162e034715908ea16ee9cf
|
3 |
+
1738945643.9848042
|
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/cpu.Dockerfile.lock
ADDED
File without changes
|
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/cpu.Dockerfile.metadata
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
2a8cee5d00b0b7cdb12bf10b2ec30ef7eec53134
|
2 |
+
522ee8147d63a9e595dcd7c4d71d246705739e4a
|
3 |
+
1738945643.500143
|
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/cuda.Dockerfile.lock
ADDED
File without changes
|
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/cuda.Dockerfile.metadata
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
2a8cee5d00b0b7cdb12bf10b2ec30ef7eec53134
|
2 |
+
974dd78a8b08a779cc2263f0f59ba57a5a4298a8
|
3 |
+
1738945644.01567
|
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/intel.Dockerfile.lock
ADDED
File without changes
|
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/intel.Dockerfile.metadata
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
2a8cee5d00b0b7cdb12bf10b2ec30ef7eec53134
|
2 |
+
af783f5e998ebac28965ed81798709d4f65ec3f1
|
3 |
+
1738945644.030537
|
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/llama-cli-cann.Dockerfile.lock
ADDED
File without changes
|
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/llama-cli-cann.Dockerfile.metadata
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
2a8cee5d00b0b7cdb12bf10b2ec30ef7eec53134
|
2 |
+
02dce501ce28629776e631f3f2e56bbf66df12a7
|
3 |
+
1738945644.513278
|
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/llama-cpp-cuda.srpm.spec.lock
ADDED
File without changes
|
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/llama-cpp-cuda.srpm.spec.metadata
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
2a8cee5d00b0b7cdb12bf10b2ec30ef7eec53134
|
2 |
+
7425d3a9d7a4027fcff13077ffec3caf6e66d02f
|
3 |
+
1738945644.1233048
|
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/llama-cpp.srpm.spec.lock
ADDED
File without changes
|
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/llama-cpp.srpm.spec.metadata
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
2a8cee5d00b0b7cdb12bf10b2ec30ef7eec53134
|
2 |
+
4d5560089816c6679bd02b673dc57a2206ebfbe7
|
3 |
+
1738945644.7103841
|
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/musa.Dockerfile.lock
ADDED
File without changes
|
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/musa.Dockerfile.metadata
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
2a8cee5d00b0b7cdb12bf10b2ec30ef7eec53134
|
2 |
+
bfd7fc1c1740fd5898a8576b5b7e13b6a79d68a0
|
3 |
+
1738945644.481749
|
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/apps.nix.lock
ADDED
File without changes
|
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/apps.nix.metadata
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
2a8cee5d00b0b7cdb12bf10b2ec30ef7eec53134
|
2 |
+
0ecf19fc56d554c69aeef8a03b253fc15338688e
|
3 |
+
1738945644.509313
|
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/devshells.nix.lock
ADDED
File without changes
|
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/devshells.nix.metadata
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
2a8cee5d00b0b7cdb12bf10b2ec30ef7eec53134
|
2 |
+
bfd304af14dcda59e284d30b9fc7491466bf71d1
|
3 |
+
1738945644.542623
|
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/docker.nix.lock
ADDED
File without changes
|
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/docker.nix.metadata
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
2a8cee5d00b0b7cdb12bf10b2ec30ef7eec53134
|
2 |
+
d607b4575772c5330e962649ff6e14e5562ecfad
|
3 |
+
1738945644.998351
|
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/jetson-support.nix.lock
ADDED
File without changes
|
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/jetson-support.nix.metadata
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
2a8cee5d00b0b7cdb12bf10b2ec30ef7eec53134
|
2 |
+
78e2e40e03864e3df046389f7b751a1fd4575656
|
3 |
+
1738945644.5733309
|
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/nixpkgs-instances.nix.lock
ADDED
File without changes
|
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/nixpkgs-instances.nix.metadata
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
2a8cee5d00b0b7cdb12bf10b2ec30ef7eec53134
|
2 |
+
90d683a713aa1f6f7b7e263549afe5edef0a4033
|
3 |
+
1738945644.6345322
|
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/package-gguf-py.nix.lock
ADDED
File without changes
|
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/package-gguf-py.nix.metadata
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
2a8cee5d00b0b7cdb12bf10b2ec30ef7eec53134
|
2 |
+
cca2f36a5bd4df214b3b7da84aa7a05d242fce0e
|
3 |
+
1738945644.985415
|
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/package.nix.lock
ADDED
File without changes
|
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/package.nix.metadata
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
2a8cee5d00b0b7cdb12bf10b2ec30ef7eec53134
|
2 |
+
043c4364b956ac30127e089216be545a499a687c
|
3 |
+
1738945645.4636748
|
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/python-scripts.nix.lock
ADDED
File without changes
|
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/python-scripts.nix.metadata
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
2a8cee5d00b0b7cdb12bf10b2ec30ef7eec53134
|
2 |
+
56ea1827887646424fa08017fe91c4bdcc82c465
|
3 |
+
1738945645.01965
|
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/scope.nix.lock
ADDED
File without changes
|
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/scope.nix.metadata
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
2a8cee5d00b0b7cdb12bf10b2ec30ef7eec53134
|
2 |
+
478e8c4228afad449c3bd396d7ed4a12759a5de3
|
3 |
+
1738945645.050901
|
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/sif.nix.lock
ADDED
File without changes
|