KBaba7 commited on
Commit
a1a4268
·
verified ·
1 Parent(s): 660a3f9

Upload 7646 files

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +103 -0
  2. llama.cpp/.DS_Store +0 -0
  3. llama.cpp/.cache/huggingface/.gitignore +1 -0
  4. llama.cpp/.cache/huggingface/download/.gitattributes.lock +0 -0
  5. llama.cpp/.cache/huggingface/download/.gitattributes.metadata +3 -0
  6. llama.cpp/.cache/huggingface/download/README.md.lock +0 -0
  7. llama.cpp/.cache/huggingface/download/README.md.metadata +3 -0
  8. llama.cpp/.cache/huggingface/download/index.html.lock +0 -0
  9. llama.cpp/.cache/huggingface/download/index.html.metadata +3 -0
  10. llama.cpp/.cache/huggingface/download/llama.cpp/.DS_Store.lock +0 -0
  11. llama.cpp/.cache/huggingface/download/llama.cpp/.DS_Store.metadata +3 -0
  12. llama.cpp/.cache/huggingface/download/llama.cpp/.clang-format.lock +0 -0
  13. llama.cpp/.cache/huggingface/download/llama.cpp/.clang-format.metadata +3 -0
  14. llama.cpp/.cache/huggingface/download/llama.cpp/.clang-tidy.lock +0 -0
  15. llama.cpp/.cache/huggingface/download/llama.cpp/.clang-tidy.metadata +3 -0
  16. llama.cpp/.cache/huggingface/download/llama.cpp/.devops/cloud-v-pipeline.lock +0 -0
  17. llama.cpp/.cache/huggingface/download/llama.cpp/.devops/cloud-v-pipeline.metadata +3 -0
  18. llama.cpp/.cache/huggingface/download/llama.cpp/.devops/cpu.Dockerfile.lock +0 -0
  19. llama.cpp/.cache/huggingface/download/llama.cpp/.devops/cpu.Dockerfile.metadata +3 -0
  20. llama.cpp/.cache/huggingface/download/llama.cpp/.devops/cuda.Dockerfile.lock +0 -0
  21. llama.cpp/.cache/huggingface/download/llama.cpp/.devops/cuda.Dockerfile.metadata +3 -0
  22. llama.cpp/.cache/huggingface/download/llama.cpp/.devops/intel.Dockerfile.lock +0 -0
  23. llama.cpp/.cache/huggingface/download/llama.cpp/.devops/intel.Dockerfile.metadata +3 -0
  24. llama.cpp/.cache/huggingface/download/llama.cpp/.devops/llama-cli-cann.Dockerfile.lock +0 -0
  25. llama.cpp/.cache/huggingface/download/llama.cpp/.devops/llama-cli-cann.Dockerfile.metadata +3 -0
  26. llama.cpp/.cache/huggingface/download/llama.cpp/.devops/llama-cpp-cuda.srpm.spec.lock +0 -0
  27. llama.cpp/.cache/huggingface/download/llama.cpp/.devops/llama-cpp-cuda.srpm.spec.metadata +3 -0
  28. llama.cpp/.cache/huggingface/download/llama.cpp/.devops/llama-cpp.srpm.spec.lock +0 -0
  29. llama.cpp/.cache/huggingface/download/llama.cpp/.devops/llama-cpp.srpm.spec.metadata +3 -0
  30. llama.cpp/.cache/huggingface/download/llama.cpp/.devops/musa.Dockerfile.lock +0 -0
  31. llama.cpp/.cache/huggingface/download/llama.cpp/.devops/musa.Dockerfile.metadata +3 -0
  32. llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/apps.nix.lock +0 -0
  33. llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/apps.nix.metadata +3 -0
  34. llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/devshells.nix.lock +0 -0
  35. llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/devshells.nix.metadata +3 -0
  36. llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/docker.nix.lock +0 -0
  37. llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/docker.nix.metadata +3 -0
  38. llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/jetson-support.nix.lock +0 -0
  39. llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/jetson-support.nix.metadata +3 -0
  40. llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/nixpkgs-instances.nix.lock +0 -0
  41. llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/nixpkgs-instances.nix.metadata +3 -0
  42. llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/package-gguf-py.nix.lock +0 -0
  43. llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/package-gguf-py.nix.metadata +3 -0
  44. llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/package.nix.lock +0 -0
  45. llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/package.nix.metadata +3 -0
  46. llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/python-scripts.nix.lock +0 -0
  47. llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/python-scripts.nix.metadata +3 -0
  48. llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/scope.nix.lock +0 -0
  49. llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/scope.nix.metadata +3 -0
  50. llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/sif.nix.lock +0 -0
.gitattributes CHANGED
@@ -33,3 +33,106 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ llama.cpp/build/bin/libggml-base.dylib filter=lfs diff=lfs merge=lfs -text
37
+ llama.cpp/build/bin/libggml-cpu.dylib filter=lfs diff=lfs merge=lfs -text
38
+ llama.cpp/build/bin/libggml-metal.dylib filter=lfs diff=lfs merge=lfs -text
39
+ llama.cpp/build/bin/libllama.dylib filter=lfs diff=lfs merge=lfs -text
40
+ llama.cpp/build/bin/libllava_shared.dylib filter=lfs diff=lfs merge=lfs -text
41
+ llama.cpp/build/bin/llama-batched filter=lfs diff=lfs merge=lfs -text
42
+ llama.cpp/build/bin/llama-batched-bench filter=lfs diff=lfs merge=lfs -text
43
+ llama.cpp/build/bin/llama-bench filter=lfs diff=lfs merge=lfs -text
44
+ llama.cpp/build/bin/llama-cli filter=lfs diff=lfs merge=lfs -text
45
+ llama.cpp/build/bin/llama-convert-llama2c-to-ggml filter=lfs diff=lfs merge=lfs -text
46
+ llama.cpp/build/bin/llama-cvector-generator filter=lfs diff=lfs merge=lfs -text
47
+ llama.cpp/build/bin/llama-embedding filter=lfs diff=lfs merge=lfs -text
48
+ llama.cpp/build/bin/llama-eval-callback filter=lfs diff=lfs merge=lfs -text
49
+ llama.cpp/build/bin/llama-export-lora filter=lfs diff=lfs merge=lfs -text
50
+ llama.cpp/build/bin/llama-gen-docs filter=lfs diff=lfs merge=lfs -text
51
+ llama.cpp/build/bin/llama-gritlm filter=lfs diff=lfs merge=lfs -text
52
+ llama.cpp/build/bin/llama-imatrix filter=lfs diff=lfs merge=lfs -text
53
+ llama.cpp/build/bin/llama-infill filter=lfs diff=lfs merge=lfs -text
54
+ llama.cpp/build/bin/llama-llava-cli filter=lfs diff=lfs merge=lfs -text
55
+ llama.cpp/build/bin/llama-llava-clip-quantize-cli filter=lfs diff=lfs merge=lfs -text
56
+ llama.cpp/build/bin/llama-lookahead filter=lfs diff=lfs merge=lfs -text
57
+ llama.cpp/build/bin/llama-lookup filter=lfs diff=lfs merge=lfs -text
58
+ llama.cpp/build/bin/llama-lookup-create filter=lfs diff=lfs merge=lfs -text
59
+ llama.cpp/build/bin/llama-lookup-stats filter=lfs diff=lfs merge=lfs -text
60
+ llama.cpp/build/bin/llama-minicpmv-cli filter=lfs diff=lfs merge=lfs -text
61
+ llama.cpp/build/bin/llama-parallel filter=lfs diff=lfs merge=lfs -text
62
+ llama.cpp/build/bin/llama-passkey filter=lfs diff=lfs merge=lfs -text
63
+ llama.cpp/build/bin/llama-perplexity filter=lfs diff=lfs merge=lfs -text
64
+ llama.cpp/build/bin/llama-quantize filter=lfs diff=lfs merge=lfs -text
65
+ llama.cpp/build/bin/llama-quantize-stats filter=lfs diff=lfs merge=lfs -text
66
+ llama.cpp/build/bin/llama-qwen2vl-cli filter=lfs diff=lfs merge=lfs -text
67
+ llama.cpp/build/bin/llama-retrieval filter=lfs diff=lfs merge=lfs -text
68
+ llama.cpp/build/bin/llama-run filter=lfs diff=lfs merge=lfs -text
69
+ llama.cpp/build/bin/llama-save-load-state filter=lfs diff=lfs merge=lfs -text
70
+ llama.cpp/build/bin/llama-server filter=lfs diff=lfs merge=lfs -text
71
+ llama.cpp/build/bin/llama-speculative filter=lfs diff=lfs merge=lfs -text
72
+ llama.cpp/build/bin/llama-speculative-simple filter=lfs diff=lfs merge=lfs -text
73
+ llama.cpp/build/bin/llama-tokenize filter=lfs diff=lfs merge=lfs -text
74
+ llama.cpp/build/bin/llama-tts filter=lfs diff=lfs merge=lfs -text
75
+ llama.cpp/build/bin/test-arg-parser filter=lfs diff=lfs merge=lfs -text
76
+ llama.cpp/build/bin/test-backend-ops filter=lfs diff=lfs merge=lfs -text
77
+ llama.cpp/build/bin/test-chat filter=lfs diff=lfs merge=lfs -text
78
+ llama.cpp/build/bin/test-chat-template filter=lfs diff=lfs merge=lfs -text
79
+ llama.cpp/build/bin/test-grammar-integration filter=lfs diff=lfs merge=lfs -text
80
+ llama.cpp/build/bin/test-json-schema-to-grammar filter=lfs diff=lfs merge=lfs -text
81
+ llama.cpp/build/bin/test-tokenizer-0 filter=lfs diff=lfs merge=lfs -text
82
+ llama.cpp/build/bin/test-tokenizer-1-bpe filter=lfs diff=lfs merge=lfs -text
83
+ llama.cpp/build/bin/test-tokenizer-1-spm filter=lfs diff=lfs merge=lfs -text
84
+ llama.cpp/build/common/CMakeFiles/common.dir/arg.cpp.o filter=lfs diff=lfs merge=lfs -text
85
+ llama.cpp/build/common/CMakeFiles/common.dir/chat.cpp.o filter=lfs diff=lfs merge=lfs -text
86
+ llama.cpp/build/common/CMakeFiles/common.dir/common.cpp.o filter=lfs diff=lfs merge=lfs -text
87
+ llama.cpp/build/common/CMakeFiles/common.dir/json-schema-to-grammar.cpp.o filter=lfs diff=lfs merge=lfs -text
88
+ llama.cpp/build/common/libcommon.a filter=lfs diff=lfs merge=lfs -text
89
+ llama.cpp/build/examples/llama-bench/CMakeFiles/llama-bench.dir/llama-bench.cpp.o filter=lfs diff=lfs merge=lfs -text
90
+ llama.cpp/build/examples/llava/CMakeFiles/llava.dir/clip.cpp.o filter=lfs diff=lfs merge=lfs -text
91
+ llama.cpp/build/examples/llava/libllava_static.a filter=lfs diff=lfs merge=lfs -text
92
+ llama.cpp/build/examples/perplexity/CMakeFiles/llama-perplexity.dir/perplexity.cpp.o filter=lfs diff=lfs merge=lfs -text
93
+ llama.cpp/build/examples/quantize-stats/CMakeFiles/llama-quantize-stats.dir/quantize-stats.cpp.o filter=lfs diff=lfs merge=lfs -text
94
+ llama.cpp/build/examples/run/CMakeFiles/llama-run.dir/run.cpp.o filter=lfs diff=lfs merge=lfs -text
95
+ llama.cpp/build/examples/server/CMakeFiles/llama-server.dir/server.cpp.o filter=lfs diff=lfs merge=lfs -text
96
+ llama.cpp/build/examples/tts/CMakeFiles/llama-tts.dir/tts.cpp.o filter=lfs diff=lfs merge=lfs -text
97
+ llama.cpp/build/ggml/src/CMakeFiles/ggml-base.dir/ggml-quants.c.o filter=lfs diff=lfs merge=lfs -text
98
+ llama.cpp/build/ggml/src/CMakeFiles/ggml-base.dir/ggml.c.o filter=lfs diff=lfs merge=lfs -text
99
+ llama.cpp/build/ggml/src/CMakeFiles/ggml-base.dir/gguf.cpp.o filter=lfs diff=lfs merge=lfs -text
100
+ llama.cpp/build/ggml/src/CMakeFiles/ggml-cpu.dir/ggml-cpu/ggml-cpu.c.o filter=lfs diff=lfs merge=lfs -text
101
+ llama.cpp/build/ggml/src/CMakeFiles/ggml-cpu.dir/ggml-cpu/llamafile/sgemm.cpp.o filter=lfs diff=lfs merge=lfs -text
102
+ llama.cpp/build/ggml/src/ggml-metal/CMakeFiles/ggml-metal.dir/__/__/__/autogenerated/ggml-metal-embed.s.o filter=lfs diff=lfs merge=lfs -text
103
+ llama.cpp/build/ggml/src/ggml-metal/CMakeFiles/ggml-metal.dir/ggml-metal.m.o filter=lfs diff=lfs merge=lfs -text
104
+ llama.cpp/build/src/CMakeFiles/llama.dir/llama-model-loader.cpp.o filter=lfs diff=lfs merge=lfs -text
105
+ llama.cpp/build/src/CMakeFiles/llama.dir/llama-model.cpp.o filter=lfs diff=lfs merge=lfs -text
106
+ llama.cpp/build/src/CMakeFiles/llama.dir/llama-sampling.cpp.o filter=lfs diff=lfs merge=lfs -text
107
+ llama.cpp/build/src/CMakeFiles/llama.dir/llama-vocab.cpp.o filter=lfs diff=lfs merge=lfs -text
108
+ llama.cpp/build/src/CMakeFiles/llama.dir/llama.cpp.o filter=lfs diff=lfs merge=lfs -text
109
+ llama.cpp/build/src/CMakeFiles/llama.dir/unicode.cpp.o filter=lfs diff=lfs merge=lfs -text
110
+ llama.cpp/build/tests/CMakeFiles/test-backend-ops.dir/test-backend-ops.cpp.o filter=lfs diff=lfs merge=lfs -text
111
+ llama.cpp/build/tests/CMakeFiles/test-chat-template.dir/test-chat-template.cpp.o filter=lfs diff=lfs merge=lfs -text
112
+ llama.cpp/build/tests/CMakeFiles/test-chat.dir/test-chat.cpp.o filter=lfs diff=lfs merge=lfs -text
113
+ llama.cpp/build/tests/CMakeFiles/test-grammar-integration.dir/test-grammar-integration.cpp.o filter=lfs diff=lfs merge=lfs -text
114
+ llama.cpp/build/tests/CMakeFiles/test-json-schema-to-grammar.dir/test-json-schema-to-grammar.cpp.o filter=lfs diff=lfs merge=lfs -text
115
+ llama.cpp/docs/development/llama-star/idea-arch.key filter=lfs diff=lfs merge=lfs -text
116
+ llama.cpp/examples/server/themes/buttons-top/buttons_top.png filter=lfs diff=lfs merge=lfs -text
117
+ llama.cpp/examples/server/themes/wild/llamapattern.png filter=lfs diff=lfs merge=lfs -text
118
+ llama.cpp/examples/server/themes/wild/wild.png filter=lfs diff=lfs merge=lfs -text
119
+ llama.cpp/gguf-py/gguf/__pycache__/quants.cpython-311.pyc filter=lfs diff=lfs merge=lfs -text
120
+ llama.cpp/media/llama0-banner.png filter=lfs diff=lfs merge=lfs -text
121
+ llama.cpp/media/llama0-logo.png filter=lfs diff=lfs merge=lfs -text
122
+ llama.cpp/media/matmul.png filter=lfs diff=lfs merge=lfs -text
123
+ llama.cpp/models/ggml-vocab-aquila.gguf filter=lfs diff=lfs merge=lfs -text
124
+ llama.cpp/models/ggml-vocab-baichuan.gguf filter=lfs diff=lfs merge=lfs -text
125
+ llama.cpp/models/ggml-vocab-bert-bge.gguf filter=lfs diff=lfs merge=lfs -text
126
+ llama.cpp/models/ggml-vocab-command-r.gguf filter=lfs diff=lfs merge=lfs -text
127
+ llama.cpp/models/ggml-vocab-deepseek-coder.gguf filter=lfs diff=lfs merge=lfs -text
128
+ llama.cpp/models/ggml-vocab-deepseek-llm.gguf filter=lfs diff=lfs merge=lfs -text
129
+ llama.cpp/models/ggml-vocab-falcon.gguf filter=lfs diff=lfs merge=lfs -text
130
+ llama.cpp/models/ggml-vocab-gpt-2.gguf filter=lfs diff=lfs merge=lfs -text
131
+ llama.cpp/models/ggml-vocab-gpt-neox.gguf filter=lfs diff=lfs merge=lfs -text
132
+ llama.cpp/models/ggml-vocab-llama-bpe.gguf filter=lfs diff=lfs merge=lfs -text
133
+ llama.cpp/models/ggml-vocab-llama-spm.gguf filter=lfs diff=lfs merge=lfs -text
134
+ llama.cpp/models/ggml-vocab-mpt.gguf filter=lfs diff=lfs merge=lfs -text
135
+ llama.cpp/models/ggml-vocab-phi-3.gguf filter=lfs diff=lfs merge=lfs -text
136
+ llama.cpp/models/ggml-vocab-qwen2.gguf filter=lfs diff=lfs merge=lfs -text
137
+ llama.cpp/models/ggml-vocab-refact.gguf filter=lfs diff=lfs merge=lfs -text
138
+ llama.cpp/models/ggml-vocab-starcoder.gguf filter=lfs diff=lfs merge=lfs -text
llama.cpp/.DS_Store ADDED
Binary file (10.2 kB). View file
 
llama.cpp/.cache/huggingface/.gitignore ADDED
@@ -0,0 +1 @@
 
 
1
+ *
llama.cpp/.cache/huggingface/download/.gitattributes.lock ADDED
File without changes
llama.cpp/.cache/huggingface/download/.gitattributes.metadata ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ 2a8cee5d00b0b7cdb12bf10b2ec30ef7eec53134
2
+ 7a3e23c37facac317a5fbc446d28a2167ace09f1
3
+ 1738945643.956943
llama.cpp/.cache/huggingface/download/README.md.lock ADDED
File without changes
llama.cpp/.cache/huggingface/download/README.md.metadata ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ 2a8cee5d00b0b7cdb12bf10b2ec30ef7eec53134
2
+ 11d013d8e5248d73e5b0732cd43d6657c84488c1
3
+ 1738945643.734035
llama.cpp/.cache/huggingface/download/index.html.lock ADDED
File without changes
llama.cpp/.cache/huggingface/download/index.html.metadata ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ 2a8cee5d00b0b7cdb12bf10b2ec30ef7eec53134
2
+ b0c4b3666032a737f3903db53e6a8a9272483e28
3
+ 1738945643.5221171
llama.cpp/.cache/huggingface/download/llama.cpp/.DS_Store.lock ADDED
File without changes
llama.cpp/.cache/huggingface/download/llama.cpp/.DS_Store.metadata ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ 2a8cee5d00b0b7cdb12bf10b2ec30ef7eec53134
2
+ b384d4d95f90147eaf6f703455c11609eeceb4be
3
+ 1738945644.0605388
llama.cpp/.cache/huggingface/download/llama.cpp/.clang-format.lock ADDED
File without changes
llama.cpp/.cache/huggingface/download/llama.cpp/.clang-format.metadata ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ 2a8cee5d00b0b7cdb12bf10b2ec30ef7eec53134
2
+ 45232b80ed8cd1c7ce3892a9d181d38b131240d7
3
+ 1738945643.594333
llama.cpp/.cache/huggingface/download/llama.cpp/.clang-tidy.lock ADDED
File without changes
llama.cpp/.cache/huggingface/download/llama.cpp/.clang-tidy.metadata ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ 2a8cee5d00b0b7cdb12bf10b2ec30ef7eec53134
2
+ 310c3d182c8f2ed2c3bc0ddfa9f6fb74f86e39ae
3
+ 1738945643.5322912
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/cloud-v-pipeline.lock ADDED
File without changes
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/cloud-v-pipeline.metadata ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ 2a8cee5d00b0b7cdb12bf10b2ec30ef7eec53134
2
+ af8c0cea6155c71c0c162e034715908ea16ee9cf
3
+ 1738945643.9848042
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/cpu.Dockerfile.lock ADDED
File without changes
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/cpu.Dockerfile.metadata ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ 2a8cee5d00b0b7cdb12bf10b2ec30ef7eec53134
2
+ 522ee8147d63a9e595dcd7c4d71d246705739e4a
3
+ 1738945643.500143
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/cuda.Dockerfile.lock ADDED
File without changes
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/cuda.Dockerfile.metadata ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ 2a8cee5d00b0b7cdb12bf10b2ec30ef7eec53134
2
+ 974dd78a8b08a779cc2263f0f59ba57a5a4298a8
3
+ 1738945644.01567
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/intel.Dockerfile.lock ADDED
File without changes
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/intel.Dockerfile.metadata ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ 2a8cee5d00b0b7cdb12bf10b2ec30ef7eec53134
2
+ af783f5e998ebac28965ed81798709d4f65ec3f1
3
+ 1738945644.030537
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/llama-cli-cann.Dockerfile.lock ADDED
File without changes
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/llama-cli-cann.Dockerfile.metadata ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ 2a8cee5d00b0b7cdb12bf10b2ec30ef7eec53134
2
+ 02dce501ce28629776e631f3f2e56bbf66df12a7
3
+ 1738945644.513278
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/llama-cpp-cuda.srpm.spec.lock ADDED
File without changes
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/llama-cpp-cuda.srpm.spec.metadata ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ 2a8cee5d00b0b7cdb12bf10b2ec30ef7eec53134
2
+ 7425d3a9d7a4027fcff13077ffec3caf6e66d02f
3
+ 1738945644.1233048
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/llama-cpp.srpm.spec.lock ADDED
File without changes
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/llama-cpp.srpm.spec.metadata ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ 2a8cee5d00b0b7cdb12bf10b2ec30ef7eec53134
2
+ 4d5560089816c6679bd02b673dc57a2206ebfbe7
3
+ 1738945644.7103841
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/musa.Dockerfile.lock ADDED
File without changes
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/musa.Dockerfile.metadata ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ 2a8cee5d00b0b7cdb12bf10b2ec30ef7eec53134
2
+ bfd7fc1c1740fd5898a8576b5b7e13b6a79d68a0
3
+ 1738945644.481749
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/apps.nix.lock ADDED
File without changes
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/apps.nix.metadata ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ 2a8cee5d00b0b7cdb12bf10b2ec30ef7eec53134
2
+ 0ecf19fc56d554c69aeef8a03b253fc15338688e
3
+ 1738945644.509313
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/devshells.nix.lock ADDED
File without changes
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/devshells.nix.metadata ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ 2a8cee5d00b0b7cdb12bf10b2ec30ef7eec53134
2
+ bfd304af14dcda59e284d30b9fc7491466bf71d1
3
+ 1738945644.542623
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/docker.nix.lock ADDED
File without changes
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/docker.nix.metadata ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ 2a8cee5d00b0b7cdb12bf10b2ec30ef7eec53134
2
+ d607b4575772c5330e962649ff6e14e5562ecfad
3
+ 1738945644.998351
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/jetson-support.nix.lock ADDED
File without changes
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/jetson-support.nix.metadata ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ 2a8cee5d00b0b7cdb12bf10b2ec30ef7eec53134
2
+ 78e2e40e03864e3df046389f7b751a1fd4575656
3
+ 1738945644.5733309
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/nixpkgs-instances.nix.lock ADDED
File without changes
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/nixpkgs-instances.nix.metadata ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ 2a8cee5d00b0b7cdb12bf10b2ec30ef7eec53134
2
+ 90d683a713aa1f6f7b7e263549afe5edef0a4033
3
+ 1738945644.6345322
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/package-gguf-py.nix.lock ADDED
File without changes
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/package-gguf-py.nix.metadata ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ 2a8cee5d00b0b7cdb12bf10b2ec30ef7eec53134
2
+ cca2f36a5bd4df214b3b7da84aa7a05d242fce0e
3
+ 1738945644.985415
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/package.nix.lock ADDED
File without changes
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/package.nix.metadata ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ 2a8cee5d00b0b7cdb12bf10b2ec30ef7eec53134
2
+ 043c4364b956ac30127e089216be545a499a687c
3
+ 1738945645.4636748
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/python-scripts.nix.lock ADDED
File without changes
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/python-scripts.nix.metadata ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ 2a8cee5d00b0b7cdb12bf10b2ec30ef7eec53134
2
+ 56ea1827887646424fa08017fe91c4bdcc82c465
3
+ 1738945645.01965
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/scope.nix.lock ADDED
File without changes
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/scope.nix.metadata ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ 2a8cee5d00b0b7cdb12bf10b2ec30ef7eec53134
2
+ 478e8c4228afad449c3bd396d7ed4a12759a5de3
3
+ 1738945645.050901
llama.cpp/.cache/huggingface/download/llama.cpp/.devops/nix/sif.nix.lock ADDED
File without changes