Text Generation
Chinese
audreyt commited on
Commit
244735f
1 Parent(s): df19a2e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -8
README.md CHANGED
@@ -24,20 +24,18 @@ quantized_by: Audrey Tang
24
 
25
  This repo contains GGML format model files for [Yen-Ting Lin's Language Models for Taiwanese Culture v1.0](https://huggingface.co/yentinglin/Taiwan-LLaMa-v1.0).
26
 
27
- They are known to work with:
28
- * [llama.cpp](https://github.com/ggerganov/llama.cpp), commit `e76d630` and later.
29
 
30
- ...and probably work with these too, but I have not tested personally:
31
- * [text-generation-webui](https://github.com/oobabooga/text-generation-webui).
32
- * [KoboldCpp](https://github.com/LostRuins/koboldcpp), version 1.37 and later.
33
- * [llama-cpp-python](https://github.com/abetlen/llama-cpp-python), version 0.1.77 and later.
34
 
35
  ## Repositories available
36
 
37
- * [2, 3, 4, 5, 6 and 8-bit GGML models for CPU+GPU inference](https://huggingface.co/audreyt/Taiwan-LLaMa-v1.0-GGML)
 
38
  * [Yen-Ting Lin's original unquantised fp16 model in pytorch format, for GPU inference and for further conversions](https://huggingface.co/yentinglin/Taiwan-LLaMa-v1.0)
39
 
40
-
41
  <!-- footer start -->
42
  <!-- footer end -->
43
 
 
24
 
25
  This repo contains GGML format model files for [Yen-Ting Lin's Language Models for Taiwanese Culture v1.0](https://huggingface.co/yentinglin/Taiwan-LLaMa-v1.0).
26
 
27
+ ### Important note regarding GGML files.
 
28
 
29
+ The GGML format has now been superseded by GGUF. As of August 21st 2023, [llama.cpp](https://github.com/ggerganov/llama.cpp) no longer supports GGML models. Third party clients and libraries are expected to still support it for a time, but many may also drop support.
30
+
31
+ Please use the GGUF models instead.
 
32
 
33
  ## Repositories available
34
 
35
+ * [2, 3, 4, 5, 6 and 8-bit GGUF models for CPU+GPU inference](https://huggingface.co/audreyt/Taiwan-LLaMa-v1.0-GGUF)
36
+ * [2, 3, 4, 5, 6 and 8-bit GGML models for CPU+GPU inference (deprecated)](https://huggingface.co/audreyt/Taiwan-LLaMa-v1.0-GGML)
37
  * [Yen-Ting Lin's original unquantised fp16 model in pytorch format, for GPU inference and for further conversions](https://huggingface.co/yentinglin/Taiwan-LLaMa-v1.0)
38
 
 
39
  <!-- footer start -->
40
  <!-- footer end -->
41