ysn-rfd commited on
Commit
72f2ef8
·
verified ·
1 Parent(s): 281985d

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +25 -40
README.md CHANGED
@@ -1,53 +1,38 @@
1
  ---
2
- base_model: apple/OpenELM-3B-Instruct
3
- license: apple-amlr
4
- license_name: apple-sample-code-license
5
- license_link: LICENSE
 
6
  tags:
 
 
 
 
 
7
  - llama-cpp
8
  - gguf-my-repo
9
  ---
10
 
11
- # ysn-rfd/OpenELM-3B-Instruct-GGUF
12
- This model was converted to GGUF format from [`apple/OpenELM-3B-Instruct`](https://huggingface.co/apple/OpenELM-3B-Instruct) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
13
- Refer to the [original model card](https://huggingface.co/apple/OpenELM-3B-Instruct) for more details on the model.
14
-
15
- ## Use with llama.cpp
16
- Install llama.cpp through brew (works on Mac and Linux)
17
-
18
  ```bash
19
- brew install llama.cpp
20
-
21
  ```
22
- Invoke the llama.cpp server or the CLI.
23
-
24
- ### CLI:
25
  ```bash
26
- llama-cli --hf-repo ysn-rfd/OpenELM-3B-Instruct-GGUF --hf-file openelm-3b-instruct-q4_0.gguf -p "The meaning to life and the universe is"
27
  ```
28
-
29
- ### Server:
30
  ```bash
31
- llama-server --hf-repo ysn-rfd/OpenELM-3B-Instruct-GGUF --hf-file openelm-3b-instruct-q4_0.gguf -c 2048
32
- ```
33
-
34
- Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
35
-
36
- Step 1: Clone llama.cpp from GitHub.
37
- ```
38
- git clone https://github.com/ggerganov/llama.cpp
39
  ```
40
-
41
- Step 2: Move into the llama.cpp folder and build it with `LLAMA_CURL=1` flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux).
42
- ```
43
- cd llama.cpp && LLAMA_CURL=1 make
44
- ```
45
-
46
- Step 3: Run inference through the main binary.
47
- ```
48
- ./llama-cli --hf-repo ysn-rfd/OpenELM-3B-Instruct-GGUF --hf-file openelm-3b-instruct-q4_0.gguf -p "The meaning to life and the universe is"
49
- ```
50
- or
51
- ```
52
- ./llama-server --hf-repo ysn-rfd/OpenELM-3B-Instruct-GGUF --hf-file openelm-3b-instruct-q4_0.gguf -c 2048
53
  ```
 
1
  ---
2
+ license: apache-2.0
3
+ datasets: Microsoft/ChatBench
4
+ language: en
5
+ base_model: microsoft/chatbench-distilgpt2
6
+ library_name: transformers
7
  tags:
8
+ - Microsoft
9
+ - ChatBench
10
+ - Interactive Benchmark
11
+ - User Simulator
12
+ - Benchmarking
13
  - llama-cpp
14
  - gguf-my-repo
15
  ---
16
 
17
+ # chatbench-distilgpt2
18
+ **Model creator:** [microsoft](https://huggingface.co/microsoft)<br/>
19
+ **Original model**: [microsoft/chatbench-distilgpt2](https://huggingface.co/microsoft/chatbench-distilgpt2)<br/>
20
+ **GGUF quantization:** provided by [ysn-rfd](https:/huggingface.co/ysn-rfd) using `llama.cpp`<br/>
21
+ ## Special thanks
22
+ 🙏 Special thanks to [Georgi Gerganov](https://github.com/ggerganov) and the whole team working on [llama.cpp](https://github.com/ggerganov/llama.cpp/) for making all of this possible.
23
+ ## Use with Ollama
24
  ```bash
25
+ ollama run "hf.co/ysn-rfd/OpenELM-3B-Instruct-GGUF:F16"
 
26
  ```
27
+ ## Use with LM Studio
 
 
28
  ```bash
29
+ lms load "ysn-rfd/OpenELM-3B-Instruct-GGUF"
30
  ```
31
+ ## Use with llama.cpp CLI
 
32
  ```bash
33
+ llama-cli --hf "ysn-rfd/OpenELM-3B-Instruct-GGUF:F16" -p "The meaning to life and the universe is"
 
 
 
 
 
 
 
34
  ```
35
+ ## Use with llama.cpp Server:
36
+ ```bash
37
+ llama-server --hf "ysn-rfd/OpenELM-3B-Instruct-GGUF:F16" -c 4096
 
 
 
 
 
 
 
 
 
 
38
  ```