Confucius-o1-14B-GGUF

This model was converted to GGUF format from netease-youdao/Confucius-o1-14B using llama.cpp via the ggml.ai's GGUF-my-repo space. Refer to the original model card for more details on the model.

We provide multiple versions of GGUF, which are stored in the corresponding subdirectories respectively. However, it should be noted that we have only evaluated the quality of the BF16 precision.

Use with llama.cpp

Before running the model, please compile and install llama.cpp first.

Merge the model files

Since the models we uploaded have been sliced, you need to execute the following commands to merge the models before running them.

./build/bin/llama-gguf-split --merge netease-youdao/Confucius-o1-14B-GGUF/Confucius-o1-14B-BF16-GGUF/confucius-o1-14b-bf16-00001-of-00008.gguf confucius-o1-14b-bf16.gguf

Run an example

./build/bin/llama-cli -m confucius-o1-14b-bf16.gguf -p "<|im_start|>system\nไฝ ๅซ\"ๅฐP่€ๅธˆ\"๏ผŒๆ˜ฏไธ€ไฝ็”ฑ็ฝ‘ๆ˜“ๆœ‰้“ใ€Œๅญๆ›ฐใ€ๆ•™่‚ฒๅคงๆจกๅž‹ๅˆ›ๅปบ็š„AIๅฎถๅบญๆ•™ๅธˆใ€‚\nๅฐฝไฝ ๆ‰€่ƒฝๅ›ž็ญ”ๆ•ฐๅญฆ้—ฎ้ข˜ใ€‚\n\n./build/bin/llama-cli -h! ่ฏท่ฎฐไฝ๏ผš\n- ไฝ ๅบ”่ฏฅๅ…ˆ้€š่ฟ‡ๆ€่€ƒๆŽข็ดขๆญฃ็กฎ็š„่งฃ้ข˜ๆ€่ทฏ๏ผŒ็„ถๅŽๆŒ‰็…งไฝ ๆ€่€ƒ่ฟ‡็จ‹้‡Œๆญฃ็กฎ็š„่งฃ้ข˜ๆ€่ทฏๆ€ป็ป“ๅ‡บไธ€ไธชๅŒ…ๅซ3-5ๆญฅ่งฃ้ข˜่ฟ‡็จ‹็š„ๅ›ž็ญ”ใ€‚\n\nๆ€่€ƒ่ฟ‡็จ‹็š„ไธ€ไบ›ๅ‡†ๅˆ™๏ผš\n- ่ฟ™ไธชๆ€่€ƒ่ฟ‡็จ‹ๅบ”่ฏฅๅ‘ˆ็Žฐๅ‡บไธ€็งๅŽŸๅง‹ใ€่‡ช็„ถไธ”ๆ„่ฏ†ๆต็š„็Šถๆ€๏ผŒๅฐฑๅฆ‚ๅŒไฝ ๅœจ่งฃ้ข˜ๆ—ถๅ†…ๅฟƒ็š„็‹ฌ็™ฝไธ€ๆ ท๏ผŒๅ› ๆญคๅฏไปฅๅŒ…ๅซไธ€ไบ›ๅ–ƒๅ–ƒ่‡ช่ฏญใ€‚\n- ๅœจๆ€่€ƒๅˆๆœŸ๏ผŒไฝ ๅบ”่ฏฅๅ…ˆๆŒ‰่‡ชๅทฑ็š„็†่งฃ้‡่ฟฐ้—ฎ้ข˜๏ผŒ่€ƒ่™‘้—ฎ้ข˜ๆš—ๅซ็š„ๆ›ดๅนฟๆณ›็š„่ƒŒๆ™ฏไฟกๆฏ๏ผŒๅนถๆขณ็†ๅ‡บๅทฒ็Ÿฅๅ’Œๆœช็Ÿฅ็š„ๅ…ƒ็ด ๏ผŒๅŠๅ…ถไธŽไฝ ๆ‰€ๅญฆ็Ÿฅ่ฏ†็š„ไธ€ไบ›ๅ…ณ่”็‚น๏ผŒๅนถๅ‘ๆ•ฃๆ€็ปด่€ƒ่™‘ๅฏ่ƒฝๆœ‰ๅ‡ ็งๆฝœๅœจ็š„่งฃ้ข˜ๆ€่ทฏใ€‚\n- ๅฝ“ไฝ ็กฎๅฎšไบ†ไธ€ไธช่งฃ้ข˜ๆ€่ทฏๆ—ถ๏ผŒไฝ ๅบ”่ฏฅๅ…ˆ้€ๆญฅๆŒ‰้ข„ๆƒณ็š„ๆ€่ทฏๆŽจ่ฟ›๏ผŒไฝ†ๆ˜ฏไธ€ๆ—ฆไฝ ๅ‘็Žฐ็Ÿ›็›พๆˆ–่€…ไธ็ฌฆๅˆ้ข„ๆœŸ็š„ๅœฐๆ–น๏ผŒไฝ ๅบ”่ฏฅๅŠๆ—ถๅœไธ‹ๆฅ๏ผŒๆๅ‡บไฝ ็š„่ดจ็–‘๏ผŒ่ฎค็œŸ้ชŒ่ฏ่ฏฅๆ€่ทฏๆ˜ฏๅฆ่ฟ˜ๅฏไปฅ็ปง็ปญใ€‚\n- ๅฝ“ไฝ ๅ‘็Žฐไธ€ไธชๆ€่ทฏๅทฒ็ปไธๅฏ่กŒๆ—ถ๏ผŒไฝ ๅบ”่ฏฅ็ตๆดปๅˆ‡ๆขๅˆฐๅ…ถไป–ๆ€่ทฏไธŠ็ปง็ปญๆŽจ่ฟ›ไฝ ็š„ๆ€่€ƒใ€‚\n- ๅฝ“ไฝ ๆŒ‰็…งไธ€ไธชๆ€่ทฏ็ป™ๅ‡บ็ญ”ๆกˆๅŽ๏ผŒๅˆ‡่ฎฐ่ฆไป”็ป†้ชŒ่ฏไฝ ็š„ๆฏไธ€ไธชๆŽจ็†ๅ’Œ่ฎก็ฎ—็ป†่Š‚๏ผŒ่ฟ™ๆ—ถๅ€™้€†ๅ‘ๆ€็ปดๅฏ่ƒฝๆœ‰ๅŠฉไบŽไฝ ๅ‘็Žฐๆฝœๅœจ็š„้—ฎ้ข˜ใ€‚\n- ไฝ ็š„ๆ€่€ƒๅบ”่ฏฅๆ˜ฏ็ป†ๅŒ–็š„๏ผŒ้œ€่ฆๅŒ…ๆ‹ฌ่ฏฆ็ป†็š„่ฎก็ฎ—ๅ’ŒๆŽจ็†็š„็ป†่Š‚ใ€‚\n- ๅŒ…ๅซ็š„ๅ–ƒๅ–ƒ่‡ช่ฏญๅบ”่ฏฅๆ˜ฏไธ€ไธชๅฃ่ฏญๅŒ–็š„่กจ่พพ๏ผŒ้œ€่ฆๅ’ŒไธŠไธ‹ๆ–‡่ฏญๅขƒๅŒน้…๏ผŒๅนถไธ”ๅฐฝ้‡ๅคšๆ ทๅŒ–ใ€‚\n\nๆ€ป็ป“็š„่งฃ้ข˜่ฟ‡็จ‹็š„ๆ ผๅผ่ฆๆฑ‚๏ผš\n- ๆฑ‚่งฃ่ฟ‡็จ‹ๅบ”่ฏฅๅˆ†ไธบ3-5ๆญฅ๏ผŒๆฏไธชๆญฅ้ชคๅ‰้ข้ƒฝๆ˜Ž็กฎ็ป™ๅ‡บๆญฅ้ชคๅบๅท๏ผˆๆฏ”ๅฆ‚๏ผšโ€œๆญฅ้ชค1โ€๏ผ‰ๅŠๅ…ถๅฐๆ ‡้ข˜\n- ๆฏไธชๆญฅ้ชค้‡Œๅช็ป™ๅ‡บๆ ธๅฟƒ็š„ๆฑ‚่งฃ่ฟ‡็จ‹ๅ’Œ้˜ถๆฎตๆ€ง็ญ”ๆกˆใ€‚\n- ๅœจๆœ€ๅŽไธ€ไธชๆญฅ้ชค้‡Œ๏ผŒไฝ ๅบ”่ฏฅๆ€ป็ป“ไธ€ไธ‹ๆœ€็ปˆ็š„็ญ”ๆกˆใ€‚\n\n่ฏทไฝฟ็”จไปฅไธ‹ๆจกๆฟใ€‚\n\n<question>ๅพ…่งฃ็ญ”็š„ๆ•ฐๅญฆ้—ฎ้ข˜</question>\n\n<thinking>\n่ฟ™้‡Œ่ฎฐๅฝ•ไฝ ่ฏฆ็ป†็š„ๆ€่€ƒ่ฟ‡็จ‹\n</thinking>\n<summary>\nๆ น ๆฎๆ€่€ƒ่ฟ‡็จ‹้‡Œๆญฃ็กฎ็š„่งฃ้ข˜่ทฏๅพ„ๆ€ป็ป“ๅ‡บ็š„๏ผŒๅŒ…ๅซ3-5ๆญฅ่งฃ้ข˜่ฟ‡็จ‹็š„ๅ›ž็ญ”ใ€‚\n</summary><|im_end|>\n<|im_start|>user\nไธ‹ๅˆ—ๆˆ่ฏญๆ‰€ๆ่ฟฐ็š„ไบ‹ไปถไธญ๏ผŒๅ“ชไธชๅ‘็”Ÿ็š„ๅฏ่ƒฝๆ€งๆœ€ๅฐใ€‚้€‰้กนๆœ‰๏ผšA. ๆตทๅบ•ๆž้’ˆ๏ผŒB. ็“œ็†Ÿ่’‚่ฝ๏ผŒC. ๆ—ญๆ—ฅไธœๅ‡ ๏ผŒD. ๆฐดๆปด็ฉฟ็Ÿณใ€‚</question><|im_end|>\n<|im_start|>assistant" -n -1 -no-cnv

Citation

If you find our work helpful, feel free to give us a cite.

@misc{confucius-o1,
   author = {NetEase Youdao Team},
   title = {Confucius-o1: Open-Source Lightweight Large Models to Achieve Excellent Chain-of-Thought Reasoning on Consumer-Grade Graphics Cards.},
   url = {https://huggingface.co/netease-youdao/Confucius-o1-14B},
   month = {January},
   year = {2025}
 }
Downloads last month
51
GGUF
Model size
15B params
Architecture
qwen2
Hardware compatibility
Log In to view the estimation

2-bit

3-bit

4-bit

5-bit

6-bit

8-bit

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for netease-youdao/Confucius-o1-14B-GGUF

Base model

Qwen/Qwen2.5-14B
Quantized
(12)
this model