Was this helpful?
<|im_system|>system<|im_middle|>You are Kimi, an AI assistant created by Moonshot AI.<|im_end|><|im_user|>user<|im_middle|>What is 1+1?<|im_end|><|im_assistant|>assistant<|im_middle|><think>curl -fsSL https://unsloth.ai/install.sh | shirm https://unsloth.ai/install.ps1 | iexunsloth studio -H 0.0.0.0 -p 8888apt-get update
apt-get install pciutils build-essential cmake curl libcurl4-openssl-dev -y
git clone https://github.com/ggml-org/llama.cpp
cmake llama.cpp -B llama.cpp/build \
-DBUILD_SHARED_LIBS=OFF -DGGML_CUDA=ON
cmake --build llama.cpp/build --config Release -j --clean-first --target llama-cli llama-mtmd-cli llama-server llama-gguf-split
cp llama.cpp/build/bin/llama-* llama.cppexport LLAMA_CACHE="unsloth/Kimi-K2.6-GGUF"
./llama.cpp/llama-cli \
-hf unsloth/Kimi-K2.6-GGUF:UD-Q2_K_XL \
--temp 1.0 \
--top-p 0.95export LLAMA_CACHE="unsloth/Qwen3.6-35B-A3B-GGUF"
./llama.cpp/llama-cli \
-hf unsloth/Kimi-K2.6-GGUF:UD-Q2_K_XL \
--temp 0.6 \
--top-p 0.95 \
--chat-template-kwargs '{"enable_thinking":false}'hf download unsloth/Kimi-K2.6-GGUF \
--local-dir unsloth/Kimi-K2.6-GGUF \
--include "*mmproj-F16*" \
--include "*UD-Q2_K_XL*" # Use "*UD-Q8_K_XL*" for full precision./llama.cpp/llama-cli \
--model unsloth/Kimi-K2.6-GGUF/UD-Q2_K_XL/Kimi-K2.6-UD-Q2_K_XL-00001-of-0008.gguf \
--mmproj unsloth/Kimi-K2.6-GGUF/mmproj-F16.gguf \
--temp 1.0 \
--top-p 0.95






