|
@@ -174,6 +174,11 @@ Successfully installed vllm-0.8.5.dev3+g7cbfc1094.d20250414
|
|
|
(ilab-25) $ *rm -rf ./vllm/*
|
|
|
----
|
|
|
|
|
|
+References:
|
|
|
+
|
|
|
+* https://github.com/triton-lang/triton[Triton Development Repository]
|
|
|
+* https://docs.vllm.ai/en/stable/getting_started/installation/cpu.html?device=apple[Building vLLM for Apple Silicon]
|
|
|
+
|
|
|
=== Converting Models to GGUF ===
|
|
|
|
|
|
You can use https://github.com/ggerganov/llama.cpp.git[`llama.cpp`] to convert models from HF, GGML, and LORA model formats to GGUF, which InstructLab can serve even on a Mac.
|