MTP support merged into llama.cpp
Multi-Token Prediction support merged into llama.cpp master branch, adding a new inference technique to the popular open-source LLM inference framework.
Excerpt
PR [22673](https://github.com/ggml-org/llama.cpp/pull/22673) has been merged into master! 🎉
Read at source: https://www.reddit.com/r/LocalLLaMA/comments/1tes1wx/mtp_support_merged_into_llamacpp/