GitHub - ggml-org llama. cpp: LLM inference in C C++ The main goal of llama cpp is to enable LLM inference with minimal setup and state-of-the-art performance on a wide range of hardware - locally and in the cloud Plain C C++ implementation without any dependencies Apple silicon is a first-class citizen - optimized via ARM NEON, Accelerate and Metal frameworks AVX, AVX2, AVX512 and AMX support for x86 architectures RVV, ZVFH, ZFH and ZICBOP
Model Context Protocol · GitHub An open protocol that enables seamless integration between LLM applications and external data sources and tools - Model Context Protocol