We are happy to announce that ggml.ai (the founding team of llama.cpp) are joining Hugging Face in order to keep future AI truly open.
Georgi and team are joining HF with the goal of scaling and supporting the ggml/llama.cpp community as Local AI continues to make exponential progress in the coming years.
ggml and llama.cpp libraries and related open-source projectsSince its foundation in 2023, the core mission of ggml.ai has continuously been to support the development and the adoption of the ggml machine learning library. Over the past 3 years, the small team behind the company has been doing its best to grow the open-source developer community and to help establish ggml as the definitive standard for efficient local AI inference. This was achieved through strong collaboration with individual contributors, as well as with partnerships with model providers and independent hardware vendors. As a result, today llama.cpp has become the fundamental building block in countless projects and products, enabling private and easily-accessible AI on consumer hardware.
Throughout this development, Hugging Face stood out as the strongest and most supportive partner of this initiative. During the course of the last couple of years, HF engineers (notably @ngxson and @allozaur) have:
ggml and llama.cppllama.cppllama.cpp into the Hugging Face Inference Endpointsllama.cppggml projects with general maintenance, PR reviews and moreThe teamwork between our teams has always been smooth and efficient. Both sides, as well as the community, have benefited from these joint efforts. It only makes sense to formalize this collaboration and make it stronger in the future.
ggml/llama.cpp, the open source project and the community?Not much – Georgi and team will continue to dedicate 100% of their time maintaining ggml/llama.cpp. The community will continue to operate fully autonomously and make technical and architectural decisions as usual. Hugging Face is providing the project with long-term sustainable resources, improving the chances of the project to grow and thrive. The project will continue to be 100% open-source and community driven as it is now. Expect your favorite quants to be supported even faster once a model is released.
Going forward, our joint efforts will be geared towards the following objectives:
Towards seamless “single-click” integration with the transformers library
Better packaging and user experience of ggml-based software
Our shared goal is to provide the building blocks to make open-source superintelligence accessible to the world over the coming years. We will achieve this together with the growing Local AI community, as we continue to build the ultimate inference stack that runs as efficiently as possible on our devices.