llama.cpp llama.cpp

llama.cpp a été créé par ggml-org il y a 2 an(s), et mis à jour pour la dernière fois il y a 24 jour(s).

LLM inference in C/C++
C++ 91.30MB MIT Github
Stars
74.7k
Fork
10.8k
Watch
581
Open Issues
718

LocalAI

30k Go MIT

:robot: The free, Open Source alternative to OpenAI, Claude and others. Self-hosted and local-first. Drop-in replacement for OpenAI, running on consumer-grade hardware. No GPU required. Runs gguf, transformers, diffusers and many more models architectures. Features: Generate Text, Audio, Video, Images, Voice Cloning, Distributed, P2P inference

1 il y a 1 an(s) il y a 3 jour(s)

LLaMA-Factory

43.2k Python Apache-2.0

Unified Efficient Fine-Tuning of 100+ LLMs & VLMs (ACL 2024)

1 il y a 1 an(s) il y a 8 jour(s)

vllm

41.5k Python Apache-2.0

A high-throughput and memory-efficient inference and serving engine for LLMs

1 il y a 2 an(s) il y a 47 minute(s)

Langchain-Chatchat

34.1k TypeScript Apache-2.0

Langchain-Chatchat(原Langchain-ChatGLM)基于 Langchain 与 ChatGLM, Qwen 与 Llama 等语言模型的 RAG 与 Agent 应用 | Langchain-Chatchat (formerly langchain-ChatGLM), local knowledge based LLM (like ChatGLM, Qwen and Llama) RAG and Agent app with langchain

1 il y a 1 an(s) il y a 28 minute(s)

llama.cpp

74.7k C++ MIT

LLM inference in C/C++

1 il y a 2 an(s) il y a 24 jour(s)

ollama

132k Go MIT

Get up and running with Llama 3.3, DeepSeek-R1, Phi-4, Gemma 3, and other large language models.

1 il y a 1 an(s) il y a 29 minute(s)