llamacpp
Here are 24 public repositories matching this topic...
Drop-in, local AI alternative to the OpenAI stack. Multi-engine (llama.cpp, TensorRT-LLM, ONNX). Powers 👋 Jan
-
Updated
Jul 18, 2024 - C++
WebAssembly binding for llama.cpp - Enabling in-browser LLM inference
-
Updated
Jul 10, 2024 - C++
Inferflow is an efficient and highly configurable inference engine for large language models (LLMs).
-
Updated
Mar 15, 2024 - C++
Inference Vision Transformer (ViT) in plain C/C++ with ggml
-
Updated
Apr 11, 2024 - C++
LLM in Godot
-
Updated
Jun 23, 2024 - C++
Getting an LLM to work with Godot.
-
Updated
Oct 11, 2023 - C++
Lightweight chat terminal-interface for llama.cpp server compilable for windows and linux.
-
Updated
Mar 1, 2024 - C++
Llama causal LM fully recreated in LibTorch. Designed to be used in Unreal Engine 5
-
Updated
Jan 5, 2024 - C++
LLM InferenceNet is a C++ project designed to facilitate fast and efficient inference from Large Language Models (LLMs) using a client-server architecture. It enables optimized interactions with pre-trained language models, making deployment on edge devices easier.
-
Updated
Jul 28, 2023 - C++
PipeInfer: Accelerating LLM Inference using Asynchronous Pipelined Speculation
-
Updated
Jul 5, 2024 - C++
Improve this page
Add a description, image, and links to the llamacpp topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with the llamacpp topic, visit your repo's landing page and select "manage topics."