llm-chat
Here are 27 public repositories matching this topic...
Start a chat room between all, or some, of your models running on Ollama. All in a single Bash shell script.
-
Updated
Aug 30, 2025 - Shell
Chat locally using leading open models built by the community, optimized and accelerated by NVIDIA's enterprise-ready inference runtime
-
Updated
Jul 25, 2024 - Python
Mem0Chat is an application showcasing an LLM-powered chat with a model-agnostic persistent memory layer powered by mem0. It enables context-aware conversations, using mem0's memory management to retain and utilize key memories across sessions/chats for a personalized, efficient user experience.
-
Updated
Jul 3, 2025 - JavaScript
Scalable implementation of Semantic search and LLM powered chat bot for online store
-
Updated
Jul 23, 2024 - Python
基于 React + TypeScript 的智能 AI 对话应用,支持流式响应、Markdown 渲染和代码高亮
-
Updated
Mar 4, 2026 - TypeScript
Unreal Engine Human Machine Interface
-
Updated
Feb 15, 2026 - C++
Chat App with Ollama
-
Updated
Dec 17, 2025 - TypeScript
Single HTML Page offering split screen LLM Chat and Web App Prototyping
-
Updated
May 22, 2025 - JavaScript
A complete, menu-driven AI model interface for Windows that simplifies running local GGUF language models with llama.cpp. This tool automatically manages dependencies, provides multiple interaction modes, and prioritizes user privacy through fully offline operation.
-
Updated
Jan 30, 2026 - PowerShell
-
Updated
May 14, 2024 - JavaScript
-
Updated
Jul 2, 2025 - Go
A modern real-time streaming chat application with FastAPI backend and React TypeScript frontend
-
Updated
Aug 31, 2025 - Python
implement llm streaming with page reload support using vercel ai sdk
-
Updated
Oct 22, 2025 - TypeScript
JavaScript templating engine based on Jinja2 (forked to experiment with LLM Chat templates in Jinja2 syntax)
-
Updated
Aug 26, 2024 - JavaScript
An ai-powered journal that determines your mood.
-
Updated
Nov 12, 2024 - TypeScript
Chat LLM local: Interfaz CLI para modelos GGUF y Transformers con compatibilidad CUDA. Permite ejecutar Llama, Mistral, Gemma, Phi y Qwen localmente con detección automática de modelos, adaptación de mensajes del sistema, soporte RAG y más.
-
Updated
Nov 5, 2025 - Python
Improve this page
Add a description, image, and links to the llm-chat topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with the llm-chat topic, visit your repo's landing page and select "manage topics."