- Notifications
You must be signed in to change notification settings - Fork169
Chat with AI large language models running natively in your browser. Enjoy private, server-free, seamless AI conversations.
License
mlc-ai/web-llm-chat
Folders and files
| Name | Name | Last commit message | Last commit date | |
|---|---|---|---|---|
Repository files navigation
WebLLM Chat is a private AI chat interface that combinesWebLLM with a user-friendly design, leveraging WebGPU to run large language models (LLMs) natively in your browser. Enjoy an unprecedented, private, and accessible AI conversation experience.
- Browser-Native AI: Experience cutting-edge language models running natively within your web browser with WebGPU acceleration, eliminating the need for server-side processing or cloud dependencies.
- Ganranteed Privacy: With the AI model running locally on your hardware and all data processing happening within your browser, your data and conversations never leave your computer, ensuring your privacy.
- Offline Accessibility: Run entirely offline after the initial setup and download, allowing you to engage with AI-powered conversations without an active internet connection.
- Vision Model Support: Chat with AI by uploading and sending images, making it easy to get insights and answers based on visual content.
- User-Friendly Interface: Enjoy the intuitive and feature-rich user interface, complete with markdown support, dark mode, and a responsive design optimized for various screen sizes.
- Custom Models: Connect to any custom language model on you local environment throughMLC-LLM. For detail, check theUse Custom Models section.
- Open Source and Customizable: Build and customize your own AI-powered applications with our open-source framework.
WebLLM Chat is a pioneering initiative that combines the robust backend capabilities of WebLLM with the user-friendly interface of NextChat. As a part of the broader MLC.ai family, this project contributes to our mission of democratizing AI technology by making powerful tools accessible directly to end-users. By integrating with NextChat, WebLLM Chat not only enhances the chatting experience but also broadens the scope for deployment of self-hosted and customizable language models.
WebLLM Chat natively supports WebLLM build-in models. You can find the full listhere.
WebLLM Chat supports custom language models throughMLC-LLM. Follow the following steps to use custom models on your local environment:
(Optional) Compile the model into MLC format by followingthe instructions.
Host REST API through MLC-LLM by followingthe instructions.
Go toWebLLM Chat, select "Settings" in the side bar, then select "MLC-LLM REST API (Advanced)" as "Model Type" and type the REST API endpoint URL from step 2.
# 1. install nodejs and yarn first# 2. config local env vars in `.env.local`# 3. runyarn installyarn dev
You can build the application as a Next.js build usingyarn build or as a static site usingyarn export. For more information, checkNext.js documentation;
docker build -t webllm_chat.docker run -d -p 3000:3000 webllm_chatYou can start service behind a proxy:
docker build -t webllm_chat.docker run -d -p 3000:3000 \ -e PROXY_URL=http://localhost:7890 \ webllm_chatIf your proxy needs password, use:
-e PROXY_URL="http://127.0.0.1:7890 user pass"WebLLM Chat thrives on community involvement. We are committed to fostering an inclusive and innovative community where developers and AI enthusiasts can collaborate, contribute, and push the boundaries of what's possible in AI technology. Join us on Discord to connect with fellow developers and contribute to the project.
WebLLM Chat is a companion project ofWebLLM and it is built upon the remarkable work ofNextChat. We extend our sincere gratitude to the developers and contributors of these projects for their invaluable efforts in advancing the field of browser-based AI and creating user-friendly chat interfaces.
Further more, this project is only possible thanks to the shoulders of open-source ecosystems that we stand on. We want to thank the Apache TVM community and developers of the TVM Unity effort. The open-source ML community members made these models publicly available. PyTorch and Hugging Face communities make these models accessible. We would like to thank the teams behind Vicuna, SentencePiece, LLaMA, Alpaca. We also would like to thank the WebAssembly, Emscripten, and WebGPU communities. Finally, thanks to Dawn and WebGPU developers.
About
Chat with AI large language models running natively in your browser. Enjoy private, server-free, seamless AI conversations.
Topics
Resources
License
Uh oh!
There was an error while loading.Please reload this page.
Stars
Watchers
Forks
Releases
Packages0
Uh oh!
There was an error while loading.Please reload this page.