Charlie Ruan (@charlie_ruan)

2024-06-13 | โค๏ธ 389 | ๐Ÿ” 92


Excited to share WebLLM engine: a high-performance in-browser LLM inference engine!

WebLLM offers local GPU acceleration via @WebGPU, fully OpenAI-compatible API, and built-in web workers support to separate backend executions.

Check out the blog post: https://blog.mlc.ai/2024/06/13/webllm-a-high-performance-in-browser-llm-inference-engine https://x.com/charlie_ruan/status/1801329345669304582/video/1

๋ฏธ๋””์–ด

video


Tags

domain-ai-ml domain-xr domain-dev-tools domain-web-graphics