| | Machine Learning Compiler (mlc.ai) |
| 1 point by selvan 5 months ago | past |
|
| | Microserving LLM Engines (mlc.ai) |
| 1 point by homarp 11 months ago | past | 1 comment |
|
| | LLM Microserving: a new RISC-style approach to design LLM serving API (mlc.ai) |
| 4 points by jinhongyii 11 months ago | past | 1 comment |
|
| | Making AMD GPUs competitive for LLM inference (2023) (mlc.ai) |
| 313 points by plasticchris on Dec 24, 2024 | past | 213 comments |
|
| | Optimizing and Characterizing High-Throughput Low-Latency LLM Inference (mlc.ai) |
| 1 point by djhu9 on Oct 11, 2024 | past |
|
| | High-Throughput Low-Latency LLM Serving with MLCEngine (mlc.ai) |
| 8 points by ruihangl on Oct 10, 2024 | past |
|
| | In-browser LLM inference engine with WebGPU and OpenAI API (mlc.ai) |
| 16 points by CharlieRuan on June 13, 2024 | past | 4 comments |
|
| | MLCEngine: Universal LLM Deployment to Both Cloud and Local Devices (mlc.ai) |
| 2 points by crowwork on June 8, 2024 | past |
|
| | Universal LLM Deployment Engine with ML Compilation (mlc.ai) |
| 17 points by ruihangl on June 7, 2024 | past | 7 comments |
|
| | MLC LLM: Universal Language Model Deployment Across Diverse Hardware and Apps (mlc.ai) |
| 1 point by georgehill on Dec 16, 2023 | past |
|
| | Scaling LLama2-70B with Multiple Nvidia/AMD GPU (mlc.ai) |
| 13 points by junrushao1994 on Oct 20, 2023 | past | 6 comments |
|
| | WebLLM: Llama2 in the Browser (mlc.ai) |
| 192 points by meiraleal on Aug 29, 2023 | past | 31 comments |
|
| | GPU-Accelerated LLM on an Orange Pi (mlc.ai) |
| 214 points by tosh on Aug 15, 2023 | past | 80 comments |
|
| | Making AMD GPUs competitive for LLM inference (mlc.ai) |
| 354 points by djoldman on Aug 9, 2023 | past | 132 comments |
|
| | Run Llama2-70B in Web Browser with WebGPU Acceleration (mlc.ai) |
| 9 points by ruihangl on July 24, 2023 | past | 6 comments |
|
| | Bringing Open Large Language Models to Consumer Devices (mlc.ai) |
| 31 points by hardmaru on May 23, 2023 | past |
|
| | Running RedPajama and other open LLMs on phones, browsers and AMD/NV/Intel GPUs (mlc.ai) |
| 11 points by junrushao1994 on May 23, 2023 | past |
|
| | Bringing Open Large Language Models to Consumer Devices (mlc.ai) |
| 11 points by shantanu_sharma on May 22, 2023 | past |
|
| | Browser-based Stable Diffusion using WebGPU (mlc.ai) |
| 3 points by Eduard on May 6, 2023 | past |
|
| | Bringing Hardware Accelerated Language Models to Consumer Devices (mlc.ai) |
| 1 point by crowwork on May 1, 2023 | past |
|
| | MLC: Bringing Hardware Accelerated Language Models to Consumer Devices (mlc.ai) |
| 8 points by junrushao1994 on May 1, 2023 | past |
|
| | What Is ML Compilation (mlc.ai) |
| 88 points by tosh on April 30, 2023 | past | 5 comments |
|
| | Vicuna on iPhone (mlc.ai) |
| 90 points by tosh on April 30, 2023 | past | 15 comments |
|
| | MLC LLM (mlc.ai) |
| 2 points by tosh on April 30, 2023 | past |
|
| | MLC LLM – Large Language Models on iPhone GPU and Many More GPU Platforms (mlc.ai) |
| 2 points by crowwork on April 29, 2023 | past |
|
| | Web LLM (mlc.ai) |
| 89 points by synergy20 on April 25, 2023 | past | 15 comments |
|
| | Web LLM: Bringing large-language models and chat to web browsers (mlc.ai) |
| 3 points by yarapavan on April 17, 2023 | past |
|
| | Web LLM – Vicuna (like LLaMA) locally inside the browser (mlc.ai) |
| 5 points by the_tli on April 15, 2023 | past |
|
| | WebLLM (mlc.ai) |
| 6 points by thm on April 15, 2023 | past |
|
| | Running Stable Diffusion fully in browser with WebGPU (mlc.ai) |
| 42 points by crowwork on March 8, 2023 | past | 19 comments |
|
|
| More |