My weekend side project: MiniLLM, a minimal system for running modern LLMs on consumer GPUs β¨ π¦ Supports multiple LLMs (LLAMA, BLOOM, OPT) βοΈ Supports NVIDIA GPUs, not just Apple Silicon π§ββοΈ Tiny, easy-to-use codebase in Python (<500 LOC) https:// github.com/kuleshov/minillm β¦
My weekend side project: MiniLLM, a minimal system for running modern LLMs on consumer GPUs β¨ π¦ Supports multiple LLMs (LLAMA, BLOOM, OPT) βοΈ Supports NVIDIA GPUs, not just Apple Silicon π§ββοΈ Tiny, easy-to-use codebase in Python (<500 LOC) https:// github.com/kuleshov/minillm β¦