Portable LLMs with llamafile
Portable LLMs with llamafile
Posted May 15, 2024 7:07 UTC (Wed) by rsidd (subscriber, #2582)In reply to: Portable LLMs with llamafile by snajpa
Parent article: Portable LLMs with llamafile
She is sending patches upstream. The whole point of open source is you can fork it for your own interests, which may not match upstream. But if upstream is interested, the fork can re-merge.
This project, as I understand it, is basically about (a) building llama.cpp with cosmocc (for improved portability: you can literally run the same output file on linux, windows, macos); (b) speeding up linear algebra (which she is feeding upstream where relevant) with the goal of (c) bundling LLM weights and llama.cpp into single, portable, fast executables that can use GPU when possible, or run with adequate speed on CPU, without user intervention or configuration.
