r/LocalLLaMA - Reddit
r/LocalLLaMA - Reddit
reddit.com
Related
Highlights
eneral-purpose models
1.1B:
TinyDolphin 2.8 1.1B
. Takes about ~700MB RAM and tested on my Pi 4 with 2 gigs of RAM. Hallucinates a lot, but works for basic conversation.
2.7B:
Dolphin 2.6 Phi-2
. Takes over ~2GB RAM and tested on my 3GB 32-bit phone via llama.cpp on Termux.
7B:
Nous Hermes Mistral 7B DPO
. Takes about ~4-5GB RAM depending on contex
...
See more
r/LocalLLaMA - Reddit
2
2
Ollama
ollama.com
https://github.com/huggingface/chat-ui
- Amazing clean UI with very good web search, my go to currently. (they added the ability to do it all locally very recently!)
https://github.com/oobabooga/text-generation-webui
- Best overall, supports any model format and has many extensions
https://github.com/ParisNeo/lollms-webui/
- Has PDF, stable diffusion...
See more
r/LocalLLaMA - Reddit
3
3
What We Learned From a Year of Building With LLMs
Bryan Bischof
oreilly.com
Unlock unlimited Related cards