Discussion
Loading...

Post

Log in
  • Sign up
  • About
  • Code of conduct
  • Privacy
  • Users
  • Instances
  • About Bonfire
Rost Glukhov
Rost Glukhov
@rosgluk@mastodon.social  ·  activity timestamp 3 days ago

Comparing LLMs performance on Ollama on 16GB VRAM GPU:
https://www.glukhov.org/post/2026/01/choosing-best-llm-for-ollama-on-16gb-vram-gpu/
#LLM #Ollama #NVidia #Hardware #SelfHosting #OpenSource #DeepLearning #AI

Rost Glukhov | Personal site and technical blog

Comparing LLMs performance on Ollama on 16GB VRAM GPU

Benchmark of 9 LLMs on RTX 4080 16GB with Ollama 0.15.2. Compare tokens/sec, VRAM usage, and CPU offloading for GPT-OSS, Qwen3, Mistral, and more.
  • Copy link
  • Flag this post
  • Block

Indieweb Studio

This is a relaxed, online social space for the indieweb community, brought to you by indieweb.social.

Please abide by our code of conduct and have a nice time!

Indieweb Studio: About · Code of conduct · Privacy · Users · Instances
Bonfire social · 1.0.2-alpha.7 no JS en
Automatic federation enabled
Log in Create account
  • Explore
  • About
  • Members
  • Code of Conduct