Hacker News
new
|
past
|
comments
|
ask
|
show
|
jobs
|
submit
login
jwr
6 months ago
|
parent
|
context
|
favorite
| on:
Qwen3-Next
I understand that, but whether it's usable depends on whether ollama can load parts of it into memory on my Mac, and how quickly.
bigyabai
6 months ago
[–]
I
really
do not suggest ollama. It is slow, missing tons of llama.cpp features and doesn't expose many settings to the user. Koboldcpp is a much better inference provider and even has an ollama-compatible API endpoint.
Guidelines
|
FAQ
|
Lists
|
API
|
Security
|
Legal
|
Apply to YC
|
Contact
Search: