| ▲ | jwillp 2 days ago | |
I've been impressed by ollama running locally for my work, involving grouping short text snippets by semantic meaning, using embeddings, as well as summarization tasks. Depending on your needs, a local GPU can sometimes beat the cloud. (I get no failures and consistent response times with no extra bill.) Obviously YMMV, and not ideal for scaling up unless you love hardware. | ||
| ▲ | duckmysick 2 days ago | parent [-] | |
Which models have you been using? | ||