nderstand2grow

joined 10 months ago
[–] nderstand2grow@alien.top 1 points 9 months ago

your comment is so insightful, thank you. if there are resources I can read/watch to learn about this stuff, I'd be happy if you could share them.

 

I'm confused by all these prefixes that appear in the finetunes of base models. Is there a glossary of all these words and similar ones?

 

(title)

[–] nderstand2grow@alien.top 1 points 10 months ago

I have ollama on my Mac (not Docker) and installed the ollama web UI. It works fine but their instruction on running ollama in a LAN network doesn't work for me. The flags they mention to add the CLI command throw an error (esp. the * part).

[–] nderstand2grow@alien.top 1 points 10 months ago

You could use M2 Ultra instead ($6500) vs. 2x$15,000+rest

[–] nderstand2grow@alien.top 1 points 10 months ago (1 children)

I think this is only available on llama.cpp. I've been using it for a while for simple structured outputs and am extremely happy with the results. With OpenAI's function calling, I always had to write validators -- first to make sure the output is indeed a JSON, and then another validator to make sure the JSON complies with my JSON schema. grammar makes all of that redundant because it is 100% guaranteed to generate the desired output (including JSON).