Remix.run Logo
alxlaz 5 hours ago

I use their API for several models, both for personal and professional use. I think their approach (smaller, specialised models that are well-adapted for specific tasks) is a very good fit for how I work. And even the more general-purpose ones, like the chat model, just... refreshingly good in a lot of ways. My "ruthless review" prompt, which I use for, well, ruthless, guided reviews of early technical drafts, has good technical results for early reviews and holy crap is it ruthless and does it know how to swear. By the time Claude or ChatGPT are done being honest about how right I am to push back and gently circling back, Mistral's large model has sent me back to the drawing board twice.

Being in the EU does smooth a lot of things in terms of compliance, payment processing and whatnot, but I also like that their data retention and privacy policies are pretty clearly spelled out. I need to know something, there's a good chance it's explained outright somewhere and I don't need to read between the EULA lines and wonder what it means.

I do hit limits in terms of capabilities sometimes, and I'm sure other providers' services offer better results for some things. But the businesses ran on top of those more capable models feel too much like a scam at this point and I'd rather not depend on them for anything I actually need.

dbl000 5 hours ago | parent [-]

That ruthless review prompt seems interesting, would you be willing to share it? I've been trying to have Claude act as a reviewer for me and it feels like it never will disagree.

alxlaz 3 hours ago | parent [-]

It's very hard to untangle it from the rest of its context (the prompt is built dynamically, from a lot of parts, some project-specific, some specific to my preferences, some built from interaction history), so I can't really share it. In any case, I don't think it's some specific prompt engineering sorcery I'm doing, it's not like I've spent any real time refining it or experimenting with various magical incantations. It's probably just some model features making it more amenable to the kind of instructions that are relevant in these cases (directness, questioning trade-offs, thoroughness etc.). My chatbot swears equally graphical in review prompts and news summarizing prompts so I'm pretty sure I'm not tickling the machine just right :)

altmanaltman 3 hours ago | parent [-]

Can you share some of its output for reference?