Preferences

As far as I know there's not really LLMs good enough that can run locally. Maybe with the R1 improvements and future derivative work that'll change.

We use about 4-6 calls per improvement and use a mix of Anthropic and OpenAI. Interestingly we really couldn't get sufficiently good performance from just one model. It's interesting how they can be good or bad at different tasks where one task doesn't seem materially harder than the other.


Thanks for the feedback, very interesting.

In the meantime I've seen this model https://huggingface.co/FPHam/Karen_TheEditor_V2_STRICT_Mistr... , I would be interested how it compares. (though it seems specifically fine-tuned for american english)

This item has no comments currently.

Keyboard Shortcuts

Story Lists

j
Next story
k
Previous story
Shift+j
Last story
Shift+k
First story
o Enter
Go to story URL
c
Go to comments
u
Go to author

Navigation

Shift+t
Go to top stories
Shift+n
Go to new stories
Shift+b
Go to best stories
Shift+a
Go to Ask HN
Shift+s
Go to Show HN

Miscellaneous

?
Show this modal