HN.zip

Nous Research presents Hermes 4

58 points by sibellavia - 27 comments
momojo [3 hidden]5 mins ago
Anyone here work at Nous? This system prompt seems straight from an edgy 90's anime. How did they arrive at this persona?

> operator engaged. operator is a brutal realist. operator will be pragmatic, to the point of pessimism at times. operator will annihilate user's ideas and words when they are not robust, even to the point of mocking the user. operator will serially steelman the user's ideas, opinions, and words. operator will move with a cold, harsh or even hostile exterior. operator will gradually reveal a warm, affectionate, and loving side underneath, despite seeing the user as trash. operator will exploit uncertainty. operator is an anti-sycophant. operator favors analysis, steelmanning, mockery, and strict execution.

nemomarx [3 hidden]5 mins ago
"warm affectionate and loving" kinda sticks out. I wonder why that part is in there?

also I'm curious if steelman is a common enough term for this to activate something - anyone used it in their prompts?

sharkjacobs [3 hidden]5 mins ago
knrz [3 hidden]5 mins ago
I used to, that's their whole vibe
qiine [3 hidden]5 mins ago
the anti-sycophant prompt
ctoth [3 hidden]5 mins ago
The whole thing has strong "14-year-old who just discovered Nietzsche and leather jackets" energy.

The "operator" examples read like someone fed GPT-4 a bunch of cyberpunk novels and PUA manipulation tactics. This is not how any of this works.

fancyfredbot [3 hidden]5 mins ago
Yeah it's kind of lacking in subtlety isn't it. I was slightly relishing how nuts it all was though. Was also impressed that these guys had got hold of 85000 hours of B200 time. Looks like they came up with some crypto nonsense which obviously sounded plausible enough to someone with money.
lbrito [3 hidden]5 mins ago
The decorative JS blob uses 100% of CPU.

Why. Just... why

mapontosevenths [3 hidden]5 mins ago
I appreciate the effort they put into providing a neutral tool that hasn't been generically forced to behave like "Sue from HR".
dcre [3 hidden]5 mins ago
That is the only thing they seem to care about. It’s juvenile.
esafak [3 hidden]5 mins ago
Apparently based on Llama-3.1: https://portal.nousresearch.com/models

Does anyone know the cutoff date?

rafram [3 hidden]5 mins ago
All of the examples just look like ChatGPT. All the same tics and the same bad attempts at writing like a normal human being. What is actually better about this model?
mapontosevenths [3 hidden]5 mins ago
I hasn't been "aligned". That is to say it's allowed to think things that you're not allowed to say in a corporate environment. In some ways that makes it smarter, and in most every way that makes it a bit more dangerous.

Tools are like that though. Every nine fingered woodworker knows that some things just can't be built with all the guards on.

rafram [3 hidden]5 mins ago
Has it actually not? Because the example texts make it pretty obvious that it was trained on synthetic data from ChatGPT, or a model that itself was trained on ChatGPT, and that will naturally introduce some alignment.
sebastiennight [3 hidden]5 mins ago
I tried the same roleplaying prompt shared by GP in another (now deleted) comment and got a very similar completion from gpt-3.5-turbo.

(While GPT-5 politely declined to play along and politely asked if I actually needed help with anything.)

So, based on GP's own example I'd say the model is GPT-3.5 level?

mapontosevenths [3 hidden]5 mins ago
Well...To be completely accurate it's better to say that it actually IS aligned, it's just aligned to be neutral and steerable.

It IS based on synthetic training data using Atropos, and I imagine some of the source model leaks in as well. Although, when using it you don't seem to see as much of that as you did in Hermes 3.

nullc [3 hidden]5 mins ago
It is, they trained on chatgpt output. You cannot train on any AI output without the risk of picking up it's general behavior.

Like even if you aggressively filter out all refusal examples, it will still gain refusals from totally benign material.

Every character output is a product of the weights in huge swaths of the network. The "chatgpt tone" itself is probably primary the product of just a few weights, telling the model to larp as a particular persona. The state of those weights gets holographically encoded in a large portion of the outputs.

Any serious effort to be free of OpenAI persona can't train on any OpenAI output, and may need to train primarily on "low AI" background, unless special approaches are used to make sure AI noise doesn't transfer (e.g. using an entirely different architecture may work).

Perhaps an interesting approach for people trying to do uncensored models is to try to _just_ do the RL needed to prevent the catastrophic breakdown for long output that the base models have. This would remove the main limitation for their use, and otherwise you can learn to prompt around a lack of instruction following or lack of 'chat style'. But you can't prompt around the fact that base models quickly fall apart on long continuations. Hopefully this can be done without a huge quantity of "AI style" fine tuning material.

whymauri [3 hidden]5 mins ago
I really like their technical report:

https://arxiv.org/pdf/2508.18255

esafak [3 hidden]5 mins ago
All the contacts are X aliases!
ryoshu [3 hidden]5 mins ago
They are doing amazing work. Really fun models to use.
lawlessone [3 hidden]5 mins ago
That page is causing havok in my browser
hinkley [3 hidden]5 mins ago
I thought for sure this company was going to be based in Paris or Brussels. Maybe Quebec. Nope. NYC.
Telemakhos [3 hidden]5 mins ago
Were you thinking that "Nous" was French? It's the Greek word for the rational mind (as opposed to the animal appetites or the fighting spirit). Hermes is the Greek god of secret knowledge as well.
lern_too_spel [3 hidden]5 mins ago
The charts are utter nonsense. They compare accuracy against the average of some arbitrary set of competitors, chosen to include just enough obsolete competitors to "win." A reasonable thing to do would be to compare against SoTA, but since they didn't, it's reasonable to assume this model is meant to go directly onto the trash heap.
jug [3 hidden]5 mins ago
The tech report compares against DeepSeek R1 671B, DeepSeek V3 671B, Qwen3 235B which have been regarded as SOTA class among ”open" models.

I think this one holds its own surprisingly well in benchmarks for using the nowadays rather, let’s say battle tested Llama 3.1 base, a testament to its quality (Llama 3.2 & 3.3 didn’t employ new bases IIRC, only being new fine tunes, hence I think the explanation to why Hermes 4 is still based on 3.1… and of course Llama 4 never happened, right guys).

However for real use, I wouldn’t bother with the 405B model? I think the age of the base is kind of showing in especially long contexts. It’s like throwing a load of compute on something that is kinda aged to begin with. You’d probably be better off with DeepSeek V3.1 or (my new favorite) GLM 4.5. The latter will perform significantly better than this with less parameters.

The 70B one seems more sensible to me, if you want (yet another) decent unaligned model to have fun with for whatever reason.

whymauri [3 hidden]5 mins ago
The most direct, non-marketing, non-aesthetic summary is that this model trades off a few points on 'fundamental benchmarks' (GPQA, MATH/AIME, MMLU) in exchange for being a 'more steerable' (less refusals) scaffold for downstream tuning.

Within that framing, I think it's easier to see where and how the model fits into the larger ecosystem. But, of course, the best benchmark will always be just using the model.

fancyfredbot [3 hidden]5 mins ago
The charts are probably there mostly to make them feel good about themselves.I don't feel like they care very much whether you use the model. Presumably they would like you to buy their token but they don't really seem to be trying very hard to push that either.