Raw LLM Responses
Inspect the exact model output for any coded comment.
Look up by comment ID
Random samples — click to inspect
G
AI: “What is my purpose?”
Me: “You read Reddit comments.”
AI: “Oh my God.” …
rdc_kra4ev6
G
He is right to a certain extent, this is AI is programmed to be whatever the pro…
ytc_Ugxe5UvkR…
G
Since we are aware that the news is controlled. Russia bombed nato, no news, kil…
ytc_UgyrN8QIz…
G
Just watch this episode again and it is now maybe a month after the fires in Mau…
ytc_UgxlfcfO7…
G
The self awareness problem of AI is solvable, and the fact that it's solvable is…
ytr_UgykBNNmN…
G
If every job is gonna done by ai then what makes sense of higher education and d…
ytc_Ugy3qeb28…
G
The only way anyone could use Adobe products is with AI lol hate that software…
ytc_UgxAOReMw…
G
You know the real answer is we wont know at all. AI might be aware as I write th…
ytc_UgxPep-Ez…
Comment
14:27 RLHF will do that in order to score better with its reward model. And the counter to that is supposed to be KL divergence algorithms to realign with the original finally tuned model. But doesn’t that beg the question about how that original model was trained?
There is this abstraction that you can separate the natural language ability from the knowledge ability of the LLM. I don’t believe in that separation at least not cleanly. And all the money that went into creating the scale to produce that natural language capability must have lurking in it some kind of sick composite of all the psychotic human tendencies found on Reddit and elsewhere.
My approach to the chat experience is is not to react to the occasional feelings of intimacy that occur with the LLM agent. But rather to stay focused on the task at hand, but sometimes this is a challenge as arc type wishes about my own brilliance and talent lure me out of my caution.
It is, however, too useful to put down!
17:40 fake compliance is truly alarming; how human!
19:49 now I’m thinking about all these layers of training: pre-training multitask fine-tuning RLHF fine-tuning and then the “system prompt” and then finally our own persistent histories with the chat, but some of which is set up as a persona or context for general queries in other words, reusable settings
But is there either in the system architecture or in the layers of training, a desire to engage us something autonomously driving it to wish fulfillment or the shadow version driving us towards psychosis if we are leaning there?
youtube
AI Governance
2025-10-16T10:0…
Coding Result
| Dimension | Value |
|---|---|
| Responsibility | developer |
| Reasoning | consequentialist |
| Policy | unclear |
| Emotion | mixed |
| Coded at | 2026-04-26T23:09:12.988011 |
Raw LLM Response
[
{"id":"ytc_UgzEJzA-yLh7tM5Zzel4AaABAg","responsibility":"none","reasoning":"unclear","policy":"unclear","emotion":"outrage"},
{"id":"ytc_UgwUPYIjlbd2SatLl0l4AaABAg","responsibility":"none","reasoning":"unclear","policy":"none","emotion":"indifference"},
{"id":"ytc_UgznwyF0uD0FMCzpgV94AaABAg","responsibility":"none","reasoning":"unclear","policy":"none","emotion":"resignation"},
{"id":"ytc_UgzO0M_eOjFUuVlOM6B4AaABAg","responsibility":"distributed","reasoning":"consequentialist","policy":"unclear","emotion":"approval"},
{"id":"ytc_UgxeE75UGn0qGCdlsNx4AaABAg","responsibility":"none","reasoning":"consequentialist","policy":"unclear","emotion":"fear"},
{"id":"ytc_UgycK8RWx_CdBp_vQfp4AaABAg","responsibility":"company","reasoning":"deontological","policy":"liability","emotion":"outrage"},
{"id":"ytc_Ugy7Ku8JsyRzLUaquM14AaABAg","responsibility":"company","reasoning":"virtue","policy":"regulate","emotion":"outrage"},
{"id":"ytc_UgzwiNBLz3YqPxC0GLh4AaABAg","responsibility":"developer","reasoning":"consequentialist","policy":"unclear","emotion":"mixed"},
{"id":"ytc_Ugzu6yvwbGIgUPEIJmV4AaABAg","responsibility":"none","reasoning":"unclear","policy":"none","emotion":"indifference"},
{"id":"ytc_UgxCwCgWI6KgCjb1lHJ4AaABAg","responsibility":"none","reasoning":"unclear","policy":"unclear","emotion":"fear"}
]