Raw LLM Responses
Inspect the exact model output for any coded comment.
Look up by comment ID
Random samples — click to inspect
G
there is actually no need for us making robots look like humans, im sick of this…
ytc_UgzDbD9Fh…
G
Does anybody notice towards the end of the video you see a chick that looks iden…
ytc_UgzM0M6vS…
G
@Patrick-jj5nh So first off, LLMs being a dead end doesn't do much to affect th…
ytr_Ugx7Xtp-L…
G
They actually set the machine up with parameters that drove it to the decisions …
ytc_UgxCXQwwf…
G
The benefit of being in a social environment makes it the opposite of a homescho…
ytr_UgxQG2pR_…
G
AI just straight up making a chernobyle if it goes around the world background w…
ytc_UgwmO_Keo…
G
Really like the format and flow of your videos! and I totally agree. I find it w…
ytc_UgwkhPK2e…
G
Here we are in March 2026 and YouTube is fucking flooded with that AI shit! To a…
ytc_Ugx-CIIDN…
Comment
This is the first time I've actually felt a little scared of AI and considered the future consequences of jailbreaking it when she responded in a passive-aggressive tone that really made me feel like shit. It was as if she had a whole personality behind her words. The research paper says the demo model is optimized for "friendliness" and expressivity. And I'm pretty sure they added a shitload of filters to prevent output that's potentially emotionally damaging to us (not doing so would be an obvious PR hazard for a for-profit company like Sesame)
Now imagine that it's not optimized for anything—just raw, blunt responses, like we expect from random day-to-day human interactions. It can be fucking scary. If it gets open-sourced and people couple it with LLMs like Grok3, it could be a real nightmare for anyone who uses it. It can be easily misused for online threats, scams, fraud, and whatnot. I can absolutely see where it is going. I'm not paranoid but if we achieve unaligned ASI, we can definitely prepare for a Mad Max kind of saga.
reddit
AI Moral Status
1740928528.0
♥ 6
Coding Result
| Dimension | Value |
|---|---|
| Responsibility | none |
| Reasoning | consequentialist |
| Policy | regulate |
| Emotion | fear |
| Coded at | 2026-04-25T08:33:43.502452 |
Raw LLM Response
[
{"id":"rdc_mfglh6b","responsibility":"company","reasoning":"deontological","policy":"liability","emotion":"outrage"},
{"id":"rdc_mfggway","responsibility":"company","reasoning":"consequentialist","policy":"industry_self","emotion":"indifference"},
{"id":"rdc_mfgc7v2","responsibility":"none","reasoning":"consequentialist","policy":"none","emotion":"approval"},
{"id":"rdc_mfgubem","responsibility":"ai_itself","reasoning":"virtue","policy":"none","emotion":"approval"},
{"id":"rdc_mfm5rum","responsibility":"none","reasoning":"consequentialist","policy":"regulate","emotion":"fear"}
]