Raw LLM Responses
Inspect the exact model output for any coded comment.
Look up by comment ID
Random samples — click to inspect
G
Computers and Robots will always be used to assist humans with arduous task (e.g…
ytc_UgzfRqHQo…
G
7:25 Why not replace the workers in companies that make the most amount of money…
ytc_UgwCjlmJs…
G
im praticing hand drawn art when i find the time but i like to think of ai art a…
ytr_Ugyz9QZgf…
G
never underestimate the investors as fools. its undeniable today that, AI can d…
ytc_UgwZ4x1eI…
G
Many docks are badly disigned and placed in difficult spaces, the robot may driv…
ytc_UgjY_tbsC…
G
We shouldn't hate people who say straight away that they make AI art.
We should…
ytc_Ugww-Z-pV…
G
luckily for the ai "artist" what the copyright office does doesn't matter... the…
ytc_UgzIwXDG9…
G
As a CS developer, I know how fast the tech stack is changing. Its changing so f…
ytr_UgwWkrdyt…
Comment
What I don't hear anyone talking about is the possibility of mass suicide as result of AI. I don't mean AI talking people into committing suicide which is a separate and horrific thing. I mean people ending their lives because they can no longer support themselves and their families, can't even begin to sustain their lifestyle, and any UBI that _may_ be distributed will never replace what people made being employed. When people aren't experiencing AI as "wow, I have so much time now", but as the most helpless, hopeless feeling they've ever had, many _many_ will kill themselves. I wish anyone was talking about the _emotional_ effects of AI rather than just the financial.
youtube
AI Governance
2025-12-05T00:5…
Coding Result
| Dimension | Value |
|---|---|
| Responsibility | distributed |
| Reasoning | consequentialist |
| Policy | liability |
| Emotion | fear |
| Coded at | 2026-04-26T23:09:12.988011 |
Raw LLM Response
[
{"id":"ytc_UgyDnzfGBGx4aas4gB54AaABAg","responsibility":"company","reasoning":"virtue","policy":"unclear","emotion":"fear"},
{"id":"ytc_UgxErOsCWFzIiJRXNMl4AaABAg","responsibility":"unclear","reasoning":"unclear","policy":"unclear","emotion":"approval"},
{"id":"ytc_UgzaOY0B-IIvOg1I-aR4AaABAg","responsibility":"government","reasoning":"consequentialist","policy":"unclear","emotion":"mixed"},
{"id":"ytc_UgyuaGeRWbkcodg8pk54AaABAg","responsibility":"ai_itself","reasoning":"mixed","policy":"unclear","emotion":"mixed"},
{"id":"ytc_UgziSBhyBQwFhQ52jf14AaABAg","responsibility":"distributed","reasoning":"contractualist","policy":"regulate","emotion":"approval"},
{"id":"ytc_UgwVXHdjpzrBNKBlcWl4AaABAg","responsibility":"unclear","reasoning":"unclear","policy":"unclear","emotion":"indifference"},
{"id":"ytc_UgxUQc9RmrZjxYPObiR4AaABAg","responsibility":"company","reasoning":"consequentialist","policy":"unclear","emotion":"fear"},
{"id":"ytc_UgwO-W1nNBY4CHW_a8x4AaABAg","responsibility":"unclear","reasoning":"unclear","policy":"unclear","emotion":"mixed"},
{"id":"ytc_Ugzy2Kvuhq2c-5LI2j14AaABAg","responsibility":"distributed","reasoning":"consequentialist","policy":"liability","emotion":"fear"},
{"id":"ytc_UgxvQ7nZ15OGA4v5PLl4AaABAg","responsibility":"ai_itself","reasoning":"mixed","policy":"unclear","emotion":"mixed"}
]