Raw LLM Responses
Inspect the exact model output for any coded comment.
Look up by comment ID
Random samples — click to inspect
G
Ironic his AI generated image depicted him as some dork while the pencil guy loo…
ytc_Ugw2iPMX_…
G
As an artist (a concept artist), I really feel seen by people like you and I rea…
ytc_Ugy6d5eaw…
G
This might be a bit rich coming from someone who can’t art yet, but I’ve heard a…
ytc_Ugz4e-eMC…
G
I know this creator. I thought everyone knew his work is ai? I remember one day …
ytc_UgxDKgGHH…
G
Tesla autopilot is on level 2 right now and it clearly says you have to pay atte…
ytc_Ugwn1L76h…
G
This has a few flaws not impossible but rare. You’re using the same complaint we…
ytr_UgwK3vSTI…
G
First of all, these data anatators are heros. But they need to be paid a LOT, an…
ytc_UgwMCxEbk…
G
the nearest future artificial intelligence is a grinder machine for digital iden…
ytc_Ugw_l3mCP…
Comment
I've gone over this very topic several times in the last few years and have come to the same conclusion each time. The only ethical way of handling accidents with self-driving cars is for each of them to prioritize the lives of people *outside* of the car. That means they will always choose to put passengers in danger before putting outsiders in danger. This way, it's the passenger's conscious decision to trust their life to a car the moment they enter it.
youtube
AI Harm Incident
2015-12-08T19:2…
♥ 323
Coding Result
| Dimension | Value |
|---|---|
| Responsibility | ai_itself |
| Reasoning | deontological |
| Policy | none |
| Emotion | approval |
| Coded at | 2026-04-27T06:24:59.937377 |
Raw LLM Response
[
{"id":"ytc_UghSiRcVXA-3FHgCoAEC","responsibility":"none","reasoning":"consequentialist","policy":"none","emotion":"indifference"},
{"id":"ytc_Ugg36gd_wQOCXHgCoAEC","responsibility":"ai_itself","reasoning":"deontological","policy":"none","emotion":"approval"},
{"id":"ytc_UggzSEiGsQNLKngCoAEC","responsibility":"none","reasoning":"consequentialist","policy":"none","emotion":"indifference"},
{"id":"ytc_UghidMHZsCybB3gCoAEC","responsibility":"ai_itself","reasoning":"consequentialist","policy":"none","emotion":"approval"},
{"id":"ytc_UgjzNTXzuzIxOngCoAEC","responsibility":"distributed","reasoning":"consequentialist","policy":"regulate","emotion":"resignation"},
{"id":"ytc_UghfmsovrnUJPXgCoAEC","responsibility":"ai_itself","reasoning":"deontological","policy":"none","emotion":"approval"},
{"id":"ytc_UgjQy7gtc5pA_XgCoAEC","responsibility":"ai_itself","reasoning":"consequentialist","policy":"none","emotion":"fear"},
{"id":"ytc_Ugio_pXgICTxCXgCoAEC","responsibility":"none","reasoning":"consequentialist","policy":"none","emotion":"approval"},
{"id":"ytc_UggKQCpjXBYZKXgCoAEC","responsibility":"developer","reasoning":"contractualist","policy":"liability","emotion":"approval"},
{"id":"ytc_UgggitcG_CbrUXgCoAEC","responsibility":"ai_itself","reasoning":"deontological","policy":"none","emotion":"approval"}
]