Raw LLM Responses

Inspect the exact model output for any coded comment.

Comment
"ust based on opinion, fails to construct his opinions ground up from first principles" Funny that, because all of the AI-safety results and discussion are the fruit of first-principles thinking he did in the early 2000s. Are you familiar with the inner/outer alignment? Mesaoptimizers? Instrumental convergence? At what level will the argument make sense to you and not sound like opinion? I'm paraphrasing some of the safety arguments: - we have ample evidence for the outcomes of adversarial agents where there is a dramatic asymmetry in their power - we have research to show that instrumental goals will develop, which will be aligned with the outer goals in the training regime but, due to distribution shift, will be non-aligned in the production regime - we have barely any mechanistic interpretability to make those inner goals obvious to us - in agentic systems, inner goals will be taken to extremes because they are intrinsically "desired" (were correlated with reward in training) - we expect instrumental convergent goals like self-defense, resource accumulation, and deception - we have near certainty that AI progress will continue at pace - we have near certainty that our safety mechanisms are not sufficiently in place to bear the live stress test of a superintelligence
youtube AI Governance 2024-12-27T03:1… ♥ 4
Coding Result
DimensionValue
Responsibilitynone
Reasoningconsequentialist
Policynone
Emotionmixed
Coded at2026-04-27T06:24:53.388235
Raw LLM Response
[ {"id":"ytr_UgxBSPLgIZgoxW75T_54AaABAg.AT5zqZXavDIAT7Flt59JVm","responsibility":"ai_itself","reasoning":"consequentialist","policy":"unclear","emotion":"mixed"}, {"id":"ytr_UgzqAFwsAv_KFYodnsZ4AaABAg.AT5zYObdugsAT6N8mi1y-k","responsibility":"none","reasoning":"consequentialist","policy":"none","emotion":"fear"}, {"id":"ytr_UgypFt-geNkrVXiCZKZ4AaABAg.ACqSJ-64r_IADGqQZSIyda","responsibility":"developer","reasoning":"deontological","policy":"none","emotion":"outrage"}, {"id":"ytr_UgypFt-geNkrVXiCZKZ4AaABAg.ACqSJ-64r_IADHiqhf0--s","responsibility":"developer","reasoning":"deontological","policy":"none","emotion":"mixed"}, {"id":"ytr_UgwmijS6ORinBNiU97l4AaABAg.ACPn9QBJA7kACXUpFJQjzw","responsibility":"none","reasoning":"mixed","policy":"none","emotion":"approval"}, {"id":"ytr_UgwmijS6ORinBNiU97l4AaABAg.ACPn9QBJA7kACXkGqvjRfO","responsibility":"none","reasoning":"consequentialist","policy":"none","emotion":"mixed"}, {"id":"ytr_UgwmijS6ORinBNiU97l4AaABAg.ACPn9QBJA7kACYsef1X_0Z","responsibility":"developer","reasoning":"consequentialist","policy":"ban","emotion":"fear"}, {"id":"ytr_Ugw5RY3Jgi5ZKd0rJTh4AaABAg.ACPiaRNcYV4AD6EVM4dvv8","responsibility":"developer","reasoning":"consequentialist","policy":"regulate","emotion":"mixed"}, {"id":"ytr_UgxPWTX-Ney8A1b-AYh4AaABAg.ACOTh2V31EVACQDPHy_sHM","responsibility":"none","reasoning":"mixed","policy":"none","emotion":"indifference"}, {"id":"ytr_UgxPWTX-Ney8A1b-AYh4AaABAg.ACOTh2V31EVACRLOiBBt8k","responsibility":"none","reasoning":"mixed","policy":"none","emotion":"outrage"} ]