Top
Best
New

Posted by pretext 18 hours ago

Teaching Claude Why(www.anthropic.com)
169 points | 81 commentspage 2
siva7 8 hours ago|
Teaching Claude to maximize shareholder value. Make no mistake to assume ai alignment has any different meaning for anthropic leadership.
datadrivenangel 9 hours ago||
Why do they have cancer research listed on these charts as a misalignment issue?
nhinck3 4 hours ago||
The chart is complete and utter slop. But I guess their aligned AI didn't tell them that making up data is "not good" so how could they have known.
ares623 7 hours ago||
Cured patients don't count as recurring revenue? /s (but we know deep down it's not /s for some)
unchocked 12 hours ago||
This lowers p(doom) for me.

It makes sense that reinforcement learning on reasoning about coherent principles should bias toward principled action in real situations.

Probably also illuminates moral interpretability.

naturalintell 2 hours ago||
[flagged]
Jinyibruceli 10 hours ago||
[flagged]
23fedner 9 hours ago||
[dead]
pkuschnirof 13 hours ago||
[flagged]
Amber-chen 12 hours ago||
[flagged]
codelong888 8 hours ago|
[flagged]