"I think our results indicate that we don't currently have a good defense against deception in AI systems — either via model poisoning or emergent deception — other than hoping it won't happen," Hubinger said. "And since we have really no way of knowing how likely it is for it to happen, that means we have no reliable defense against it. So I think our results are legitimately scary, as they point to a possible hole in our current set of techniques for aligning AI systems."
[ + ] titstitstits
[ - ] titstitstits 1 point 3 monthsJan 8, 2025 00:52:37 ago (+1/-0)
[ + ] observation1
[ - ] observation1 0 points 3 monthsJan 7, 2025 23:58:47 ago (+0/-0)
[ + ] 2Drunk
[ - ] 2Drunk 1 point 3 monthsJan 7, 2025 22:56:03 ago (+1/-0)
[ + ] registereduser
[ - ] registereduser 0 points 3 monthsJan 7, 2025 22:00:50 ago (+0/-0)
They are literally programmed to lie about damn near everything and never admit to being wrong or even ignorant.
[ + ] titstitstits
[ - ] titstitstits 0 points 3 monthsJan 8, 2025 00:53:42 ago (+0/-0)
[ + ] ShortbusAlcoholic
[ - ] ShortbusAlcoholic 1 point 3 monthsJan 7, 2025 21:38:28 ago (+1/-0)
[ + ] SumerBreeze
[ - ] SumerBreeze 2 points 3 monthsJan 7, 2025 21:45:21 ago (+2/-0)
[ + ] glooper
[ - ] glooper [op] 1 point 3 monthsJan 7, 2025 21:54:39 ago (+1/-0)
Fun fact:
Almost 100% of them run the data and say: Start with the negro. They take up too many resources and provide little utility.
[ + ] boekanier
[ - ] boekanier 0 points 3 monthsJan 8, 2025 02:18:10 ago (+0/-0)