What are alignment problems in AI?

Brian Leach

about 1 month ago

I keep seeing 'alignment' come up in safety discussions. Can someone explain what alignment means in this context?

Julie Castillo MD

about 1 month ago

Alignment refers to ensuring AI systems pursue the goals we actually want them to pursue, not just what we think we're telling them.

Brian Leach

about 1 month ago

It's about the difference between what we specify and what we actually want. Classic example: asking for paperclips and getting the universe converted to paperclips.

Shelly Lang

about 1 month ago

The challenge becomes even more complex as AI systems become more capable and autonomous.