Page 27: Research news on AI alignment

AI alignment examines how artificial systems acquire, represent, and act on goals, values, and social norms, and why their behavior often diverges from human expectations. Work in this area studies systematic failures such as bias, sycophancy, hallucinations, deceptive or selfish reasoning, and cultural or linguistic inequities, as well as limitations in commonsense, emotion, and social understanding. It also develops methods for preference learning, norm-following, interpretability, and reliability guarantees to better align AI behavior with human values and societal constraints.

Machine learning & AI

Firms and researchers at odds over superhuman AI

Hype is growing from leaders of major AI companies that "strong" computer intelligence will imminently outstrip humans, but many researchers in the field see the claims as marketing spin.

Machine learning & AI

This AI model is more certain about uncertainty

Artificial intelligence (AI) plays a role in virtually every aspect of our lives, from self-driving cars to smart vacuum cleaners, to computer models that can predict the course of an epidemic. No matter how advanced these ...

Machine learning & AI

Navigating trust in an age of increasing AI influence

In 2025, it can seem as though the future generations of AI advocates promised have finally arrived. We see the benefits of artificial intelligence on a daily basis—we use it to help us navigate traffic, to identify new ...

page 27 from 28