Computer Sciences

Large sequence models for sequential decision-making

Transformer architectures have facilitated the development of large-scale and general-purpose sequence models for prediction tasks in natural language processing and computer vision, for example, GPT-3 and Swin Transformer.

Machine learning & AI

Understanding attention in large language models

Chatbot users often recommend treating a series of prompts like a conversation, but how does the chatbot know what you're referring back to? A new study reveals the mechanism used by transformer models—like those driving ...

Computer Sciences

Study shows ChatGPT writes better school essays than students

In a study published in Scientific Reports, a research team from the University of Passau compared the quality of machine-generated content with essays written by secondary school students. The upshot: The AI-based chatbot ...

Computer Sciences

Generative AI explained

A quick scan of the headlines makes it seem like generative artificial intelligence is everywhere these days. In fact, some of those headlines may actually have been written by generative AI, like OpenAI's ChatGPT, a chatbot ...

page 2 from 11