Ask a large language model (LLM) such as ChatGPT to summarize what people are saying about a topic, and although the model might summarize the facts efficiently, it might give a false impression of how people feel about the topic. LLMs play an increasingly large role in research, but rather than being a transparent window into the world, they can present and summarize content with a different tone and emphasis than the original data, potentially skewing research results. Ask a large language model (LLM) such as ChatGPT to summarize what people are saying about a topic, and although the model might summarize the facts efficiently, it might give a false impression of how people feel about the topic. LLMs play an increasingly large role in research, but rather than being a transparent window into the world, they can present and summarize content with a different tone and emphasis than the original data, potentially skewing research results. Social Sciences Political science Phys.org – latest science and technology news stories
AI’s emotional blunting effect: Researchers find LLMs can neutralize sentiments of original text
