"When a user says, 'Hey, ChatGPT, what happened at Davos today?' we would like to display content, link out to brands of places like the New York Times or the Wall Street Journal or any other great publication and say, 'Here's what happened today' and then we'd like to pay for that. We'd like to drive traffic for that," said Altman, adding it's not a priority to train models on that data, just display it.
In future, LLMs will be able to take smaller amounts of higher quality data during their training process and think harder about it and learn more.
When content is used for training, Altman said we need new economic models that would compensate content owners.
I think this is really interesting, because search engines fought tooth and nail about having to pay the sources that they linked to, whereas it is a natural extension of large language model providers to have to pay for the content that they are training on and serving. Generative AI might save journalism to some extent, as the robot must consume to be useful.
"When a user says, 'Hey, ChatGPT, what happened at Davos today?' we would like to display content, link out to brands of places like the New York Times or the Wall Street Journal or any other great publication and say, 'Here's what happened today' and then we'd like to pay for that. We'd like to drive traffic for that," said Altman, adding it's not a priority to train models on that data, just display it.
In future, LLMs will be able to take smaller amounts of higher quality data during their training process and think harder about it and learn more.
When content is used for training, Altman said we need new economic models that would compensate content owners.