In an unprecedented move, The New York Times has announced that it will no longer allow the use of its content for artificial intelligence (AI) model training.
The decision follows recent reports that IBM was using New York Times articles to train an AI model for its Natural Language Understanding platform. According to reports, the data set was primarily comprised of articles published prior to 2018, but the Times says it was never authorized to use the newspaper’s content in this manner.
The decision signals a growing trend of organizations restricting access to their data for ethical reasons. The Times’ newsroom ethics statement says that its content should “not be used to train algorithms and AI models that are developed for commercial purposes” and that its material should not be used to “exert power over populations or large groups of people.”
The move raises important questions about the ethical considerations at play in the development of AI. While companies may be eager to take advantage of content from major news organizations to develop powerful AI tools, media organizations are increasingly aware of the potential for misuse of their material.
There are also questions about what role large media organizations should play in the development of AI tools. With access to millions of words of content, organizations like The New York Times are in a unique position to shape the standards for ethical use of their material. The Times’ decision suggests that it wants to lead the way in establishing ethical standards for using its content in the development of AI models.
The decision is likely to have a ripple effect as other major media organizations consider their own ethical standards for using their content in the development of AI tools. While data may be essential for AI development, organizations should be expected to take measures to protect the ethical use of their material. The New York Times has made a bold move that will set a precedent for other news organizations to follow.