AlgoSoc / Publication
October 05, 2023

Foundation models and the privatization of public knowledge

To protect the integrity of knowledge production, the training procedures of foundation models such as GPT-4 need to be made accessible to regulators and researchers. Foundation models must become open and public, and those are not the same thing", writes AlgoSoc PI José van Dijck with researchers Fabian Ferrari and Antal van den Bosch in the September issue of Nature Machine Intelligence.

Their article discusses concerns surrounding ChatGPT, an AI developed by OpenAI. These concerns encompass safety, plagiarism, bias, and accuracy, with a particular focus on how large language models (LLMs) owned by private entities may impact knowledge production. The authors notes that OpenAI, despite its nonprofit origins, has transformed partly into a capped-profit company with Microsoft as a major investor. The issue of whether regulators and scientists will have access to the inner workings of deep neural networks, including their training data and procedures, remains under-addressed. The lack of transparency could hinder thorough inspection, replication, and testing, potentially compromising the integrity of public knowledge.

In their article, Van Dijck et al. also examine the training data and procedures of foundation models like GPT-4 and PaLM. While some information is available about the datasets used, details about their specific composition and sources remain undisclosed. The author highlights that making foundation models open and public is crucial. "Open" refers to making them available for detailed inspection and replication, while "public" implies treating them as utilities accessible to all. Van Dijck et al. suggest that ensuring transparency, implementing technical safeguards, and enacting legal and regulatory measures are essential steps. Moreover, they emphasize the need for improved AI literacy among citizens to foster a democratic understanding of these technologies in the context of knowledge creation.

Cite this article: Ferrari, F., van Dijck, J. & van den Bosch, A. Foundation models and the privatization of public knowledge. Nat Mach Intell 5, 818–820 (2023). https://doi.org/10.1038/s42256...

More results /

/ algosoc
Universiteit, maak je los van Big Tech

By Albert Meijer • José van Dijck • February 24, 2025

De concentratie van macht over AI is ondemocratisch

By Natali Helberger • February 13, 2025

(Hoe) blijven universiteiten baas over hun bestanden in de cloud?

By Seda Gürses • Roel Dobbe • February 13, 2025

/ justice
The rise of technology courts

By Natali Helberger • March 06, 2025

AI in the justice sector: a force to be reckoned with

By Kätliin Kelder • June 18, 2024

/ health
From dialogue to decision

By Leonie Westerbeek • November 22, 2024

Discriminerende algoritmes

By Corien Prins • May 18, 2021

/ media
Misinformatie in de gezondheidszorg

By José van Dijck • February 26, 2025

Subscribe to our newsletter and receive the latest research results, blogs and news directly in your mailbox.

Subscribe