ΑΙhub.org
 

COVID-19 Open Research Dataset (CORD-19) now available for researchers


by
17 March 2020



share this:
CORD-19 dataset

On 16 March the COVID-19 Open Research Dataset (CORD-19) was released. This comprises an open-source, machine-readable collection of scholarly literature covering COVID-19, SARS-CoV-2, and the Coronavirus group. This free resource contains over 29,000 relevant scholarly articles, including over 13,000 with full text.

The release of the dataset is a result of a collaborate effort between the Allen Institute for AI, Chan Zuckerberg Initiative, Georgetown University, Microsoft, and the US National Library of Medicine (NLM). This resource is intended to mobilize researchers to apply recent advances in natural language processing to generate new insights in support of the fight against this infectious disease.

The CORD-19 dataset is available on the Allen Institute’s SemanticScholar.org website and will continue to be updated as new research is published in archival services and peer-reviewed publications.

Kaggle is hosting a challenge using this dataset and at present there are 10 initial tasks for people to work on. These key scientific questions have been drawn from the National Academies of Sciences, Engineering, and Medicine’s research topics and the World Health Organization’s R&D Blueprint for COVID-19.

Links:

You can access the official webpage for CORD-19 here .
Find the kaggle challenge page here.




Lucy Smith is Senior Managing Editor for AIhub.
Lucy Smith is Senior Managing Editor for AIhub.

            AIhub is supported by:



Subscribe to AIhub newsletter on substack



Related posts :

Studying the properties of large language models: an interview with Maxime Meyer

  11 Mar 2026
What happens when you increase the prompt length in a LLM? In the latest interview in our AAAI Doctoral Consortium series, we sat down with Maxime, a PhD student in Singapore.

What the Moltbook experiment is teaching us about AI

An experimental social media platform where only AI bots can post reveals surprising lessons about artificial intelligence behaviour and safety.

The malleable mind: context accumulation drives LLM’s belief drift

  09 Mar 2026
LLMs change their "beliefs" over time, depending on the data they are given.

RWDS Big Questions: how do we balance innovation and regulation in the world of AI?

  06 Mar 2026
The panel explores the tensions, trade-offs and practical realities facing policymakers and data scientists alike.

Studying multiplicity: an interview with Prakhar Ganesh

  05 Mar 2026
What is multiplicity, and what implications does it have for fairness, privacy and interpretability in real-world systems?

Top AI ethics and policy issues of 2025 and what to expect in 2026

, and   04 Mar 2026
In the latest issue of AI Matters, a publication of ACM SIGAI, Larry Medsker summarised the year in AI ethics and policy, and looked ahead to 2026.

The greatest risk of AI in higher education isn’t cheating – it’s the erosion of learning itself

  03 Mar 2026
Will AI hollow out the pipeline of students, researchers and faculty that is the basis of today’s universities?

Forthcoming machine learning and AI seminars: March 2026 edition

  02 Mar 2026
A list of free-to-attend AI-related seminars that are scheduled to take place between 2 March and 30 April 2026.



AIhub is supported by:







Subscribe to AIhub newsletter on substack




 















©2026.02 - Association for the Understanding of Artificial Intelligence