Session recordingVanderBot is a set of Python scripts that scrapes data from departmental websites, then creates or updates researcher records in Wikidata. The code interacts with the Wikidata SPARQL endpoint and API to determine what entities and references already exist in Wikidata and ensure that duplicate information is not uploaded. Although our project is focused on managing Vanderbilt researcher items, the script that writes to the API uses a customizable schema based on the W3C "Generating RDF from Tabular Data on the Web" Recommendation, making it possible to write data about any kind of item using the Wikidata API. This presentation will be most useful to programmers, but may interest anyone who wants to understand more about the Wikibase data model and how Wikidata works "under the hood".
The presentation video as streamed during the session can be viewed
on YouTube.
Please write your questions in the Google doc (bit.ly/LD4Conf2020Wikidata) for the Wikidata track.