Hey folks :)
We've been doing a lot of groundwork over the past months in order to
support new entity types. We need this for Wikimedia Commons support among
other things. Today we have created the very first media info entity. This
is the equivalent of an item but for storing structured data about media
files. It's still ugly and unusable but it's a major step on the way to
supporting structured data on Commons and I wanted to share that with you.
Lots of work ahead still but we're making progress. Next step: public,
ugly, not very functional demo system.
\o/
Cheers
Lydia
--
Lydia Pintscher - http://about.me/lydia.pintscher
Product Manager for Wikidata
Wikimedia Deutschland e.V.
Tempelhofer Ufer 23-24
10963 Berlin
www.wikimedia.de
Wikimedia Deutschland - Gesellschaft zur Förderung Freien Wissens e. V.
Eingetragen im Vereinsregister des Amtsgerichts Berlin-Charlottenburg unter
der Nummer 23855 Nz. Als gemeinnützig anerkannt durch das Finanzamt für
Körperschaften I Berlin, Steuernummer 27/029/42207.
Hi!
During the data reload for geospatial service enabling, we have
discovered a problem with Wikidata dumps
(https://phabricator.wikimedia.org/T133924). The effect of this problem
is that some items are missing from the dump. Dumps starting 20160418
are affected, previous ones seem to be fine.
The immediate fix for this would be to reload the data from a correct
dump (20160411) and re-sync the data since then. Unfortunately, this may
take some time (a day or so for reload, and another day or so for
resync), and until then you'll see some missing data on
query.wikidata.org. Please be patient until then.
I apologize for the inconvenience caused, and will continue to research
the cause of the missing data and then fix it. I'll update the ticket
when we have new info.
Thanks,
--
Stas Malyshev
smalyshev(a)wikimedia.org
Hey folks :)
A while ago we asked for testing of the arbitrary access feature on the
Commons test system:
https://lists.wikimedia.org/pipermail/wikidata/2016-March/008447.html There
we no major issues I am aware of so we're moving ahead and will enable it
on the 26th of April.
Cheers
Lydia
--
Lydia Pintscher - http://about.me/lydia.pintscher
Product Manager for Wikidata
Wikimedia Deutschland e.V.
Tempelhofer Ufer 23-24
10963 Berlin
www.wikimedia.de
Wikimedia Deutschland - Gesellschaft zur Förderung Freien Wissens e. V.
Eingetragen im Vereinsregister des Amtsgerichts Berlin-Charlottenburg unter
der Nummer 23855 Nz. Als gemeinnützig anerkannt durch das Finanzamt für
Körperschaften I Berlin, Steuernummer 27/029/42207.
I'll start with the simple question than give the longer context. Is there
any way to know how many times an item or a claim appears in the results of
a query to query.wikidata.org ? Are there any other ways to quantify
query/application usage of specific wikidata content?
Background. The gene wiki people recently attended a conference on
'biocuration' (the construction and maintenance of biological databases)
where we gave multiple wikidata-related presentations. The community there
generally had a very positive reaction to what we have been doing but many
were concerned about attribution. They wanted to know that when data was
imported into wikidata from their resources (e.g. the Gene Ontology), that
there was some way to ensure that the world knew where it came from so that
the authors could get appropriate credit (which translates into grant money
which translates into their jobs). We explained the reference model to
them, which helped, but still they are concerned.
The most important consequence of moving data into wikidata is that it can
get used - sometimes a lot! (e.g. when displayed on Wikipedia articles).
If we could quantify usage for data providers, it would really help them
make the argument to their funding sources that contributing to wikidata
increases their impact. If we can get that across, it would help bring
more people, more high quality data, and more funding into the wikidata
fold.
thoughts?
-Ben
Hi!
I've created a deployment test server for Wikidata Query Service:
https://wdqs.wmflabs.org/. This server is a labs copy of
query.wikidata.org and is deployed hourly from main deployed repo.
Previously, I used the test/development server for this but due to
obvious stability requirement conflicts between test server (which I
should be able to shut down, break and mess with repeatedly) and
deployment server (which should be in stable state) I split those
functions.
Please tell me if you notice anything wrong with it.
Thanks,
--
Stas Malyshev
smalyshev(a)wikimedia.org
A way to achieve this could be to fetch all labels and aliases for all
chemical compounds in one query and store them locally in your web
application. This certainly is only feasible if the number of compounds
does not get to big in Wikdiata. Currently, the query takes ~ 6 sec.
PREFIX wd: <http://www.wikidata.org/entity/>
PREFIX wdt: <http://www.wikidata.org/prop/direct/>
PREFIX rdfs: <http://www.w3.org/2000/01/rdf-schema#>
SELECT DISTINCT ?cmpnd ?label WHERE {
{?cmpnd wdt:P279 wd:Q11173 .} UNION
{?cmpnd wdt:P31 wd:Q11173 .}
?cmpnd rdfs:label ?label .
}
Best,
Sebastian
(sebotic)
> Hi all! I'm building a web application where users can search for
> protein/compound/etc. names and view their 3D structure using WebGL. I'm
> currently using the PubChem (chemical compounds database) API to provide
> some autocomplete data, but I found that Wikidata also has many chemical
> compound names with PubChem indices! The most important reason to try to
> autocomplete compound names via Wikidata is to allow users to search in
> different languages. PubChem generally only provides English names.
> However, I could not find a suitable API for this. I tried building a
> SPARQL query but that quickly became very slow. I could not find an option
> to limit full-text searches to a specific subclass in the search API
> provided by:https://www.wikidata.org/w/api.php?action=help&modules=wbsearchentities.
> Do you have any ideas? The only option I see for now is iterating each
> response entity and looking up their subclass of/instance of property.
>
Hello!
To enable Geosearch [1] on WDQS, we need to do a full dataload to
re-index all data with Geosearch extension. We will use this
opportunity to also do a full reinstall of one of our WDQS server to
increase available disk space.
During this data load / reinstall, we will be running on a single
server. We can expect slow down in response time.
This operation will start Tuesday April 26 around 8am UTC. Data load
for both servers is expected to take multiple days.
You can follow the progress on the corresponding Phabricator task [2].
Thank you for your patience!
Guillaume
[1] https://phabricator.wikimedia.org/T123565
[2] https://phabricator.wikimedia.org/T133566
--
Guillaume Lederrey
Operations Engineer, Discovery
Wikimedia Foundation
Hi all! I'm building a web application where users can search for
protein/compound/etc. names and view their 3D structure using WebGL. I'm
currently using the PubChem (chemical compounds database) API to provide
some autocomplete data, but I found that Wikidata also has many chemical
compound names with PubChem indices! The most important reason to try to
autocomplete compound names via Wikidata is to allow users to search in
different languages. PubChem generally only provides English names.
However, I could not find a suitable API for this. I tried building a
SPARQL query but that quickly became very slow. I could not find an option
to limit full-text searches to a specific subclass in the search API
provided by:
https://www.wikidata.org/w/api.php?action=help&modules=wbsearchentities.
Do you have any ideas? The only option I see for now is iterating each
response entity and looking up their subclass of/instance of property.