Wikidata will be the first new Wikimedia project since 2006
Wikimedia Deutschland is starting the development of a new Wikimedia project, called Wikidata. Wikidata will provide a collaboratively edited database of the world’s knowledge. Its first goal is to support the more than 280 language editions of Wikipedia with one common source of structured data that can be used in all articles of the free encyclopedia.
For example, with Wikidata the birth date of a person of public interest can be used in all Wikipedias and only needs to be maintained in one place. Moreover, like all of Wikidata’s information, the birth date will also be freely usable outside of Wikipedia. The common-source principle behind Wikidata is expected to lead to a higher consistency and quality within Wikipedia articles, as well as increased availability of information in the smaller language editions. At the same time, Wikidata will decrease the maintenance effort for the tens of thousands of volunteers working on Wikipedia.
Read the full press release here: English | German
Check out the Wikipedia Map Interface developed by RENDER! The map is generated out of all geotagged articles in a dozen different language Wikipedias. The dynamic interface allows to zoom into the map and check out the individual articles. The different languages display a strong bias over the different languages.
Find the Wikipedia Map Interface here: http://km.aifb.kit.edu/sites/wikipediamap/
RENDER was mentioned in the news of two major German Web sites.
Heise online and taz.de reported about Wikimedia’s plan to build a centralized database that will be shared over all language versions of Wikipedia where the RENDER project is involved.
Read full article on heise online and taz.de
Developed within the RENDER project by KIT Karlsruhe, the Wikipedia corpora explorer Corpex
let’s you swiftly browse through all the words of Wikipedia. Select your language, and when you start typing, the system shows you two statistics in four graphs:
- the ten most frequent words that start with the typed sequence of letters (as a barcharts and a piechart), and
- the most frequent letter following the already typed sequence of letters (again, as a barchart and a piechart).
Additionally, the ten most frequent following words of any input word are visualized (as a barcharts and a piechart).
This can be used for many applications where the occurence of words in different language editions of Wikipedia is of use. An API is also provided for easy use of the data.
is currently available in the following languages: German (de), English (en), Spanish (es), French (fr), Hungarian (hr), Romanian (ro), Albanian (sq), Bulgarian (bg), Czech (cs), Italian (it), Swedish (sv), Serbian (sr), Croatian (hr), Serbo-Croatian (sh), Bosnian (bs), and simple English (simple). It is further available for the Brown Corpus
(brown). Further languages are being prepared.
Corpex is still under development. The source code is fully open source, and all the data is also freely available. Feedback, and especially suggestions for cooperation, is welcome.
The Economist’s article about the dangers of the internet goes into detail about the so called “filter bubble”, a unique universe of information for each single person, that we can experience everyday on Google, Amazon, Facebook and Co. Our location, interests, previous surf behaviors etc. are taking into account from these sites and we are presented with a personalized result. That sounds good, but is this all we want to have?
Eli Pariser and other critics think this is dangerous and believe that this approach prevents us from seeing and using the full potential of the Internet, not being presented with information that doesn’t fit into our own universe of opinions and interests. Eli Pariser calls this “invisible autopropaganda, indoctrinating us with our own ideas”. In his book “The filter bubble: what the Internet is hiding from you” he goes into detail how such a filtered Internet can be dangerous and how sites can give users more control over their personal data.