News feed platforms, such as Facebook and Twitter, are growing continuously. Their primary requirements are scalability, low request latencies and high availability  for read and write requests. This requires to scale out the system to multiple machines . STOU and Graphity were reported as high performant algorithms to power a news feed system.
In the last years, crowdsourcing has become a popular technique for evaluating systems, as well as for cleansing, enhancing and labeling data. While there are many success stories both in research and industry, there are still many challenges to overcome.
We look at a very popular construct in probabilistic modelling: the Dirichlet process (DP). Starting with the Chinese restaurant metaphor, we see how context can be modelled using DPs and understand some of the underlying assumptions.
Bis vor wenigen Jahren konnte zwischen virtuellen Spielen und Spielen in der physischen Welt klar unterschieden werden. Die wachsende Verbreitung von Mobiltelefonen mit hoher Rechenleistung, verschiedenen Sensoren und ständiger Internetverbindung ermöglicht es, Spiele zu entwickeln, die virtuelle und physische Umgebung vereinen.
Quite often, Linked Open Data (LOD) applications pre-fetch data from the Web and store local copies of it in a cache for faster access at runtime. Yet, recent investigations have shown that data published and interlinked on the LOD cloud is subject to frequent changes.
In anticipation of RDF graphs exceeding one trillion triples, the W3C tested RDF stores whether they can deal with such huge graphs. This amount of data can be stored in a cloud at a reasonable price. But storing a graph in a cloud consisting of several individual computers raises several issues like the triple placement or the efficient processing of interactive queries.
In this talk I will review and evaluate models of network evolution based on the notion of structural diversity. I show that diversity is an underlying theme of three principles of network evolution: the preferential attachment model, connectivity and link prediction. I show that in all three cases, a dominant trend towards shrinking diversity is apparent, both theoretically and empirically.
Identifying suitable data sets is a crucial task in several fields of application, like data analysis, but the task itself is highly complicated and mostly heavily related to manually skimming through a vast amounts of data.
An claim made quite often is that types in RDF and the Semantic Web differ fundamentally from the types as they exist in programming languages. The claim is usually aimed towards the way modern, existing languages treat data types.
Wikidata, the free knowledge base of Wikipedia, is one of the largest collections of human-authored structured information that are freely available on the Web. It is curated by a unique community of tens of thousands of editors who contribute in up to 400 different languages.