In the last years, scalable RDF stores in the cloud have been developed, where graph data is distributed over compute and storage nodes for scaling efforts of query processing and memory needs. One main challenge in these RDF stores is the data placement strategy that can be formalized in terms of graph covers. These graph covers determine whether (a) the triples distribution is well-balanced over all storage nodes (storage balance) (b) different query results may be computed on several compute nodes in parallel (vertical parallelization) and (c) individual query results can be produced only from triples assigned to few - ideally one - storage node (horizontal containment).
In the last few years, machine learning techniques have been successfully applied to many application areas such as information retrieval , e-commerce, image processing , computational biology, and chemistry. To understand and explore the real datasets, we often apply machine learning techniques such as clustering or classification in a high dimensional space. However, developing these machine learning models on large data sets can be very time-consuming because of its high dimensionality. Dimensionality reduction is the most important technique in unsupervised learning, to get a meaningful structure or previously unknown patterns in the multivariate data.
Koldfish aims at providing means for consuming Linked Open Data comfortably. However, at present it is unable to incorporate semantic data stemming from, e.g., schema.org-annoted web sites. Rectifying this technical shortcoming may also open up opportunities for a refined conceptual indexing of data elements. This talk will layout the proposed change to the Koldfish system and discuss potential implications for its Schema Index.
Ein solcher Ansatz für das Reasoning in der Abstrakten Argumentationssystemen nach Dung sind Backtracking-Algorithmen, die Extensionen aufzählen. Zur Optimierung dieser Algorithmen sollen verschiedene Heuristiken verglichen werden.
For semantic analysis of activities and events in videos, it is important to capture the spatio-temporal relation among objects in the 3D space. This presentation introduces a probabilistic method that extracts 3D trajectories of objects from 2D videos, captured from a monocular moving camera. Compared to existing methods
that rely on restrictive assumptions, the presented method can extract 3D trajectories with much less restriction by adopting new example-based techniques which compensate the lack of information. Here, the focal length of the camera is estimated based on similar candidates, and used afterwards to compute depths of detected objects.
Approximate reasoning is regarded one of the most convincing approaches to reasoning with ontologies and knowledge graphs in applications. This talk has two parts. In the first part, I will explain why approximate reasoning might work and how to perform faithful approximate reasoning, i.e, approximate reasoning with some level of quality control. In the second part, I will share some further thoughts towards a new roadmap of approximate reasoning in the era of Knowledge Graph.
The concept of relevance was proposed to model different temporal effects in networks (e.g. "aging effect", i.e. how the interest of nodes decay over time), where the traditional preferential attachment (PA) model fails to explain. We analyze the citation data provided by American Physical Society (APS). We group papers by their final in-degrees (# of papers citing them), and do not observe an obvious decline of citations for the most cited papers (indegree>1000). This might be due to the fact that the size of the network (total # of papers) grows exponentially, which compensates the decay of papers' relevance. For the next step, we want to analyze different citation networks (ACM, DBLP).
Gaze-based virtual keyboards provide an effective interface for text entry by eye movements. The efficiency and usability of these keyboards have traditionally been evaluated with conventional text entry performance measures such as words per minute, keystrokes per character, backspace usage, etc. However, in comparison to the traditional text entry approaches, gaze-based typing involves natural eye movements that are highly correlated with human brain cognition. Employing eye gaze as an input could lead to excessive mental demand, and in this work we argue the need to include cognitive load as an eye typing evaluation measure.
Detection and handling of misinformation is one of nowadays' hot topics, which is being specially discussed and researched with the existence of social networks and exponential spreading of political fake news in the web. The main point here is to first understand (1) what is misinformation, (2) which structure does it contain (present), (3) how can it be discovered (misinformation prediction) and (4) how can it's spreading be prevented (misinformation prevention)?! In general, I would define whole this as a general term of "Misinformation Analytics (MA)". The first step in the MA is to understood what and how machine learning algorithms could support autonomous fake news detection and handling. Unsupervised learning is one of candidate solutions in this regard.
SoLiD is the Social Linked Data platform developed by Tim Berners Lee's Group at the Distributed Information Group
at MIT, which extends the web to move us from a hyper text browser web to a hyper data application web. Henry will
give an overview of the architecture of this project, philosophical issues that spawn it, and technical issues that he
discovered while building both server and client applications for it.