In this paper, we propose a new approach to reduce this gap by
constructing, in an unsupervised manner, an intermediate layer
between low-level information (tracked objects from video) and
high-level interpretation of activity (e.g. cooking, eating and
sitting). Our method is a novel approach allowing the detection of
complex activities with long duration in an unstructured scene. We
have developed a complete vision-based framework that enables us
to model, discover and recognise activities online while monitoring
a patient. The two main contributions of this work are as follows:
a patient. The two main contributions of this work are as follows: