Nowadays, the majority of the human-computer interaction (HCI) is based on mechanical devices such as
keyboards, mouses, joysticks or gamepads. In recent years there has been a growing interest in a class of
methods based on computational vision due to its ability to recognise human gestures in a natural way [1].
These methods use as input the images acquired from a camera or from a stereo pair of cameras. The main
goal of these algorithms is to measure the hand configuration in each time instant.