Theoretical Framework
PyEyesWeb inherits from the rich tradition of computational movement analysis initiated by the EyesWeb project 123 and grounds on the multi-layered computational framework of qualities in movement developed in the DANCE project 4.
Conceptual Model
Warning
This is the result of an open field of research. As such, certain aspects of the model may be provisional or subject to refinement. Some concepts are open to interpretation, and current limitations are actively being addressed in the ongoing work.
The framework describes how raw sensor data can be progressively transformed into meaningful descriptions of expressive movement qualities and is organized into four layers.
The layers represent a conceptual model and not a strict processing pipeline.
Layers Overview
Attention!
The concept of timescale is crucial in this framework, and each layer operates at different temporal scales.
As an example, a key distinction from Layers 2 and 3 is moving from instantaneous or short-window features (~0.5s) to longer windows (0.5-3s) or movement units (e.g., a specific sport gesture, a choreographic phase).
One same feature can occur at different layers and yield different interpretations depending on the timescale of analysis.
Layer 1 – Physical Signals : raw data captured by virtual sensors, i.e., physical devices (motion capture, accelerometers, video, RGB-D cameras, physiological sensors, etc.) enriched with preprocessing (denoising, filtering, extraction of trajectories, silhouettes, respiration, etc.).
Foundation for all higher layers. → Learn more
Layer 2 – Low-Level Features : instantaneous or short-window (0.5s) descriptors computed from physical signals. Examples are velocity, acceleration, kinetic energy, balance, smoothness, etc.
Represented as time-series. → Learn more
Layer 3 – Mid-Level Features : operates on movement units or longer windows. Examples are directness, lightness, suddenness, fluidity, repetitiveness.
Introduce amodal descriptors across modalities. → Learn more
Layer 4 – Expressive Qualities : focuses on what an observer perceives from movement: emotional expression, saliency, attraction/repulsion, hesitation, predictability. Involves memory and context, influencing how movement is interpreted.
Requires context and ML mappings. → Learn more
Analysis Primitives : Core computational tools applied across all layers. Includes: statistical moments, entropy, shape descriptors (peaks, slopes), synchronization, time-frequency transforms, predictive and physical models (e.g., mass–spring).
Provide the building blocks for extracting meaningful features. → Learn more
References
-
Camurri, A., Mazzarino, B., & Volpe, G. (2003, April). Analysis of expressive gesture: The eyesweb expressive gesture processing library. In International gesture workshop (pp. 460-467). Berlin, Heidelberg: Springer Berlin Heidelberg. ↩
-
Camurri, A., Coletta, P., Massari, A., Mazzarino, B., Peri, M., Ricchetti, M., ... & Volpe, G. (2004, March). Toward real-time multimodal processing: EyesWeb 4.0. In Proc. AISB (pp. 22-26). ↩
-
Volpe, G., Alborno, P., Camurri, A., Coletta, P., Ghisio, S., Mancini, M., ... & Sagoleo, R. (2016). Designing multimodal interactive systems using EyesWeb XMI. In CEUR Workshop Proceedings (pp. 49-56). CEUR-WS. ↩
-
Camurri, A., Volpe, G., Piana, S., Mancini, M., Niewiadomski, R., Ferrari, N., & Canepa, C. (2016, July). The dancer in the eye: towards a multi-layered computational framework of qualities in movement. In Proceedings of the 3rd International Symposium on Movement and Computing (pp. 1-7). ↩