Detailed Information on Publication Record
2013
MUSE framework 1.1.1
RUSŇÁK, Vít, Lukáš RUČKA, Martin ĽUPTÁK and Petr HOLUBBasic information
Original name
MUSE framework 1.1.1
Authors
RUSŇÁK, Vít (203 Czech Republic, guarantor, belonging to the institution), Lukáš RUČKA (203 Czech Republic, belonging to the institution), Martin ĽUPTÁK (703 Slovakia, belonging to the institution) and Petr HOLUB (203 Czech Republic, belonging to the institution)
Edition
2013
Other information
Language
English
Type of outcome
Software
Field of Study
10201 Computer sciences, information science, bioinformatics
Country of publisher
Czech Republic
Confidentiality degree
není předmětem státního či obchodního tajemství
References:
RIV identification code
RIV/00216224:14330/13:00072055
Organization unit
Faculty of Informatics
Keywords in English
multi-touch; human-computer interaction; HCI; interaction framework; gesture recognition; group collaborative environment
Technical parameters
Kontaktní osoba: Vít Rusňák, Fakulta Informatiky, Masarykova univerzita, Botanická 68a, Brno, e-mail:xrusnak@fi.muni.cz; Software je šířen pod BSD licencí. Uživatel software souhlasí s podmínkami této licence.
Změněno: 14/1/2015 11:16, RNDr. Vít Rusňák, Ph.D.
Abstract
V originále
MUSE is a framework for developing cost-affordable interactive environments. It allows for rapid development of interactive systems based on tabletops and interactive tiled-display walls. It serves as a testing environment for a development of new interaction approaches for large-scale interactive systems. The framework provides different interaction possibilities of multimodal user interfaces which are made of multiple low-cost commodity devices (e.g., single- and multi-touch overlay panels and foils, web cameras, depth sensors). Main features of the framework: a) coupling of multiple low-cost commodity multi-touch sensors which are represented as a single seamless interface; b) functions for distinguishing and continuous association of users and touch operations they performed; c) hand-tracking for identification and distinguishing users' body parts (e.g., hands) used in association with touch input events; d) integration of gesture recognition algorithms Protractor, $1 and $N to provide experimental gesture recognition of touch events (libreco library) e) extended semantic description of input events enabling their personalisation (e.g., personalized gesture recognition); (libdtuio library) f) highly configurable toolkit and application base. Supported HW: majority of existing resistive and capacitive single- and multi-touch sensor overlay panels and devices, MS Kinect depth tracker.