RUSŇÁK, Vít, Lukáš RUČKA, Martin ĽUPTÁK and Petr HOLUB. MUSE framework 1.1.1. 2013.
Other formats:   BibTeX LaTeX RIS
Basic information
Original name MUSE framework 1.1.1
Authors RUSŇÁK, Vít (203 Czech Republic, guarantor, belonging to the institution), Lukáš RUČKA (203 Czech Republic, belonging to the institution), Martin ĽUPTÁK (703 Slovakia, belonging to the institution) and Petr HOLUB (203 Czech Republic, belonging to the institution).
Edition 2013.
Other information
Original language English
Type of outcome Software
Field of Study 10201 Computer sciences, information science, bioinformatics
Country of publisher Czech Republic
Confidentiality degree is not subject to a state or trade secret
WWW Domovská stránka projektu MUSE Framework
RIV identification code RIV/00216224:14330/13:00072055
Organization unit Faculty of Informatics
Keywords in English multi-touch; human-computer interaction; HCI; interaction framework; gesture recognition; group collaborative environment
Technical parameters Kontaktní osoba: Vít Rusňák, Fakulta Informatiky, Masarykova univerzita, Botanická 68a, Brno, e-mail:xrusnak@fi.muni.cz; Software je šířen pod BSD licencí. Uživatel software souhlasí s podmínkami této licence.
Changed by Changed by: RNDr. Vít Rusňák, Ph.D., učo 172757. Changed: 14/1/2015 11:16.
Abstract
MUSE is a framework for developing cost-affordable interactive environments. It allows for rapid development of interactive systems based on tabletops and interactive tiled-display walls. It serves as a testing environment for a development of new interaction approaches for large-scale interactive systems. The framework provides different interaction possibilities of multimodal user interfaces which are made of multiple low-cost commodity devices (e.g., single- and multi-touch overlay panels and foils, web cameras, depth sensors). Main features of the framework: a) coupling of multiple low-cost commodity multi-touch sensors which are represented as a single seamless interface; b) functions for distinguishing and continuous association of users and touch operations they performed; c) hand-tracking for identification and distinguishing users' body parts (e.g., hands) used in association with touch input events; d) integration of gesture recognition algorithms Protractor, $1 and $N to provide experimental gesture recognition of touch events (libreco library) e) extended semantic description of input events enabling their personalisation (e.g., personalized gesture recognition); (libdtuio library) f) highly configurable toolkit and application base. Supported HW: majority of existing resistive and capacitive single- and multi-touch sensor overlay panels and devices, MS Kinect depth tracker.
PrintDisplayed: 15/7/2024 13:21