Escolar Documentos
Profissional Documentos
Cultura Documentos
VERITAS
Virtual and Augmented Environments and Realistic User
Interactions To achieve Embedded Accessibility DesignS
247765
Status F (Final)
3 Draft version created and sent for peer review on May 2012.
Table of Contents
Executive Summary....................................................................... 10
1 Introduction ............................................................................. 12
2 Architecture of the Multimodal Interfaces ................................. 13
2.1 Positioning in the VERITAS Global Architecture ................................. 13
2.2 Local Architecture Analysis ................................................................. 16
3 Methodology ............................................................................ 20
3.1 Adaptation Stage ................................................................................. 21
3.1.1 Assignment of Multimodal Interfaces to the adapted environment ......... 22
3.1.2 Analysis of the Simulation Model based on the Multimodal Interfaces
Models 22
3.1.3 Definition of the alternative tasks ........................................................... 22
3.2 Runtime Stage .................................................................................... 23
3.2.1 Analysis of the Virtual User Models disability model.............................. 23
3.2.2 Modality Compensation and Replacement process ............................... 27
4 Integration Status of the Multimodal Interfaces Architecture
Components .................................................................................. 31
4.1 Multimodal Interfaces Manager ........................................................... 31
4.1.1 Functional Description ........................................................................... 31
4.1.2 Data Flow .............................................................................................. 32
4.1.3 Integration Status .................................................................................. 33
4.1.4 Development Status .............................................................................. 35
4.2 Speech Recognition Module ............................................................... 35
4.2.1 Functional Description ........................................................................... 35
4.2.2 Data Flow .............................................................................................. 35
4.2.3 Integration Status .................................................................................. 36
4.2.4 Development Status .............................................................................. 37
4.3 Speech Synthesis Module ................................................................... 37
4.3.1 Functional Description ........................................................................... 37
4.3.2 Data Flow .............................................................................................. 38
4.3.3 Integration Status .................................................................................. 38
4.3.4 Development Status .............................................................................. 39
4.4 Haptics Module ................................................................................... 39
4.4.1 Functional Description ........................................................................... 39
4.4.2 Data Flow .............................................................................................. 40
June 2012 iv CERTH/ITI
VERITAS D2.8.2 PU Grant Agreement # 247765
List of Figures
Figure 1: The global architecture of Veritas. The box with the red dashed line
depicts the WP2.8 products, i.e. the Multimodal Interfaces Models and the
Multimodal Interfaces Manager (VerMIM). ....................................................... 14
Figure 2: The local architecture of the Multimodal Interfaces. .......................... 17
Figure 3: Portion of a VUM file containing the <disabilityModel> section;
The depicted disability model describes a user with two disabilities:
Gonarthrosis and Rheumatoid Arthritis. ........................................................... 18
Figure 4: In this figure the Virtual User Model information that is valuable to the
VerMIM component is depicted; the disabilitys ICF and its affected task
information are used in the modality compensation step. ................................. 27
Figure 5: Data flow of the modality compensation process for the automatic
simulations, i.e. for simulations that a real user is not present. ........................ 29
Figure 6: Data flow of the Modality compensation process in the immersive and
interactive simulations. ..................................................................................... 30
Figure 7: Data flow and implementation diagram of the Multimodal Interfaces
Manager. .......................................................................................................... 33
Figure 8: Data flow of the Speech Recognition module. ................................... 36
Figure 9: Data flow of the Speech Synthesis module. ...................................... 38
Figure 10: Data flow of the Speech Synthesis module. ................................... 40
Figure 11: Data flow of the Sign Language Synthesis module. ........................ 42
Figure 12: Data flow of the Symbolic Module. .................................................. 43
Figure 13: Data flow of the Screen Reader module. ........................................ 46
Figure 14: Data flow of the Screen Magnifier module. ...................................... 48
List of Tables
Table 1: Table that represents which tools are connected with the VerMIM and
what features are provided to the tools after the integration process. .............. 16
Table 2: The two stages of adding and using the Multimodal Interfaces in the
VERITAS applications. ..................................................................................... 21
Table 3: The Multimodal Interfaces adaptation process steps.......................... 22
Table 4: ICF codes, their descriptions and the affected user modalities. ......... 26
Table 5: The Multimodal Interfaces Components. ............................................ 31
Table 6: Integration status of the Multimodal Interfaces Manager. ................... 34
Table 7: Development status of the Multimodal Interfaces Manager. ............... 35
Table 8: Integration status of the Speech Recognition Module. ....................... 37
Table 9: Development status of the Speech Recognition Module. ................... 37
Table 10: Integration status of the Speech Synthesis Module. ......................... 39
List of Abbreviations
Abbreviation Explanation
OSG OpenSceneGraph
PoI Point of Interest
RoM Range of Motion
RRT Rapidly-exploring Random Tree
SP Sub-project
WP Work-package
VR Virtual Reality
Executive Summary
The D2.8.2 manuscript describes the integration process of the VERITAS
WP2.8 Multimodal Interfaces tools with the Simulation Framework and more
precisely, with the tools that have been implemented as part of the conducted
work of WP2.1.
The main component developed as part of WP2.8 is the VERITAS Multimodal
Interfaces Manager, or shortly VerMIM. Its primary objective is to provide a set
of Multimodal Interfaces tools and methods for testing, during all the
development stages, the accessibility of multimodal interface solutions
addressed to users with special needs, such as older adults and people with
different types of impairments. VerMIM exports to the simulation framework the
functionality of managing the Mulimodal Interfaces Models (A2.8.2). Moreover, it
contains the modality compensation methods (A2.8.3) that have been created
to improve the interaction efficiency and accessibility for the users with
disabilities.
More specifically, the inclusion of the Multimodal Interfaces toolset into the
VERITAS simulation framework tries to achieve three core objectives: a) to
support the interaction process related to user interfaces of different modalities,
b) to formalize and manage properly the user interaction process via the
multimodal interface models, and c) to provide the functionality of an advanced
modality compensation process based on prioritisation rules to the rest
VERITAS accessibility assessment tools.
The conducted work of VERITAS WP2.8 has been devoted to the development
of tools which will allow the engineers to design multimodal interfaces
specifically devoted to the demands of persons with special needs (older people
and people with disabilities). The generated Multimodal Interface Models are
able to simulate the various steps of the interaction process of mono- and
multimodal interfaces and link them to the sensorial capabilities of the users.
These models have been integrated into the VERITAS Framework and are
used in the Modality Compensation process.
Additionally, various modality interfaces have been developed and connected to
the VerMIM component. The interfaces include speech synthesisers and
recognisers, haptic devices manipulators, sign language synthesisers, alerting
modules, screen readers and screen magnifiers. The inclusion of these
interfaces into the VerMIM has taken into account not only the integration of
these tools to the VERITAS platform, but also to existing simulation platforms.
Other interfaces have been taken from 3rd party open source tools and others
have been developed from scratch. The variety of the different interface
characteristics has been homogenized under one common Multimodal
Interfaces application layer interface. The supported unimodal interfaces may
run simultaneously to produce interfaces of multimodal nature.
The deliverable starts with an introductory section. After that it describes the
Multimodal Interfaces Manager placement inside the global VERITAS
architecture is given. The methodology that has been followed for performing
the Modality Compensation process is then depicted and details concerning the
data flow, the integration status and the development status are provided for
each of the Multimodal Interfaces components. The document continues with
June 2012 10 CERTH/ITI
VERITAS D2.8.2 PU Grant Agreement # 247765
the descriptions of the basic classes of the provided Application Layer Interface;
classes which concern the management of the interfaces and the models are
shortly described with their most important functions. Such descriptions are
valuable to the developers of 3rd party tools that will intent to be connected with
the VerMIM.
Concluding, in their current phase most of the tools are at their final stage and
the integration process in all of them has been almost completed. Extra
functionality may be added to the tools as the integration and optimization
process continues as part of A2.8.5. Additionally, minor changes may have to
be applied to the tools after their validation of their evaluation performance as
part of WP3.6/7.
1 Introduction
The main objective of the D2.8.2 deliverable is to describe the methodology that
has been followed in order to integrate the Mutlimodal Interfaces components
with the VERITAS simulation framework. A bibliographic review of multimodal
interfaces and assistive devices, as well as their addressed user groups, is
omitted from this manuscript, as it is included in the deliverable D2.8.1 [1], the
prequel of this document.
This document contains information about the integration process of the WP2.8
products with the products of the WP2.1 [2]. The integration process resulted
into one core component in which seven interface modules have been pre-
integrated. The core component is the Multimodal Interfaces Manager, or
VerMIM, which is responsible for managing the Multimodal Interfaces Models,
for performing the modality compensation process and for providing access to a
list of seven assistive modules that are implementation of interfaces tools.
The seven interfaces modules are mostly 3rd party tools that when combined
provide to the VERITAS framework the multimodal interfacing layer. The
modules include interfaces based on speech recognition, speech synthesis,
alerting methods, sign language synthesis, haptics support, screen magnifiers
and screen readers. The integration of these interfaces and their management
is crucial for the improvement of the user interactions with the tested product,
especially when they have vision, hearing, motor or cognitive impairments.
The present manuscript is split into six sections, including this one.
In Section 2, the architecture of the Multimodal Interfaces is presented. Initially
the positioning of the VerMIM inside the VERITAS global architecture is given.
Then, an analysis of the local VerMIM and its components is explained.
Section 3 describes the methodology of that is hidden under the modality
compensation process and how the various WP2.8 data are used in the
process. Specifically, this section describes the process in two stages: the
adaptation and the runtime. The first is irrelevant of the Virtual User model and
basically is performed as an adaptation step of the Multimodal Interfaces in the
3D Scene or the GUI interface that is going to be tested. The second step is
automatic and initiates every time a simulation session takes place.
The Multimodal Interfaces developed components are described in Section 4.
For each of the eight components a functional description, data flow, integration
and implementation status is given. The architecture diagrams are also
presented and the data input / output are depicted.
In Section 5, the Multimodal Interface Application Layer Interfaces is described.
The integration of the VerMIM not only targets to the rest VERITAS tools, but
also targets to provide its functionality to 3rd party simulation applications. This
section contains information of the provided classes and their functions.
Finally, in Section 6 conclusions regarding the work performed and the
capabilities of the developed tools are reported and future plans are given.
Figure 1: The global architecture of Veritas. The box with the red dashed line depicts the
WP2.8 products, i.e. the Multimodal Interfaces Models and the Multimodal Interfaces
Manager (VerMIM).
GUI Simulation Viewer (VerSim-GUI): this tool is used for evaluating the
accessibility of a graphical user interface. It supports two modes:
automatic and interactive.
o In the automatic mode the program drives automatically the
mouse pointer based on the parameters of the Virtual User Model
and applies any motor, vision, hearing and cognitive impairments
to the interface. The whole process is driven by a pre-recorded
scenario file captured by the VerSEd-GUI tool. The scenario file is
created by an ideal GUI user, i.e. a user that performs the whole
sequence without any task deviations or mistakes. In this case,
the VerMIM is used to provide alternative task sequences in cases
of scenario failure, based on the interfaces that have been
assigned to the GUI by the VerSEd-GUI tool.
o In the immersive mode the VerSim-GUI takes as input the actual
mouse movements of a real user and applies motor, visual,
hearing and cognitive filtering to the users actions. The VerMIM
can be used in such cases to provide alternative interfacing
capabilities to the user, such as GUI navigation via speech
recognition.
Interaction Manager (VerIM): this VERITAS component is used to drive
the VR Interaction Tools from WP2.7. It is used in combination with the
3D Immersive Simulation Viewer (IVerSim-3D) for simulation sessions in
which an immersed user interacts with the VR environment. In this case,
VerMIM can provide multimodal interfacing tools, such as speech
synthesis, recognition, haptics functionality, etc, to the VR environment.
After the integration process of the VerMIM with the above-mentioned tools
various feature will be added to them. Table 1 summarises the added features.
More about the description and nature of the added features can be read in
Section 3.
Assignment of alternative
interfaces to the 3D scene objects.
Simulation adaptation
VerSEd-3D of a 3D scene/product. Alternative task sequences
definition (based on the Multimodal
Interfaces Models).
Modality compensation and
replacement process.
Automatic simulation in
VerSim-3D Usage of alternative interfaces or
3D environments
alternative task-sequences when
the avatar fails in a task.
VerSEd-GUI Simulation adaptation Assignment of alternative
Immersive
Core Simulation VerSEd-3D/GUI IVerSim-3D
Simulation
Platform VerSim-3D/GUI VerIM
Platform
Speech Recognition
Module
Virtual
User Haptics
Model Module
(UsiXML)
Speech Synthesis
Module
Simulation Multimodal Interfaces
Model Manager (VerMIM)
(UsiXML) Sign Language
Synthesis Module
Symbolic Module
Multimodal
Interfaces
Models
Repository Screen Reader
Module
Screen Magnifier
Module
Haptics module: used for converting the users motor sensory functions
into tactile feedback. Used in cases of sever vision impaired users.
Speech synthesis module: produces speech from text. Useful for users
with vision impairments.
Sign language synthesis module: generates animation sequences of a
humanoid moving its hands. Useful for the hearing impaired users.
Symbolic module: generates multimodal alerts, i.e. sounds, visual
signals. Useful for users with cognitive, perceptual problems.
Screen reader module: reads the text of various windows applications
using speech synthesis.
Screen magnifier module: magnifies a portion of the screens desktop.
Figure 3: Portion of a VUM file containing the <disabilityModel> section; The depicted
disability model describes a user with two disabilities: Gonarthrosis and Rheumatoid
Arthritis.
Finally, depending on the simulation type (Table 1), VerMIM is connected with a
VERITAS tool and adds extra functionality to it.
3 Methodology
This section describes the methodology of adding the Multimodal Interfaces
functionality to the rest VERITAS tools. The methodology steps are split into two
main stages:
Adaptation stage: this stage contains the users assignment of
alternative interfaces to the simulated environment objects and
components, either 3D scene or GUI, e.g. the assignment of a speech
recognition interface to a virtual telephone device. Moreover this stage
requires the definition of the alternative tasks contained in the Multimodal
Interfaces Models.
Runtime stage: this stage contains the analysis of the Virtual User Model
file and identification of the virtual users affected modalities. It also
contains the prioritization process on the selection of the alternative task
sequences. The modality compensation and replacement process also
takes place in this stage.
The two stages along with the tools involved and the actions that are performed
are presented in Table 2.
A typical example of a disability model and the information that is parsed from it
is depicted in Figure 4.
Figure 4: In this figure the Virtual User Model information that is valuable to the VerMIM
component is depicted; the disabilitys ICF and its affected task information are used in
the modality compensation step.
Then the VerMIM reads the Simulation model and via the usage of the
Multimodal Interfaces Models creates a list of task sequences. Multiple task
sequences are created based on the derived combinations of the available
Multimodal Interfaces Models. Each task sequence has its tasks annotated with
modalities that are necessary for it to be performed successfully.
Finally, the VerMIM performs the Modality Compensation process by sorting the
generated task sequences. The prioritisation that takes places takes into
account the fact that the sequences containing the lesser of the users affected
modalities, as well as affected tasks, are sorted first. Then the list of the
generated sequences is passed to the VerSim-GUI or VerSim-3D tools in order
to perform the simulation sessions.
ICF Affected
Codes Tasks
Available
Interfaces
Map ICF to modalies
(Multimodal Intefaces
Manager)
Multimodal
Interfaces Affected
Models Modalities
(A2.8.2)
TaskSequenceA
TaskSequenceB
TaskSequenceC
...
Modality Compensation
(Multimodal Intefaces
Manager)
Legend Prioritisation
TaskSequenceA 1st alternative
Input
Runtime TaskSequenceB 2nd alternative
File
Action Data TaskSequenceC 3rd alternative
...
Figure 5: Data flow of the modality compensation process for the automatic simulations,
i.e. for simulations that a real user is not present.
ICF Affected
Codes Tasks
Available
Simulation Interfaces
Map ICF to modalies
Model (Multimodal Intefaces
(WP2.2~ Manager)
WP2.6)
Affected
Modalities
Creation of the default Multimodal
task sequence Interfaces
(Core Simulation Platform) Models
(A2.8.2)
Default TaskSequence
Modality Compensation:
No Find an alternative task
Run Task Success?
(Multimodal Intefaces
Manager)
Yes
Yes
Legend
Exit Input
File Runtime
Action Data
Figure 6: Data flow of the Modality compensation process in the immersive and
interactive simulations.
tools that integrate the simulation platforms and connects to them indirectly. The
VerMIM has been implemented as Dynamic Link Library and thus it has to be
integrated into other executable tools to run. The DLL implementation allows the
Multimodal Interfaces Manager to easily be integrated into 3 rd party simulators.
Speech Recognition
Module
Veritas
Tool (EXE):
VerSEd-3D/GUI
VerSim-3D/GUI Haptics
IVerSim-3D Module
Modality
Virtual Compensation Speech Synthesis
User Process Module
Model
(UsiXML)
Multimodal
Sign Language
Interfaces
Synthesis Module
Manager
(VerMIM DLL)
Simulation
Model Symbolic Module
(UsiXML) Module
Screen Reader
Module
Selection of
Multimodal Management
Multimodal
Interfaces of Multimodal
Interfaces
Models Interfaces Screen Magnifier
Models
Repository Module
Figure 7: Data flow and implementation diagram of the Multimodal Interfaces Manager.
Programming C++
language
Filename MultimodalInterfacingTools.dll
Other None
adjustments
Pending None
integration
actions
Development Final
status
Under development
The VerMIM component provides the audio signal to the Speech Recognition
module and the latter converts into text strings. The text is then transferred to
the simulation platform via the VerMIM API (Figure 8). As it is depicted in the
Figure, the module is communicating directly only with the VerMIM component.
Audio
VerMIM signal
Prerecorded
Avatar speech file
Immersive
SimulImmeation
Platform
Core
Simulation
Platform
Programming C++
language
Other None
adjustments
Dependencies The Module requires the CMU Sphinx libraries to run as well
the language model, the acoustic model and a dictionary file of
the language that is interpreted.
Pending None
integration
actions
Development Final
status
Under development
Progress up to The Speech Recognition Module has been fully integrated with
date the Multimodal Interfaces Manager.
Speech
Simulation Text Text
VerMIM Synthesis
Platform
Module
Speech Immersed
Audio User
VERITAS
tools and
components
Other None
adjustments
Pending None
integration
actions
Development Final
status
Under development
Progress up to The Speech Recognition Module has been fully integrated with
date the Multimodal Interfaces Manager.
position,
orientation
Immersed Haptics
User Module
force
feedback
Haptic device
Simulation
VerMIM
Platform
Programming C++
language
other
VERITAS
tools and
components
Other None
adjustments
Dependencies The Chai3D libraries are not required, because they have been
already included in the VerMIM DLL.
Pending None
integration
actions
Development Final
status
Under development
Progress up to The Haptics Module has been fully integrated with the
date Multimodal Interfaces Manager.
Annotated
Text VerMIM text
Programming C++
language
Filename ATLASPlayer.dll
Other None
adjustments
Dependencies OGRE [15] runtime libraries are needed for rendering the 3D
Avatar.
Pending None
integration
actions
Development Final
status
Under development
Progress up to The Sign Language Module has been fully integrated with the
date Multimodal Interfaces Manager.
Text alert
Sound alert
Programming C++
language
Filename None as the Symbolic Module has been created as part of the
VerMIM and is fully integrated with the MultimodalInter-
facingTools.dll library.
Other None
adjustments
Dependencies OpenCV libraries are needed for generating the visual and text
messages.
Pending None
integration
actions
Development Final
status
Under development
Progress up to The Symbolic Module has been fully integrated with the
date Multimodal Interfaces Manager.
Content, mainly
Textual, sent to
standard output
Request
for screen Screen Reader
VerMIM
reading Module
Braille
Data
Synthetic
Speech
Braille
Device
Immersed
User
Programming C++
language
Integration The NVDA screen reader is bundled with the Screen Reader
short Module and it is called as an external application by the
description VerMIM.
Filename ScreenReader.dll
Other None
adjustments
Pending None
integration
actions
Development Final
status
Under development
Progress up to The Screen Reader Module has been fully integrated with the
date Multimodal Interfaces Manager.
The Module works in online mode, being useful for immersive and interactive
simulations. In both cases, an image depicting a portion of the screens desktop
is captured in real time, which is centred to the position of the mouse pointer
over the screens desktop. The Module takes this screen portion as input,
together with a parameter indicating the desired magnification level (e.g. 2x,
3x). The given portion of the screen is magnified according to the given
magnification level. The result, an image depicting the magnified screen portion
is provided as output to the immersed user.
The screen magnification implementation is based on the OpenCV framework
[9].
Screen
Portion
Image
Magnified
Screen Magnifier Screen Immersed
VerMIM Request
Module Portion User
Image
Programming C++
language
Filename ScreenMagnifier.dll
Other None
adjustments
Pending None
integration
actions
Development Final
status
Under development
Progress up to The Screen Magnifier Module has been fully integrated with the
date Multimodal Interfaces Manager.
MInteractionModel MInteractionModel()
Stores information of a The class constructor. Each constructor creates an
Multimodal Interaction empty task tree and stores it in the object.
Model. Functions for
parsing the UsiXML ParseFromUsiFile()
files and analysing their Parses the Multimodal Interfaces Model UsiXML file
task trees are provided. and fills the tree of the object.
GetStrName()
Returns the name of the Model.
GetStrUsi()
Returns the name of the UsiXML file.
GetTaskById()
Returns a task stored in the model using its Id.
Class Function
SetApplicationArea(),
GetApplicationArea()
Sets and gets the application area type. The
application area type can be any of the following:
ApplicationArea_Generic,
ApplicationArea_SmartLivingSpaces,
ApplicationArea_Automotive,
ApplicationArea_Workplace,
ApplicationArea_Infotainment,
ApplicationArea_PersonalHealthcare.
GetRootTask()
Returns the task that resides on the task trees root.
GetNextPossibleNextTasks()
Gets as input a task and returns a list of the next task
candidates.
MInteraction- MInteractionModelManager()
ModelManager
Constructor of the manager. The VerMIM uses only
Class that manages the one manager. The manager is capable of managing all
Multimodal Interfaces the models contained in the repository.
Models. Contains
functions for parsing ParseXmlFile()
and retrieving the Parses an xml index file that contains the list of the
models from the repository models. The index categorises the models
repository. per their application area. After parsing the index each
model is parsed and stored in the manager.
SearchModel()
Searches a model by its name and its application area
and returns a list of matched models.
GetModels()
Returns all the models of a specific application area.
MTask MTask()
Represents the task Constructor of a Multimodal Interfaces Models Task.
element of a Multimodal
Interfaces Model. Each GetId()
model contains a task- Returns the id of the task. The id of a task is unique.
Class Function
GetMeans()
Returns the means that are needed for performing the
task action. Most of the times, the means parameter is
a body part, such hand, finger, etc.
GetObject()
Returns the object that is required for performing the
task action. An example of an object could be the
speech recogniser device.
GetType()
Returns the type of the task. The type can be any of
the following:
Type_Abstraction: used for complex tasks
that do not contain any interactions, but can be
decomposed into primitives tasks. The actions
of these tasks are contained in their children
tasks.
Type_Interaction: used for describing the
task which contain the interfaction.
Class Function
GetParent()
Retrieves the parent task of the task.
SearchTaskById()
Performs recursive search of the task in the tree by
using its id.
GetTaskEnabling()
Returns the next task that is enabled after the current
task has been completed.
GetFirstEnablingChild()
Returns the first child task that is enabled after the
current task has been completed.
AddTaskChoiceBelow(),
GetTaskChoiceBelow(),
GetNumTaskChoicesBelow()
Functions that create (and get) task choices. A task
can lead to several other task choices (alternative
tasks).
MTaskSequence MTaskSequence()
Class that produces The constructors of task sequence.
objects of task
sequences. A task AddTask()
sequence is series of Adds a task to the end of the sequence.
task objects. The
VerMIM analyses a GetTask()
simulation model and
creates several task Get a task a specific task location.
sequence, which GetNumTasks()
actually are all the
possible task Gets the number of tasks contained in the sequence.
combinations.
ICFElement ICFElement()
Used to represent a Constructor of an ICF element.
mapping of an ICF code
to its affected SetCode(), GetCode()
modalities. Each Sets and returns the ICF code of an element.
element contains three
parts: a) the ICF code, SetName(), GetName()
b) the name of the body
function type, and c) the Sets and returns the name of an element.
Class Function
ICFParser ICFParser()
This class provides the Constructor of the class. Accepts as input the filename
functionality of reading of the xml mapping file. The constructor performs the
an xml file which maps parsing of the file in its initialisation.
the ICF codes to their
respective modalities. SearchICFElementByCode()
Searches an ICF element by its code.
Module Module()
Class Module is an The constructor of the Module Class. The constructor is
abstract class that is called only by the derived classes objects constructors.
used for basic
management of an AsHapticsModule()
interface. AsSignLanguageSynthesisModule()
Contains initialisation AsSpeechSynthesisModule()
and termination
routines for the AsSpeechRecognitionModule()
interfaces. AsSymbolicModule()
AsScreenReaderModule()
AsScreenMagnifierModule()
Casts the module object into its derived class. Used
when it is needed to provide the special functionality of
the derived class objects.
GetName()
Returns the name of the Module.
Class Function
GetType()
Returns the type of the Module. It can be one of the
following:
Type_SpeechSynthesis,
Type_SpeechRecognition,
Type_Haptics,
Type_SignLanguageSynthesis,
Type_Symbolic,
Type_ScreenReader,
Type_Symbolic.
SetDataDir()
Set the directory where the Modules specific data reside.
Initialize()
Initialises the module this function must be called
before any other request to the module.
Terminate()
Terminates the module this function must be called
after the usage of a module in order to free any resources
that are occupied by it.
ModuleHaptics SelectDevice()
Class that is used for Selects a haptic device to manipulate.
manipulation of a
haptic device. The GetNumDevices()
class objects that Returns the number of the connected haptic devices.
need to be created
have to be equal to GetPosition()
the number of the
haptic devices Returns the 3D position of the haptic devices end
connected to the effector.
system.
GetRotation()
Return the 3D orientation of the haptic devices end
effector.
ApplyForce()
Sends a request to the device to generate a 3D force.
Class Function
GetForce()
Receives the current force that is generated from the end
effector.
GetLinearVelocity()
Receives the current linear velocity of the end effector.
GetAngularVelocity()
Receives the ends effectors angular velocity.
ModuleSpeechSyn- ModuleSpeechSynthesis()
thesis
Class constructor: receives as input the speech synthesis
Class for managing engine type, which can be one of the following:
the Speech Synthesis
SubType_Festival_v2_1
module. Two speech
synthesis engines are SubType_eSpeak_v1_45_04
supported: Festival
(high quality, but SetVoice()
slow) and eSpeak
Sets the voice of the speech synthesizer.
(fast, but lower
quality). If necessary, SayText()
two objects of this
class, each one Converts the input text into speech.
assigned to each
engine, can be used
simultaneously.
ModuleSpeechRec- ModuleSpeechRecognition()
ognition
Class Constructor.
Class that is used for
manipulating the SetDictionary()
speech recognition Sets the dictionary file. The dictionary contains all the
engine. Supports two words that are recognisable. Decreasing the dictionary
modes of speech size, increases the recognition accuracy.
recognition: a) offline,
i.e. from reading an SetLanguageModel()
audio file a file and b)
via receiving the Sets the language model. The language model contains
users speech from the rules that form the languages syntax.
the microphone.
SetAcousticModel()
Sets the acoustic model, i.e. the transcription of a word to
its phonemes.
Class Function
RecognizeRawDataFromFile()
Takes as input the filename of a raw audio data file and
converts it to text.
RecognizeFromMicrophone()
Receives speech audio from the microphone and
converts into text.
SetSamplesPerSecond()
Sets the sampling frequency of the input audio data.
SetMaximumRecordingDuration()
Sets the sizes of the audio buffers (in sec).
ModuleSignLangu- ModuleSignLanguageSynthesis()
ageSynthesis
Class constructor
Class that is used for
the Sign Language LoadSingleSentence()
Module. Currently Loads an xml file that contains the annotated text to be
implemented as a animated as sign language gestures.
demo application.
Play()
Plays the animated sign language sequence of the
loaded text.
ModuleSymbolic ModuleSymbolic()
Class that is used for Class constructor
managing the
Symbolic module. Wait()
Pauses the module; useful for playing a sequence of
events.
PlaySoundFromFile()
Loads a plays an audio file.
StopAllSounds()
Stop all audio playback.
ShowWindow(), HideWindow(),
SetWindowSize(), SetWindowColor()
Window manipulation functions, for visual alerts.
Class Function
DisplayText()
Displays text on screen.
ModuleScreenRead Start()
er
Initiates the screen reading module. The module parses
Class for the screens text that is under the mouse pointer and
manipulating the converts it into speech.
screen reader.
Stop()
Stops the screen reading.
ModuleScreenMagn Start()
ifier
Initiates the screen magnification process.
Class for
manipulating the SetDimensions()
Screen Magnifier Sets the dimensions, i.e. width and height, of the screen
Module. portion that will be magnified.
SetZoom()
Sets the magnification zoom level.
Stop()
Stops the screen magnification process.
References
[1] Panagiotis Moschonas, Dimitrios Tzovaras, George Ghinea, VERITAS
Deliverable D2.8.1 First prototypes of the multimodal interface tool set,
December 2011.
[5] Carnegie Mellon University (CMU) Sphinx, Open Source Toolkit For
Speech Recognition, http://cmusphinx.sourceforge.net/