Rules extraction from neural networks trained on multimedia data
The result's identifiers
Result code in IS VaVaI
<a href="https://www.isvavai.cz/riv?ss=detail&h=RIV%2F67985807%3A_____%2F19%3A00512089" target="_blank" >RIV/67985807:_____/19:00512089 - isvavai.cz</a>
Alternative codes found
RIV/68407700:21240/19:00334955 RIV/68407700:21340/19:00334955
Result on the web
<a href="http://ceur-ws.org/Vol-2473/paper4.pdf" target="_blank" >http://ceur-ws.org/Vol-2473/paper4.pdf</a>
DOI - Digital Object Identifier
—
Alternative languages
Result language
angličtina
Original language name
Rules extraction from neural networks trained on multimedia data
Original language description
Since the universal approximation property of artificial neural networks was discovered in the late 1980s, i.e., their capability to arbitrarily well approximate nearly arbitrary relationships and dependences, a full exploitation of this property has been always hindered by the very low human-comprehensibility of the purely numerical representation that neural networks use for such relationships and dependences. The mainstream of attempts to mitigate that incomprehensibility are methods extracting, from the numerical representation, rules of some formal logic, which are in general viewed as human-comprehensible. Many dozens of such methods have already been proposed since the 1980s, differing in a number of diverse aspects. Due to that diversity, and also due to a close connection of the semantics of extracted rules to the repsective application domain, no rules extraction methods have ever become a standard, and it is always necessary to select a suitable method for the considered domain. Here, rules extraction from trained neural networks is employed for multimedia data, which is an increasingly important but also increasingly complex kind of data. Three particular rules extraction methods are considered and applied to the modalities recognized text data and the speech acoustic data, both of them with different subsets of features. A detailed comparison of the performance of the considered methods on those datasets is presented, and a statistical analysis of the obtained results is performed.n
Czech name
—
Czech description
—
Classification
Type
D - Article in proceedings
CEP classification
—
OECD FORD branch
10201 - Computer sciences, information science, bioinformathics (hardware development to be 2.2, social aspect to be 5.8)
Result continuities
Project
<a href="/en/project/GA18-18080S" target="_blank" >GA18-18080S: Fusion-Based Knowledge Discovery in Human Activity Data</a><br>
Continuities
I - Institucionalni podpora na dlouhodoby koncepcni rozvoj vyzkumne organizace
Others
Publication year
2019
Confidentiality
S - Úplné a pravdivé údaje o projektu nepodléhají ochraně podle zvláštních právních předpisů
Data specific for result type
Article name in the collection
ITAT 2019: Information Technologies – Applications and Theory
ISBN
—
ISSN
1613-0073
e-ISSN
—
Number of pages
10
Pages from-to
26-35
Publisher name
Technical University & CreateSpace Independent Publishing
Place of publication
Aachen
Event location
Donovaly
Event date
Sep 20, 2019
Type of event by nationality
EUR - Evropská akce
UT code for WoS article
—