Mostra i principali dati dell'item

dc.contributor.authorCorchado Rodríguez, Juan Manuel 
dc.contributor.authorMacDonald, Donald
dc.contributor.authorFyfe, Colin
dc.date.accessioned2017-09-06T09:16:40Z
dc.date.available2017-09-06T09:16:40Z
dc.date.issued2002
dc.identifier.citationNeural Networks for Signal Processing, 2002. Proceedings of the 2002 12th IEEE Workshop on. pp. 327 - 336 .
dc.identifier.isbn0-7803-7616-1 (Print)
dc.identifier.urihttp://hdl.handle.net/10366/135113
dc.description.abstractWe investigate an extension of Hebbian learning in a principal component analysis network which has been derived to be optimal for a specific probability density function(PDF). We note that this probability density function is one of a family of PDFs and investigate the learning rules formed in order to be optimal for several members of this family. We show that, whereas previous authors have viewed the single member of the family as an extension of PCA, it is more appropriate to view the whole family of learning rules as methods of performing exploratory projection pursuit (EPP). We explore the performance of our method first in response to an artificial data type, then to a real data set.
dc.format.mimetypeapplication/pdf
dc.language.isoen
dc.publisherIEEE
dc.rightsAttribution-NonCommercial-NoDerivs 3.0 Unported
dc.rights.urihttps://creativecommons.org/licenses/by-nc-nd/3.0/
dc.subjectComputer Science
dc.titleClassification and ICA using maximum likelihood Hebbian learning
dc.typeinfo:eu-repo/semantics/conferenceObject
dc.rights.accessRightsinfo:eu-repo/semantics/openAccess


Files in questo item

Thumbnail

Questo item appare nelle seguenti collezioni

Mostra i principali dati dell'item

Attribution-NonCommercial-NoDerivs 3.0 Unported
Excepto si se señala otra cosa, la licencia del ítem se describe como Attribution-NonCommercial-NoDerivs 3.0 Unported