Mostrar el registro sencillo del ítem

dc.contributor.authorUllah, Rafi
dc.contributor.authorKhan, Ayaz H.
dc.contributor.authorEmaduddin, S.m.
dc.date.accessioned2020-06-23T11:12:50Z
dc.date.available2020-06-23T11:12:50Z
dc.date.issued2019-08-14
dc.identifier.citationADCAIJ: Advances in Distributed Computing and Artificial Intelligence Journal, 8 (2019)
dc.identifier.issn2255-2863
dc.identifier.urihttp://hdl.handle.net/10366/143315
dc.description.abstractk-Nearest Neighbor (k-NN) is a non-parametric algorithm widely used for the estimation and classification of data points especially when the dataset is distributed in several classes. It is considered to be a lazy machine learning algorithm as most of the computations are done during the testing phase instead of performing this task during the training of data. Hence it is practically inefficient, infeasible and inapplicable while processing huge datasets i.e. Big Data. On the other hand, clustering techniques (unsupervised learning) greatly affect results if you do normalization or standardization techniques, difficult to determine "k" Value. In this paper, some novel techniques are proposed to be used as pre-state mechanism of state-of-the-art k-NN Classification Algorithm. Our proposed mechanism uses unsupervised clustering algorithm on large dataset before applying k-NN algorithm on different clusters that might running on single machine, multiple machines or different nodes of a cluster in distributed environment. Initially dataset, possibly having multi dimensions, is pass through clustering technique (K-Means) at master node or controller to find the number of clusters equal to the number of nodes in distributed systems or number of cores in system, and then each cluster will be assigned to exactly one node or one core and then applies k-NN locally, each core or node in clusters sends their best result and the selector choose best and nearest possible class from all options. We will be using one of the gold standard distributed framework. We believe that our proposed mechanism could be applied on big data. We also believe that the architecture can also be implemented on multi GPUs or FPGA to take flavor of k-NN on large or huge datasets where traditional k-NN is very slow.
dc.format.mimetypeapplication/pdf
dc.language.isoeng
dc.publisherEdiciones Universidad de Salamanca (España)
dc.rightsinfo:eu-repo/semantics/openAccess
dc.subjectComputación
dc.subjectInformótica
dc.subjectComputing
dc.subjectInformation Technology
dc.titleck-NN: A Clustered k-Nearest Neighbours Approach for Large-Scale Classification
dc.typeinfo:eu-repo/semantics/article


Ficheros en el ítem

Thumbnail

Este ítem aparece en la(s) siguiente(s) colección(ones)

Mostrar el registro sencillo del ítem