Package rknn. June 9, 2015

Similar documents
Package rpst. June 6, 2017

Package sprinter. February 20, 2015

Package FWDselect. December 19, 2015

Package MixSim. April 29, 2017

Package freeknotsplines

Package sspline. R topics documented: February 20, 2015

Package listdtr. November 29, 2016

Package RAMP. May 25, 2017

Package cwm. R topics documented: February 19, 2015

Package optimus. March 24, 2017

Package cumseg. February 19, 2015

Package clustering.sc.dp

Package flam. April 6, 2018

Package mixphm. July 23, 2015

Package svmpath. R topics documented: August 30, Title The SVM Path Algorithm Date Version Author Trevor Hastie

Package randomforest.ddr

Package RCA. R topics documented: February 29, 2016

Package gains. September 12, 2017

Package datasets.load

Package texteffect. November 27, 2017

Package DPBBM. September 29, 2016

Package GLDreg. February 28, 2017

The simpleboot Package

Package mlegp. April 15, 2018

Package munfold. R topics documented: February 8, Type Package. Title Metric Unfolding. Version Date Author Martin Elff

Package Rramas. November 25, 2017

Package subsemble. February 20, 2015

Package cgh. R topics documented: February 19, 2015

Package obliquerf. February 20, Index 8. Extract variable importance measure

Package uclaboot. June 18, 2003

Package glinternet. June 15, 2018

Package longclust. March 18, 2018

Package ridge. R topics documented: February 15, Title Ridge Regression with automatic selection of the penalty parameter. Version 2.

Package SAM. March 12, 2019

Package nplr. August 1, 2015

The supclust Package

Package gppm. July 5, 2018

Package maboost. R topics documented: February 20, 2015

Package RNAseqNet. May 16, 2017

Package crmn. February 19, 2015

Package glmnetutils. August 1, 2017

Package Daim. February 15, 2013

Package kernelfactory

Package manet. September 19, 2017

Package clusterrepro

Package DTRlearn. April 6, 2018

Package FastKNN. February 19, 2015

Package flexcwm. May 20, 2018

Package ICEbox. July 13, 2017

Package replicatedpp2w

Package TANDEM. R topics documented: June 15, Type Package

Package clustmixtype

Package msda. February 20, 2015

Package Modeler. May 18, 2018

2 cubist.default. Fit a Cubist model

Package SSRA. August 22, 2016

Package SmoothHazard

Package OLScurve. August 29, 2016

Package EnsembleBase

Package wskm. July 8, 2015

Package Mondrian. R topics documented: March 4, Type Package

Package isopam. February 20, 2015

Package missforest. February 15, 2013

Package ClustGeo. R topics documented: July 14, Type Package

Package biosigner. March 6, 2019

Package SAENET. June 4, 2015

Package sgpls. R topics documented: August 29, Type Package Title Sparse Group Partial Least Square Methods Version 1.

Package swcrtdesign. February 12, 2018

Package areaplot. October 18, 2017

Package CALIBERrfimpute

Package rfsa. R topics documented: January 10, Type Package

Package CuCubes. December 9, 2016

Package Numero. November 24, 2018

Package NNLM. June 18, 2018

Package lga. R topics documented: February 20, 2015

Package ipft. January 4, 2018

Package GAMBoost. February 19, 2015

Package lspls. July 27, 2018

Package ICSOutlier. February 3, 2018

Package caretensemble

Package madsim. December 7, 2016

Package abe. October 30, 2017

Package EFS. R topics documented:

Package cutoffr. August 29, 2016

Nonparametric Classification Methods

Package restlos. June 18, 2013

Package ConvergenceClubs

Package ranger. November 10, 2015

Package svapls. February 20, 2015

Package qvcalc. R topics documented: September 19, 2017

Package simex. R topics documented: September 7, Type Package Version 1.7 Date Imports stats, graphics Suggests mgcv, nlme, MASS

Package mlvar. August 26, 2018

Package alphashape3d

Package Cubist. December 2, 2017

Package edrgraphicaltools

Package SSLASSO. August 28, 2018

Package flsa. February 19, 2015

Package FisherEM. February 19, 2015

Package PropClust. September 15, 2018

Package birch. February 15, 2013

Transcription:

Type Package Title Random KNN Classification and Regression Version 1.2-1 Date 2015-06-07 Package rknn June 9, 2015 Author Shengqiao Li <lishengqiao@yahoo.com> Maintainer Shengqiao Li <lishengqiao@yahoo.com> Depends R (>= 2.14), gmp (>= 0.5-5) Suggests Hmisc, Biobase, genefilter, golubesets, chemometrics Random knn classification and regression are implemented. Random knn based feature selection methods are also included. The approaches are mainly developed for highdimensional data with small sample size. License GPL (>= 2) LazyLoad yes NeedsCompilation yes Repository CRAN Date/Publication 2015-06-09 00:14:51 R topics documented: rknn-package........................................ 2 bestset............................................ 3 confusion.......................................... 4 cv.coef............................................ 5 eta.............................................. 5 fitted............................................. 6 lambda............................................ 7 normalize.......................................... 7 plot backward elimination................................. 8 plot rknn support...................................... 8 predicted.......................................... 9 PRESS............................................ 10 1

2 rknn-package print.rknn.......................................... 10 print.rknnbe........................................ 11 print.rknnsupport...................................... 11 r............................................... 12 rknn............................................. 13 rknnbeg........................................... 14 rknnsupport......................................... 15 rsqp............................................. 16 varused........................................... 16 Index 17 rknn-package Random KNN Classification and Regression Random KNN Classification and Regression Details Package: rknn Type: Package Version: 1.1 Date: 2013-08-05 Depends: R (>= 2.15.0), gmp Suggests: Hmisc, Biobase, genefilter, golubesets, chemometrics Imports: class, FNN License: GPL (>=2) LazyLoad: yes Packaged: 2013-08-5 Index: PRESS begknn bestset confusion cv.coef eta fitted.randomknn knn.reg knn.reg.cv lambda Predicted Residual Sum of Squares Backward Elimination Feature Selection with Random KNN Extract the best subset of feature from selection process Classification Confusion Matrix and Accuracy Coefficient of Variation Coverage Probability Extract Model Fitted s KNN Regression KNN Regression Cross-Validation Compute Number of Silent Features

bestset 3 normalize.decscale plot.begknn plot.supportrknn predicted print.knnregcv print.beknn print.randomknn print.supportrknn r randomknn rknn-package rsqp supportrknn varused Data Normalization Plot Function for Recursive Backward Elimination Feature Selection Plot Function for Support Criterion Prediced From a Linear Model Print Method for KNN Regression Cross-validation Print Method for Recursive Backward Elimination Feature Selection Print method for Random KNN regression cross-validation Print Method for Random KNN Support Criterion Choose number of KNNs Random KNN Classification and Regression Random KNN Classification and Regression Predicted R-square Support Criterion Features Used or Not Used in Random KNN Shengqiao Li Maintainer: Shengqiao Li <lishengqiao@yahoo.com> References Shengqiao Li, E James Harner and Donald A Adjeroh. Random KNN feature selection - a fast and stable alternative to Random Forests. BMC Bioinformatics 2011, 12:450. http://www. biomedcentral.com/1471-2105/12/450 bestset Extract the Best Subset of Feature from Selection Process Extract the best subset of feature from selection process. bestset(x, criterion=c("mean_accuracy", "mean_support")) prebestset(x, criterion=c("mean_accuracy", "mean_support")) x criterion An object returned by rknnbeg or rknnbel. either uses mean_accuracy or mean_support for best.

4 confusion A character vector of feature names. See Also rknnbeg, rknnbel Examples ## Not run: bestset{x} confusion Classification Confusion Matrix and Accuracy Compute classification confusion matrix and accuracy confusion(obs, pred) confusion2acc(ct) obs pred ct A vector of observed classes. A vector of predicted classes. A table retruned from confusion confusion return a matrix of cross classification counts. confusion return a overall classification accuracy. Examples obs<- rep(0:1, each =5); pre<- c(obs[3:10], obs[1:2]) confusion(obs, pre) confusion2acc( confusion(obs, pre))

cv.coef 5 cv.coef Coefficient of Variation Compute coefficient of variation. cv.coef(x) x A numeric vector. A number within [0, 1]. eta Coverage Probability Calculate the coverage probability eta(p, m, r, method = c("binomial", "poisson", "exact")) p m r method Total number of available features. Number of features to be drawn by each KNN. Number of KNN to be generated. Either binomial approximation, poisson approximaton or exact method. Details exact method needs gmp pakage and is slow for large p.

6 fitted The coverage probability is returned. See Also r Examples eta(1000, 32, 100) fitted Extract Model Fitted s Extract Random KNN fitted values. ## S3 method for class rknn fitted(object,...) object A rknnobject.... Additional arguments. A vector of fitted values. Examples ## Not run: golub.train<- t(scale(golubtrain)); golub.test<- t(scale(golubtest)); golub.rnn<- randomknn(data=golub.train, newdata=golub.test, y=golubtrain.cl, r=821, mtry=55); fitted(golub.rnn) ## End(Not run)

lambda 7 lambda Compute Number of Silent Features Compute number of silent features for Random KNN lambda(p, m, r) p m r Total number of available features. Number of features to be drawn for each KNN. Number of KNN to be generated. A scalar for the mean silent features. normalize Data Normalization Data matrix normalization procedures. normalize.decscale(data) normalize.sigmoidal(data) normalize.softmax(data) normalize.unit(data) data A data matrix to be normalized. A normalized data matrix.

8 plot rknn support plot backward elimination Plot Function for Recursive Backward Elimination Feature Selection Plot the recursive backward elimination feature selection process. ## S3 method for class rknnbeg plot(x, col = "springgreen4", xlab = "no. of features", ylab = "mean accuracy",...) ## S3 method for class rknnbel plot(x, col = "springgreen4", xlab = "no. of features", ylab = "mean accuracy",...) x col xlab ylab An object returned from rknnbel or rknnbeg. Plot line color. Plot x label. Plot x label.... Addional plot parameters. Shengqiao Li Maintainer: plot rknn support Plot Function for Support Criterion Plot support of the important features. ## S3 method for class rknnsupport plot(x, n.var = min(30, length(x$support)), main = deparse(substitute(x)), bg = "gold", lcolor = "blue",...)

predicted 9 x a list with support returned from rknnsupport. n.var number of variables to be displayed. main text for main title. bg background color. lcolor line color.... addtional plot arguments. predicted Prediced From a Linear Model Extraceted predicted values from a linear model. predicted(obj) obj A linear model. A vector of ppredicted values.

10 print.rknn PRESS Predicted Residual Sum of Squares Predicted Residual Sum of Squares PRESS(obj) obj A linear model or knn regression returns predicted residual sum of squares print.rknn Print method for Random KNN Print method for Random KNN ## S3 method for class rknn print(x,...) x A rknn object.... Additonal print arguments.

print.rknnbe 11 print.rknnbe Print Method for Recursive Backward Elimination Feature Selection Print summary of recursive backward elimination feature selection. ## S3 method for class rknnbe print(x,...) x A beknn object returned by rknnbeg or rknnbel.... Addtional arguments to print method. print.rknnsupport Print Method for Random KNN Support Criterion Print support summary of the features. ## S3 method for class rknnsupport print(x,...) x A list returned from rknnsupport or rknnregsupport.... Additonal print arguments.

12 r r Choose number of KNNs Choose number of KNNs r(p, m = floor(sqrt(p)), eta = 0.99, nu = 20, rmax = p, nsim = 1000, lambda = 0.01, method = c("binomial", "poisson", "nu", "geo.independent", "geo.sim", "geo.dependent", "lambda")) p m eta nu rmax nsim lambda method Total number of available features. Number of features to be drawn by each KNN. Coverage Probability. mean mutiplicity of a feature number of series terms for independent geometric approximation number of simulations for geometric simulation. mean number of silient features. one of binomial, poisson, nu, geo.independent, geo.sim, geo.dependent, lambda Details Method binomial and poisson are approximation method for a given eta value. Method nu computes r for a given nu. Method geo.independent, geo.sim and geo.dependent compute r using geomerical random variables z until eachh is at least drawn once. The difference is that geo.independent ignores dependency,geo.sim is a simulation method and geo.independent is an exact method using gmp package. Method lambda computes r for a given lambda. An integer. Examples r(100, 10);

rknn 13 rknn Random KNN Classification and Regression Random KNN Classification and Regression rknn(data, newdata, y, k = 1, r = 500, mtry = trunc(sqrt(ncol(data))), cluster = NULL, seed = NULL) rknn.cv(data, y, k = 1, r = 500, mtry=trunc(sqrt(ncol(data))), cluster=null, seed = NULL) rknnreg(data, newdata, y, k=1, r=500, mtry=trunc(sqrt(ncol(data))), cluster=null, seed=null) data newdata y k r mtry cluster seed A training dataset. A testing dataset. A vector of responses. Number of nearest neighbors. Number of KNNs. Number of features to be drawn for each KNN. An object of class c("sockcluster", "cluster") An integer seed. Return a RandomKNN object.

14 rknnbeg rknnbeg Backward Elimination Feature Selection with Random KNN Recursive Backward Elimination Feature Selection with Random KNN rknnbeg(data, y, k = 1, r = 500, mtry = trunc(sqrt(ncol(data))), fixed.partition = FALSE, pk = 0.5, stopat = 4, cluster=null, seed = NULL) rknnbel(data, y, k = 1, r = 500, mtry = trunc(sqrt(ncol(data))), fixed.partition = FALSE, d = 1, stopat = 4, cluster=null, seed = NULL) data y k r mtry An n x p numeric design matrix. A vector of responses. For a numeric vector, Random Knn regression is performed. For a factor, Random classification is performed. An integer for the number of nearest neighbors. An integer for the number of base KNN models. Number of features to be drawn for each KNN. fixed.partition Logical. Use fixed partition of dynamic partition of the data into training and testing subsets for each KNN. pk d stopat cluster seed A real number between 0 and to indicate the proportion of the feature set to be kept in each step. A integer to indicate the number of features to be dropped in each step. an integer for the minimum number of variables. An object of class c("sockcluster", "cluster") An integer seed.

rknnsupport 15 rknnsupport Support Criterion Compute support criterion using Random KNN classification or regression rknnsupport(data, y, k = 1, r = 500, mtry = trunc(sqrt(ncol(data))), fixed.partition = FALSE, cluster=null, seed = NULL) rknnregsupport(data, y, k = k, r = 500, mtry = trunc(sqrt(ncol(data))), fixed.partition = FALSE, cluster=null, seed = NULL) data y k r The input dataset. A vector of responses. Number of nearest neighbors. Number of KNNs. mtry Number of features to be drawn for each KNN. fixed.partition Logical. Use fixed partition of dynamic partition of the data into training and testing subsets for each KNN. cluster seed An object of class c("sockcluster", "cluster") An integer seed. A supportknn object.

16 varused rsqp Predicted R-square Computed predicted R-square rsqp(obj) obj A regression model. returns predicted R-square varused Features Used or Not Used in Random KNN List the features used or not used in Random KNN modeling. varused(x, by.knn = FALSE, count = TRUE) varnotused(x) x by.knn count A randomknn object. Logical. Should list features used in each KNN? Logical. Should list number of times each used. varnotused returns a list features not used. varused returns a list of feartures used.

Index Topic classif confusion, 4 rknnbeg, 14 Topic distribution eta, 5 lambda, 7 r, 12 Topic hplot plot backward elimination, 8 plot rknn support, 8 Topic manip normalize, 7 Topic models fitted, 6 predicted, 9 rknn, 13 Topic multivariate bestset, 3 rknn, 13 rknnbeg, 14 rknnsupport, 15 varused, 16 Topic package rknn-package, 2 Topic print print.rknn, 10 print.rknnbe, 11 print.rknnsupport, 11 Topic regression PRESS, 10 rknnbeg, 14 rsqp, 16 Topic univar cv.coef, 5 eta, 5 fitted, 6 lambda, 7 normalize, 7 plot backward elimination, 8 plot rknn support, 8 plot.rknnbeg (plot backward elimination), 8 plot.rknnbel (plot backward elimination), 8 plot.rknnsupport (plot rknn support), 8 prebestset (bestset), 3 predicted, 9 PRESS, 10 print.rknn, 10 print.rknnbe, 11 print.rknnsupport, 11 r, 6, 12 rknn, 13 rknn-package, 2 rknnbeg, 4, 14 rknnbel, 4 rknnbel (rknnbeg), 14 rknnreg (rknn), 13 rknnregsupport (rknnsupport), 15 rknnsupport, 15 rsqp, 16 varnotused (varused), 16 varused, 16 bestset, 3 confusion, 4 confusion2acc (confusion), 4 cv.coef, 5 17