DDMarkerFAST: DDMarker FeAture SelecTion

Description Arguments Details Value Author(s) References See Also Examples

Description

The Feature Selection Method of Diagnose and Detect Markers in Extracellular Circulating

results = DDMarkerFAST(data, ...);

Arguments

data

Data matrix, the first column must contain the feature types, eg. BLOOD, URINE, ... , and the following ones should be the features just like the references [1] and [2] did.

method

-method should be ( "svm", "adaboost", "bayes", "cart", "c5" ), "svm" denotes Support Vector Machines, the most stable method in biomarkers diagnosis, "adaboost" denotes AdaBoost, also known as the the most efficiency and accuracy method, "bayes" denotes Naive Bayes Cluster, "cart" denotes Recursive Partitioning and Regression Trees, "c5" denotes Decision Trees.
default: "svm"

Details

models = DDMarkerFAST(data = data, method = "svm");

Value

The R function, DDMarkerFAST returns an object of list:

class shows the method used.
names shows the model attributes, which can be used in the prediction method, the more details will be found in references.
vardep.summary, xlevels, ylevels adaboost and cart will return the elements, the more details will be found in references.

Author(s)

Yu Shang (JLU & UGA) yushang@uga.edu
Qiong Yu (JLU & UGA) yuqiong@uga.edu yujoan_2001@163.com
Huansheng Cao (UGA) hshcao@uga.edu
Guoqing Liu (IMUST & UGA) gqliu@uga.edu gqliu1010@163.com
Xiufeng Liu (GZUCM & UGA) xfliu@uga.edu liu_xf@gzucm.edu.cn
Hao Wu (BIT & UGA) wuhao@uga.edu wuhao@bit.edu.cn
Yan Wang (JLU & UGA) wy6868@hotmail.com
Ying Xu (JLU & UGA) xyn@uga.edu xyn@bmb.uga.edu

Maintainer: Yu Shang (JLU & UGA) yushang@uga.edu

References

citation("DDMarkerFAST");
[1] Juan Cui, et al. (2011) An integrated transcriptomic and computational analysis for biomarker identification in gastric cancer. Nucleic Acids Research, 39: 1197-1207
[2] Juan Cui, et al. (2008) Computational prediction of human proteins that can be secreted into the bloodstream. BIOINFORMATICS, Vol.24 no. 20 2008 pages 2370-2375
[3] http://bioinfosrv1.bmb.uga.edu/DMarker/
[4] Breiman L. (1999) Prediction games and arcing classifiers. Neural Comput 11(7):1493:1517
[5] Breiman L, et al. (1984) Classification and regression trees. Wadsworth, Belmont
[6] CheungDW, et al. (1996) Maintenance of discovered association rules in large databases: an incremental updating technique. In: Proceedings of the ACM SIGMOD international conference on management of data, pp. 13:23
[7] Dietterich TG. (1997) Machine learning: Four current directions. AI Mag 18(4):97:136
[8] Domingos P. (1999) MetaCost: A general method for making classifiers cost-sensitive. In: Proceedings of the fifth international conference on knowledge discovery and data mining, pp 155:164
[9] Domingos P, et al. (1997) On the optimality of the simple Bayesian classifier under zero-one loss. Mach Learn 29:103:130
[10] Fix E, et al. (1951) Discriminatory analysis, nonparametric discrimination. USAF School of Aviation Medicine, Randolph Field, Tex., Project 21-49-004, Rept. 4, Contract AF41(128)-31, February 1951
[11] Freund Y, et al. (1997) A decision-theoretic generalization of on-line learning and an application to boosting. J Comput Syst Sci 55(1):119:139
[12] Friedman JH, et al. (1977) An algorithm for finding best matches in logarithmic time. ACMTrans.Math. Software 3, 209. Also available as Stanford Linear Accelerator Center Rep. SIX-PUB- 1549, February 1975
[13] Friedman JH, et al. (1996) Lazy decision trees. In: Proceedings of the thirteenth national conference on artificial intelligence, San Francisco, CA. AAAI Press/MIT Press, pp. 717:724
[14] Friedman N, et al. (1997) Bayesian network classifiers. Mach Learn 29:131:163
[15] Hand DJ, et al. (2001) Idiot Bayes not so stupid after all. Int Stat Rev 69:385:398
[16] Friedman J, et al. (2000) Additive logistic regression: a statistical view of boosting with discussions. Ann Stat 28(2):337:407
[17] Herbrich R, et al. (2000) Rank boundaries for ordinal regression. Adv Mar Classif pp 115:132
[18] Hunt EB, et al. (1966) Experiments in induction. Academic Press, New York
[19] Inokuchi A, et al. (2005) General framework for mining frequent subgraphs from labeled graphs. Fundament Inform 66(1,2):53:82
[20] Messenger RC, et al. (1972) A model search technique for predictive nominal scale multivariate analysis. J Am Stat Assoc 67:768:772
[21] Morishita S, et al. (2000) Traversing lattice itemset with statistical metric pruning. In: Proceedings of PODS 00, pp 226:236
[22] Olshen R. (2001) A conversation with Leo Breiman. Stat Sci 16(2):184:198
[23] Quinlan JR. (1979) Discovering rules by induction from large collections of examples. In: Michie D (ed), Expert systems in the micro electronic age. Edinburgh University Press, Edinburgh
[24] Quinlan R. (1989) Unknown attribute values in induction. In: Proceedings of the sixth international workshop on machine learning, pp. 164:168
[25] Quinlan JR. (1993) C4.5: Programs for machine learning. Morgan Kaufmann Publishers, San Mateo
[26] Reyzin L, et al. (2006) How boosting the margin can also boost classifier complexity. In: Proceedings of the 23rd international conference on machine learning.

See Also

DDMarkerFAST-package DDMarkerP-method DDMarker NAR

Examples

1
models = DDMarkerFAST();

yu-shang/DDMarkerFAST documentation built on May 4, 2019, 5:34 p.m.