Description Arguments Details Value Author(s) References See Also Examples
The Feature Selection Method of Diagnose and Detect Markers in Extracellular Circulating
results = DDMarkerFAST(data, ...);
data |
Data matrix, the first column must contain the feature types, eg. |
method |
-method should be ( "svm", "adaboost", "bayes", "cart", "c5" ), "svm" denotes |
models = DDMarkerFAST(data = data, method = "svm");
The R function, DDMarkerFAST
returns an object of list
:
class | shows the method used. |
names | shows the model attributes, which can be used in the prediction method, the more details will be found in references . |
vardep.summary, xlevels, ylevels | adaboost and cart will return the elements, the more details will be found in references . |
Yu Shang (JLU & UGA) yushang@uga.edu
Qiong Yu (JLU & UGA) yuqiong@uga.edu yujoan_2001@163.com
Huansheng Cao (UGA) hshcao@uga.edu
Guoqing Liu (IMUST & UGA) gqliu@uga.edu gqliu1010@163.com
Xiufeng Liu (GZUCM & UGA) xfliu@uga.edu liu_xf@gzucm.edu.cn
Hao Wu (BIT & UGA) wuhao@uga.edu wuhao@bit.edu.cn
Yan Wang (JLU & UGA) wy6868@hotmail.com
Ying Xu (JLU & UGA) xyn@uga.edu xyn@bmb.uga.edu
Maintainer: Yu Shang (JLU & UGA) yushang@uga.edu
citation("DDMarkerFAST");
[1]
Juan Cui, et al. (2011) An integrated transcriptomic and computational analysis for biomarker identification in gastric cancer. Nucleic Acids Research, 39: 1197-1207
[2]
Juan Cui, et al. (2008) Computational prediction of human proteins that can be secreted into the bloodstream. BIOINFORMATICS, Vol.24 no. 20 2008 pages 2370-2375
[3]
http://bioinfosrv1.bmb.uga.edu/DMarker/
[4]
Breiman L. (1999) Prediction games and arcing classifiers. Neural Comput 11(7):1493:1517
[5]
Breiman L, et al. (1984) Classification and regression trees. Wadsworth, Belmont
[6]
CheungDW, et al. (1996) Maintenance of discovered association rules in large databases: an incremental updating technique. In: Proceedings of the ACM SIGMOD international conference on management of data, pp. 13:23
[7]
Dietterich TG. (1997) Machine learning: Four current directions. AI Mag 18(4):97:136
[8]
Domingos P. (1999) MetaCost: A general method for making classifiers cost-sensitive. In: Proceedings of the fifth international conference on knowledge discovery and data mining, pp 155:164
[9]
Domingos P, et al. (1997) On the optimality of the simple Bayesian classifier under zero-one loss. Mach Learn 29:103:130
[10]
Fix E, et al. (1951) Discriminatory analysis, nonparametric discrimination. USAF School of Aviation Medicine, Randolph Field, Tex., Project 21-49-004, Rept. 4, Contract AF41(128)-31, February 1951
[11]
Freund Y, et al. (1997) A decision-theoretic generalization of on-line learning and an application to boosting. J Comput Syst Sci 55(1):119:139
[12]
Friedman JH, et al. (1977) An algorithm for finding best matches in logarithmic time. ACMTrans.Math. Software 3, 209. Also available as Stanford Linear Accelerator Center Rep. SIX-PUB- 1549, February 1975
[13]
Friedman JH, et al. (1996) Lazy decision trees. In: Proceedings of the thirteenth national conference on artificial intelligence, San Francisco, CA. AAAI Press/MIT Press, pp. 717:724
[14]
Friedman N, et al. (1997) Bayesian network classifiers. Mach Learn 29:131:163
[15]
Hand DJ, et al. (2001) Idiot Bayes not so stupid after all. Int Stat Rev 69:385:398
[16]
Friedman J, et al. (2000) Additive logistic regression: a statistical view of boosting with discussions. Ann Stat 28(2):337:407
[17]
Herbrich R, et al. (2000) Rank boundaries for ordinal regression. Adv Mar Classif pp 115:132
[18]
Hunt EB, et al. (1966) Experiments in induction. Academic Press, New York
[19]
Inokuchi A, et al. (2005) General framework for mining frequent subgraphs from labeled graphs. Fundament Inform 66(1,2):53:82
[20]
Messenger RC, et al. (1972) A model search technique for predictive nominal scale multivariate analysis. J Am Stat Assoc 67:768:772
[21]
Morishita S, et al. (2000) Traversing lattice itemset with statistical metric pruning. In: Proceedings of PODS 00, pp 226:236
[22]
Olshen R. (2001) A conversation with Leo Breiman. Stat Sci 16(2):184:198
[23]
Quinlan JR. (1979) Discovering rules by induction from large collections of examples. In: Michie D (ed), Expert systems in the micro electronic age. Edinburgh University Press, Edinburgh
[24]
Quinlan R. (1989) Unknown attribute values in induction. In: Proceedings of the sixth international workshop on machine learning, pp. 164:168
[25]
Quinlan JR. (1993) C4.5: Programs for machine learning. Morgan Kaufmann Publishers, San Mateo
[26]
Reyzin L, et al. (2006) How boosting the margin can also boost classifier complexity. In: Proceedings of the 23rd international conference on machine learning.
DDMarkerFAST-package
DDMarkerP-method
DDMarker
NAR
1 | models = DDMarkerFAST();
|
Add the following code to your website.
For more information on customizing the embed code, read Embedding Snippets.