DDMarkerFAST-package: DDMarkerFAST

Description Details Author(s) References See Also

Description

Diagnose and Detect Markers in Extracellular Circulating is a homo sapiens deductive system solving the markers in extracellular circulating. It entails the symbols of markers, like the genes, the proteins, the micro RNAs, and the isoforms, whether can be diagnose and detect in extracellular circulating, especially the blood serum and the urine for the biological and medicine significance. With the help of a homo sapiens annotation database in DDMarkerData package, DDMarker can even diagnose and detect the sequence among the genes, the proteins, the micro RNAs, and the isoforms. There are two main function in this package, the ddmarker, and the DDMarkerMMC, short for Minimal Metabolize Circulation. DDMarkerMMC entails the markers among the minimal metabolize circulation.

FAST package is the FeAture SelecTion method of DDMarker, the main function is DDMarkerFAST() and DDMarkerP(). The more details you can find in DDMarkerFAST-method and DDMarkerP-method

Details

Package: DDMarkerFAST
Type: Package
Version: 1.0
Date: 2016-07-12
Depends: R (>= 3.0.3), e1071, adabag, rpart, C50
License: GPL (>= 2)
LazyLoad: yes
LazyData: true

Author(s)

Yu Shang (JLU & UGA) [email protected]
Qiong Yu (JLU & UGA) [email protected] [email protected]
Huansheng Cao (UGA) [email protected]
Guoqing Liu (IMUST & UGA) [email protected] [email protected]
Xiufeng Liu (GZUCM & UGA) [email protected] [email protected]
Hao Wu (BIT & UGA) [email protected] [email protected]
Yan Wang (JLU & UGA) [email protected]
Ying Xu (JLU & UGA) [email protected] [email protected]

Maintainer: Yu Shang (JLU & UGA) [email protected]

References

citation("DDMarkerFAST");
[1] Juan Cui, et al. (2011) An integrated transcriptomic and computational analysis for biomarker identification in gastric cancer. Nucleic Acids Research, 39: 1197-1207
[2] Juan Cui, et al. (2008) Computational prediction of human proteins that can be secreted into the bloodstream. BIOINFORMATICS, Vol.24 no. 20 2008 pages 2370-2375
[3] http://bioinfosrv1.bmb.uga.edu/DMarker/
[4] Breiman L. (1999) Prediction games and arcing classifiers. Neural Comput 11(7):1493:1517
[5] Breiman L, et al. (1984) Classification and regression trees. Wadsworth, Belmont
[6] CheungDW, et al. (1996) Maintenance of discovered association rules in large databases: an incremental updating technique. In: Proceedings of the ACM SIGMOD international conference on management of data, pp. 13:23
[7] Dietterich TG. (1997) Machine learning: Four current directions. AI Mag 18(4):97:136
[8] Domingos P. (1999) MetaCost: A general method for making classifiers cost-sensitive. In: Proceedings of the fifth international conference on knowledge discovery and data mining, pp 155:164
[9] Domingos P, et al. (1997) On the optimality of the simple Bayesian classifier under zero-one loss. Mach Learn 29:103:130
[10] Fix E, et al. (1951) Discriminatory analysis, nonparametric discrimination. USAF School of Aviation Medicine, Randolph Field, Tex., Project 21-49-004, Rept. 4, Contract AF41(128)-31, February 1951
[11] Freund Y, et al. (1997) A decision-theoretic generalization of on-line learning and an application to boosting. J Comput Syst Sci 55(1):119:139
[12] Friedman JH, et al. (1977) An algorithm for finding best matches in logarithmic time. ACMTrans.Math. Software 3, 209. Also available as Stanford Linear Accelerator Center Rep. SIX-PUB- 1549, February 1975
[13] Friedman JH, et al. (1996) Lazy decision trees. In: Proceedings of the thirteenth national conference on artificial intelligence, San Francisco, CA. AAAI Press/MIT Press, pp. 717:724
[14] Friedman N, et al. (1997) Bayesian network classifiers. Mach Learn 29:131:163
[15] Hand DJ, et al. (2001) Idiot Bayes not so stupid after all. Int Stat Rev 69:385:398
[16] Friedman J, et al. (2000) Additive logistic regression: a statistical view of boosting with discussions. Ann Stat 28(2):337:407
[17] Herbrich R, et al. (2000) Rank boundaries for ordinal regression. Adv Mar Classif pp 115:132
[18] Hunt EB, et al. (1966) Experiments in induction. Academic Press, New York
[19] Inokuchi A, et al. (2005) General framework for mining frequent subgraphs from labeled graphs. Fundament Inform 66(1-2):53:82
[20] Messenger RC, et al. (1972) A model search technique for predictive nominal scale multivariate analysis. J Am Stat Assoc 67:768:772
[21] Morishita S, et al. (2000) Traversing lattice itemset with statistical metric pruning. In: Proceedings of PODS 00, pp 226:236
[22] Olshen R. (2001) A conversation with Leo Breiman. Stat Sci 16(2):184:198
[23] Quinlan JR. (1979) Discovering rules by induction from large collections of examples. In: Michie D (ed), Expert systems in the micro electronic age. Edinburgh University Press, Edinburgh
[24] Quinlan R. (1989) Unknown attribute values in induction. In: Proceedings of the sixth international workshop on machine learning, pp. 164:168
[25] Quinlan JR. (1993) C4.5: Programs for machine learning. Morgan Kaufmann Publishers, San Mateo
[26] Reyzin L, et al. (2006) How boosting the margin can also boost classifier complexity. In: Proceedings of the 23rd international conference on machine learning.

See Also

DDMarkerFAST-method DDMarkerP-method DDMarker NAR


yu-shang/DDMarkerFAST documentation built on May 4, 2019, 5:34 p.m.