Abstract
We investigate the potential of artificial neural networks in diagnosing thyroid diseases. The robustness of neural networks with regard to sampling variations is examined using a cross‐validation method. We illustrate the link between neural networks and traditional Bayesian classifiers. Neural networks can provide good estimates of posterior probabilities and hence can have better classification performance than traditional statistical methods such as logistic regression. The neural network models are further shown to be robust to sampling variations. It is demonstrated that for medical diagnosis problems where the data are often highly unbalanced, neural networks can be a promising classification method for practical use.
Similar content being viewed by others
References
M. Bahemuka and H.M. Hodkinson, Screening for hypothyroidism in elderly inpatients, British Medical Journal 2 (1975) 601-603.
B. Cheng and D.M. Titterington, Neural networks: A review from a statistical perspective, Statistical Sciences 9 (1994) 2-54.
A. Ciampi and Y. Lechevallier, Statistical models as building blocks of neural networks, Communications in Statistics: Theory and Methods 26 (1997) 991-1009.
G. Cybenko, Approximation by superpositions of a sigmoidal function, Mathematical Control Signals Systems 2 (1989) 303-314.
R.O. Duda and P. Hart, Pattern Classification and Scene Analysis (Wiley, New York, 1973).
P. Gallinari, S. Thiria, F. Badran and F. Fogelman-Soulie, On the relations between discriminant analysis and multilayer perceptrons, Neural Networks 4 (1991) 349-360.
L.A. Gavin, The diagnostic dilemmas of hyperthyroxinemia and hypothyroxinemia, Advances in Internal Medicine 33 (1988) 185-204.
F.E. Harreli and K.L. Lee, A comparison of the discriminant analysis and logistic regression under multivariate normality, in: Biostatistics: Statistics in Biomedical, Public Health, and Environmental Sciences, ed. P.K. Sen (North-Holland, Amsterdam, 1985).
G.E. Hinton, How neural networks learn from experience, Scientific American (September, 1992) 145-151.
K. Hornik, Approximation capabilities of multilayer feedforward networks, Neural Networks 4 (1991) 251-257.
K. Hornik, M. Stinchcombe and H. White, Multilayer feedforward networks are universal approximators, Neural Networks 2 (1989) 359-366.
M.S. Hung, M.Y. Hu, B.E. Patuwo and M. Shanker, Estimating posterior probabilities in classification problems with neural networks, International Journal of Computational Intelligence and Organizations 1 (1996) 49-60.
C.C. Klimasauskas, Applying neural networks. Part 3: Training a neural network, PC-AI (May/June, 1991) 20-24.
A. Papoulix, Probability, Random Variables, and Stochastic Processes (McGraw-Hill, New York, 1965).
E. Patuwo, M.Y. Hu and M.S. Hung, Two-group classification using neural networks, Decision Science 24 (1993) 825-845.
S.J. Press and S. Wilson, Choosing between logistic regression and discriminant analysis, Journal of American Statistical Association 73 (1978) 699-705.
J. Quinlan, Simplifying decision trees, International Journal of Man-Machine Studies 27 (1987) 221-234.
M.D. Richard and R.P. Lippmann, Neural network classifiers estimate Basyesian a posteriori probabilities, Neural Computation 3 (1991) 461-483.
M. Riedmiller and H. Braun, A direct adaptive method for faster backpropagation learning: The Rprop algorithm, in: Proceedings of the IEEE International Conference on Neural Networks, 1993.
B.D. Ripley, Statistical aspects of neural networks, in: Networks and Chaos—Statistical and Probabilistic Aspects, eds. O.E. Barndorfe-Nielsen, J.L. Jensen and W.S. Kendall (Chapman and Hall, London, 1993) pp. 40-111.
B.D. Ripley, Neural networks and related methods for classification (with discussion), Journal of Royal Statistical Society, Series B 56 (1994) 409-456.
D.E. Rumelhart, G.E. Hinton and R.J. Williams, Learning internal representations by error propagation, in: Parallel Distributed Processing: Explorations in the Microstructure of Cognition, eds. D.E. Rumelhart and J.L. Williams (MIT Press, Cambridge, 1986).
W.S. Sarle, Neural networks and statistical models, in: Proceedings of the 19th Annual SAS Users Group International Conference, 1994.
SAS Institute Inc., SAS/STAT User's Guide, Version 6, 4th ed., Vol. 2 (SAS Institute Inc., Cary, NC, 1990).
M. Schumacher, R. Robner and W. Vach, Neural networks and logistic regression: Part I, Computational Statistics & Data Analysis 21 (1996) 661-682.
P.K. Sharpe, H.E. Solberg, K. Rootwelt and M. Yearworth, Artificial neural networks in diagnosis of thyroid function from vitro laboratory tests, Clinical Chemistry 39 (1993) 2248-2253.
H.C. Sox, M.A. Blatt, M.C. Higgins and K.I. Marton, Medical Decision Making (Butterworths, Boston, 1988).
W.M. Tunbridge et al., The spectrum of thyroid disease in a community: the Eidkham survey, Clinical Endocrinology 7 (1977) 481-493.
W. Vach, R. Robner and M. Schumacher, Neural networks and logistic regression: Part II, Computational Statistics & Data Analysis 21 (1996) 683-701.
B. Warner and M. Misra, Understanding neural networks as statistical tools, The American Statistician 50 (1996) 284-293.
H. White, Learning in artificial neural networks: A statistical perspective, Neural Computation 1 (1989) 425-464.
H. White, Some asymptotic results for learning in single hidden layer feedforward networks, Journal of American Statistical Association 84 (1989) 1008-1013.
T.J. Wilke, Estimation of free thyroid hormone concentrations in the clinical laboratory, Clinical Chemistry 32 (1986) 585-592.
E.T. Wong and M.W. Steffes, A fundamental approach to the diagnosis of diseases of the thyroid gland, Clinical Laboratory Medicine 4 (1984) 655-670.
Author information
Authors and Affiliations
Rights and permissions
About this article
Cite this article
Zhang, G.(., Berardi, V.L. An investigation of neural networks in thyroid function diagnosis. Health Care Management Science 1, 29–37 (1998). https://doi.org/10.1023/A:1019078131698
Issue Date:
DOI: https://doi.org/10.1023/A:1019078131698