Phys. Med. Biol. 38 (1993) 959-970. Printed in the UK A neural network approach to automatic chromosome c1assfication Anne M Jennings and Jim Graham Depanment of Medical Biophysics, Universiw of Manchester, UK Received 5 October 1992, in final form 1 April 1993 Abstract. Classification of banded metaphase chromosomes is an important step io automated clinical chromosome analysis. We have conducted a preliminary investigation of the application of aaificial neural networks to this process, making use of a natural representation of the banding pattern. Two different network architectures have been compared the Kohonen %IF-organizing feature map and the multi-layer perceptron (MLP). For each of these a search of their respective parameter spaces over a limited range has resulted in configurations of modest dimension which achieve creditable classification mes. The MLP in particular shows promise of being a useful classifier. When size and shape features are supplied as inputs to the MLP in addition to a low-resolution banding profile, misclassificatinn lates are obtained which are comparable with those of a well developed statistical classifier. 1. Introduction In a normal human cell there are 46 chromosomes which, at an appropriate stage of cell division (metaphase), can be observed as separate objects using high-resolution light microscopy. Appropriately stained they show a series of bands along their length and a characteristic constriction called the centromere. Figure l(a) shows a typical metaphase cell, stained to produce the most commonly used banding appearance (G banding). Chromosome analysis, which involves visual examination of these cells, is routinely undertaken in hospital laboratories, for example for pre-natal diagnosis of genetic abnormality or monitoring of cancer treatment. This visual analysis involves counting the chromosomes and examining them for structural abnormalities. To determine the significance of both numerical and structural abnormality it is necessary to classify the chromosomes into 24 groups on the basis of their size, the pattern of bands and the centromere position. Twenty two of these groups normally contain two homologous (structurally identical) chromosomes. The other two groups contain the sex chromosomes X and Y. In the case of a normal male cell, the X and Y groups contain one chromosome each; in a female cell there is a homologous pair of X chromosomes and the Y group is empty. The time consuming nature of chromosome analysis has resulted in considerable interest in the development of automated systems based on machine vision. A number of such systems are now in routine use in many hospitals (see, e.g. Graham 1987, Graham and Pycock 1987; for a review see Lundsteen and Martin 1989). The chromosome classification performance of these systems depends on the type of material used, but is at best in the range 6 1 8 % misclassification (Piper and Granum 1989). This compares poorly with visual classification by cytotechnician, which was estimated by Lundsteen etal (1976) to result in 0031-9155/93/070959+12$07.50 @ 1993 IOP Publishing Ltd 959 960 A M Jennings and J Graham Figure 1. Chromosomes and chromosome feahlres. ( a ) A cell at metaphase. The individual chromosomes show the banding paftem (G banding) produced by staining. ( b ) Schematic drawing of a chromosome showing the position of the centromere. The density profile (helaw) is formed by projecting the density onto the curved centreline. a classification error rate of 3% for inspection of isolated chromosomes, dropping to 0.1% when all chromosomes in a cell could be examined together. All automated systems in clinical use operate interactively, allowing an expert operator to correct machine errors in image segmentation, feature extraction and classification, resulting in useful performance (Graham and Piper 1993). However, there is clear scope for improvement in automatic classification. An important issue for automatic classification is the representation of the banding pattern. Several different classifiers have been reported using statistical or syntactic approaches (Granlund 1976, Granum 1982, Lundsteen et al 1981, Groen et al 1989, Thomason and Granum 1986). Each of these involves the extraction of a number of intuitively defined features, usually associated with the chromosome’s density profile. The density profile is a one-dimensional pattern obtained by projecting the chromosome’s density onto its centre line (figure l(b)),and reflects the largely linear organization of the chromosome structure. The processing involved in extracting features from the profiles involves the risk of losing information, a risk which may he eliminated by using the density profile itself as the handing representation. This type of one-dimensional pattern is a natural form of input for artificial neural networks, in which the classification features are selected automatically by a process of training from examples. Neural networks have been successfully used for a range of pattern recognition applications (Kohonen 1988, Sejnowski and Rosenburg 1987, Haykin and Deng 1991). Their advantages lie in their ability to correctly classify noisy or incomplete data and the relative ease with which they can be retrained for classification of new types of data. These properties are likely to he useful for chromosome classification. Specimen preparation techniques in routine use evolve very rapidly, resulting in changes in chromosome appearance. In particular, there is an increasing clinical requirement to use higher-resolution handing for diagnostic purposes, resulting in routine examination of longer (prometaphase) chromosomes. This will result in the need for greater adaptability in automated karyotyping systems. In this paper we present a preliminary investigation into the feasibility of using neural networks as chromosome classifiers, using the density profile as the representation of the handing pattern. We compare two commonly used network configurations and assess their suitability for this problem. The data used for this study are a set of 2904 profiles from G-handed chromosomes from Rigshospitalet, Copenhagen (Gerdes and Lundsteen 1981). each annotated with the chromosome’s length, centromere position and classification Neural networks for chromosome classifcafion 961 assigned by an experienced cytogeneticist. The profiles were carehlly measured by densitometry from photographic negatives, care being taken to avoid overlapped or badly bent chromosomes. For this reason many of the cells in this data set are incomplete (there are, on average, 38.2 chromosomes per cell) which does not present a problem for this study as we are interested solely in classification of isolated chromosomes. However, the sex chromosomes, particularly the Y chromosomes, are represented by a very small number of examples. For this reason we have omitted the Y chromosomes from the study, only using data from the remaining 23 classes. 2. Neural nets There is a variety of different network architectures, each with its appropriate training algorithm (see, e.g. Wasserman 1989, or the review in Physics in Medicine and Biology by Clark (1991)). Two types of network have been predominantly used for classification problems: the Kohonen self-organizing feature map or Kohonen net (Kohonen 1990) and the multi-layer perceptron (hl~~).These nets accept continuous-valued rather than binary inputs and have been shown to generate low classification error rates in a range of other application domains (Kohonen 1988, Bisset et al 1989). output layer (2-0 array of nodes) hidden layer input layer (0) Figure 2. Neural nenvork architectures. (a)The Kohonen self-organizing feature map. Each input is connected by a weighted link to each of the nodes in a two-dimensional anay of outputs. (b) The MLP. The nodes are arranged in layers. Each node in a given layer is connected by weighted links to each node in the succeeding layer. The example shows a network with n input nodes, W output nodes and a single hidden layer, as used in this study. Figure Z(a) shows the topology of a Kohonen net. All input nodes have links to all nodes in the two-dimensional array of outputs. Associated with each link is a weight, initially assigned a random value. The value of an output node is the sum of the weighted inputs on each link and these output values are manipulated by a competitive leaming algorithm to arrive at an unsupervised clustering of output nodes (Kohonen 1990). As each training pattern is presented at the input, the output node whose pattern of weights most closely matches the input pattern is selected as the winner in the competition; its weights, and those of its neighbours, are updated to be closer to the input pattern, making it more likely to be selected and updated when a similar pattern is presented. In this way after a number of presentations of a representative set of training patterns, the output nodes form a map of regions each of which respond to different patterns in the training set. These regions can be 962 A M Jennings and J Graham labelled according to the true class of the patterns to which they respond. The learning is unsupervised since the correct classifications of the input pattern a e not used in adjusting the net parameters, merely in labelling the clusters after training. The MLP topology is illustrated in figure 2(b). Nodes are connected in layers, all the nodes in each layer being connected via weighted links to all nodes in the succeeding layer. Each node performs a non-linear transformation of the sum of its inputs, and the output of this transformation is sent along all of the node’s links to the next layer. As it passes along each link, this value is multiplied by the weight on that link before becoming one of the inputs to the next node. Each node may have an additional input, in the form of a bias or threshold which ensures that the summed inputs lie within an appropriate range for the transforming function. The input pattern is fed into the input layer, passing through several transformations before generating a pattern of outputs. There may be either one or two hidden layers of nodes between the inputs and the outputs, which act as feature detectors. In the experiments which we report below, only one hidden layer is used. For a given input, the output pattern is determined by the pattern of weights on the links. This pattern of weights is obtained by supervised training; each input pattern on training generates an output which is compared with the correct output pattern for its class, and the weights are adjusted so that the observed output is closer to the desired output. The bias values are also adjusted as part of the same process. The most commonly used algorithm for adjusting the weights is error back propagation (Rumelhart et a1 1986). The error signal at the output is used to propagate adjustments to the weights back through the network so that the resulting pattern of weights produces outputs closer to those appropriate for the input pattern. 3. The classification experiments The classification experiments involved training each network with approximately half of the chromosome data and using the remainder as ‘unseen’ data to test the network‘s classification performance, reversing the roles of the training and unseen set and averaging the classification rates. Several network parameters were varied for each network to investigate the effect on classification performance. One feature common to the two networks was the form of the input. We wished to use the banding profile to represent the banding pattern, i.e. to assign one profile sample to each input node. Cells are measured at different stages of contraction during metaphase, so that the sum of the lengths of all the chromosomes in a cell varies considerably. Furthermore, chromosomes do not contract uniformly along their lengths so that not only do longer chromosomes show more bands than shorter ones but this increase in banding resolution occurs unevenly along the chromosome. There is therefore considerable inter-cell variation in the appearance of chromosomes of the same class. One aspect of this variation is in the number of samples making up a profile, as these are taken at equal spacings along the centreline, typically at pixel positions in the metaphase image. This variation can be reduced by normalization of the sum of the chromosome lengths. In our experiments we use the median cell length as a normalization measure to compensate for the fact that many of the cells had missing chromosomes. To ensure that each input to the network corresponds to a consistent feature in the profiles, all chromosomes were presented using the same number of profile values. The number of profile value inputs needed to achieve best classification performance was one of the parameters investigated for each network. The longest profiles in our data set consisted of 116 elements. Shorter input vectors were generated by adding consecutive elements Neural networks for chromosome clars@cation 963 togethe: without averaging, a procedure which retains the total chromosome density in the profile, to produce profiles 58, 29 and 15 elements long. Additional inputs representing chromosome length and centromeric index were used in the case of the MLP. In each experiment the measure of performance was the percentage of alI chromosomes misclassified in both the training set and the unseen test set. A misclassification in this case was taken to be the assignment of a chromosome to a class other than that to which it had been assigned by the cytogeneticist when the data set was collected. The important misclassification rate is, of course, that for the unseen portion of the data, which provides information on the network's ability to generalize; it is this value we seek to minimize by selecting appropriate network parameters. 3.1. Kohonen self-organizing map 3.1.1. Mefhodr. The Kohonen map was investigated using only the banding data as input. Three experiments were conducted. (i) The number of input nodes. In this experiment the size of the map was fixed at 81 nodes (a 9 x 9 array). Each network, with 116, 58 and 29 input nodes, respectively, was trained using 30 examples of each of the 23 classes and tested on 50 unseen examples noting the classification performance. The number of output nodes was chosen to be similar to that which had been used successfully in a problem of similar magnitude, in which 96 output nodes were used to recognize 18 phonemes (Kohonen 1988). (ii) The number of training examples. The selection of 30 examples for training in experiment (i) was rather arbitrary, and the effect on the selected network of altering the number of training examples was therefore investigated. Classification performance was noted using training sets of 10, 20, 30 and 40 chromosomes. (iii) The size of the output map. Finally the effect on classification performance of the size of the output map was considered. The size of the output map was varied from 7 x I to 20 x 20 nodes using eight passes of 40 examples of each chromosome for training. 3.1.2. Results. (i) The number of input nodes. Table 1 shows that the classification error rate on both training and unseen data is relatively insensitive to the number of inputs used (i.e. to the resoluton at which the banding profile is presented). As both training time and labelling time are highly dependent on the number of inputs (and hence the number of weighted links), there is a clear advantage in the use of shorter input vectors. Experiments (ii) and (iii) were conducted using 29 input nodes. Table 1. Variation in the average classification error of a Kohonen net as the number of input values repnenting the banding pattern is varied. Output map size: 9 x % tmining set: 30 examples of each Chromosome class: test set: 50 examples of each class. Average classification mor (94) Number of input values Training set (690 chromosomes) Unseen set (1150 chromosomes) 116 24.0 22.8 37.4 37.0 37.1 58 29 22.4 (ii) The number of training examples. Figure 3 demonstrates the effect of the number of training examples on the net's ability to classify and to generalize to unseen data. Training A M Jennings and J Graham 964 passes using different numbers of example chromosomes show that when 40 examples are used (figure 3(d))not only is best classification performance achieved, but the performance on unseen data matches that on training data, indicating that the net is 'generalizing', or responding to general features of the training data, rather than specifically modelling the training examples. classification error classification errw loi(a) 0 0 , , , , 2 4 6 8 6 8 1 10 (6) 0 0 2 4 6 8 10 5040- (4 0 0 (4 0 2 4 training cycle 0 2 4 6 8 training cycle F i y r c 3. Training the Kohonen net with different numbers of examples of each class. (U) 10, (b) 20, ( c ) 30, ( d ) 40 examples in the training set. The curVeS show the percentage of chromosomes incorrectly classified in the training se1 (broken) and the unseen set (full) as the number of passes of the trainins data is increased. There is a clear improvement in both absolute classification performance and ability to generalize. (iii) The size of the output map. The effect on classification of the size of the output map is shown in table 2. A consistent training regime (eight passes of 40 examples, following the result of experiment (ii)) was used for each map, and under this regime, best 965 Neural networks for chromosome classification performance was obtained using an 18 x 18 may of output nodes. As with the smaller maps, only a few passes of training data were required for the network to settle down to a fairly stable configuration (figure 3), after which the misclassification rates oscillate as the number of training passes is increased. The minimum misclassification rate observed with this configuration occurred after six passes of the training data, giving values of 11.8% on training data and 16.7% on unseen data. Table 2. Variation in the average classification error of a Kohonen net with the size of the output map. Input vector size: 29 values; mining set: 40 examples of each class. Average classification enor (%) Size of output map 7x7 9x9 11 x 11 13 x 13 15 x 15 17 x 17 18 x 18 19 x 19 20 x 20 ~ ~. Training set (920 chromosomes) Unseen set (920 chromosomes) 35.4 26.7 39.5 25.9 23.6 23.3 20.3 20.1 18.9 22.3 20.5 22.6 17.2 15.4 14.2 12.7 14.1 8.9 3.2. Multi-her perceptron 3.2.1. Methods. The MLP net was trained using the back-propagation algorithm of Rumelhart er a1 (1986), the weights and bias values being updated after every presentation of a training profile. The back-propagation algorithm provides two parameters, gain (or learning rate) and momentum, controlling the rate at which weights change in response to error signals. We wished to investigate the effect on classification of varying the gain and momentum values, as well as the size of the input vector and the number of hidden nodes. For simplicity we reshicted our study to the use of a single layer of hidden nodes. As we were not aware of any similar studies in the literature which might give us reasonable stating values for OUT experiments (as had been the case with the Kohonen net) we conducted some pilot experiments, similar to those reported below using chromosomes from groups 1-5 only (the longest chromosomes). The results of these experiments indicated that the number of inputs representing the banding pattern could be reduced to 15 without loss of classification ability, and that 15 hidden nodes were adequate. Experimental determination of appropriate gain and momentum values gave figures of 0.3 and 0.7 respectively (values which have been used successfully in other net simulations; see, e.g. Lippmann 1987). All the nets tested had 23 output nodes, one for each class (Ychromosomes being omitted). The desired output is therefore close to unity at the output node corresponding to the correct class, and zero elsewhere. The classification produced by the network on being presented with a test pattern was taken to be the node with the highest output, even if this was significantly lower than unity. (i) Gain and momentum. The gain term q determines the rate at which weights are altered in response to an observed error signal. A large value of gain (near 1.0) produces fast training and is useful in moving weights rapidly away from their initial random values. However, once the weight configuration is close to a minimum in error space, relatively large weight shifts can have the effect of ‘jumping’out of a global minimum and into a nearby 966 A M Jennings and J Graham local minimum. Small values of gain are more appropriate later in training and a useful approach is to successively reduce the learning rate as training proceeds. Two measures of network performance may be used to determine the points at which the gain term is reduced. These are the total net error (the sum of the absolute values of the differences between the desired and observed net outputs over a txaining pass) and the classification performance on training data. Both of these measures reflect the net’s classification ability, but they are not identical, due to the fact that only the output node with the highest value affects classification. At different stages of the learning process one may be more useful than the other. This difference between these two measures has been noted in other applications (Dah1 1987). The weight change at a node at a given training cycle, determined by the observed error signal and the gain, may be further augmented by a fraction of the weight change which occurred at the previous cycle. In this way successive weight changes in the same direction are encouraged, whereas weight changes in successively different directions are damped, reducing oscillatory behaviour. The momentum term 01 specifies the fraction of the previous weight change which is added at each cycle. We investigated the best choices for momentum and initial gain, as well as the effect of reducing gain using the two criteria described above. (ii) The number of input nodes. The result of our pilot experiment, that the profiles could be reduced to 15 samples (one eighth of the length of the longest profile), was tested by measuring classification performance on all 23 classes using a network with 15 hidden nodes, using gain and momentum values determined in experiment (i). (iii) The number of hidden nodes. The classification performance was determined using different numbers of nodes in the hidden layer. (iv) The use of additional chromosome features. Experiments (i)-(iii) were designed to examine the feasibility of configuring an MLP to use the density profile as a banding representation for classification. As noted in the introduction, chromosome size and centromere position are also important classification features. To test the overall classification performance of our network, two additional input nodes were included for the normalized length and the centromeric index. me centromere divides the chromosome into a long ‘arm’ and a short ‘arm’ (figure l(b)). The centromeric index is the ratio of the length of the short arm to the whole chromosome length.) The effect on classification of adding these features was examined. Results. (i) Gain and momentum. The effect of reducing gain during training is demonstrated in figure 4, which shows a succession of training passes for a 15-15-23 network. Each training cycle involved the presentation of 1150 chromosomes. Figure 4(u) compares the network convergence on training with and without a reduction in gain. In this case gain was halved whenever a 10% increase in net error occurred or if the total number of correctly classified chromosomes had not increased by at least two over the previous presentation of the training data. The gain reduction method clearly results in improved convergence. Figure 4(b) demonstrates the effectiveness of using both tota1 net error and classification performance as criteria for reducing gain. The starting value of gain was 0.3 and momentum was 0.7, as determined by the pilot study. Figure 5(a) and (b) shows training passes using different values of momentum and initial gain confirming that 0.7 and 0.3 are the most appropriate for this configuration. (ii) The number of input nodes. Table 3 confirms that the classification performance is fairly insensitive to the number of values used to represent the banding profile (and hence, the numbers of inputs). No loss in classification accuracy is observed when an input vector of 15 values is used. 3.2.2. Neural networks for chromosome clussi&ution 961 classification error ..- x - no reduction in gain term 10 (4 0 0 2 4 8 6 10 12 14 16 18 20 ." Gain term reduced using A - net error only + - net error and Classification performance 40 30 .. . . . . . . . . . . . . . . . . . . . . . . . . . 0 8 1 6 ' 2 4 3 2 4 0 4 8 5 6 6 4 7 2 8 0 8 8 training cycle Figure 4. The effect of reducing gain during vaining of the MW.A 15-15-23 net was used with momentum (I = 0.7 and initial value of gain q = 0.3. The graphs show the percentage of the training set incorrectly classified as the data are successively presented to the network.. (a)Comparison of no gain reduction with gain reduction based on totll network enor. (b) Comparison of gain reduction criteria. Total network error alone compared with network emx and classification performance combined. (iii) The number of hidden nodes. The choice of 15 nodes in the single hidden layer is confirmed by figure 5(c) which shows training passes of 15-10-23.15-15-23 and 15-20-23 networb. A M Jennings and J Graham 968 classiocation emor classification error M b 4 8 I6 12 20 24 (4 0 0 4 8 I2 20 16 training cycle 0 4 8 12 16 20 24 Figure 5. The effect on MLP training of varying networlt parameters The percentage of the training set incorrectly classified is shown as training proceeds. ( 6 ) Varying the value of the momenluin term (I. The gain term q is initially set to 0.3 in each case and reduced during mining as described in the text. x. a = 0.9; A, a = 0.7 and +, (I = O S . (b)Varying the starring value of the gain term 4. The momentum term (I is set to 0.7 in each case. x. q = 0.5: A, 4 = 0.3 and +. q = 0.1. (c) Varying the number of nodes in the hidden layer. = 0.7, q = 0.3. x, 10 nodes; A, 15 nodes and +, 20 nodes. Table 3. Variation in classification e m r rate of the MLP with the number of input vdues. The network had 15 hidden nodes and 23 output nodes. Number of inout values 15 58 116 Avenge classification error (%) , , Training set (1150 chromosomes) U n y n set (690 chromosomes) 2.2 3.9 3.0 10.1 10.0 10.3 (iv) The use of additional chromosome features. Table 4 shows the effect on classification rates of using additional inputs for normalized chromosome length and centromeric index using (16 or 17)-15-23 networks. These features clearly contribute siznificantly to classification performance. The overall best misclassification rates observed were 2.6% on training data and 6.6% on unseen data. Table 4. The contributions of chromosome features to classification by the M W Features used for cIassiR ctltion Grey level profile Grey level profile and normalized length Grey level profile, normalized length and ana centromeric index Average classifiuIion error (5%) . , , , Training set (990 chromosomes) Unseen set (990 chromosomcs) 4.3 4.0 10.9 8.8 2.6 6.6 Neural networks for chromosome classij'ication 969 4. Discussion This preliminary study has investigated the feasibility of using neural network classifiers as part of an automated chromosome analysis system. The motivation for investigating neural networks in this application is their potential adaptability to changes in the data arising from evoluton in the sample preparation techniques and the desire to use a classifier making use of a more natural representation of the banding pattern than has been used in previous studies. The MLP shows considerablepromise as a classifier. By varying the operating parameters we have found a configuration which is capable of achieving useful classification rates. Our search of the parameter space was not broad and, even within its own narrow limits, not exhaustive. The final configuration of parmeters may not therefore be the best obtainable, but it is gratifying to note that a 'good' combination was found fairly rapidly. It will be the subject of further study to investigate whether a more thorough search of parameter space will improve the network further. For comparison, the best published classification rates to be found in the literature (to our knowledge) are those reported by Piper and Granum (1989) who achieve 5.9% misclassifications on a superset of the profile data used here. The additiond data used by these authors included a number of badly bent and some overlapping chromosomes (excluded from our data). Their results, however, include the effect of a further classification step which imposes the constraint that each chromosome class contains at most two chromosomes. While our results are not directly comparable with theirs, we are encouraged to achieve similar classification performance on similar data. Our experiments with the MLP demonstrate that considerable advantage is to be gained by careful adjustment of network parameters and mining conditions. Table 5. Comparison of the results obtained for chromosome classification using a Kohonen net with parameters and results for two other studies. The classification accuracy reported for phoneme recognition including post-processing using a context-sensitive grammar. Number of inputs Output map size Number of classes Number of training examples Error for unseen set (%) Recognition of phonemes Chromosome classification Recognitition of the spoken words 'yes' and 'no' 29 18 x 18 23 40 16.7 15 15 7x7 2 798 8x12 6.9 3-8 18 50 It is unsurprising that the performance of the Kohonen map is inferior to the MLP in this application. For one thing, supervised training is much more appropriate to this data than unsupervised training. Table 5 shows a comparison of our results with two other applications of the Kohonen map on phoneme recognition (Kohonen 1988) and on recognition of the spoken words 'yes' and 'no' (Lucas and Kittler 1989). The results on phoneme recognition include a final supervised training phase, and refer to the classification accuracy of speech after processing the neural net output using a context-sensitive grammar (similar in its effect to the rearrrangment included in the results of Piper and Granum). It is tempting to speculate that better classification results could be obtained using the Kohonen map for chromosome classification, for example by increasing the size of the training set. This option was not available to us in this study, and the more promising performance of the MLP suggests that development of that route is more appropriate as a practical approach to this problem. 970 A M Jennings and J Graham Acknowledgments This work was supported by funding from the Science and Engineering Research Council. It was greatly facilitated by the exchange of materials within the Concerted Action of Automated Cytogenetics Groups, supported by the European Community (project II.1.1113). References Bisset D L, Filho E and Fairhunt M C 1989 A comparative study of neural network stmcNres for practical application in a pattern recognition environment Is1 IEE Int C o d on Arlificiul Neural Networks (London, 19891 (London: pp 37&82 Clark J W 1991 Neural network modelling Phvs. Med B i d . 36 1259-317 Dah1 D E 1987 Accelerated learning usingthegeneralised delta mle lE&E I s l l n l . Can$ on Neural Neworkr vol2 (New York: IEEE)pp 523-30 Gerdes T and Lundsteen C 1981 Documenlalion elhe Rigshmpilal Chromosome Densiry Profile Dura Base Department of Obstetrics and Gynaecology YA and Department of Paediatrics, RigshospitaleL University of Copenhagen Graham J 1987 Automation of routine c l i n i d chromosome analysis I. Karyotyping by machine Anal. Quonr. Cylol. Hisfol. 9 383-90 Graham J and Piper J 1993 Automatic karyotype analysis Chromasome Analysis ProMcols ed I R Gosden (Cliffon, NI: Humana) at press Graham J and Fycock D 1987 Automation of routine clinical chromosome analysis 11. metaphase finding Anal. Quonr. Cytol. H i d . 9 391-7 Gmlund G H 1976 Identification of human chromosomes using.integrated density ~rofilesIEEE Truu, Biomed. . Eng. BME-23 183-92 G m u m E 1982 Amlicakion of statistical and syntactical methods of analysis to classification of chromosome data Pallern Reeo&irion TheorynndApplicot& ed J Kirtler. K S Fu and L F Pan NATO AS1 (Dordrecht: Reidel) pp 373-98 Groen F C A, tenKate T K,Smeulders A W M and Young I T 1989 Human chromosome classification based on local band descriptors Panem Recognition lelt 9 21 1-22 Havkin S and Dens! C 1991 Classification of radar clutter usinn - neural networks IEEE Trans.Neural Networks "-2589600 Kohonen T 1988 The neural phonetic typewriter Compurer 21 11-22 - 1990 Se~Orgmtsmonid Arsocralive Memory.3rd edn (Bcrlm: Springer) Lippmmnn R P 1987 An introduction 10 computing with neunl ne& /€&E ASSP Mag. 4 h?? Luws A E and Kittler J I989 A compmtive study of the Kohonen and multiedit neuml net I m g dgorithm Irr IEE Inr Conf on Anificiol .Seeural Vcnvorkr ( L o n d m 19891 (London: LEE) pp 7-1 I Lundsrecn C. Gxdcs T. G m u m E and PNlip J 1981 Aaom3lic ehromosomc analysis n. Kllryoryping of banded human chromosomes using band lnnsition sequences Ciin. Genu. 19 2 6 3 6 Lundsteen C, Lind A-M and Cmnum E 1976 Ksud classitcanon of banded human chromosomes 1. Kqoryping mmpxcd with classification of isolxcd chromosomes Am. 1. Hum Genei. 40 81-97 Lundsteen C and h l m n A 0 1989 On the selection of systems lor nutomarcd cylogenetic mdysis A n 1. Med. Gener. 32 72-80 hpcr J 3nd G m u m E 1989 On fully aulomatic measuiemcnt for banded chromosome classification Cyiomerry 10 242-85 Rumelhart D E. Hinton G E and Willims R J 1986 Leaming mternd rcprecenwuons by m o r propagation Parollel Dislribured Pmcerringr Erpiorortom in !he Microrrrueiurerof Cognilion YOI I , Foundztions sd D E Rumclhart and J L McCellmd (Cambridge. MA. MlT press) pp 51842 Sejnosski T J and Rosenburg C R 1987 Parallel networks that I- to pronounce Engllsh text ComplexSyre,ns 111848 Thomason .U G m d G m u m E 1986 Dynmically prognmmcd inrexnce or hlxkov networks from finire sets or sample swing IEEE Tmm. P..MI P&\n-8 491-801 \h'~ssermanP 1949 .Veurol C o m p i i n g - T k q ond Prmiice (New York. Yon Noswand Reinhold) i~m ~~ .
© Copyright 2026 Paperzz