American Journal of Psychology and Behavioral Sciences 2015; 2(2): 33-40 Published online March 10, 2015 (http://www.openscienceonline.com/journal/ajpbs) A Review of Computational Classical Conditioning Models Mehmet Emin Tagluk1, Omer Faruk Ertugrul2, * 1 2 Department of Electrical and Electronic Engineering, Inonu University, Malatya, Turkey Department of Electrical and Electronic Engineering, Batman University, Batman, Turkey Email address [email protected] (M. E. Tagluk), [email protected] (O. F. Ertugrul) To cite this article Mehmet Emin Tagluk, Omer Faruk Ertugrul. A Review of Computational Classical Conditioning Models. American Journal of Psychology and Behavioral Sciences. Vol. 2, No. 2, 2015, pp. 33-40. Abstract Classical conditioning (CC), which is a basic learning phenomenon, used to explain basic emotions, such as fear, phobia, and reflexes. It was introduced by Pavlov in 1927, and since then it has been investigated in psychological, behavioral, memory, neuroscience, and neurobiology perspectives. It has been used for investigating consumer behavior, conditional fear acquisition and extinction, response extinction and robotic control. Additionally, a large number of methods were proposed to model the CC learning stage. Unfortunately, none of them could model all outcomes of CC. In this study, these computational models, their usages and also the papers about their comparisons are reviewed. It obvious from this review that there is a high requirement to a model, which has a capability to model each feature of CC. Keywords Classical Conditioning, Pavlov, Computational Model, Behavioral Learning 1. Introduction The classical conditioning (CC) is at the heart of our understanding of human learning. CC, which was first demonstrated experimentally by Pavlov in 1927 [1] named as a conditioned reflex, is based on the association of two stimuli; a stimulus (S), which evokes either no or a weak response, usually unrelated to the response that eventually will be learned, and an unconditioned stimulus (US), which consistent a response called the unconditioned response (UR) [2]. In CC, a temporal association between S and US is learned [4] by presenting the S before the US. After a successful conditioning, a specific response R will be observed whatever the presence or absence of the US [5] and the probability to observe and the strength of CR increases over multiple training sessions [2], i.e. can be defined as been strengthening the associative link [6]. Finally, an association is formed between S and US as a result of the time series pairing (S, US) and S is termed as CS, anymore. Similarly, this association can be weaker while having CS with non-occurrence of the expected US followed for a long time [7]. Additionally, behavioral learning theories, in which learning is defined as a change in behavior that occurs as a result of experience, are based on findings about CC. The major outcome of CC is about the main knowledge about human. According to Descartes and the most of other thinkers in the past several centuries, the human body is a marvelous machine and all of its functions or properties are unknown. Therefore they only interested in mind. On the other hand, Pavlov [1] showed that body is a living machine not only a machine or robot, which being touched in the world, by experiments, and this is similar to Aristotle view. For example, Miguez et al. reviewed the results of stress, and they showed that it can produce conditioned analgesia or conditioned hyperalgesia, which can be explained by CC [8], Lindquist et al. investigates the effect of ethanol in CC delay time by using the results of experiments on rats [9] and Dalla and Shors and Brom et al. reviewed the sex differences in learning processes [10, 11]. These examples showed that, the human body is a living machine and his learning is also depending on the physical properties of him. CC is dealing with learned reflex; therefore it becomes a very important issue to understand human or animal behavior, to control behaviors or emotions of them. Classical conditioning can be seen as a basic learning scheme; the learning experiments are sometimes done by human subject [12, 13, 14], but, generally with animals [13, 15, 16, 17, 18, American Journal of Psychology and Behavioral Sciences 2015; 2(2): 33-40 19]. Additionally, it is important to recognize that the performance of animals in a conditioning experiment based on the memory and emotion processes of that animal. Therefore the results of experiments provide a rich knowledge that can be used for studying basic learning, memory, and emotion processes in animals. The past decade, researchers have shown an increased interest in CC and they investigated this stage in various perspectives, such as: psychological [20], behavioral [21], long and short term memory [22], neuroscience behind CC [23], neurobiology [20, 21, 24], the role of the Cerebellum in CC [25], CC of motor responses [26], the activity of EEG signals in theta band during CC [18, 27], biological neurons [2, 7, 28, 29, 30] and also in Pulsed Neural Chip [31, 32]. Drew et al. investigate the effect of psychological operant and classical on Central Pattern Generator Neural Circuit modeled by Hodgkin-Huxley. They reported that the simulations failed to fully reproduce empirical observations [28]. Moustafa et al. proposed a biologically plausible model, which is able to simulate the behavioral effects: including latent inhibition, acquired equivalence, sensory preconditioning, negative patterning, context shift effects, the effect of the number of training trials on blocking and overshadowing [30]. Magri et al. improved the biological neural network of McGregor to assess the psychological classical conditioning in biological network [33]. As a summary, the CC cannot explain the human thoughts; it is dealing with the emotional responses [34], reflexes [35], phobia and superstitions [36]. New CR cannot be produced by CC, only the probability of occurrence of a CR and its strength can be changed [34]. However, the researchers were showed that a S pairing cannot be created for all types of S, it depends on the requirements of the organism [37]. The strength of R and more permanent pairings is correlated with the requirements and expectation of the organism [34, 37]. Additionally, the CC learning also depends to experiences of organism [38] and the statistical dependency of S [39]. The purpose of this paper is to review the researches, applications about CC. This paper has been divided into five parts. The first part deals with the CC learning stage, the second part is about its applications, the third one is about the proposed computational models, the fourth part reviewed the comparison papers and the last one concluded the paper. 2. Applications of CC There is a large volume of published studies describing the role of CC in consumer behavior [40, 41, 42, 43], network security by dynamic policy-based management [44, 45], robotic control [5, 7, 29], response extinction in rats [46] and conditional fear acquisition and extinction [15, 16, 19, 47, 48]. Additionally, the effect of CC in obesity is also investigated [49]. Allen and Madden (1985) reported that classical conditioning is an important tool for understanding and producing advertising effects, and their study is a general review of the theories of classical conditioning in consumer 34 behavior [40]. These types of consumer studies mentioned that the consumer behavior can be manipulated by CC and experiments were evaluated on this [41, 42, 43]. Guoshi et al. developed a biophysical network model of the lateral amygdala neurons to investigate the underlying mechanisms for acquisition and extinction of conditioned fear. They used Hodgkin-Huxley formalism to model neurons and Hebbian type synaptic plasticity to model the learning process. They reported that the network model, they proposed, can replicate the neuronal behaviors successfully [47]. Additionally, several attempts have been made to investigate fear conditioning by using human subjects [14, 49] and animal experiments; Yoshida et al. used goldfish by using fear-related classical heart rate in their experiments [15], Yamada made his experiments on mice [16] and Baldi and Bucherelli used rat subjects in their experiment [19]. In 2007, Erasmus demonstrated CC by using RealNeuron networks based on associative learning. They reported that the RealNeuron networks are simple and can be easily implemented to standard discreet-logic integrated circuits [7]. Salotti and Lepretre presented that the results of CC in robot training are still very far from what can be obtained with animals. They proposed a new model that integrates both classical and operant conditioning mechanisms, which may be a source of inspiration to control the basic interaction mechanisms of robots [5]. Lehmann showed the applicability of CC by the Edinburgh Classical Conditioning Pulsed Neural Chip that implements a drive-reinforcement learning algorithm for continuous-time pulsed networks with nonlinear synapses. He reported that the circuit shows good results [31, 32]. However, a major problem with this kind of application is that these applications cannot model each particular feature of the CC, because of the proposed methods. 3. Computational Models Furze analyzed the machine learning applications of CC in his PhD thesis [50]. In [51], Ludvig et al. presented the reinforcement learning methods. Dawson published a book about connectionism in CC [52]. Additionally, Schmajuk published two books about mechanisms and computational methods in CC [53, 54]. Furthermore, there is a large and growing literature in CC, but unfortunately, there is not any proposed method that can explain all features of CC yet [55], e.g. spontaneous recovery. One observer, Balkenius and Morén, have already drawn attention to that, the CC can be seen a very simple and basic phenomena, but unfortunately, there is not any model that capable of explaining all parts of CC for the last century. Therefore, they pointed out that the CC is one of the hardest questions in learning [55]. Lack of enough experimental results because of the ethical rules, has existed as a main reason of this problem, but which may have important outcomes, for instance, Chester defined classical conditioning as a form of temporal learning that may be useful in intelligent control [56]. Several studies modeling CC have been carried out, such as: LA Pyramidal Cell based on the Hebbian learning method [47], Bayesian method [57, 58], 35 Mehmet Emin Tagluk and Omer Faruk Ertugrul: A Review of Computational Classical Conditioning Models Hidden Markov method [59], Bayesian confidence propagating neural networks (BCPNNs) that employed Hebbian learning [60] and artificial neural network based methods [4, 44, 45, 61]. Gershman and Niv presented a Bayesian based method, which is based on the idea of that animals learn the hidden structure of the environment from their experience and they employ this internal model of the environment to make predictions about unobserved or future variables [58]. Courville et al. proposed a framework based on Bayesian model to explain how animals cope with uncertainty about contingencies in CC including acquisition, negative and positive patterning, and forward and backward blocking [57]. Austermann and Yamada demonstrated the CC learning stage by using a robot that can learn by rewards. In this method, there is a cooperative learning through the robot and its user [59]. Additionally, some special computational models were developed to define the CC stage [62], which can be categorized into two parts: reinforcement and artificial neural network based approaches, and they were explained in next section briefly. 3.1. Reinforcement Learning Based Models The popular reinforcement learning based models [51] are Sutton-Barto (SB) [63], SOP [64], TD [65], DR [66], Balkenius Modeli (BM) [67] and CP [68] models [6, 55]. The output in reinforcement learning method is calculated as follows: the weight adjustments are: ∆ ∑ = = (1) ̅ , Table 1. Eligibility trace ( ̅ ). ̅ +1 = ̅ DR CP SOP ̅ = ̅ & +1 = ̅ ̅ ̅ +1 = +1 = +1 = +1 = ̅ | #& # + ̅ ̅ + + | − − − +% 1− ̅ + 1− ̅ − −& −1 , −1 , >0 ≤0 ∗ +% ̅ +1 = +1 = +1 = #, , ̅ #& ∗ ∗ ∗ +1 − − + − −1 − +% 1− ̅ + %# 1 − ̅ −1 ∗ −1 −1 ∗ −& −& (7) (8) −1 (9) (10) %, % and %# are learning coefficients inside 3.2. Artificial Neural Network Based Models The popular artificial neural network based model is Klopf [69] model. The other models are policy management model [PM] [44, 45], adaptive threshold learning model (ATL) [4, 70], and our model, ET [71]. The equations are listed in Table 3. Table 3. Artificial Neural Network Based Methods ( Equation Klop f ∆ = =∆ * += PM ( | −) − |∆ , -. * + + ,/0 1. ∆, = ∗ 2 ∗ * + ∗ * + − 4 1− 0< − 4 26 2 =3 − 4 1.5 − 6< − 6 +∆ . −< ; = +: 2; . −< < +∆ . −< . =< + = 2= . −< . ?@A ?BA => 1−% ?@A ?BA -. . = tanh ?@A +% − OP N = O N + N = Q N−∆ O N+ 4 . . (12) −< −< −< >0 −< <0 -. + ?@A 1. − ∆ C. .M N ∗RO N + 100S (11) ≤6 ?BA -. + ?@A + I 4JK ∆L . − 1 .M N = . N − . N − ∆ ET (6) ̅ & where , 0 and 1. (4) (5) ̅ SOP ATL − −1 ̅ CP −1 + +% TD (3) − − SB / DR Equation SB / TD ). Equation (2) where, denotes weights, ∆ , shows weight adjustment, is input, ̅ denotes eligibility trace of , is reinforcement term which is depend on the relation between reinforcement output ( ) and real output ( ) [6]. The eligibility trace and reinforcement terms are different in each method and they are summarized in Table 1 and 2. , Table 2. Reinforcement Term ( >0 <0 (13) 1. (14) is learning coefficient in Klopf model [69]. T is cache cycle, j is time instant, US is the initial value (x0), CS is input (xi) and c is a small coefficient [44, 45]. ; , 2; , = , 2= are American Journal of Psychology and Behavioral Sciences 2015; 2(2): 33-40 learning coefficients, % forgetting coefficient shows eligibility, Hi is the threshold and Si is output in the ATL model [4]. In ATL, both threshold values and weights are updated. The basic properties of CC, such as: delay and trace conditioning, different ISI effects, blocking, overshadowing, and compound stimulus, were demonstrated in their study [4]. The (0 < < 1) and (0 < < 1) are association and extinction coefficients [71]. In [44, 45] the typical network security cases are presented to simulate how network management policies, which are described in this study as a conditioned reflex, for a dynamic learning process based on classical conditioning. They simulated the acquisition, extinction, reacquisition, secondary conditioning phenomena in network security policies. Therefore the proposed approach is dynamic and it can be employed for self-management, which is a better policy scheme than using static rules. The ATL method has the ability of modeling trace and delay conditioning, acquisition, extinction, reaquisition, blocking, conditioned inhibition, secondary conditioning and facilitation, but it cannot model spontaneous recovery [71]. 36 Table 4. Comparison of CC models [55] Feature Trace Conditioning Delay Conditioning ISI Curve S Shaped Acquisition Extinction Reacquisition Blocking Secondary Conditioning Spontaneous Recovery Conditioned Inhibition Facilitation SB TD Klopf Balkenius SD + + + + + - + ± + + - + - ± + ± + + - + + + + + ± + + + + + + ± ± + + - - - - - - + + + + + + + + + + Malaka and Hammer reviewed real CC models; Sutton/Barto (1981), SOP (Wagner, 198l), temporal difference (Sutton & Barto, 1990), the drive-reinforcement (Klopf, 1989) and CP (Malaka, 1995). The results of their comparison are sorted in Table 5 [6]. 4. Comparison of CC Models Table 5. Comparison of CC models [6] Chester compared Sutton-Barto (1981), Klopf (1989) and Grossberg-Schmajuk (1989) models in his study. The models he assessed are based on Hebbian learning, but they differ in how events are remembered. He reported that the models he compared fall far short of being good learners of temporal associations. Sutton-Barto, Klopf and the Hebbian models associate CS with R without the US and the Klopf model is a better associator than the Sutton- Barto model. Also, he presented that the Grossberg- Schmajuk model is more suited to learning temporal associations than the others, on the other hand, it cannot be respond with a sharp memory of a transient event and recover quickly after the situation has returned to normal [57]. Cheung et al. demonstrated the real time models of CC, such as; the unsupervised spatio-temporal single-neuron models of Sutton-Barto (1981, 1982), Rescorla-Wagner (1972), and Klopf (1988, 1991). They reported that Klopf model can be used for demonstrating CC features such as delay and trace conditioning, inter-stimulus interval effects, second-order conditioning, conditioned inhibition, extinction, reacquisition, backward conditioning, blocking, overshadowing, compound conditioning, and discriminatory stimulus effects [70]. Balkenius and Moren compared the most popular CC models, which are Sutton-Barto (1981), temporal difference (Sutton-Barto, 1987), Klopf (Drive-Reinforcement, 1982), Balkenius (1995, 1996, 1998), Schmajuk-DiCarlo (1992) models [72]. These comparisons are based on the applicability of the CC features, such as: acquisition, inter-stimulus interval effects, extinction, reacquisition, blocking, conditioned inhibition and secondary conditioning. The results of their comparison are sorted in Table 4 [55]. Positive Forward Conditioning Inhibition Backward Conditioning without Background Inhibition Backward Conditioning with Background Extinction Blocking Delay Conditioning Conditioned Inhibition Unpaired Inhibition Second Order Conditioning Complete Serial Compound Conditioning SB DR TD CP SOP CP/ TD + + + + + + + + - - + + + + + - + + + + + - + + + + - + + ± + - + + + + - + + ± + ± + + + + ± + ± + + - + - - + + + + Additionally, Vogel et al. [73] described the quantitative models of Pavlovian conditioning and Malaka [74] analyzed CC models, such as: Sutton and Barto (1981), Drive-reinforcement (Klopf, 1989; Klopf and Morgan, 1990), Temporal difference (TD) models (Sutton and Barto, 1990), Sometimes opponent processes (SOP) (Wagner, 1981; Donegan et al., 1989; Wagner and Donegan, 1989) and CP model (Malaka et al., 1995a, b; Malaka and Hammer, 1996) with respect to their mathematical basis. The comparison results that he obtained were summarized in Table 6 and 7 [74]. 37 Mehmet Emin Tagluk and Omer Faruk Ertugrul: A Review of Computational Classical Conditioning Models Table 6. Eligibility term [74] Conditioning paradigm Positive Forward Conditioning Inhib. backward conditioning (IBC) IBC with backgroundb Conditioned inhibition Convergence of multi-trial learning Monotonicc ∆wi (xi ) Extinction (short CS) Extinction of long CSs and Extinction of CI possibled Non-extinction of CI possibled Unpaired inhibitione Delay-conditioning (DC) ineffective for long CS One-trial effectiveness in DCg Discrimination/generalization proportional to CS overlaps Differential conditioning Compound conditioningh Over shadowingh Second-order conditioning Blocking CSD prediction CSC prediction SB/TD + + + + +b + + + + + + + + + + + + + + + + where a Only for a short range of ISIs possible, b multi-trial learning can be described by trial-based models, c highest xi lead to largest weight changes, d extinction and non-extinction of conditioned inhibitors depend on the restriction of y in the reinforcement term, e is modeled as inhib. backward DR + + + + + + + + + + + + + + + + + + + + SOP + + + + + + + + + + + + + + + + ± + + CP + + + + +b + + + + + + + + + + + + + + + + CP-CP + + + + +b + + + + + + + + + + + + + + + + CP+SB + + + + +b + + + + + + + + + + + + + + + + conditioning with long negative ISI, f only possible with background and appropriate eligibility, g for long CSs, the learning result after multiple trials is the same as after one trial and h depends on stimulus overlaps and initial weights, with SOP eligibility additionally ISI-dependent. Table 7. Reinforcement Term [74] Conditioning paradigm Positive Forward Conditioning Inhib. backward conditioning (IBC) IBC with backgroundb Conditioned inhibition Convergence of multi-trial learning Monotonicc ∆wi (xi ) Extinction (short CS) Extinction of long CSs and Extinction of CI possibled Non-extinction of CI possibled Unpaired inhibitione Delay-conditioning (DC) ineffective for long CS One-trial effectiveness in DCg Discrimination/generalization proportional to CS overlaps Differential conditioning Compound conditioningh Overshadowingh Second-order conditioning Blocking CSD prediction CSC prediction SB/TD + +-a +-a + + + + + + + + + + + + + + - where a Only for a short range of ISIs possible, b multi-trial learning can be described by trial-based models, c highest xi lead to largest weight changes, d extinction and non-extinction of conditioned inhibitors depend on the restriction of y in the reinforcement term, e is modeled as inhib. backward conditioning with long negative ISI, f only possible with DR + ±f + + + + + + + + ±f + + + + + + + + + + SOP + + + + + + + + + + + + + + ± + + CP + + + + + + + + + + + + + + + + + + + + + + CP-CP + + + + + + + + + + + + + + + + ± - CP+SB + + + + + + + + + + + + + + + + + + + + + + background and appropriate eligibility, g for long CSs, the learning result after multiple trials is the same as after one trial and h depends on stimulus overlaps and initial weights, with SOP eligibility additionally ISI-dependent. As seen, all the previously mentioned methods suffer from some serious limitations as can be seen in Table 4-7. None of American Journal of Psychology and Behavioral Sciences 2015; 2(2): 33-40 them can provide all features of CC. There maybe two possible reasons for this situation. One of them is lack of experimental data. The other one is, it may be because of deficiencies in theory of CC. This paper may be used by researchers to determine which model will be more suitable for their studies. 5. Conclusion This study set out with the aim of reviewing computational models of CC, and its applications. The CC, which explains human body as a living machine, is an important phenomenon in the human learning that explains basic behaviors, such as: fear, reflex, emotions and phobia. The CC learning stage is analyzed in various areas because the results of experiments provide a rich knowledge that can be used for studying basic learning, memory, and emotion processes. Basically, it was used for investigating consumer behavior, network security, robotic control, response extinction, conditional fear acquisition and extinction and obesity. As a summary, there is not any model that capable of explaining all outcomes of CC, which can be seen that the hardest questions in learning. As seen, all the previously mentioned methods suffer from some serious limitations as can be seen in Table 4-7. This paper may be used by researchers to determine which model will be more suitable for their studies. References [1] Pavlov, Ivan Petrovich. DoverPublications. com, 1927. [2] Martin Inderbitzin, Ivan Herreros-Alonso and Paul F.M.J. Verschure, “An Integrated Computational Model of the Two Phase Theory of Classical Conditioning”, IEEE, 2010 [3] Dugatkin, LA., “Principles of animal behavior” New York, New York, USA: WW Norton, 2004. [4] Raja L. Clouse, Soowon Kim, and Manjula B. Waldron, “An Adaptive Threshold Learning Algorithm for Classical Conditioning”, Proceedings - 19th International Conference IEEE/EMBS Oct. 30 - Nov. 2, 1997 Chicago, IL. USA, pg. 1380-1382. [5] [6] Conditioned reflexes. Salotti, Jean Marc, and Florent Lepretre. "Classical and Operant Conditioning as Roots of Interaction for Robots." Proceedings of the Workshop from Motor to Interaction Learning in Robots, Conference on Intelligent Robotics Systems, Nice. 2008. Malaka, Rainer, and Martin Hammer. "Real-time models of classical conditioning." Neural Networks, 1996, IEEE International Conference on. Vol. 2. IEEE, 1996. [7] Louwrence D Erasmus, “Classical Conditioning Implementation with RealNeuron Networks”, IEEE, 2007. [8] Gonzalo Miguez, Mario A. Laborda and Ralph R. Miller, “Classical conditioning and pain: Conditioned analgesia and hyperalgesia”, Acta Psychologica 145 (2014) 10–20. [9] Derick H. Lindquist, Greta Sokoloff, Joseph E. Steinmetz, “Ethanol-exposed neonatal rats are impaired as adults in 38 classical eyeblink conditioning at multiple unconditioned stimulus intensities”, Brain Research 1150 (2007) 155 – 166. [10] Christina Dalla, Tracey J. Shors, “Sex differences in learning processes of classical and operant conditioning”, Physiology & Behavior 97 (2009) 229–238. [11] Mirte Brom, Stephanie Both, Ellen Laan,Walter Everaerd, Philip Spinhoven, “The role of conditioning, learning and dopamine in sexual behavior: A narrative review of animal and human studies”, Neuroscience and Biobehavioral Reviews 38 (2014) 38–59. [12] Ole Åsli, Magne Arve Flaten, “Conditioned facilitation of the unconditioned reflex after classical eyeblink conditioning”, International Journal of Psychophysiology 67 (2008) 17–22. [13] Tomer Shechner, Melanie Hong, Jennifer C. Britton, Daniel S. Pine, Nathan A. Fox, “Fear conditioning and extinction across development: Evidence from human studies and animal models”, Biological Psychology 100 (2014) 1–12. [14] Bram Vervliet, Maarten Geens, “Fear generalization in humans: Impact of feature learning on conditioning and extinction”, Neurobiology of Learning and Memory 113 (2014) 143–148. [15] Masayuki Yoshida, Izumi Okamura, Kazumasa Uematsu, “Involvement of the cerebellum in classical fear conditioning in goldfish”, Behavioural Brain Research, 153 (2004) 143–148. [16] Kazuyuki Yamada, “Strain differences of selective attention in mice: Effect of Kamin blocking on classical fear conditioning”, Behavioural Brain Research 213 (2010) 126–129. [17] Thomas M. Ricarta, Xilu Jiao, Kevin C.H. Pang, Kevin D. Beck, Richard. J. Servatius, “Classical and instrumental conditioning of eyeblink responses inWistar–Kyoto and Sprague–Dawley rats”, Behavioural Brain Research 216 (2011) 414–418. [18] Miriam S. Nokia, Markku Penttonen, Tapani Korhonen, Jan Wikgren, “Hippocampal theta (3–8 Hz) activity during classical eyeblink conditioning in rabbits”, Neurobiology of Learning and Memory 90 (2008) 62–70. [19] Elisabetta Baldi, Corrado Bucherelli, “Entorhinal cortex contribution to contextual fear conditioning extinction and reconsolidation in rats”, Neurobiology of Learning and Memory 110 (2014) 64–71. [20] Andrew R. Delamater, R. Frederick Westbrook, “Psychological and neural mechanisms of experimental extinction: A selective review”, Neurobiology of Learning and Memory 108 (2014) 38–51. [21] Travis P. Todd, Drina Vurbic, Mark E. Bouton, “Behavioral and neurobiological mechanisms of extinction in Pavlovian and instrumental learning”, Neurobiology of Learning and Memory 108 (2014) 52–64. [22] M.E. Bouton, E.W. Moody, “Memory processes in classical conditioning”, Neuroscience and Biobehavioral Reviews 28 (2004) 663–674. [23] Luis Aguado, “Neuroscience of Pavlovian Conditioning: A Brief Review”, The Spanish Journal of Psychology , 2003, Vol. 6, No. 2, 155-167. [24] Jonathan D. Raybuck, K. Matthew Lattal, “Bridging the interval: Theory and neurobiology of trace conditioning”, Behavioural Processes 101 (2014) 103–111. 39 Mehmet Emin Tagluk and Omer Faruk Ertugrul: A Review of Computational Classical Conditioning Models [25] R. F. Thompson and J. E. Steinmetz, “The Role of the Cerebellum in Classical Conditioning of Discrete Behavioral Responses”, Neuroscience 162 (2009) 732–755. [40] Chris T. Allen, Thomas J. Madden, “A Closer Look at Classical Conditioning”, The Journal of Consumer Research, Vol. 12, No: 3, December 1985, pp. 301-315. [26] Germund Hesslow, Dan-Anders Jirenhed, Anders Rasmussen, Fredrik Johansson, “Classical conditioning of motor responses:What is the learning mechanism?”, Neural Networks 47 (2013) 81–87. [41] Elnora W. Stuart, Terence A. Shimp, Randall W. Enge, “Classical Conditioning of Consumer Attitudes: Four Experiments in an Advertising Context”, Journal of Consumer Research, Vol. 14, December 1987, pp. 334-349. [27] Stephen D. Berry, Loren C. Hoffmann, “Hippocampal theta-dependent eyeblink classical conditioning: Coordination of a distributed learning system”, Neurobiology of Learning and Memory 95 (2011) 185–189. [42] Chanthika Pornpitakpan, “A critical review of classical conditioning effects on consumer behavior”, Australasian Marketing Journal 20 (2012) 282–296. [28] Drew C. Thompson, Irving Z. Basanez, Douglas A. Baxter and John H. Byrne, “A Computational Study of Operant and Classical Conditioning in a Central Pattern Generator Neural Circuit”. [29] Blackburn, Michael R. "A general architecture for classical conditioning of perceptual-motor sequences." Neural Networks, 1990., 1990 IJCNN International Joint Conference on. IEEE, 1990. [30] Ahmed A. Moustafa, Catherine E. Myers, Mark A. Gluck, “A neurocomputational model of classical conditioning phenomena: A putative role for the hippocampal region in associative learning”, Brain Research 1276 (2009) 180 – 195. [31] Torsten Lehmann, “Classical Conditioning with Pulsed Integrated Neural Networks: Circuits and System”, IEEE Transactions on Circuits and Systems—II: Analog and Digital Signal Processing, 45(6), (1998), 720-728. [32] Lehmann, Torsten. "Eccopunch: the Edinburgh classical conditioning pulsed neural chip." Circuits and Systems, 1997. ISCAS'97., Proceedings of 1997 IEEE International Symposium on, 1997. [33] Magri, S., Pepe, I., Traverso, S., & Nicolini, C. (1993, October). A model of associative learning and classical conditioning in Aplysia californica: a preliminary study. In Neural Networks, 1993. IJCNN'93-Nagoya. Proceedings of 1993 International Joint Conference on Vol. 3, pp. 2694-2697, IEEE. [34] B. Yeşilyaprak, B. Aydın, M. Bilgin, G. Can, K. Ersanlı, M. Kılıç, İ. Kısaç, İ. Korkmaz, H. Küçükkaragöz, Ş. Küllahoğlu, B. Öztürk ve E. Uçar, Eğitim Psikolojisi Gelişim-Öğrenme-Öğretim, Ankara: Pegem Akademi, 2012. [35] R. L. Veznedaroğlu ve A. O. Özgür, “Öğrenme Stilleri: Tanımlamalar, Modeller ve İşlevleri”, İlköğretim Online, 4(2), (2005), pp. 1-16. [36] Olson, M. H., & Hergenhahn, B. R. (2009). An introduction to theories of learning. Pearson/Prentice Hall. [37] Garcia, J., Koelling, R. A. (1966). Relation of cue to consequences in avoidance learning. Psychonomic Science, 4, 123-124. [38] Kamin, L. J. (1968). Attention–like processes in classical conditioning. In Jones, M. R. (Ed.) Miami symposium on the prediction of behavior: aversive stimulation (pp. 9–31). Miami: University of Miami Press. [39] Rescorla, R. A. (1968). Probability of shock in the presence and absence of CS in fear conditioning. Journal of comparative physiological psychology, 66, 1–5. [43] Yi Fen Chen ve Jian Han Jiang, “Effects Of Classical and Operant Conditioning On Online Consumer Purchase and Repurchase Intention”, Effects Of Classical and Operant Conditioning On Online Consumer Purchase and Repurchase Intention. [44] Suping Liu, Yongsheng Ding, “A Classical Conditioning Model for Policy-based Management”, 2009 International Conference on Networks Security, Wireless Communications and Trusted Computing, DOI 10.1109/NSWCTC.2009.129, pp. 249-252. [45] Suping Liu, Yongsheng Ding, “An Adaptive Network Policy Management Framework Based on Classical Conditioning”, Proceedings of the 7th World Congress on Intelligent Control and Automation, June 25 - 27, 2008, Chongqing, China, pg. 3336-3340. [46] R. Prueckl, A.H. Taub, I. Herreros, R. Hogri, A. Magal, S.A. Bamford, A. Giovannucci, R. Ofek, Y. Shacham-Diamand, P.F.M.J. Verschure, M. Mintz, J. Scharinger, A. Silmon, C. Guger, “Behavioral Rehabilitation of the Eye Closure Reflex in Senescent Rats using a Real-Time Biosignal Acquisition System”, 33rd Annual International Conference of the IEEE EMBS, Boston, Massachusetts USA, August 30 - September 3, 2011, Pg. 4211-4214. [47] Guoshi Li, Gregory J. Quirk, and Satish S. Nair, “Modeling Acquisition and Extinction of Conditioned Fear in LA Neurons using Learning Algorithm”, Proceedings of the 2007 American Control Conference Marriott Marquis Hotel at Times Square New York City, USA, July 11-13, 2007, Pg. 552-557. [48] Susan Shi Yuan Li, Gavan P. McNally, “The conditions that promote fear learning: Prediction error and Pavlovian fear conditioning”, Neurobiology of Learning and Memory 108 (2014) 14–21. [49] Sven Benson, Joswin Kattoor, Jennifer S. Kullmann, Sarah Hofmann, Harald Engler, Michael Forsting, Elke R. Gizewski, Sigrid Elsenbruch, “Towards understanding sex differences in visceral pain: Enhanced reactivation of classically-conditioned fear in healthy women”, Neurobiology of Learning and Memory 109 (2014) 113–121. [50] Furze T. A., “The Application of Classical Conditioning to the Machine Learning of a Commonsense Knowledge of Visual Events”, PhD Thesis, The University of Leeds, School of Computing. [51] Elliot A. Ludvig, Marc G. Bellemare and Keir G. Pearson, “A Primer on Reinforcement Learning in the Brain: Psychological, Computational, and Neural Perspectives” in Eduardo Alonso, Esther Mondragón, “Computational Neuroscience for Advancing Artificial Intelligence: Models, Methods and Applications”, Medical Information Science Reference, Hershey • New York, 2011. American Journal of Psychology and Behavioral Sciences 2015; 2(2): 33-40 [52] Dawson M. R. W., “Connectionism and Classical Conditioning”, A peer-reviewed monograph published by Comparative Cognition and Behavior Reviews on behalf of the Comparative Cognition Society. 2008 doi: 10.3819/ccbr.2008.30008. [53] Schmajuk N., “Mechanisms in Classical Conditioning: A Computational Approach”, Cambridge University Press, The Edinburgh Building, Cambridge CB2 8RU, UK, 2010. [54] Schmajuk N., “Computational Models of Conditioning”, Cambridge University Press, The Edinburgh Building, Cambridge CB2 8RU, UK, 2010. [55] Balkenius, Christian, and Jan Morén. "Computational models of classical conditioning: a comparative study." Proceedings of the fifth international conference on simulation of adaptive behavior on From animals to animats. Vol. 5. 1998. [56] Daniel L. Chester, “A Comparison of some Neural Network Models of Classical Conditioning”, 1990 IEEE, pp.1163-1168. [57] A. C. Courville, N. D. Daw, G. J. Gordon ve D. S. Touretzky, “Model Uncertainty in Classical Conditioning”, Advances in neural information processing systems. 2003. [58] Samuel J. Gershman & Yael Niv, “Exploring a latent cause theory of classical conditioning”, Learn Behav (2012) 40:255–268 DOI 10.3758/s13420-012-0080-8. [59] Anja Austermann, Seiji Yamada, “Learning to Understand Multimodal Rewards for Human-Robot-Interaction using Hidden Markov Models and Classical Conditioning”, 2008 IEEE Congress on Evolutionary Computation (CEC 2008), Pg. 4096-4103. [60] Christopher Johansson ve Anders Lansner, “An Associative Neural Network Model of Classical Conditioning”, Report from Studies of Artificial Neural Systems, (SANS), Numerical Analysis and Computer Science (Nada), Royal Institute of Technology (KTH) , S-100 44 STOCKHOLM, Sweden. theory of adaptive networks: expectation prediction." Psychological review 88.2 (1981): 135. 40 and [64] Wagner, A.R. (1981) SOP: A model of automatic memory processing in animal behavior. In Spear, N.E. and R.R. Miller, editors, Informatzon Processing in Animals: Memory Mechaieisms, volume 85, chapter 1, pages 5-44. Erlbaum, New Jersey. [65] Sutton, R.S. and A .G. Barto (1990) Time-derivative models of Pavlovian reinforcement. In Gabriel, A. and J. Moore, editors, Learning and Computational Neuroscience: Foundations and Adaptive Networks, chapter 12, pages 497-537. MIT Press, Cambridge, Mass and London. [66] Klopf, A.H. (1989) Classical conditioning: Phenomena predicted by a drive-reinforcement model of neural function. In Byrne, J.H. and W.O. Berry, editors, Neural Models of Placticity: Experimental and Theoretical Approaches, chapter 6, pages 94-103. Academic Press. [67] Christian Balkenius, Jan Mor´en, “Dynamics of a Classical Conditioning Model”, Autonomous Robots 7, 41–56 (1999). [68] Malaka, R., R. Lange, and M. Hammer (1995) A constant prediction model for classical conditioning. In Elser, N. and R. Menzel, editors, Learning and Memory: Proceedings of the 23rd Gottingen Neurobiology Conference, volume 1, page 75. Thieme-Verlag, Stuttgart, New York. [69] Klopf, A. H. (1988). A neuronal model of classical conditioning. 16, 2, 85–125. [70] John Y. Cheung, David C. Chance, and Asa Lawton, “An Examination of Real-Time Neuronal Models in a Classical Conditioning Framework”, 1992 IEEE, Pg. III-504 – III-509. [71] Ertugrul, O. F., & Tagluk, M. E. (2014, April). Learning with classical conditioning. In Signal Processing and Communications Applications Conference (SIU), 2014 22nd (pp. 927-930). IEEE. doi: 10.1109/SIU.2014.6830382. [61] H. Hassan Ve M. Watan, “On Mathematical Analysis of Pavlovian Condltioning Learning Process Using Artificial Neural Network Model”, 10th Mediterranean Electrotechnical Conference, MEleCon 2000, Vol. II, Pg. 578-581. [72] Nestor A. Schmajuk and James J. DiCarlo, “Stimulus Configuration, Classical Conditioning, and Hippocampal Function”, Psychological Review, 1992, Vol. 99, No. 2,268-305. [62] Hiroyuki Ishii, Masaki Nakasuji, Motonori Ogura, Hiroyasu Miwa, Atsuo Takanishi, “Accelerating Rat’s Learning Speed Using a Robot - The robot autonomously shows rats its functions –”, Proceedings of the 2004 IEEE International Workshop on Robot and Human Interactive Communication, Kurashiki, Okayama Japan September 20-22,2004. [73] Edgar H. Vogel, Maria E. Castro, Maria A. Saavedra, “Quantitative models of Pavlovian conditioning”, Brain Research Bulletin 63 (2004) 173–202. [63] Sutton, Richard S., and Andrew G. Barto. "Toward a modern [74] Rainer Malaka, “Models of Classical Conditioning Models of Classical Conditioning”, Bulletin of Mathematical Biology (1999) 61, 33–83, Article No. bulm.1998.0074.
© Copyright 2026 Paperzz