J Neurophysiol 93: 2279 –2293, 2005; doi:10.1152/jn.01042.2004. Object Motion Computation for the Initiation of Smooth Pursuit Eye Movements in Humans Julian M. Wallace,1 Leland S. Stone,2 and Guillaume S. Masson1 1 Institut de Neurosciences Cognitives de la Méditerranée, Centre National de la Recherche Scientifique, Unité Mixte de Recherche 6193, Marseille France; 2Human Information Processing Research Branch, Human Factors Research and Technology Division, National Aeronautics and Space Administration, Ames Research Center, Moffett Field, California Submitted 4 October 2004; accepted in final form 8 November 2004 INTRODUCTION When an object moves across the visual field, smooth pursuit eye movements act to stabilize its image on the retina by smoothly rotating the line of sight at the same speed and direction as the object. The signal driving the initiation of smooth pursuit eye movements is dominated by the visual motion of the target (Rashbass 1961). Reconstructing the target’s motion is a complex task because of the aperture problem (Wallach 1996): early stages of visual motion processing sense local changes in the image through detectors with a limited field of view and spatiotemporal properties such that they respond best to motion orthogonal to the local onedimensional (1D) edge (Albright 1984). As individual 1D motion measurements are ambiguous and can be different from Address for reprint requests and other correspondence: G. S. Masson, Team DyVA, INCM–CNRS UMR-6193, 31 Chemin Joseph Aiguier, 13402 Marseille, France (E-mail: [email protected]). www.jn.org the actual two-dimensional (2D) object motion, the latter is reconstructed by combining these piecewise local signals. Several simple “two-stage” computational solutions have been proposed to solve the aperture problem and compute the 2D object motion. Vector averaging (VA) of the various 1D motions associated with each 1D edge of the object is one simple way of combining the 1D information. However, this approach does not always yield the correct target motion (Ferrera and Wilson 1990). The Intersection of Constraints (IOC) rule (Adelson and Movshon 1982; Fennema and Thompson 1970) computes the true translation vector in velocity space as the intersection of the different constraint lines formed by the 1D edge motions. Finally, feature tracking (FT) strategies can also yield the correct translation vector by considering the velocity of unchanging features of the object (such as corners, line-ends, texture elements. . .) the luminance profile of which is intrinsically 2D and therefore the 2D motion of which is unambiguous (e.g., Gorea and Lorenceau 1991; Lorenceau et al. 1993; Mingolla et al. 1992). It remains unresolved whether any of these three rules best describes the neural 2D motion computation underlying human perception or behavioral performance. Indeed, more recently probabilistic computational theories have been put forward to account for motion integration (e.g., Koechlin et al. 1999; Rao 2004; Weiss et al. 2002). We have recently demonstrated that the initiation of pursuit depends on the shape of simple, line-drawing diamonds moving along horizontal or vertical trajectories (Masson and Stone 2002). In humans, pursuit was always initiated in the direction predicted by the vector average of the four motion signals orthogonal to the four diamond edges. More specifically, when this VA prediction and the actual object motion differed by 44°, tracking direction slowly evolved from the oblique VA prediction to the cardinal veridical object-motion direction (e.g., IOC or FT prediction) over a period of 200 ms. Furthermore, once steady-state tracking direction matched the actual object-motion direction, transiently (100 ms) blanking the object had no impact on the eye-movement direction, indicating that the aperture problem was solved visually once at target motion onset and that sensorimotor feedback was not required for pursuit to converge onto the correct target motion. Previous studies documented the same basic finding for human motion perception and monkey pursuit and neural responses. Using The costs of publication of this article were defrayed in part by the payment of page charges. The article must therefore be hereby marked “advertisement” in accordance with 18 U.S.C. Section 1734 solely to indicate this fact. 0022-3077/05 $8.00 Copyright © 2005 The American Physiological Society 2279 Downloaded from http://jn.physiology.org/ by 10.220.32.247 on June 15, 2017 Wallace, Julian M., Leland S. Stone, and Guillaume S. Masson. Object motion computation for the initiation of smooth pursuit eye movements in humans. J Neurophysiol 93: 2279 –2293, 2005; doi: 10.1152/jn.01042.2004. Pursuing an object with smooth eye movements requires an accurate estimate of its two-dimensional (2D) trajectory. This 2D motion computation requires that different local motion measurements are extracted and combined to recover the global object-motion direction and speed. Several combination rules have been proposed such as vector averaging (VA), intersection of constraints (IOC), or 2D feature tracking (2DFT). To examine this computation, we investigated the time course of smooth pursuit eye movements driven by simple objects of different shapes. For type II diamond (where the direction of true object motion is dramatically different from the vector average of the 1-dimensional edge motions, i.e., VA ⫽ IOC ⫽ 2DFT), the ocular tracking is initiated in the vector average direction. Over a period of less than 300 ms, the eye-tracking direction converges on the true object motion. The reduction of the tracking error starts before the closing of the oculomotor loop. For type I diamonds (where the direction of true object motion is identical to the vector average direction, i.e., VA ⫽ IOC ⫽ 2DFT), there is no such bias. We quantified this effect by calculating the direction error between responses to types I and II and measuring its maximum value and time constant. At low contrast and high speeds, the initial bias in tracking direction is larger and takes longer to converge onto the actual object-motion direction. This effect is attenuated with the introduction of more 2D information to the extent that it was totally obliterated with a texture-filled type II diamond. These results suggest a flexible 2D computation for motion integration, which combines all available one-dimensional (edge) and 2D (feature) motion information to refine the estimate of object-motion direction over time. 2280 J. M. WALLACE, L. S. STONE, AND G. S. MASSON METHODS We used the same behavioral paradigm and the same basic stimuli as Masson and Stone (2002), but we have now investigated a full set of parameters such as speed, contrast, and relative weight of 1D and 2D features to better understand the 2D object motion computation underlying smooth pursuit initiation. Visual stimuli and behavioral paradigm The stimuli were line-figures (line luminance: 60cd/m2) of equal area (71deg2) but of different shapes and tilts (Beutter and Stone 2000), moving either in horizontal (rightward or leftward) or vertical (upward or downward) directions at different speeds over a black (⬍0.1cd/m2) untextured background. The two shapes were a square diamond (main axes: 11.9°) and a counter-clockwise (CCW) elonJ Neurophysiol • VOL gated and tilted diamond (28.3 ⫻ 5° parallelogram, tilted ⫾45° with respect to vertical with internal angles of 14.5 and 165.5°). These square and elongated diamonds have properties equivalent to those of type I and II moving plaids, respectively (Wilson et al. 1992). For the square diamond, the vector average of the edge motions (i.e., the motions orthogonal to the edges) is co-linear with the object-motion direction. For the elongated diamonds, it is biased ⫾44° away from the object’s direction (see Beutter and Stone 2000). Several parameters of the visual motion stimuli were changed for both square and tilted diamonds. For each parameter, corresponding type I and II visual stimuli were interleaved. In the first experiment, the speed of object motion was varied from 2.5 to 20°/s. In the second experiment, the luminance of the edges was changed. For this sole experiment, objects were projected over a gray background of constant luminance so that object contrast was varied from 10 to 90%. The next two experiments were designed to probe the role of 1D (edges) and 2D (corners) motion cues. In the third experiment, the relative salience of edges, corners, and line endings was manipulated applying a circular Gaussian filter ( ⫽ 5 pixels, for square edges; ⫽ 12.5 pixels, for shorter diamond edge; ⫽ 16 pixels, for longer diamond edge) centered either on the center of the edges or on the corners themselves. In the fourth experiment, the line figures were filled with a texture made of a random dot pattern with 50% dot density. In a control experiment, we checked for a possible role of expectation on motion computation by statically presenting either a square or a tilted diamond for 0, 50, 100, 200, or 400 ms before motion onset. Observers had their head stabilized by chin and forehead rests. Each trial started with the presentation of a stationary fixation point. Observers were required to fixate for 300 ⫾ 150 ms within a 1 ⫻ 1° window. The fixation point was then extinguished and a moving object presented. In the control experiment, the object was presented as static for a given duration together with the fixation point. The fixation point was then extinguished as the object was set into motion. In all cases, the object moved for 450 ⫾ 50 ms. Observers were instructed to track the object center and trials were aborted if eye position did not stay within 2° of the object center (⬃2% of trials, independent of condition). All conditions were randomly interleaved to minimize cognitive expectations and anticipatory pursuit. We collected ⬃60 trials per condition and observer over several days. Eye-movement recordings and analysis Eye movements were recorded from three observers (1 naı̈ve) using methods described in detail elsewhere (Masson et al. 2000). Briefly, a PC running REX controlled stimulus presentation and data acquisition. Stimuli were generated with an SGI Octane workstation and back-projected along with the red fixation point onto a large translucent screen (80 ⫻ 60°) using a video-projector (1,280 ⫻ 1,024 pixels at 76 Hz). The position of the right eye was sampled at 1 kHz using the scleral search-coil technique (Collewijn et al. 1975). Data were then stored on disk for off-line analysis. Eye-position data were linearized, smoothed with a spline interpolation (Busettini et al. 1991) and then differentiated to obtain vertical and horizontal eye-velocity profiles. After visual inspection using an interactive graphic software, we used a conjoint velocity and acceleration threshold to detect and remove saccades (Krauzlis and Miles 1996). The latency of each trial was computed for both horizontal and vertical eye-velocity profiles using an objective method (Krauzlis and Miles 1996; Masson and Castet 2002). Trials were then realigned so that time 0 was now response onset (Fig. 1A). To compute tracking direction, mean vertical and horizontal eye positions were computed over 40-ms bins, regularly spaced from 0 (response onset) to 320 ms (steady-state tracking). The change in eye position between successive bins was then computed for each trial, for both type I and II stimuli. Thus we generated new time bins representing the mean direction over an 80-ms period. For each object motion 93 • APRIL 2005 • www.jn.org Downloaded from http://jn.physiology.org/ by 10.220.32.247 on June 15, 2017 oblique moving lines, Pack and Born (2001) found that monkeys initiate ocular tracking in the direction orthogonal to the lines. Lorenceau and colleagues (Castet et al. 1993; Lorenceau et al. 1993) found that humans manifest similar biases in motion direction and speed estimation for short presentation of a single moving tilted line. Moreover, the alignment of the initial voluntary pursuit with the VA prediction is consistent with our previous work on ocular-following responses in humans. We showed that reflexive tracking is always initiated in the VA direction of all 1D motion signals, albeit with a latency of ⬃85 ms (shorter than the usual 100-ms latency of voluntary pursuit) but that the later portion of the ocular responses is influenced by a number of factors including the presence of unambiguous 2D motion signals moving in the actual patternmotion direction (Masson and Castet 2002; Masson et al. 2000). Our preliminary brief report (Masson and Stone 2002) about pursuit of line-drawing objects did not fully resolve a number of important questions. First, our earlier estimate of the time course of pursuit direction was ⬃90 ms, which is close to the internal latency of the visuomotor feedback loop (Goldreich et al. 1992), suggesting that negative feedback loop might still be playing a role in reducing the initial tracking error. Herein, our goal was to address this question more closely by offering a better quantification of the time course of pursuit direction and by examining it over a large range of stimulus parameters. Second, we previously observed the initial VA bias in pursuit initiation only using very slow speeds and high contrast. Biases in perceived direction were observed under similar conditions but not with high speed and strong contrast (Lorenceau et al. 1993). By varying target speed and contrast, we could determine whether or not the initiation of pursuit follows the same pattern and therefore shares the same solution to the aperture problem as motion perception, as has been shown for steadystate tracking (Beutter and Stone 2000; Stone et al. 2000). Third, we examined the role of local 1D edges and 2D features (corners) by manipulating their relative salience in the images and measuring the effect on the initial pursuit direction error and its time course of recovery. An additional goal of this last experiment was to examine to what extent our findings could be extended to more complex objects, such as textured surfaces where multiple motion cues coexist that could improve the accuracy of visual motion signal driving the initiation of pursuit. 2D OBJECT MOTION FOR PURSUIT EYE MOVEMENTS 2281 direction, the coordinate system was rotated so that object-motion direction was equal to zero and the vector average prediction was of ⫹44°. Thus tracking directions could be averaged across different object motion directions. Figure 2A plots these directions, in bins spaced 40 ms apart centered from 40 to 280 ms (i.e., the 1st bin centered on 40 ms represents direction from 0 to 80 ms, the final bin centered in 280 ms represents the direction from 240 to 320 ms). For each bin, tracking direction errors (⌰e) were computed by subtracting the mean tracking direction obtained for type I diamonds from that obtained for type II diamonds (Fig. 2B). This estimate reflects the biased tracking direction of responses to tilted diamonds and the relationship between time and tracking direction errors reflect the temporal dynamics of smooth eye movements to 2D object motion. This relationship was then fit with the sum of two exponential functions using the following formula ⌰ e ⫽ A关exp共t/i兲 ⫹ exp共 ⫺ t/d兲兴 ⫹ B where i and d are time constants for the early (growing) and late (decaying) exponential functions, respectively (Fig. 2C). The goodness of fit was estimated by the r2 values, which ranged between 0.61 and 0.99 [0.9562 ⫾ 0.0607 (SD)] and were always significant (P ⬍ 0.05). The estimated peak of the function indicates the maximal deviation of the initial ocular response. The d and B parameters give a precise estimate of the time course over which this initial tracking direction error converges toward the true object-motion direction and of the residual error once steady-state pursuit is reached. These parameters were then compared across conditions to investigate the dependence of the temporal dynamics of ocular responses on visual stimulus parameters such as speed, contrast, and so on. Of particular interest is the tracking direction error at the point in time equal to the mean estimate of the vertical and horizontal latencies (dashed vertical line). This value indicates by how much the initial tracking bias has been reduced before the closing of the oculomotor control loop. J Neurophysiol • VOL FIG. 2. Computation of tracking direction errors. A: relationship between mean tracking directions for a rightward motion of either type I (F) or type II (E) diamonds. Tracking directions are computed over 40-ms bins. - - -, the true object-motion direction (0°) and the vector average prediction (44°) for type II diamonds. B: tracking direction errors are computed by subtracting the tracking directions observed for the type I to the type II diamonds. C: the average tracking direction errors across 4 motion directions are fitted with an doubleexponential function. This descriptive function captures the sharp initial increase in the tracking direction error, followed by a regular decay toward the asymptotic value corresponding to the object motion direction. Parameters of interest are the peak direction error and its time of occurrence as well as the decay time constant (). 93 • APRIL 2005 • www.jn.org Downloaded from http://jn.physiology.org/ by 10.220.32.247 on June 15, 2017 FIG. 1. Temporal dynamics of pursuit eye movements. A: horizontal (ėh) and vertical (ėv) velocity profiles of smooth pursuit responses to type I and II moving diamonds. Eye movements have been realigned relative to pursuit onset. Rightward (leftward) type II moving diamond elicited large transient vertical components in the upward (downward) direction. B: distribution of tracking direction errors across trials for 3 different time bins and each type of diamond. Vertical dotted lines indicate the object motion direction (defined as 0°) and the vector average direction (⫹44° with respect to the true object-motion direction). 2282 J. M. WALLACE, L. S. STONE, AND G. S. MASSON RESULTS Figure 1A illustrates the horizontal and vertical eye velocity profiles of smooth pursuit responses to either type I or II moving diamonds, realigned to eye-movement onset. As previously reported by Masson and Stone (2002), initial tracking of type II diamonds is along an oblique direction, as evidenced by the presence of an initial eye acceleration in both horizontal and vertical directions (Fig. 1A, right). A transient vertical component reached its peak ⬃150 ms after pursuit onset and then progressively declined to 0 in ⬃150 ms. Thus ⬃300 ms after pursuit onset (⬃420 ms after target motion onset), the tracking direction was now aligned with the pure object-motion direction. This transient vertical component was absent in responses to type I diamonds (Fig. 1A, left). To examine this phenomenon more thoroughly than in our preliminary report, the tracking error analysis was done on a trial-by-trial basis to determine how consistent the initial bias was across trials. Figure 1B plots the distributions of tracking direction for three different time bins spanning the time period of the phenomeJ Neurophysiol • VOL non (see METHODS). It is evident that for type I diamonds (dark bars), initial tracking is aligned with object-motion direction. Over the initial eye acceleration period, the distribution narrowed, but the average tracking direction did not change. For the type II diamond (white bars), the whole initial distribution is shifted toward the VA prediction. As tracking becomes more reliable across trials, the whole distribution progressively narrows and shifts toward the true object-motion direction (0°). Overall, it is evident that the initial responses were consistent across trials. In particular, bimodal distributions were never observed for any of our three observers. Interestingly, the variance of the initial distributions for type I and II were similarly broad, although the type II variance appears somewhat larger for the early bins. Figure 2A plots the time course of the tracking direction averaged across trials (see METHODS). The mean tracking direction stayed close to the true object-motion direction (0°) for the type I diamond (closed symbols). However, for the Type II diamond the initial tracking direction grew to 93 • APRIL 2005 • www.jn.org Downloaded from http://jn.physiology.org/ by 10.220.32.247 on June 15, 2017 FIG. 3. Effects of object speed. A: for each subject, mean horizontal (ėh) and vertical (ėv) velocity profiles of ocular tracking to type II diamond drifting rightward at 5 different speeds indicated by the numbers at the right end of the curves. B, top: mean tracking direction errors, across 4 motion directions, for a 20°/s motion stimulus, as a function of time. —, best-fitting double-exponential functions. Bottom: the same data obtained with different target speeds. - - -, the mean pursuit latency, estimating the time at which the oculomotor loop closes. Thus the yellow shaded area correspond to the open-loop period of the responses. 2D OBJECT MOTION FOR PURSUIT EYE MOVEMENTS TABLE 2283 1. Dependence of pursuit direction on object-motion speed Speed, °/s GM 2.5 5 10 15 20 Error max, ° T peak 34 32 40 41 41 Latency, ms 62 70 70 67 68 126 128 124 121 119 , ms 62 71 70 68 69 38 ⫾ 4 26 25 37 38 32 67 ⫾ 3 46 51 56 55 50 123 ⫾ 3 122 123 119 116 115 68 ⫾ 3 46 51 56 55 50 2.5 5 10 15 20 32.6 ⫾ 6 29 33 38 38 39 51.6 ⫾ 4 104 98 78 78 76 119 ⫾ 3.5 127 118 118 117 114 52 ⫾ 4 104 41 78 77 33 Mean 35.4 ⫾ 4.3 86.8 ⫾ 13.2 118.8 ⫾ 4.4 66.6 ⫾ 29.2 Average 34.9 ⫾ 5.2 68.6 ⫾ 16.7 120.6 ⫾ 4.5 62.1 ⫾ 17.6 29.73 ⫾ 3.76 30.26 ⫾ 4.22 38.06 ⫾ 1.87 38.92 ⫾ 1.94 37.48 ⫾ 4.47 70.67 ⫾ 29.96 73.00 ⫾ 23.64 68.00 ⫾ 11.14 66.67 ⫾ 11.50 64.67 ⫾ 13.32 124.82 ⫾ 2.69 122.97 ⫾ 4.66 120.37 ⫾ 3.00 118.06 ⫾ 2.74 115.94 ⫾ 2.56 70.96 ⫾ 29.72 54.17 ⫾ 14.97 68.18 ⫾ 11.01 66.92 ⫾ 11.16 50.68 ⫾ 18.06 Mean IV Mean 2.5 5 10 15 20 ⬃30° before slowly decaying (open symbols). Tracking directions for type I and II stimuli were not distinguishable ⬃250 ms after pursuit onset. Figure 2B plots the tracking direction errors as computed by subtracting results of type I from type II for a given object direction. In Fig. 2C, the data points are now the tracking error averaged across the four possible object directions, and — shows the best-fitting double exponential function (see METHODS). This plot shows that the tracking direction error reached a maximum ⬃80 ms after pursuit onset and then decayed over ⬃200 ms to reach a value where eye and object trajectories appropriately coincide. In the three subjects, we ran a control experiment to check if expectation or anticipation could explain these results. We ran an experiment where the same stimuli were first presented as stationary objects (for a number of different durations), and then set into motion for 450 ms. We found that static presentation had no impact whatsoever on pursuit initiation. With the same three subjects, we then conducted a set of experiments to investigate the effect of varying a number of visual parameters on the initial pursuit response. Dependence on object-motion speed Figure 3A shows, for each subject, the vertical and horizontal velocity profiles obtained with type II objects moving rightward with speeds ranging from 2.5 to 20°/s. Increasing the speed of the object motion resulted in an increase of the initial, horizontal eye velocity. The initial transient in the vertical direction was similarly scaled as target speed increased, and, for all target speeds, this vertical component was reduced to 0 after ⬃300 ms of pursuit. Figure 3B, top, illustrates for each J Neurophysiol • VOL subject the relationship between the mean tracking error (across trials and object-motion directions) and time after pursuit onset, for 20°/s object speed. The - - - indicates the mean latency values of the responses, and therefore the yellow shaded areas plot observer-specific estimates of the open-loop period of pursuit. One can see that the peak of the tracking error and associated beginning of the exponential decay occurred before the closing of the oculomotor feedback loop. Indeed, the mean peak error was reached at 68.6 ⫾ 16.7 ms (⫾SD across subjects), while the mean latency (a measure of the open-loop duration) (see Lisberger and Westbrook 1985) was 120 ⫾ 4.3 ms. Figure 3B, bottom, shows, for all three subjects, the time course of their mean tracking direction error for all object speeds tested. It is evident that the maxima of the best-fitting functions occur at the same point in time and that the decay rates of the error were similar across speeds. However, the two slowest target speeds (2.5 and 5°/s) did yield smaller maximum direction errors. Table 1 summarizes for each subject the best-fitting parameters of the double exponential functions illustrated in Fig. 3 by the continuous lines. The mean peak direction errors ranged from 29.7 ⫾ 3.8° (at 2.5°/s) to 38.9 ⫾ 1.9° (at 15°/s) with a tendency toward larger errors for faster object motions (⬎10°/s). A one-way repeated-measures ANOVA showed a significant effect of target speed on peak direction error [F(2,4) ⫽ 15.22, P ⬍ 0.008]. Figure 4A shows this relationship for each subject, where increasing the target speed between 2.5 and 10°/s yielded larger initial bias, whereas no clear increase was observed for speeds ⬎10°/s. At these higher speeds, the peak bias was, on average, ⬃39°, which is only 5° away from the vector average prediction (44°). However, the time to peak for 93 • APRIL 2005 • www.jn.org Downloaded from http://jn.physiology.org/ by 10.220.32.247 on June 15, 2017 2.5 5 10 15 20 Mean JW 2284 J. M. WALLACE, L. S. STONE, AND G. S. MASSON the direction error was found to be essentially constant across all target speeds, ranging from 65 ⫾ 13 to 73 ⫾ 24 ms. Therefore the correction of the direction error is already well underway before the end of the open-loop period with the magnitude of the correction during this period ranging from ⬃30% at the lowest speed tested to ⬃17% at the highest speed. The decay of the tracking error was not affected by the target speed. Figure 4B plots the time constants of the tracking error decay for each speed and each subject. No systematic trend was found over the investigated range of target speeds [1-way repeated-measures ANOVA, F(2,4) ⫽ 0.75; P ⫽ 0.59]. Thus the time course of the decaying initial tracking error appears to be rather independent of target speed, averaging from 54 ⫾ 15 (at 5°/s) to 68 ⫾ 11 ms (at 10°/s). In brief, our results show that target speed has a limited, but significant, impact on the magnitude of the initial tracking error but no effect on its temporal dynamics. While object speed did not modulate the dynamics of the effect, the fact that the reduction of the initial error began before the closing of the visuo-motor control loop was evident across all speeds tested. Effects of stimulus contrast Next, we investigated the effect of target contrast on pursuit initiation. Figure 5 shows the results obtained in the same three J Neurophysiol • VOL 93 • APRIL 2005 • www.jn.org Downloaded from http://jn.physiology.org/ by 10.220.32.247 on June 15, 2017 FIG. 4. Effects of object speed on pursuit dynamics. A: best-fitting estimates of peak direction errors, as a function of object-motion speed, for each subject. - - -, the vector average prediction, 44° away from the actual objectmotion direction. B: for the same 3 subjects, the best-fitting decay time constants are plotted against object motion speed. observers in the same format as Fig. 3. For all target contrast values, the initial transient onset of the vertical component can be seen, albeit with different amplitudes across subjects. As we used a relatively slow target (10°/s), the amplitude of this transient component was particularly small in subject JW but consistent with the results plotted in Fig. 3A, middle. For all subjects, at the lowest contrast, initial horizontal eye acceleration was reduced, the transient vertical eye onset was larger (producing a larger deviation of the initial tracking direction toward the oblique axis), and the initial tracking error took longer to subside. Furthermore, horizontal eye velocity during steady-state tracking was also slightly reduced. Figure 5A, top, illustrates the mean tracking direction errors as a function of time for the highest contrast target (90%). Results were identical to those found in the previous experiments (compare with Fig. 3B). The bottom panels plot the same relationship, but now for all five contrasts tested. The light blue curves illustrate the time course of the tracking error for the lowest contrast (10%). It is evident that this lower contrast object produced a larger peak tracking direction error and a slower decay toward the true 2D object trajectory for all three observers. On average, across target-motion directions, this initial tracking error was not fully resolved after 300 ms of pursuit, indicating a large residual mis-estimation of the actual 2D target trajectory. For all other contrasts, the initial tracking error decays with a faster time constant, and it was resolved ⬃200 ms after pursuit onset. The - - - indicates the mean pursuit latency across contrast values, as no significant effect of target contrast was found on the response onset (see Table 2). The yellow shaded area thus indicates the mean open-loop period (106 ⫾ 5 ms, averaged across subjects and conditions). The latencies are somewhat lower than the previous experiment, suggesting an effect of the mean luminance (gray background) on pursuit. For all contrasts tested, the direction error begins to decay before the end of the open-loop period. Figure 6A shows the raw direction error computed in three time bins as a function of contrast for each of the three observers. The three time bins span a range from before and after the maximum error and before and after the end of the open-loop period. The error decreases as contrast is increased for all three time bins. This effect is similar in the two earlier bins and somewhat larger in the latest bin. Figure 6, B and C, summarizes the relationship between the two main parameters of interest (the peak direction error and the time constant of error decay) of the double-exponential functions and target contrast for each subject. Corresponding values (with the other estimated parameters) are given in Table 2, together with mean values across subjects. Figure 6B shows that for each subject, increasing target contrast resulted in a smaller peak direction error. At very low contrast (10%), the initial tracking direction perfectly matched the vector average prediction (44°, - - -) for each subject (mean across subjects: 43 ⫾ 3°). The initial tracking bias was reduced to ⬃32 ⫾ 2° when increasing the target contrast from 10 to 40%. Further increase in contrast did not lead to obvious changes in the initial tracking direction, which reached an asymptote around 30°. A one-way repeatedmeasures ANOVA indicated that the relationship between peak direction error and contrast was highly significant [F(2,4) ⫽ 37.94; P ⬍ 0.0001]. 2D OBJECT MOTION FOR PURSUIT EYE MOVEMENTS 2285 Unlike target speed, target contrast had some impact on the temporal dynamics of the direction error correction, as shown in Fig. 6C. By the end of the open-loop period, the initial error was reduced by only 8.5% for the lowest contrast condition but by 32 ⫾ 1% for the higher contrast conditions, suggesting an increase in the time to correction for the lowest contrast needed. For all three subjects, increasing target contrast elicited a shortening of the decay time constant from 168 ⫾ 84 ms (at 10% contrast) to an asymptotic value ⬃60 ms for contrasts ⬎40%. This relationship was highly significant [1-way repeatedmeasures ANOVA, F(2,4) ⫽ 9.141; P ⬍ 0.0044]. Overall, this experiment indicated that at 10% target contrast, the initial tracking direction is aligned with the VA prediction and a longer time is required (compared with higher contrast conditions) to reduce this larger initial error and to align the eye and object-motion trajectories. In comparison, at higher contrast values both the initial error and time to correction are reduced, increasing the contrast from 10 to 53% yielded a 26% decrease in the peak direction error and a 280% decrease in the decay time constant. J Neurophysiol • VOL Relative contribution of edges and corners Changing either target speed or contrast has a profound effect on the reliability of local motion, either 1D (edges) or 2D (corners). To further probe the relative contribution of these different types of local motion cues, we filtered the object image with circular Gaussian filters centered either on the corners or on the center of each of the four edges. Centering the filter onto the edges decreases the contribution of 1D edge cues and also adds eight new smoothed, line-ending features. Therefore in this “filter on edge” condition there are in total 12 line-ending features (the 4 corners plus the 8 newly introduced endings). Centering the filters on the corners removes the corners and introduces eight new smoothed, line-endings features at each end of the four disconnected edges. Therefore in this “filter on corner ” condition there are in total eight line-ending features. Introducing new 2D “intrinsic” features moving in the true object motion direction by filtering corners is somewhat similar to the effect of masking corners with invisible occluders parallel to the object trajectory (Shiffrar et 93 • APRIL 2005 • www.jn.org Downloaded from http://jn.physiology.org/ by 10.220.32.247 on June 15, 2017 FIG. 5. Smooth pursuit responses to a 10°/s object motion: effects of object contrast. Layout is the same as for Fig. 2. A: mean velocity profiles of pursuit eye movements driven by a type II diamond drifting rightward for a range of object contrasts. For this experiment, the background was set to a gray level corresponding to the mean luminance level. B: mean tracking direction errors as a function of time for a single contrast condition (90%, top) and for a range of object contrasts (bottom). The same color code is used for A and B. 2286 TABLE GM JW IV 2. Dependence of pursuit direction on object contrast C, % Error max, ° T peak, ms Latency, ms , ms 10 30 53 72 90 10 30 53 72 90 10 30 53 72 90 10 30 53 72 90 46 39 34 33 32 43 34 32 29 33 41 31 30 29 31 43.0 ⫾ 2.6 34.6 ⫾ 4.3 32.0 ⫾ 2.3 30.5 ⫾ 2.4 32.0 ⫾ 1.0 76 55 52 50 51 63 45 41 40 39 106 83 79 72 72 81.7 ⫾ 22.1 61.0 ⫾ 19.7 57.3 ⫾ 19.6 54.0 ⫾ 16.4 54.0 ⫾ 16.7 114 109 107 105 104 101 101 100 99 99 112 109 108 107 106 109.23 ⫾ 7.03 106.56 ⫾ 4.66 104.98 ⫾ 4.15 103.83 ⫾ 3.79 103.00 ⫾ 3.61 156 57 52 50 52 91 45 41 41 40 257 148 85 92 94 168.0 ⫾ 83.8 83.7 ⫾ 56.3 59.6 ⫾ 22.8 61.0 ⫾ 26.9 62.0 ⫾ 28.3 al., 1995). Filtered objects were interleaved with unfiltered objects, for both type I and type II stimuli. In the “no filter” condition as in the previous experiments, there are only four line-ending features (corresponding to the corners of the object). Figure 7A illustrates the horizontal and vertical velocity profiles obtained for each subject with each type of stimulus, always moving rightward at 10°/s. With the unfiltered type II diamond, a large initial, transient, vertical eye acceleration was again found, replicating the results from experiments 1 and 2. When filters were applied to the centers of the edges (filter on edge, green curve), the initial vertical transient was greatly reduced in at least two subjects, for this particular object motion direction. When filters were applied to the corners (filter on corner condition, blue curves), the initial transient vertical response was reduced relative to the no filter condition. Figure 7B plots the mean tracking direction error for the unfiltered condition only (top) or for all three conditions together (bottom). It can be seen that peak direction error was largely reduced in the filter on edge condition and arrived earlier in the temporal course. The initial tracking error was then reduced to 0 within ⬃200 ms of pursuit, that is ⬃100 ms earlier than in the fully visible condition. The results from the filter on corner conditions were intermediate between the two other condition, both in terms of peak direction error and time course. Table 3 summarizes the best-fitting parameters obtained for each subject and each image filtering condition. It is worth noting that smooth pursuit eye movements were initiated in response to each image type with the usual latencies (121.8 ⫾ 7.4 ms on average), and again, the reduction of the initial error was initiated well before the closing of the loop (58.6 ⫾16.1 ms on average). Pairwise comparison revealed no significant difference between image types. Mean values of the parameters of interest are plotted in Fig. 8. For the unfiltered condition, mean peak direction error was of 33.3 ⫾ 2.5°, a value similar to those found in the first two experiments. Removing edge information (and thus adding unambiguous 2D features) significantly reduced this peak direction error to 12.7 ⫾ 2.3° [t-test, t(4) ⫽ 10.47; P ⬍ 0.0004]. By comparison, adding 2D J Neurophysiol • VOL features alone induced a smaller, albeit significant reduction in the peak direction error [mean : 22.7 ⫾ 3.1°, t-test, t(4) ⫽ 4.7, P ⫽ 0.01]) as compared with the no filter condition. By the closing of the loop, these peak errors were already reduced by 20.17% (no filter), 65.37% (filter on edge), and 47.67% (filter on corners). The mean time constants of the exponential decay of the tracking direction error showed similar differences to those of the peak direction error. With the unfiltered stimulus, the mean was of 72 ⫾ 17 ms, similar to those reported above. The shortest values were found in the filter on edge condition [mean: 49 ⫾ 13 ms, t-test, t(4) ⫽ 12.4; P ⬍ 0.002] as compared with no filter. The filter on corner condition elicited an intermediate time course [mean: 56 ⫾ 12 ms, t-test, t(4) ⫽ 5.98; P ⬍ 0.003], as compared with the unfiltered condition. Pursuing texture-filled objects In this experiment, we manipulated the proportion of 1D (edges) and 2D (features) motion signals by filling the object with a high-density random pattern of small (0.2 ⫻ 0.2°) dots. For comparison, these textured diamonds were interleaved with the line-drawing figures and with a solid gray rhombus of same geometry. This latter filled stimulus controls for the effect of FIG. 6. Effects of object contrast on pursuit dynamics. A: relationships between mean (across direction) tracking direction errors and target contrast, for 3 different time bins and each subject. B: best-fitting peak direction errors, as a function of object contrast, for each subject. - - -, the vector average prediction, 44° away from the actual object-motion direction. C: for the same 3 subjects, the best-fitting decay time constants are plotted against contrast. 93 • APRIL 2005 • www.jn.org Downloaded from http://jn.physiology.org/ by 10.220.32.247 on June 15, 2017 Mean J. M. WALLACE, L. S. STONE, AND G. S. MASSON 2D OBJECT MOTION FOR PURSUIT EYE MOVEMENTS 2287 having luminance information distributed within the object. Figure 9A illustrates, for the three subjects, the vertical and horizontal mean velocity profiles obtained with each stimulus, for 10°/s rightward motion. Empty and uniformly filled objects yielded identical ocular responses both for the regular horizontal component and for the transient vertical component. However, when filled with a random pattern texture, the same object gave no transient component along the vertical direction and brisker initial eye acceleration along the horizontal direction. Figure 9B, top, TABLE 3. Effects of Gaussian filtering on pursuit dynamics Filters GM No filter Filter on Filter on JW No filter Filter on Filter on IV No filter Filter on Filter on Mean No filter Filter on Filter on corner edge corner edge corner edge corner edge Error max,° T peak, ms Latency, ms , ms 36 66 115 65 26 64 121 65 14 52 122 52 33 57 129 58 22 42 131 42 14 34 132 35 31 91 111 91 20 61 117 62 10 60 118 61 33.3 ⫾ 2.5 71.3 ⫾ 17.6 118.6 ⫾ 7.6 71.7 ⫾ 17.2 22.7 ⫾ 3.1 55.7 ⫾ 11.9 122.9 ⫾ 5.9 56.0 ⫾ 12.2 12.7 ⫾ 2.3 48.7 ⫾ 13.3 124.3 ⫾ 5.8 49.0 ⫾ 12.8 J Neurophysiol • VOL shows the mean tracking direction error as a function of time, for a line-drawing figure moving at 10°/s. Yellow shaded areas indicate the open-loop period from the latency measurements obtained in this experiments (average latency: 115 ⫾ 7 ms), and again the peak direction error was reached (63 ⫾ 13 ms) before the oculomotor control loop was closed. Results are similar to those observed with similar visual conditions in the previous experiments. Bottom plots show the same data but now together with the mean tracking direction errors observed with a texturefilled and a uniformly filled rhombus. First, there is no difference between the transiently biased responses to empty and filled objects. Second, for all three observers, the transient bias toward the VA prediction is absent (or at least dramatically reduced) with the texture-filled objects. The decay time constant could not be estimated as the time course of the tracking direction error was almost flat, remaining close to the veridical object-motion direction across time. DISCUSSION This study extends our previous finding that the initiation of smooth pursuit eye movements reflects the dynamics of early motion integration. By manipulating the contrast, speed, texture filling, and the shape of simple, line-drawings targets, we 93 • APRIL 2005 • www.jn.org Downloaded from http://jn.physiology.org/ by 10.220.32.247 on June 15, 2017 FIG. 7. Smooth pursuit responses to a 10°/s object motion: effects of edges, corners, and line endings. Layout is the same as for Figs. 3 and 5. A: mean velocity profiles of pursuit eye movements driven by 3 different types of tilted diamonds drifting rightward. Insets: the 3 different types of images used with the color code for the velocity profiles. B: mean tracking direction errors as a function of time for a single control condition (unfiltered diamonds, top) and for the 3 different filtered image types (no filter, filter on edge, filter on corner, bottom). The same color code is used for A and B. 2288 J. M. WALLACE, L. S. STONE, AND G. S. MASSON Pursuit as a probe of the temporal dynamics of motion integration FIG. 8. Effects of edges and corners upon pursuit dynamics. Top and bottom: the mean (⫾SD, across subjects) Peak direction errors and time constants, respectively, for each image type.* and ** indicate 0.05 and 0.001 significance level for pair-wise comparisons. were able to measure the influence of these parameters on the tracking direction accuracy, and in particular to examine the time course over which the neural direction signal driving pursuit develops to match the veridical motion direction of the pursued object. Our main findings are the following. First, the initiation of pursuit to empty line-drawing objects is launched in the direction of the VA solution computed by averaging the local motions orthogonal to the edges, despite the fact that for, type II diamonds, this estimate is ⬃44° away from the actual 2D trajectory of the object. This initial mis-estimation gradually disappears within ⬃300 ms such that steady-state tracking is then aligned with the true object-motion direction. However, we found that the reduction of the initial tracking error starts before the closing of the oculomotor control loop, indicating that the initial refinement of the direction estimate is purely visually based, as opposed to being a consequence of a visuomotor correction driven by negative visual feedback. Second, the phenomenon is clearly manifested and has similar dynamics over a large range of target speeds and contrasts although the peak error and decay time constant do depend on object J Neurophysiol • VOL Because pursuit is a closed-loop negative feedback system, when interpreting the time course of the initial smooth eyemovement responses to visual motion, one must be vigilant to dissociate the dynamics of the underlying visual motion signals driving pursuit (the open-loop sensory drive) from that caused by the dynamic properties of negative feedback (the sensorimotor feedback signal). In our earlier study (Masson and Stone 2002), we addressed this issue indirectly by transiently blanking the visual stimulus during steady-state pursuit. The postblanking re-appearance of the stimulus did not engender a re-initiation of the tracking error. Instead, the steady-state response to the same type II diamond motion immediately drove pursuit back in the correct object-motion direction. While this result did show that the steady-state pursuit response is latched by an initial visual decision process that does not need to be re-performed, this finding did not resolve the more pointed question of whether or not the initial pursuit dynamics were dominated by feedforward visual or feedback sensorimotor signals. Herein, we resolved this question by demonstrating that the initial tracking error begins to converge toward the veridical direction before the closing of the oculomotor control loop, that is, before feedback signals could possibility affect the response. The estimated direction error typically peaked after only ⬃60 ms, which is ⬃30 – 60 ms before feedback is thought to affect pursuit (Lisberger and Westbrook 1985). At the time of closing the feedback loop, the initial direction was already reduced by ⬃20 –30%. Moreover, under high-contrast and high-speed conditions, the time constant of the decaying tracking error was as fast as ⬃60 ms, which is about twice as fast as the internal time constant of pursuit pathways (Goldreich et al. 1992). Finally, at low target contrast, initial eye speed was reduced by ⬃20%, but the peak direction error was increased by ⬃15%, and the mean time constant increased by ⬃200%. This indicates that, at low contrast, the much longer time it takes to converge to the true object-motion direction cannot be accounted for by the more sluggish pursuit responses. We therefore can conclude with confidence that the decay of the tracking error is largely reflective of the time course of the visual refinement of the neural object-motion signal and is largely independent of any 93 • APRIL 2005 • www.jn.org Downloaded from http://jn.physiology.org/ by 10.220.32.247 on June 15, 2017 contrast and speed. In particular, lower contrast and higher speed produce larger initial deviations toward the VA solution. Third, varying the relative strength of 1D (edges) and 2D (corners) motion cues has a profound impact on the initial tracking behavior. The main finding herein is that when enough nonambiguous 2D cues (such as surface texture elements) are provided, the initial tracking responses is no longer aligned with the VA prediction, but instead follows immediately—and at the same latency—the true object motion trajectory. In summary, this study shows that the time course of pursuit onset is a powerful tool that can be used to probe the dynamics of human cortical motion processing. More specifically, our results shed light on the relationship between the visual signals driving perception and pursuit, provide insight into the dynamics of the neural representation of object motion direction, and challenge current models of primate cortical motion processing as well as of pursuit control. 2D OBJECT MOTION FOR PURSUIT EYE MOVEMENTS 2289 pursuit-driven feedback signals. As an extension of the work of Lisberger and Westbrook (1985) with small spots, our results show that examination of the initial open-loop pursuit response to judiciously constructed stimuli is a powerful probe to reveal the temporal dynamics of the neural processing of 2D visual motion signals. Comparison with perceptual studies Previous attempts to measure the temporal dynamics of motion integration have examined the effects of stimulus duration on perceived direction. Apart from the agreement that for very short durations (⬍100 ms), the perceived direction matches the VA solution (Lorenceau et al. 1993; Yo and Wilson 1992), there is no clear consensus on the time course over which the percept shifts to the actual 2D direction of the pattern/object. For high-contrast type II plaids, human observers report seeing the correct pattern direction for durations ⬎150 ms (Yo and Wilson 1992). For multiple moving tilted lines, for which the direction of motion orthogonal to the line differs from the true motion direction of the line endings, asymptotic performance was reached with stimulus durations J Neurophysiol • VOL ⬎400 ms (Lorenceau et al. 1993). Our pursuit findings are broadly consistent with these perceptual findings. For a better comparison, we fit a single-exponential function to the psychophysical data from Lorenceau et al. (1993) and Yo and Wilson (1992) and found a large variability in estimated decay time constants: ⬃200 and ⬃30 ms, respectively. The latter estimate reflects the change in the actual perceived direction over time and is therefore more directly comparable to our behavioral results. Thus the psychophysical temporal dynamics are in reasonable agreement with the present results, consistent with a shared visual direction signal for both perception and pursuit (Beutter and Stone 2000; Krukowski and Stone 2004; Stone and Krauzlis 2003; Stone et al. 2000; Watamaniuk and Heinen 1999). For measuring response dynamics, the approach of examining pursuit responses provides however a number of advantages over forced-choice psychophysical techniques. First, it is much more time efficient in that a single pursuit trial provides data for the entire range of time points, whereas forced-choice or adjustments psychophysical measures must be repeated for each time point studied. Second, even with brief stimulus presentations, the psychophysical decision may not be limited to the stimulus duration alone: observers have until they 93 • APRIL 2005 • www.jn.org Downloaded from http://jn.physiology.org/ by 10.220.32.247 on June 15, 2017 FIG. 9. Smooth pursuit responses to a 10°/s object motion: effects of surface 2D texture elements. Layout is the same as for Figs. 3, 5, and 7. A: mean velocity profiles of pursuit eye movements driven by 3 different types of tilted diamonds drifting rightward. Insets: the 3 different types of images used with the color code for the velocity profiles. B: mean tracking direction errors as a function of time for a single control condition (line-figure diamonds, top) and for the 3 different fill patterns (none, uniform, and textured, bottom). The fits obtained with the textured diamond were not significant. The same color code is used for A and B. 2290 J. M. WALLACE, L. S. STONE, AND G. S. MASSON provide a response to continue processing the visual information collected during the stimulus presentation. Third, while forced-choice psychophysics provide a measure of the signalto-noise value inferred from the percentage of trials where object-motion direction is reported, pursuit data simultaneously provide separate measures of signal (the mean pursuit direction) and noise (the variance in that direction) that allow for clearer interpretation. Indeed, in this study, we found a smooth transition over time for the ensemble of trials as shown by the gradual shift of a monotonic distribution of tracking directions (Fig. 1B). This is easily distinguishable from a stochastic shift between discrete VA and IOC decisions, which would have yielded bimodal distributions of pursuit directions. Our quantitative analysis revealed three main characteristics of motion integration. First, at low contrast (10%) and high speed (20°/s), the peak direction error is almost identical to the VA solution (⬃43 and 38°, respectively). Second, when the surface object is filled with high-spatial frequency 2D local information, the peak direction error is always reduced to small, residual biases (⬍5°), meaning that the earliest part of pursuit onset is already aligned with the true object-motion direction. Third, intermediate peak direction errors were found for a variety of conditions, including high contrast, slow speeds, filtered objects with edge motion, and/or increased line-ending motion. Across all these conditions, mean peak direction errors ranged between 20 and 30°, that is, in between VA and IOC directions. In particular, when circular Gaussian filters were centered onto the object corners (filter on corner condition), the peak direction error was lowered down to ⬃20°, likely due to the introduction of new 2D features at the end of the four edges, which move in the object-motion direction (Shiffrar et al. 1995). When circular Gaussian filters were centered onto the edges (filter on edge condition), thereby removing much of the edge motion information while also introducing new line endings, this reduced the peak direction error even more, to ⬃10° (Fig. 7B) as now pursuit is driven predominantly by the motion of 2D features (the corners and the line endings). Interestingly, the latency of pursuit responses remained remarkably constant across all these conditions. At first glance, the fact that under some conditions the initial pursuit direction can be aligned with the true object/pattern motion direction might appear to be contradictory with our earlier studies on the temporal dynamics of short-latency ocular following responses in humans (Masson 2004). We have shown previously that reflexive tracking eye movements always start first in the direction of 1D (grating or edge) motion such as with the barber-pole stimulus (Masson et al. 2000) or in the VA direction of multiple 1D motion signals such as with type I plaid patterns (Masson and Castet 2002). However, the latency of these 1D-driven reflexive responses is only ⬃85 ms, which is shorter than the latencies found in the present series of experiment (⬃100 –120 ms). Furthermore, we found that 2D motion cues such as line endings in barber-pole stimuli or blobs in plaid patterns had an influence on tracking direction ⬃25 ms later, i.e., with a latency of ⬃110 –120 ms after visual motion onset. Such temporal dynamics were found over a large range of pattern speeds and contrasts (Masson and Castet 2002; Masson et al. 2000). These results on reflexive, short-latency J Neurophysiol • VOL 93 • APRIL 2005 • www.jn.org Downloaded from http://jn.physiology.org/ by 10.220.32.247 on June 15, 2017 Motion-integration computation for pursuit initiation tracking responses provide two insights into the present results on voluntary pursuit. First, 1D edge motions are extracted first and vector averaged to determine the potentially biased initial estimate of global motion. Second, at high contrast, ⬃110 ms after stimulus onset, that is at the time of smooth pursuit onset, 2D motion features have been extracted and are now being integrated into the computation of the 2D object motion. Thus at the time of voluntary pursuit onset, (i.e., ⬃100 –110 ms after visual motion onset), the relative weighting between 1D and 2D motion can explain the continuum we observed for the peak direction errors, lying in between the VA and true objectmotion directions. Classical views of motion integration postulate that motion signals orthogonal to an elongated edge are the most salient information present in the image, particularly at low contrast (Simoncelli et al. 1991; Wilson et al. 1992) and therefore dominate early perceptual responses. Our results show that this assumption applies to voluntary smooth eye movements. At low contrast, the initial estimate of global motion relies mostly on the simple VA solution and 2D feature motion has very little, if any, impact. Increasing the target speed had similar effects as sweeping an edge or a feature more rapidly over a restricted receptive field would increase the level of measurement uncertainty at the level of the edge-motion detectors. Moreover, at low contrast, the latency of visual processing of 2D features is delayed (Masson and Castet 2002), and therefore the earliest phase of tracking is driven nearly exclusively by 1D edge motion signals. Thus the initial pursuit direction nearly perfectly matches the VA prediction. At high contrast, 2D features are integrated more rapidly and therefore can start influencing the tracking direction earlier (Wilson et al. 1992). Weiss et al. (2002) proposed a probabilistic model of motion integration where a most probable velocity field is computed from the local velocity likelihood at every image location (Simoncelli 2003; Simoncelli et al. 1991), which are combined with an a priori distribution favoring slow and smooth motions. At low contrast and high speed, i.e., high noise levels, the product of the prior and the local velocity distribution yields an a posteriori distribution whose maximum occurs at the VA solution. At high contrast, however, the maximum of the posterior probability is found at, or near, the IOC prediction. Such a Bayesian model captures some of the main characteristics of 2D motion perception, such as a bias toward the VA solution at low contrast and a gradual shift from VA to IOC solutions as contrast (or, equivalently, duration) increases. For the present study, it can predict the initial bias toward the VA solution and more particularly the larger peak direction errors found at low contrast and high speed. However, the simple model proposed by Weiss and coworkers (2002) falls short in two critical ways. First, for linedrawing objects, pursuit always starts in a direction with a minimum bias of ⬃30°, which is much closer to the VA solution than any other solution. And this 30° deviation is observed at high contrast target and slow speeds as well. This finding clearly differs from the predictions of their model that was tuned to accommodate perceptual results where the IOC solution prevails at high contrast (e.g., Lorenceau et al. 1993; Yo and Wilson 1992). Within their theoretical framework, our results could, however, be accounted for by the temporal integration window of the early spatiotemporal filtering being much longer than the classical pursuit latency (⬃100 ms). 2D OBJECT MOTION FOR PURSUIT EYE MOVEMENTS J Neurophysiol • VOL ing on contrast, speed, and spatial filtering, such a model would give different weights to 1D and 2D motion measurements as a function of their reliabilities. The prior favoring slow speed would still dominate motion integration at shorter latencies when the uncertainty on the 1D motion information is high, i.e., with low contrasts and high speeds. On the contrary, at high contrast, 2D feature motion would start to dominate and therefore lower the impact of the prior. Moreover, as the latencies of 2D motion detection increases with lower contrast, one would expect more sluggish dynamics in the gradual shift between VA and true object-motion direction. Neural solutions of dynamical motion integration Area MT is a key structure in the complex neural network underlying smooth pursuit eye movements (see Krauzlis 2004 for a review). Tracking eye movements get their main visual input from a population of area MT neurons that represent both target direction and speed (Lisberger and Movshon 1999; Priebe et al. 2003, Priebe and Lisberger 2004). MT neurons have many interesting properties relevant to the present results. When presented with single oriented bars or gratings, the direction selectivity of many neurons is normal to the bar/ grating orientation present in the receptive field (Albright 1984). When presented with two overlapping gratings, such as in plaid patterns, these neurons respond to one component motion direction. They have been called “component-selective” neurons (Movshon et al. 1985). There is a subgroup of MT neurons that respond to a moving contour independently from its orientation (Pack and Born 2001) or to a moving plaid independently to the orientation/direction of its components (Movshon et al. 1985; Rodman and Albright 1989). It has been argued that these ”pattern-selective“ cells form the basis for 2D motion perception (Movshon et al. 1985). There is recent neurophysiological evidence pattern-selectivity gradually emerges over ⬃100 ms. Pack and Born (2001) showed that when presented with tilted moving bars, the earliest responses of MT neurons are highly dependent on stimulus orientation, i.e., their direction selectivity tuning curves point toward the vector normal to the bar orientation. This dependence decreases gradually so that within ⬃150 ms after stimulus motion onset, MT cells primarily encode the actual true 2D objectmotion direction of the moving bar, independently of their orientations. To quantify the neuronal temporal dynamics, we fit their “neuronal direction selectivity error” data, i.e., the difference between preferred directions for oblique and upright bars (Pack and Born 2001) (Fig. 2C) with our double-exponential function. We obtained a best-fitting estimate for the decay time constant of ⬃40 ms. Such neuronal dynamics are strikingly consistent with our behavioral observation on pursuit. More recently, Pack et al. (2001) and Smith et al. (2001) showed similar temporal dynamics for the emergence of true “patternselective” cells in response to moving plaids. Again, it took ⬎100 ms of neuronal response before an accurate estimate of 2D pattern motion emerged. It is still unclear why macaque MT cells, as well as smooth pursuit eye movements in both monkeys (Pack and Born 2001) and humans (Masson and Stone 2002) exhibit such temporal dynamics. Pack et al. (2004) suggested a specific role for 2D features such as line endings, which can be extracted through specific detectors such as V1 end-stopped cells (Pack et al. 93 • APRIL 2005 • www.jn.org Downloaded from http://jn.physiology.org/ by 10.220.32.247 on June 15, 2017 Under this assumption, a high-contrast target can still generate large measurement uncertainty and the prior would still dominate the a posteriori distribution. However, this seems unlikely given that earlier psychophysical studies have shown that the optimal motion stimulus duration for speed and direction discrimination is ⬍100 ms (DeBruyn and Orban 1988; Masson et al. 1999; McKee 1981; Watson and Turano 1995). This estimate is consistent with a recent neural information theory analysis by Osborne et al. (2004) showing that information about motion direction builds up very rapidly, reaching ⱕ80% of maximum information in the first 100 ms of neuronal responses from a small population of area MT neurons. Second, increasing the number of 2D features lowered the initial peak direction error, suggesting that one should consider the relative weight of 1D and 2D motion cues to explain why—and when— the initial tracking direction is deviated toward the VA solutions. This scheme departs from the model of Weiss et al. (2002) as it stresses the role of 2D features in a parallel computation of object-motion direction with the output no longer dictated by the relative probabilities of the VA and the IOC signals, which are both based only on the 1D motion signals extracted from the image. Furthermore, previous studies have shown that different motion percepts and steady-state pursuit responses can be obtained from identical velocity space information, demonstrating a clear need to consider other factors (Lorenceau and Alais 2001; Stone et al. 2000). The final two experiments showed that manipulating the relative amount of 1D and 2D signals available has a large effect on the directional error of pursuit. In fact, when a large proportion of 2D features is provided, as with the textured object in Fig. 9, the initial tracking direction is in the direction of the true object motion. (Note that this is not true for a uniform gray stimulus, indicating it is indeed the texture elements driving the object response, not simply the addition of luminance to the interior of the stimulus). This result suggests that there is a competition between 1D edge and 2D feature motion signals and that the most reliable signals dominate the direction estimate. Interestingly, recent findings of Lindner and Ilg (2000) have demonstrated a similar effect, in which initial pursuit responses are driven in the average direction of first- and second-order information before the true object direction wins the direction estimate (over a similar time course as found in the present study). Finally, our results suggest that the 2D features become reliable at high contrast and low speeds, whereas 1D information is more reliable at low contrast and high speeds, consistent with the views espoused in several psychophysical studies (Castet et al. 1993; Lorenceau et al. 1993; Mingolla et al. 1992; Shiffrar and Lorenceau 1996). To explain all these results, an alternative view is that 1D and 2D motion signals have different latencies and therefore influence the global estimate of object motion at different points in time. Such a scheme of parallel processing between 1D and 2D motion cues can be easily encorporate a Bayesian framework with two parallel channels having different dynamics in which both contribute independent likelihood fields to the final distributed representation object motion (Perrinet et al. 2004). By having two different latencies (⬃85 and 110 ms, respectively), the same model can render some dynamical aspects of both short-latency ocular following (latencies ⬍110 ms) and smooth pursuit responses (for latencies ⬎110 ms) via parallel, asynchronous inputs to the integration stage. Depend- 2291 2292 J. M. WALLACE, L. S. STONE, AND G. S. MASSON Conclusions We have shown that the initiation of smooth pursuit eye movements exhibits interesting dynamic properties of 2D motion integration similar to that previously observed for perception and for neural responses in macaque area MT. All of these results call for the development of models of 2D motion integration wherein an initial estimate of object motion is first extracted by combining the different motion signals available at different points in time within a 2D integration stage and then refines its output over time. ACKNOWLEDGMENTS We are grateful to D. Paleressompoule and A. De Moya for technical assistance. We thank I. Vanzetta for carrying on the experiments and E. Castet and L. Perrinet for insightful discussions and comments about the manuscript. GRANTS This work was supported by the Centre National de la Recherche Scientifique and by the European Commission through the Perception and Recognition for Action Research and Training Network (HPRN-CT-2002-00226). L. Stone was supported by NASA’s Biomedical Research and Countermeasures (111-10-10) and Aviation Operation Systems (727-05-30) programs. REFERENCES Adelson EH and Movshon JA. Phenomenal coherence of moving visual patterns. Nature 300: 523–525, 1982. Albright TD. Direction and orientation selectivity of neurons in visual area MT. J Neurophysiol 52: 1106 –1130, 1984. Beutter BR and Stone LS. Motion coherence affects human perception and pursuit similarly. Vis Neurosci 17: 139 –153, 2000. Bowns L. Evidence for a feature tracking explanation for why type II plaids move in the vector sum directions at short directions. Vision Res 36: 3685–3694, 1996. Busettini C, Miles FA, and Schwarz U. Ocular responses to translation and their dependence on viewing distance. II. Motion of the scene. J Neurophysiol 66: 865– 878, 1991. Castet E, Lorenceau J, Shiffrar M, and Bonnet C. Perceived speed of moving lines depends on orientation, length, speed and luminance. Vision Res 33: 1921–1936, 1993. Collewijn H, van der Mark F, and Jansen TC. Precise recordings of human eye movements. Vision Res 15: 447– 450, 1975. De Bruyn B and Orban GA. Human velocity and direction discrimination measured with random dot patterns. Vision Res 28: 1323–1335, 1988. Fennema CL and Thompson WB. Velocity determination in scenes containing several moving images. Comp Graph Image Proc 9: 301–315, 1970. J Neurophysiol • VOL Ferrera VP and Wilson HR. Perceived direction of moving two-dimensional patterns. Vision Res 32: 1015–1031, 1990. Goldreich D, Krauzlis RJ, and Lisberger SG. Effect of changing feedback delay on spontaneous oscillations in smooth pursuit eye movements of monkeys. J Neurophysiol 67: 625– 638, 1992. Goréa A and Lorenceau J. Directional performances with moving plaids: component-related and plaid-related processing modes coexist. Spatial Vis 5: 231–252, 1991. Koechlin E, Anton JL, and Burnod Y. Bayesian inference in populations of cortical neurons: a model of motion integration and segmentation in area MT. Biol Cybern 80: 25– 44, 1999. Krauzlis RJ. Recasting the smooth pursuit eye movement system. J Neurophysiol 91: 591– 603, 2004. Krauzlis RJ and Miles FA. Initiation of saccades during fixation and pursuit: evidence in humans for a single mechanisms. J Neurophysiol 76: 4175– 4179, 1996. Krukowski A and Stone LS. Expansion of direction space around the cardinal axes revealed by smooth pursuit eye movements. Neuron In press Linder A and Ilg UW. Initiation of smooth-pursuit eye movements to first-order and second-order motion stimuli. Exp Brain Res 133: 450 – 456, 2000. Lisberger SG and Movshon JA. Visual motion analysis for pursuit eye movements in area MT of macaque monkeys. J Neurophysiol 19: 1907– 1916, 1999. Lisberger SG and Westbrook LE. Properties of visual inputs that initiate horizontal smooth pursuit eye movements in monkeys. J Neurosci 5: 1662–1673, 1985. Löffler G and Orbach HS. Computing feature motion without feature detectors: a model for terminator motion without end-stopped cells. Vision Res 39: 859 – 871, 1999. Lorenceau J and Alais D. Form constraints in motion binding. Nat Neurosci 4: 745–751, 2001. Lorenceau J, Shiffrar M, Wells N, and Castet E. Different motion sensitive units are involved in recovering the direction of moving lines. Vision Res 33: 1207–1217, 1993. Masson GS. From 1D to 3D via 2D: dynamics of surface motion segmentation for gaze stabilization in primates. J Physiol (Paris) 98: 35–52, 2004. Masson GS and Castet E. Parallel motion processing for the initiation of short-latency ocular following in humans. J Neurosci 22: 5149 –5163, 2002. Masson GS, Mestre DR, and Stone LS. Speed tuning of motion segmentation and discrimination. Vision Res 39: 4297– 4308, 1999. Masson GS, Rybarczyk Y, Castet E, and Mestre DR. Temporal dynamics of motion integration for the initiation of tracking eye movements at ultra-short latencies. Vis Neurosci 17: 753–767, 2000. Masson GS and Stone LS. From following edges to pursuing objects. J Neurophysiol 88: 2869 –2873, 2002. Miles FA. The neural processing of 3D visual information: evidence from eye movements. Eur J Neurosci 10: 811– 822, 1998. McKee SP. A local mechanism for differential velocity detection. Vision Res 21: 491–500, 1981. Mingolla E, Todd J, and Norman J. The perception of globally coherent motion. Vision Res 32: 1015–1031, 1992. Movshon JA, Adelson EH, Gizzi MS, and Newsome WT. The analysis of visual moving patterns. In: Pattern Recognition Mechanisms, edited by Chagas C, Gattass R, and Gross C. New York: Springer, 1985, p. 117–151. Movshon JA and Newsome WT. Visual response properties of striate cortical neurons projecting to area MT in macaque monkeys. J Neurosci 16: 7733–7741, 1996. Osborne LC, Bialek W, and Lisberger SG. Time course of information about motion direction in visual area MT of macaque monkeys. J Neurosci 24: 3210 –3222, 2004. Pack CC, Berezovskii VK, and Born RT. Dynamic properties of neurons in cortical area MT in alert and anesthetized macaque monkeys. Nature 414: 905–908, 2001. Pack CC and Born RT. Temporal dynamics of a neural solution to the aperture problem in visual area MT of macaque brain. Nature 409: 1040 – 1042, 2001. Pack CC, Gartland AJ, and Born RT. Integration of contour and terminator signals in visual area MT. J Neurosci 24: 3268 –3280, 2004. Pack CC, Livingstone MS, Duffy KR, and Born RT. End-stopping and the aperture problem: two-dimensional motion signals in macaque V1. Neuron 39: 671– 680, 2003. Perrinet L, Castet E and Masson GS. A model of motion processing underlying early ocular following. Perception ECVP Abstracts 33: 70b, 2004. 93 • APRIL 2005 • www.jn.org Downloaded from http://jn.physiology.org/ by 10.220.32.247 on June 15, 2017 2003), although these cell types are not known to feed directly through to MT (Movshon and Newsome 1996). Nevertheless, V1 end-stopped cells may play a crucial role in the effects we found. End-stopping is known to weaken at low contrast (Sceniak et al. 1999) and, at high contrast, is delayed ⬃20 –30 ms after the initial edge driven response (Pack et al. 2003), corresponding well with that found in the ocular following study previously discussed of Masson et al. (2000). The frontend end-stopped response could drive pursuit in the correct direction when there is an abundance of texture information, while in situations with more equally available 2D and 1D information the delayed end-stopped response could underlie the temporal dynamics of motion integration we have observed. However, several other schemes remain plausible such as computing a delayed IOC-based 2D signal from different 1D motion signals (Adelson and Movshon 1982; Movshon et al. 1985) or extracting separately first- and second-order motion signals before vector averaging them (Löffler and Orbach 1999; Wilson et al. 1992). 2D OBJECT MOTION FOR PURSUIT EYE MOVEMENTS J Neurophysiol • VOL Smith MA, Majaj N, Bair W, and Movhson JA. Dynamics of responses to grating and plaids in macaque area MT. Soc Neurosci Abstr 21: 680.2, 2001. Stone LS, Beutter BR, and Lorenceau J. Visual motion integration for perception and pursuit. Perception 29: 771–787, 2000. Stone LS and Krauzlis RJ. Shared motion signals for human decisions and oculomotor actions. J Vis 3: 725–736, 2003. Wallach H. Uber Visuell Wahrgenommene Bewegungrichtung. Psychol Forsch 20: 325–380, 1935 translated by Wuerger S and Shapley B. Perception 11: 1317–1367, 1996. Watamaniuk SJN and Heinen SJ. Human smooth pursuit discrimination. Vision Res 39: 59 –70, 1999. Watson AB and Turano K. The optimal motion stimulus. Vision Res 35: 325–336, 1995. Weiss Y and Adelson EH. Slow and smooth: a Bayesian theory for the combination of local motion signals in human vision. Technical Report AIM-1624, MIT, Feb. 1998. Weiss Y, Simoncelli EP, and Adelson EH. Motion illusions as optimal percepts. Nature Neurosci 5: 598 – 604, 2002. Wilson HR, Ferrera VP, and Yo C. A psychophysically motivated model for two-dimensional motion perception. Visual Neurosci 9: 79 –97, 1992. Yo C and Wilson HR. Perceived direction of moving two-dimensional patterns that depends on duration, contrast and eccentricity. Vision Res 32: 135–147, 1992. 93 • APRIL 2005 • www.jn.org Downloaded from http://jn.physiology.org/ by 10.220.32.247 on June 15, 2017 Priebe NJ, Cassanello CR, and Lisberger SG. The neural representation of speed in macaque area MT/V5. J Neurosci 23: 5650 –5661, 2003. Priebe NJ and Lisberger SG. Estimating target speed from the population response in visual area MT. J Neurosci 24: 1907–1916, 2004. Rashbass C. The relationship between saccadic and smooth tracking eye movements. J Physiol 159: 326 –338, 1961. Rao RPN. Bayesian computation in recurrent neural circuits. Neural Comp 16: 1–38, 2004. Rodman and Albright TD. Single-unit analysis of pattern-motion selective properties in the middle temporal visual area (MT). Exp Brain Res 75: 53–64, 1989. Sceniak MP, Ringach DL, Hawken MJ, and Shapley R. Contrast’s effect on spatial summation by macaque V1 neurons. Nat Neurosci 2: 733–739, 1999. Simoncelli EP. Local analysis of visual motion. In: Visual Neuroscience, edited by Chalupa LM and Werner JS. Cambridge, MA: MIT Press, 2003, vol. II, ch. 109, p. 1616 –1623. Simoncelli EP, Adelson EH, and Heeger D. Probability distributions of optical flow. Proc. IEEE Conf Comput Vis Pattern Recogni 310 –315, 1991. Shiffrar M, Li X, and Lorenceau J. Motion integration across differing image features. Vision Res 36: 2137–2146, 1995. Shiffrar M and Lorenceau J. Increased motion linking across edges with decreased luminance contrast, edge width and duration. Vision Res 36: 2061–2067, 1996. 2293
© Copyright 2026 Paperzz