IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 12, NO. 4, APRIL 2003 409 Jointly Registering Images in Domain and Range by Piecewise Linear Comparametric Analysis Frank M. Candocia, Member, IEEE Abstract—This paper describes an approach whereby comparametric analysis is used in jointly registering image pairs in their domain and range, i.e., in their spatial coordinates and pixel values, respectively. This is accomplished by approximating a camera’s nonlinear comparametric function with a constrained piecewise linear one. The optimal fitting of this approximation to comparagram data is then used in a re-parameterized version of the camera’s comparametric function to estimate the exposure difference between images. Doing this allows the inherently nonlinear problem of joint domain and range registration to be performed using a computationally attractive least squares formalism. The paper first presents the range registration process and then describes the strategy for performing the joint registration. The models used allow for the pair-wise registration of images taken from a camera that can automatically adjust its exposure as well as tilt, pan, rotate and zoom about its optical center. Results concerning the joint registration as well as range-only registration are provided to demonstrate the method’s effectiveness. Index Terms—Comparagram, comparametric function, domain registration, image registration, joint registration, least squares, photoquantity, piecewise linear, range registration. I. INTRODUCTION R ESEARCH into the registration of images has been ongoing for several decades [1]. Results of this research have been applied in fields such as photogrammetry [2], computer vision [3] and medical imaging [4] and to many problems such as image superresolution [5], [6], depth extraction [7], video compression [8], [9], video indexing [10] and the construction of large mosaics [11]. This large body of research, however, has been slanted toward domain-only registration; that is, registration between image functions’ spatial coordinates. A reading of comprehensive survey papers on image registration and the references they list provides such evidence [1], [4]. These registration approaches have typically assumed a specific geometric relationship between images’ spatial coordinates, for example translational, affine or perspective relationships have been common [11]–[13]. Also assumed have been more complex relationships involving the computation of the optical flow field [14] or the modeling of lens distortion [15]. More recently, there is a growing trend Manuscript received June 27, 2002; revised January 23, 2003. This work was supported in part by the Telecommunications and Information Technology Institute of Florida International University under Grant BA01-20. The associate editor coordinating the review of this manuscript and approving it for publication was Prof. Bruno Carpentieri. The author is with the Department of Electrical and Computer Engineering, Florida International University, Miami, FL 33174 USA (e-mail: [email protected]). Digital Object Identifier 10.1109/TIP.2003.811497 in optical imaging involving range-only registration; that is, registration between image functions’ pixel values [16]–[20]. This research seeks to construct high dynamic range maps of an imaged scene from several images captured at different optical settings. In this way, the typical imaged scene resolution of 8 bits/pixel can be significantly increased and the multiple captured images can be combined into a single picture where all objects are now properly exposed [21], [22]. The registration of optical images in both domain and range has been an area that has seemingly received less attention from the research community. In medical imaging, the difficulty of accurately modeling intensity relationships between inter-modality images has led to variational approaches for performing the registration [23], [24]. With optical images, this modeling has been less complicated. As such, approaches that have drawn from the aforementioned range-only registration research have utilized spatially varying optical filters on cameras so as to obtain multiple measurements of scene points under differing optical settings [25], [26]. By characterizing these filters responses, a mechanism has been created that can provide a greater dynamic scene description and this, in turn, can be used to create image mosaics with increased dynamic range. The work presented in this paper is also concerned with registering images where scene points have been multiply imaged at different optical settings. However, it registers captured images from a camera needing no special lenses or attachments. That is, as today’s typical cameras automatically adjust their exposure depending on the scene imaged and/or ambient lighting and this effect on the image produced can be accurately modeled, it should be accounted for during the registration process if indeed this is warranted. Now, there are situations where this change can be neglected, for example when the amount of variation is practically less than camera’s noise power or when brightness independent features are used in the registration process. But when this variation in pixel intensity is no longer negligible, as in luminance-based, i.e., featureless, registration approaches, it should be accounted for during registration. To the author’s knowledge, the lone work that has contributed to this type of joint registration has been that of Mann [19], [27]. This range registration work has established models that can be used to describe the nonuniform biasing of pixels that an image undergoes when a camera’s exposure settings are automatically adjusted in response to the amount of light sensed. The nonuniform biasing is due to the nonlinear dynamic range compression that an image sensor’s output voltage undergoes and which is subsequently quantized to form a pixel. This range compression is described by the camera’s response function or can, alternately, be described by its associated comparametric 1057-7149/03$17.00 © 2003 IEEE 410 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 12, NO. 4, APRIL 2003 function [19]. Owing to the comparametric function’s nonlinear nature, exposure estimation between images is typically a nonlinear estimation problem given the pixel samples between two images. When images are also to be registered in their spatial coordinates, all but the simplest of domain models also result in a nonlinear parameter estimation problem. Any effective linearization or procedure whereby the registration problem can be solved using a least squares formalism would present a computationally attractive alternative to this aforementioned nonlinear problem—whether it be in domain registration, range registration or both. The approach taken herein will perform the registration of a set of images by pair-wise registering adjacent images in the sequence and then utilizing the group structure of both the domain and range models to relate each image to a common domain and range reference. For each image pair, the joint estimation of parameters of a homographic (perspective) mapping in domain and a comparametric mapping in range is performed. The range model utilizes a piecewise linearization of the associated nonlinear comparametric function and this is incorporated into a domain model that is approximated using a series expansion. This results in a straightforward procedure where now, images can be registered jointly in domain and range using linear least squares and this can be done using any camera response function. In the remainder of this paper, Section II will address the camera response function and how this is used in registering two images in range. A piecewise linear approximation will be presented so as to solve this range registration using least squares and some range registration examples will be shown. In Section III, the established range model will be incorporated into an optimization procedure whereby both the domain and range model parameters can be determined using least squares computations. Results utilizing this framework on a real video sequence will be described in Section IV. The paper will conclude in Section V with a brief summary and paper-related comments. II. CAMERA RESPONSE FUNCTION AND REGISTRATION PROCESS THE RANGE The formation of each observed pixel value of an image is a process whereby the integration of photons over a finite period of time (the camera’s exposure time) occurs over a finite sensor area. This integration over time and space results in a quantifiable unit regarded as a photoquantity [19] and will be denoted by . The photoquantity is then subject to intrinsic sensor noise and finally goes through a dynamic range compression for limiting the range of possible values to a finite interval which is then quantized to form a pixel (typically 8 bits/pixel). We can express the formation of an image pixel then as related to the final formation of the pixel (which includes quantization, electronics and/or image compression related noise). Also, and represents the two-dimensional sensor’s spatial coordinates. Notice that the photoquantity lies in the interval where signifies that no light has been absorbed. The is a monotonically increasing camera response function function that is “normalized” such that over the domain of , , hence the pixel value varies in direct relation with the measured photoquantity. If we now capture an image over the same field of view but where its exposure time is different from that of , the photoquantity measured (per pixel) in is directly proportional to that measured in , i.e., is now expressed as (2) when there is an increase in exposure time in and where with a decrease in exposure time image relative to , when the exposure times are equivalent. from to and To register images in range, one must estimate the exposure difference between images typically using only the observed and , differ only in pixel values. Now, let two images, exposure. From (1) and (2) we see the only difference between these two images is parameter , which as mentioned, quantifies the exposure difference we seek to estimate. By normalizing , , we all pixel values in and such that can model our camera using any appropriate unicomparametric function described by Mann. The estimation process then involves the computation of a comparagram, i.e., a matrix used to relate the pixel values of two differently exposed images of the same subject matter, from pixel values of and followed by the solving for in the unicomparametric equation that provides an optimal fit to the comparagram data [19]. A. Piecewise Linear Modeling can be Any nonlinear unicomparametric function . modeled with a constrained piecewise linear function Notice here that and are variables representing the normalized (to interval [0,1)) range, i.e., pixel values, of two images and that the explicit relation to the spatial coordinates and photoquantity upon which images and depend is purposely represents omitted for the time being. Also note that the dependence of on the value of as well as the exposure parameter . With that said, a piecewise function is defined , as with linear segments over the interval .. . (3) (1) is the photoquantigraphic measure at spatial coorwhere represents the intrinsic noise dinate of the planar sensor, is the nonlinear camera response funcrelated to the sensor, is noise tion performing the dynamic range compression and , is a linear segment where represents prespecified knots in , . That and and are the knots at the interval’s endpoints and , is, are the interior knots. In order that CANDOCIA: JOINTLY REGISTERING IMAGES IN DOMAIN AND RANGE BY PIECEWISE LINEAR COMPARAMETRIC ANALYSIS best approximates in this work, certain constraints are employed. First, the endpoints of adjacent linear segments must be equal. This is satisfied by the constraint 411 where note that (4) and which accounts for the endpoints where “interior” linear segments. Second, the constraints of the and to be on the two unaccounted endpoints enforced follow from the unicomparametric equations that are and . being approximated here, namely that , This leads to simple constraints in our model where and , . Using the constraints in , the “ -intercepts” in each linear (4) together with becomes a function of the slopes and knots segment defining each previous linear segment: (10) . That is, for each of the constrained linear and the partial derivative of this function with segments in depends on the linear segment of respect to any parameter interest. Also, note that the equivalence in (9) must hold for all . This being the case, (9) can be written where in the form of a linear system of equations is the vector of unknown parameters is an matrix and is to estimate and length vector. Please note that for each value of , an , (9) can equivalently be expressed as (5) , the summation in (5) is not performed Notice that when in linear since would range from 1 to 0. This leads to as it should. When the constraint is segment is seen to be a function of slope included, slope parameter and the knots . We can therefore write parameters (6) . This makes dewhere since the knots are fixed. pend only on parameters Thus, we have chosen to parameterize the model in terms of , the slopes of the first ( ) linear segments , and we can express each linear segment , of , in terms of these free parameters as (11) Now substituting (3), whose constrained linear segments are described using (7), into (11) followed by simplification, we can determine the values in and . We find that each element in vector is (7) . pairs of samples ( , ), , Given from the comparagram of our two input images, we choose the best-fit piecewise linear function to be the one that minimizes the cost function (12) (8) more compactly, we write In order to express where with unTo minimize this cost function with respect to the we evaluate , known parameters , to get a system of simultaneous equations unknowns: in (9) otherwise (13) 412 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 12, NO. 4, APRIL 2003 and with otherwise where that fall in the interval with , (14) is the number of samples . Also, ; otherwise and (15) where data pairs ( Once and are constructed from the can be solved using, for example, least squares. Fig. 1. Two images to register in range. There is a difference in exposure of k = 16 in (b) relative to (a). for values of near 0 and near 1 and inclusion of such values from noise in the average unjustly biases the final estimate. The last point to make regarding (17) is that the integral is easily evaluated using any simple numerical integration technique. We have used Simpson’s rule in our calculations [28]. C. Range Registration Example and Results , (16) ), B. Exposure Difference Estimation One can notice that the piecewise linear model established in (3) is not parameterized with respect to our parameter of interest, , which quantifies the exposure difference we seek between images and . Instead, the estimate of , which will be denoted , is obtained as the following average: (17) is the unicomparaA few points are now addressed. First, metric function that has been re-expressed so that is a function of and now serves as the parameter, i.e., we’ve written . however has been replaced with its piecewise linear which in turn is a function of , knot loapproximation and slopes so that in (17) cations . Second, this inrepresents tegral arose from initially considering the minimization of the cost function (18) with respect to which simply results in . Rather than obtaining an average over some finite set of samples of (which naturally leads to the question, “How many samples of is enough?”), we can integrate over the entire range , and forego the previous issue. Third, of , i.e., the interval note that continually ranges between [0,1) but constants and in (17) are typically chosen such that is slightly greater than 0 is very sensitive and is slightly less than 1. This is because It is now instructive to illustrate range registration results using a simple example before proceeding to the joint registration work of the paper. In these results, a SONY DCRTV510 camcorder was fixed on a tripod and a series of 5 images were captured. The field of view in each of these images was the same but each image was captured at an exposure level twice that of the preceding image. Fig. 1 illustrates the first and last images of the captured sequence. Notice that, because of the exposure difference, it’s much easier to make out objects outside of the window in Fig. 1(a) when compared to Fig. 1(b). Conversely, the interior of the room is better seen in the longer exposed image of Fig. 1(b) when compared to Fig. 1(a). The range model between images that was utilized was the “preferred” unicomparametric equation given in [19] which relates the pixel values between images and by . By plotting the “preferred” unicomparametric equation for different values of , we can appreciate the nonlinear relationship between pixel values of two images that differ in exposure. This is illustrated in the comparagram of Fig. 2. Note that parameters and are fixed for a given camera. Once these are estimated for any camera, they remain fixed. We found values for our camera of and . To solve for the exposure difference between the two images in Fig. 1, we created a comparagram between the pixel values in these images and then used an linear segment model in (3) with equispaced knots in order to populate matrix and vector , as described in Section II-A so that of the first we could solve for the slopes linear segments using least squares. The was then estimated and the true value using (17). The estimated value was was known to be 16. We compared this value (obtained using linear least squares) to a nonlinear least squares optimization rather than the pieceusing the unicomparametric equation . Here, we also found wise linear approximation to it, . Fig. 3 illustrates the comparagram plot of the data of and have our two images of Fig. 1 and the best fit been superimposed on this plot. The values listed in Table I describe the estimates obtained by registering the image in Fig. 1(a) with the other four images CANDOCIA: JOINTLY REGISTERING IMAGES IN DOMAIN AND RANGE BY PIECEWISE LINEAR COMPARAMETRIC ANALYSIS 413 ^ = 13:9 Fig. 4. Range estimated images computed from Fig. 1(a). (a) Using k obtained from g (f ). (b) Using the a priori known true k = 16. TABLE I COMPARISON BETWEEN TRUE AND ESTIMATED VALUES THE CAPTURED SEQUENCE Fig. 2. Plot of “preferred” unicomparametric function exposure values k (with a, c fixed). g (f ) OF k FROM for different utilized does a good job of bringing Fig. 1(a) into range correspondence with Fig. 1(b). This is seen in Fig. 4(a) where the image of Fig. 1(a) has had its exposure adjusted using the “pre. ferred” unicomparametric model with the estimated For comparative purposes, Fig. 4(b) illustrates the exposure-ad. Nojusted image of Fig. 1(a) using the a priori known tice that the resulting images are practically indistinguishable and now closely resemble the captured image in Fig. 1(b). D. An Alternate Form To facilitate the derivation of the joint registration in the next section, we will re-express (3) into a more compact form. Before doing this, it is useful to define a set of orthogonal “basis images” in range which, in essence, indicate those spatial coordinates whose pixel values are within a range interval of interest: for all (19) otherwise Fig. 3. Comparagram and comparametric functions for the images in Fig. 1. Comparagram data is indicated by shading in the figure where darker shading indicates a higher density of samples. Superimposed are the estimated piecewise linear comparametric function g (f ) indicated as a solid line and the nonlinear comparametric function g (f ) indicated as a dashed line. in our five image “range sequence.” The results are seen to be strikingly similar regardless of whether the unicomparametric model or its piecewise linear approximation was used. Notice that the “preferred” model being utilized approximates the camcorder’s response function well but it is not a highly accurate model as the comparagram’s data indicates a saturation of image pixel values (here normalized to the interval [0,1)) at around 0.93. This is the reason for the error difference between the true and estimated values of in Table I. Nonetheless, the model and where the dependence of image on the for photoquantity and spatial coordinates [that have purposely been omitted in (3)–(7)] have been explicitly written for the sake of clarity in defining (19). Also, recall the ’s in (19) represent the knot locations into which is partitioned. Using (19) and on and dropping, once again, the dependencies of and for notational simplicity we can express the piecewise linear function of (3) as (20) where represents multiplication for any spatial coordinate in our image; otherwise, it would represent element-wise matrix multiplication if the set of spatial coordinates of our image is 414 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 12, NO. 4, APRIL 2003 viewed as a matrix to be operated on. By substituting (7) into (20) and simplifying, we arrive at a compact form for in terms of the free parameters : and using the pseudo-perspective model described in [13], we can express the spatial displacement as (21) where the ’s are nonlinearly related to the eight parameters describing the perspective mapping. Also, note and where (28) (22) is approximated as the time differ. By approxience between frames and will be denoted mating (27) with the piecewise linear model of (21) and utilizing the approximations mentioned above, (26) can be rewritten as III. JOINT DOMAIN AND RANGE REGISTRATION A. Models Used and Their Group Structure The process of jointly estimating the domain and range transformations between two images that is presented here utilizes the eight-parameter homographic, or perspective, mapping (23) (29) (24) on and and that of where the dependence of , , and and on have been dropped for notational conciseness. Note that in the domain and the single parameter range mapping which describes the change in exposure between images. The group structure of these domain and range models allows for a simple computing of the composition of joint domain and range transformations, specifically that (25) , where , and scalar and we also have . Thus, knowledge of transformations between consecutive pairs of images suffices to create a composite image. B. Joint Optimization and be two images, captured at different Let times, that are related by a rotation, panning, tilting, zooming and/or exposure change, all about the camera’s optical center (we can equivalently have arbitrary camera motion along with exposure change when imaging a planar surface). The objective is to jointly determine the domain and range transformations that register with . To this end, the cost function to be minimized is (30) can be approximated with the difand that ference operator in the -direction. A similar argument . The camera response function implemented holds for here is the “preferred” function described in [19] which is and for which the partial derivative of in (30) is . We now list all free parameters to estimate in a vector and paramperform the minimization of with respect to the eters in . To do so, we find the partial derivative of in (29) with respect to each of the parameters in and equate each partial derivative to 0. This reequations in as many unknowns: sults in a system of (31) where (32) (26) with and now and where where is minimum at that both images are registered in domain and range, respectively. Let us define so that denotes the domain-registered image. We can then see that (27) is now the requirement for a jointly registered image. Using a first order Maclaurin series expansion we can write , , and , . We have that , , . The elements are shown in (33)–(35) at the bottom in the submatrices of of the next page. Note that since , we write to denote elements that aren’t necessary to fill in. Also, the denotes summation over all and . After estimating from (31) using least squares, we must relate the parameters to the eight parameters of the domain’s perspective transformation. This portion is accomplished as described in [13]. Also, of our piecewise linear we must relate the slopes model to the exposure estimate . This is done as described in Section II-B. CANDOCIA: JOINTLY REGISTERING IMAGES IN DOMAIN AND RANGE BY PIECEWISE LINEAR COMPARAMETRIC ANALYSIS C. Registration Process As stated earlier, the adjacent images in a sequence will be registered pair-wise and then all of them will be brought to a final domain and range reference using our models’ group structure as described in Section III-A. So, given two images and , we proceed to iteratively solve for the true parameters of our 415 domain and range models, i.e., the , , and in (23) and (24). At each iteration step of the pair-wise registration, we first solve for vector in (31) and then we relate these approximate model parameters to the true parameters of our models as described in Sections II and III. To begin the iterations, we have used phase correlation [29] to estimate the translational component between and . That is, initially, we set to the identity (33) and .. . .. . .. . (34) and (35) 416 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 12, NO. 4, APRIL 2003 Fig. 5. Sequence of images to register in both domain and range. The camera panned across the room from left to right yielding the time-ordered images (a)–(s). Image (i) in the figure denotes the one whose domain and range coordinates are used as the reference for the composite image created in Fig. 6. matrix (no rotation or scaling), to the translation estimate from the phase correlation, and to zero (no perspective transformation). Also, we initially assume no exposure difference between images so is set to 1. Using these initial values we solve for a new set of model parameters and, using (25), we update our true parameter estimates. This iteration process continues until the percent decrease among successive iterations for the sum of squared and is differences between the parameter transformed smaller than some preset value. We have not used a hierarchical or multistage approach to performing the registration in this paper as done in many other registration approaches, but this approach is certainly amenable to that. Finally, once all adjacent images of our sequence are pair-wise registered, we choose one of the images in the sequence to serve as the “global” domain and range reference. Thus, using (25), we can use the local (pair-wise) domain and range transformations computed to determine the global transformations that take us to our desired coordinates in domain as well as range for each image in the sequence considered. IV. JOINT REGISTRATION EXPERIMENTAL RESULTS To test the approach presented, a sequence consisting of 19 images, each of size 240 320, was used. These images are depicted in Fig. 5. Please note that the camera, its parameters CANDOCIA: JOINTLY REGISTERING IMAGES IN DOMAIN AND RANGE BY PIECEWISE LINEAR COMPARAMETRIC ANALYSIS 417 TABLE II REGISTRATION DATA FOR THE IMAGE SEQUENCE IN FIG. 5 and the models used were the same as for the experiment described in Section II-C. Also, this sequence was chosen to illustrate the algorithm’s performance in the presence of reasonable model errors. That is, as the camera was hand-held and panned roughly about its optical center, a reasonable approximation for relating the spatial coordinates between images in the sequence is the homographic transformation. In range, our use of Mann’s “preferred” comparametric equation also yields a reasonable approximation to the camera’s actual response—please refer to Fig. 3 for an example. The sequence in question is of a room containing a sliding door with a brightly lit exterior. The camera begins inside the room and proceeds to pan across the sliding door from left to right, i.e., the time-ordered sequence begins with image (a) of Fig. 5 and continues until image (s). Because the inside of the room is dark relative to the brightly lit exterior, the camera automatically initially adjusts its exposure to capture the interior of the room in the top left image. As the camera begins to pan across the sliding door, the exposure time is automatically diminished so that now objects outside the sliding door can be discerned. However, this now leads to a “darkening” or underexposure of objects inside the room. Table II illustrates data obtained from the registration of the image pairs in this sequence. Note that the table contains results using domain-only registration as well as joint domain and range registration. The domain-only registration is easily performed within the framework presented by simply fixing in simply (27). That is, as an exposure isn’t being estimated, represents one of the two images being registered. As there are no parameters being associated with , any derivatives associated with it are zero and the resulting optimization is modified accordingly. Please note that image pair 1 referred to in Table II consists of images (a) and (b) of Fig. 5, image pair 2 refers to images (b) and (c), and so on. In both registration approaches, the mean absolute difference (MAD) between pixels of the registered image pairs is shown. Note that for all pairs, the MAD for domain-only registration is greater than when the joint registration is performed. This is intuitively satisfying because we expect a better match between pixels in the case when exposure differences—that we know exist in this sequence—are accounted for. This exposure difference, denoted by , is also listed in the table. This value describes that additional exposure into range correspondence with required to bring image image . For example, consider image pair 1. In order to register image (b) in Fig. 5 with image (a), we need to adjust the exposure of image (b) by a factor of 1.37, i.e., image (b) would have needed to be exposed 1.37 times more than it was in order to have it registered in range with image (a). In general, we can see that, as our camera panned across the sliding door, the exposure times between adjacent images continued diminishing, just as we would expect. We can see that, using our joint model’s group structure of (25), to bring image (s) in our sequence into range correspondence with image (a), we would need an exposure 653 times greater than that used in capturing image (s) (just multiply all the values of found in Table II). The additional accuracy in registration from the algorithm does come at a computational price. Notice that the domain-only registration process requires an average of about 5 seconds per image pair. This is in contrast to the joint registration process whose registration time per image pair varied between approximately one to two minutes in time. The algorithms were written and run using the Matlab 6.0 R12 package in the Windows 2000 environment on a Dell 330 Precision PC with a 1.5 GHz Pentium IV processor. To visually illustrate the results of the presented approach, the jointly registered sequence in Fig. 5 is shown in Fig. 6. Note that the domain and range reference being utilized is that of image (i) in Fig. 5. The gray lines superimposed on the figure are for illustrating the domain mapping that was computed for each image and how each of these images where overlain to produce the final composite. We can also notice that since these images were also registered in range, the images whose interior is darker (those captured after our reference image was) have now had their exposures synthetically increased so that they can match with that of the reference image. These now appear brighter in the final composite. Accordingly, the images captured before the 418 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 12, NO. 4, APRIL 2003 Fig. 6. Composite image resulting from the joint registration of the sequence in Fig. 5. The gray lines illustrate the mappings performed to obtain domain registration. The global reference used, both for domain and for range registration, is image (i) depicted in Fig. 5. reference image was (those with greater exposure times than the reference) have had their exposures decreased to match with that of the reference image. These images appear darker in the final composite. reviewers for providing helpful comments that were used in preparing the final manuscript. REFERENCES V. SUMMARY AND CONCLUSIONS An approach has been described whereby the joint registration of images in domain as well as range is accomplished using piecewise linear comparametric analysis. The attractiveness of this formalism lies in its ability to convert an inherently nonlinear estimation problem in both domain and range into one involving linear estimation. In both the domain and range models, an intermediate set of parameters is solved for using least squares and these are then easily related to the true parameters we seek to estimate. The method for estimating the exposure difference between images used a constrained piecewise linear model for the camera’s comparametric function. The accuracy with which the piecewise linear function approximates the comparametric equation depends on the number of segments one selects to use. We have empirically found that six equispaced segments provides good results while keeping the computational complexity low. Results indicate that the estimation of the exposure using the linear least squares approach presented here is as effective as estimation using nonlinear least squares with the original comparametric function. Experimental results have been conducted over many sets of images and illustrate that this procedure works well. ACKNOWLEDGMENT The author would like to thank the Telecommunications and Information Technology Institute at Florida International University for supporting this work as well as the anonymous [1] L. G. Brown, “A survey of image registration techniques,” ACM Comput. Surv., vol. 24, no. 4, pp. 325–376, Dec. 1992. [2] E. M. Mikhail, J. S. Bethel, and J. C. McGlone, Introduction to Modern Photogrammetry. New York: Wiley, 2001. [3] L. G. Shapiro and G. C. Stockman, Computer Vision. Englewood Cliffs, NJ: Prentice-Hall, 2001. [4] J. B. A. Maintz and M. A. Viergever, “A survey of medical image registration,” Med. Image Anal., vol. 2, no. 1, pp. 1–37, 1998. [5] R. Schultz and R. Stevenson, “Extraction of high-resolution frames from video sequences,” IEEE Trans. Image Processing, vol. 5, pp. 996–1011, June 1996. [6] M. C. Chiang and T. E. Boult, “Efficient image warping and super-resolution,” in Proc. IEEE Workshop on Applications of Computer Vision (WACV ’96), 1996, pp. 56–61. [7] U. Dhond and J. Aggarwal, “Structure from stereo—A review,” IEEE Trans. Syst., Man, Cybern., vol. 19, no. 6, pp. 1489–1510, 1989. [8] S. Baron and W. Wilson, “MPEG Overview,” SMPTE J., pp. 391–394, June 1994. [9] M. Irani, S. Hsu, and P. Anandan, “Video compression using mosaic representations,” Signal Process.: Image Commun., vol. 7, pp. 529–552, 1995. [10] H. S. Sawhney and S. Ayer, “Compact representation of videos through dominant multiple motion estimation,” IEEE Trans. Pattern Anal. Machine Intell., vol. 18, no. 8, pp. 814–830, 1996. [11] H. Y. Shum and R. Szeliski, “Systems and experiment paper: Construction of panoramic image mosaics with global and local alignment,” Int. J. Comput. Vis., vol. 36, no. 2, pp. 101–130, 2000. [12] M. Irani and S. Peleg, “Improving resolution by image registration,” Comput. Vis., Graph., Image Process.: Graph. Models Image Process., vol. 53, no. 3, pp. 231–239, 1991. [13] S. Mann and R. Picard, “Video orbits of the projective group: A simple approach to featureless estimation of parameters,” IEEE Trans. Image Processing, vol. 6, pp. 1281–1295, Sept. 1997. [14] S. Baker and T. Kanade, “Superresolution optical flow,” Robotics Institute, Carnegie Mellon Univ., Pittsburgh, PA, Tech. Rep. CMU-RI-TR-99-36, 1999. [15] H. S. Sawhney and R. Kumar, “True multi-image alignment and its applications to mosaicing and lens distortion correction,” IEEE Trans. Pattern Anal. Machine Intell., vol. 21, pp. 235–243, Mar. 1999. CANDOCIA: JOINTLY REGISTERING IMAGES IN DOMAIN AND RANGE BY PIECEWISE LINEAR COMPARAMETRIC ANALYSIS [16] S. Mann and R. W. Picard, “On being ‘undigital’ with digital cameras: Extending dynamic range by combining differently exposed pictures,” in Proc. of IS&T 48th Annual Conference, May 1995, pp. 422–428. [17] P. E. Debevec and J. Malik, “Recovering high dynamic range radiance maps from photographs,” in Proc. SIGGRAPH, 1997, pp. 369–378. [18] T. Mitsunaga and S. Nayar, “Radiometric self calibration,” in Proc. IEEE Conf. Computer Vision and Pattern Recognition, 1999, pp. 374–380. [19] S. Mann, “Comparametric equations with practical applications in quantigraphic image processing,” IEEE Trans. Image Processing, vol. 9, pp. 1389–1406, Aug. 2000. [20] S. Mann and R. Mann, “Quantigraphic imaging: Estimating the camera response and exposures from differently exposed images,” in Proc. IEEE Conf. Computer Vision and Pattern Recognition, 2001, pp. 842–849. [21] R. Fattal, D. Lischinski, and M. Werman, “Gradient domain high dynamic range compression,” ACM Trans. Graph., vol. 21, no. 3, pp. 249–256, 2002. [22] S. K. Nayar and T. Mitsunaga, “High dynamic range imaging: Spatially varying pixel exposures,” in Proc. IEEE Conf. on Computer Vision and Pattern Recognition, vol. 1, June 2000, pp. 472–479. [23] C. ChefD’Hotel, G. Hermilloso, and O. Faugeras, “A variational approach to multi-modal image matching,” in Proc. IEEE Workshop on Variational and Level Set Methods in Computer Vision, 2001, pp. 21–28. [24] A. Roche, A. Guimond, J. Meunier, and N. Ayache, “Multimodal elastic matching of brain images,” in Proc. of 6th Eur. Conf. on Computer Vision, Dublin, Ireland, June 2000. [25] Y. Y. Schechner and S. K. Nayar, “Generalized mosaicing,” in Proc. IEEE Conf. Computer Vision and Pattern Recognition, vol. 1, 2001, pp. 17–24. [26] M. Aggarwal and N. Ahuja, “High dynamic range panoramic imaging,” in Proc. IEEE Int. Conf. Computer Vision, vol. 1, 2001, pp. 2–9. [27] S. Mann, “Pencigraphy’ with AGC: Joint parameter estimation in both domain and range of functions in same orbit of the Projective-Wyckoff Group,” in IEEE Int. Conf. Image Processing (ICIP ‘96), vol. 3, Los Alamitos, CA, 1996, pp. 193–196. 419 [28] E. Kreyszig, Advanced Engineering Mathematics, 7th ed. New York: Wiley, 1993. [29] B. S. Reddy and B. N. Chatterji, “An FFT-based technique for translation, rotation and scale-invariant image registration,” IEEE Trans. Image Processing, vol. 5, pp. 1266–1271, Aug. 1996. Frank M. Candocia (M’94) received the B.S. and M.S. degrees in electrical engineering in 1991 and 1993, respectively, from Florida International University, Miami. In 1998, he received the Ph.D. degree in electrical and computer engineering from the University of Florida, Gainesville, where his work focused on optical image superresolution. During his doctoral studies, he also investigated superresolution of synthetic aperture radar imagery at MIT Lincoln Labs. From 1998 to 2000, he was a Senior Systems Engineer for Raytheon Systems Company, Bedford, MA, where he was responsible for the design and implementation of the search and acquisition portion of the National Missile Defense program’s X-band radar. He is currently an Assistant Professor with the Department of Electrical and Computer Engineering, Florida International University, and is the founder of the image understanding lab (http://iul.eng.fiu.edu). His research interests include all phases of the image understanding problem as well as the tools needed toward their eventual solution. Dr. Candocia has been the recipient of several scholarships and fellowships including a University of Florida graduate minority fellowship and a Motorola fellowship.
© Copyright 2026 Paperzz