IMAGING & THERAPEUTIC TECHNOLOGY 1769 The AAPM/RSNA Physics Tutorial for Residents Fluoroscopy: Optical Coupling and the Video System1 Michael S. Van Lysel, PhD In fluoroscopic/fluorographic systems, an image intensifier is optically coupled to recording cameras. The optical distributor is responsible for transmitting a focused image from the output phosphor of the image intensifier to the focal planes of the cameras. Each camera has an aperture, which is used to control the level of light reaching its focal plane. The aperture setting determines the patient x-ray exposure level and the image noise level. Increasing the x-ray exposure reduces image noise; reducing the x-ray exposure increases image noise. Fluoroscopic/ fluorographic systems always include a video camera. The functions of the video system are to provide for multiple observers and to facilitate image recording. The camera head contains an image sensor, which converts the light image from the image intensifier into a voltage signal. The device used to generate the video signal is a pickup tube or a charge-coupled device sensor. The method used is raster scanning, of which there are two types: progressive and interlaced. The vertical resolution of the system is primarily determined by the number of scan lines; the horizontal resolution is primarily determined by the bandwidth. Frame rate reduction can be a powerful tool for exposure reduction. Abbreviations: AEC = automatic exposure control, CCD = charge-coupled device, DSA = digital subtraction angiography, II = image intensifier, MTF = modulation transfer function Index terms: Fluoroscopy • Physics • Video systems RadioGraphics 2000; 20:1769–1786 1From the Departments of Medicine (Cardiovascular Medicine Section) and Medical Physics, University of Wisconsin, H6/333 Clinical Science Center, 600 N Highland Ave, Madison, WI 53792. From the AAPM/RSNA Physics Tutorial at the 1999 RSNA scientific assembly. Received June 29, 2000; revision requested July 18 and received August 14; accepted August 16. Address correspondence to the author (e-mail: vanlysel@facstaff .wisc.edu). ©RSNA, 2000 1770 November-December 2000 Introduction The fluoroscopic/fluorographic imaging chain is used in the performance of many diagnostic and interventional radiographic procedures. In this system, an image intensifier (II) is coupled to film cameras (cine, photospot) and a video camera. Video applications have become increasingly important. Combined with a digital image processor, video is replacing film in many applications. The purpose of this article is to discuss two components of this imaging chain: the optical distributor and the video system. After reading this article, the reader will be able to (a) describe how the optical distributor forms an image for the cameras to record; (b) describe how the camera aperture sets the patient x-ray exposure level and the image noise level; (c) describe the operation of tube-based and charge-coupled device (CCD)– based video cameras; (d) calculate the resolution limitations imposed by the video camera; and (e) explain the video terms lag, gamma, progressive scanning, interlaced scanning, fields, frames, standardline, high-line, and upscanning. Optical Coupling Overview The optical distributor is responsible for coupling the light image on the output phosphor of the II to the recording cameras. The cameras present on a fluoroscopic/fluorographic system can include a cine camera (35-mm motion picture film) or a photospot camera (100-mm cut film or 105-mm roll film). Systems always include a video camera. Initially, video cameras were present merely to assist in and monitor a procedure, which ultimately was recorded on one of the film cameras (or on a spot film or film changer placed at the entrance to the II). However, the quality and flexibility of video images have improved dramatically over the past 2 decades. Of particular note is the introduction of digital image processors, which rely on the video camera to provide an input signal. The result is that film cameras are becoming less common. This trend is especially true in cardiac imaging, in which most new systems are cineless; the recording and archival functions of the cine camera are replaced by the video/digital system. RG ■ Volume 20 • Number 6 The optical distributor is responsible for three functions: (a) transmission of a focused image from the II output phosphor to the focal planes of all cameras present on the system; (b) ensuring that the intensity (ie, brightness) of the images at the camera focal planes is correct for each camera; and (c) sampling the light level of the II output phosphor and transmitting this information to the automatic exposure control (AEC) circuitry of the x-ray generator. Optical Distributor Components Let us look at the components of an optical distributor that has both a video camera and a cine film camera (Fig 1). During cine operation, a semisilvered, beam-splitting mirror simultaneously provides images to both cameras. During coronary angiography, for example, this mirror allows realtime monitoring of the injection with the video camera while images are recorded on film. (It is typical to also videotape the video signal as backup against loss of the film during processing.) During fluoroscopy, the mirror is removed from the beam path so that all of the light passes to the video camera. The beam path from the II output to each camera contains two lenses. The collimating lens (also referred to as the objective lens), which collects light emitted from the II output phosphor, is common to all paths. Each camera has its own camera lens. Each camera also has its own aperture, which is used to control the light level reaching the focal plane of the camera. Other names used for the aperture include f-stop, diaphragm, and iris. Also included in the beam path is a small mirror or prism for directing light to the AEC sensor. The sensor can be either a photodiode or a scintillating crystal coupled to a photomultiplier tube. A lens is used to form an image, then a field stop is used to prevent light from the periphery of that image from reaching the sensor. It is not uncommon for the edge of the image to consist of unattenuated radiation. If the AEC were to include this region in its determination of exposure level, the central portion of the image, which is presumably the area of interest, would be underexposed (ie, dark and noisy). If only the central 50% (for example) of the image is sent to the AEC sensor, this problem is avoided. Optical Considerations The optical elements in the distributor are always arranged in a particular manner. The distance between the II output phosphor and the collimating RG ■ Volume 20 • Number 6 Van Lysel 1771 Figure 1. Components of an optical distributor servicing a video camera, a cine camera, and the AEC of the generator. FL = focal length. lens is equal to the focal length of the collimating lens (ie, the phosphor is in the focal plane of the collimating lens) (Fig 1). As a result, the light leaving the lens is collimated (ie, the light rays that were emitted from a single point on the phosphor are parallel after they leave the lens). The image is said to be “focused at infinity.” The video camera target is placed in the focal plane of the video camera lens, and the film in the cine camera is placed in the focal plane of the cine camera lens (Fig 1). Because the light received by each camera lens is collimated, the image formed by the lens is at the focal plane. Cameras can go out of focus over time as vibrations cause the distance from the lens to the target or film to change. Periodic quality control measurements are necessary to detect this problem. A property of collimated light is that every region within the beam path contains light from all points on the output phosphor (however, there are limits to this statement that, if not heeded by the optical designer, will lead to vignetting, or darkening of the image periphery). As a result, when objects such as an aperture or the small mirror for the AEC sensor are placed in the beam path, these objects do not show up in the image. The intensity of the image is reduced due to the reduction in light reaching the image, but this reduction is uniform across the entire image. This fact is why apertures are always placed between the two lenses. In contrast, the field stop in the beam path for the AEC sensor is placed after the second lens. At this location, after the image has been formed, the field stop serves to occlude light from the image periphery; as described earlier in this article, such occlusion of light is desirable for operation of this device. The most important practical matter to understand regarding the optical distributor is the function of the camera apertures. To understand that, we need to discuss the concept of the speed of the optical system. The term speed is used here in the same manner that it is used for a screen in screen-film radiography. For a fixed dose rate, the faster the optical system, the less total x-ray exposure is required to properly expose the camera. The f-number of the optical system (often written f/#) designates the system’s speed. Amateur photographers will recognize that it is the lens f/# that they vary to adjust for the brightness of the scene they are photographing. The f/# of a lens is defined as the ratio of its focal length to its diameter (technically, the diameter of the entrance pupil): f /# = focal length . diameter (1) 1772 November-December 2000 RG ■ Volume 20 • Number 6 Figure 2. Relationships governing the speed of a lens. FL = focal length. These two properties determine the brightness of the image formed by the lens (Fig 2). The size of the image is in direct proportion to the focal length; longer focal lengths form larger images. Because the lens collects a fixed amount of light, larger images must necessarily be dimmer, since the same number of light photons are spread over a larger area. Doubling the focal length quadruples the area of the image, which is then only onefourth times as bright. The tandem lens system forms an image at the image receptor plane (ie, the video camera target or the film plane). The size of the image is a function of the focal lengths of the collimating and camera lenses. The system designer chooses the focal lengths of the two lenses to match the image size with the size of the receptor. Image size selection is not a straightforward issue because the image of the II output phosphor is round, but the film frame and video scanning pattern are rectangular (or square). Thus, the image can be made large, improving resolution but truncating the edges, or it can be made small, allowing the entire II image to be viewed but significantly limiting resolution. Usually, a medium position is chosen. This issue of framing, in the context of cine, is discussed in detail in reference 1. Although the focal lengths of the lenses are fixed and thus can be ignored during everyday Figure 3. Two different aperture styles: a fixed aperture (left) and a variable, motor-driven, “iris-style” aperture (right). Also included with the variable aperture is a neutral density filter. operation of the system, the aperture can be adjusted to change the speed of the system. Introduction of an aperture that is smaller than the lens diameter reduces the brightness of the resulting image (Fig 2). The reduction in brightness is in direct proportion to the reduction in the effective area of the lens (thus it is also in proportion to the square of the diameter). Apertures allow the system to operate over a wide range of x-ray exposure levels, as described later in this article. Two different aperture styles are used in fluoroscopic/fluorographic systems (Fig 3). Generally RG ■ Volume 20 • Number 6 Van Lysel 1773 Figure 4. Sample calculation of aperture f/# versus II entrance exposure (Exp). D = aperture diameter, TV = television camera. speaking, clinical procedures that make use of a film camera are performed at the same “detected” x-ray exposure level for all patients (eg, 16 mR [4.1 ´ 10-9 C/kg] per frame incident on the II for coronary angiography in the 17-cm mode); thus, a fixed aperture is used for these cameras. When the angiographic unit is installed, the service engineer determines the aperture size that will produce a properly exposed film at the desired x-ray exposure level. Periodically, this aperture may need to be replaced with one of a different size as the II gain falls with age, if the II is replaced, or if the film type or the desired x-ray exposure level changes. In contrast to film cameras, video cameras must operate over a wide range of x-ray exposures: from a low of 2 mR (5.2 ´ 10-10 C/kg) per frame for fluoroscopy to potentially 1,000 mR (2.6 ´ 10-7 C/kg) per frame for digital subtraction angiography (DSA). This large range in x-ray exposure is accompanied by an equivalent range in II output brightness. This range is too great for the video camera, so the video camera aperture is typically configured as a motor-driven iris. The iris is opened up for fluoroscopy, when the II output phosphor is dim; it is “stopped down” during photospot imaging or DSA, when the phosphor is bright. At very high exposures, it may be necessary to stop down the aperture to such a small opening that it is difficult to precisely control. In this case, a neutral density filter (Fig 3) can be swung into the opening to provide additional light attenuation. A sample calculation illustrates the relationship between f/# and the x-ray exposure incident on the II (Fig 4): If the f/# of the video camera lens is f/2 during fluoroscopy performed at 2 mR per frame (Exp1), what is the proper f/# for the lens during cine angiography performed at 16 mR per frame (Exp2)? We answer this question by noting that an eightfold increase in entrance exposure results in an eightfold increase in II output brightness. Therefore, to keep the light intensity reaching the video camera constant, the area of the camera aperture must be decreased by eight times. Working through the calculation (Fig 4), we see that to decrease the light transmitted by the aperture by eight times, we need to increase the f/# (note the reciprocal relationship) by Ö8, to f/5.6. The reason we increase x-ray exposure, of course, is to reduce the x-ray noise amplitude in RG ■ Volume 20 • Number 6 1774 November-December 2000 a. b. Figure 5. Comparison of x-ray noise amplitudes in coronary angiograms acquired at fluoroscopic (2 mR per frame) (a) and angiographic (16 mR per frame) (b) exposure levels. Figure 6. Major components of a fluoroscopic/fluorographic video system. ADC = analog-to-digital converter, DAC = digital-to-analog converter, OD = optical distributor, VCR = videocassette recorder. our images. The noise amplitude is inversely proportional to the square root of the II entrance exposure. Thus, in the example given in the preceding paragraph, an eightfold exposure increase produces a 0.35 times noise amplitude. When coronary angiograms are obtained at fluoroscopic (2 mR per frame) and cine angiographic (16 mR per frame) exposure levels, the higher level of noise in the fluoroscopic image is apparent (Fig 5). Video System Overview When IIs were introduced in the early 1950s, the operator viewed the output phosphor “directly” through a system of lenses and mirrors. Video systems were soon introduced. The functions of the video system are to provide for multiple observers and to facilitate image recording. In addition, the video system provides the input to the RG ■ Volume 20 • Number 6 Van Lysel 1775 Figure 7. Operation of a video camera pickup tube. E = electric field, G4 = electrode for removing excess electrons, Ibeam = focused electron beam, Isig = signal current, VOUT(t) = video signal output, +VT = target voltage. digital image processor, which is now present in virtually all fluoroscopic systems. Digital imaging has driven major improvements in video system image quality to the point where digital video is displacing both photospot and cine cameras in most new installations. The components of a closed-circuit video system used for fluoroscopy/fluorography are as follows (Fig 6): The camera head is attached to the optical distributor, allowing it to view the II output phosphor. The camera head contains an image sensor that converts the light image from the II into a voltage signal that can be transmitted to other components of the video system, where it can be modified, displayed, and recorded. The camera control unit, which is usually located in one of the generator’s electronics racks, processes the video signal from the camera head so that it is suitable for display and digitization. The camera control unit also synchronizes the scanning of all of the components of the system by generating and transmitting horizontal (H sync) and vertical (V sync) synchronization signals. Signal Generation Two distinctly different devices are now in use for generation of the video signal voltage from the incident light image. The traditional device is a vacuum tube sensor referred to as a pickup tube. A pickup tube operates as follows (Fig 7): Light photons absorbed by the photoconductive target generate electron-hole pairs in the target material. The holes travel to the inside surface of the target under the influence of an electric field (E), which is placed across the target by the target voltage (+VT). These holes collect on the inside target surface, where they represent a latent image. Hole density is in proportion to the intensity of the light image falling on the target at that location. To read the charge-density spatial distribution, a focused electron beam (Ibeam) is scanned across the inside surface of the target under control of the deflection coils. Electrons from the beam land on the target and combine with the holes. The number of electrons that land on the target is equal to the number of holes at that location. Excess electrons are removed by an electrode (G4), which is located just above the target surface. Electrons landing on the target cause a proportional current to flow in an external load resistor (Isig). The resulting voltage drop across this resistor constitutes the video signal output from the pickup tube (VOUT[t]). The other technology now used to generate a video signal replaces the pickup tube with a CCD sensor. A CCD is a solid-state device. As with the target of a pickup tube, absorption of light photons in the CCD semiconductor material generates electron-hole pairs. In this case, the electrons are held at the location where they are generated by a two-dimensional array of potential RG ■ Volume 20 • Number 6 1776 November-December 2000 Figure 8. Readout operation of a CCD video camera. VOUT(t) = video signal output. Figure 9. Video camera lag. Because of lag, the video camera output (VOUT[t]) does not respond instantaneously to an increase or decrease in light level. wells set up in the material. Electrons collect in the potential wells in proportion to the number of light photons incident at that location. The frame-transfer method (Fig 8) is a common method used to read out the CCD chip. (Although the example shows pulsed x-ray generation, continuous x rays can also be used.) After termination of the x-ray pulse, the charge in the wells is rapidly transferred to an adjacent storage region. The storage region consists of a second array of potential wells similar to the sensor region. However, the storage region is shielded from the incident light falling on the sensor region. Charge is transferred in a method that has been referred to as a “bucket brigade.” Each well Figure 10. Effect of camera lag. Angiogram of a rapidly moving coronary artery shows a trailing “ghost” due to excessive camera lag (the direction of travel is from right to left). transfers charge to an adjacent well. This charge transfer operation continues until all of the charge resides in the storage array. The sensor region is now ready to accept the next image, whereas readout of the storage region can extend over the entire frame period. CCD cameras have been available for generalpurpose fluoroscopy systems for several years. In general, they have not been of sufficient quality to allow their use in high-resolution angiography suites, but recently they have begun to appear in these systems as well. The advantages of CCDs over pickup tubes include small size, low power RG ■ Volume 20 • Number 6 Van Lysel 1777 Figure 11. Signal transfer curve. Graph shows video signal as a function of light input to the camera for gamma values of 0.7 and 1 (linear response). For a gamma of less than 1, small signals in the dark parts of the image are amplified more than equivalently sized signals in the bright parts of the image. consumption, ruggedness, no geometric distortions, no lag or burn-in phenomenon, long lifetime, and no setup or maintenance tuning. Pickup tube performance degrades over time. For example, the expected lifetime of an $8,000 Plumbicon pickup tube in an angiography suite is 18 months. Proper pickup tube operation requires careful setup by an experienced service technician. Two operating characteristics of a video system that affect image quality and that users should understand are lag and linearity. Video camera lag is described as follows (Fig 9): When the light level falling on any region of the video sensor changes (either increasing or decreasing), it is desirable for the video signal out of the sensor to change instantaneously to its new value. The term lag refers to the condition in which the video signal output rises or falls to the new value more slowly than the changing light input. It is sometimes said that the camera is “sticky.” The magnitude of lag varies greatly between different types of video cameras, depending on the details of target construction and materials. CCD cameras have no lag. At the other extreme, a photoconductive pickup tube called a vidicon has a very high degree of lag. The vidicon uses SbS3 as the target material. In the middle are several pickup tubes (eg, Plumbicon, Saticon, Chalnicon, Newvicon) that have different target configurations but all deliver about the same amount of lag. These “other” tubes are referred to as “low-lag” cameras but actually are low lag only when compared with the vidicon. In fact, the amount of lag for these cameras is appreciable. Lag is signal dependent; it is much higher in the dark parts of the image than in the bright parts. Lag increases as a pickup tube ages. Tube replacement is often prompted because the tube has become excessively “laggy.” One clinically relevant effect of camera lag is blurring of objects in motion, since a moving ob- ject leaves behind a trail of decaying signal (Fig 10). As a result, the vidicon pickup tube cannot be used in angiography suites, where motion blurring would seriously degrade the spatial resolution of moving vessels. On the beneficial side, lag provides a degree of noise integration, reducing x-ray noise amplitudes in the resulting images. The increase in noise levels on a generalpurpose fluoroscopy system when a vidicon is replaced by a CCD camera is profound. Artificial lag can be provided in these cases by a recursive filter implemented in the digital image processor. The second video characteristic that should be understood by users is the video signal transfer curve. Transfer curves describe the relationship between light input and video signal output. Because transfer curves can be described by the following expression: V Vo æ ö = ççç L ÷÷÷÷ èL ø g , (2) o where V is signal voltage and L is input light intensity, curves are usually characterized by their gamma value (2). The vidicon has a gamma of about 0.7, whereas all of the other pickup tubes and CCD cameras have a gamma of unity ( g = 1 indicates a linear response). However, because a gamma of less than 1 has desirable properties, many systems with linear cameras impose a nonlinearity on the video signal. Imposition of nonlinearity can be done by either the camera control unit or the digital image processor. Small signals in the dark parts of an image are amplified by a transfer curve with a gamma of less than 1, allowing better visualization of these difficult-to-see signals (Fig 11). RG ■ Volume 20 • Number 6 1778 November-December 2000 Figure 12. Horizontal line in raster scanning. Each horizontal scan line is commenced and terminated by a horizontal synchronization pulse (H sync) generated by the camera control unit. TH = horizontal line period. Video Scanning The light image falling onto the video camera target is two-dimensional. The camera transforms this two-dimensional image into a one-dimensional time-varying voltage by raster scanning. The pickup tube scans the target with an electron beam. The CCD camera scans by shifting the accumulated charge in shift-register fashion out of the sensor region. The fundamental unit of raster scanning is the horizontal line (Fig 12). Each line is terminated (and each new line is commenced) by a horizontal synchronization (H sync) pulse generated by the camera control unit. Video images are scanned in the same manner in which we read a book. A line is slowly read from left to right. When we reach the end of the line, we rapidly shift our eyes back to the far left, and a little bit down, to commence reading the next line. When we’ve read the last line on the page, we rapidly shift our gaze to the upper left corner of the next page. In a video system, the H sync pulse instructs the defection circuitry in both the pickup tube and the monitor to rapidly retrace to the beginning of the next line. In a pickup tube, the electron beam is turned off (the beam is said to be “blanked”) while this retrace occurs, so that the yet-to-be-read charge on the target is not disturbed. Retrace results in dead time each horizontal line during which no image information is transmitted. This dead time, which amounts to about 15% of the total horizontal period, will be important when we calculate video bandwidths later in this article. Similarly, the camera control unit transmits a vertical synchronization (V sync) pulse to signal all of the components in the system that it is time to perform vertical retrace back up to the top of the image. Vertical retrace results in dead time equal to about 7% of the total vertical period. The book-reading analogy suffers a bit when we consider vertical scanning in more detail. There are two ways in which vertical scanning can be performed. Progressive scanning (also known as sequential scanning) does conform to the analogy. In progressive scanning, lines are scanned in consecutive order. Line 2 is scanned after line 1. Line 3 is scanned after line 2. When we get to the bottom of the image, we will have scanned all of the lines. Interlaced scanning is a bit more complicated. After scanning line 1 we scan line 3, skipping line 2. After scanning line 3 we scan line 5, skipping line 4. We scan only the odd lines, reaching the bottom of the image in one-half the time required for progressive scanning. However, having read only half of the lines, we must go back to the top of the image and read the even lines in succession (lines 2, 4, 6, etc). The odd and even fields are interlaced to produce one frame. In this discussion, it is important to carefully note the difference between fields and frames. Interlaced scanning was devised in the early days of broadcast television as a means to reduce the required bandwidth of video transmissions, allowing more channels to be broadcast within a market. Successful video transmission relies on the fact that the eye-brain system of the observer perceives the scanned image as continuous. In reality, the monitor phosphor begins to decay in brightness as soon as the electron beam moves on (see the discussion of monitor operation later in this article). After a few hundred microseconds, the phosphor is dark again. If your visual system were fast enough, you would see a bright bar, representing the image, sweeping down the face of a predominantly dark monitor. The monitor refresh rate is set so that your visual system blends the scanning bright bar into a continuous, bright image. The “flicker-fusion frequency” at which this blending occurs is a variable that depends on image brightness but generally requires a refresh rate of at least 50 images per second. RG ■ Volume 20 • Number 6 Van Lysel 1779 Figure 13. Scan conversion. Progressive-to-interlaced scan conversion writes a progressively scanned image into one memory while reading the previous frame, stored in a second memory, in interlaced fashion for display. ADC = analog-todigital converter, TV = television camera. The standard scanning mode employed by conventional broadcast video is described as 30 frames per second, 525 lines per frame, 2:1 interlaced video. The 2:1 interlace process was described earlier; two fields are interlaced to produce one frame. The 525 lines per frame means that there are 525 horizontal lines making up one frame (one “image”). Thirty of these 525-line images are transmitted every second. We know that 30 frames per second is not fast enough to avoid a flickering image. However, when viewed from a sufficient distance, the individual fields, presented at 60 fields per second, blend together enough (because of the limited resolution of the human eye) to greatly attenuate the perception of flicker. For many years, the video standard used in fluoroscopic/fluorographic systems was the same as that used in conventional broadcast television, 30 frames per second, 525 lines per frame, 2:1 interlaced. This system is still used today for inexpensive systems. Substantial cost savings are enjoyed by using off-the-shelf video components (eg, cameras, monitors, tape recorders). However, as the use of video has shifted from merely monitoring a procedure to being the primary diagnostic medium, and also the means to perform sophisticated interventions, the quality of standard video is no longer sufficient. One improvement is greater use of progressive scanning. Although interfield flicker is usually not apparent under casual observation of an interlaced display, close inspection of the image (as is likely to occur during medical procedures) does reveal an annoying flicker, especially at a horizontally oriented boundary between bright and dark regions. This flicker is due to the fact that the 30 frames per second display rate is, in fact, below the flickerfusion frequency. The best systems now employ progressive scanning monitors with display rates of 60 frames per second or above (as opposed to a 60 fields per second, 30 frames per second, interlaced scanning monitor). More important than display considerations, interlaced scanning by the video camera during acquisition can cause significant image artifacts. The nature of the artifact depends on the manner of x-ray production. The most common artifacts are distortions of moving objects, such as coronary arteries and guide wires. For both continuous fluoroscopy and pulsed fluoroscopy/fluorography at a rate of 60 pulses per second, the moving object is in two different locations for the even and the odd fields. The resulting image of the artery in the interlaced frame will have a serrated appearance (3). For pulsed operation, reduction of the x-ray pulse rate to 30 pulses per second eliminates the serration artifact, but a more severe artifact is produced. In this case, the intensity of the two fields is very different, resulting in severe flicker. Cardiac angiographers who were practicing before the advent of digital imaging are familiar with the flashing that occurred on the video monitor during 30 frames per second cine-film imaging. This flashing occurs with a pickup tube system because, while the scanning electron beam is reading the odd field, it inevitably reads a great deal of charge that properly belongs to the even field, depressing the intensity of the even field when it is read next if a new x-ray pulse does not replenish the lost charge. A very common method used in modern fluoroscopic video systems to provide both the superior performance of progressive scanning of the video camera during acquisition and the economy of an interlaced display is scan conversion (Fig 13). Scan conversion requires a digital image processor to buffer the data received from 1780 November-December 2000 RG ■ Volume 20 • Number 6 Figure 14. MTFs of the video system, the II and x-ray tube focal spot (“everything else”), and the total system for two clinically relevant examples. Assumed II mode, focal spot (fs) size, and number of video lines are indicated. In the peripheral DSA case, the focal spot size (0.3 mm) is appropriate for fluoroscopy during the DSA procedure; the angiographic images acquired during this procedure would be performed with a larger focal spot. the camera until it is time for display. In an analog system without a buffer, the signal read from the camera is transmitted immediately to the display monitor. Without a short-term storage method, both the camera and the monitor must scan in lockstep. Therefore, if the monitor is a conventional interlaced scanning model, the camera target must be read in interlaced fashion. However, with a digital buffer, acquisition and display are decoupled. The camera reads progressively, and the data are stored in a digital memory. Once one frame has been read, the camera begins to fill a second memory, freeing the first memory to be read out in interlaced fashion for display. Spatial Resolution In this section, we will calculate the spatial resolution limitations placed on the radiographic image by the video system. However, it is important to understand that there are many factors that can contribute to the final resolution in a fluoroscopic image. These include the video system, the digital system, the II, the x-ray tube focal spot, and patient motion. Under some circumstances, the video system is the major limiting component to resolution. Under other circumstances, the video system may impose no significant limitations because one or more of the other components are limiting. These two cases can be demonstrated by plotting the modulation transfer functions (MTFs) of a video system, the other components (eg, the II and x-ray tube focal spot), and the total system (Fig 14). The total MTF is the product of the video and “everything else” MTFs: MTFtotal (u) = MTFvideo (u) ´ MTFeverything else (u) , (3) where u is the spatial frequency in cycles per millimeter. In the coronary angiography example, the MTF of the high-line video system is significantly better than the MTF of the II and 1-mm focal spot. The result is that the additional resolution loss imposed on the total imaging system by the video component is small. In contrast, in the peripheral DSA example, MTFvideo and MTFtotal are virtually identical, indicating that the resolution limitation of this system is dominated by the video system. In such a case (specifically, when a large-II format is used), a standard 525-line video system is insufficient to provide adequate resolution. The reader should understand this point after the following discussion. The spatial resolution of the video system is determined by separate considerations in the vertical and horizontal directions (Fig 15). Vertical resolution is primarily determined by the number of horizontal lines the system uses. Horizontal resolution is primarily determined by the bandwidth of the video system. Unlike IIs and screenfilm systems, which tend to have rotationally symmetric resolution properties, the resolution of the video system can potentially be very different in the vertical and horizontal directions. However, the video system designer usually chooses to make the resolution in the two directions approximately equal. Vertical Resolution.—Let us perform a simple calculation of the limiting vertical resolution of a conventional 525-line video system, which we might measure using a line-pair gauge. First, we must understand a peculiarity of video terminology. There are not 525 visible lines in a 525-line system. Remember that approximately 7% of the vertical period is spent retracing from the bottom of the image back up to the top. As a result, there are 525(1 - 0.07) . 490 active scan lines available to transmit image data. The remaining 35 lines RG ■ Volume 20 • Number 6 a. b. Figure 15. Primary determinants of video spatial resolution: number of scan lines (a) and video system bandwidth (b). Vertical resolution is the resolution you would measure when a linepair gauge is positioned as in a. Horizontal resolution is the resolution you would measure when a line-pair gauge is positioned as in b. Van Lysel 1781 Figure 16. Degradation of vertical resolution by vertical sampling. Figure 17. Calculation of the video resolution referenced to the object plane. mag = magnification, res = resolution, SID = source-to-image distance, SOD = source-to-object distance, TV = television camera. (or, more accurately, the remaining time that would have been spent scanning these 35 lines) are spent in vertical retrace. With 490 lines, we can represent 490/2 = 245 line pairs. We can therefore say that our vertical resolution is 245 line pairs per scan height. Actually, we should say that the maximum achievable vertical resolution is 245 line pairs per scan height. Several factors conspire to reduce the typical vertical resolution from that which we just calculated. For one, the electron-beam spot in the pickup tube is not pillbox-shaped (with sharp edges) but tends to be more Gaussian, resulting in overlap between the lines (especially during interlaced scanning). More significantly, how the line-pair gauge lines up with the raster scan (ie, its phase) can significantly affect the outcome of our observation (Fig 16). The calculation performed earlier assumes that the spaces and bars of the line-pair gauge line up exactly with the scan lines of the video system. If the line-pair gauge is shifted by half the width of a line (90° phase shift), contrast (modulation) in the resulting image disappears, since the signal read by each video line represents half a bar and half a space. In the interest of quoting a single specification for limiting resolution, it is common to reduce the maximum achievable vertical resolution by an amount referred to as the Kell factor, after R. D. Kell, who researched this issue in the 1930s. A typically quoted value for the Kell factor is 0.7, resulting in a practical vertical resolution for our 525-line system of 245(0.7) . 170 line pairs per scan height. Line pairs per scan height is a bit removed from what we really want to know: What is our resolving power (eg, in line pairs per millimeter) in the object plane (ie, the patient)? This conversion is performed by referencing resolution to the object plane (Fig 17). Line pairs per scan height specifies our resolution for objects located on the RG ■ Volume 20 • Number 6 1782 November-December 2000 Figure 18. Determination of bandwidth by proper selection of the video system components. The horizontal resolution of a properly operating video system is set by the system bandwidth. The frequency-response characteristics of the video electronics are represented here by a low-pass filter. The consequences of bandwidth that is too low or too high are shown as well. CCU = camera control unit, TV = television camera, VOUT(t) = video signal output. video camera target. We can translate this value into line pairs per millimeter at the face of the II by simply dividing by the diameter of the relevant II mode. We then translate that value into line pairs per millimeter at the object plane by multiplying by the relevant geometric magnification. The geometric magnification, which is a consequence of the fact that the x-ray field is diverging, is calculated as follows: mag = SID , SOD (4) where SID is the distance from the source to the II face (source-to-image distance) and SOD is the distance from the source to the object plane (source-to-object distance). As an example, calculate the object-plane resolution of a 525-line video system when the 23-cm II mode is used. Assume that the geometric magnification is 1.2 and the Kell factor is 0.7: res = 490 æç line pairs ö÷ ÷ ´ 0.7 ´ 1 ´1.2 ç 2 çè scan height ø÷ 230 mm = 0.9 ççèæç lp ÷÷÷öø mm . (5) The Table presents video resolution as a function of II mode. Resolution is seen to decrease with increasing II size, since a fixed number of video lines are stretched across an increasing field of view. What is striking about this result is how poor the video resolution is. Compare a video resolution of approximately 1 lp/mm with the ap- Vertical Object-Plane Limiting Resolution of Video Systems as a Function of II Mode Limiting Resolution (lp/mm) II Diameter (cm) 525-Line* System 1,049-Line† System 36 23 17 13 0.6 0.9 1.2 1.6 1.1 1.8 2.4 3.2 Note.—Assumptions include a geometric magnification of 1.2 and a Kell factor of 0.7. *490 active lines. † 980 active lines. proximately 5 lp/mm limiting resolution of the II. Film resolution (either cine or photospot film) is so high that the imaging chain resolution for these modalities is limited to that provided by the II. As the role of the video system expanded in the 1980s and 1990s, it became clear that video resolution had to be improved. The result was the adoption of high-line video for angiography and interventional systems. Like the current plans for high-definition television for broadcast video, high-line video systems typically double the number of lines (eg, from 525 to 1,049 lines), along with a concurrent increase in video bandwidth (as discussed later in this article). Doubling the number of video lines doubles the vertical resolution of the video system. As seen in the Table, the combination of a high-line system and a small II mode produces a respectable resolution (eg, 3 lp/mm). However, the application in which high-line systems are crucially important is large II formats. RG ■ Volume 20 • Number 6 Horizontal Resolution.—Resolution in the horizontal direction is limited by the bandwidth of the video system. The concept of video bandwidth can be explained by drawing an analogy with the stereo systems that many people have in their homes. When designing the electronics for an audio system, the designer makes use of the fact that the human ear is capable of detecting audio frequencies of up to 20,000 cycles per second (20 kHz). The designer therefore makes sure that the electronics are capable of transmitting voltage waveforms of up to 20 kHz (ie, that the bandwidth is 20 kHz). If the designer does not do so, the listener will not hear the high-frequency tones in the music because they were not passed from the source to the speakers by the electronic components (such as the amplifier). Conversely, any money or effort spent to pass frequencies above 20 kHz is wasted because the listener cannot hear them. This analogy can be extended to video systems by imagining a low-pass filter at the end of the amplifier chain that transmits the signal from the camera to the monitor (Fig 18). A low-pass filter passes all frequencies up to a cutoff frequency, attenuating frequencies above the cutoff frequency. We can equate the video system bandwidth with this cutoff frequency. It is the task of the system designer to make sure that this filter has the appropriate frequency-transfer characteristics. The consequences of improper bandwidth selection are significant (Fig 18). Say we desire to transmit the image of a line-pair gauge. If the bandwidth is set too low, a rounded-off, blurry representation of the gauge will be passed to the video monitor. We will have degraded the horizontal resolution of the video system. However, we also want to make sure that we do not set the bandwidth too high because there is a great deal of electronic noise, generated by the video system, at high frequencies. Passing these frequencies through to the monitor will result in an excessively noisy image. The appropriate bandwidth is the one that yields a horizontal resolution equal to the vertical resolution. Since the vertical resolution is dictated by the number of video lines, it is straightforward to calculate this bandwidth. Step 1 is to determine how many cycles we require along a horizontal line. What we mean by “cycles” are sinusoidal variations from bright to dark (or from Van Lysel 1783 high voltage to low voltage). For our purposes, we can equate cycles with line pairs, even though the equivalence is not exact. Say we wish to calculate a bandwidth for the standard 525-line video system. We know from the earlier discussion that the maximum vertical resolution is 245 line pairs per scan height. Therefore, to achieve equal horizontal and vertical resolution, we require that the video system pass up to 245 cycles per scan line. Step 2 is to determine how much time is spent scanning one line. In our example, we have a 525 lines per frame system that scans 30 frames per second. Therefore, the period of a single line is as follows: TH = 1 æ lines ö÷ æç frames ö÷ çè frame ø÷´ ÷ 30 ççè sec ÷÷ø 525 çç = 63.5 m sec . (6) However, we learned earlier that 15% of this period is spent in horizontal retrace, during which no information is transmitted by the system. Therefore, we have only 63.5(1 - 0.15) = 54 msec to transmit our 245 cycles. Step 3 in our bandwidth calculation is to divide the required number of cycles per scan line (245 cycles) by the time we have to transmit them (54 msec): 245 cycles = 4.5 ´106 cycles = 4.5 MHz . (7) sec 54 ´ 10-6 sec The only other video scanning mode typically encountered in a fluoroscopy system is the highline mode discussed earlier. The required bandwidth for a 1,049 lines per frame, 30 frames per second scanning mode is calculated as follows: BW = æ lines ö÷ 980 æç cycles ö÷ ÷ 1, 049 ççç ´ ÷ ç ÷ ç 2 è line ø è frame ø÷ ´ 30 æçççè frames ÷÷÷öø ´ sec » 18 MH z . 1 (1 - 0.15) (8) RG ■ Volume 20 • Number 6 1784 November-December 2000 In this calculation, it is assumed that there are 980 active scan lines (69 are lost to vertical retrace) and that the active horizontal period is reduced by 15% due to horizontal retrace. The reader will note that the high-line bandwidth is four times that of the standard-line mode. This increase is a result of two facts: (a) We require twice as many cycles per scan line to increase the horizontal resolution by the same amount as the improved vertical resolution (the equal vertical and horizontal resolution criterion) and (b) the horizontal period is half as long (31.75 msec vs 63.5 msec) because we need to scan twice as many lines in the same 1/30th of a second frame period. What if the bandwidth of a particular system is less than that required? We are now in a position to calculate the resulting resolution. Let us assume that the video system in question has a bandwidth of only 3 MHz. If we are imaging in the 23-cm II mode, the limiting resolution at the face of the II is calculated as follows: æ cycles ÷ö ´54 ´10èç sec ÷÷ø = 0.7 æçççè lp ÷÷÷öø . mm res = 3 ´ 106 çç 6 çççæ sec ÷÷÷ö ´ 1 çççæ lines ÷÷÷ö è line ø 230 è mm ø (9) If the geometric magnification was 1.2, then the object plane resolution would be 0.7 ´ 1.2 = 0.85 lp/mm. Frame Rate.—Inspection of the relationships in the previous section indicates a direct proportionality between frame rate and bandwidth. Doubling the frame rate doubles the required bandwidth because it halves the horizontal line period. However, a point that is sometimes misunderstood is that lower x-ray pulse rates do not typically reduce the required video bandwidth. It is becoming increasingly common to perform both fluoroscopy and fluorography at reduced pulse rates (eg, 15 or 7.5 pulses per second) to reduce x-ray dose (as discussed later in this article). It is common to refer to these rates as 15 or 7.5 “frames” per second. However, the implication that the video system is operating at a reduced frame rate is usually incorrect. For example, both coronary angiography and pulsed fluoroscopy are now often performed at “15 frames per second.” Figure 19. Relationship between x-ray pulses and video camera scanning for reduced frame rate coronary angiography (also applicable to reduced frame rate pulsed fluoroscopy) and for slow frame rate (eg, 1 frame per second) DSA. The “high” state indicates “x rays on,” and the “low” state indicates “video signal blanked.” fps = frames per second. The typical manner in which this rate is achieved is to produce x-ray pulses at 15 pulses per second but continue to scan the video camera at 30 frames per second. The first camera scan after the x-ray pulse contains the video information from that pulse (Fig 19 [middle waveform]). This frame is both sent to the video monitor for display and stored in a digital frame buffer. The second time the camera target is scanned, there is no new information because there was not an x-ray pulse. Therefore, the first frame, stored in the frame buffer, is displayed again. This process retains the necessary 30 frames per second display rate (to avoid display flicker). At the beginning of the third frame, a new x-ray pulse provides new video information. When the x-ray pulse rate is less than 30 pulses per second, a less often used alternative is to use the additional time to reduce the frame rate at which the camera is scanned. The advantages of this slow-scan mode include a lower bandwidth (and thus less video noise) and reduced demand on the analog-to-digital converter of the digital image processor. When the x-ray pulse width becomes greater than about 10 msec, other methods must be used to scan the camera. This situation occurs in low frame rate (eg, 1 frame per second) DSA and digital photospot imaging. In this case, the detected x-ray exposure can run from 100 mR to 1,000 mR per frame (at the face of the II) to reduce the x-ray noise amplitude in the resulting RG ■ Volume 20 • Number 6 Van Lysel 1785 Figure 20. Operation of a video monitor. image. To achieve this exposure level, x-ray pulse widths on the order of 0.1 second or greater must be used. The x-ray pulse then extends over several video frame periods. Two different methods can be employed to scan this signal. One method is to operate the video camera continuously in the normal 30 frames per second scanning mode. Video frames that span the x-ray pulse period are integrated (summed) by the digital image processor to produce one image. This image is stored in the image processor and displayed repeatedly until the next x-ray pulse provides a new image. The other method is as follows (Fig 19): For a pickup tube–based camera, the scanning electron beam is blanked, just as it is during retrace, during the x-ray pulse. After termination of the x-ray pulse, the beam is turned back on (synchronous with V sync) to read the image. CCD camera operation is similar; the shift operations are suspended during the x-ray pulse. As mentioned earlier, it is becoming increasingly common to encounter systems that offer “frame rate reduction” as a means to decrease patient and staff x-ray exposure. Such systems require that the x-ray source be pulsed. Cardiac angiography (ie, cine) has always been pulsed, but in the past most fluoroscopy was performed with “continuous” x rays. Frame rate reduction also requires that scanning of the video camera be progressive rather than interlaced. However, it is not true, as is often claimed, that pulsed-progressive fluoroscopy, in and of itself, results in exposure reduction. Rather, pulsed-progressive fluoroscopy enables frame rate reduction, which results in exposure reduction. Exposure reduction is not as straightforward as it might at first appear. If the frame rate is dropped from 30 to 15 frames per second, the x-ray exposure drops by a factor of two only if the milliampere-seconds per pulse remains constant. However, the vendor might raise the milliampere-seconds per pulse to “compensate” for the reduction in frame rate, reducing or negating the exposure reduction. If it is assumed that the milliampere-seconds per pulse remains fixed, it is important to understand that, although the exposure does drop by two times, image quality also drops. This effect occurs because, when viewing a rapid succession of images, the eye integrates the noise content of all images presented within a period of approximately 0.2 seconds. If fewer images are presented within this “eye integration period,” the observer perceives an increase in noise. Research shows that if image quality is held constant by increasing the milliampere-seconds per pulse as the frame rate is reduced, frame rate reduction results in only modest exposure reduction (4). However, as a practical matter, frame rate reduction can be a powerful tool for exposure reduction. In the past, changing the fluoroscopic exposure rate was an involved process requiring a service engineer. However, with pulsed-progressive fluoroscopy, the physician can be given tableside control of the fluoroscopic frame rate. If it is assumed that the vendor maintains a constant milliampere-seconds per pulse, the physician now has the ability to decrease or increase the exposure level in response to the demands of the procedure. Image Display The image that is read from the video camera target is displayed on a video monitor (Fig 20). The similarities in operation between the monitor and the camera pickup tube are conspicuous. RG ■ Volume 20 • Number 6 1786 November-December 2000 In a monitor, an electron beam is produced by thermionic emission from a hot cathode. This beam is accelerated toward the face of the monitor by a high-potential anode. When the electron beam strikes the fluorescent screen on the monitor face, light is emitted. Deflection coils cause the electron beam to be scanned across the monitor screen in the required manner (usually in a 2:1 interlaced fashion). The magnitude of the electron beam current is determined by the voltage applied to the control grid, which in turn is determined by the video signal originating from the camera. If the point at which the electron beam is currently striking the monitor face is to be dark, the control grid reduces the electron beam current. If the image is to be bright, more electrons are allowed to pass through the control grid to reach the phosphor. During horizontal and vertical retrace, the control grid blanks the electron beam. In the absence of a digital image processor, the scanning electron beams in the pickup tube and the monitor are precisely synchronized (by V sync and H sync). The presence of an image processor allows decoupling of camera and monitor scanning. Earlier in this article, scan conversion to allow progressive acquisition and interlaced display was discussed. A second common scan conversion operation is upscanning. The most typical upscanning method is to convert a 525-line frame acquired by the camera into a 1,049-line frame for display. The digital processor interpolates the acquired data to double the number of lines. Although this procedure does not increase the spatial resolution of the video system (that would require 1,049-line acquisition), it does provide an improved display (5). Close inspection of a 525-line display shows dark gaps between adjacent lines. Upscanning fills in the gaps between the lines, producing a more continuous and higher-contrast display. An alternative method to achieve the same result, defocusing of the scanning electron beam of the monitor, results in an unacceptable loss of resolution in the horizontal direction. Conclusions Although it is unnecessary for the operators of fluoroscopy equipment to be familiar with all of the technical details pertaining to the imaging system, there are some principles that should be understood. For example, the x-ray exposure level between particular systems can vary substantially, primarily due to choices made by the installer or vendor but also due to maintenance issues. Routine quality assurance testing addresses the latter issue. The former issue is best addressed by active participation by the user in the installation process. Responsible use of x-ray equipment requires the physician to be aware of the x-ray exposure level employed on his or her machine. If this level is higher than average, questions should be asked. To ensure that the answers received are valid, it is important to understand how the optical distributor works. Proper performance of the video system is critically important to good image quality. Ensuring optimum performance starts with the equipment purchase; here again, a sound understanding of system operation will allow intelligent choices. Once the system is in operation, parameters such as edge enhancement, lag, and gamma, which can often be adjusted through the digital system, can have significant effects on the perception of image detail. Finally, modern systems often afford the operator the option of reducing “frame rate” to reduce radiation exposure. It is beneficial to understand how such reduction is accomplished. References 1. Friesinger GC, Adams DF, Bourassa MG, et al. Optimal resources for examination of the heart and lungs: cardiac catheterization and radiographic facilities—Examination of the Chest and Cardiovascular System Study Group. Circulation 1983; 68: 891A–930A. 2. Sandrik JM. The video camera in medical imaging. In: Fullerton GD, Hendee WR, Lasher JC, Properzio WS, Riederer SJ, eds. Electronic imaging in medicine. Medical Physics Monograph no. 11. College Park, Md: American Institute of Physics, 1984; 145–183. 3. Seibert RA, Barr DH, Borger DJ, Hines HH, Bogren HG. Interlaced versus progressive readout of television cameras for digital radiographic acquisitions. Med Phys 1984; 11:703–707. 4. Xue P, Wilson DL. Pulsed fluoroscopy detectability from interspersed adaptive forced-choice measurements. Med Phys 1996; 23:1833–1843. 5. Holmes DR Jr, Wondrow MA, Reeder GS, et al. Optimal display of the coronary arterial tree with an upscan 1,023-line video display system. Cathet Cardiovasc Diagn 1989; 18:175–180.
© Copyright 2026 Paperzz