ICCP ICCP #5 #5 ICCP 2014 Submission #5. CONFIDENTIAL REVIEW COPY. DO NOT DISTRIBUTE. 000 001 002 003 004 005 006 Phase Messaging Method for Time-of-flight Cameras 007 008 009 010 011 Paper ID 5 012 013 014 015 016 017 018 019 020 021 022 023 024 025 026 027 028 029 030 031 032 033 034 035 036 037 038 039 040 041 042 043 044 045 046 047 048 049 050 051 052 053 Anonymous ICCP submission 1. Introduction As cameras become ubiquitious, communication between displays and cameras has become an active research area. Visible-Light Communication (VLC) has been developed where standard cameras detect patterns intentionally placed within images. These patterns can be fixed as in QRcodes [14], dynamic as in Visual MIMO, [4, 27] or LEDto-LED applications [22], and coded using OFDM as in Pixnet [17, 21]. The main theme in these methods, is communication by modulating light intensity from an observed 059 060 061 062 063 064 065 066 067 068 069 Abstract Ubiquitous light emitting devices and low-cost commercial digital cameras facilitate optical wireless communication system such as visual MIMO where handheld cameras communicate with electronic displays. While intensity-based optical communications are more prevalent in camera-display messaging, we present a novel method that uses light phase for messaging and time-of-flight (ToF) cameras for receivers. With intensity-based methods, light signals can be degraded by reflections and ambient illumination. By comparison, communication using ToF cameras is more robust against challenging lighting conditions. Additionally, the concept of phase messaging can be combined with intensity messaging for a significant data rate advantage. In this work, we design and construct a phase messaging array (PMA), which is the first of its kind, to communicate to a ToF depth camera by manipulating the phase of the depth camera’s infrared light signal. The array enables message variation spatially using a plane of infrared light emitting diodes and temporally by varying the induced phase shift. In this manner, the phase messaging array acts as the transmitter by electronically controlling the light signal phase. The ToF camera acts as the receiver by observing and recording a time-varying depth. We show a complete implementation of a 3 × 3 prototype array with custom hardware and demonstrating average bit accuracy as high as 97.8%. The prototype data rate with this approach is 1 Kbps that can be extended to approximately 10 Mbps. 054 055 056 057 058 070 071 072 073 074 075 076 077 078 079 080 081 082 083 084 085 086 087 Figure 1. An example application of “smart signage” that communicates to cameras in the environment with infrared LED’s that send invisible messages with light phase for communication with time-of-flight depth cameras. These cameras are increasingly more commonplace and can be expected in mobile robotics and devices. The LEDs are depicted large for clarity but would be unnoticeable in a real application. image. Recently, high quality depth cameras using timeof-flight (ToF) have become readily available at low-cost, creating an entirely new opportunity for communications. Creating a depth-pattern is a more challenging optical task when compared to intensity patterns. However, since depth is measured with ToF cameras using the phase of reflected light, manipulating the phase of the reflected light signal effectively changes the measured depth, resulting in a phasebased message. The applications for depth cameras are numerous since most intelligent sensing applications can benefit from scene geometry information. We develop a novel method for com- 088 089 090 091 092 093 094 095 096 097 098 099 100 101 102 103 104 105 106 107 ICCP ICCP #5 #5 ICCP 2014 Submission #5. CONFIDENTIAL REVIEW COPY. DO NOT DISTRIBUTE. 108 120 121 122 123 124 125 ToF Camera ToF Camera Distance d Distance d (a) ToF Camera ToF Camera Customized Phase Customized MessagingPhase Array Messaging Array d1 d1 d2 d2 dn dn (b) 126 127 128 129 130 131 Figure 2. Phase Messaging Array (PMA) method. (a) A customized PMA with n LEDs is physically at distance d away from a ToF camera. (b) By modulating phases for LEDs on PMA, LEDs “seem” at individual distances away in depth images from the ToF camera. Messages are represented by these measured distances. 132 133 134 135 136 137 municating with depth cameras using the phase of infrared (IR) light. The transmitter arrays can be used as beacons, tags or communication channels that interact with RGBD cameras for real world applications. One can envision “smart signage” with invisible light messaging to cameras in the environment (see Figure 1); vehicle id tags transferring information to road cameras or cameras on other vehicles; robots that gather task information like access keys stored in fixed beacons; product tags with information (shelf life, expiration date and price) that are read during scene depth-imaging so that the product can also be localized. While visible light communication has become a popular research area, ambient illumination and reflections in real world environments can degrade the signal quality. Messaging with light phase instead of intensity has distinct advantages. First, time-of-flight cameras are designed to measure the phase of the source light signal (typically 15-30 MHz) by correlating the source and the reflected signal. This design means that depth cameras are tuned to their source light signal and are relatively unaffected by environment lighting. Therefore, reflections and ambient lighting do not degrade the quality of the signal. This is a major problem for operations of intensity-based systems. Additionally, the time-of-flight camera measures time delays on the order of nanoseconds, and this short time interval makes the method particularly useful for mobile scenarios where the camera and messenger are in motion (e.g. handheld camera, automotive applications). The time interval of the light path is orders of magnitude smaller than that of the system motion, so that motion blur is not an issue. On the contrary, 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 intensity-based messaging is vulnerable to degradation by motion blur. Furthermore, the use of IR light enables messaging without visibility to human observers. In this work, we design and construct a phase messaging array (PMA), which is the first of its kind, to communicate to a depth camera by manipulating the phase of the depth camera’s own light signal (see Figure 2). The light signal is replicated and the phase is shifted by our custom circuitry. Replicating the camera’s own light signal is a key design component of the system. The time-of-flight camera is generally unresponsive to light signals that are not exactly matched in frequency to their source signal. Sensing and replicating the original camera signal with a message encoded in the phase shift ensures independence of the transmitter (PMA) and receiver (camera) in the communications channel. A spatial distribution of IR light emitting diodes allows messaging units to be packed in the x-y plane for an increased data rate. The array driving circuitry measures the depth camera’s source signal and effectively delays that signal according to the desired message. In this manner, the depth camera observes a time-varying depth that is electronically controllable. We show a prototype implementation of a 3×3 array with 4 phase symbols and messaging accuracies around 97.8%. The data rate of this 3 × 3 array approaches 1 Kbps which can be readily extended to a data rate more than 10 Mbps (see our discussion in Section 4.2). …… 115 116 117 118 119 Customized Phase Customized MessagingPhase Array Messaging Array …… 109 110 111 112 113 114 2. Related Work Visible-Light Communications (VLC) is a general term for methods that use visible light instead of radio frequencies for communication. A theme in VLC in recent years has been to extend existing infrastructure to create new communication channels. Li-Fi [9] uses light-bulbs that not only illuminate the room, but can also transmit temporally dynamic data by modulating light intensity. The data rate of Li-Fi with a single color LED reaches 1.67 Gbps; however, specialized equipment is needed for both the receiver and transmitter. Increasing the field-of-view in visible light commuincations has been addressed in [25] with the use of hemispherical lenses at the receivers. However, the use of ordinary cameras provides the benefit of an inherently wide field-of-view. Visual MIMO [3, 4] is a method that uses any light-emitting spatial array such as an LCD display as the transmitter and uses ordinary cameras as the receivers. This method is a type of Camera-Display Communication (CDC) since the framework enables cameras and electronic displays to communicate. Other CDC methods include QR-codes [1, 14], Bokode [16] and Pixnet [21]. QR-codes are comprised of black-white block patterns, and are quite common in recent years applied to existing ads or signage. Unlike Visual MIMO, the QR-code is passive so the information is static and the pattern is visible to the human observer. Similar to Visual MIMO, QR-codes are designed for use with general purpose cameras, allowing 2 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 ICCP ICCP #5 #5 ICCP 2014 Submission #5. CONFIDENTIAL REVIEW COPY. DO NOT DISTRIBUTE. 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 broad real-world applicaton. Bokode and Pixnet also display message codes with light-emitting devices as transmitters. The difference is that Bokode exploits the bokoh effect of ordinary camera lenses for long-distance static tag reading, while Pixnet generalizes the popular OFDM transmission algorithms [2, 15] to address the unique characteristics of the LCD-camera link. Visual MIMO [27, 28] proposes a photometric modeling for hidden imagery and can realize a dual use the electronic display: to display an image for human observation while simultaneously transmitting hidden bits for decoding by a camera. The heirarchy of this work in optical communications in real-world environments progresses from barcodes [18] that are one-dimensional and static, to QR-codes that are two-dimensional and static, to the various CDC methods that use controllable intensity patterns and are two dimensional and time-varying. Our method adds to this framework with phase messaging that uses controllable phase patterns for 2D time-varying messaging. We continue the theme of using ordinary cameras for messaging since low-cost unmodified ToF cameras are used for the receiver. 270 271 272 273 274 4 s(t) r(t) 3.5 τ 3 2.5 Signal Strength 216 2 275 276 277 278 279 280 1.5 1 0.5 0 x1 x0 x2 x3 −0.5 −1 0 1 2 3 4 5 6 Time(t) Figure 3. Four bucket method. Get four equidistant samples of the received signal r(t) within one cycle. Their cross-correlations with the sent signal s(t) are c(x0 ), c(x1 ), c(x2 ), c(x3 ). They can be used to solve the depth, which is proportional to the delay τ . IR LEDs, since they are readily available at low cost. 1 Denote the modulation frequency as fm . The sent signal s(t) and the received signal r(t) with delay τ are given by Ordinarily, ToF cameras are used to directly estimate 3D scene structure complementing traditional multiple-view reconstruction methods [10,11]. Applications using ToF cameras include 3D scene reconstruction [6, 23], reflectance capture using ultrafast imaging [20], frequency analysis to address multipath interference [5, 13, 26] and human pose detection/tracking [7, 8]. These depth cameras are increasingly common in various application domains such as intelligent vehicles, robotic environment mapping and surveillance. Our novel framework deploys ToF depth cameras as receivers for optical communications. Compared with intensity-based messaging methods, this phase-based messaging is more robust against background illumination as shown in our experimental results. Also high measurement precision in phases (for example, Softkinetic ToF cameras have 32768 phase levels) can create a larger coding space for an increased data rate. s(t) = r(t) = a cos (2πfm t) , A cos (2πfm (t − τ )) + B (1) (2) where A and B are the unknown scale and offset of the reflected signal and these parameters are estimated at the camera. For s(t) and r(t), the cross-correlation is Z T2 r(t)s(t + x) dt c(x) = lim T →∞ − T2 aA cos(2πfm x + 2πfm τ ) + B. (3) 2 There are three unknown variables in Eqn 3: amplitude A, offset B and delay τ . These variables can be solved by “the four-bucket method” [12, 19]. The basic principle is to obtain 4 equidistant samples at locations (x0 , x1 , x2 , x3 ) within a modulation cycle as in Figure 3. The crosscorrelations of these four outputs with the source signal s(t) is denoted by c(x0 ), c(x1 ), c(x2 ), c(x3 ). The three unknown variables can be solved based on these correlations in closed form. The amplitude A can be solved as q 2 2 [c(x3 ) − c(x1 )] + [c(x0 ) − c(x2 )] A= . (4) 2 The offset B is mainly attributed to background illumination and can be solved as = 3. Methods 3.1. Introduction to Time-of-flight Camera We summarize the basic operation of ToF cameras as presented in [12, 19]. While most intensity-based digital camera are passive, ToF cameras rely on integrated light sources. During imaging, photons emitted from this light source travel to objects in the scene. After surface reflection, the returning light is sensed by the ToF camera. The light travels a distance equal to twice the object depth relative to the camera. There are two main types of ToF cameras: (1) pulse-modulation (PM) (2) continuous-wave (CW) For this work, we use a continuous wave ToF camera using B= 1 Creative c(x0 ) + c(x1 ) + c(x2 ) + c(x3 ) . 4 (5) Interactive Gesture Camera Developer Kit from Intel costs $149. DepthSence311 from SoftKinetic costs $299. (December 2013) 3 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 ICCP ICCP #5 #5 ICCP 2014 Submission #5. CONFIDENTIAL REVIEW COPY. DO NOT DISTRIBUTE. 324 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359 360 361 362 363 364 365 366 367 368 369 370 371 372 373 374 375 376 377 ToF Camera Analog circuit@15MHz FIFO A0 ToF Camera LED Multiplexer Am-1 p0 pn-1 …... 331 332 333 334 335 Message Stream Shift Register 325 326 327 328 329 330 Data Format Converter …... Microprocessor LEDs of the ToF Camera Photo Detector sc(t) sac(t) Amplifier Phase Modulation Figure 4. Basic Operating Principle of the Phase Messaging Array (PMA). For clear illustration of the method, only one transmitter LED is shown. Data flow is labeled with red arrows and control flow is in green. As a communications channel, the LED’s on the PMA are the message transmitters and the ToF camera is the receiver. The PMA senses the signal from the ToF LED light source using a photo detector. This copy of the light signal is denoted sc (t). After amplification, sac (t) is used to create multiple phase-delay versions p0 , · · · , pn−1 . Delays range in (0, 2π] and are achievable in our design. These phase delays are symbols sent to multiplexers as output candidates. The control flow is the bit version of the message stream. At the multiplexer, the combination [Am−1 , · · · , A0 ] selects the corresponding phase-delayed candidate signals to sequentially drive the transmitter LED in the next cascade. The delay τ is the time of flight for light sent from camera LED, reflected from object’s surface and received by the ToF camera and can be solved as c(x3 ) − c(x1 ) τ = arctan . (6) c(x0 ) − c(x2 ) camera using a photodiode to convert the light signal into the electrical signal sc (t). This copied light signal is amplified to sac (t) and delayed to multiple phase-delay versions p0 , · · · , pn−1 . Delays range in (0, 2π] and are controllable in our design. In this manner, we create a dictionary of phase symbols and that are sent to the multiplexers as output candidates. The signals sc (t), sac (t) and pi (t) comprise the data flow component of the P M A circuit. The control flow is the bit version of the message stream. At the multiplexer, the combination [Am−1 , · · · , A0 ] selects corresponding phase symbol in a sequential way to drive the transmitter LED in the next cascade2 . There is a simple intuitive method of achieving the phase delay that fails in practice: applying a signal generator or crystal to produce signals at the same modulation frequency fm of the ToF camera. For the “four bucket method”, it is critical to make sure the sent signal s(t) and the received signal r(t) have the same modulation frequency fm . Though labeled with the fixed frequency value fm , a signal generator or crystal always produces outputs with some variance around fm . This frequency error converts into a delayed phase error and makes the measured depth values random. We have shown this experimentally in Figure 5(a) with the following procedure: A signal generator produces a sinusoid waveform at 15MHz. This source drives an IR LED facing to a receiver ToF camera working at the same modulation frequency fm = 15MHz. As expected, the The value of τ is proportional to the distance d between the ToF camera and the captured object as τ = 2d c , where c is light speed. Then the delayed phase between s(t) and r(t) can be denoted as φ = 2πfm τ = 2πfm 2d c . Since φ ranges in (0, 2π], the measured distance d reaches its maximum dmax when φ = 2π. For a ToF camera working at fm = 15MHz, dmax =10m. 3.2. Phase Messaging Array Method Controllable Phase Delay The design of our proposed PMA method is illustrated in Figure 4. For simplicity, the figure shows the basic operating principle with a single LED on the phase messaging array. As a communications channel, the PMA is a message transmitter and the ToF camera is the message receiver. The goal is to create phase delays at the PMA circuit that mimic depth variations in order to message to the ToF depth camera. Delaying light directly would typically require variation of the refractive index of the medium, and controllable delays in this manner are not practical. However, as an electical signal, we can readily introduce delays using an appropriately designed cascade of logic gates. This approach is a key innovation in the design. We sense the signal from the source LED of the ToF 2A m−1 is the Most Significant Bit (MSB), while A0 is the Least Significant Bit (LSB). 4 378 379 380 381 382 383 384 385 386 387 388 389 390 391 392 393 394 395 396 397 398 399 400 401 402 403 404 405 406 407 408 409 410 411 412 413 414 415 416 417 418 419 420 421 422 423 424 425 426 427 428 429 430 431 ICCP ICCP #5 #5 ICCP 2014 Submission #5. CONFIDENTIAL REVIEW COPY. DO NOT DISTRIBUTE. 350 300 300 250 250 200 150 sac(t) 456 457 458 459 460 461 462 463 464 465 466 467 468 469 470 471 472 473 474 475 476 477 478 479 480 481 482 483 484 485 Inv1 Inv2 100 100 50 50 0 0 C2 Inva Invb p0 0 0.5 1 1.5 2 2.5 3 360⁰ C3 p1 180⁰ Invc p2 p3 (a) Phase modulation scheme 0 0.2 0.4 0.6 0.8 1 1.2 1.4 0 ⁰ (b) Four phase images 1.6 Time(second) Time(second) (a) Source: signal generator Inv3 C1 Inv4 150 350 350 (b) Source: copied signal 300 Phase histogram 300 444 445 446 447 448 449 450 451 452 453 454 455 Inv0 200 Phases(deg) 439 440 441 442 443 350 Phases(deg) 433 434 435 436 437 438 Phases(deg) 432 250 200 150 100 0 200 400 600 800 1000 Time(s) (c) Plot of phases (c) Signal amplification 200 150 100 50 50 0 250 1200 1400 0 0 50 100 150 200 250 300 350 Phases(deg) (d) Histograms of phases Figure 6. Phase modulation. (a) In the phase modulation scheme, “deterioration prevention” and “phase delay control” are two basic considerations. (b) Four phase images of an LED displayed with delayed signals p0 , p1 , p2 and p3 . Phase values are mapped to colors for visualization. (c) Plot of measured phases for an LED displayed with four delayed signals in turn. (d) Histogram of measured phases in (c). Red dotted lines are hard thresholds for phase symbol assignment. Figure 5. Comparison of two sources. (a) Measured phases of an IR LED driven by a 15MHz output from a signal generator are random, because of the frequency noise of the signal generator. (b) Measured phases of an IR LED by our proposed PMA method are stable. (c) Amplification of sc (t) in blue to sac (t) in red. measured phases of the LED from the depth image are random. Our solution is to sense and copy the source signal from LED of the receiver ToF camera. An example is illustrated in Figure 5(c): the copied signal sc (t) in blue has an amplitude of 400mV with unit voltage 200mV; the amplified signal sac (t) in red has an amplitude of 5V with unit voltage 2V. This amplified signal sac (t) drives an LED facing the same ToF camera. Compared with Figure 5(a), the measured phases of the LED in Figure 5(b), are stable enough for messaging. from the receiver ToF camera. Though the location of the LED is fixed, ToF camera measures it at four possible distances, corresponding to four phases p0 , p1 , p2 and p3 . For visualization, these four candidate phases are mapped to individual colors. These four delayed signals are displayed sequentially on an IR LED controlled by a timer as in Figure 6(c). The histograms of all measured phases are shown in Figure 6(d). The histograms show the presence of outliers from the phase symbols due to transition states. We currently remove outliers by using an indicator LED to temporally identify transition times. Omitting transition times, the measured phase has a standard deviation σp = 1.2 from an individual LED. The phases can be separated by hard thresholds indicated by the red dotted lines. Phase Modulation For messaging, the mimicing signal of an LED from the receiver ToF camera is amplified into sac (t). In our proposed scheme in Figure 6(a), there are four phase-delayed candidates: p0 , p1 , p2 and p3 . Delays range in (0, 2π] and are controlled by capacitors C1 , C2 and C3 . Each phase candidate is connected to the same channel of all multiplexers and each multiplexer drives one transmitter LED. Since each multiplexer works as a “switch”, the configuration is a heavy load to output inverters Ia , Ib , Ic and Id , and may degrade the output signals severely. To prevent signal deterioration between cascades, the output signal pi for each cascade and the one sent to the next cascade are not the same. Additionally, inverter I0 is used to increase driving ability of the input signal. Figure 6(b) illustrates four phase images of an IR LED Messaging and Signal Selection Each of four delayed phases p0 , p1 , p2 and p3 (in red) are connected to the same channel of all multiplexers. Every multiplexer drives one transmitter LED facing towards the receiver ToF camera. The bit streams from the control flow (in green) determines which candidate to select. One multiplexer case is shown in Figure 7(a). By the truth table in Figure 7(b) [24], signal A1 A0 = 102 selects the signal with phase p2 . In the decoding scheme, messages are extracted by converting the measured phase value into the corresponding binary code using the truth table in Figure 7(b). 5 486 487 488 489 490 491 492 493 494 495 496 497 498 499 500 501 502 503 504 505 506 507 508 509 510 511 512 513 514 515 516 517 518 519 520 521 522 523 524 525 526 527 528 529 530 531 532 533 534 535 536 537 538 539 ICCP ICCP #5 #5 ICCP 2014 Submission #5. CONFIDENTIAL REVIEW COPY. DO NOT DISTRIBUTE. 540 541 542 543 544 545 546 547 548 549 550 551 552 553 554 555 556 557 558 559 560 561 562 563 564 565 566 567 568 569 output A0 Multiplexer A1 p0 p1 p2 p3 A1 A0 output 0 0 1 1 0 1 0 1 p0 p1 p2 p3 ure 7(b) and labels p2 p1 p0 p3 p2 p1 p0 p3 p1 are decoded to the binary form “100100111001001101”. 360⁰ (b) Truthtable 0 ⁰ Figure 7. Signal selection for messaging. (a) A 4×1 multiplexer with four channels (in red) and two bits A1 A0 (in green) for selection. (b) The truthtable of the multiplexer in (a). (a) Snapshot of 9 LEDs 4. Experiments In our proposed scheme, the receiver is a CW ToF camera SoftKinect DS311. Light from LEDs of this camera is IR with wavelength of 850nm and modulation frequency fm =15MHz. The depth sensor has a resolution of 160×120. The PMA transmitter is illustrated in Figure 8. There are five basic components: (1) photo detector and signal amplifier; (2) phase modulator; (3) multiplexers; (4) an array of 9 LEDs; (5) shift register to convert the serial control flow into the parallel one for all multiplexers. The control flow is converted into a binary stream and buffered at a microprocessor MSP430G in this experiment. For the PMA transmitter circuit, it is critical to select component chips with delay far less than one period, i.e. 66ns. 593 323.5 172.4 p2 p3 p0 232.9 276.8 327.1 p1 p2 p3 182.3 236.7 242.3 p0 p1 p1 (c) Phase label assignment Figure 9. Phase symbol assignment for one phase image. (a) A phase image with 9 LEDs. For visualization, phase values are mapped to colors. (b) Mean phase values of 9 LEDs in degree. (c) Phase symbol assignment with p0 -p3 by the threshold method. Of the nine LEDs in the prototype, one LED is used as an indicator to signal the frame transition. The other eight LEDs provide 16 message bits per captured frame. We send 8000 random bits using the PMA and extract messages at the ToF camera. The average accuracy of five independent experiments is 97.8% with a standard deviation of 1.3%. The camera frame rate for experiments is 24 fps (framesper-second). In practice, the transmission rate at the PMA is modified to allow for repetitive frames to assist the synchronization and the measured number of repetitive frames has a mean of 6.3, resulting in an effective frame rate of 4 fps. With 16 bits per frame (2 bits per LED), the effective message rate is approximately 64 bps. By addressing synchronization without using repetitive frames, the data rate for this method can reach 960 bps by setting 60fps, i.e. the maximum frame rate of SoftKinect DS311. An accuracy of 97.8% corresponds to an uncoded bit error rate of 2.2%. Appropriate error correction codes such as Reed-Solomon for burst error and convolutional codes for random errors can be used to improve performance of accuracy, at the tradeoff of bit rate. The appropriate channel characterization and coding is reserved for future work. 575 576 577 578 579 580 587 588 589 590 591 592 272.6 (b) Phase values in degree 570 571 572 573 574 581 582 583 584 585 586 599 600 601 602 603 604 180⁰ (a) Multiplexer Figure 8. Phase messaging array (PMA) circuit for experiments. We test the system in an indoor environment by sending random bits from PMA and measure the accuracy (or recognition rate) at the receiver side (i.e. ToF camera). Here, recognition rate is given by the number of correctly recognized bits divided by the number of transmitted bits. One snapshot of the captured phase image is illustrated in Figure 9(a). For message extraction, we get average phase values for each LED as in Figure 9(b) and assign phase labels by the threshold method. Refer to the truthtable in Fig- 4.1. Robustness We compare the influence of ambient background illumination for 1) phase-based messaging with a ToF camera (SoftKinetic DS311) as the receiver and 2) intensitybased messaging using typical cameras (smart phone Samsung SIII and consumer digital camera Sony DSC-RX100) as receivers. The transmitter for phase-messaging is our 6 594 595 596 597 598 605 606 607 608 609 610 611 612 613 614 615 616 617 618 619 620 621 622 623 624 625 626 627 628 629 630 631 632 633 634 635 636 637 638 639 640 641 642 643 644 645 646 647 ICCP #5 ICCP #5 ICCP 2014 Submission #5. CONFIDENTIAL REVIEW COPY. DO NOT DISTRIBUTE. 648 649 650 651 652 653 654 655 656 657 658 659 660 661 662 663 664 665 666 667 668 669 670 671 672 673 674 675 676 677 678 679 680 681 682 683 684 685 686 687 688 689 690 691 692 693 694 695 696 697 698 699 700 701 array, and a 60 frames-per-second camera, the expected data rate would be 600000 × 2 × 8 = 9, 600, 000 ≈ 10 Mbps . PMA displaying four phases sequentially; while the transmitter for the intensity-based messaging is a visible light LED switching between on and off and controlled by a simple electronic timer. Plots of captured intensity and phases are shown in Figure 10. For the normal indoor condition, both ToF camera and intensity-based cameras have switching states as expected. We then apply a bright ambient illumination (from an ordinary fluorescent office light) that is manually switched on and off. With this strong timevarying background illumination, the captured values from intensity-based camera is significantly degraded. We repeat sending 8000 random bits under varying background illumination for five times. The average bit accuracy is 95.8% with a standard deviation of 2.1%, for the phase-based messaging. The intensity-based messaging gives an accuracy of 100% with no background illumination, but the accuracy dips to approximately 70% with random blinking background illumination. As expected, the phase-based messaging method is much more robust against ambient or background illumination than intensity-based messaging. 5. Conclusion In this work, we propose and develop a communication methodology with a low-cost continuous-wave ToF camera and low-cost custom circuitry. The novel method, called Phase Messaging Array (PMA), is the first of its kind that allows messaging to the ToF camera using electronically controllable phase. The message bits can be encoded temporally via phase delays and spatially in the x-y coordinate frame. We demonstrate high accuracy messaging with the PMA operating independently from the ToF camera, suggesting applications where PMA’s in scenes can act as messaging tags, beacons or landmarks for auxilliary communication to mobile systems employing depth cameras. A key demonstrated advantage of phase-messaging is the performance under various lighting conditions allowing operation in outdoor and bright light environments. Messaging accuracies for all cases are above 95%. 4.2. Data Rate Discussion References The system we present in this paper is a complete implementation of the proof of concept 3 × 3 phase-messagingarray with nine transmitter LEDs. In the current implementation, each LED can emit a time varying signal taking on values from a set of four phase delays, i.e. a 2 bit binary message stream. The basic concept can be easily extended to increase the number of LEDs packed in the spatial plane and to increase the number of phase symbols. Of the nine LEDs in the prototype, eight messaging LEDs provide 16 message bits per captured frame. The ToF camera works at the framerate of 24 fps so that the prototype can transmit up to 384 bps with synchronization, and 960 bps at 60 fps. An extension of the current approach can be made by increasing the number of phase symbols to span the 360-degree range. With the current level of phase noise (σp = 1.2) approximately (360/2.4) = 150 symbols or approximately 7 bits can be implemented and still be separated enough for discrimination. By also increasing the spatial density of the LEDs to a 100 × 100 array, and operating at 60 fps, the extended data rate would be 4.2 Mbps. Further modification of the hardware of the PMA to reduce the phase variance so that 256 symbols can be used would enable full utilization of the 8-bit depth camera range and give a data rate of 4.8 Mbps. A key interesting point is that this communications channel is achievable while the depth camera continues its normal operation as a depth camera. No additional hardware on the ToF camera is necessary for its role as the message receiver. As future work, phase-based messaging can be combined with varying the amplitude of the light signal to achieve a combination of both intensity-based and phase-based messaging (in environments with low ambient light). For example, with 8-bits of intensity variation, a 100 × 100 pixel [1] ISO/IEC 18004: 2006 Information technology Automatic identification and data capture techniques QR code 2005 bar code symbology specification, 2006. [2] J. Armstrong. Ofdm for optical communications. J. Lightwave Technol., 27(3):189–204, Feb 2009. [3] A. Ashok, M. Gruteser, N. Mandayam, T. Kwon, W. Yuan, M. Varga, and K. Dana. Rate adaptation in visual mimo. Proceedings of IEEE Conference on Sensor, Mesh and Ad Hoc Communications and Networks, pages 583–591, 2011. [4] A. Ashok, M. Gruteser, N. Mandayam, J. Silva, M. Varga, and K. Dana. Challenge: mobile optical networks through visual mimo. Proceedings of the sixteenth annual international conference on Mobile computing and networking (MobiCom), pages 105–112, 2010. [5] A. Bhandari, A. Kadambi, R. Whyte, L. Streeter, C. Barsi, A. Dorrington, and R. Raskar. Multifrequency time of flight in the context of transient renderings. In ACM SIGGRAPH, pages 46:1–46:1, 2013. [6] Y. Cui, S. Schuon, D. Chan, S. Thrun, and C. Theobalt. 3D shape scanning with a time-of-flight camera. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 1173–1180, 2010. [7] M. V. den Bergh and L. V. Gool. Combining RGB and ToF cameras for real-time 3D hand gesture interaction. In IEEE Workshop on Applications of Computer Vision (WACV), pages 66–72, 2011. [8] V. Ganapathi, C. Plagemann, D. Koller, and S. Thrun. Real time motion capture using a single time-of-flight camera. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 755–762, 2010. [9] H. Haas. High-speed wireless networking using visible light. SPIE Optoelectronics and Communications, 2013. [10] M. Hansard, S. Lee, O. Choi, and etc. Time of Flight Cameras: Principles, Methods, and Applications. Springer, 2013. 7 702 703 704 705 706 707 708 709 710 711 712 713 714 715 716 717 718 719 720 721 722 723 724 725 726 727 728 729 730 731 732 733 734 735 736 737 738 739 740 741 742 743 744 745 746 747 748 749 750 751 752 753 754 755 ICCP ICCP #5 #5 ICCP 2014 Submission #5. CONFIDENTIAL REVIEW COPY. DO NOT DISTRIBUTE. 756 774 775 776 777 778 779 780 781 782 783 784 785 786 787 788 789 790 791 792 793 794 795 796 797 798 799 800 801 802 803 804 805 806 807 808 809 Samsung SIII 350 810 811 812 813 814 RX100 200 80 180 70 160 250 60 Intenisity 140 Intenisity Phases(deg) 300 120 100 80 200 50 40 60 30 40 150 0 50 100 150 200 250 300 20 350 0 50 100 150 Time(s) 200 250 300 350 400 20 450 0 100 200 Time(s) 350 300 400 500 600 Time(s) 200 80 180 250 60 Intenisity Intenisity 140 120 100 80 200 50 40 60 30 40 150 0 50 100 150 200 250 300 20 0 50 100 150 Time(s) Phase 200 250 300 Time(s) Intensity 350 400 450 20 0 100 200 300 400 500 600 Time(s) Intensity Figure 10. Influence of ambient background illumination. We compare the influence of time-varying ambient light for phase-based messaging with a ToF camera (SoftKinetic DS311) and intensity-based messaging (using a smart phone camera Samsung SIII and a consumer digital camera Sony DSC-RX100). We can observe that phase-based messaging is more robust against time-varying ambient illumination than intensity-based cameras. [11] R. I. Hartley and A. Zisserman. Multiple View Geometry in Computer Vision. Cambridge University, 2004. [12] R. P. Horaud. Tutorial on 3D Sensors, Apr 2013. [13] A. Kadambi, R. Whyte, A. Bhandari, L. Streeter, C. Barsi, A. Dorrington, and R. Raskar. Coded time of flight cameras: Sparse deconvolution to address multipath interference and recover time profiles. ACM Trans. Graph., 32(6):167:1– 167:10, Nov. 2013. [14] K. Kamijo, N. Kamijo, and G. Zhang. Invisible barcode with optimized error correction. In IEEE International Conference on Image Processing, pages 2036 –2039, oct. 2008. [15] X. Li, J. Vucic, V. Jungnickel, and J. Armstrong. On the capacity of intensity-modulated direct-detection systems and the information rate of aco-ofdm for indoor optical wireless applications. IEEE Transactions on Communications, 60(3):pp. 799–809, 2012. [16] A. Mohan, G. Woo, S. Hiura, Q. Smithwick, and R. Raskar. Bokode: Imperceptible visual tags for camera based interaction from a distance. ACM Trans. Graph., 28(3), July 2009. [17] M. R. H. Mondal and J. Armstrong. Impact of linear misalignment on a spatial OFDM based pixelated system. In Asia-Pacific Conference on Communications, 2012. [18] A. Q. Morton. Packaging history: The emergence of the Uniform Product Code (UPC) in the United States. History and Technology, 11(1):101111, 1994. [19] C. D. Mutto, P. Zanuttigh, and G. M. Cortelazzo. Time-ofFlight Cameras and Microsoft Kinect(TM). Springer Publishing Company, Incorporated, 2012. [20] N. Naik, S. Zhao, A. Velten, R. Raskar, and K. Bala. Single view reflectance capture using multiplexed scattering and [21] [22] [23] [24] [25] [26] [27] [28] 8 815 816 817 818 819 820 821 822 823 824 825 70 160 300 Phases(deg) 768 769 770 771 772 773 No Blinking Amb. Light 763 764 765 766 767 ToF Camera With Blinking Amb. Light 757 758 759 760 761 762 time-of-flight imaging. In Proceedings of SIGGRAPH Asia Conference, pages 171:1–171:10, 2011. S. D. Perli, N. Ahmed, and D. Katabi. Pixnet:designing interference-free wireless links using lcd-camera pairs. ACM Int. Conf. on Mobile Computing and Networking, 2010. S. Schmid, G. Corbellini, S. Mangold, and T. R. Gross. Ledto-led visible light communication networks. In Proceedings of the fourteenth ACM international symposium on Mobile ad hoc networking and computing (MobiHoc), 2013. A. Velten, T. Willwacher, O. Gupta, A.Veeraraghavan, M.G.Bawendi, and R. Raskar. Recovering three-dimensional shape around a corner using ultrafast time-of-flight imaging. Nature Communication, 3:745+, 2012. J. Wakerly. Digital Design: Principles and Practices (4th Edition). Prentice-Hall, Inc., 2005. T. Q. Wang, Y. A. Sekercioglu, and J. Armstrong. Analysis of an optical wireless receiver using a hemispherical lens with application in mimo visible light communications. J. Lightwave Technol., 31(11):1744–1754, Jun 2013. D. Wu, M. O’Toole, A. Velten, A. Agrawal, and R. Raskar. Decomposing global light transport using time of flight imaging. In IEEE CVPR, pages 366–373, 2012. W. Yuan, K. Dana, A. Ashok, M. Varga, M. Gruteser, and N. Mandayam. Photographic steganography for visual mimo: A computer vision approach. IEEE Workshop on the Applications of Computer Vision, pages 345–352, 2012. W. Yuan, K. Dana, A. Ashok, M. Varga, M. Gruteser, and N. Mandayam. Spatially varying radiometric calibration for camera-display messaging. IEEE Global Conference on Signal and Information Processing (GlobalSIP), 2013. 826 827 828 829 830 831 832 833 834 835 836 837 838 839 840 841 842 843 844 845 846 847 848 849 850 851 852 853 854 855 856 857 858 859 860 861 862 863
© Copyright 2026 Paperzz