10
Date of publication xxxx 00, 0000, date of current version xxxx 00, 0000. Digital Object Identifier 10.1109/ACCESS.2017.DOI Touch Sensing for a Projected Screen Using Slope Disparity Gating MAYUKA TSUJI 1 , HIROYUKI KUBO 2 , (MEMBER, IEEE), SUREN JAYASURIYA 3 , (MEMBER, IEEE), TAKUYA FUNATOMI 1 , (MEMBER, IEEE), AND YASUHIRO MUKAIGAWA 1 (MEMBER, IEEE) 1 Nara Institute of Science and Technology (NAIST), 8916-5 Takayama, Ikoma City, Nara 630-0192, JAPAN 2 Tokai University, 2-3-23 Takanawa, Minato City, Tokyo 108-8619, JAPAN 3 Arizona State University, Tempe, Arizona State 85287, USA Corresponding author: Mayuka TSUJI (e-mail: [email protected]). This work was supported by JSPS Kakenhi(JP19H04138), JST CREST (JPMJCR1764), JST FOREST(JPMJFR206I), JST SICORP (JPMJSC2003) and NSF grant (IIS-1909192). ABSTRACT This paper introduces a novel touch sensing system for a projected screen with a single camera. To achieve touch sensing with camera image, there are two challenges to overcome. One is to establish robustness against image processing being misled by projected contents. The other is to realize both touch detection and fingertips localization with a single-camera image. In this work, we use Slope Disparity Gating and capture only the region slightly above the projected screen. This selected region is set geometrically so that a part of the finger is captured when touching, and no finger is captured when not touching. A key feature of this study is that the projector can enable this selective imaging capability in the camera while still displaying a normal projected image to the user’s eyes. The projector is both a programmable light source for capturing system and an interface display device for the users. We built a prototype system and touch sensing algorithm, evaluated the touch accuracy, demonstrated robustness to a projected content. INDEX TERMS touch sensing, human-computer interface, image sensing system, projector, projector- camera system I. INTRODUCTION T OUCH operation is one of the most intuitive and com- mon forms of human-computer interaction exemplified by the ubiquitous smartphone display. However, large touch displays are expensive and occupy a large physical space. To overcome these limitations, a visible light projector can convert various flat surfaces into virtual displays, decoupling the virtual display size from the surface area. This projector can display video onto ordinary flat surfaces such as a desk or wall, transforming them into a virtual touch screen. Since flat surfaces that can be used as a projected screen are generally not equipped with a touch-sensing mechanism like a smartphone display, fingertip sensing is needed. One approach is to mount a device on the fingertip for acquiring and transmitting visual information through haptic channels. If one wants to avoid the user from wearing the device on their hands or fingers, another approach is to use image processing techniques to track fingertip position from an observing camera. Previous work extracted fingertip position from the scene image via segmentation regions for objects [1]–[5]. However, there are two main challenges to realize touch sensing using the approach. The first challenge is that the projected image can interfere with the finger localization in the camera image due to its textures and displayed scene clutter. As the finger touches the projected screen, the finger’s original color and contour can change as the displayed image overlaps the finger. The second challenge is simultaneous touch detection and fingertip localization from a single camera image. Typical cameras project a 3D scene onto a 2D plane. Thus even if the fingertip position can be detected, it is difficult to detect whether this fingertip touches the surface or is hovering above it. In this paper, we introduce a novel touch sensing system for the virtual display using a synchronized projector-camera. VOLUME 4, 2016 1

Touch Sensing for a Projected Screen Using Slope Disparity

  • Upload
    others

  • View
    2

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Touch Sensing for a Projected Screen Using Slope Disparity

Date of publication xxxx 00, 0000, date of current version xxxx 00, 0000.

Digital Object Identifier 10.1109/ACCESS.2017.DOI

Touch Sensing for a Projected ScreenUsing Slope Disparity GatingMAYUKA TSUJI1,HIROYUKI KUBO2, (MEMBER, IEEE),SUREN JAYASURIYA3, (MEMBER, IEEE),TAKUYA FUNATOMI1, (MEMBER, IEEE),AND YASUHIRO MUKAIGAWA 1(MEMBER, IEEE)1Nara Institute of Science and Technology (NAIST), 8916-5 Takayama, Ikoma City, Nara 630-0192, JAPAN2Tokai University, 2-3-23 Takanawa, Minato City, Tokyo 108-8619, JAPAN3Arizona State University, Tempe, Arizona State 85287, USA

Corresponding author: Mayuka TSUJI (e-mail: [email protected]).

This work was supported by JSPS Kakenhi(JP19H04138), JST CREST (JPMJCR1764), JST FOREST(JPMJFR206I), JST SICORP(JPMJSC2003) and NSF grant (IIS-1909192).

ABSTRACT This paper introduces a novel touch sensing system for a projected screen with a singlecamera. To achieve touch sensing with camera image, there are two challenges to overcome. One is toestablish robustness against image processing being misled by projected contents. The other is to realizeboth touch detection and fingertips localization with a single-camera image. In this work, we use SlopeDisparity Gating and capture only the region slightly above the projected screen. This selected region isset geometrically so that a part of the finger is captured when touching, and no finger is captured whennot touching. A key feature of this study is that the projector can enable this selective imaging capabilityin the camera while still displaying a normal projected image to the user’s eyes. The projector is both aprogrammable light source for capturing system and an interface display device for the users. We built aprototype system and touch sensing algorithm, evaluated the touch accuracy, demonstrated robustness to aprojected content.

INDEX TERMS touch sensing, human-computer interface, image sensing system, projector, projector-camera system

I. INTRODUCTION

TOUCH operation is one of the most intuitive and com-mon forms of human-computer interaction exemplified

by the ubiquitous smartphone display. However, large touchdisplays are expensive and occupy a large physical space.To overcome these limitations, a visible light projector canconvert various flat surfaces into virtual displays, decouplingthe virtual display size from the surface area. This projectorcan display video onto ordinary flat surfaces such as a deskor wall, transforming them into a virtual touch screen.

Since flat surfaces that can be used as a projected screenare generally not equipped with a touch-sensing mechanismlike a smartphone display, fingertip sensing is needed. Oneapproach is to mount a device on the fingertip for acquiringand transmitting visual information through haptic channels.If one wants to avoid the user from wearing the device ontheir hands or fingers, another approach is to use image

processing techniques to track fingertip position from anobserving camera. Previous work extracted fingertip positionfrom the scene image via segmentation regions for objects[1]–[5]. However, there are two main challenges to realizetouch sensing using the approach.

The first challenge is that the projected image can interferewith the finger localization in the camera image due to itstextures and displayed scene clutter. As the finger touchesthe projected screen, the finger’s original color and contourcan change as the displayed image overlaps the finger.

The second challenge is simultaneous touch detection andfingertip localization from a single camera image. Typicalcameras project a 3D scene onto a 2D plane. Thus even ifthe fingertip position can be detected, it is difficult to detectwhether this fingertip touches the surface or is hoveringabove it.

In this paper, we introduce a novel touch sensing systemfor the virtual display using a synchronized projector-camera.

VOLUME 4, 2016 1

Page 2: Touch Sensing for a Projected Screen Using Slope Disparity

Author et al.: Preparation of Papers for IEEE TRANSACTIONS and JOURNALS

A key feature of our system is that the camera selectivelycaptures only the area near the projected screen as opposedto the entire scene within the lens field of view. To captureonly the selected area, we adopt a disparity gating method,which synchronizes the projector with the camera [6], [7].O’Toole et al. [6] first realized this method, which capturesphotos only in a specific depth range by combining a laserscanning projector and a rolling shutter camera. Ueda etal. [7] generalized this approach to capture photos along asloped planar volume in the scene, controllable in its depth,tilt, and thickness. In our work, we apply this projector-camera system to capture only the selected region to detectwhen the finger is touching the projected screen and whenit is absent and merely hovering over the projected screen.Using our region-selected images, we develop a dedicatedalgorithm to realize fingertip localization via simple butefficient image processing.

We make several contributions in this paper. Our systemgeometrically determines whether a finger has touched thescreen surface from a single camera image. Since our methodcaptures only the region slightly above the screen, it is robustto the projected image and does not suffer from visual inter-ference. Our touch sensing system consists of one projector,one camera, and one microcontroller, and does not requireany additional cameras, depth sensors, or light sources. Theprojector operates in a dual role: projecting images onto thesurface as well as the light source for imaging only slightlyabove the projected screen synchronously with the camera.Our method has an advantage in terms of privacy preser-vation compared with other camera-based methods becauseour device only captures light from the space just above aprojected surface. In other words, privacy-sensitive parts suchas human faces or other identifying details are not capturedby the system outside of the region-of-interest.

Finally, we demonstrate that a simple image processingalgorithm is enough to realize touch sensing as opposed toconventional methods that segment fingertips from complexscenes. This is achieved by geometrically constrained capturethat eliminates most of the imaged scene clutter unnecessaryfor the touch sensing beforehand.

II. RELATED WORKTouch operation has been actively studied as a means ofinteraction due to its intuitive nature. We discuss varioustechnologies to enable touch sensing and compare them toour proposed solution.

A. DIRECT SENSING

The most common method of touch sensing is to detectforce on the touch surface itself. Capacitive touchscreenshave been widely used in our daily life, namely in our smartphones. They also have been actively studied in the work ofLee et al. [8], MTMR [9], Smartskin [10]. In addition tocapacitance, vibrational sensing has been used effectively todetect touch. Skinput [11] uses an armband sensor to detect

vibrations transmitted from the body. This allows the user tomanipulate the image projected onto the skin by touch.

There are other methods to add sensing functions to thetouch surface, such as using a grid of wired [12] or acousticsensors [13]. These devices have static screen sizes and arenot suitable for dynamic screen resizing. In addition, largedisplays with these sensors are still expensive and physicallydifficult to install.

In contrast, our method requires no specialized surfaceelectronics and can scale to large display sizes effortlessly.

B. FINGER INPUT DEVICESAnother technique is to mount special sensors on the objector finger used for input. This is effective when it is difficultto mount sensors onto the touch surface itself. Escritoire [14]uses an ultrasonic pen that enables touch control of theprojector image. Sensetable [15] uses a physical mice-likedevice in which an electromagnetic sensor is embedded andrealizes interaction with the projector image.

In recent years, however, much research has focused onenabling intuitive hand operation rather than an input deviceto the screen. Unless fine strokes are required such as indrawing, bare-handed operation provides more flexibility andnatural interaction for the user. 3DTouch [16] realize fingerwearable devices using an optical laser sensor and a 9-DOF inertial measurement unit. Shi et al. [17] attached anaccelerometer to the tip of a fingernail and used deep learningto recognize the vibration patterns for touch operation.

In a similar fashion to these works, our system also enablesintuitive hand operation with our virtual display.

C. PASSIVE CAMERA SENSINGAnother thrust of research utilizes passive sensing with cam-era images. This allows cheap systems to be implementedin practice. Letessier et al. [18] detected fingertips on thedisplay using a single camera image, although it did notdetermine whether a finger touched a surface. A single cam-era cannot provide depth information, so some studies haveused multiple cameras to perform stereo matching [1], [19].Alternatively, Marshall et al. using a discriminator of thecolor-changes of the fingernail when touching [20] to sensedepth from a single camera image. However, many of thesestudies require the ability to robustly distinguish betweenfingers and the displayed projector image, and thus can beconfused by visual clutter in the virtual scene.

D. ACTIVE SENSINGA final avenue of research on touch sensing uses activesensing with active light. Many active sensing use an ad-ditional infrared light source to perform fingertip’s depthsensing without interference with projected content. Suchdepth information is remarkably effective for touch sensingand has been shown by Wilson et al. [21], DIRECT [2],Omnitouch [3], and Dante vision [4]. The use of LiDAR asan infrared light source has been investigated such as La-seiWall [22], SurfaceSight [23] and Short-Throw Interactive

2 VOLUME 4, 2016

Page 3: Touch Sensing for a Projected Screen Using Slope Disparity

Author et al.: Preparation of Papers for IEEE TRANSACTIONS and JOURNALS

Display Engine [24]. As an exceptional study that does notuse an infrared light source, Touchscreen Everywhere [5]used special structured illumination from a projector. Ourmethod uses visible light from a projected image to achieveactive sensing.

We summarize the comparison between existing methodsand our method in Table 1. We can realize touch sensing witha single camera without the need for additional infrared de-vices, and our method has the advantage of being inexpensiveto build. There is minimal processing of the projected imageas the region of interest is selectively captured optically, andthis makes our system have a low computational cost. Also,since we do not use a machine learning approach, we donot require any training data. Finally, there is no need toattach any device to the finger, nor require any special displayelectronics for the surface itself. All these features make oursystem a compelling approach for realizing touch sensing forprojected screens.

III. METHODFigure 1 shows our user interaction situation of interest. Theprojector presents the displayed video onto an everyday-surface (for example a table surface in Figure 1). Users per-form touch operations on the images with their fingers. Oursystem performs touch sensing by identifying the fingertipposition in an image of the scene captured with the camera.

FIGURE 1: User situation involving touch sensing and oursystem, where a user’s finger interacts with the projectedcontents on a flat surface. In this case, we used a projectoronto a 150cm × 75cm table. We set a camera directly belowthe projector for touch sensing. The distance from our deviceto the center of the projected surface is about 50cm.

In this paper, to realize touch sensing, we have developeda method to selectively capture only the region slightlyabove the projected screen. To achieve selected capture ofthe desired region, we adopt the method of Ueda et al.synchronizing a projector with a camera [7]. In this section,we first discuss the significance of selective capture, which isthe basis of our approach, then review Ueda et al.’s projector-camera system utilized in this research, and finally describethe specific algorithm for touch sensing.

A. CAPTURING ONLY SELECTIVE REGIONS FORTOUCHED FINGERTIPIn situations where touch sensing is performed on a projectedscreen as shown in Figure 1, there are two main challenges toovercome. The first challenge is that the visual informationof the projected image interferes with hand detection. In par-ticular, projected colors overlap on a fingertip’s skin whichmakes extracting the fingertip area from the image difficult.Further, if the projected image content includes real handsas can occur if there are people in the displayed content,the system cannot distinguish between the fake hand in theprojected image and the real hand that is touching the screen.One way to overcome this challenge is to use additionaldevices such as infrared cameras or depth sensors to detectthe real hand region as a non-planar object. However, suchadditional hardware increases the complexity of the deviceconfiguration. The second challenge is performing both touchdetection and fingertip localization based on images froma single fixed camera. While multiple cameras can triangu-late the 3D coordinates of a finger, this also increases thescale and computational complexity of the interaction device.However, the use of multiple cameras increases the scale ofthe interaction device and the computational complexity.

By capturing only slightly above the projected screen andeliminating superfluous information in the scene, we canovercome the above two difficulties.

Our key insight is that interaction with a physical surfacerequires touch detection and fingertip localization at theactual plane of the physical surface itself. However, this isdifficult to do without a depth camera or physical sensors as aconventional camera captures all visible scene points withinthe field of view. Scenes of touch sensing situation containa lot of wasteful information, not intrinsically necessary fortouch sensing. In Figure 2, for example, the following partsare unnecessary for sensing: the projection contents, the handarea other than the fingertips, the surrounding scenery, thehovering fingertip, and the fingertip touching outside thescreen range. To selectively capture only this plane and alloptical interactions with it, we utilize the recently developedmethod of Slope Disparity Gating [7].

B. SLOPE DISPARITY GATINGSlope Disparity Gating is a measurement technology that cancontrol the depth, tilt, and thickness of a selected region tobe optically captured in the scene by synchronizing a singleprojector and a single camera. Before reviewing this method,we first introduce the projector-camera system of O’Toole etal. [6], which was the precursor to Slope Disparity Gating.

O’Toole et al. [6] realized a method to capture onlyreflected light from the projector by synchronizing a singlelaser scanning projector and a single rolling shutter camera.In the laser scanning projector, the image is not all projectedat once but is sequentially projected on a horizontal line.Similarly, the rolling shutter of the camera exposes not theentire sensor but rather one row at a time to capture photos.By synchronizing the laser projector’s row scanning with the

VOLUME 4, 2016 3

Page 4: Touch Sensing for a Projected Screen Using Slope Disparity

Author et al.: Preparation of Papers for IEEE TRANSACTIONS and JOURNALS

TABLE 1: Table comparing several existing methods as well as our method with respect to various desirable properties for atouch sensing system.

Method Additional infraredlight source Camera device Special

projection Learning data Uniqueinput device

Uniqueinput plane

Capacitive touchscreen [8] [9] not needed not used not needed not needed not needed requiredSkinput [11] not needed not used not needed not needed not needed requiredDiamondTouch [12] not needed not used not needed not needed not needed requiredEscritoire [14] not needed not used multiple project not needed required not neededSensetable [15] not needed not used not needed not needed required requiredReady, Steady, Touch! [17] not needed not used not needed required required not neededKats et al. [1] not needed multiple camera not needed not needed not needed not neededTouchLight [19] required multiple camera not needed not needed not needed requiredTouchscreen Everywhere [5] not needed single camera Structured light not needed not needed not neededPressing the Flesh [20] not needed single camera not needed not needed not needed not neededWilson et al. [25] required multiple kinect not needed not needed not needed not neededDIRECT [2] required kinect not needed not needed not needed not neededOmnitouch [3] required kinect not needed not needed not needed not neededDante vision [4] required thermal & kinect not needed not needed not needed not neededSurfaceSight [23] required not used(LiDAR) not needed required not needed not neededOur method not needed single camera not needed not needed not needed not needed

FIGURE 2: Touch sensing intrinsically require only the fin-ger that touches the projected screen. A regular camera cap-tures an entire scene which contains unnecessary informationfor touch sensing, and requires advanced computer vision toperform handtracking.

camera’s rolling shutter, O’Toole et al. showed that only lighttriangulated from a plane some depth away is captured.

Ueda et al. [7] further developed the method of O’Toole etal. and realized only the reflected light from objects withina specific depth, tilt, and thickness region is captured. Inparticular, they leverage hardware parameters to specify thisgeometry of the region. This region was showed to be asloped, near-planar surface.

Figure 3a shows an overview of Slope Disparity Gating.As shown in Figure 3a, Slope Disparity Gating captures onlythe objects on the intersection (shown in red) of the scanningplane of a laser-scanning projector (shown in yellow) and thescanning plane of a rolling shutter camera (shown in blue).In addition, these scanning lines repeatedly sweep up anddown, and the intersecting lines form a plane (the surfacerepresented in red, called the imaging plane). This planecan be set to an arbitrary depth and tilt by controlling thesynchronous timing and scanning speed of the projector and

camera (Figure 4).Slope Disparity Gating has many useful properties in-

cluding real-time optical capture of the selected region androbustness to ambient light.

C. UTILIZING SELECTIVE IMAGING IN OURINTERACTIVE SYSTEMWe employed Slope Disparity Gating (SDG) as a means ofidentifying only the fingertip that is touching the screen.We calibrated the camera parameters (delay, pixel clock,and exposure time) to set the capturing region using SDG.More specifically, we selectively capture the selected regionslightly above the screen plane, horizontal to the screenplane, and a few centimeters thick, as shown in Figure 3b,because this is the interaction region of interest. For theparameter settings, we followed the method of Ueda et al. [7]to determine the values of delay, pixel clock, and exposuretime.

In this way, the camera captures the fingertips only whenthey touch the plane without complex image recognitiontechniques to track them. Also, since our capturing regiondoes not include the projected screen area, we can realizetouch sensing without being affected by surface texture orroughness. We do require a roughly planar surface withno clutter or objects that would protrude into the opticallycaptured region above the surface.

The clear difference between Slope Disparity Gating andour system is in the projection role of the projector. In SlopeDisparity Gating, the projector is just a programmable lightsource device. In our touch-input sensing, the projector isboth a programmable light source and an interface displaydevice.

One notable feature of our system is that the projector canpresent different visuals to each the human-eye and the cam-era simultaneously. The users can see the image projected bythe projector naturally whereas the camera captures only theregion slightly above the screen for touch sensing. In otherwords, the projector simultaneously takes a role of a display-

4 VOLUME 4, 2016

Page 5: Touch Sensing for a Projected Screen Using Slope Disparity

Author et al.: Preparation of Papers for IEEE TRANSACTIONS and JOURNALS

Projector scanline

Imaging plane

Laser scanning projector

Rolling shutter CMOS camera

Camera scanline

(a)

Laser scanning projector

Rolling shutterCMOS camera

Surface

Projection light

Imaging area

(b)

FIGURE 3: (a) The concept of Slope Disparity Gating which allows a single plane in 3D space to be captured by the camera [7].(b) Imaging of a selected region. Although the projector projects the image surrounded by blue lines in perspective projection,our team’s imaging technique enables us to capture only the region surrounded by red lines.

camera scanline

projector scanlinecamera

projector

(a) standard

Imaging plane

(d) thickness control(c) tilt control(b) depth control

FIGURE 4: To control imaging regions, our system utilizes adjusting the parameters in Slope Disparity Gating [7]. Figureshows a side view of Slope Disparity Gating. (a) We can capture the line of the trajectory where the camera scanline and theprojector scanline intersect. (b) Depth can be controlled by adding a delay to the scanning motion of the camera. (c) The tiltcan be controlled by changing the pixel clock of the camera. (d) Thickness can be controlled by changing the exposure time ofthe camera.

interface device for users and a role of the light source for thecamera to selectively illuminates the touch sensing area.

D. TOUCH SENSING ALGORITHM WITH SIMPLE IMAGEPROCESSINGFigure 5 shows the system flowchart. First, using the princi-ple of Slope Disparity Gating, we capture the selected regionslightly above the touch surface, as shown in Figure 3b. Thealgorithm consists of touch detection and fingertip localiza-tion. In the remaining parts of this section, we will show howwe can achieve touch detection and fingertip localization withsimple image processing.

In our method, the system can determine whether a fingeris touching the surface or not using the images from a singlecamera. Figure 6 (a) and (b) show pictures taken with anormal capturing system. In (a), the user touches the surface,and in (b), the user does not touch the surface. Figure 6(c) and (d) show pictures taken with a SDG system, whichcaptured same scene as (a) and (b), respectively. Note that we

use the same camera figures (a) through (d). This approachgives us clear discriminative power whether a fingertip is nearthe screen.

Our imaging method acquires only reflected light from alimited region. Conversely, if there is no reflected light inthe imaging region, the light is not measured and appearsblack in the image. Here, we consider an ideal situationwhere no occluder or object is blocking the surface, and theregion is a free space that only fingers may enter into. Inthis case, theoretically, only the finger parts reflect light fromthe projector and have a non-zero brightness in the capturedimage. Then, if the captured image has some luminancevalue, the finger exists. If the luminance value of the capturedimage is zero, the finger does not exist.

However, when capturing with SDG, some pixels in thespace where the object does not exist may have small pixelvalues due to the influence of ambient light. This ambientlight is relatively weak because the exposure time of ourcapturing system is extremely short. To remove the effect

VOLUME 4, 2016 5

Page 6: Touch Sensing for a Projected Screen Using Slope Disparity

Author et al.: Preparation of Papers for IEEE TRANSACTIONS and JOURNALS

Capture image usingour existing imaging system

Finger touch?

Calculate coordinateson camera image

Homograpy transformation

Touch event

yes

no

FIGURE 5: Algorithm diagram of our touch sensing system

of the small pixel values, a threshold was applied to theimage. Note that this threshold depends on the exposure time,gain, and environment light. After applying the threshold, oursystem calculates the total luminance value of the image andcarries out touch detection. If the total luminance value isnon-zero, the finger touches touched; if it is zero, the imageis not touched.

The center of luminance gravity in the entire image repre-sents the finger position. As shown in Equation (1), we canrepresent the position of a finger in the camera image as thecoordinates of the center of gravity (xc, yc) of the image,which is weighted by the pixel values mi,j at the coordinates(i, j) of the camera image.

(xc, yc) =

∑i,j

mi,j(i, j)∑i,j

mi,j

(1)

Since our touch sensing method is based on a cameraimage, it is necessary to convert the coordinates of a fingertipon the screen into the coordinates of a fingertip estimatedfrom the captured image. Because we assume the screen isflat, we utilize a homography matrix for this transformation.Let C = (xc yc 1)

> and H be respectively the homogeneouscoordinate of the fingers obtained in camera and the homog-raphy matrix, then the fingertips in the screen coordinatesystem C ′ = (xs ys 1)

> can be expressed as Equation (2). xs

ys1

= H

xc

yc1

(2)

To estimate the homography matrix, we collect N (≥ 4)corresponding points between the camera and the screencoordinates. We project N points that uniformly cover theprojection area, then ask the user to touch the points one

by one. Then, we estimate the corresponding finger positionfrom the captured image. This process allows us to collectthe pairs {Cn, C

′n}Nn=1 that can be used to estimate the

homography matrix. The image used in our method wascapturing the selected region slightly above the touch surface,so the coordinates obtained from the image do not mean anaccurate fingertip position. This homographic transformationalso serves as a correction to this fingertip position.

The image processing we have used is simple, as shown inEquations (1) and (2). This is because we no longer requirefingertip segmentation as the fingertip is sensed directly byits intersection with the red region.

IV. EVALUATIONA. CAPTURING SETUPWe illustrate a prototype touch sensing system as shown inFigure 7. We use a Sony MP-CL1A laser scanning projectorwith 1280×720 resolution and 60Hz refresh rate. We alsouse an IDS UI-3250CP-C-HQ RGB camera. This camerasupports both rolling shutter and global shutter, and weuse rolling shutter mode for Slope Disparity Gating. Theresolution of this camera is 1600 × 1200. The hardwaredesign places the projector and camera close to one anotherfor compactness of space. To do this, both a projector and acamera were securely mounted onto a 3D-printed fixed plat-form supported by a tripod. Since the positional relationshipbetween the screen and the system device is fixed, we fixedthe aperture, shutter speed, and focus manually. We open theaperture all the way up so that the finger has a brighter pixelvalue. This is because using SDG produces dark images asambient light is rejected and only limited photons from theregion of interest are captured. In this case, the F# is 1.6.When the aperture is fully open, the depth of field becomesshallow. Our target of the capturing area is slightly above theprojected screen, so we focused the camera on the screenitself. The camera and the projector were synchronizedthrough an Arduino configured to receive and analyze thesignals from the MEMS mirror of the projector and syncthe camera’s rolling shutter to them. This synchronizationmethod is based on the work of Kubo et al. [26]. Forsoftware, our camera and projector are connected to a laptoprunning C# code to implement the touch sensing algorithm.

B. TOUCH EVENTTo demonstrate that our prototype can realize touch sensing,we configured a simple interactive system. In this interactivesystem, when a finger touches the desk surface (Figure 8a),which is the projected screen, a program is implemented todisplay the touch position calculated by the touch sensingsystem as a red circle. In other words, if the red circlecorresponds to the actual touch position, the touch sensingis correctly performed. As a result, as shown in Figure 8b,the calculation result of the center of luminance gravitycorrectly shows the part of the finger. The coordinates of thefinger on the camera image obtained as shown in Figure 8bwas transformed to the coordinates on the projected image

6 VOLUME 4, 2016

Page 7: Touch Sensing for a Projected Screen Using Slope Disparity

Author et al.: Preparation of Papers for IEEE TRANSACTIONS and JOURNALS

FIGURE 6: A comparison between a normal caputuring system and the Slope Disparity Gating (SDG) system, in which a fingeris close to the projected screen. (a) finger touches projected screen. (b) finger is hovering. (c) the exact same scene as (a), butcaptured with SDG. (d) the exact same scene as (b), but captured with SDG. In the (a) and (b) cases, both images look like thefinger is touching, but in the (c) and (d) case, grounding can be determined by whether or not the finger is in the picture.

FIGURE 7: Slope disparity gating hardware prototype withraster-scanning projector aligned with rolling-shutter cameraalong with synchronization electronics via an Arduino.

through the homography transformation and displayed as ared circle on the image. As shown in Figure 8c, the red circleindicating the touch position is projected on the fingertippart, indicating that the touch position of the fingertip canbe identified by our touch sensing system.

(a) (b) (c)

FIGURE 8: (a) Adjusting the angle of the hardware toproject the image onto the desk surface, (b) calculating theluminance center of gravity of the SDG image when a fingertouches the surface, (c) calculating the touch position byapplying the homography transformation shown as a redcircle.

To evaluate the accuracy of touch sensing, we designed atask in which the user touches the center of a crosshair on aknown coordinate. We used a target image with 11 crosshairs(5 pixel line width) on a 1280×720 pixel size canvas, whichis the same as the resolution of the projector, as shown inFigure 9. The user was instructed to touch all the intersectionsof the 11 crosshairs with their bare hands.

FIGURE 9: Target image of crosshairs projected for touchaccuracy evaluation.

For each trial, we calculated the error between thecrosshairs’ position and the calculated touch position of thefinger for a total of 110 cumulative touch trials with fivepeople in their twenties. As a result, an average horizontalerror was 16.8 pixels towards the side where the user wasstanding, and an average vertical error was 4.2 pixels towardsthe side where the projector was set. Note that this error is incrosshair image coordinates. If we consider this at the scaleof the projected image, we can see that the size of the fingerbelly affects the error (Figure 10). Simply put, the error inour method is in a range that is not unnatural to the user.

FIGURE 10: Adapting the error range to the crosshair image.(a) the error scale on the crosshair mage, (b) project (a) in aenvironment of touch-accuracy experiment and compare it toa finger scale, (c) Looking at the touch of the crosshairs, theerror does not deviate from the range of the finger’s belly.

C. STRESS TESTINGWe did experiments with negative vectors: low light and

VOLUME 4, 2016 7

Page 8: Touch Sensing for a Projected Screen Using Slope Disparity

Author et al.: Preparation of Papers for IEEE TRANSACTIONS and JOURNALS

defocus from the camera.We calculated the difference between the results of touch

sensing at 710 lux with daily lighting and touch sensing at 5lux with no lighting at all. An average horizontal error was0 pixels towards the side where the user was standing, andan average vertical error was 4.8 pixels. This proves that ourdevice is robust to lighting conditions.

On the other hand, when the camera is out of focus, theaccuracy of touch sensing is significantly reduced. This isbecause it is difficult to control the thickness of the capturedarea as the depth of field becomes shallower, and parts of thehand other than the fingertip are also captured, which affectsthe coordinate calculation.

D. HAND DETECTIONTo verify the robustness of this touch sensing system to video,we examine the case in which the projected contents includehands image. When preparing the image to be projectedfrom the projector, we used a realistic image including ahand quoted from unsplash, which is a stock photographyservice. As a typical example of hand detection, we usedHandtrack [27].

We captured a scene where a user interacts with projectedcontent that includes fake hands as a test of Handtrack’s handrecognition. Handtrack detected the projected fake hand butcould not detect the real hand, as shown in Figure 11(b). Thisresult shows the brittleness of Handtrack in distinguishingbetween real and fake hands.

In contrast, our method eliminates the projected contentoptically as shown in Figure 11(c). Therefore, compared toconventional methods using general camera imaging, oursystem can make touch sensing performance robust to theprojected scene content.

V. CONCLUSION AND FUTURE WORKA. CONCLUSIONWe have developed a novel touch sensing system leveragingSlope Disparity Gating. The important aspect of this studyis to capture only the region slightly above the projectedscreen necessary for sensing. Based on this point of view,we synchronized a projector with a rolling-shutter camerafor the selected region capturing. By not capturing the screensurface in the first place, we realized touch sensing that isrobust to the visual information of projected contents. Also,By eliminating capturing scene unnecessary for the sensing,we realized touch sensing algorithm that can both touch de-tection and fingertip localization from a single camera image.Our touch sensing method can be realized by simple imageprocessing. Our system leverages the projector as both theactive illumination for touch sensing as well as to display usercontent. there is no need for multiple cameras and additionalinfrared devices.

B. LIMITATION AND FUTURE WORKThere are several limitations to our proposed approach. It isimpossible to do multiple planes of sensing at once with-

(a)

(b)

(c)

(d)

FIGURE 11: Robustness to scene content: (a) the scene thatfinger touches a projected image of many synthetic hands,(b) the result of hand detection by HandTrack [27] (detectingonly fake hand), (c) the image taken with the selectivecapturing technique in our system, (d) visualization of thedetected finger in the original image.

out changing hardware parameters. This prevents dynamictracking applications. Further, occlusion cannot be overcomewith our single camera system. Our system needs calibrationat the beginning to align to the surface of interest. If thedepth of field of the camera becomes too shallow and thecamera is not focused on the surface, it becomes difficult tocontrol the thickness of the captured area, leading to a loss ofaccuracy. In this study, a false positive does not occur if theimaging area is set appropriately in an environment with noobject other than fingers placed on the screen. However, if anobstacle other than fingers exist on the projected screen, suchas a cup placed, a false positive may occur. In practice, weexpect our system to be used for surfaces with no clutter oroccluding objects to allow full visibility for the user.

There are also several future works. In a practical scene,several artifacts such as dropping frames can sometimeshappen because of the camera hardware issues. Since ourdetection algorithm runs on a frame-by-frame basis inde-pendently, our method could suffer from temporal jitter.However, this issue could be easily solved by applying atemporal smoothing filter which should be a future work ofour paper.

In this paper, we consider only single touch for our userinteraction. Multi-touch is possible in principle using ouralgorithm, but we leave it to future work to fully characterizemultitouch interaction for the system.

We will aim to address the above challenges, but alsoenable touchless operations such as airborne operation andgesture operation.

8 VOLUME 4, 2016

Page 9: Touch Sensing for a Projected Screen Using Slope Disparity

Author et al.: Preparation of Papers for IEEE TRANSACTIONS and JOURNALS

REFERENCES[1] Itai Katz, Kevin Gabayan, and Hamid Aghajan. A multi-touch surface

using multiple cameras. In Proceedings of the 9th International Conferenceon Advanced Concepts for Intelligent Vision Systems, ACIVS’07, page97–108, Berlin, Heidelberg, 2007. Springer-Verlag.

[2] Robert Xiao, Scott Hudson, and Chris Harrison. Direct: Making touchtracking on ordinary surfaces practical with hybrid depth-infrared sensing.In Proceedings of the 2016 ACM International Conference on InteractiveSurfaces and Spaces, ISS ’16, page 85–94, New York, NY, USA, 2016.Association for Computing Machinery.

[3] Chris Harrison, Hrvoje Benko, and Andrew D. Wilson. Omnitouch:Wearable multitouch interaction everywhere. In Proceedings of the 24thAnnual ACM Symposium on User Interface Software and Technology,UIST ’11, page 441–450, New York, NY, USA, 2011. Association forComputing Machinery.

[4] E. N. Saba, E. C. Larson, and S. N. Patel. Dante vision: In-air and touchgesture sensing for natural surface interaction with combined depth andthermal cameras. In 2012 IEEE International Conference on EmergingSignal Processing Applications, pages 167–170, 2012.

[5] J. Dai and C. R. Chung. Touchscreen everywhere: On transferring anormal planar surface to a touch-sensitive display. IEEE Transactions onCybernetics, 44(8):1383–1396, 2014.

[6] Matthew O’Toole, Supreeth Achar, Srinivasa G. Narasimhan, and Kiri-akos N. Kutulakos. Homogeneous codes for energy-efficient illuminationand imaging. ACM Trans. Graph., 34(4), July 2015.

[7] T. Ueda, H. Kubo, S. Jayasuriya, T. Funatomi, and Y. Mukaigawa. Slopedisparity gating using a synchronized projector-camera system. In 2019IEEE International Conference on Computational Photography (ICCP),pages 1–9, 2019.

[8] SK Lee, William Buxton, and K. C. Smith. A multi-touch three dimen-sional touch-sensitive tablet. SIGCHI Bull., 16(4):21–25, April 1985.

[9] D. Wei, S. Z. Zhou, and D. Xie. Mtmr: A conceptual interior design frame-work integrating mixed reality with the multi-touch tabletop interface. In2010 IEEE International Symposium on Mixed and Augmented Reality,pages 279–280, 2010.

[10] Jun Rekimoto. Smartskin: An infrastructure for freehand manipulation oninteractive surfaces. In Proceedings of the SIGCHI Conference on HumanFactors in Computing Systems, CHI ’02, page 113–120, New York, NY,USA, 2002. Association for Computing Machinery.

[11] Chris Harrison, Desney Tan, and Dan Morris. Skinput: Appropriatingthe body as an input surface. In Proceedings of the SIGCHI Conferenceon Human Factors in Computing Systems, CHI ’10, page 453–462, NewYork, NY, USA, 2010. Association for Computing Machinery.

[12] Paul Dietz and Darren Leigh. Diamondtouch: A multi-user touch tech-nology. In Proceedings of the 14th Annual ACM Symposium on UserInterface Software and Technology, UIST ’01, page 219–226, New York,NY, USA, 2001. Association for Computing Machinery.

[13] Robert Xiao, Greg Lew, James Marsanico, Divya Hariharan, Scott Hudson,and Chris Harrison. Toffee: Enabling ad hoc, around-device interactionwith acoustic time-of-arrival correlation. In Proceedings of the 16thInternational Conference on Human-Computer Interaction with MobileDevices and Services, page 67–76, New York, NY, USA, 2014. Associ-ation for Computing Machinery.

[14] M. Ashdown and P. Robinson. Escritoire: A personal projected display.IEEE MultiMedia, 12(01):34–42, jan 2005.

[15] James Patten, Hiroshi Ishii, Jim Hines, and Gian Pangaro. Sensetable:A wireless object tracking platform for tangible user interfaces. InProceedings of the SIGCHI Conference on Human Factors in ComputingSystems, CHI ’01, page 253–260, New York, NY, USA, 2001. Associationfor Computing Machinery.

[16] A. Nguyen and A. Banic. 3dtouch: A wearable 3d input device for 3dapplications. In 2015 IEEE Virtual Reality (VR), pages 373–373, 2015.

[17] Yilei Shi, Haimo Zhang, Kaixing Zhao, Jiashuo Cao, Mengmeng Sun, andSuranga Nanayakkara. Ready, steady, touch! sensing physical contact witha finger-mounted imu. Proc. ACM Interact. Mob. Wearable UbiquitousTechnol., 4(2), June 2020.

[18] Julien Letessier and François Bérard. Visual tracking of bare fingers forinteractive surfaces. In Proceedings of the 17th Annual ACM Symposiumon User Interface Software and Technology, UIST ’04, page 119–122,New York, NY, USA, 2004. Association for Computing Machinery.

[19] Andrew D. Wilson. Touchlight: An imaging touch screen and displayfor gesture-based interaction. In Proceedings of the 6th InternationalConference on Multimodal Interfaces, ICMI’04, page 69–76, New York,NY, USA, 2004. Association for Computing Machinery.

[20] Joe Marshall, Tony Pridmore, Mike Pound, Steve Benford, and BorianaKoleva. Pressing the flesh: Sensing multiple touch and finger pressureon arbitrary surfaces. In Proceedings of the 6th International Conferenceon Pervasive Computing, Pervasive’08, page 38–55, Berlin, Heidelberg,2009. Springer-Verlag.

[21] Andrew D. Wilson and Hrvoje Benko. Combining multiple depth camerasand projectors for interactions on, above and between surfaces. In Pro-ceedings of the 23nd Annual ACM Symposium on User Interface Softwareand Technology, UIST’10, page 273–282, New York, NY, USA, 2010.Association for Computing Machinery.

[22] J. A. Paradiso, K. Hsiao, J. Strickon, J. Lifton, and A. Adler. Sensorsystems for interactive surfaces. IBM Syst. J., 39(3–4):892–914, July2000.

[23] Gierad Laput and Chris Harrison. Surfacesight: A new spin on touch, user,and object sensing for iot experiences. In Proceedings of the 2019 CHIConference on Human Factors in Computing Systems, CHI ’19, page1–12, New York, NY, USA, 2019. Association for Computing Machinery.

[24] Microvision. Product brief:short-throw interactive display module &starter kit. http://www.microvision.com/wp-content/uploads/2020/01/DB0140010_MV-2407sti-421_Product_Brief.pdf, 2020. (Accessed onJune 2021).

[25] Andrew D. Wilson. Playanywhere: A compact interactive tabletopprojection-vision system. In Proceedings of the 18th Annual ACMSymposium on User Interface Software and Technology, UIST’05, page83–92, New York, NY, USA, 2005. Association for Computing Machinery.

[26] Hiroyuki Kubo, Suren Jayasuriya, Takafumi Iwaguchi, Takuya Funatomi,Yasuhiro Mukaigawa, and Srinivasa G. Narasimhan. Programmable non-epipolar indirect light transport: Capture and analysis. IEEE Transactionson Visualization and Computer Graphics, 27(4):2421–2436, 2021.

[27] Dibia Victor. Handtrack: A library for prototyping real-time hand tracking-interfaces using convolutional neural networks. GitHub repository, 2017.

MAYUKA TSUJI is currently a Ph.D. studentat division of Information Science, Nara Instituteof Science and Technology (NAIST), Japan since2021. She received M.S. degrees in Engineeringfrom Nara Institute of Science and Technology,Japan, in 2021. Her research interests includecomputer vision, computational photography, andcomputer graphics.

HIYOYUKI KUBO is currently an specially ap-pointed associate professor at Tokai university,Japan since 2020. He was an assistant professor atNara Institute of Science and Technology (NAIST)until 2020. He received his M.S. and Ph.D. degreesfrom Waseda University, in 2008 and 2012, re-spectively. His research interests include computergraphics and computer vision.

VOLUME 4, 2016 9

Page 10: Touch Sensing for a Projected Screen Using Slope Disparity

Author et al.: Preparation of Papers for IEEE TRANSACTIONS and JOURNALS

SUREN JAYASURIYA is an assistant professor atArizona State University (ASU), USA since 2018.He was a postdoctoral fellow at the Robotics Insti-tute at Carnegie Mellon University, USA in 2017.He received his B.S. degree in mathematics anda B.A. degree in philosophy from the Universityof Pittsburgh, PA in 2012, and received his Ph.D.degree in Electrical and Computer Engineeringfrom Cornell University, USA in 2017. His re-search interests are in computational photography

and imaging, computer vision, and image sensors. He is a member of IEEE.

TAKUYA FUNATOMI is currently an associateprofessor at division of Information Science, NaraInstitute of Science and Technology (NAIST),Japan since 2015. He was an assistant professor atKyoto University, Japan from 2007 to 2015, and avisiting assistant professor at Stanford University,USA in 2014. He received B.S. degree in Engi-neering, M.S. and Ph.D. degrees in Informaticsfrom Graduate School of Informatics, Kyoto Uni-versity, Japan, in 2002, 2004, and 2007, respec-

tively. His research interests include computer vision, computer graphics,and pattern recognition. He is a member of IEEE.

YASUHIRO MUKAIGAWA is currently a pro-fessor at division of Information Science, NaraInstitute of Science and Technology (NAIST),Japan since 2014. He was a research associate atOkayama University, Japan from 1997 to 2003,an assistant professor at University of Tsukuba,Japan from 2003 to 2004, an associate professorat Osaka University, Japan from 2004 to 2014. Hereceived his ME and Ph.D. degrees from Univer-sity of Tsukuba in 1994 and 1997, respectively.

His research interests include photometric analysis and computational pho-tography. He is a member of IEEE.

10 VOLUME 4, 2016