Labor Marketplace Applications: Human Computer Interaction KAIST KSE Uichin Lee 9/26/2011

Preview:

Citation preview

Labor Marketplace Applications: Human Computer Interaction

KAIST KSEUichin Lee9/26/2011

Soylent: A Word Processor with a Crowd Inside

Michael Bernstein, Greg Little, Rob Miller, David Karger, David Crowell, Katrina Panovich

MIT CSAILBjörn Hartmann, Mark Ackerman

UC Berkeley University of MichiganUIST 2010

Part of slides are from http://projects.csail.mit.edu/soylent/

Soylent: a word processing interface

• Uses crowd contributions to aid complex writing tasks.

Challenges for Programming Crowds

• The authors have interacted with ~9000 Turkers on ~2000 different tasks

• Key Problem: crowd workers often produce poor output on open-ended tasks– High variance of effort: Lazy Turker vs. Eager Beaver– Errors made by Turkers; e.g., “Of Mice and Men” “Of

Mice” or calling it a movie..• 30% Rule: ~30% of the results from open-ended

tasks will be unsatisfactory• Solution: find-fix-verify

(at least 20% of the Turkers agree upon)

(e.g., given to 10 Tuckers)

(e.g., given to 5 Tuckers)

(e.g., given to 5 Tuckers)

Find-Fix-Verify Discussion• Why split Find and Fix?

– Force Lazy Turkers (tend to choose easiest work) to work on a problem of our choice– Allows us to merge work completed in parallel

• Why add Verify?– Quality rises when we place Turkers in productive tension– Allows us to trade off lag time with quality

• Timeout at each stage for responsive operations (e.g., 10-20 minutes)

Evaluation

• Implementation– Microsoft World plug-in using MS Visual Studio

Tools for Office (VSTO) and Windows Presentation Foundation (WPF)

– TurKit Mechanical Turk toolkit • Evaluation:– Shortn, Crowdproof, Human Macro

• Metrics: quality, delay, cost

Shortn Evaluation

• Setting:– Find: 6-10 workers ($0.08 per Find)– Fix/Verify: 3-5 workers ($0.05 per Fix, $0.04 per Verify)– Delay: wait time, work time

• Results:– 78%-90% reduction.. – Wait time of all stages: median 18.5 minutes– Work time: median 118 seconds – Average costs: $1.41 ($0.55 + $0.48 + $0.38)– Caveats: some modifications are grammatically appropriate,

but stylistically incorrect

Crowdproof Evaluation

• Tested 5 input texts– Manually labeled all spelling, grammatical and style

errors in each of the five inputs (total 49 errors)• Ran Crowdproof w/ a 20-minute stage timeout; and

measure how many corrections Crowdproof makes– Task setup: Find: $0.06, Fix: $0.08, Verify: $0.04

• Soylent: 67% vs. MS Word (grammar check): 30%– Combined: 82%

Human Macro Evaluation• Give 5 prompts (Input and Output) to users (CS students, Admin, Author)• Ask to generate descriptions (Request) used for Human Macro tasks

Discussion

• Delay in interface outsourcing: minutes to hours..

• Privacy? Confidential documents?• Legal ownership?• Lack of domain knowledge or shared context

to usefully contribute

VizWiz: Nearly Real-time Answers to Visual Questions

Jeffrey P. Bigham, Chandrika Jayant, Hanjie Ji, Greg Little, Andrew Miller,

Robert C. Miller, Robin Miller, Aubrey Tatarowicz, Brandyn White, Samual White, and Tom Yeh. UIST

2010.

Part of slides are from: http://husk.eecs.berkeley.edu/courses/cs298-52-sp11/images/8/8d/Vizwiz_soylent.pdf

VizWiz

• “automatic and human-powered services to answer general visual questions for people with visual impairments.”

• Lets blind people use mobile phone to:1. Take a photo2. Speak a question3. Receive multiple spoken answers

Motivation

• Current technology uses automatic approaches to help blind people access visual information– Optical Character Recognition (OCR) – Ex) Kurzweil knfbReader: ~$1,000

• Problems: error-prone, limited in scope, expensive– Ex: OCR cannot read graphic labels, handwritten

menu, street sign

Motivation

• Solution: ask real people• Can phrase questions naturally– “What is the price of the cheapest salad?” vs. OCR

reading the entire menu• Feedback– Real people can guide blind people to take better

photos• Focus on blind people’s needs, not current

technology

Human-Powered Services

• Problem: Response latency• Solution: quikTurkit (and some tricks)– “First attempt to get work done by web based

workers in nearly real-time”– Maintain a pool of workers to answer questions

quikTurkit

• “requesters create their own web site on which Mechanical Turk workers answer questions.”

• “answers are posted directly to the requester’s web site, which allows [them] ... to be returned before an entire HIT is complete.”

• “workers are required to answer multiple previously-asked questions to keep workers around long enough to possibly answer new questions”

Answering Site

Deployment• Setting: 11 blind iPhone users• quikTurkit:

– Reward range: $0.01 (1/2 of jobs), $0.02/3 (1/2)• Results: 86.6% of first answers “correct”

– Average of 133.3s latency for first answer

Problems: Photos too dark or too blurry and thus unanswerable.

VizWiz 2.0 detects and alerts users if photo is too dark or blurry

VizWiz: LocateIt

• Combine VizWiz with computer vision to help blind people locate objects

LocateIt mobile: Sensor (zoom and filter) and sonification modules

Web interface

(angle)

(distance)

Future Work

• Expand worker pools beyond Mechanical Turk (e.g. social network)

• Reduce cost by using game, volunteers, friends

• Improve interface to make photo-taking easier for blind people

• Combine automatic approaches to improve delay

Discussion

• Resource usage vs. delay (wasting resources for better responses or near real-time services?)– Any better approach than quikTurkit?

• Quality control? How do we make sure that the workers correctly identified the photos?– How do systems accept/reject the submissions? (if

it’s becoming a large scale service?)• Other application scenarios with quikTurkit? • Adding inertial sensors to LocateIt?

Recommended