More about HKUST
A Survey of CrowdSourcing: Query Processing with People
PhD Qualifying Examination
Title: "A Survey of CrowdSourcing: Query Processing with People"
by
Mr. Zhou ZHAO
Abstract:
Some complex problems, such as image tagging and sentiment analysis, are
challenge for computers where even the state-of-the-art technologies are
far from perfect. For instance, given a set of hotels with their locations
and photos, the user may want to know ”which are the good hotels”.
Fortunately, Crowdsourcing markets like Amazon’s Mechanical Turk (MTurk)
make it possible to divide these complex jobs into small tasks and assign
to workers (humans) to judge. However, the workers are obviously different
from machines. Nevertheless, their performance are uncertain in the
latency, expected reward and the quality of the work. Furthermore, the
users may even pose time or monetary budget on their crowdsourced jobs.
Recently, considerable research has been conducted on the Crowdsourcing
problems in the database re- search community. Because the subjective
database operators are involved in the applications of the Crowd-
sourcing. In this survey, we first introduce different state-of-the-art
Crowdsourcing databases. Then we review different Crowdsourcing techniques
to tackle the database problems such as join and max. Besides, we study
the Crowdsourcing problems in the social network area. Finally, we
conclude this survey by giving some further research directions from the
recent work.
Date: Wednesday, 22 August 2012
Time: 10:30am - 12:30pm
Venue: Room 3501
lifts 25/26
Committee Members: Dr. Wilfred Ng (Supervisor)
Prof. Dik-Lun Lee (Chairperson)
Dr. Qiong Luo
Dr. Raymond Wong
**** ALL are Welcome ****