Difference between revisions of "Winter Milestone 8"

From crowdresearch
Jump to: navigation, search
(Created page with "'''Due date (PST): 8:00 pm 6th March 2016 for submission, 12 pm 7th March 2016 for peer-evaluation.''' This week, we will refine methods and systems proposed last week furthe...")
(Topic background)
Line 16: Line 16:
=== Topic background ===
=== Topic background ===
'''Michael's summary from last week's hangout:'''
Check out our '''[https://docs.google.com/document/d/1ada3U8fUZKp9emoFpOPRubFI0C5ryZSxdw1d4JU0rCQ/edit?usp=sharing system proposal]''', mainly first few pages and comments to catch up on what we've been working on and thinking about.
We have aligned on a specific vision for a smarter, more informative task feed by incentivizing workers and requesters to share information they might not otherwise share, or not share accurately. There are three main components to this:
'''The task feed hangouts from previous weeks:'''
*Reputation. It shows more accurate reputation information, by influencing which workers get your future tasks (if you're a requester), and which requesters show up at the top of your feed (if you're a worker). The individual incentive: you get better workers, or better requesters, by reporting honestly. The global win: the reputation scores more directly reflect individual incentives. This is Boomerang as previously described.
*Hourly rate. The goal is to for the task feed to show an estimate of how much you'd make with each task (e.g., $9/hr). To do so, it asks workers who just completed the task to estimate how long it took them to do it. The individual incentive: it uses the worker's responses to build a model to estimate their effective hourly rate for all the other tasks in the marketplace. The global win: those workers' responses are used to produce estimates shown to all other workers.
*Rejection information. It shows the % of tasks for workers like you that get rejected, by influencing which workers get the requesters' future tasks based on rejection information. The individual incentive: the more of a worker's tasks a requester accepts, the earlier they get access to their future tasks. (This is a smaller effect than the reputation feedback above, but does have an impact.) This prevents "accept all submissions" degenerate behavior. The global win: workers can now see the % of tasks accepted for workers like them.
''' Boomerang: Incentivizing Information Disclosure in Paid Crowdsourcing Platforms'''
There is a massive amount of information necessary for a healthy crowdsourcing marketplace — for example accurate reputation ratings, skill tags on tasks, and hourly wage estimates for tasks — that is privately held by individuals, but rarely shared. We introduce Boomerang, an interactive task feed for a crowdsourcing marketplace, that incentivizes accurate sharing of this information by making the information directly impact their future tasks or workers. Requesters' ratings of workers, and their skill classifications of tasks, are used to give early access to workers who that requester rates highly and who are experts in that skill, so giving a high rating to a mediocre worker dooms the requester to more mediocre work from that worker. Workers' ratings of requesters are used to rank their high-rated requesters at the top of the task feed, and their estimates of active work time are used to estimate their hourly wage on other tasks on the platform.
'''The task feed hangouts from last week:'''
* Youtube link of the task feed meeting 1: [http://www.youtube.com/watch?v=OeBoiA07BBk watch]  
* Youtube link of the task feed meeting 1: [http://www.youtube.com/watch?v=OeBoiA07BBk watch]  
* Youtube link of the task feed meeting 2: [http://www.youtube.com/watch?v=42d5H0yoPPc watch]
* Youtube link of the task feed meeting 2: [http://www.youtube.com/watch?v=42d5H0yoPPc watch]

Revision as of 19:59, 29 February 2016

Due date (PST): 8:00 pm 6th March 2016 for submission, 12 pm 7th March 2016 for peer-evaluation.

This week, we will refine methods and systems proposed last week further:

  • Youtube link of the meeting today: watch
  • Winter Meeting 8 slideshow: coming soon...

Task Feed

We have 3 goals for this week:

  • Refining our system proposal. Please make comments directly on the google doc. We want to address any holes in our incentive compatible structure and make any concrete decisions regarding things like the # of workers that report time spent on a task before we start displaying it on the taskfeed.
  • Design tweaks in order to capture and display the new information we need. This includes a task element to record the time a worker spent and small modifications to the taskfeed to display the rejection rate and effective wage. Please post low-fi or high-fi mocks to the #taskfeed channel.
  • Engineering plan of action and division of work. Led by @dmorina.

Topic background

Check out our system proposal, mainly first few pages and comments to catch up on what we've been working on and thinking about.

The task feed hangouts from previous weeks:

  • Youtube link of the task feed meeting 1: watch
  • Youtube link of the task feed meeting 2: watch
  • Youtube link of the task feed meeting 3: watch

Michael's synthesized needs:

  • to find new tasks that will maximize income (reduce uncertainty in payment, rejection, maximize certainty in what will be asked of me and how quickly I can do it)
  • to find new tasks that fit my expertise profile
  • to refind old requesters' new tasks, since I know I like them

In addition...

  • to identify tasks I can do on my own time
  • to learn new skills

Task Authoring

Our goal for this week is to run experiments, then analyze the findings and write up a report. By the end of the week, we want to produce a report with findings from our pilot experiments - this will help us in making future decisions.

First half of the week (Volunteer to be a requester): Pilot participants should try to finish their HTML task designs, that takes about an hour. Please message @catherine.mullings for that. Once we have your task designs, @dmorina will post them on Mechanical Turk for you. The taskauthoring folks are looking for a couple volunteers to create some tasks for four datasets that we have. We give you the task and some example input/outputs, and you try to write a task interface in Mechanical Turk that will get workers to produce the right answers.

Second half of the week (analyze+report): We analyze the results from the experiment, conduct a Google hangout with @michaelbernstein, and produce a report on the findings. We can start making contributions to the report here on Google doc.

Open Gov and Design

Check out this week's meeting, and based on the open gov discussions here - create a mock, minimal design within Daemo. Think about questions like: how does it work as a system? how it would fit in Daemo? Like, walk us through. I’m a new worker on Daemo. What do I do? Am I already part of a guild? How do I get into one? How do I get work once I’m in one? What if the requester doesn’t like what I do? And how does all this solve the reputation problem? Try to pitch a story/wireframe with specific design that has as few moving parts as possible.

Helpful links:

You can use balsamic or Google slides to give shape to your ideas. Design folks, come join and help move this effort forward. Once you're done, post here: http://crowdresearch.meteor.com/category/open-gov