Due date (PST): 11:59 pm 18th March 2015 for submission, 9 am 20th March 2015 for peer-evaluation.
This week, we will take the set of needs that we collectively identified in the previous milestone and use those insights to generate design ideas.
- 1 Needs from Milestone 2
- 2 Recommended Readings
- 3 Initial Brainstorm
- 4 Dive Deeper into Specific Ideas
- 5 Dark Horse idea
- 6 Submitting
Needs from Milestone 2
We synthesized the main needs groups identified in Milestone 2 in the following table:
|Workers need to be able to quickly find tasks they'd want to work on||Monday evening panel workers cite the challenge in identifying or distinguishing tasks because of poor tagging, Reddit discussion also cited the exorbitant amount of time that they spend trying to find tasks and do the mental calculations to find the opportunities that match them best (e.g. time to complete the task on average, average $ per minute on task, requirements to complete the task).||Finding a task that matches well with the worker's skillset and pays well takes a significant amount of (unpaid) time.|
|Workers need to feel they are being fairly compensated for their work.||Reddit discussion cites that the payment system for HITs is not adaptive and does not take into account changing marketplace conditions (supply/demand) and pricing of tasks based on those changes.||Monetary compensation is the primary motivator for many crowd workers.|
|Workers need to feel like they are treated fairly and respectfully, and have a voice in the platform||Comment on Turkopticon: "Got a mass rejection from some hits I did for them! Talked to other turkers that I know in real life and the same thing happened to them. There rejection comments are also really demeaning. Definitely avoid!"||Unreasonable rejections and low payments lead workers to feel disrespected. The implicit assumption on MTurk is that workers are unskilled and replaceable. They can do little if their work is rejected.|
|Workers need to be able to expose their skills so they can get work they are qualified for and advance their skills||Monday evening panel workers from oDesk cite that most employers will not work with them until they have enough feedback or past work on the platform||If users cannot get new work without feedback, this makes it difficult for new users to establish their reputation and get jobs that will help develop their skillsets.|
|Requesters need to get their HITs completed (quickly / correctly)||Requester asking on forum why nobody is doing his HITs (7-minute, 25-cent surveys - a very low wage)||Requesters want their HITs done, and when nobody's doing them, they do not know the reason why (e.g. it is because he is underpaying workers)|
|Requesters need to be able to trust the results they get||Requesters will often rely on previous workers whose results they can trust, and add mechanisms to detect spammers, or manually verify some results.||If spammers are not caught, this brings the correctness of results into question. If requesters are not sure the results are correct, they may need to discard the data.|
|Requesters need to have workers who have the appropriate skills and demographics do their tasks||Requesters worry that they are not able to verify self-reported demographics for surveys.||Workers' self-reported skills and demographics are often not viewed as trustworthy. This is a problem for surveys, which need to have correct demographic data to be useful.|
|Requesters need to be able to easily generate good tasks||Companies hire full-time developers to deal with the complexities of posting microtasks on MTurk. Requesters often develop their own tools and workflow systems on top of Amazon's.||The process of authoring HITs is currently difficult and makes crowd-work inaccessible to potential requesters|
|Requesters need to price their tasks appropriately||Requesters asking on forums about the appropriate amount they should pay for their HITs||Requesters often don't have a good intuition of what the appropriate wage for their task would be in terms of price per HIT.|
|Requesters need workers to trust them||Requesters say they are reluctant to reject work, because they fear they might get bad reviews.||Workers are more likely to do HITs if the requester seems trustworthy. Requesters do not want bad reviews, because they may result in workers ignoring the requester's HITs|
Michael Bernstein's synthesis
These needs boil down to two main issues: 1) trust, and 2) power.
- How do I trust who you say you are?
- How do I trust that the results I get are results that will be good?
- How do I trust that you’ll respect me as a worker, and pay me accordingly?
- Who has the power to post work?
- To edit other peoples’ posted work?
- To return results to the requester? Can I, as a worker, send it back myself, or does someone else need to vet it?
As we brainstorm, we should be thinking about solutions that holistically address these issues of power and trust, not just surface fixes that get at micro-elements of the system.
Coming up with good, novel visions and ideas is a crucial part of doing successful research, and reading other researchers' visions and ideas can help you come up with better ideas yourself. These readings discuss visions and ideas that crowdsourcing researchers have thought of related to future crowd marketplaces. This week's readings are optional and don't have a deliverable, but are highly recommended.
Design notes for a future crowd work market
Design notes for a future crowd work market - This is a Medium post written by researchers involved with Turkopticon in response to hearing about this research project. It discusses their vision for a future crowd marketplace, where workers are more involved in the management of the marketplace.
The future of crowd work
Kittur A, Nickerson J V, Bernstein M, et al. The future of crowd work. Proceedings of the 2013 conference on Computer supported cooperative work. ACM, 2013: 1301-1318. - This is a vision of a future crowd marketplace that emphasizes on workers' long-term development, and where people can be proud to be workers. It is a long paper. Feel free to focus on just the parts that particularly interest you.
Now it's time for your team to brainstorm some ideas based on these needs.
Work with your team to brainstorm as many ideas as you can under two headings: trust and power. Here are some examples of " How might we" questions (a technique which can inspire specific brainstorms) which can drive the generation of your ideas.
- “ How might we” enable workers to trust the requester’s intention to pay?”
- “How might we” enable requesters to trust the results they get back?
Use whatever tools you need - if you can get together in-person, whiteboards and sketchbooks are great tools, while if you're a remote team, services like Google Docs and sketchboard.io should help you with the brainstorming process.
Sketch out enough ideas until you find a set that you’re inspired to explore further. It should be at least 20 ideas total - 10 ideas for trust, and 10 ideas for power This brainstorm should be wild and broad. Focus not on usability patches, but deeper design innovations.
The ideas you brainstormed, (at least 10 ideas for trust, and at least 10 ideas for power). Provide them in whatever format you want - diagrams, sketches, descriptions, or a combination (the wiki supports images, see here for instructions on uploading them).
Dive Deeper into Specific Ideas
Select 2 ideas per heading (trust, power) that you would like to pursue, and expand on them a bit further. Tell us:
- What are the goals of the design? For example, Google's Android design goals are: delight me in surprising ways, simplify my life, and make me amazing (e.g., grant me special powers).
- Which aspects of your design reflect each goal? How does your design solution addresses the users' needs?
For each of the 4 ideas (2 for trust, 2 for power), describe (using diagrams, sketches, storyboards, or some combination) the ideas in further detail.
Dark Horse idea
Now that you've identified some design directions you like, it's time to change tack and toss in a dark horse idea. A dark horse, in horse racing, is a contender who most people don't think will win, but may turn in an unexpectedly strong performance and produce a huge payoff. Dark horse ideas are intended to be something far out there or nearly impossible. In the best case, your dark horse ideas might end up winning the race. However, even in the worst case, they can give us tremendous design insight and prevent design fixation, where the design space shrinks too rapidly.
There are three requirements for dark horse ideas. First, they must be "dark": they must explore a space that is risky, radical, infeasible, and/or in a direction orthogonal to previously explored solutions. They should feel slightly uncomfortable. Second, they must be brainstormed after the more traditional ideas — you can't have a dark path without a traditional "light" path to contrast it against. Third, they must be refined enough that they could be prototyped and objectively tested. That is, it cannot be infeasible: it needs to be something that we could put in front of real people to see whether it would work.
If you have dark horse ideas that came up in your initial brainstorm, you can use that. If you're not satisfied, brainstorm some! Try using Powers of Ten and other techniques to push further and generate even more. After you brainstorm and sketch out dark horse ideas, choose one that you'd like to include among your set of two top candidates from before. Expand on your dark horse idea like you did in the previous section.
Describe your dark horse idea (using diagrams, sketches, storyboards, text, or some combination)
Please start early, so you have time to ask questions or figure out Crowdgrader system. Thanks.
Create a Wiki Page for your Team's Submission
Please create a page for your team's submission at http://crowdresearch.stanford.edu/w/index.php?title=Milestone_3_YourTeamName&action=edit (substituting in YourTeamName with the team name), copy over the template at Milestone 3 Template . If you have never created a wiki page before, please see this or watch this.
Submit on CrowdGrader and do Peer Evaluations
After you have put your team's submission on the wiki, post the link to the wiki page you created on CrowdGrader!
Step 1 for everyone - For the most of you, you don't have to enroll. I have done it for you. You can directly go to: http://www.crowdgrader.org/crowdgrader/venues/view_venue/879 . However, if you cannot access it, please self-enroll using this link: http://www.crowdgrader.org/crowdgrader/venues/join/879/dufipo_fivuvy_tunyge_qedumy
Step 2 for team leaders: [[[Repeats every week]]] Make sure all of your team members have enrolled into the system (though I have done it for you, please double check). Now add them to your group/team - there's an option to add collaborators (your team members might get an email, and have to confirm before it shows up as your collaborators). Yes, you have to repeat the process, we're working with Crowdgrader, so you don't have to do it every week. However, for now, please give yourself enough time, so you can add collaborators into your team. You CANNOT add collaborators after you make your submission.
Step 3 for team leaders: Make the submission and represent your team, only team leaders should make the submission (unless, its not possible for him/her).
Step 4 for everyone: Begin peer-evaluation. We will NOT send any email notification for this, please check back on Crowdgrader to find submissions to evaluate. Everyone will be randomly assigned 3 submissions to grade (5 in total, you can skip 2), and 25% of your grades depend on your duty to peer-grade others - check Crowdgrader to find and grade the submissions.
Please comment and justify why you gave this score, and point out good/bad points about the submission. For this week, look for the most interesting and insightful needs, see if you can find or infer some of it, and synthesize it with your feedback as something which can be shared with all of us. Team leaders, please make sure that every member of your team grades the submissions.
Milestone 3 Submissions
To help us track all submissions and browsing through them, once you have finished your Milestone 3, go to the link below and post the link: