The platform belongs to those who work it:
The Amara case study

by David Rozas

David Rozas. P2P Models Geek Sociologist

Co-authors: Elena Martínez Vicente, Paulo Colombo, Jorge Saldivar, Eve Zelickson

 Data Range: 2018-20

  Region: global

Methodologies

Social research

• Semi-structured Interviews
• Focus group
• Participant observation
• Documentary Analysis

 Technology

• Ethereum
• Infura
• React
• IPFS
• Aragon/Aragon Connect
• The Graph

 Design

• Lean Design
• Design Thinking
• Co-creation
• Workshops Prototype & Iteration
• User test

 Github repository

Related papers/preprints

Talk is Silver, Code is Gold? Beyond ‘ObjectCentric’ Notions of Contribution in Peer Production

Amara is a project of the Participatory Culture Foundation (PCF), a nonprofit working to build a more inclusive media ecosystem since 2006. Amara is an open and collaborative subtitling platform [1]. Examples of organizations that use the Amara platform to create subtitles through volunteer engagement include Khan Academy,  Scientific American and California Academy of Sciences [2]. As a research project, however, our focus is on the use of the Amara subtitling platform as an on-demand and paid service: Amara On Demand (AOD). AOD was launched in 2013 [3] as a result of the success of the Amara platform [1]. Over the past few years, AOD has grown from just a few linguists to more than nine hundred at the time of writing. The work of linguists in AOD is remunerated, in contrast to the volunteer subtitling supported by the Amara platform in the aforementioned case of Khan Academy. AOD is inspired by cooperative and commoning practices [4], presenting a remarkable contrast when compared with the market-based logic of other crowdsourcing platforms.

Why did we choose Amara?

We are framing AOD as part of the broader phenomenon of crowdsourcing. Nowadays collaboratively creating knowledge is the economy’s primary productive force [5,6], and digital platforms increasingly mediate contemporary working practices. On crowdsourcing platforms, work is “taskified” [7]. Entering receipts into expense reports, curating data to train artificial intelligence software, translating texts and tagging words and images are all examples of work that can be easily “taskified”. These tasks are carried out by an invisible workforce of “humans in the loop” [4]. Through platforms such as Amazon Mechanical Turk (AMT), this volatile and globally distributed crowd of workers with varying degrees of expertise and backgrounds [8] compete with each other to find tasks to work on.

Hansson et al. [9,10] explored the different modes of production mediated by crowdsourcing platforms by studying 21 different cases. Crowdsourcing platforms such as AOD and AMT were categorised as “Human Computing”, i.e. those in which “users  do micro-tasks that do not require much expertise, such as transcribing audio and video files, translating texts, or tagging maps” [9].

To further our understanding of how crowdsourcing platforms might support social relationships in these contexts, instead of merely capitalizing on them [9], we decided to study, establish a collaboration with and develop a set of experimental tools with AOD. The strong cooperative values embedded in AOD [4] offer, in this respect, an opportunity to experiment with fostering more relational crowd models, in which the producer is also the owner of the means of production and the products created are an expression of self-realization [9]. AOD contrasts with platforms such as AMT [4], which understand workers as random passive objects from which tasks and items can be sourced [9].

Our vision: “The platform belongs to those who work it”

Our ultimate goal consists of experimenting with a gradual shift in decision-making power (governance) over the rules that determine how value is distributed to the members of crowdsourcing platforms. Our vision can be summarized by the motto: “The platform belongs to those who work it.”1

Photo by Chris Montgomery on Unsplash

We aim for this research to have an impact beyond this case study and to influence a broader range of areas of the collaborative economy. Our goal is to increase worker participation and ownership of the platforms that mediate their work. Similar decentralized platforms owned by worker cooperatives, which distribute tasks and value according to their own agreements, could be envisioned for a myriad of areas. An example might be a cooperative of taxi drivers whose organization is mediated by a blockchain-based platform that they control. The platform could distribute and monitor rides and payments according to rules defined by the workers within their specific context. Ultimately, we expect to explore and contribute to the development of worker-centric crowdsourcing platforms [7] and, more generally, to that of platform cooperativism [13] within the context of blockchain-based tools that embed commoning practices [14].

What are we experimenting with?

As a research group, we are experimenting with blockchain technologies to help collaborative economy organizations work in a more decentralized way. Our aim is to explore the potentialities and limitations of blockchain technologies to experiment with alternative value distribution models and more participatory decision-making mechanisms.  In this case study, we first needed to identify the best point of intervention to do so.

1. We have adapted this motto inspired by the motto coined by Teodoro Flores in the context of the Mexican Revolution (1910-1920): “La tierra es para quien la trabaja” (The land belongs to those who work it). The motto captures the revolutionaries’ vision for land reform compiled in the Plan of Ayala [11], written by Emiliano Zapata together with other Mexican revolutionaries [12].

Stage 1. In-depth analysis of AOD

Stage 1 (October 2018 – March 2019) focused on an in-depth analysis of the inner workings of AOD from the perspective of the linguists. From the onboarding process to daily operations, this first iteration of data collection and analysis provided a rich picture of the experiences, needs and workflow vision of participating AOD linguists. The research involved online participant observation and 15 semi-structured qualitative interviews. The selection of interviewees during this stage aimed to reflect the diversity of AOD linguists in terms of language group, experience level and degree of engagement. The primary outcomes during this stage of research were the mapping of AOD workflow and the identification of an initial set of communitarian needs, which led us to identify several points of intervention as potential candidates to experiment with the development of blockchain-based tools.

Mapping the Workflow

Groups organize themselves in AOD based on language direction, i.e. which languages they can translate to or from English. For example, if a customer requires a set of videos in German to be subtitled into Spanish, this will involve the groups German → English and English → Spanish. 

Once a linguist is admitted into the community and agrees to the rules outlined in the Amara guidelines, they are added to their language team’s mailing list. Project Managers send job orders by email, which linguists use to sign up for tasks. These tasks are allocated on a “first-come, first-served” basis. Figure 1 and Table 1 provide an overview of the overall workflow in AOD, as identified and mapped by Zelickson [3] at the time2 the research was conducted (October 2018 – April 2019). While linguists do not have direct control over the creation and modification of organizational rules, they do help to enforce them when performing quality assurance on others’ work [3].

Figure 1 provide an overview of the overall workflow in AOD, as identified and mapped by Zelickson [3] at the time the research was conducted (October 2018 – April 2019).

Evaluating the points of intervention

Drawing on the analysis of this initial set of semi-structured interviews and a literature review on Amara and broader forms of crowdsourcing, we identified several possible “points of intervention” for the development of blockchain-based tools which could potentially help our case study.  Among all of them, we finally chose task allocation.

The “first-come, first-served” logic embedded in the platform is the main component of task allocation. This policy creates a competitive dynamic between workers in assigning tasks to themselves. As a result, some linguists find themselves continually checking for work, an issue identified more generally in other crowdsourcing projects [4]. Overall, this logic needs to be understood in an environment with an inconsistent workload. 

The results of this analysis led us to define task allocation as our main object of experimentation with blockchain-based tools, with the aim of exploring and implementing new ways to allocate tasks according to communitarian needs. To further our understanding of this point of intervention, we carried out a second stage of data collection and analysis.

Photo by Daria Nepriakhina on Unsplash

Photo by Daria Nepriakhina on Unsplash

2. As is usual in platform cooperatives and peer production communities, Amara’s workflows and processes are dynamic and constantly evolving.

Stage 2. Prototyping with blockchain

Stage 2 (March 2019 – September 2020) involved a new iteration of data collection and analysis focused on core members of AOD. It involved several months of participant observation, nine semi-structured interviews and documentary analysis of materials generated and posted in the official channels of AOD. This data collection concerned two primary goals. The first was to extend the data concerning organizational processes of the workflow and the changes experienced in it over time. The aim was to include all of the different perspectives of the actors involved in the platform in order to supplement those gathered during stage 1 (focused on linguists). The second goal was to enrich our understanding of the organizational characteristics surrounding the selected point of intervention: task allocation. After analyzing the data, this stage was concluded by organizing a workshop to begin the development of initial prototypes with members of AOD.

Understanding the core of AOD

AOD’s core team facilitates and oversees the whole of the production process, coordinating and sustaining the infrastructure required to achieve the successful creation of subtitles and captioning. The core team operates as a central node in AOD, although their members are globally distributed. During this phase, we carried out participant observation, and conducted semi-structured interviews with key members of AOD’s core team: project managers, developers, members of the finance team, and project leaders, amongst others. We interviewed at least one member of each area, and we decided to explore in even greater depth the vision of project managers. The reason for this was that they are the core team members with the closest contact to our point of intervention: they coordinate each project, prepare and send job orders and monitor overall progress. Having achieved an in-depth picture of the current workflow from the myriad perspectives of AOD members and a richer understanding of the factors surrounding our main point of intervention, stage 2 concluded with the organization of an online focus group to inform, develop and test the initial prototypes.

The first co-designing workshop

The first workshop focused on the identification of alternative task allocation models together with AOD linguists. From all of the groups in AOD (organised by language direction), we chose the Portuguese-Brazilian group as the target one, due to its organizational complexity. The workshop consisted of two parts. The first part involved a two-day online focus group, during which general needs and alternative models for the allocation of tasks were explored. The second part consisted of individual sessions (lasting three days) to carry out user testing with an experimental prototype3. The prototype employed decentralized technologies to embed the same “first-come, first-served” logic for task allocation currently implemented in the official platform. This prototype was aimed at helping to identify the impact and limitations of these decentralized experimental tools before implementing alternative models that emerge from the focus groups. For each part, key insights resulting from the data analysis were as follows:

DISTRIBUTION OF TASKS

We identified three alternative models. These are to be understood as ideal types, whose characteristics can be adapted and combined:

13_icon_round_robin

1. ROUND ROBIN

A model characterized by a balanced allocation of work according to the amount of work previously carried out by the linguist. Tasks are offered to linguists during a window of time for acceptance/refusal. If it is rejected, the task is then offered to another linguist.

14_icon_reputation

2. ALLOCATION BY REPUTATION

The tasks are offered to the linguist according to the quality of work done in previous tasks, based on the feedback received by their peers.

15_icon_content_preferences

3. PRE-SET ASSIGNMENTS ACCORDING TO CONTENT PREFERENCES

Tasks are pre-assigned to the linguist according to their self-reported preferences regarding the content of the videos themselves

The sovereignty for decision-making regarding which models to employ should reside within the language group. Regarding the former, for example, the Brazilian-Portuguese group may decide to work with a Round Robin model. In contrast, the Japanese group may choose to keep on using a “first-come, first-served” model.

Similarly, language groups should be autonomous in deciding the parameters that characterize the selected model. Continuing with the previous example, the Brazilian-Portuguese group should have the autonomy to determine the parameters of their Round Robin model, e.g. how long the window of time is to accept/reject a task, or what the parameters of the workload are. Several models for decision-making (e.g. one person, one vote; plural voting; quadratic voting [15]) could be employed to make collective decisions about these parameters.

Re-allocation of resources between language groups: future blockchain-based interventions could consist of a distribution of resources between different language groups federally. For example, suppose the Brazilian-Portuguese group suffers from a scarcity of work during a certain period. In that case, resources could be allocated from other groups which have had a higher workload during that time.

Specific workshops will be required to refine the parameters for each of the models, as well as to identify more models within different language groups.

User testing of the blockchain prototype

In general, linguists perceive a higher degree of complexity when using the blockchain-based prototype to assign themselves tasks than they do with the current platform. However, they feel a higher degree of ownership, and they were willing to accept this higher degree of complexity to a certain extent.

Pay for work“: The use of gas (a type of cryptocurrency) to pay for the assignment of the tasks generated uncertainty and was at times rejected. Further research should continue to explore the technical and social aspects surrounding this issue.

The use of a wallet (specifically the Metamask browser plugin) was the key factor in the aforementioned higher degree of perceived complexity. The learning curve for using a wallet is steep at first, although it becomes more comprehensible with time. Precise and clear instructions were critical to being able to carry out testing.

Ongoing research and next steps

Over the next months, we will focus our efforts on two lines of research:

Quantitavive analysis

We will carry out a longitudinal quantitative analysis of the relationship between users and their activities, drawing on the data already available on AOD’s platform. Language direction will be the unit of analysis used to map the relationship between demographic  characteristics and activities (e.g. creating subtitles, creating revisions) on a per-group basis.

Workshops

We will organize several workshops, similar to the one described in section 4.2, with different groups according to language direction. The results of our quantitative analysis will inform the selection of the specific groups. Additionally, we are planning to organize more focus groups with the Brazilian-Portuguese group, aiming to dive deeper into the parameters and characteristics of the models already identified during the co-designing workshop with this group. For each of these models, we will develop a new prototype.

More information

Related info

Previous
Next

References

  1. Jansen D, Alcala A, Guzman F. Amara: A Sustainable, Global Solution for Accessibility, Powered by Communities of Volunteers. In: Stephanidis C, Antona M, editors. Universal Access in Human-Computer Interaction Design for All and Accessibility Practice. Springer International Publishing; 2014. pp. 401–411.
  2. Video Translation and Accessibility at Scale – Amara Enterprise. [cited 27 Oct 2020]. Available: https://amara.org/
  3. Zelickson E. A Balancing Act: Hybrid Organizations in the Collaborative Economy Empirical Evidence from Amara on Demand. University of Brown. 2019.
  4. Gray M, Suri S. Ghost work: how to stop Silicon Valley from building a new global underclass. Eamon Dolan Books; 2019.
  5. Faraj S, Jarvenpaa SL, Majchrzak A. Knowledge Collaboration in Online Communities. Organization Science. 2011;22: 1224–1239.
  6. Harhoff D, Lakhani KR. Revolutionizing Innovation: Users, Communities, and Open Innovation. MIT Press; 2016.
  7. Amer-Yahia S, Basu-Roy S. Toward Worker-Centric Crowdsourcing. 2016 [cited 31 Jan 2019]. Available: http://sites.computer.org/debull/A16dec/p3.pdf
  8. Ross J, Irani L, Silberman MS, Zaldivar A, Tomlinson B. Who are the crowdworkers?: shifting demographics in mechanical turk. Proceedings of the 28th of the international conference extended abstracts on Human factors in computing systems – CHI EA ’10. New York, New York, USA: ACM Press; 2010. p. 2863.
  9. Hansson K, Ludwig T, Aitamurto T. Capitalizing Relationships: Modes of Participation in Crowdsourcing. Comput Support Coop Work. 2018 [cited 31 Jan 2019]. doi:10.1007/s10606-018-9341-1
  10. Hansson K, Aitamurto T, Ludwig T. From alienation to relation: Examining the modes of production in crowdsourcing. 2017 [cited 31 Jan 2019]. doi:10.18420/ecscw2017-13
  11. López LE. Plan de Ayala: Un siglo después. Instituto Nacional de Antropología e Historia; 2018.
  12. Mcneely JH. Origins of the Zapata Revolt in Morelos. Hisp Am Hist Rev. 1966;46: 153–169.
  13. Scholz T. Platform Cooperativism. Challenging the Corporate Sharing Economy. 2016.
  14. Rozas D, Tenorio-Fornés A, Díaz-Molina S, Hassan S. When Ostrom Meets Blockchain: Exploring the Potentials of Blockchain for Commons Governance. Sage Open. 2021;(in production). doi:10.1177/21582440211002526
  15. Cossar S, Berman P. Democracy Earth Foundation Digital Decision-Making Pilots Report. Democracy Earth; 2020 Oct. Available: https://www.dropbox.com/s/udcg0og64wlcqft/Democracy%20Earth%20Pilots%20Report.pdf?dl=0