Keywords

1 Introduction

As organizations grow larger, the layers of administrative, legal, and managerial roles tend to grow proportionately. These added layers inevitably produce increased complexity and increased risk of organizational inefficiencies. As a result, large organizations often hire external assistance to help them operate more efficiently. Public and private organizations spend billions of dollars annually assessing their performance and investing in external entities to identify and resolve organizational inefficiencies. External consulting firms are expensive, require extensive employee time, and only see a snapshot in time of the organization’s dynamics. Collaborative crowdsourcing systems can enable workers to actively and continuously participate in improving their own organization, saving costs and increasing worker empowerment.

We have developed GECKOS (Generating Employee Crowdsourced Knowledge for Organizational Solutions) to enable large organizations to leverage their own workers to continuously identify and resolve inefficiencies. GECKOS leverages the latest research findings on collaborative crowdsourcing to enable workers to anonymously contribute issues they are experiencing at work and participate in resolving them. Workers can collaborate with other workers to prioritize which issues to address first, generate ideas to address the issue(s), select highest-quality ideas that can be used to resolve the issue, and synthesize good ideas into holistic solution proposals.

2 Background: Collaborative Crowdsourcing Research

The development of GECKOS was motivated by recent innovations in collaborative crowdsourcing, which presented a unique opportunity to address the inefficiency challenges faced by large, hierarchical organizations. Our research question lies at the intersection of this problem and opportunity: How can we design a collaborative-crowdsourcing system that successfully leverages workers within an organization to enhance its efficiency?

Crowdsourcing has gained popularity as an approach to enable large numbers of contributors (often from outside an organization) to contribute to solve problems or complete tasks in a variety of domains. Research has found that a crowd can in some instances outperform domain experts, even when crowd members individually do not possess any expertise and are not particularly accurate [1]. Historically, crowdsourcing approaches have primarily focused on eliciting individual, independent inputs which are either aggregated or used to select the best individual contribution within the crowd (e.g., [2]). However, recent research has demonstrated the value of algorithms and workflows to structure and support collaboration among contributors.

Examples of workflows supporting collaboration include MicroTalk, which has been shown to enhance the quality of generated responses by structuring collaboration among contributors [3]. The MIT Collaboratorium provides a platform to support collaborative deliberation around complex challenges [4]. The Collaboratorium provides a structure to organize, think about, and select issues and ideas to address issues. Bag of Lemons is a relatively simple innovation that posits that it is easier for humans to select the worst ideas (i.e., bag of lemons) than it is to select the best ideas (i.e., bag of stars) and this can be leveraged when crowdsourcing idea selection [5]. The Mechanical Novel workflow provides support to enable crowd contributors to collaborate to complete complex, generative tasks like writing a fictional novel [6].

Another set of relevant research and innovations comes from anonymity/privacy research. In order for workers to feel comfortable contributing their honest input, they need to trust that they can contribute without fear of repercussion. AnonRep provides an approach to keep users anonymous while also holding them accountable [7]. Anonymous communication networks can also help ensure that workers trust that their contributions cannot be tracked by management. These networks include free and open-source options (e.g., torproject.org) as well as experimental options [8].

3 GECKOS – Generating Employee Crowdsourced Knowledge for Organizational Solutions

In designing GECKOS, we leveraged existing innovations and research to create a sound, functional tool to support large organizations. As we developed our initial design concepts, we received feedback from partner organizations and from our sponsor. We also interviewed workers at partner organizations to better understand how they identified and resolved issues. Our initial GECKOS concept is shown in Fig. 1.

Fig. 1.
figure 1

Main four stages of the original GECKOS design.

We incrementally developed an experimental testbed and conducted a series of experiments with real workers, real issues and ideas, real solutions, and in real organizations. These experiments were designed to provide value to the company, while concurrently providing sound experimental results to contribute to the scientific body of knowledge. It was critical that the issues (i.e., stimuli) presented to workers (i.e., participants) were real issues that workers would report if the system was implemented in their organization. We elicited over 100 issues from workers at our partner organization and populated an experimental task to mimic what participants would have generated with a real GECKOS implementation. Ensuring that the results of our experiments provided value to the leadership of our partner organization turned out to be critical to develop an increasingly efficient testbed to research collaborative crowdsourcing in organizations.

Currently, we have a complete functional prototype of GECKOS (Fig. 2). The existing system leverages many of Majchrzak and Malhotra’s insights [9], especially when it comes to visibility of other users’ input and its importance to both engagement and collaboration. The laboratory versions of these workflows proved too constraining and brittle to be feasible within GECKOS, but we incorporated features of research innovations (e.g., argumentation framework, Bag of Lemons, Mechanical Novel). Priority was given to enabling users to engage and disengage as desired while supporting high-quality collective outputs.

4 Challenges and Lessons Learned

In this section, we share some of the lessons we learned from developing and testing GECKOS. We believe our experience and lessons learned can provide valuable information to researchers, leaders, and workers interested in leveraging collaborative crowdsourcing to enhance organizational efficiency.

Fig. 2.
figure 2

Screenshots showing the current design of GECKOS. The home page is shown on the top left panel. The system includes three distinct stages: issue management (top right), idea generation (bottom left), and solution development (bottom right).

4.1 Designing a Crowdsourcing System that Can Successfully Support Collective Action by Workers to Improve Their Organization

Translating Laboratory Research Findings into Applied/Practical Design for Organizations.

When we began designing GECKOS, our priority was to support interaction among contributors and to maximize the quality of the collective output. We leveraged innovations like the MicroTalk workflow [3], which is proven to enhance the quality of the collective output when compared to not using the argumentation workflow. For example, in our original design we had a stage in which users who reported similar issues would work together in small groups (e.g., three users) to frame the issue for the rest of the organization. We tested our implementation of the MicroTalk logic to support issue framing internally with our research team and it worked well. While conceptually sound, our original approach required a level of commitment from users that made it too brittle and impractical. These requirements are too restrictive for a collaborative crowdsourcing system in which users self-select to contribute and may choose to stop contributing at any time. When that is the case, we must design for resilience. We achieved this by creating a more open, less structured interface that allowed users to dynamically see contributions from other users and add their inputs in a self-initiated way at a time of their own choosing.

Our original design also prevented users from seeing the input of other participants in an organic manner and constrained when they were able to contribute. When designing a system in which users must self-select to contribute, it is key to design for engagement. Further, being able to see other users’ input, even when one is not being asked to contribute, makes the system more attractive and engaging to users. This visibility allows workers to self-select to contribute to individual topics of interest to the user.

A more open, less structured interface also has drawbacks. For example, seeing other users’ input upfront can bias the contributions of subsequent users [10]. While workflows can help strike the balance between minimizing bias and benefiting from the responses from others, they need to be flexible enough to allow users to disengage without affecting the outcome. This can be a difficult balance to strike in collaborative-crowdsourcing systems. In one of our experiments, we explored a workflow that incorporated features of the bag-of-lemons approach [5]. Users in the experimental condition completed the workflow before accessing the workspace. The workflow provided some benefits when it comes to idea quality while not reducing engagement after they accessed the workspace in a substantial manner.

When considering self-selection situations, designing for resilience and engagement become higher priority in this context than designing for maximizing output quality. There are ways in which workflows to enhance quality can be leveraged, but they cannot be too restrictive. The workflow we found useful just involved an extra step prior to accessing the open interface.

Balancing User Anonymity with Some Users’ Desire for Recognition.

The ability to contribute anonymously is critical for a system like GECKOS. Users must be able to contribute anonymously and they must trust that this is the case. In our interviews, workers expressed skepticism about the anonymity of existing methods used by their organizations (e.g., employee surveys). While we did not implement all of the anonymity measures during this phase of the effort, we did create a system that prevented us from being able to identify contributors.

While we erred on the side of ensuring anonymity, some users expressed their desire to be identified so that their leaders could recognize them for their contributions. To the extent that inability to be recognized for contributions could negatively impact participation, this should be taken into consideration. In our current version, we intentionally were not able to track users in order to reward them or share their names with leadership. While innovations exist that could enable us to do this (e.g., AnonRep [7]), it is also important to note that tracking users to identify them for recognition is going to have a negative impact on the trust of users worried about anonymity.

Challenge of Synthesizing Ideas into a Comprehensive Solution.

Collaborative crowdsourcing is best suited to elicit information from contributors, prioritize, rank, rate, or even deliberate and argue. However, synthesizing information across ideas into a comprehensive solution is a challenging task to complete through a collaborative crowdsourcing process. A workflow like Mechanical Novel involves too many steps and iterations with temporal constraints to make it practical in organizational settings. In this effort, we did explore a couple of alternatives, including adding a step which is completed by leadership or smaller teams of organization workers who would create a set of proposals to present to the crowd to vote or prioritize. This is better than most existing civic-engagement processes, which typically gather ideas from contributors who do not then get a voice on selecting among proposed solutions. More research is needed in the development of lean workflows to support solution development.

4.2 Conducting Research with, and Providing Value to, Our Partner Organizations and Their Workers

Establishing Partnerships and Developing an Experimental Testbed.

One of the accomplishments of this effort was to establish an experimental testbed to conduct collaborative-crowdsourcing research with a partner organization. Majchrzak and Malhotra [9] described how challenging and time-consuming it is to create and maintain an experimental testbed for this type of research. We established partnerships with multiple separate organizations and were able to a successful experimental testbed with the smaller of the three (i.e., less than less than 1000 workers). And its size likely played a role in the success of this partnership. In the other two cases, a health-care system with over 10,000 workers and a Fortune 100 company, the same inefficient processes which this effort is trying to address (e.g., legal departments with rigid criteria that do not accommodate for unusual or innovative agreements) likely played a role in the partnership not reaching the experimental testbed stage.

Developing and nurturing this partnership involved a gradual process of gaining buy-in from leadership and workers at the partner organization. Buy-in was hardest at first, without a fully developed system. As we experimented using GECKOS to prioritize issues encountered by their employees, generate ideas to resolve top-priority issues, and so on, the value of the system became apparent to leadership and the level of support and enthusiasm also grew. This was very insightful in terms of how to develop a fruitful collaboration and set up an experimental testbed within a limited amount of time. Currently, we have a process that allow us to configure and conduct experiments with over 100 workers at this partner organization within a few weeks. This capability is invaluable when you are developing and testing crowdsourcing tools.

Research Limitations.

There are a few challenges that relate specifically to conducting research with workers within organizations. First, follow-up experiments to explore variations of the same experiment and answer questions about what the results mean cannot be done with the same issues and/or ideas as in the first experiment. It would need to be at least a separate issue or set of issues. Otherwise, participants would see redundancy in the task content and would likely not remain as genuinely engaged. For example, if we conducted experiments while workers were generating ideas to, let us say, improve the communication strategy within the company, and we wanted to test a follow-up hypothesis with a modified workflow, we would not be able to use a modified workflow on the same issue. While the value obtained from the stimuli being realistic and motivating to employees is worth the sacrifice, this limitation makes answering questions definitely more challenging with an experimental testbed like this one. Ideally, this challenge could be addressed by having a large number of partner organizations as part of the testbed.

Another important challenge is assessing the quality of the collective output. We were interested in understanding which approaches result in better collective outputs, but establishing output quality is not straightforward. For example, participants may generate ideas to resolve an issue they are experiencing at work or prioritize issues reported by their peers for resolution. What is the ground truth to which to compare the collective output? We addressed this problem by using a new sample from the same population of workers (i.e., a panel of judges) to rank the collective outputs obtained from both methods based on perceived quality (e.g., idea quality, prioritization quality). The results showed some promise. For example, panel of judges used in the pre-tests we conducted with Amazon Mechanical Turk® workers and the experiments with workers seemed to show consistent results. In one experiment, we created a ‘ground truth’ for issue priority based on the frequency with which users reported the issue. The results from the panel of judges were also consistent with this ‘ground truth’. The panel-of-judges approach provides some value, but additional measures of collective-output quality useable with real-world stimuli are needed.

5 The Future of Collaborative Crowdsourcing in the Workplace

The work described in this paper highlights both the inherent challenge of developing a successful collaborative-crowdsourcing system for large organizations and its potential to democratize management/leadership and support more efficient and adaptive organizations. We present the implications of this work to researchers/technologists, leaders/managers, and workers at organizations separately.

Implications for Researchers and Technologists.

Conducting research on collaborative-crowdsourcing systems using the real issues and ideas of real workers of real organizations poses challenges, limits experimental control, and constrains design. However, establishing a partnership with mid-size organizations that have more flexibility than extremely large organizations can enable researchers and technologists establish a testbed to use for experimentation. The nature of the stimuli used and the experiments can be such that the organization’s leadership can see the value of the partnership even before the promised system is provided to them. Once workers are engaged and issues and ideas that they recognize are important are presented, it is much easier to efficiently engage both leadership and workers.

Implications for Leaders and Managers.

The success of collaborative-crowdsourcing systems within organizations is not merely a technological issue. Leaders and managers within organizations play a key role in enabling these systems to achieve their potential or preventing them from doing so. In the book Team of Teams, General McChrystal discusses what a more distributed, decentralized, adaptive leadership might look like [11]. In order to truly grasp the impact that these systems could have when it comes to managing/leading large organizations, leaders and managers will have to be willing to give up some decision-making/problem-solving control and trust in the collective wisdom of their workers. Without this mindset in leadership/management, the impact will be dampened and the potential of these systems will not be fully realized.

Implications for Workers and Worker Trust.

The way things are typically done in organizations and systemic/cultural issues (e.g., individualistic society) also present challenges regarding how individual users contribute to the system. For example, we need to ensure that workers can trust that their input is anonymous and that it is safe to contribute. Mainstream management practices that localize the decision-making authority and responsibility in a few hands engender mistrust and limit the ability of workers to truly participate in important decisions. And workers are used to operating in this environment. Just like placing a fishbowl fish in a bathtub does not mean that the fish is going to swim through the whole bathtub, the virtual walls of the ways in which workers are used (and expected) to engage within organizations can limit the impact of collaborative-crowdsourcing systems. The success of these systems also relies on workers giving them a genuine try, donning a collectivistic mindset, and working to advance the collective wellbeing and efficiency of their organization.

Ultimately, the goal of GECKOS is to democratize organizational management in a way that serves both workers and leadership. New technological advances and scientific discoveries can support more inclusive, participative, democratic, and cooperative organizations. Collaborative-crowdsourcing systems like GECKOS can help realize this more inclusive, participative, democratic, and cooperative future.