Value Sensitive Design and power in socio-technical ecosystems

Mattis Jacobs, Ethics in Information Technologies, Universität Hamburg, Germany, mattis.jacobs@uni-hamburg.de
Christian Kurtz, IT Management and Consulting, Universität Hamburg, Germany, christian.kurtz@uni-hamburg.de
Judith Simon, Department of Informatics, Universität Hamburg, Germany, simon@informatik.uni-hamburg.de
Tilo Böhmann, Department of Informatics, Universität Hamburg, Germany, tilo.boehmann@uni-hamburg.de

PUBLISHED ON: 30 Sep 2021 DOI: 10.14763/2021.3.1580

Abstract

Recent European policy papers call for the consideration of human values in the design of information technology. Value Sensitive Design (VSD) provides a framework for systematically accounting for values in the design of technical artefacts. This paper examines how the distribution of power within socio-technical ecosystems poses a challenge for the application of VSD. It identifies four crucial factors determining the effect of the distribution of power on VSD: the level of decentralisation of the ecosystem; if VSD is applied at the core or periphery; when power can be exercised (temporality); and the phase of VSD (conceptual, empirical, and technical) that power can be exercised in. Based on these factors, it outlines how the challenge of accounting for power can be addressed.
Citation & publishing information
Received: November 12, 2020 Reviewed: April 1, 2021 Published: September 30, 2021
Licence: Creative Commons Attribution 3.0 Germany
Funding: Mattis Jacobs, Christian Kurtz, Judith Simon, and Tilo Böhmann acknowledge financial support from the Hamburg Ministry of Science, Research and Equality in the project Information Governance Technologies—Ethics, Policies, Architectures, Engineering—with reference LFF-FV 34. Mattis Jacobs and Judith Simon acknowledge financial support from the German Federal Ministry of Education and Research in the project GOAL—about algorithmic behaviour control and artificial intelligence—with reference 01IS19020.
Competing interests: The author has declared that no competing interests exist that have influenced the text.
Keywords: Value sensitive design, Power, Sociotechnical ecosystems, Platforms, Blockchain
Citation: Jacobs, M. & Kurtz, C. & Simon, J. & Böhmann, T. (2021). Value Sensitive Design and power in socio-technical ecosystems. Internet Policy Review, 10(3). https://doi.org/10.14763/2021.3.1580

This paper is part of Governing “European values” inside data flows, a special issue of Internet Policy Review guest-edited by Kristina Irion, Mira Burri, Ans Kolk, Stefania Milan.

1. Introduction

Recent European policy papers call for the consideration of human values 1 in the design of information technology (European Commission, 2020a, 2019; HLEG-AI, 2019; Datenethikkommission, 2019). Approaches such as Value Sensitive Design (VSD) promote the idea that human values can be accounted for in the development of technological artefacts and provide a framework for systematically analysing, weighing, and operationalising them (Friedman et al., 2008; Friedman & Hendry, 2019). However, while VSD is well received in the academic context and attracts attention from various disciplines such as computer science and information systems (Friedman et al., 2008; Friedman & Hendry, 2019; Winkler & Spiekermann, 2018; Mueller & Heger, 2018), computer ethics (Brey, 2010; Introna, 2005), healthcare (Walton & DeRenzi, 2009), urban design (Borning et al., 2008; Waddell et al., 2008) and others, there are challenges barring the path to widespread adoption.

Reflecting on such “grand challenges”, Friedman and Hendry (2019) name “accounting for power” as one of them. 2 In this paper, the term “power” refers to “the ability of agents […] to realize a certain outcome” (Brey, 2008, p. 75) 3—specifically design decisions—“even against resistance” (Weber, 2019, p. 134). Friedman and Hendry (2019) elaborate on the challenge: VSD “has not yet explicitly addressed how to handle differences in power among […] stakeholders [and how] best to account for power relations within a value sensitive design framing remains an open question.”

This paper argues that this challenge is exacerbated in many cases by the integration of technical artefacts in increasingly vast and complex socio-technical ecosystems defined as “a dynamic community of competing and interdependent people, organizations, and computing systems operating in a complex, capricious environment” (McConahy et al., 2012, p. 1). In such socio-technical ecosystems, the power over the variety of independent design decisions which, in their totality, define the shape of an artefact is often distributed over various actors. Furthermore, due to their socio-technical nature, containing a “technical, social, political, and economic” (van House, 2004, p. 18) as well as “organizational […] and business” (Feiler et al., 2006, p. 27) domain, power can manifest in various forms in these ecosystems. 4 In order to account for all the domains in which power can manifest itself, the paper adopts a systemic view on power, which “regards power as the property of broader social, economic, cultural, and political networks, institutions, and structures” (Sattarov, 2019, p. 20) and focuses on how “systems confer differentials of dispositional power on agents, thus structuring possibilities for action” (Haugaard, 2010, p. 425; see also Sattarov, 2019).

The remainder of this paper explores the following questions: 1) how and to what extent does the “grand challenge” of accounting for power in VSD get exacerbated by the integration of technical artefacts in increasingly vast and complex socio-technical ecosystems; 2) how does the organisational structure of socio-technical ecosystems affect the challenge of accounting for power in VSD; 3) how can this challenge be addressed; and 4) are there positive effects for VSD if the approach is applied in settings in which the power to make design decisions is distributed over various actors.

Section ‎2 provides an overview of VSD. Section ‎3 further elaborates on the actors involved in developing technical artefacts in different types of socio-technical ecosystems and their respective leverage over how developers can account for specific values. It discusses two exemplary types of socio-technical ecosystems that differ in the degree of decentralisation and, thus, the way power is distributed within them: platform ecosystems such as Apple’s iOS ecosystem (section ‎3.1) and blockchain-based systems such as Bitcoin and Ethereum (section ‎3.2). Section ‎4 outlines how adopting a power-sensitive ecosystem perspective can foster a more pronounced understanding of the challenge of accounting for power. Based on these observations, section ‎5 derives some reference points for addressing the challenge of accounting for power in socio-technical ecosystems. Lastly, section ‎6 concludes.

2. The Value Sensitive Design approach

According to Friedman and Hendry (2019, p. 3), VSD “seeks to guide the shape of being with technology”. Directed at “researchers, designers, engineers, policy makers, and anyone working at the intersection of technology and society […], it provides theory, method, and practice to account for human values in a principled systematic manner throughout the technical design process”. To account for values in the design process of technical artefacts, VSD is structured in three phases of action: conceptual, empirical, and technical investigations. Pertinent literature maps out the respective phases primarily by determining what practitioners should aim to achieve in them. The approach deliberately refrains from prescribing specific methods in the individual phases, allowing VSD practitioners to select and integrate methods tailored to the respective context of application on a case-by-case basis (Friedman et al., 2008; Friedman & Hendry, 2019).

Conceptual investigations “comprise analytic, theoretical, or philosophically informed explorations of the central issues and constructs under investigation” (Friedman & Hendry, 2019, p. 12). They address issues such as the identification of direct and indirect stakeholders, the nature of the respective stakeholder’s implication, the conceptualisation of values, and dealing with value conflicts (Friedman et al., 2008). Regarding value conflicts, it is important to note that such conflicts can also exist between human and instrumental values (Friedman & Hendry, 2019) and between values that are directly affected and values whose preservation is being put at risk only in the future (see Czeskis et al., 2010). In a more recent publication, Friedman and Hendry (2019) also include developing a framework for evaluating a successful application of VSD into this phase.

Empirical investigations employ quantitative and qualitative social sciences methods to determine the stance of (groups of) stakeholders towards values and their respective weighing (Simon, 2016). Additionally, practitioners can deploy empirical methods in a later stage to “evaluate the success of a particular design” with regards to whether it supports the realisation of a particular value as intended (Friedman et al., 2008, p. 72). Empirical investigations of this second form aim to answer whether the objectives defined in the conceptual investigation have been achieved. In both early- and late-stage empirical investigations, developers apply advanced survey and interview methods to disclose discrepancies between the espoused practice of stakeholders with their actual practice (Friedman et al., 2008). Thus, empirical investigations provide “a more situated understanding of the socio-technical system” in question and facilitate “the observation of stakeholders’ usage and appropriation patterns, but also whether the values envisioned in the design process are fulfilled, amended, or subverted” (Simon et al., 2020, p. 4).

Technical investigations also take two different forms in VSD. The first form comprises a retrospective analysis of existing technological artefacts and aims at disclosing “underlying mechanisms [that] support or hinder human values” (Friedman et al., 2008, p. 73). This form thus corresponds roughly to what Brey (2010) and Introna (2005) refer to as “Disclosive Computer Ethics”. The second form of technical investigations “involve[s] the proactive design of systems to support values identified in the conceptual investigation” (Friedman et al., 2008, p. 73). Practitioners here address how the respective conceptualisation and weighing of values can be operationalised and accounted for in the design process, i.e., how they can be translated into code.

The three phases of Value Sensitive Design repeat iteratively. Neither a starting point nor an order is prescribed. The respective phases are intended “to inform and shape and reshape each other” through the iterations (Friedman & Hendry, 2019, p. 35).

Because VSD does not prescribe the use of specific methods in the respective phases, recent overview articles and literature reviews (Friedman et al., 2017; Friedman & Hendry, 2019; Winkler & Spiekermann, 2018) provide practitioners of VSD with heuristics on how to proceed by invoking exemplary case studies. They instance methods such as stakeholder analyses (Friedman et al., 2006), value scenarios (Nathan et al., 2007), ethnographically informed inquiries (Nathan, 2012), multi-lifespan timelines (Yoo et al., 2016), and others. Additionally, pertinent literature provides further heuristics such as lists “of human values with ethical import that are often implicated in system design” as a tangible basis for practical application (Simon et al., 2020, p. 4; see also Friedman et al., 2008).

3. Accounting for power in socio-technical ecosystems

In contrast to the development of independently working, stand-alone, monolithic technical artefacts, the development of artefacts integrated into socio-technical ecosystems have a much more constrained scope for design. To enable coordination among providers of components of a socio-technical ecosystem, the component’s design must account for the existing technical and non-technical features of the ecosystem. Thus, the actors (co-)determining these features set restrictions to design decisions for novel components of an ecosystem (McConahy et al., 2012), including attempts to account for human values.

The following sections showcase two types of socio-technical ecosystems, characterised by different actor constellations, to reveal the specific manifestation of the challenges for applying VSD in the respective contexts. Section ‎3.1 focuses on platform-based ecosystems such as Apple’s iOS, Google’s Android, and the Facebook ecosystem, whereas section ‎3.2 focuses on blockchain-based ecosystems. The cases differ in how power is distributed in the respective types of ecosystems, how it manifests, how visible its distribution is, and the available modes of governance to address power-related issues. The dissimilarity of the cases allows to take a wide-angle perspective and to identify various facets of the challenge of accounting for power in VSD.

3.1 Platform-based ecosystems

The majority of digital interactions occur in ecosystems, often facilitated and connected over a digital platform (Lusch & Nambisan, 2015). The term “platform” is often a source of confusion due to a variety of definitions. In this paper, the term refers to a software-based system as the core with an extensible codebase that enables functionality for users through additional software subsystems in the form of peripheral applications—or modules—that interoperate with it (Baldwin & Woodard, 2009; Reuver et al., 2018; Tiwana et al., 2010). Peripheral services are provided by developers to enable the provision of functionality, service, or content (Constantinides et al., 2018), which can be accessed by the user via the platform (Ghazawneh & Henfridsson, 2013).

Previous research has already addressed some challenges for applying VSD (or accounting for human values more generally) in platform-based ecosystems (Shilton & Greene, 2016; Shilton & Greene, 2019; van Dijck et al., 2018; Warnier et al., 2015). For instance, focusing on deliberations of mobile application developers in developer forums on how to account for values, especially privacy, Shilton and Greene (2016) demonstrate the extent to which platform providers can assert their ideas without actively engaging in design processes. Shilton conceptualises the platform provider’s means to do so as “value levers” (Shilton, 2012) and, together with Greene, provides comparative studies on the use of these levers in different platform ecosystems (Shilton & Greene, 2019). However, a more elaborate ecosystem perspective—as developed in information systems research—could provide a means to refine such an analysis.

According to information systems literature, the platform provider’s central role in platform-based ecosystems is a facilitating one (van Alstyne et al., 2016). To scale a platform, the platform provider needs to attract external actors into the platform ecosystem that engage in interactions. In platform-based ecosystems, mainly actors of four different groups come together: users, platform providers, app providers, and third parties. Platforms enable external developers to contribute to the ecosystem by providing so-called boundary resources (Constantinides et al., 2018; Tiwana & Konsynski, 2010). Boundary resources are socio-technical manifestations of the platform provider’s power to influence a platform ecosystem (Ghazawneh & Henfridsson, 2013), such as application programming interfaces (API), software development kits (SDK), legal guidelines, and application approval processes (Eaton et al., 2015; Karhu et al., 2018).

As control points for a platform provider, boundary resources facilitate an arm’s length relationship between the platform provider and service providers (Ghazawneh & Henfridsson, 2013). They offer the providers of peripheral applications access to a platform’s resources while allowing the platform provider to retain influence over the platform (Eaton et al., 2015). Using boundary resources, a platform provider orchestrates its platform ecosystems and enables service providers to participate in and contribute to the platform’s development (Eaton et al., 2015). Designing and implementing boundary resources is a balancing act of retaining power while supporting service providers to create independent platform-based innovation (Eaton et al., 2015). Thus, platform providers hold the privileged position to exercise power by determining the design of boundary resources and thereby influence the actions of service providers and third parties involved with the platform (Eaton et al., 2015; Ghazawneh & Henfridsson, 2013), with direct implications for how these actors can account for values.

This indicates that platform providers can serve a decisive role in encouraging (or discouraging) design decisions that support the realisation of human values. For instance, in the redesign of its boundary resources via iOS 13, Apple introduced fine-grained user configuration options regarding the usage of location data by apps (Apple, 2019). In previous iOS versions, users could choose among the three options ‘Never’, ‘While Using the App’, and ‘Always’ (Apple, 2019). iOS 13 introduced the additional option ‘Ask Next Time’. Users and the developers of applications in Apple’s ecosystem are directly affected by such decisions. The configuration options have a considerable impact on user information privacy since the user can make case-by-case decisions on whether or not to grant access to their location data to an app. App developers, on the other hand, have to consider these case-by-case decisions in the expected user behaviour. Platform providers mostly prescribe such changes in boundary resource design unilaterally. Users and developers of peripheral applications are often regarded as passive recipients of these changes. Although the developers of peripheral applications can generate some degree of pressure through public criticism (Hestres, 2013) or building coalitions to achieve their goals (Perez, 2020), the decision-making power lies with the platform provider, in this case, Apple.

In another instance, Apple changed the data interface design for apps to access the MAC (Media Access Control) addresses of the devices an iPhone is connected to. Various applications misused this interface to bypass restrictions on location data access. They approximated the location data by using these MAC addresses in combination with publicly available databases that offered the specific locations of the devices that hold the respective MAC addresses. With the update to iOS 11 in 2017, access to these network data was disabled (Butts, 2017). However, the interface design also blocked data access for app providers that offer network services. As a consequence, these apps were no longer functioning. Thus, Apple restricted the scope for design of app providers, ruling out an operationalisation of privacy that would maintain the existing functionality.

Lastly, digital service providers and related services may also be influenced by the necessity of maintaining and keeping up with platform updates by the platform provider, such as APIs or framework refinements (Ausloos & Veale, 2020). For instance, in OS 14, Apple established the framework App Tracking Transparency, playing out privacy features more prominently than in earlier versions of iOS. Due to these changes, app developers have to request user authorisation to access app-related data for tracking the user or the device (Apple, 2020). In the future, Apple intends to ban applications that track users without permission and thus violate the new requirements and respective guidelines (Leswing, 2021). In consequence, Apple’s decision to establish third-party transparency has a significant influence on how developers of peripheral applications can conceptualise and operationalise data protection and information privacy.

3.2 Blockchain-based ecosystems

A blockchain is a distributed, encrypted, chronological database of transactions recorded by a distributed network of computers (Morabito, 2017; Wright & De Filippi, 2015). It contains “every transaction that has been carried out and shared among those participating in the network” (Morabito, 2017, p. 4). The entries are “encrypted and organized” in “smaller datasets referred to as ‘blocks,’” each of which references “to the preceding block in the blockchain” (Wright & De Filippi, 2015, p. 7). A consensus mechanism warrants the integrity of each transaction over the network. Contrary to other approaches in computer security, in open, permissionless blockchains, the consensus mechanisms are not based on access control, i.e., on “carefully vetting participants and excluding bad actors” (Antonopoulos, 2014). Instead, they rely on economic incentive systems that aim at motivating actors—referred to as miners (Alsindi & Lotti, 2021)—to participate in the validation process and ensuring that it is “more profitable and attractive [for them] to contribute to the network than to attack it” (Brekke & Alsindi, 2021, p. 2). As a result of this approach, “the key characteristics of a blockchain […] are that it is: distributed, decentralized, public or transparent, time-stamped, persistent, and verifiable.” (DuPont & Maurer, 2015, p. 2). Moreover, the blockchain technology is not restricted to the record-keeping function it utilises in its origin in cryptocurrencies. More recently developed blockchain-based systems such as Ethereum incorporate Turing-complete virtual machines that allow executing not only simple transactions but also more complex operating steps. In turn, this enables running decentralised second-layer applications (DApps) as services on top of the system.

In line with discourses around earlier decentralised technical systems such as the internet (Bodó et al., 2021), developers, scholars, and the broader community discussed blockchain technology in value-related terms from the very beginning. In this discourse, a libertarian reading of the technology is dominant (De Filippi, 2017; Werbach, 2018; Wright & De Filippi, 2015). Furthermore, trustworthiness (Becker & Bodó, 2021; Hawlitschek et al., 2018; Jacobs, 2021; Werbach, 2018) and sustainability (Alsindi & Lotti, 2021; Giungato et al., 2017) are discussed prominently as values that should be accounted for in the technology’s technical design. Notably, design decisions regarding comparably subtle changes in a system’s protocol—such as a change in the number of transactions aggregated in one block—are debated by the community in terms of values embodied in the respective design decision (Werbach, 2018).

The technical properties of blockchain-based systems affect the applicability of VSD. In contrast to the platform-based ecosystems discussed in the previous section, there is no central entity controlling the system that can unilaterally determine the design of technical interfaces (Antonopoulos, 2017). Thus, consent between several (groups of) actors is necessary to implement protocol amendments successfully. These are, first and foremost, the developers themselves, but also a significant share of miners, cryptocurrency exchanges, and token holders.

The software protocols of open and permissionless blockchains are maintained as open source projects, i.e., the code is publicly available, everyone can propose or recommend code changes and amendments, and a “mix of volunteer and paid software developers write and update the software” (Walch, 2019, pp. 60–61). However, while there are no explicit boundaries to participating in the design process in many blockchain-based systems, there are still groups of core developers with additional rights that guide and oversee the design processes in most larger systems (Walch, 2019). Thus, as Werbach (2018, p. 104) notes, “developers have more power than they let on. […] And even in an open-source project, a single individual can exercise significant authority.”

However, core developers only propose updates. Ultimately, the actors running the network need to “adopt and run [these] implementation[s]” (Antonopoulos, 2017, p. 259). Since the developers do not operate the system, they only create a new version of the system’s protocol in a software repository, i.e., they create a “software fork”. The respective nodes, miners, and wallet-holders individually decide whether or not they use client software with the updated version of the protocol, i.e., create a “network fork” (Antonopoulos, 2017).

As many protocol upgrades lead to consensus rules that are not “forward compatible” (Antonopoulos, 2017; see also Swan, 2015), i.e., they are incompatible with the pre-upgrade version’s ones, miners continuing to proceed according to the old rules and miners proceeding according to the new rules from this point on participate in diverging ledgers. In such a hard fork, “two chains evolve independently” from one another (Antonopoulos, 2017, p. 257). If, in the long run, either all or no miners follow the core-developers advice to adhere to the new version of the protocol, one respective branch of the fork perishes. If a sufficiently large group of miners adheres to either version of the protocol, the network splits, with both ledgers persisting. These share the same history but are henceforth dissociated from one another (Antonopoulos, 2017; for a discussion of several cases see DuPont, 2019).

Furthermore, cryptocurrency exchanges, too, need to adopt the new rules for them to be successfully introduced (Antonopoulos, 2017). While exchanges do not directly engage in maintaining or running blockchain-based systems and exist merely at their fringes, they nevertheless impact the incentives that drive the more central actors. For instance, cryptocurrency exchanges need to decide which ledger they list, i.e., for which of the ledgers they offer exchanges to customers and thus provide easy access to the system as a whole. While a delisting on one exchange platform potentially only has negligible effects, a coordinated effort to delist a system by several major platforms can hinder access to the system and diminish the economic incentives to participate in it (Orcutt, 2019).

The actors involved in blockchain-based systems thus do not only potentially hold different values or have diverging preferences and incentives regarding the conceptualisation or operationalisation of values but wield sufficient power to unilaterally intervene in design decisions that concern the realisation of a specific value. Moreover, various recent examples showcase that this is not a mere theoretical possibility, but that diverging preferences regarding values in practice do entice these actors to make use of these means.

Regarding core developers, one of the most prominent instances took place in the aftermath of a hack—commonly referred to as “TheDAO hack”—in which an attacker was able to gain hold of assets worth around “$55 million at the time” (De Filippi & Wright, 2018, p. 141) and siphon them to a fund under its control. However, before the attacker was able to move the assets further or sell them on a cryptocurrency exchange (Botsman, 2017), the core developers of Ethereum pushed through a code update to ultimately void the illicit transactions and “recover the funds from the attackers” (De Filippi & Wright, 2018, p. 141). Because this update affected not just the general principles and functionality of the technology but also individual transactions, commentators commonly use the example of this update to illustrate the power of core developers in the system (De Filippi & Wright, 2018; Walch, 2019; Werbach, 2018). By voiding the transactions, the core developers made a value judgment in that they favoured the restoration of trust in the community over the ledger’s integrity, understood as the immutable nature of ledger entries. This is because the reversal of the transactions “meant that Ethereum transactions were not truly immune from centralized interference” (Werbach, 2018, p. 68). 5

Peculiarly, the case of “TheDAO” hack also serves as a curious case highlighting the role that significant miners take in the process of incorporating code amendments, as they independently decide whether or not to follow the core developers’ advice to update their client software. In the case of the TheDAO hack, the community split (DuPont, 2018, 2019). While most miners followed the core developers’ advice, a minority stuck to the old protocol and thereby created an incompatible version of the shared ledger called “Ethereum Classic” (Werbach, 2018). This example demonstrates that only by convincing large proportions of significant miners to adopt the updated implementation developers can turn a software fork into a network fork (Antonopoulos, 2017). Therefore, as a stakeholder group, miners cannot be overruled or circumvented in design decisions regarding protocol changes. 6

Lastly, the coordinated approach of various cryptocurrency exchanges to delist the Bitcoin-Cash spin-off Bitcoin SV highlights the capability of cryptocurrency exchanges to engage in the negotiation process on how a blockchain-based system should account for human values. Here, two groups of stakeholders proposed different code upgrades for the Bitcoin Cash protocol. One group, surrounding “the developers of the most popular Bitcoin Cash software client, called Bitcoin ABC, proposed a series of upgrades, including smart contract capability.” In contrast, another group, including a mining pool controlling “more than 15 percent of all Bitcoin Cash mining,” proposed a divergent upgrade without such fundamental changes to the system’s capabilities (Orcutt, 2018). The advocates of this alternative upgrade claimed that it adheres more closely to the original ideals of Bitcoin as outlined in early white papers.

Consequently, a hard fork occurred, establishing Bitcoin SV as a spin-off of Bitcoin Cash, followed by turmoil within the community and what Orcutt (2019) calls “social media-fueled coin delistings”. Major cryptocurrency exchanges like Kraken or Binance released statements criticising the team behind the newly established Bitcoin SV. KRAKENFX (2019) announced that the behaviour of “the team behind Bitcoin SV” in the aftermath of the fork was incongruent with the values held by “Kraken and the wider crypto community”. Binance (2019) questioned whether Bitcoin SV “continues to meet the high level of standard” they expect. Consequently, the two exchanges—among others—stopped exchanging Bitcoin SV on their platforms, which, in turn, lead to “a substantial drop in [Bitcoin SV’s] value” (Orcutt, 2019), restricted access to the system for its users, and diminished economic incentives to participate in the system for Bitcoin SV miners.

4. Findings

Sections ‎‎3.1 and ‎3.2 offer several insights into the challenge of accounting for power in VSD. These allow identifying four crucial factors that co-determine the effects of the distribution of power in socio-technical ecosystems on the applicability of VSD.

Level of decentralisation

Juxtaposing platform-based ecosystems and blockchain-based ecosystems suggests that considering the level of decentralisation of the ecosystem is of paramount importance for determining the kind of issues that might occur when accounting for power in the application of VSD. As Shilton and Greene (2019) demonstrate, actors like platform providers at the centre of more centralised ecosystems can assert their ideas of conceptions, weighings, and operationalisations of values to a large extent. Using boundary resources as value levers, they do not just enforce these conceptualisations, weighings, and operationalisations onto the core components of the ecosystem, which they directly control, but also onto the design of peripheral applications.

Conversely, in organisationally more decentralised ecosystems, there is, by definition, no central actor who can similarly assert itself. As shown in the case of open and permissionless blockchains, many actors have the power to impact decisions in the context of how human values are conceptualised, weighed, and operationalised in the design processes. The distribution of power in such ecosystems makes some form of deliberation and coordination inevitable to avoid gridlocks.

Core/periphery

Different issues arise in design decisions concerning an ecosystem’s core components and design decisions concerning peripheral applications. The design of the core components, for the most part, affects more stakeholder groups than the design of peripheral services. Accordingly, negotiation processes are often more complex and conflictual.

Conversely, on the side of peripheral applications, fewer actors are involved. Instead, VSD practitioners have to consider boundary resources that constrain their scope for design. They need to account for the ecosystem’s technical and non-technical infrastructure and the actors in charge of it. Section ‎‎3.1 underlines the resulting power imbalance in platform-based ecosystems. Platform providers determine the design of boundary resources largely independently and, as a result, determine the scope for design of developers of peripheral services.

Temporality

As demonstrated in the two cases, how actors can exercise power varies considerably. One key difference is temporality. Some means of exercising power function ex-ante, i.e., actors suppress potential design decisions from being implemented in the first place. Examples of ex-ante exercises of power are the design of technical interfaces that predetermine how data can be accessed and managed, how users and peripheral services can interact, and more generally, which criteria peripheral services have to meet in order to be compatible with an ecosystem’s technical infrastructure. By utilising technical interfaces, central actors can predetermine how human values like privacy can be conceptualised and operationalised in the entire ecosystem. Other modes of exercising power function ex-post, i.e., they interfere with a technical artefact’s deployment or usage after an undesired design decision is implemented. Examples of means to exercise power ex-post are, for instance, app-store approval processes that central actors can use to exclude specific applications or services from a platform, or the decision of significant miners in blockchain-based ecosystems to omit using a new version of a blockchain protocol after developers deployed it in a software repository.

Phase of VSD

The two cases demonstrate that accounting for power is relevant for making decisions in the conceptual, empirical, and technical phase of VSD. They reveal that the way power is situated in the “broader social, economic, cultural, and political networks, institutions, and structures” (Sattarov, 2019, p. 20) of an ecosystem affects how human values are conceptualised (as demonstrated in the analysis of iOS and Android developer forums by Shilton and Greene (2019)), weighed (as demonstrated by Ethereum’s core developers’ value-judgement leading to recovering the funds after TheDAO hack), operationalised (as demonstrated by Apple’s move to change the data interface design for apps to access the MAC addresses of devices), and how the overall process of accounting for values can be evaluated (as illustrated by “social media-fueled coin delistings”). Accounting for power thus concerns VSD practitioners in all phases of VSD.

5. Discussion

These observations allow deriving some points of reference for addressing the challenge of accounting for power. They suggest that the general applicability of VSD and its potential to address power-related issues varies tremendously depending on features of the ecosystem and the role that a given artefact is supposed to play in it. While the distribution of power within an ecosystem, in some cases, hinders the application of VSD, it appears to accommodate the approach in other cases. Though this is true for both more centralised and more decentralised systems, the decisive factors differ.

In more centralised ecosystems like platform-based ecosystems, boundary resources function as obligatory passage points (Law & Callon, 1992; see also Callon, 1984) for peripheral applications and constrain the application’s design process. These constraints cut back on the agency of developers and can either obstruct or compel design decisions that promote or demote the realisation of specific values. 7 Shilton and Greene (2019) outline discussions from iOS and Android developer forums that illustrate this lack of agency of developers of peripheral applications in platform-based ecosystems. Here, developers are often required to interpret and realise a concept of privacy that is predefined and manifests, e.g., in the platform’s boundary resource design. As Greene and Shilton (2017, p. 16) note, “platforms govern design by promoting particular ways of doing privacy, training devs on those practices, and (to varying degrees) rewarding or punishing them based on their performance”. Thus, while developers of peripheral applications can always decide to collect less data, privacy here, for the most part, is whatever the platform providers define as privacy. A meaningful application of VSD is virtually non-viable for developers of peripheral applications in such settings.

While Hestres (2013) and van Dijck et al. (2018) outline how concerted efforts of various stakeholder groups can principally have success in appealing to platform providers and lead to changes in boundary resource design, this commonly is not part of the design process of individual technical artefacts and thus outside the scope of VSD. Nevertheless, grassroots efforts proved effective in many cases and should be considered a tool to create the necessary environment for an application of VSD by VSD practitioners in more centralised socio-technical ecosystems. 8

However, if Apple’s boundary resource design is assessed regarding its direct impact on the realisation of values, it’s apparent that it gives users more autonomy by allowing them to configure the location data usage (Apple, 2019) and to protect their privacy by eradicating access to a device’s MAC address (Butts, 2017). Thus, this case shows that powerful actors such as platform providers can also encourage “ethical practice within their ecosystems” (Shilton & Greene, 2019, p. 144) by making use of a carefully considered boundary resource design. Regarding privacy, Shilton and Greene (2016, n.p.) describe this phenomenon as “a ‘trickledown privacy’ effect in which platform providers exercise strong power over privacy definitions”. As platform providers can shape the conceptualisation of values within the ecosystem more generally, similar effects can be realised with other values (Shilton & Greene, 2019). 9Therefore, if VSD is used in boundary resource design, it enables the value sensitive shaping of ecosystems.

While the means by which the platform providers exert power in the discussed examples are primarily technical on the surface, they also affect developers of peripheral applications economically. For instance, as Ausloos and Veale (2020, p. 138) note, platform providers can utilise a restrictive API design to “break an entire set of business models” that rely on specific data streams through the respective APIs (see also Bucher, 2013; Leerssen et al., 2019). Thus, platform providers can use technical means such as API design to exert economic pressure on other actors within the respective ecosystem by affecting the economic viability and potential profitability of business models behind peripheral applications. Such strategic exploitation of the API design as a tool “to exclude certain business or functionality from integration” (Ausloos & Veale, 2020, p. 138) can establish economic constraints on the scope for design of VSD practitioners at the periphery of ecosystems.

In more decentralised ecosystems, the challenge of accounting for power manifests differently. Here, the negotiation processes among the involved actors on how to account for human values can lead to gridlock. This is because various actors with divergent incentives and interests may disagree as to how human values are conceptualised, weighed, or operationalised. Applying VSD in such cases could ensure that various stakeholder groups are represented in decision-making processes and balance the interests of different stakeholder groups without calling into question the ecosystem's decentralised nature. However, in more decentralised ecosystems, VSD practitioners need to ensure that actors who are not directly involved in design decisions, but affected by them, are also taken into account.

These differences in ecosystems suggest that if there is freedom of choice, the selection of the ecosystem that VSD practitioners embed an artefact in has a significant effect on how human values can be accounted for in the artefact’s design. For instance, Atzori and Ulieru (2017) argue that research on platformisation, i.e., “the penetration of the infrastructures, economic processes, and governmental frameworks of platforms in different economic sectors and spheres of life” (Poell et al., 2019, pp. 5–6) is indicating that “the concept of distributive justice / distributive efficiency [is] strongly dependent on platform architectural design and they are unlikely to be achieved in centralized, two-sided markets” (Atzori & Ulieru, 2017, pp. 4–5). However, due to the quasi-monopolistic position of many platforms (Eaton et al., 2015), such a choice does not always exist for developers of peripheral applications if they want to attract a larger user group. Furthermore, that developers of peripheral applications need to consider not only a platform’s current features, but also the platform provider’s means (technical, economic, or other) to exert power in the future, further complicates the selection process.

Moreover, as the means of different actors to exercise power concern various phases of the design process and can come into play even after deployment, VSD practitioners have to consider the matter continuously: from early conceptualisations of values to the process of operationalising and implementing values to the deployment of artefacts and the evaluation of the design decisions related to values later on. More specifically, since both the development of core components of an ecosystem and peripheral applications often continue after deployment in the form of a constant redesign (Eaton et al., 2015), VSD similarly has to incorporate continuous monitoring of the respective ecosystem’s modifications, updates, developments and related effects on the distribution of power within the ecosystem. While the foundational texts of VSD in principal already set out the approach as extending over all of these phases (Friedman et al., 2008; Friedman & Hendry, 2019), in practice, most practitioners do not perform several iterations of the three phases over the entire length of the design process (Winkler & Spiekermann, 2018). Therefore, it is essential to stress the importance of a continuous application of VSD once more.

Furthermore, the findings of this paper suggest that the spectrum of tasks involved in VSD expands if applied in the design of artefacts embedded in vast and complex socio-technical ecosystems. Some tasks, such as monitoring the distribution of power in the ecosystem over extended periods of time or dealing with platform monopolies, appear to be too extensive to be addressed by individual VSD practitioners or even small development teams. Therefore, the range of tasks needs to be distributed over more actors if they are to remain manageable. Regulatory authorities, in particular, must play a role in addressing some of these challenges. In particular, challenges arising due to 1) (quasi-) monopolistic players, 2) the complexity of continuously monitoring the manifold actor constellations and distribution of power within an ecosystem, and 3) boundary resource design that prevents developers of peripheral applications from accounting for values surpass the capabilities of VSD practitioners and the scope of VSD in a traditional sense.

Dealing with (quasi-) monopolistic actors, especially platform providers, is in the domain of antitrust and competition authorities (Crémer et al., 2019; Kommission Wettbewerbsrecht 4.0, 2019; Monopolkommission, 2015), which therefore play a crucial role in ensuring the applicability of VSD. Furthermore, the European Commission’s recent proposal for the Digital Markets Act (European Commission, 2020c) contains several propositions for concrete regulatory measures aiming to curb the quasi-monopolistic standing of many platform providers. Relevant here are, for instance, the proposed requirements for gatekeepers to “allow business partners,” such as the providers of peripheral applications in a platform-based ecosystem, “to offer the same products or services to end users through third party online intermediation services at prices or conditions that are different from those offered through the online intermediation services of the gatekeeper” 10 (European Commission, 2020c, art. 5 (b)) or to “provide effective portability of data generated through the activity of a business user or end user […]” (European Commission, 2020c, art. 6 (h)). Thereby, regulation built on the European Commission’s proposal for a Digital Markets Act could help to counteract “winner-takes-all dynamics” (Anderson & Mariniello, 2021) that favour the development of (quasi-)monopolies. In turn, such a development would provide more choices to select a suitable platform (or suitable platforms) for developers of peripheral applications and make it a more viable option to integrate this selection process in the application of VSD.

Additionally, establishing oversight institutions like the recently launched AI Observatory of the German Federal Ministry of Labour and Social Affairs (Bundesregierung, 2020) or a competence centre for algorithmic systems, as proposed by the German Data Ethics Commission (Datenethikkommission, 2019), could play a crucial role in the monitoring of platform-based ecosystems. The proposal for the Digital Markets Act outlines further supportive measures. Especially the requirement for gatekeepers “to refrain from preventing or restricting business users from raising issues with any relevant public authority relating to any practice of gatekeepers” (European Commission, 2020c, art. 5 (d)) is crucial here, as it would allow for a closer collaboration of regulatory authorities and developers of peripheral applications. If cooperating closely, oversight institutions and developers of peripheral applications could identify the most problematic practices of platform providers jointly in a bottom-up approach and lay the groundwork for possible future regulation and governance that addresses the most urgent issues for VSD practitioners.

Furthermore, in future regulatory frameworks, regulatory authorities should consider an ecosystem’s boundary resource composition when determining. As outlined above, iOS is a closed platform. Apple controls and governs the unique distribution channel and, thus, establishes itself as an obligatory passage point. Developers of peripheral applications need to follow Apple’s guidelines closely since there is no alternative distribution channel for applications on iOS devices. This is a conscious decision by Apple which brings the company a wide range of business benefits. Yet, from a regulatory perspective, this decision could also be linked to stricter obligations for Apple since it constrains the scope for design of developers of peripheral applications and predetermines to what degree they can account for human values in design decisions. If platform providers use their power to exert influence over design decisions and limit access to alternative distribution channels for developers of peripheral applications, it seems reasonable to link these activities to a stricter regime of obligations. Suppose platform providers engage in exercising control over how developers of peripheral applications account for human values in the technical design of their applications and shape the ecosystem more proactively. In that case, this value sensitive shaping of the ecosystem should be subject to increased scrutiny by regulators. Conversely, if platforms refrain from exercising control over how developers of peripheral applications account for human values in the technical design of their applications, the focus of regulators should shift more to the actors at the ecosystem’s periphery.

6. Conclusion

Technical design in accordance with human values is increasingly considered a building block for shaping the digital future. VSD is a long-standing and well-established approach for achieving design in accordance with human values. This paper shows that the integration of technical artefacts in increasingly vast and complex socio-technical ecosystems with power distributed over various actors affects the applicability of VSD in multiple ways. Several factors determine how this challenge manifests in practice. This paper identifies 1) the level of decentralisation of the ecosystem in question, 2) whether VSD is applied regarding the design of components of an ecosystem’s core or periphery, 3) the temporality of the exercise of power, and 4) the phase of VSD in which power is exercised in as four of these factors.

Adopting a power-sensitive ecosystem perspective provides some reference points for addressing the challenge of accounting for power. While in some constellations, the application of VSD appears to be less applicable since the scope for design of developers is restricted, other constellations appear to accommodate the approach. On the one hand, these are cases where a multitude of assertive actors engage in decision-making processes regarding specific design choices that result in conflicts or even gridlock. Here, VSD can provide a structured approach that supports resolving these conflicts and balances the interests of different stakeholder groups. VSD also appears to be of particular importance for the design of an ecosystem’s core components, such as boundary resources. Here, individual design decisions can shape entire ecosystems in accordance with human values (see Shilton, 2012). For this reason, highly centralised ecosystems are also potentially more attractive to regulatory authorities because important nodes and actors in such ecosystems are more easily identifiable and addressable.

Dealing with (quasi-)monopolistic players, accounting for the complexity of continuously monitoring the complex actor constellations and distribution of power within an ecosystem, and addressing boundary resource design preventing developers of peripheral applications from accounting for values emerge as significant novel challenges when applying VSD in vast and complex socio-technical ecosystems. However, recent proposals for establishing new oversight institutions (Bundesregierung, 2020; Datenethikkommission, 2019) and new regulatory approaches such as the Digital Markets Act and the Digital Services Act (European Commission, 2020c, 2020b) indicate that regulatory authorities can support VSD practitioners in overcoming these challenges. Furthermore, in the future, close cooperation between oversight institutions and VSD practitioners can reveal problematic practices of powerful actors in socio-technical ecosystems and thereby lay the foundation for further regulatory action.

Lastly, a lesson that can be drawn for further research in the field of VSD is that developing a unified framework for dealing with power imbalances between stakeholders in socio-technical ecosystems does not seem to be an attainable goal because the way that power manifests in different ecosystems varies substantively. Thus, instead of aiming for a unified framework, practitioners need to make calls on adequate procedures on a case-by-case basis. Future research, therefore, should aim at advancing the understanding of the actor constellations in socio-technical ecosystems and the distribution of power within them. In particular, in-depth comparative analyses of various socio-technical ecosystems, the distribution of power among the actors involved in them, and the human values expressed in the design of their boundary resource design could provide valuable and more readily applicable insights for VSD practitioners.

References

Alsindi, W. Z., & Lotti, L. (2021). Mining. Internet Policy Review, 10(2). https://doi.org/10.14763/2021.2.1551

Alstyne, M., Parker, G., & Choudary, S. P. (2016). Pipelines, platforms, and the new rules of strategy. Harvard Business Review, 94(4), 54–62.

Anderson, J., & Mariniello, M. (2021). Regulating big tech: The Digital Markets Act [Blog post]. Bruegel. https://www.bruegel.org/2021/02/regulating-big-tech-the-digital-markets-act/

Antonopoulos, A. (2014). Bitcoin security model: Trust by computation. O’Reilly Radar. http://radar.oreilly.com/2014/02/bitcoin-security-model-trust-by-computation.html

Antonopoulos, A. (2017). Mastering Bitcoin: Programming the open Blockchain (2nd ed.). O’Reilly.

Apple. (2019). Turn Location Services and GPS on or off on your iPhone. Apple Support. https://support.apple.com/en-us/HT207092

Apple. (2020). App Tracking Transparency: Request user authorization to access app-related data for tracking the user or the device. Apple Developer Documentation. https://developer.apple.com/documentation/apptrackingtransparency

Atzori, M., & Ulieru, M. (2017). Architecting the eSociety on Blockchain: A Provocation to Human Nature. SSRN Electronic Journal. Advance Online Publication. https://doi.org/10.2139/ssrn.2999715

Ausloos, J., & Veale, M. (2020). Researching with Data Rights.

Baldwin, C., & Woodard, C. J. (2009). The Architecture of Platforms: A Unified View. In A. Gawer (Ed.), Platforms, Markets and Innovation (pp. 19–46). Edward Elgar.

Becker, M., & Bodó, B. (2021). Trust in blockchain-based systems. Internet Policy Review, 10(2). https://doi.org/10.14763/2021.2.1555

Binance. (2019, April 15). Binance Will Delist BCHSV. Binance Latest News. https://binance.zendesk.com/hc/en-us/articles/360026666152

Bodó, B., Brekke, J. K., & Hoepman, J. ‐H. (2021). Decentralisation: A multidisciplinary perspective. Internet Policy Review, 10(2). https://doi.org/10.14763/2021.2.1563

Borning, A., Waddell, P., & Förster, R. (2008). Urbansim: Using Simulation to Inform Public Deliberation and Decision-Making. In R. Sharda, S. Voß, H. Chen, L. Brandt, V. Gregg, R. Traunmüller, S. Dawes, E. Hovy, A. Macintosh, & C. A. Larson (Eds.), Integrated Series In Information Systems. Digital Government (Vol. 17, pp. 439–464). Springer US. https://doi.org/10.1007/978-0-387-71611-4_22

Botsman, R. (2017). Who can you trust? How technology brought us together and why it might drive us apart (1st ed.). Public Affairs.

Brekke, J. K., & Alsindi, W. Z. (2021). Cryptoeconomics. Internet Policy Review, 10(2). https://doi.org/10.14763/2021.2.1553

Brey, P. (2008). The Technological Construction of Social Power. Social Epistemology, 22(1), 71–95. https://doi.org/10.1080/02691720701773551

Brey, P. (2010). Values in technology and disclosive computer ethics. In L. Floridi (Ed.), The Cambridge handbook of information and computer ethics (pp. 41–58). Cambridge Univerity Press. https://doi.org/10.1017/CBO9780511845239.004

Bucher, T. (2013). Objects of intense feeling: The case of the Twitter API. Computational Culture, 3.

Bundesregierung. (2020). Technologie soll dem Menschen dienen: KI-Observatorium nimmt Arbeit auf [Press release]. Presse- und Informationsamt der Bundesregierung. https://www.bundesregierung.de/breg-de/aktuelles/ki-oberservatorium-1726794

Butts, J. (2017). Thanks to Misuse, Apps Can’t View Mac Addresses on iOS 11. Mac Observer. https://www.macobserver.com/news/product-news/apps-cant-view-mac-addresses-on-ios-11/

Callon, M. (1984). Some Elements of a Sociology of Translation: Domestication of the Scallops and the Fishermen of St Brieuc Bay. The Sociological Review, 32(1_suppl), 196–233. https://doi.org/10.1111/j.1467-954X.1984.tb00113.x

Constantinides, P., Henfridsson, O., & Parker, G. G. (2018). Introduction—Platforms and Infrastructures in the Digital Age. Information Systems Research, 29(2), 381–400. https://doi.org/10.1287/isre.2018.0794

Crémer, J., Montjoye, Y.-A., & Schwitzer, H. (2019). Competition Policy for the Digital Era (Report KD-04-19-345-EN-N). Publications Office of the European Union. http://doi.org/10.2763/407537

Czeskis, A., Dermendjieva, I., Yapit, H., Borning, A., Friedman, B., Gill, B., & Kohno, T. (2010). Parenting from the pocket. In L. F. Cranor (Ed.), Proceedings of the Sixth Symposium on Usable Privacy and Security—SOUPS ’10 (p. 1). https://doi.org/10.1145/1837110.1837130

Datenethikkommission. (2019). Gutachten der Datenethikkommission [Report]. Bundesministerium des Innern, für Bau und Heimat. https://www.bmi.bund.de/SharedDocs/downloads/DE/publikationen/themen/it-digitalpolitik/gutachten-datenethikkommission.pdf

De Filippi, P. (2017). In Blockchain we Trust: Vertrauenslose Technologie für eine vertrauenslose Gesellschaft. In Rudolf-Augstein-Stiftung (Ed.), Reclaim Autonomy: Selbstermächtigung in der digitalen Weltordnung (edition suhrkamp, Vol. 2714, pp. 53–81).

De Filippi, P., Mannan, M., & Reijers, W. (2020). Blockchain as a confidence machine: The problem of trust & challenges of governance. Technology in Society, 62. https://doi.org/10.1016/j.techsoc.2020.101284

De Filippi, P., & Wright, A. (2018). Blockchain and the law: The rule of code. Harvard University Press.

DuPont, Q. (2018). Experiments in Algorithmic Governance: A history and ethnography of ‘The DAO,’ a failed Decentralized Autonomous Organization. In M. Campbell-Verduyn (Ed.), Bitcoin and beyond: Cryptocurrencies, blockchains and global governance. Routledge. https://doi.org/10.4324/9781315211909-8

DuPont, Q. (2019). Cryptocurrencies and blockchains. John Wiley & Sons.

DuPont, Q., & Maurer, B. (2015, June 23). Ledgers and Law in the Blockchain. Kings Review. https://www.kingsreview.co.uk/essays/ledgers-and-law-in-the-blockchain

Eaton, B., Elaluf-Calderwood, S., Sørensen, C., & Yoo, Y. (2015). Distributed Tuning of Boundary Resources: The Case of Apple’s iOS Service System. MIS Quarterly, 39(1), 217–243. https://doi.org/10.25300/MISQ/2015/39.1.10

European Commission. (2019). Building Trust in Human-Centric Artificial Intelligence. https://ec.europa.eu/digital-single-market/en/news/communication-building-trust-human-centric-artificial-intelligence

European Commission. (2020). On Artificial Intelligence—A European approach to excellence and trust (White Paper COM(2020) 65 final). https://ec.europa.eu/info/sites/info/files/commission-white-paper-artificial-intelligence-feb2020_en.pdf

Proposal for a REGULATION OF THE EUROPEAN PARLIAMENT AND OF THE COUNCIL on a Single Market For Digital Services (Digital Services Act) and amending Directive, COM(2020) 825 final (2020). https://eur-lex.europa.eu/legal-content/EN/TXT/PDF/?uri=CELEX:52020PC0825&from=en

Proposal for a REGULATION OF THE EUROPEAN PARLIAMENT AND OF THE COUNCIL on contestable and fair markets in the digital sector (Digital Markets Act), COM(2020) 842 final (2020). https://eur-lex.europa.eu/legal-content/EN/TXT/PDF/?uri=CELEX:52020PC0842&from=en

Feiler, P., Sullivan, K., Wallnau, K., Gabriel, R., Goodenough, J., Linger, R., Longstaff, T., Kazman, R., Klein, M., Northrop, L., & Schmidt, D. (2006). Ultra-Large-Scale Systems: The Software Challenge of the Future. Software Engineering Institute, Carnegie Mellon University.

Friedman, B., & Hendry, D. G. (2019). Value Sensitive Design: Shaping Technology with Moral Imagination. MIT Press.

Friedman, B., Hendry, D. G., & Borning, A. (2017). A Survey of Value Sensitive Design Methods. Foundations and Trends® in Human–Computer Interaction, 11(2), 63–125. https://doi.org/10.1561/1100000015

Friedman, B., Kahn, P. H., & Borning, A. (2008). Value Sensitive Design and Information Systems. In K. E. Himma & H. T. Tavani (Eds.), The Handbook of Information and Computer Ethics (pp. 69–101). John Wiley & Sons, Inc. https://doi.org/10.1002/9780470281819.ch4

Friedman, B., Kahn, P., Hagman, J., Severson, R., & Gill, B. (2006). The Watcher and the Watched: Social Judgments About Privacy in a Public Place. Human-Computer Interaction, 21(2), 235–272. https://doi.org/10.1207/s15327051hci2102_3

Gartenberg, C. (2020, September 24). Spotify, Epic, Tile, Match, and more are rallying developers against Apple’s App Store policies: As the ‘Coalition for App Fairness’. The Verge. https://www.theverge.com/2020/9/24/21453745/spotify-epic-tile-match-coalition-for-app-fairness-apple-app-store-policies-protest

Ghazawneh, A., & Henfridsson, O. (2013). Balancing platform control and external contribution in third-party development: The boundary resources model. Information Systems Journal, 23(2), 173–192. https://doi.org/10.1111/j.1365-2575.2012.00406.x

Giungato, P., Rana, R., Tarabella, A., & Tricase, C. (2017). Current Trends in Sustainability of Bitcoins and Related Blockchain Technology. Sustainability, 9(12), 2214. https://doi.org/10.3390/su9122214

Greene, D., & Shilton, K. (2018). Platform Privacies: Governance, Collaboration, and the Different Meanings of “Privacy” in iOS and Android Development. New Media & Society, 20(4), 1640–1657. https://doi.org/10.1177/1461444817702397

Haugaard, M. (2010). Power: A ‘family resemblance’ concept. European Journal of Cultural Studies, 13(4), 419–438. https://doi.org/10.1177/1367549410377152

Hawlitschek, F., Notheisen, B., & Teubner, T. (2018). The limits of trust-free systems: A literature review on blockchain technology and trust in the sharing economy. Electronic Commerce Research and Applications, 29, 50–63. https://doi.org/10.1016/j.elerap.2018.03.005

Hestres, L. (2013). App neutrality: Apple’s app store and freedom of expression online. International Journal of Communication, 7, 1265–1280. https://ijoc.org/index.php/ijoc/article/view/1904

High-Level Expert Group on Artificial Intelligence. (2019). Ethics guidelines for trustworthy AI [Report]. European Commission. https://digital-strategy.ec.europa.eu/en/library/ethics-guidelines-trustworthy-ai

Introna, L. D. (2005). Disclosive Ethics and Information Technology: Disclosing Facial Recognition Systems. Ethics and Information Technology, 7(2), 75–86. https://doi.org/10.1007/s10676-005-4583-2

Jacobs, M. (2020). How Implicit Assumptions on the Nature of Trust Shape the Understanding of the Blockchain Technology. Philosophy & Technology. https://doi.org/10.1007/s13347-020-00410-x

Karhu, K., Gustafsson, R., & Lyytinen, K. (2018). Exploiting and defending open digital platforms with boundary resources: Android’s five platform forks. Information Systems Research : ISR : An Information Systems Journal of the Institute for Operations Research and the Management Sciences, 29(2), 479–497.

Kommission Wettbewerbsrecht. (2019). A New Competition Framework for the Digital Economy: Report by the Commission ‘Competition Law 4.0’ [Report]. https://www.bmwi.de/Redaktion/EN/Downloads/a/a-new-competition-framework.pdf?__blob=publicationFile&v=2

KRAKENFX. (2019). Kraken is Delisting BSV [Blog post]. Kraken. https://blog.kraken.com/post/2274/kraken-is-delisting-bsv/

Law, J., & Callon, M. (1992). The Life and Death of an Aircraft: A Network Analysis of Technical Change. In W. E. Bijker & J. Law (Eds.), Inside technology. Shaping technology/building society: Studies in sociotechnical change (pp. 21–52). MIT Press.

Leerssen, P., Ausloos, J., Zarouali, B., Helberger, N., & Vreese, C. H. (2019). Platform Ad Archives: Promises and Pitfalls. Internet Policy Review, 8(4), 1–21. https://doi.org/10.14763/2019.4.1421

Leswing, K. (2021). Apple exec warns it may remove apps that track users without permission. https://www.cnbc.com/2020/12/08/apple-may-remove-apps-that-track-users-without-permission-in-2021.html

Lusch, R. F., & Nambisan, S. (2015). Service Innovation: A Service-Dominant Logic Perspective. MIS Quarterly, 39(1), 155–175. https://doi.org/10.25300/MISQ/2015/39.1.07

McConahy, A., Eisenbraun, B., Howison, J., Herbsleb, J. D., & Sliz, P. (2012). Techniques for monitoring runtime architectures of socio-technical ecosystems. Workshop on Data-Intensive Collaboration in Science and Engineering (CSCW 2012).

Monopolkommission. (2015). Competition policy: The challenge of digital markets. Special Report by the Monopolies Commission pursuant to section 44(1)(4) of the Act Against Restraints on Competition (Special Report No. 68). Monopolies Commission. https://www.monopolkommission.de/images/PDF/SG/s68_fulltext_eng.pdf

Morabito, V. (2017). Business Innovation Through Blockchain: The B3 Perspective. Springer International Publishing. https://doi.org/10.1007/978-3-319-48478-5

Mueller, M., & Heger, O. (2018). Health at any Cost? Investigating Ethical Dimensions and Potential Conflicts of an Ambulatory Therapeutic Assistance System through Value Sensitive Design. 39th International Conference on Information Systems (ICIS). https://aisel.aisnet.org/icis2018/healthcare/Presentations/17/

Nathan, L. P. (2012). Sustainable information practice: An ethnographic investigation. Journal of the American Society for Information Science and Technology, 63(11), 2254–2268. https://doi.org/10.1002/asi.22726

Nathan, L. P., Klasnja, P. V., & Friedman, B. (2007). Value scenarios: A technique for envisioning systemic effects of new technologies. CHI ’07 Extended Abstracts on Human Factors in Computing Systems, 2585–2590. https://doi.org/10.1145/1240866.1241046

Orcutt, M. (2018). Chain Letter #102: You can go your own way. https://mailchi.mp/technologyreview/chain-letter-767541?e=93dc606e34&utm_campaign=chain_letter.unpaid.engagement&utm_source=hs_email&utm_medium=email&utm_content=71892724&_hsenc=p2ANqtz--2S3Tqvwcm1BQc_Eb-ArlTKDA-f9cAdBdc6Lxq67nmBy3Y2Z48OyMOEMW__mczpT4YRw2w-7sUtAvryWgLnYMUo_VkOxYs6DqYzXyno1pRK7AGncM&_hsmi=71892724

Orcutt, M. (2019). Chain Letter #139: On social media-fueled coin delistings [Blog post]. Cryptocurrency News Now! https://cryptocurrency-news-now.blogspot.com/2019/04/139-on-social-media-fueled-coin.html

Perez, S. (2020). Coalition for App Fairness, a group fighting for app store reforms, adds 20 new partners. TechCrunch. https://techcrunch.com/2020/10/21/coalition-for-app-fairness-a-group-fighting-for-app-store-reforms-adds-20-new-partners/

Poell, T., Nieborg, D., & van Dijck, J. (2019). Platformisation. Internet Policy Review, 8(4). https://doi.org/10.14763/2019.4.1425

Reuver, M. de, Sørensen, C., & Basole, R. C. (2018). The Digital Platform: A Research Agenda. Journal of Information Technology, 33(2), 124–135. https://doi.org/10.1057/s41265-016-0033-3

Sattarov, F. (2019). Power and technology: A philosophical and ethical analysis. Rowman et Littlefield.

Shilton, K. (2012). Values Levers. Science, Technology, & Human Values, 38(3), 374–397. https://doi.org/10.1177/0162243912436985

Shilton, K., & Greene, D. (2016, March 15). Because privacy: Defining and legitimating privacy in ios development. iConference 2016. https://doi.org/10.9776/16229

Shilton, K., & Greene, D. (2019). Linking Platforms, Practices, and Developer Ethics: Levers for Privacy Discourse in Mobile Application Development. Journal of Business Ethics, 155(1), 131–146. https://doi.org/10.1007/s10551-017-3504-8

Simon, J. (2016). Values in Design. In J. Heesen (Ed.), Handbuch Medien- und Informationsethik (pp. 357–364). J.B. Metzler.

Simon, J., Wong, P. ‐H., & Rieder, G. (2020). Algorithmic bias and the Value Sensitive Design approach. Internet Policy Review, 9(4). https://doi.org/10.14763/2020.4.1534

Swan, M. (2015). Blockchain: Blueprint for a new economy (First edition.). O’Reilly.

Tiwana, A., & Konsynski, B. (2010). Complementarities Between Organizational IT Architecture and Governance Structure. Information Systems Research, 21(2), 288–304. https://doi.org/10.1287/isre.1080.0206

Tiwana, A., Konsynski, B., & Bush, A. A. (2010). Research Commentary—Platform Evolution: Coevolution of Platform Architecture, Governance, and Environmental Dynamics. Information Systems Research, 21(4), 675–687. https://doi.org/10.1287/isre.1100.0323

van Dijck, J., Poell, T., & De Waal, M. (2018). The platform society: Public values in a connective world. Oxford University Press. https://doi.org/10.1093/oso/9780190889760.001.0001

van House, N. A. (2004). Science and technology studies and information studies. Annual Review of Information Science and Technology, 38, 3–86. https://doi.org/10.1002/aris.1440380102

Waddell, P., Wang, L., & Liu, X. (2008). UrbanSim: An evolving planning support system for evolving communities. In R. K. Brail (Ed.), Planning Support Systems for Cities and Regions (pp. 103–138). Lincoln Institute for Land Policy.

Walch, A. (2019). In Code(rs) We Trust: Software Developers as Fiduciaries in Public Blockchains. In P. Hacker, I. Lianos, G. Dimitropoulos, & S. Eich (Eds.), Regulating Blockchain: Techno-Social and Legal Challenges (pp. 58–82). Oxford University Press. https://doi.org/10.1093/oso/9780198842187.003.0004

Walton, R., & DeRenzi, B. (2009). Value-Sensitive Design and Health Care in Africa. IEEE Transactions on Professional Communication, 52(4), 346–358. https://doi.org/10.1109/TPC.2009.2034075

Warnier, M., Dechesne, F., & Brazier, F. (2015). Design for the Value of Privacy. In J. Hoven, V. P. E., & I. van de Poel (Eds.), Handbook of ethics, values, and technological design: Sources, theory, values and application domains (pp. 431–445). Springer. https://doi.org/10.1007/978-94-007-6970-0_17

Weber, M. (2019). Economy and Society: A New Translation. Harvard University Press.

Werbach, K. (2018). The Blockchain and the New Architecture of Trust. MIT Press.

Winkler, T., & Spiekermann, S. (2018). Twenty years of value sensitive design: A review of methodological practices in VSD projects. Ethics and Information Technology, 18(4), 185. https://doi.org/10.1007/s10676-018-9476-2

Wright, A., & De Filippi, P. (2015). Decentralized blockchain technology and the rise of lex cryptographia. https://doi.org/10.2139/ssrn.2580664

Yoo, D., Derthick, K., Ghassemian, S., Hakizimana, J., Gill, B., & Friedman, B. (2016). Multi-lifespan Design Thinking. In J. Kaye, A. Druin, C. Lampe, D. Morris, & J. P. Hourcade (Eds.), Proceedings of the 2016 CHI Conference on Human Factors in Computing Systems (pp. 4423–4434). https://doi.org/10.1145/2858036.2858366

Footnotes

1. This paper uses Friedman and Hendry’s (2019, p. 24) working definition of the term “human value,” referring to “what is important to people in their lives, with a focus on ethics and morality”. While the term has been criticised as being both under- and over-defined, it provides an appropriate balance for the practical application in the context of Value Sensitive Design. For an in-depth discussion of existing critique on the definition as well as the advantages and disadvantages of various alternative definitions see Friedman and Hendry (2019) and Brey (2010).

2. The identification of the “grand challenges” for Value-Sensitive Design that Friedman and Hendry refer to took place at two workshops in 2015 and 2016 organised by Batya Friedman, David Hendry, Jeroen van den Hoven, Alina Huldtgren, Catholijn Jonker, Aimee van Wynsberghe, and Maike Haarbers in Aarhus, Denmark, and Leiden, The Netherlands, respectively. The workshops aimed at “Charting the next decade” for the approach (Friedman & Hendry, 2019).

3. Accordingly, the conceptualisation of power used in this paper does not capture the ability to exercise control over other agents (Brey, 2008).

4. See, e.g., van Dijck et al. (2018) for political and economic manifestations of power, Shilton and Greene (2019) for technical manifestations of power, and De Filippi et al. (2020) for social manifestations of power.

5. Walch (2019) lists further examples of core developers wielding power in design processes.

6. Highlighting the power of significant miners in design processes manifested in the discourse on the Bitcoin block-size, which resulted in the hard fork of Bitcoin and Bitcoin Cash in 2017 (DuPont, 2019).

7. This issue only comes into play where design decisions or other actions by platform providers are in conflict with (the operationalisation of) a value that VSD practitioners aim to account for. Applying VSD to account for other values is still possible for developers of peripheral applications.

8. For a current example, observe the current dispute between Apple and the Coalition for App Fairness (Gartenberg, 2020).

9. When using VSD in the design of individual technical artefacts, platform providers might have to deal with value conflicts during attempts to engage in the value-sensitive shaping of an ecosystem. These conflicts may arise between two or more human values or between a human value—such as privacy—and instrumental values—such as cost-efficiency or usability.

10. Note that the term “gatekeeper” is defined more narrowly by the European Commission than the term “platform provider” that is used in this paper (see European Commission, 2020c)

Add new comment