“This is not how we imagined it”: Technological affordances, economic drivers, and the Internet architecture imaginary

The Internet architecture is widely perceived as engine for innovation by providing the equal opportunity to deploy new protocols and applications. This view reflects an imaginary that guides the co-production of policy and technology that can be traced back to the early days of the Internet, which is still prominent among the engineers in one of the main governance bodies of the Internet, the Internet Engineering Taskforce (IETF). After the privatization of the Internet architecture in the 1990s, the interplay between the architectural principles of end-to-end, permissionless innovation, and openness subverted equality among Internet users and hampered their ability to redesign the Internet. I draw on media studies, science and technology studies and international political economy, and use a combination of qualitative and quantitative methods to show how the Internet architecture’s affordance structure got reconfigured, and how this facilitated the prioritization of corporate interests over the interests of end users.


Introduction
When in the early 1990s, the Internet was released from the labs and found its way to millions of users, it was widely perceived as an engine for innovation (Van Schewick, 2012), an information highway (Flichy, 2007), and a tool for democratization (Castells, 2009). These expectations and aspirations accompanied the development of the Internet architecture and were operationalized through three main architectural principles, namely, end-to-end, permissionless innovation, and openness. In this article, I show how the interplay between these principles, after the privatization of the Internet in the early 1990s, undermined the equality of users and the ability of individuals, researchers, and small companies to redesign the Internet.
The Internet architecture is co-produced by corporations, state actors, researchers, and advocates in a self-regulatory industry standards body called the Internet Engineering Taskforce (IETF). 1 Self-regulation has been the general paradigm for the governance of the Internet, because it is assumed to be most suited to the transnational and quickly evolving nature of the Internet (Price and Verhulst, 2000). The evolution of the architecture of the Internet is taking place through the development of open and voluntary standards that facilitate interoperability between the products of network operators, equipment vendors, content and service providers, and software developers. Because of the nature of the standard setting process, it has been described as a "wild mix of politics and economics" (Shapiro and Varian, 1998) and "politics by other means" (Abbate, 1999). While the standards and protocols that are developed in the IETF are largely hidden from the larger public, they shape our behavior (Chadwick, 2006), determine vectors of control over user data flows (Galloway, 2006), how users access information (DeNardis, 2014), and how users can exercise their rights online (Lessig, 2006).
To understand the standard setting process, I use the terms "sociotechnical imaginary," "co-production," and "technological affordance." A sociotechnical imaginary is the combination of visions, symbols, and futures that exist in groups and society. It influences behavior, individual, and collective identity as well as the development of narratives, policy, and institutions (Jasanoff and Kim, 2015). A sociotechnical imaginary guides the process in which people co-create knowledge, technology, and order, a process that Jasanoff (2004) calls co-production. Technology, which is an inherent part of this co-production process, inhibits and stimulates human behavior. Hutchby (2001) describes this "constraining, as well as enabling, materiality of artifacts" (p. 441) as technological affordances. I leverage these terms to show how the Internet architecture's sociotechnical imaginary and its technological affordances got reconfigured and subverted during three decades of co-production following the commercialization and privatization of the Internet. This compounded theoretical lens allows me to jointly take into account the shaping of institutional configurations, technological orderings, economic drivers, and the collaboration among disparate groups and competitors facilitated by a joint vision. This approach enables me to analyze the Internet architecture as a site of contestation (ten Oever, 2019), as an assemblage of power (DeNardis, 2014), and "as a normative 'system of systems,' and to unpack 'the micro practices of governance as mechanisms of distributed, semi-formal or reflexive coordination, private ordering, and use of internet resources'" (Epstein et al., 2016), without defaulting to a reductionist approach.
I will provide an overview of the relevant literature, then I will provide an overview of the methods used in this research, after which I will provide an analysis in which I will establish the Internet architecture imaginary, and subsequently show how it got subverted. Finally, I will offer some thought about what this means for self-regulatory governance models and avenues for further research.

An imaginary space between a technological dream and an economic reality
Instead of looking at the content of datastreams, which is like the "juicy piece of meat carried by the burglar to distract the watchdog of the mind" (McLuhan, 2013: 19), I argue we should rather look at the preconditions, shapes, and characteristics of data streams, the Internet architecture. Before elaborating on this, I will first describe the process through which the Internet architecture is co-produced and introduce the concept of the sociotechnical imaginaries as a lens to understand this process and give an overview of recent academic debates pertaining the Internet architecture. Contemporary debates in media studies, science and technology studies, and governance studies that discuss the Internet architecture focus on (1) the values, or lack thereof, that are enshrined in the internetworking protocols (Braman, 2012b;Flanagin et al., 2010), (2) how the Internet infrastructure is used to exercise control (Musiani et al., 2015;Van Eeten and Mueller 2013), and (3) consolidation and market concentration in the Internet architecture (Easterling, 2014;McKelvey, 2018;Mansell, 2013).
Technical standards, of which networking protocols are a subset, are rules, procedures, and formats that facilitate communication between two or more parties. The Internet architecture consists of "standards which make up the technical back-bone of an information infrastructure" (Hanseth and Monteiro, 1997: 183) that, through its affordance structure, dynamically shapes our information societies. The Internet architecture is co-produced in governance bodies and standards developing organizations such as the IETF. Whereas, theoretically, participation in the IETF is open for everyone, it is dominated by employees of transnational corporations. The most common affiliations of the authors of IETF output documents, the so-called Request for Comments (RFC-) series, are Cisco, Huawei, Ericsson, Google, Juniper, IBM, Nokia, Microsoft, AT&T, and BBN. 2 The RFC-series should not only be understood as a series of technical documents, but also as policy documents (Braman, 2013), which describe the values, processes, and procedures for co-production and, therefore, are relevant for understanding the sociotechnical imaginary of the Internet architecture. While most RFCs are written by authors with an affiliation in the private sector, there are also many RFCs that have been authored by researchers, and even some by members of civil society organizations.
The Internet architecture's sociotechnical imaginary revolves around doing things that are "for the good of the Internet" (Mathew, 2014: 160), sustaining a "generative Internet" (Zittrain, 2008: 6), and is underpinned by three specific engineering principles, namely, end-to-end, permissionless innovation, and openness (Internet Society, 2012).
The Internet architecture's imaginary is rooted in the idea that the network is a general purpose "common carrier network" (Davies et al., 1967: 3), where "all hosts are equal" (Mogul et al., 1984: 1), meaning that they can function as general purpose end nodes (cf. Braden, 1989;Carpenter, 1996;Padlipsky, 1982), and that everyone has the freedom to shape their own traffic by deploying new protocols between end nodes, and thus redesigning the Internet. The ability to freely deploy protocols fits very well with the "ideology of open standards" (Russell, 2014: 21) and the voluntary nature of the Internet standards developed in the IETF. In this article, I interrogate this sociotechnical imaginary because "[m]yths are important for what they reveal (including a genuine desire for community and democracy) and for what they conceal (including the growing concentration of communication power in a handful of transnational media businesses)" (Mosco, 2005: 19).
DeNardis (2009) explores the complex process of the co-production of the Internet architecture by describing how the mere transition from one protocol to another caused a significant amount of contestation because of its geopolitical interests and impacts. This process of contestation is described by Clark et al. (2005), who argue that there are different adverse interests at work in defining the Internet architecture, that this "tussle" should be accommodated because "it is crucial to the evolution of the network's technical architecture" (Clark et al. 2005: 65), and that rigid designs which do not accommodate this tussle will not survive the passage of time and will be broken. Braman (2011) convincingly shows that social policy issues such as rights and freedoms have always been part of the Internet standards deliberations. Braman (2011), however, did not address how or whether these discussions actually resulted in changes in the technical materiality of the network. Davidson et al. (2004) foresaw Braman's findings, but argue that while "many technologists within the leading standards bodies are public-minded, few have explicit expertise in policy-making or at interpreting the public interest. Standards organizations have always (appropriately) emphasized technical goals over societal ones" (p. 4). Since the call of Davidson et al. to assess the impact of protocols in the IETF, there have been several efforts to better understand the relationship between values and networks (Orwat and Bless, 2016), develop guidelines to integrate human rights considerations in protocol design (ten Oever and Cath, 2017), and calls for the IETF to "enable the actualisation of human rights through the protocols and standards it designs by implementing a responsibility-by-design approach to engineering" (Cath and Floridi, 2017: 449).
The IETF has not operationalized any structural assessment of the impacts of their standards and protocols. The lack of integration of impact assessments in the standards process and the intentional undermining of technical standards has led to a discussion about the legitimacy and adequacy of the self-governing technical standards bodies to deliver a trustworthy Internet architecture (Rogers and Eden, 2017). Internet shutdowns during political events, such as elections, foregrounded how the Internet architecture is used as a domain of control and showed how infrastructure is used by governments to realize their objectives. Research has established a trend of the enactment of governance objectives through and by private parties rather than governments (Arpagian, 2016;DeNardis and Musiani, 2014). Levinson and Cogburn (2016) remark that this process is tightly connected with the privatization of the governance of the Internet architecture (p. 219). While the privatization of the Internet architecture was supposed to lead to competition and innovation (Cowhey et al., 2009;Van Schewick, 2012), this article argues ten Oever that actually led to the subversion of equality between hosts and the freedom to deploy new protocols.
As described above, the Internet architecture has been an object of research in different fields, but analyses that take into account the combination of the guiding sociotechnical Internet architecture imaginary, the materiality of the technology, and economic drivers, are still quite rare, but are gaining traction (e.g. DeNardis, 2014;McKelvey, 2018;Mathew, 2014). Analyses generally take into account institutional and technical aspects, such as Dourish (2018) has done in his analysis of IPv6; or rather, it takes economic and institutional aspects into account, such as Van Schewick (2012), Russell (2014), and Smyrnaios (2018) have done. Arguably, both approaches undervalue the dynamic interplay between technological materiality, institutional configurations, and economic drivers. I argue that it is exactly this interplay that creates new orderings and affordances. My theoretical contribution is to reveal how economic drivers prompted an interplay among architectural principles, which led to a reconfiguration and subversion of technological affordances and the Internet architecture's sociotechnical imaginary. This seeks to overcome an economic, legal, or technological reductionist approach in the analysis of the Internet architecture.

Methods
My research into the IETF started with a long-standing fascination for RFCs: their language, particular formatting, and authoritative standing for everyone interested in computer networking. The institutions, people, and processes behind the production of the RFCs, their infrastructure so to say, only became apparent when I started participating in the IETF and its surrounding environs. This participation is the basis for my ethnographic memoir, which developed into participant observation when I formalized my research plans. The research period spanned between March 2014 and July 2018, during which I participated in 11 tri-annual IETF meetings and actively participated on mailing lists. I participated in several working groups, and served in several leadership positions. This experience provided me with a firsthand account of the practices in the most prominent Internet standards body, as well as access that an external observer might not otherwise have. In qualitative research, the researcher is an inherent part of the creation of meaning (Denzin et al., 2006), part of a critical ethnographic practice is, therefore, "an ongoing commitment to re-thinking and re-doing one's work as ethnographer and activist" (Lave, 2011: 2). Part of this process was to address my particular situatedness in the fieldsite (Haraway, 1988), namely, as an activist-engineerresearcher. To gather and seek to understand different points of views, I employed a mixed method approach, to triangulate and validate my findings, and in that process to create an opportunity for reflection on research context, the relationships with the community I researched and was situated in, and the power dynamics in the process of knowledge production (Haraway, 1991).
To analyze the evolution and emergence of explicit values in the large body of data on transnational governance of the Internet infrastructure, I engaged in the quantitative analysis of IETF mailing lists and the IETF's technical documents published in the socalled RFC-series. In my analysis, I focused on prevalence and development over time of language related to society, ethics, and rights, as well as trends in the professional affiliation of document authors, guided by intuitions that arose from participant observation. I obtained these documents from the IETF website, after which I undertook a quantitative document analysis of the RFC-series using the JavaScript-based tool rfcanalysis, 3 and engaged in quantitative mailing list analysis using the Python-based tool BigBang, 4 to gain a deeper insight into cases, trends, and interactions. The outcome of these quantitative analyses informed the creation of the questionnaire I developed for the interviews, and helped me focus on architectural principles and specific protocols. I engaged in 25 semi-structured interviews with IETF leadership and RFC authors through a purposive sample of seasoned and visible members of the Internet protocol (IP) community. The audio was transcribed and analyzed using qualitative methods informed by thematic analysis, with which themes were identified and coded across interviews. Through the identification of themes, concepts, practices, and activities, I analyzed the interview data to understand the ways in which the Internet architecture fundamentally changed from the early 1990s up to now, and how that affected the equality of users and their ability to design and deploy new protocols. The last step of triangulation and validation was to see how my findings on the Internet architecture compared with the description of the Internet architecture in the RFC-series. Therefore, I engaged in the qualitative analysis of a specific subset of RFCs. I made a purposive sample of 20 RFCs that mention "architecture" and made a chain sample of 20 RFCs that are referred to in the aforementioned 20 RFCs to add background understanding for the architectural issues that are being referred to and that they are responding to. Finally, I analyzed five RFCs that specifically got mentioned during the interviews.

Analysis-on the idea of smart endpoints and the dumb pipes
I will first describe the Internet architecture imaginary, then I will describe the challenges in the form of the rise of the middlebox and the accompanying reconfiguration of the affordances of the architecture, and the subsequent iterative responses to overcome the obstacles presented by this new ordering.

The Internet architecture imaginary
The sociotechnical Internet architecture imaginary emerged during the early phases of the development of the Internet, while it was still a research network. I focus on the stabilization of this imaginary that started with the privatization of the Internet in the early 1990s when the US government ceded direct control over the Internet. This gave way to an increased amount of self-regulation through private governance bodies such as the IETF. When asking engineers in the IETF about the central architectural values or principles of the IP community, their answers have a significant amount of overlap. I will describe the imaginary as a category, or ideal type, based on the research data, which in reality can appear less monolithic and will have fuzzy edges. Nonetheless in the interviews, documents, and observations, the imaginary turned out to be remarkably consistent. The end-to-end principle (Saltzer et al., 1984), permissionless innovation, and openness (Russell, 2014) get mentioned time and again in interviews as well as in technical and policy documents (Internet Society, 2012). These three architectural principles shaped a sociotechnical imaginary which is rooted in the equality of "internet host computers" (Deering, 1989;Mogul et al., 1984), the ability to design and deploy new protocols between these computers, and to increase and grow the Internet with more computers and more users (Braman, 2012a). The architectural principles have both sociotechnical and sociopolitical conceptions which play important roles in the coproduction of the Internet. I will discuss the three architectural principles in depth, because they played a central role in the demise of the Internet architecture's imaginary.
The first architectural principle, the end-to-end principle, appeared as a central pillar of the values and principles of the architecture in nearly every interview, RFC3724 even calls it "the core architectural guideline of the Internet" (Internet Architecture Board, 2004). The principle describes where to put functionality in the network, namely, at the edges (Carpenter, 1996), and let the network be "dumb pipes" 5 that solely transport data. The end-to-end principle allowed for a "tremendous amount of agency in individuals and anyone who could put a server anywhere. Anybody could make arrangements to have a prototype protocol pair that you could talk to with each other from anywhere to anywhere." 6 This principle was infra-structurally a revolution because it contrasted so strongly with the communication networks that preceded the Internet. Endpoints that are controlled by their owners can be altered quickly, and thus allow for freedom and flexibility. Changes in the infrastructure are far more cumbersome, or in the words of a former telecommunications engineer turned Internet engineer, "we have the end-to-end principle because so you can do things really quickly on the infrastructure, but [.] if you have to change the infrastructure, that takes a long time." 7 This captures the importance of the end-to-end principle for innovation, but it has further implications as a sociopolitical conceptualization, one engineer mentioned that: There are other folks who take that principle to be more than an engineering principle, but rather an ethical or values driven principle which says that the role of the network is to enable parties to communicate with each other, and not to enable the network itself as a form of control, centralized control. I take both views. 8 The end-to-end principle provides users with the freedom to shape and create their own networking experience. This had a tremendously empowering effect on engineers: there was "a desire to go your own way, um, and a kind of idea that we can invent our own rules and we don't need too many rules, but the ones that we want, we can invent." 9 The second principle, the principle of permissionless innovation, describes that there should be no barriers to the deployment of new protocols. In other words: you don't need to negotiate with any entity in the middle of the network to get your new thing deployed. [ . . . ] [Y]ou don't need to negotiate with any entity in the middle of the network in order to transport your packets. 10 Negotiating is meant here in both technical and sociopolitical terms; permissionless innovation depends on the fact that there is no authority that can sanction what protocols can or should be used. "[A] typical example is the Web. Tim Berners-Lee did not ask permission from anyone, he invented something, went back, built it, and then it was downloaded and no one [ . . . ] had anything to say about it." 11 This sociotechnical conceptualization has clear sociopolitical implications: there should be no limitations on the ability to deploy new technologies on the network. With this principle, participants in the IETF also limit their own authority and responsibility, summarized in an often repeated phrase among long-time participants: "we're not the protocol police." 12 The third principle, the principle of openness, is described as the property "that you can reach from any point of the Internet to any other point of the Internet without your packets been hampered or they'd been stopped or so on," 13 it furthermore means that new computers can be added to the network.
The sociotechnical conception of openness is directly coupled with a conception of connectivity, access, as well as their explicit sociopolitical consequences. Except for the sociotechnical conception of openness of the network, openness is also often associated with the sociopolitical consensus approach to standards development, which fits into an "ideology of open standards" (Russell, 2014: 21) that "linked the open standards-making process with the ideals of participatory democracy, open markets, individual autonomy, and social progress" (Rogers and Eden, 2017: 804). Similar to the end-to-end principle and permissionless innovation, openness is associated not only with a technical ability, in this case, to add nodes to the network, but also with open communications, open standards, and open governance (Internet Society, 2013). The ideal of participatory democracy is also reflected in the IETF's unofficial motto: "We reject: kings, presidents, and voting. We believe in: rough consensus and running code" (Clark, 1992, 543), a credo which was minted during the Internet-Open Systems Interconnection (OSI) standards war (DeNardis, 2009;Russell, 2006), when the governance model of the IETF was heavily tested and further refined. This process strengthened the organizational practice that individuals opinions should seriously be considered and discussed, and cannot simply be overruled by an authority or a majority (Resnick, 2014). This shows the strong interrelation between the technology, the institutional organization of the IETF, and the community that co-produces the architecture. The makeup of the community participating in the IETF, however, has changed over the years. In the early days of its work, the Internet architecture was produced largely by network researchers that were working at universities and as government contractors. Since the privatization of the Internet in the early 1990s, there has been an exodus of researchers (Ding et al., 2013), whose ranks have been filled by contributors from the private sector who now dominate the IETF. This can, for instance, be observed in increasing preeminence of private sector affiliations among authors of RFCs. 14 While the IETF community makes explicit statements about values and principles, its website says, "We try to avoid policy and business questions, as much as possible." 15 This is quite a remarkable statement for self-regulatory body of a US$44 billion networking-infrastructure market. 16 Interestingly, the architectural principles that have strong sociotechnical and sociopolitical conceptions, at the same time, anchor the architectural imaginary and obfuscate the socioeconomic reality.

Cracks in the imaginary I: firewalls, NATs, and network management
The first threat to the end-to-end principle, the openness of the Internet, and permission innovation took the form of middleboxes. Middlebox is a shorthand for "intermediary ten Oever device[s] performing functions other than the normal, standard functions of an IP router on the datagram path between a source host and destination host" (Carpenter and Brim, 2002). Middleboxes can have many different functions, such as firewalls, network address translation (NAT) routers, IP tunnels (such as virtual private networks), and network management devices. The introduction of middleboxes formed a paradigmatic shift in the functioning of the network (McKelvey, 2018). Whereas, the network was previously supposed to function as a "dumb pipes", 17 as outlined by the end-to-end principle, functionality was added to the network. This happened because of the following three issues that resulted from the rapid growth of the network: (1) an increased need for security, (2) the depletion of IP addresses, and (3) increasing business interests (Internet Architecture Board, 2004). I will shortly describe the reactions to these issues below.
To be able to connect to the network, every device needs a unique number, an IP address. It was never envisaged that so many devices would be connected to the network, so when more devices were connected, IP addresses were running out and a new addressing scheme needed to be developed. This was especially pressing since adding new nodes to the network is an inherent part of the principles of openness, one of the Internet's architectural principles. However, there was no direct replacement addressing scheme ready, and there were projections that IP addresses would run out by 1994. 18 This led to the introduction of the "temporary solution" of NAT, which allowed a network of computers to share one public IP address (Francis and Egevang, 1994) from the pool of IP addresses. While this was an efficient short-term solution, this directly went against end-to-end principle according to which "packets [should] flow unaltered through the network" (Carpenter, 2000). NATs interrupted the packet flow because the IP address of the end-device is not known to the network or the recipient and needed to be added by the middlebox, thus adding functionality to the network.
When the network grew beyond a group of researchers, there was the need to introduce firewalls "which screens network traffic in some way, blocking traffic it believes to be inappropriate, dangerous, or both" (Freed, 2000: 2). Firewalls were installed on end-devices, home routers, as well as inside larger networks and thus not only found at the edges of the Internet. A regularly implemented functionality of firewalls is directionality. This means that a network, and the computers connected to it, are "protected" from receiving connections from an outside computer that it did not request. This is a sound security measure on one hand, but on the other hand, it creates a difference between servers and clients. If your computer is located behind a directional firewall (or NAT), the computer cannot function as a server because other clients cannot reach you, traffic can only be initiated from one end of the connection. While many smartphones currently have more processor capacity and storage space than early webservers, they cannot function as a server because the network is imposing a onedirectional ordering. This is how NATs and firewalls create the difference between producers and consumers. Network operators, with the help of equipment vendors, inscribed boundaries into the Internet architecture and attempted "thereby to configure the user such that s/he can only meaningfully encounter the technology on the company's terms" (Hutchby, 2001: 451). This represents the first step in the creation of inequality between Internet hosts, and thus creation of a class of mere users. Network management is used by network operators to optimize network performance. There are different ways for approaching network management, a contested approach is the differentiation and prioritization of specific services over others, or even the blocking of some services or providers for economic reasons. This discussion is more commonly known as the net neutrality debate (Crowcroft, 2007). One probably would not notice if you would receive an email a few milliseconds later, but if there is a delay in the delivery of a videostream, this might cause irritating hiccups. It might seem efficient to prioritize video content over mail traffic, there is, however, a fine line between network management and discrimination between kinds of traffic. If one prioritizes a specific kind of traffic or traffic from a specific provider, this could pose a barrier for alternative streams and providers to grow and develop, since competitors would have a distinct advantage. Violations of network neutrality also interfere with the end-to-end principle and the idea that the network should just transport packets.
The introduction of middleboxes in the network solved some immediate problems, such as security issues, delayed other problems, such as the shortage of IP addresses, and create some economic incentives, in the case of the prioritization of services. The response to the issues of security and the lack of IP addresses could also be understood as response to the architectural principle of openness, because if these issues would not be addressed, it would hamper the connectivity of existing and new nodes. The changes in network management could be interpreted as enactments of permissionless innovation, but all responses inherently violated the end-to-end principle.

Cracks in the imaginary II: the advent of ossification and the failure of stream control transmission protocol
While middleboxes improved performance of specific kinds of traffic, they also negatively impacted the ability to alter protocols through a process called ossification (Thaler, 2011). Ossification is the decreasing flexibility of the network which results in the inability to deploy a new protocol or protocol extensions due to the unchangeable nature of infrastructure components that have come to rely on a particular feature of the current protocols (Clark, 2018). NATs and firewalls ossify around specific protocol characteristics. If these middleboxes receive traffic with other, and thus unknown, characteristics they will reject the traffic. While middleboxes seek to optimize the network, they actually hamper the ability to deploy new protocols. Or in the words of a senior network operator: So at the moment there's a whole industry of middleboxes that basically break [ . . . ] end-toend connections.
[T]hey end up ossifying the internet itself [ . . . ] because these are boxes that are trying to operate transparently and sort of invisibly you don't know that they exist or where they exist. You can't point to them even. They don't have an address. You can't do anything. They are bumps in the wire. 19 Actually, ossification by middleboxes sometimes turns out to be a lot more than a proverbial bump in the wire by actually obstructing the deployment of new protocols as the following example shows.

ten Oever
The stream control transmission protocol (SCTP) was developed as an evolution to transport more data in a faster way than was possible up to then. Initially, it was standardized for telephone networks in 2000 (Taylor et al., 2000) and was adapted to be a general purpose IP in 2004 and after that has received updates for over a decade. Nonetheless, SCTP never significantly worked on the Internet. SCTP worked perfectly in the lab and lived up to all of its design expectations, but it would not work in the wild, on the actual Internet, because middleboxes added inflexibility to the network, in other words, ossification. In the words of a former SCTP developer: [Y]ou can run [SCTP] on your own network when you control all of the middleboxes, but if you try to run it across the public internet there's some non-trivial points that the traffic won't get through because there will be some boxes like: "SCTP, what's that?." NAT middleboxes are a classic example there. [ . . . ] [Y]ou can't really run SCTP across the public network. We tried that and there's too many things in the way. 20 Middlebox induced ossification changed the Internet from an environment where equal hosts could deploy their own protocols, to a network where to design for the future, protocols need to look like the past. Foundational architectural principles of the Internet imaginary cannibalized themselves: in order to safeguard openness, permissionless innovation in the network was leveraged. This undercut the end-to-end principle, which in turn undermined permissionless innovation.
The introduction of middleboxes reconfigured the affordances of the network, with pivoted the locus of control from the endpoints to the network operators (Minar and Hedlund, 2001). The latter were enabled in this endeavor by networking equipment vendors. There were clear incentives for both the network operators and the equipment vendors: the network operators wanted more control over their networks, and offer better performance to their customers. Equipment vendors wanted to sell the network operators equipment. The way they did this was adding more intelligence to the network, which was a relatively low investment for the operators which yielded results on the short term, and benefited the network equipment vendors. An Internet pioneer who was on the forefront of connecting new countries and continents formulated it this way: There seems to have been the development that there is now more, some would say, "intelligence" in the network now. Well, this is a bunch of shit from a bunch of basket cases like Cisco with a willing set of co-conspirators called network operators because in the telephone world they were the center of the universe. [ . . . ] The network folks looked at this [the Internet] and said, no, no, and they found a willing co-conspirator in Cisco and instead of having 15 line router that just switched packets, now they have something with 50,000,000 lines of code. 21 Freedom, agency, and control were taken from the endpoints by network operators, with devices that were provided by equipment vendors. As I have shown, this had both technical and economic reasons, which jointly surmounted to a reordering of the affordances of the network. This reordering largely benefited network operators and equipment vendors, not so much the people that were operating services on the endpoints, because they were hampered in the deployment of new protocols, such as SCTP. Thus, a response from the latter group was to be expected.

The return of the strong endpoints: the rise of Quick UDP Internet Connections
The limitations introduced by middleboxes accrued quite some resentment in the IP community because it constricted the freedom to deploy new protocols in the network. The fact that middleboxes do not announce themselves, and thus make the troubleshooting issues harder, added to the frustration. For quite some time, protocol developers could not find a solution: SCTP developers had worked on it for almost a decade and did not solve it. For content providers, it became increasingly important to have a protocol that would deliver content in a faster manner over the networks, because of the increasing demand in streaming video and media rich websites. This finally became possible with the development of the Quick UDP Internet Connections (QUIC) protocol, a connection-based stream protocol which supports multiple streams. QUIC functioned in a way similar to SCTP, with some extra features. A quintessential difference between SCTP and QUIC, however, was that the latter was developed by Google that already had a fast global content distribution network and developed the most-used browser in the world, Google Chrome. Thus, Google held two important pieces of the puzzle, but needed a protocol to connect the two pieces, "Google is very invested in this [QUIC] because they make a lot of money off of making sure that no one gets in the path between them and the user, and they centralize all that power." 22 QUIC would allow Google to serve their content faster, and ensure that user data would not be shared with other parties, such as network operators. Both have significant economic implications for a company that makes most of its money via targeted advertising. But except for motive, Google also had the network control and capacity to develop this. In the words of a longtime protocol developer: the reason that QUIC [.] can do what it can do is because the two endpoints are controlled by the same people, so they [Google] can, they can do like dark releases and AB-testing and all that that we can't do. 23 Google started developing QUIC in 2012 and in January 2018 between 2.1% and 9.1% of all Internet traffic was using QUIC, which is dominated by Google that uses it for 42.1% of its traffic (Rüth et al., 2018). Google was able to gain much better results than SCTP, because it could do testing between its network and its browsers, and because it had significant resources to invest. Network operators would also think twice about blocking Google's faster services because it would negatively impact many of their users: Google's big enough that it's very hard to stop in the sense that when you think about blocking Google you're blocking access to search and peoples' email and all of the different services that they provide, a huge number of different services. 24 This was another large non-technical but rather economic difference between QUIC and SCTP-QUIC already had a large market share through its user-base: Google's users. Google did not keep QUIC for itself as a proprietary protocol, Google brought QUIC to the IETF for standardization, which would increase the chances of broader adoption of the protocol, and therefore, further ensure that new and updated middleboxes would not block QUIC traffic.
The implementation of QUIC will lead to a reordering of the network. QUIC was built to penetrate middleboxes and provide as little control as possible for network operators to shape, filter, or access data streams. QUIC was built to reconfigure the affordances of the architecture: it will reinstate the end-to-end principle and re-enable permissionless innovation but only as long as the QUIC protocol is used, creating a new path dependency. The cause and the effect are clear for protocol developers: "the incentive for QUIC was to try and prevent ossification in the network, but I think the implication is that it's going to take power away from the network." 25 The reasons for this are the limitation incurred by the ossified network and power imbalance: "I do think that there's a massive power differential that exists between people who run the network and the end users," 26 and now "the pendulum is swinging the opposite way" 27 back to the end users. While QUIC restores the end-to-end principle, it cannot overcome the differentiation between users and providers introduced through NAT directionality, and therefore, it does not restore equality between all hosts. This brings us back to the initial conception of the Internet architecture's imaginary, wherein all hosts were equal and one could freely deploy protocols, strutted by architectural principles like end-to-end, permissionless innovation, and openness. While the limitations of ossification have partially been overcome through QUIC, this has only been possible by a significantly resourced transnational corporation that also controlled large parts of the network, and controlled the world's most-used browser, and could hire the best engineers, some of whom previously extensively worked on SCTP.
In other words, a precondition to restore part of the Internet architecture imaginary, was a significant economic incentive and technical and economic concentration, which contributes to an even further consolidated technological and economic reality. The increasing dominance of socioeconomic considerations over sociopolitical considerations is illustrated a member of the senior IETF leadership who confirmed that "you need to play into some of the operators or vendors earning models in order to get something deployed." 28 This reflects demographic changes in the IETF: whereas the IETF used to be dominated by researchers, the overwhelming majority of participants now are representing the private sector. Deploying new protocols is still possible on higher layers of the networking stack, but less so in the lower layers of the architecture, unless one can gather resources like the one Google could muster, as we have seen in the example of SCTP. One simply needs to abide by the rules set by transnational corporations. In the words of a long-time participant in the IP community, "[t]he mantra of the Internet enterprise is simple: 'Get Big or Get Bought!'" (Huston, 2017: 5): It is probably the same phenomenon we see in other industries. When it is brand new [.] you have more freedom to think exactly about how you want this thing to work and not worry about how much money you're going to make, because you're you will just make a lot of money. Now, a lot of people have made a lot of money off the Internet, there is still more, gobs and gobs need to be made, but it is a little bit crowded. We are heading a bit into the "winner-take-all"-phase. 29 While this startling socioeconomic reality, with significant impact on the material affordances of the Internet architecture, is widely recognized in interviews, the official IETF position is "to avoid policy and business questions, as much as possible." 30 This new socioeconomic reality, which is in part produced by the networks affordances, has cannibalized the Internet architectural imaginary that is still being professed. The hollowed out sociotechnical imaginary actually functions as a cover, or even an implicit justification, for this consolidation of communication power.

Conclusion
The Internet architecture is hailed as an architecture in which all hosts are equal and everyone has the freedom to deploy their own protocols. This sociotechnical imaginary is anchored in the principles of end-to-end, permissionless innovation, and openness, and was operationalized through a process of co-production in Internet governance institutions such as IETF. When the Internet architecture was privatized a tussle over control over datastreams ensued between networks operators, enabled by equipment vendors, and content providers. This tussle led to the reconfiguration of the technical affordances of the Internet architecture. In this reconfiguration, the equality of hosts and the ability to deploy new protocols between hosts has been subverted.
The sociotechnical Internet architecture imaginary and its self-regulatory governance model have not been able to cement the equality of Internet hosts and the freedom of researchers, small companies or individuals to function deploy new protocols, especially on lower layers of the protocol stack. Previously, central sociopolitical conceptions and considerations that were part and parcel of the architecture's sociotechnical imaginary effectively faded into the background, while socioeconomic considerations have acquired a far more prominent place in the shaping of the Internet's technological affordances. The Internet architecture imaginary, that is still professed in IETF, obscures the socioeconomic reality in which interoperation between transnational corporations has overshadowed the practice and ethos of doing things for "the good of the Internet" (Mathew, 2014). This dynamic arguably has contributed to the relative absence of the Internet architecture in current academic and policy debates on government regulation of the Internet, whereas the much discussed platforms and search engines are only a part of the Internet power assemblage.
The preceding analysis has shown how economic drivers spurred iterative changed in the affordances and materiality of the Internet architecture as well as its sociotechnical imaginary, illustrating the dynamic interrelation between economic drivers and technological affordances. This analysis contributes to the debates in governance studies by concluding that the self-regulation of the Internet architecture undermined the very design goals of the Internet architecture, changed its sociotechnical imaginary, and facilitated the prioritization of corporate interests.
Future research could focus on how standards development and self-regulatory governance bodies can take explicit values, such as the equality and freedoms of users, structurally into account as a formalized part of their processes, and what internal or external incentive structures would be needed to achieve this.

Funding
The author(s) disclosed receipt of the following financial support for the research, authorship, and/or publication of this article: This project has benefited from funding