Paradigms for Studying the Internet

From Technologies and Politics of Control
Jump to navigation Jump to search

February 4

Before we can even begin exploring the who's, what's, and why's – we need to answer the critical question of how. Indeed, the phrase "studying the web" could embrace a staggering world of possible routes to explore, even before beginning to examine its relationship with society and culture. We need something to guide us through this massive field of (very interesting!) foxholes, and link the ideas we encounter into a consistent piece. We need some kind of structure to allow us to understand what we are looking at, the same way a chemist thinks of things in terms of atoms and molecules, or a philosopher can think about things in terms of schools of thought.

This class will explore different frameworks for studying the web, which will structure both the discussion and topic matter covered in the course, as well as the methodology that you should apply to your assignments. The second hour of the class will focus on applying these concepts to Wikipedia, and teeing up the final project for the class, where we will discuss the research prompt, talk about some successful projects from prior years, and plot out the deadlines for the rest of the semester.


Readings

Mechanisms of control
The effects of control

Optional Readings


Assignment 1

Assignment 1 is due before next week's class (February 11th). Details of the assignment will be discussed in today's class; see this page for further information. You can submit the assignment here.

Videos Watched in Class

Links

Class Discussion

Please remember to sign your postings by adding four tildes (~~~~) to the end of your contribution. This will automatically add your username and the date/time of your post, like so: Andy 11:49, 8 November 2013 (EST)




Therefore, to maintain order, ensure efficient government, and improve social justice, kings, presidents, and prime ministers must be the chief architect of their country's internet code. They must be multi-skilled or have the support of a talented and scholarly team.

Ichua 13:39, 31 January 2014 (EST)




It seems lots more fun to watch than just read: http://www.youtube.com/watch?v=o7UlYTFKFqY

Ichua 03:30, 2 February 2014 (EST)

Zittrain's talks are always a lot of fun! But we chose the two chapters in order to focus on a few of the specific things we'd like to dive into for this class. His book talk is much more general. Andy 08:46, 2 February 2014 (EST)




The Zittrain chapters give a good overview of how the Internet had been developed up to circa 2008, but there have been some significant changes--and possible reversals of the "generative" model since that time. The increasing role of SaaS platforms, centralized APIs, and operating platforms with a much more pervasive level of control relative to older operating systems (e.g., IOS, Android, and social networking platforms like the Facebook developer platform) have reintroduced an aspect of large, single-point-of-failure, commercially controlled systems. Whereas Cluetrain envisioned a future of "small pieces loosely joined," the Internet of today might be better described as "lots of small pieces largely dependent on a few large, commercially-controlled pieces." These few large pieces raise concerns in terms of limiting the potential for innovation, negotiation with gatekeepers (which, as rightly discussed in the Zittrain chapters, was one of the things that killed innovation on earlier mobile platforms) and the shifting of business opportunities across the market from creators to platform owners. Will there be another wave of generative platforms that will wear down the the current trend to centralization, and if not, how can we best ensure continuous innovation on the Internet? Jradoff 20:27, 3 February 2014 (EST)


At an event last night Prof. Zittrain mentioned another possible enclosure movement for generativity I hadn't thought of before: many web services are finding themselves at the receiving end of DDoS Attacks for one reason or another. As a result, services are moving from their own servers capable of withstanding such attacks - primarily Amazon Web Services, but there are a few others as well. If all of the Internet moves to just one of three or four web servers, that gives those servers tremendous power to cut off something they may not like. That's a form of "contingent generativity" that could cut off a lot of the social good that both Zittrain and Benkler flag in their articles. Andy 09:28, 4 February 2014 (EST)




Coming off of last week's reading (specifically John Perry Barlow's "A Declaration of Independence of the Internet"), I found danah boyd's essay "White Flight in Networked Publics?" particularly interesting. Even before reading boyd's piece, Barlow's "Declaration" seems hilariously naive in 2014, though I can certainly appreciate the utopian vision it's based on. The idea that the world that we exist in (the physical reality described by Orin Kerr) won't intrude on the virtual world of the Internet seems impossible. (Did they really not believe that the best AND worst parts of us would be present?) The role of the Internet in our everyday social lives has, of course, increased exponentially since 1996, so it only makes sense that who we are and how we behave in the physical world will translate to equivalent behavior on the Internet. The ways in which behavior on the Internet effects people in the physical reality of their lives (particularly when it comes to harassment, threatening behavior, etc.) lends a great sense of urgency to figuring out how we should think about the Internet and the law. Jkelly 23:18, 3 February 2014 (EST)

Both boyd and Hargittai use a lot of pre-Internet scholarship in their writings for this course - a nice reminder that new technology does not necessarily mean new approaches to scholarship. But as Benkler notes, it is not that we are simply repeating the 20th century with shinier objects. There is something different about the way that information travels today that changes the ecology of information and cultural production. We can either adopt that change or legislate/architect it away. Andy 10:36, 4 February 2014 (EST)




I agree with Ichu's remarks about the need to somehow maintain order and to do so utilizing a talented and scholarly task force. My question would then be how this team would be selected/elected? Another potential issue would be how to ensure justice in a system where internet code is controlled by one's government or sole government official/king/president? In our reading by Orin Kerr, he highlights how these conflicting external and internal perspectives on the internet add fuel to the problem of internet law. The internet has two personalities in its vast internal cyberspace and also in acting as a physical network; striking a balance between the two and incorporating both identities into a legal system continues to evade and frustrate authorities.

In response to Megan Garber's reading on Wikipedia, I find that Wikipedia often does not get the credit or praise it deserves. Admittedly, no online community-built encyclopedia can be fool-proof, but the reason why Wikipedia has prevailed is its relative reliability. I have used the site extensively and it has provided me with a quick summary of events on a particular debate or issue. Garber's reasons for Wikipedia's success are logical in that familiarity is the cornerstone for many website's success rates. The ease of navigating the site and the non-committal method of editing or adding to the work encourages more users to contribute. I would also argue that, beyond the cultural/socio-economic/racial influences that cause users to migrate from site to site (such as from myspace to facebook), the constantly changing platform of facebook has led many to stray from the site. This is difficult to prove, of course, but when I had a Facebook account I recall many complaints from my peers about all of the changes that kept happening occurring on the site. It seemed that every week we had to ajust to a new feature or re-learn how to navigate. Accordingly with Garber's theory, the "familiarity" factor was diminishing for users and people tend to resist change especially on a site that they have grown accustomed to.

--AmyAnn0644 04:08, 4 February 2014 (EST)

I was also interested in Megan Garber's point that the authorless structure of Wikipedia lowers the pressure of contributing. It certainly makes sense to me (and, I'm sure, to anyone who has read the comment section of any news article or blog post ever written...) that anonymity can encourage participation. When there's lower pressure to perform and you aren't faced with high stakes when you get involved, it's easier to bring yourself to contribute. This seems to tie in to Zittrain's point about the success of Wikipedia: it developed somewhat un-self-consciously and organically, rather than as a top-down "knowledge project" initiated by large universities. Oversight of the development of new technologies would presumably put a damper on this type of growth at any and all levels. I think this is nicely addressed by Zittrain's point that we're not looking at choosing between technology and non-technology, but a hierarchy and polyarchy.

Jkelly 12:48, 4 February 2014 (EST)


Lawrence Lessig’s article focuses on liberty in Cyberspace and how various modes of regulation effect that liberty. He focuses on four different ways that the web can be regulated,1)the Law, 2)social norms, 3) the market, and 4)architecture. Lessing tries to get us to think differently, more critically, about different mechanisms that can lead to restriction of freedom on the Web.

For instance, with the architecture of the Web, Lessing asserts that the written code of programs inherently can either provide more freedom, or restrict freedom, and access. And when it comes to the law, Lessing points out that “The efficient answer may well be unjust.” He gives an example of the law requiring life sentences for stealing car radios.

We all would probably agree that that is overboard and excessive. And, with that absurdity planted in our minds, Lessing then shows how a coder could easily put a restriction in the radios code that would make stealing the radio less desirable for thieves. Which would in turn make it unnecessary for such a draconian law of life sentences for car radio thieves. This example makes me think about Aaron Swartz, a friend of Lessigs, whom took his own life in 2013. Aaron was prodigy kid who helped create RSS feed, and Reddit at a young age. He later became what you might call an internet activist, and made enemies in the federal government for some hacking activities. He was eventually charged with multiple felonies by the Federal government for hacking MIT’s JSTOR server. Lessing talks about how law and code can either liberate or restrict the Internet. I believe Swartz’s case shows how the MIT/JSTOR rules of access, restricted information on the Web, and how federal laws were excessive and restricted innovation and liberty for Web users. And lastly, Swartz’s case shows how one coder tried to use hacktivism, to liberate information on the web.Mikewitwicki 12:58, 4 February 2014 (EST)