Hackers can’t solve Surveillance

Médecins Sans Frontières (MSF), also known as Doctors without Borders, is an organization that saves lives in war-torn and underdeveloped regions, providing health care and training in over 70 different countries. MSF saves lives. Yet, nobody thinks that doctors can “solve” healthcare. It’s widely understood that healthcare is a social issue, and universal health care can not be achieved by either the voluntary work of Doctors or by way of donations and charity alone.

Just as Doctors can’t solve healthcare, Hackers can’t solve surveillance. Doctors can’t make human frailty disappear with some sort of clever medical trick. They can help mitigate issues, fight emergencies, they can be selfless, heroic. but they can’t, on their own, solve healthcare.

One of the ways that Hackers can fight surveillance is to develop better cryptographic communications tools, and train people how to use them.. This is certainly critical work that hackers can contribute to, but we can’t, on our own, solve surveillance.

Nothing that Hackers can do on their own can eliminate surveillance. Just as universal healthcare is only something that can be achieved by social means, privacy respecting mass communications platforms can only be achieved by social means. Safe mass communications platforms can not be created by private interests, neither commercially, nor voluntarily.

As we well know, private medical provisioning provides unequal health care. The reason is obvious, health needs and the ability to pay are not usually corelated. Private provisioning means that those who can’t pay, wont be served by profit-driven institutions, and though this can be mitigated by voluntarism and charity, it can’t be fully overcome.

Likewise, mass communications that are built for the profit motive either need to charge a fee, and thereby be exclusive, or be advertising supported. Other options can exist for connected and technically savvy users, but these will be niche by necessity. For the masses, the main options available will always be well funded platforms with employees to do support, development, and marketing, without wich, it’s impossible to build-up a mass user base.

The lucrativeness of advertising-based platforms, makes it difficult even for fee-based systems to compete, since they don’t generally produce enough revenue to invest significantly in support, development and marketing, which makes them less attractive even to users who could or would pay, but the major issue that kills such platforms is that the fee means that some people will not be able to use it at all.

Thus, commercial mass platforms tend to be advertising driven. This means that the business of platform operators is selling audience commodity. Commodities are sold by measure and grade. You can buy 10lbs of Fancy Grade Granny Smith Apples, or two dozen Grade A free range eggs. Or 2 million clicks from age 18-35 white males.

Audience commodity, the users of the platform, are sold to advertisers, by measure of clicks or conversion, and by grade. For advertisers, audience is graded by specifications that include age, sex, income level, family composition, location, ethnicity, home or automobile ownership, credit card status, etc. The Demographics, as they say.

Since an advertising funded platform must grade audience commodity, it must collect data on it’s users in order to grade them. This means that the one thing such a platform can not offer its users is privacy. At least not privacy from the platform operators and their advertisers.

And so long as the platform operators collect such data, there is no way that this data will not be made available to local and foreign intelligence agencies.

This hard reality has been hard to grapple with, especially for a hacker community who saw the Internet as a new realm, as John Perry Barlow wrote in the Declaration of the Independence of Cyberspace: “We are creating a world where anyone, anywhere may express his or her beliefs, no matter how singular, without fear of being coerced into silence or conformity.” His colleague, John Gilmore, famously claimed “The Net interprets censorship as damage and routes around it.”

Those two quotations, born of the 90s hey-day of net.culture, contrast starkly with what Adam Curtis describes in his BBC documentary All Watched Over By Machines of Loving Grace:

“The original promise of the Californian Ideology, was that the computers would liberate us from all the old forms of political control, and we would become Randian heroes, in control of our own destiny. Instead, today, we feel the opposite, that we are helpless components in a global system, a system that is controlled by a rigid logic that we are powerless to challenge or to change”

Oddly, the film doesn’t credit Richard Barbrook and Andy Cameron who coined the term the “Californian Ideology” in their seminal 1995 text, which was among the first to identify the libertarian ideology endemic in Silicon Valley culture.

The visions of a free, uncensorable cyberspace envisioned by Barlow, Gilmore and others was incompatible with the needs of Capital, and thus the libertarian impulses that drives Silicon valley caused a change in tune. Cyberspace was no longer a new world, declared independent with its own unalienable rights, it was now an untamed frontier, a wild-west where spooks and cypherpunks do battle and your worth is measured by your crypto slinging skills and operational security. Rather than united denizens of a new terrain, we are now crypto individualists homesteading in hostile territory.

This, as Seda Gurses argues, leads to Responsibilization, “Information systems that mediate communications in a way that also collects massive amounts of personal information may be prone to externalizing some of the risks associated with these systems onto the users.”

Users themselves are responsible for their privacy and safety online. No more unalienable rights, no more censorship resistant mass networks, no more expressing beliefs without fear of being silenced. Hack or be hacked.

Since libertarian ideology is often at odds with social solutions, holding private enterprise as an ideal and viewing private provisioning as best, the solutions presented are often pushing more entrepreneurship and voluntarism and ever more responsibilization. We just need a new start-up, or some new code, or some magical new business model! This is what Evgeny Morozov calls Solutionism, the belief that all difficulties have benign solutions, often of a technocratic nature. Morozov provides an example “when a Silicon Valley company tries to solve the problem of obesity by building a smart fork that will tell you that you’re eating too quickly, this […] puts the onus for reform on the individual.”

Karl Marx makes a similar argument in Eighteenth Brumaire of Louis Bonaparte:

“The proletariate […] gives up the task of revolutionizing the old world with its own large collective weapons, and, on the contrary, seeks to bring about its emancipation, behind the back of society, in private ways, within the narrow bounds of its own class conditions, and, consequently, inevitably fails.”

Solutionism underestimates social costs and assumes that social issues can be solved by individuals and private interests, and some may be, but where universality, equality and fairness need to be provided regardless of skill or wealth this is not the case. These sorts of things can only be provided socially, as public goods.

Many Hackers have always known this. In a excellent Journal of Peer Production essay Maxigas quotes Simon Yiull:

“The first hacklabs developed in Europe, often coming out of the traditions of squatted social centres and community media labs. In Italy they have been connected with the autonomist social centres, and in Spain, Germany, and the Netherlands with anarchist squatting movements.”

Early hacklabs didn’t view their role as being limited to solutionism, though hackers have alway helped people understand how online communications works and how to use it securely, hackers where embedded within social movements, part of the struggle for a fairer society. Hacker saw themselves as part of affinity groups fighting against privatization, war, colonialism, austerity, inequality, patriarchy and capitalism, they understood that this was the way to a new society, working shoulder to shoulder with mass movements fighting for a new society, and that here their knowledge of networks and communications systems could be of service to these movements.

Yet, as Maxigas goes on to argue,, “hackerspaces are not embedded in and not consciously committed to an overtly political project or idea.”

Instead, hackerspaces often focus on technological empowerment, which is certainly beneficial and important, but like community health centers that teach health maintenance practices are beneficial, they can’t solve larger social issues, such each-one-teach-one projects can not, on their own, solve social issues like privacy or health.

Hackers need to understand that there is no business model for secure mass communications. In order to achieve a society where we can expect privacy we need more hackers and hackerspaces to embrace the broader political challenges of building a more equal society.

Crypto Santa, Onion Gift Wrapping and Postal Remailing

Today I’m releasing the first public beta of Crypto Santa.

It’s planner for organizing Secret Santas that scrambles your list of participants, providing instructions for each person to wrap their present three extra times, making it difficult to find out who is giving gifts to who. Anonymity through Onion Wrapping!

I was thinking about onion routing and old-school postal remailing services and, well, it being nearly christmas, Secret Santas

Onion routing is the technique Tor uses to anonymize web browsing. Postal remailler’s are services that would accept mail and then forward it to another address. There are many reason’s these are used, especially remaining anonymous, when you don’t want the received to know who the sender is, and perhaps if you’re in country where a vender refused to deliver, due to their own delivery policy or export regulations.

I quickly realized that Secret Santas where a great opportunity to get people to talk about, well, secrecy, or more to the point privacy and anonymity. Crypto Santa is a Secret Santa for the post-Snowden era. Before I knew it, I was madly coding the site and getting my colleagues to help with graphics, copy, testing, cramming to get the site ready for Christmas, and viola, here it is:

http://cryptosanta.xyz

In addition to being a way to introduce questions of Internet privacy and surveillance to holiday parties, and helping people understand how Onion Routing works by doing it with wrapping paper, Crypto Santa plays with interesting possibilities.

What could be done in the future versions of Crypto Santa? Imagine if we added mailing addresses, and it was not just a party that met once, but a club of sorts, kind of like Mail Art communities, and imagine if recipients where not randomly chosen, but chosable by the sender. You would have a pretty interesting system of anonymous postal delivery. Of course, such a system should not be a centralized web-based system, but based on peer-peer software, like Tor is. It also be interesting to actually encode each label on each layer of wrapping so that only the addressee can decode the next address in the circuit. Today, Crypto Santa is just a great way to get people talking.

If you’re planning a secret santa with your friends, famility or colleagues, take it up a notch and try Crypto Santa! And spread the word, if we’re going to really find a solution for the issue of Internet surveillance we need people to get together and talk about it, especially where they normally wouldn’t, like for instance your office Christmas party!

It’s the first release, so expect bugs, when you encounter one, try refreshing the page, and please tell us about it. As well as types, copy errors and what new features we should add.

Today is Stammtisch, I’m Toronto, so I’ll be at The Cafe Pamenar at 307 Augusta Ave in Kensington Market, come and say hi!

Merry merry. Happy Crypto Santa!

In Soviet Russia we had spies, informants and bugs, in Modern Internet we have pageTracker._setVar() and message bidRequest {}

Though in principal, I’m against the use of “Soviet Russia” or “The Stasi” as shorthand for state surveillance, as this is just propaganda that seeks to portray USSR-aligned intelligence as a disproportionate response to western intelligence acvtivity, which they understate and characterizing as heroic rather than sinister, but something about reading the documentation for Google’s realtime-bidding protocol and good old Yakov Smirnoff’s voice popped into my head saying “In Soviet Russia we had spies, informants and bugs, in Modern Internet we have pageTracker._setVar() and message bidRequest {}”.

When we think of surveillance we imagine trench coated creepers with holes in their newspapers spooking around train stations. Bugs installed in wall clocks. Glaring bureaucrat bullying incriminating testimony out of hapless informants. Yet, for the modern advertiser who wants to know the location, browsing habbits, gender, and other demographic data of Web users, this information magically comes to them in a bidRequest message when they participate in an online advertising auction.


message BidRequest {

[…]

// A hyperlocal targeting location when available.
//
message Hyperlocal {
// A location on the Earth’s surface.
//
message Point {
optional float latitude = 1;
optional float longitude = 2;
}

// The mobile device can be at any point inside the geofence polygon defined
// by a list of corners. Currently, the polygon is always a parallelogram
// with 4 corners.
repeated Point corners = 1;
}

message HyperlocalSet {
// This field currently contains at most one hyperlocal polygon.
repeated Hyperlocal hyperlocal = 1;

// The approximate geometric center of the geofence area. It is calculated
// exclusively based on the geometric shape of the geofence area and in no
// way indicates the mobile device’s actual location within the geofence
// area. If multiple hyperlocal polygons are specified above then
// center_point is the geometric center of all hyperlocal polygons.
optional Hyperlocal.Point center_point = 2;
}

// Hyperlocal targeting signal when available, encrypted as described at
// https://developers.google.com/ad-exchange/rtb/response-guide/decrypt-hyperlocal
optional bytes encrypted_hyperlocal_set = 40;

// The offset of the user’s time from GMT in minutes. For example, GMT+10 is
// timezone_offset = 600.
optional int32 timezone_offset = 25;

// List of detected user verticals. Currently unused.
repeated int32 user_vertical = 30 [
packed=true
];

// Demographic data provided by the publisher.
message UserDemographic {
enum Gender {
UNKNOWN = 0;
MALE = 1;
FEMALE = 2;
}
// Gender
optional Gender gender = 1 [
default = UNKNOWN
];

// Age interval
optional int32 age_low = 2 [
default = 0];
optional int32 age_high = 3 [
default = 999];
}
optional UserDemographic user_demographic = 50;

What intelligence agency would not smack their lips at the prospect of that kind of dossier! In real time! And where are the informants that collect all this data? Not the KGB nor the NSA, but google who uses cookies and other techniques to track your browsing, and it’s publishers, who include code snippets in their Google Analytics code to snitch on their visitors.

pageTracker._setVar(‘Male18-24′);

Millions and millions of websites use google analytics and similar services, and pass information like the above about you to advertising platforms like google.

To channel Yakov again, “In America, spies don’t need to spy, we spy on ourselves to help us shop! What a country!”

Miscommunication Technologies with @dmytri & @baruch at @BerlinAtonal

Baruch Gottlieb and I will be giving a talk about Miscommunication technologies at Berlin Atonal today.

http://www.berlin-atonal.com/

Below is a text written by us about the series of artworks originally published in “Disrupting Business,” Edited by Tatiana Bazzichelli & Geoff Cox and published by Autonomedia – Data Browser 05.

http://networkingart.eu/2013/10/disrupting_business/

Miscommunication Technologies
Telekommunisten Artworks 2009-2013
Dmytri Kleiner, Baruch Gottlieb

The development of communication technologies is not merely a neutral process driven by discovery, progress and innovation, but an intensely social and political process where choices are made in ways that fundamentally influence the reproduction of the class conditions of the societies that produce these technologies. Communications technologies embody and perpetuate the social relations of their mode of production.

The Miscommunication Technologies series of artworks by Telekommunisten explore these social relations by creating technologies that don’t work as expected, or work in unexpected ways. The artworks in the series allow the embedded social relations to be critically experienced and confronted. The series employs parody, juxtaposition, exaggeration and reductio ad absurdum to bring aspects of these relations which are normally hidden from view, into the foreground.

The Miscommunication Technologies artworks illustrate some of the real world challenges faced by anyone or any group which would like to challenge the dominance of capitalist models of production. Miscommunication Technologies take a light-hearted approach to an intractable reality: capitalism is not only the system by which maximum value is extracted from social production, it is also the current global system which, in its unsatisfactory yet somewhat reliable manner, provide vital services we depend on every day. Any challenge to capitalist hegemony must be prepared to provide for the same social needs which will persist any system.

The illusions of the early Internet as a panacea platform for the emancipation of human intelligence and collaborative spirit emerged because it was financed for use-value, not exchange-value. It’s early developers were universities, NGOs, hobbyists and, prominently, the military. The contributors to the early Internet built the platform according to what could be seen as a product of a communist credo, “from each according to ability, to each according to need.”

As Richard Barbrook described in “The::Cyber.Com/munist::Manifesto” “Within the Net, people are developing the most advanced form of collective labour: work-as-gift.” Information and software spread freely across the network. This, to many people, created the impression that a new society was emerging, for instance, “The Declaration of the Independence of Cyberspace” by John Perry Barlow stated “We are creating a world where anyone, anywhere may express his or her beliefs, no matter how singular, without fear of being coerced into silence or conformity.” Barlow’s EFF co-founder, John Gilmore claimed that “The Net interprets censorship as damage and routes around it.” implying that The Net existed beyond the jurisdiction of States, or even the organisations that operate it, as it can simply “route around” those that would seek to interfere with the freedom of exchange on the network.

This might have held true to some extent during the initial stages of commercialisation of the Internet, since the first commercial ventures, “Internet Service Providers” or “ISPs” did not develop their own communications technologies, but only provided access to the public Internet, and the decentralized, open technologies that ran on it, such as email and usenet. The exchange value these ISPs were capturing was collectively created. Each ISP was independently earning income by being a part of a common platform, not owned by anybody as a whole, but composed of the mutual interconnections of the participants. Though made up of parts owned by public and private organizations, the platform as a whole functioned as a commons, a common stock of productive assets used independently by the ISPs and their users.

In parallel to the Internet, “Online Services” like CompuServ emerged from the capitalist imagination, they were financed for exchange value, by profit seeking investors, and as such did not employ a mesh topology like the Internet, but rather employed a star topology. Users could not communicate directly with each other, but only through the central servers of the operator, which could not be “routed around.” This was required by profit-oriented business models, since control of user interaction and user data is required to monetise the platform, for instance by charging fees or selling advertising.

Part of what fed the illusion of the emancipatory potential of the then-possible Internet was the fact that the platform made Capitalist-funded “Online Services” like CompuServ and AOL obsolete. This happened largely because of the explosive growth made possible by it’s distributed infrastructure, allowing the ISP industry to develop as a kind of petit-bourgeois industry of small producers. ISPs were a cottage industry of mom and pop telecoms of sorts. The design of the Internet allowed anybody with a connection to the internet, to provide a connection to others, thus the barrier of entry to becoming an ISP was relatively small, just an upstream connection, some computers, modems and telephone lines.

During the early days of the public Internet the communistic petit bourgeois ISPs prevailed over the feudalistic haute bourgeois Online Services, making it seem momentarily that the superior technical architecture of the Internet, combined with the cultures of sharing and gift economies would be able to surpass and even transcend Capital.

Both personal and commercial users migrated en masse to the internet. For instance, in a letter to their customers that is still available online the software company BASIS international, “The Big Little Software Company,” writes “By the end of 1997, BASIS plans to move completely off CompuServe (CSi) and onto the Internet. This is a logical consequence of the many changes that have taken place in the online world over the past few years.” In their letter, BASIS spells out a lot of these changes; “While our CSi presence has served the company well in the past, its pay-to-access structure is increasingly harder to justify with the Internet providing almost limitless content at a negligible incremental cost. People are moving away from CSi in significant numbers, making it a less effective platform from which to address our current and future customers. We believe that moving our existing support infrastructure from CSi to the Internet will give us better access to our customers and our customers better access to us” and goes on to explain how it will now use open platforms like email, Usenet and IRC instead of CompuServ’s proprietary and centralized applications. How ironic that now web 2.0 platforms have companies and individuals returning to centralized, proprietary systems for their support and communications. The reason for this is not because centralized platforms where superior all-along, but because they are they only kind of systems that are funded by capitalists.

While ISPs invested in bringing Internet access into households and offices worldwide, they did little to actually develop the communications platforms used on the network, these were largely developed within the gift economy of the users themselves. The ISPs were even less able to take over the provision of long-haul data transmission, dominated by international telecommunications conglomerates. Most ISPs got their start by simply connecting shelves full of consumer grade modems to consumer grade computers running free software, providing connectivity to an upstream internet provider for end-users who were using freely available communications platforms.

Thus, while the emergence of the ISPs and the rapid mainstream adoption of the Internet were spectacular, they were not able to capture enough profit to scale up and take over the more investment-heavy infrastructure of Internet provision. The end was already apparent in beginning. Well-financed telecommunications conglomerates would soon replace the mom & pop ISPs, either buying them up, or driving them out of business by providing “broadband” services which delivered internet to the home along with telephone service, leaving the remaining ISPs as just resellers, providing service over telecom managed circuits.

As Internet usage grew, technically-oriented users became the minority. The general Internet user became what Clay Shirkey eventually called “everybody”. This had a significant impact on the culture of sharing and tolerance. The first wave of “everybody” to arrive was when AOL, in an effort to remain relevant, allowed it’s users to access the Internet, this epoch has been called “The Eternal September” since then. The Jargon File, a glossary of hacker slang, describes this as “All time since September 1993. One of the seasonal rhythms of the Usenet used to be the annual September influx of clueless newbies who, lacking any sense of netiquette, made a general nuisance of themselves. This coincided with people starting college, getting their first internet accounts, and plunging in without bothering to learn what was acceptable. These relatively small drafts of newbies could be assimilated within a few months. But in September 1993, AOL users became able to post to Usenet, nearly overwhelming the old-timers’ capacity to acculturate them; to those who nostalgically recall the period before, this triggered an inexorable decline in the quality of discussions on newsgroups.”

The Jargon File, mentions “Netiquette,” a quaint term from the innocent times of net.culture, yet Netiquette was not simply a way of fitting-in, like table manners at an exclusive dinner party. The cultural context of that Internet that made acculturation necessary was its relative openness and lack of stratification.
Netiquette was required, because the network had relatively little constraints built into it, the constraints needed to be cultural for the system to work. There was much more to this culture than teaching new users how to not abuse resources or make a “general nuisance of themselves.” Netiquette was not so much about online manners, it was rather about how to share. Starting from the shared network resources, sharing was the core of the culture, which not only embraced free software and promoted free communications, but generally resented barriers to free exchange, including barriers required to protect property rights and any business models based on controlling information flow.

As dramatic as the influx of new users was to the old-timers’ net.culture, the influx of capital investment and its conflicting property interests quickly emerged as an existential threat to the basis of the culture. Net.culture required a shared internet, where the network itself and most of the information on it was held in common. Capital required control, constraints and defined property in order to earn returns on investment. Lines in the sand were drawn, the primitive communism of the pre-September Internet was over. The Eternal September began, and along with it, the stratification of the internet began.

Rather than embracing the free, open, platforms where net.culture was born, like Usenet, EMail, IRC, etc, Capital embraced the Web. Not as the interlinked, hypermedia, world-wide-distributed publishing platform it was intended to be, but as a client-server private communications platform where users’ interactions were mediated by the platforms’ operators. The flowering of “Web 2.0” was Capital’s re-engineering of the web into an internet accessible version of the online services they were building all along, such as the very platforms whose mass user bases where the influx that started the Eternal September. CompuServ and AOL most notable among them.

The gift-economy model of software development that developed platforms like email and usenet was unable to compete with a quickly growing Venture Capital start-up scene pushing Web 2.0 platforms. Like the profit-oriented Online Services before them, these start-ups were also compelled by the the profit motives of their investors to implement a centralised topology, a star topology, because once again, the central control of user data and interaction was required to monetise the platforms. We have moved from a world CompuServ and AOL to a world of Amazon and Facebook. Scratch off the Facebook logo and you’ll find the CompuServ logo underneath.

The OCTO P7C-1 prototype premiered at transmediale 2013 was produced by Jeff Mann, Jonas Frankki, Diani Barreto, Baruch Gottlieb and Dmytri Kleiner with raumlaborberlin. OOCTO exemplified this problematic. OCTO, the fictional venture capitalist start-up promised to build the next dimension of the Internet, the physical dimension of communication through a pervasive pneumatic tube network. The utopian rhetoric of the OCTO boosters is exuberantly cliché, promising all manner of human empowerment and positive transformation, and conveniently leaving behind in the shadow of bold promises the fact that this technology will be completely centralised and completely transfused with invasive security and monitoring technologies.

OCTO P7C-1 presented the situation on several parallel levels. First, the actual working prototype, the P7C-1 allowed visitors to send capsules around the entire Haus der Kulturen der Welt. The P7C-1 stations were integrated everywhere at transmediale and used by staff and visitors alike. Use of the system was purposefully complicated, every capsule having to be sent through a central station in coordination and at the mercy of the operators positioned there. P7C-1’s cumbersome, labour-intensive and privacy-agnostic factuality flew in the face of the transcendent promises unflaggingly issued from the fictional directorship of the fictional OCTO company. The constant work of managing the central station, end-stations and tube network is labour theatre, unlike the internet where the physical labour is hidden, the labour in OCTO P7C-1 is presented as a central theatrical aspect of the work. OCTO the company, provided the second layer, the social fiction, constantly driving home the lesson that there is a price for the convenience of every new technological utopia under capitalism, and the price will be extracted from those who are promised to benefit.

We have moved from administering our own email to using the centralized email services of giant entities like Google and Yahoo, which, as part of their mere functioning, parse and analyse private contents. Massive data sets have proven as useful for optimising AI applications such as automatic translation as any improvement from the (academic) information science community. Access to these storehouses of real-time contextual semantic data is the nec plus ultra of contemporary web profit models.

The revolutionary Internet that inspired Barbrook, Barlow, Gilmore and many others has become a dystopia, a platform whose capabilities and pervasiveness of surveillance and behavioural conditioning and influence surpass the wildest dreams of the tyrants and technocrats of previous eras. As we will see again and again, despite claims that culture and economy has gone ‘immaterial’, the rules of access to the physical technology of the internet conditions the forms of services which are eventually at the disposal of users.

Whereas OCTO is the archetypal network startup with a unabashed agenda of market sector conquest, Thimbl appears as the light at the end of the long dark tunnel of centralised hegemonic corporate dominance of the Internet. Developed by Dmytri Kleiner, Jonas Frankki, Rico Weise and Mike Pearce with contributions from a small community that developed around it, including Anthony Shull, Silja Neilson, Mark Carter and Fernando Guillen ,Thimbl is made out to be a distributed, peer-to-peer alternative to microblogging platforms such as Twitter. Thimbl appears as an analogue of projects like Diaspora, also launched in 2010 like Thimbl, Diaspora is a purely altruistic project with no profit motive and only the idealism of freedom of information.

The tragedy of projects like Diaspora is that they are not really a viable replacements for capital-funded projects like Facebook, for economic and political reasons, not technical reasons. Therein lies the message of Thimbl. Anyone who has some understanding of the elementary server architecture of the Internet can use Thimbl, because it is based on a protocol originally developed in the 1970’s called Finger which allowed users to post public “project” and “plan” messages akin to status updates. The free-access, non-commercial functionality of finger harkens back to the period when the Internet was still being developed for use value. By retrieving finger, Thimbl indicates how users today are allowing corporations to benefit from the value of their social interactions for services which, in principle, could be used freely and for free. Thimbl shows that all that is necessary to provide a microblogging experience like Twitter is available for free and built in to the Internet right now, but, precisely because they are freely available, technologies based on protocols like finger will never be developed to the extent that they offer the satisfactory user experience of competitive commercial platforms.
Unlike the highly centralized OCTO, capital will never fund a project like Thimbl because it will not generate sufficient ROI. Thimbl is an economic fiction or social fiction. Making it work is not the greatest challenge, making it financially viable is. Thimbl does not provide investors with the ability to control it’s users or their data, and as Thimbl’s Manifesto states “This control is required by the logic of Capitalist finance in order to capture value. Without such control profit-seeking investors do not provide funds.”

For Thimbl, or any other platform with a similar vision, to become a real alternative to the capitalist financed platforms like Facebook and Twitter, we need more than running code, even more than a small, perhaps dedicated, user base. To get beyond this and actually break the monopolizing grip of centralized social media we need to match their productive capacities. We need financing on a similar scale. so that the development, marketing, and operations budgets are comparable and sufficient to compete. Just like science fiction becomes reality when science transcends the limitations that existed when the fiction was imagined, for economic fiction like Thimbl to become reality society will need to transcend the political and economic limitations that we currently face. We can write code, we can write texts, we can create artworks, but as a small network of artists and hackers, we can’t change the economic conditions we work in by ourselves.

Free, distributed platforms are very practically suited to the work of radical communities, both symbolically as a matter of solidarity, and also practically, since support for privacy and cryptography is often desirable. These platforms should, in a meritocratic economy of technological product, become prevalent, but instead they are marginalized by the current ‘owners’ of the Internet. Free, distributed platforms cannot provide the same ease-of-use, the so-called user-experience (UX) provided by capitalist platforms because they simply lack the work-time to generate such quality. The result is that radical programmers pride themselves on the superiority of the software and bemoaning the state of things which prevents that such software become prevalent. Radical programmers are motivated to campaign on the level of code for a freer, anarchist, egalitarian Internet, but they are not motivated to confront the political and economic realities which prohibit the social adoption of these technologies. This generates much frustration and defensiveness, rather than the commitment to dedicate some small quanta of their formidable imaginations and intelligences to the problem of ownership.

Miscommunication Technologies show-up the improvisatory economic structures of network-optimism in the way they inevitably ‘fail’ to deliver the seamless networked experience they provocatively advertise. The schism between the promise of utopia and the reality of a system which requires much spontaneous effort on the part of users even to provide a modicum of functionality, playfully points to the immense work still needed to produce conditions which will support a radically different model of industrial communications as it prioritizes the generation and cultivation of direct interpersonal engagement between a community of users.

General concern regarding the censorship and surveillance on commercial online platforms is growing, and these concerns are opportunities to to introduce political topics by arguing that these features are not unintended side-effects of these platforms, but central to their business models, and that platforms that do not surveille or control can not and will not be financed by capital, but only by collective or public undertaking as an expression of priorities which diverge from capitalism. Once this becomes clearer, concern over privacy settings on Facebook can be directed towards capitalism itself, instead of the idiosyncrasies of that platform or it’s founders.

Privacy and surveillance, at the same time, become wedge issues to de-legitimize alternative networks and services for the general public. Under the banner of security and ‘quality’, corporations have lobbied governments to favour centralized ‘unfree’ network applications built on the still free but ever fading-from-view Internet. We have seen often enough how products like Bitcoin can be impugned to ‘enable elicit activity’, cast as disreputable, until completely controlled and regulated by capital-concerned governments. Without acknowledging the systemic necessity, under the capitalist financing regime, of a centralized Internet, citizens’ legitimate concerns about corporate encroachment into private and personal spheres is co-opted to generate unfavorable opinions about technologies which could help disrupt the dominance of capitalist priorities of control.

It is worthwhile to re-emphasize that the Internet itself is not immaterial. The Internet is only accessible through hardware which needs to be built according to unfree and often unfair industrial production rules. The industrial production of electronics is a quintessentially capital-intensive undertaking requiring global flows of materials, which, under capitalism take place in extreme conditions of competition and extraction of labour value. Any challenge to how the Internet is run, or what it is available to be used for must also challenge how it is produced and reproduced.

iMine, an experimental art-app/game produced by Baruch Gottlieb in 2011 with Horacio González Diéguez and Cocomoya, prior to Baruch’s work with Telekommunisten, is now integrated into the Miscommunication Technologies series. iMine is a game that can be played on a smartphone building the reality of labour exploitation in the mining industries needed to produce the minerals required to make the device being used to play the game into the experience of playing the game. iMine does not try to make the gameplay enjoyable or directly educational but seeks to create an experience of bleakness and drudgery, true to that of the mine workers, not to entertain the user with the story of the mining, but evoke the experience of the miner. At the heating heart of the emancipatory digital device, are highly hierarchic systems of production and control. iMine is dismalware.

The gameplay is designed from the start to be stripped down to the mere basics Someone who wants to play first creates a new miner giving it a unique name and a country. After this simple registration the only thing left to do is repeatedly thrust the phone as if it were a shovel into the ground. The website keeps track of the global iMining action going on at any particular time, and also features an extensive resource section with information on mining and the political and economic enjeu in global supply chain for minerals necessary in portable computing device production. After having been developed and premiered at LABoral,

Miscommunication Technologies thus indicates that there can be no uniquely technological fix. Colonial wars and security states, corporate rule and centralization will persist despite the best intentions of emancipatory technologists, and worse, the best and most innovative technologies are not only appropriated to perpetuate capital but to this end they are incomparably better funded than had been the visionary projects of their emancipatory inventors. The technologies which become dominant, become dominant in the form dictated by the prevailing conditions of capitalist production under which we labour today.

The free, distributed platforms, that can not be controlled or censored, can not exist on any large scale under capitalism. Not for technical reasons, in fact the technology that enables such interaction is in many cases well-described and readily available, but for social and political reasons. The productive capacity that is required to build and support them will not be provided by Capital, thus so long as Capital is the dominant mode of production, it will produce platforms that reproduce itself, thus platforms than enable the accumulation of wealth by engineering control and extraction into communications systems.

R15N, originally developed as Jessycom by Dmytri Kleiner during a residency at the Israeli Center for Digital Art, was premiered as R15N in collaboration with Jonas Frankki, Jeff Mann, Baruch Gottlieb, Rico Weise and Mike Pearce at transmediale 2011. R15N is a project which pushes to absurdity the emancipatory rhetoric of mobile networked computing. Events like the antiglobalization protests in Copenhagen or the political upheaval often referred to as ‘arab spring’ generate much enthusiastic hyperbole about how new realtime networks employing mobile devices can become an unstoppable democratising force. R15N points to the economic predilections built into the provision of network connectivity may work against such emancipatory agendas.

R15N retrieves an obsolete form of social networking, the ‘telephone tree’ and dresses it up as the lastest thing in robust circumventionist networking. Perfect for planning a flash mob, R15N easily becomes a nuisance as phone calls multiply rendering the commitment one made to one’s community by joining the network a near-constant obligation to participate.

Whereas iMine proposes that critical games or critical media can only do so much to challenge the economic exigencies underlying an unacceptable status quo, and that the materiality (itself) of networked utopia is the key to understanding its injustices, R15N suggests that circumventionism will not fundamentally challenge intolerable social conditions without the concurrent care and effort being dedicated to actually building up strong communities which have committed to working together toward transforming society, as users of R15N are constantly reminded, the system depends on your competence and diligence.

Miscommunication Technologies are artworks with a principal purpose, that of engaging people in provocative networked experiences in which they inadvertently but necessarily confront the unadorned material and economic conditions under which such experiences are made possible.

Werkstatt TechOps Report #1

ThoughtWorks Werkstatt Berlin hosts many different working groups, including several Cryptoparties, The Kids’ Hacker Club, and the Marx-Engels Werkshau group. In order for the groups to plan and stay in touch with each other in between their meetings at Werkstatt, we have implemented Werkstatt Groups, an online discussion forum based on NodeBB.

Creating a discussion channel for Werkstatt is tricky, since working group participants range from Tor project contributors, who are very knowledgable and concerned about technology and privacy issues, to kids, to political activists, who have other interests and areas of focus, and may be still learning about technology and privacy issues. So the Werkstatt Groups platform needs to be something that is usable across the spectrum, to be a place where privacy experts and privacy novices can intereact online.

Looking at the options available, a simple web forum became the most reasonable choice. With the many working groups at Werkstatt, managing dozens of mailing lists seems unworkable. Usenet, alas, has become entombed behind paywalls, and is inaccessable to most people, except through untrusted interfaces like Google Groups. Platforms that offer groups functionality like Facebook obviously have privacy issues, among many others, and old favourites like IRC and Jabber are not particularly suitable for asynchronous group discussion.

So how to set up a web forum that respects privacy? Run it on a Tor hidden service!

Before I explain how this was done, I need to start with a disclaimer: Werkstatt Groups makes no guarantees of privacy or anonymity, Tor is designed to provide anonymity. However, identifying all the possible ways in which the software running the forum may leak information is not easy, so use caution and report any issues or potential issues to us.

There are two ways to access this site, the recommended way is Tor Browser. Downloading and installing Tor Browser Bundle takes seconds and ensures that all your browser traffic goes over Tor and that your browser doesn’t leak any information and is difficult to fingerprint.

Using Tor Browser, you can access Werkstatt Groups using this url: http://vgnx2fk2co55genc.onion. Note HTTPS is not used, this is because the connection is already encrypted by Tor.

The other way of accessing it is by way of the public URL, http://groups.werkstatt.tw, which links to HTTPS when you access the forum. This is a reverse proxy running on a different server than the one that hosts the hidden service, accessing the hidden service over the tor network, thus making the site publicly accessible outside of the Tor network by way of a public url, while at the same time not revealing the location of the hidden service.

The NodeBB platform itself is a very dynamic, responsive platform which makes heavy use of websockets by way of socket.io, this is very advantageous over Tor, as a request to a hidden service needs to traverse 6 different servers, making page loads very expensive. Minimizing page loads by way of websocket requests compensates for this.

However, NodeBB also has some drawbacks, the platform uses Gravatar and Google Fonts, and socket.io includes a Flash fallback option, so a small Flash object is loaded in the site. All these issues are fixable, and are on our isssues list, however the best way to defend against these kinds of issues is to use Tor Browser. This way, even requests to Gravatar and Google Fonts go over Tor, and potentially dangerous plugins like Flash are blocked. However, JavaScript running in the browser is always a security concern, as exploits are possible. Also, NodeBB is beta software in very active development, and we are running the bleeding-edge head-of-branch, so expect glitches and some downtime.

OK, OK, so with all that out of the way, here is how the setup works. If all you want to do is use the forum, just get started here: http://groups.werkstatt.tw, however if you want to know how the setup works, keep reading. This assumes a relatively expert knowledge of server setup, including node, tor, nginx and iptables.

NodeBB

NodeBB installation instructions
for various platforms are available. However Werkstatt Groups uses git and npm to install NodeBB. So the steps are:

– Install and run Redis
– Pull head
git clone https://github.com/NodeBB/NodeBB.git
– Enter the directory, i.e:
cd /srv/http/NodeBB
– Build dependencies
npm install
– run the setup script
./nodebb setup
– change bind address in config.json to local interface only
"bind_address": "127.0.0.1",
– If all that runs successfully, start NodeBB
./nodebb start
– Check the log
./nodebb log
This line should appear if all is good:
info: NodeBB is now listening on: 127.0.0.1:4567

Tor Hidden Service

– Install and run Tor
– Configure the Tor hidden service in /etc/tor/torrc

HiddenServiceDir /var/lib/tor/hidden_service/
HiddenServicePort 80 127.0.0.1:4567

– restart Tor and find out onion address

cat /var/lib/tor/hidden_service/hostname
vgnx2fk2co55genc.onion

Your onion address will be different, of course.

To reduce the chances of the server revealing it’s address to other services based on outbound requests, add iptables rules to ensure that requests that come from the server go over Tor, here is an example of REDIRECT rules in the nat table OUTPUT chain configuration on Werkstatt Groups

*nat
-A OUTPUT -p icmp -j REDIRECT --to-ports 9040
-A OUTPUT -s xxx.xxx.xxx.xxx/32 -p tcp -m owner ! --uid-owner 43 -j REDIRECT --to-ports 9040
-A OUTPUT -p udp -m udp --dport 53 -j REDIRECT --to-ports 5353

‘xxx.xxx.xxx.xxx’ is the IP address of the hidden service and 43 is Tor’s userid, this means that all requests that originate fom this ip address that are not Tor itself are redirected over Tor’s Transparent Proxy, which I’ve configured to run on 9040. DNS Requests are redirected over Tor. For good measure, ping is short circuited as ell.

– Enable DNSPort and TransPort in /etc/tor/torrc

DNSPort 5353
TransPort 9040

Restart Tor again and visit the onion address in Tor Browser, you will see your NodeBB forum! Hooray!

Reverse Proxy

The reverse proxy runs nginx and tor.

In order to set up the public https server we need to use a different server. The IP address of the hidden service should not be listed anywhere, so it can not be used in your DNS zone.

So on this other server
– install and run tor
– install and run nginx
– make an ssl certificate and set up an https server with nginx
– set up proxy_pass in nginx for your onion node with websocket support, i.e.

proxy_pass http://vgnx2fk2co55genc.onion/;
proxy_http_version 1.1;
proxy_set_header Upgrade $http_upgrade;
proxy_set_header Connection "upgrade";

Now we need to set up Tor to transparently proxy requests for Tor hidden services.

– In /etc/tor/torrc/

DNSPort 5353
TransPort 9040
VirtualAddrNetwork 10.192.0.0/10
AutomapHostsOnResolve 1

In addition to setting up the DNSPort and TransPort, The above code maps hidden services to the network 10.192.0.0/10, which we can then use in our iptables rules, as follows:

*nat
-A OUTPUT -p udp --dport 53 -j REDIRECT --to-ports 5353
-A OUTPUT -d 10.192.0.0/10 -p tcp -j REDIRECT --to-ports 9040

And viola! Once you restart Tor and nginx, if you navigate to your https server, you should see your hidden service!

Questions and comments very welcome!

Dear #NETMundial, Governance is cool and all, but we need to DEMAND IPv6 NOW! cc #OurNetMundial

Many of my friends and colleagues where in Sao Paulo last week for NETMundial, the Multi-stakeholder Meeting on the Future of Internet Governance. Dilma Rousseff, President of Brazil, convened this initiative to “focus on principles of Internet governance and the proposal for a roadmap for future development of this ecosystem.”

NETMundial was originally motivated by revelations from Edward Snowden about mass surveillance conducted by the US and UK governments, including spying on President Rouseff herself. These revelations prompted Mrs Rousseff to state “In the absence of the right to privacy, there can be no true freedom of expression and opinion, and therefore no effective democracy” in a speech to the UN at the 68th General Assembly.

Yet, as important as Internet governance is for our future, and as valuable any effort to address this is, it is unlikely to do much, if anything, about the right to privacy online. Why? Because surveillance is not an issue of Internet governance, but of the way the Internet is financed. The vast amount of consumer data amassed by private companies like Google, Facebook and Verizon is not the result of IANA or ICANN policy, but of the business models of these companies which seek to generate profits by way of this data. It is inconceivable that these companies could amass such vast amounts of consumer data, use it for marketing purposes, sell and share access to it with other companies, and yet, somehow keep it out of the hands of the NSA and similar intelligence agencies. Likewise, the extraordinary hacks, mods and exploits the NSA has conducted, as revealed by Snowden, would not be thwarted by any IANA regulation. Aggression by the US is not an Internet problem, and Internet governance can not do away with it, any more that it can do away with drone strikes and regime change projects.

Yet, there is lots that governments can do to ensure the right to privacy, and they can do so today, even absent any change in global Internet governance.

Governments have the ability to regulate the way Telecomms and Internet companies operate within their countries, indeed, the government is no stranger to creating regulation. Government regulation ensures buildings are built correctly, structurally sound, follow the fire code, etc. Governments create rules that make sure highways, roads, and sidewalks are used safely. Governments pass laws to prevent consumers from being defrauded, create statuary warranties, labour standards, regulate broadcast media, etc. Governments can pass regulations to protect the right to privacy. The idea that the Governments such as Brazil, Germany and the others participating in NETMundial need reforms to IANA and friends before they can work towards guaranteeing their own citizens’ right to privacy is absurd.

To guarantee the right to privacy, communication systems must implement the end-to-end principle, which states that functionality ought to reside in the end hosts of a network rather than in intermediary nodes. The term “end-to-end” principle was coined in a 1981 paper by J.H. Saltzer, D.P. Reed and D.D. Clark at the MIT Laboratory for Computer Science, “End-to-End Arguments in System Design,” in which they specifically address privacy.

In the section titled “Secure transmission of data,” the authors argue that to ensure “that a misbehaving user or application program does not deliberately transmit information that should not be exposed,” the “automatic encryption of all data as it is put into the network […] is a different requirement from authenticating access rights of a system user to specific parts of the data.” This means that to protect the users’ rights to privacy, it is not sufficient to encrypt the network itself, or even the platform, as this does not protect against the operators of the network, or other users who have access to the platform. What is needed, the authors argue, is the “use of encryption for application-level authentication and protection,” meaning that only the software run by the user on the end-node, or their own personal computer, should be able to encrypt and decrypt information for transmission, rather than any intermediary nodes, and only with the user’s own login credentials.

The end-to-end principle is a key concept in the design of the Internet itself, the underlying “Transmission Control Protocol,” one of the core protocols of the Internet protocol suite (TCP/IP), exemplifies the end-to-principle, and allows applications running on remote nodes to use the Internet for the reliable communication of arbitrary data across the network, without requiring any of the intermediary nodes to know or understand the purpose of the data being transmitted.

In principle, therefore, there is absolutely nothing technically stopping everybody from employing private communications on the Internet. So then, how do we get into this mess we’re in now? Why did the Internet, which has the end-to-end principle in it’s core architecture, become host to the most large scale mass surveillance in history?

Two reasons: Capitalism and IPv4. Let’s start with IPv4.

Internet Protocol Version 4 (IPv4) was created in 1981, the same year the Saltzer, Reed, and Clark paper was published. IPv4 provides approximately 4.3 billion addresses, which sounds like a lot, until you realize the every device that connects to the Internet needs at least one. Running out was not presumed to be a big issue at the time, as this version was originally presumed to be a test of DARPA’s networking concepts, and not the final addressing scheme for the global Internet. In 1981 4.3 billion addresses seemed like an awful lot, but when the public Internet began to take off in the Nineties, it became clear that this would not be nearly enough. In 1998 RFC 2460 was released, this document is the specification for IPv6, an addressing scheme that allows for a near limitless number of addresses, trillions of trillions for each person on earth. Yet, as NETMundial was taking place in Brazil, nearly 16 years since the protocol was invented, Google reports that about 3% of visits to its services use IPv6. The “World IPv6 Launch” site, which promotes IPv6 adoption, estimates that more than half Internet users around the world will have IPv6 available by 2018. In other words, 20 years after the design of the protocol, nearly half of all Internet users will not have access. It’s important to note that it is not hardware adoption that is holding things up, it’s highly doubtful that many device made in the last 10 years could not support IPv6, it’s rather that the owners of the networks do not configure their networks to support it.

As everybody knows, 20 years is effectively infinity in Internet years. With IPv6 a far away utopia, and with IPv4 addresses still the currency of Internet service, NAT was developed. The vast majority of devices available to users where not assigned public IP addresses, but only private ones, separated from the public internet by “Network Address Translation” (NAT), a system that allowed the sharing of public IP addresses by many end-nodes, this was an effective solution to IPv4 address exhaustion, but introduced a bigger problem, the network was no longer symmetric, software running on users’ computers can reach central Internet resources, but can not reach other users, who are also on private address space, without some intermediary service providing access.

What this means is that so long as users’ are on private address space, any communication system they use requires centralized resources to bridge connections between users, and what’s more, the scale of these central resources must grow in proportion to the the number of users it has. In order for the end-to-end principle to be respected, these intermediary services need to support it.

And this where we get to the Capitalism part: Building, maintaining and scaling these resources requires money. In the case of “web scale” platforms, lots of money.

By and large, this money comes from Venture Capital. As Capitalists must capture profit or lose their capital, these platforms require business models, and while many business models are possible, the most
popular today, the one presumed to be the most lucrative by investors, is big data. Thus, instead of respecting the end-to-end principle and engineering functionality into the end hosts of a network, capitalists instead only invest in applications where core functionality is built into the intermediary nodes, that can capture user data and control user interaction, which is how they make money.

Capitalist platforms grow and collect data around these intermediary nodes in the same way the mould grows around leaky pipes. In order to give alternative platforms that respect the right to privacy a fighting chance and rid the Internet of the mould of centralize data-collecting platforms, we must fix the pipes, we need to remove the asymmetry in the network.

We can not allow private initiative alone to push adoption of IPv6, and wait however many years or decades it takes to get it. If governments want to promote their citizens right to privacy, they need to mandate adoption of IPv6, to ensure their citizens are able to use software that respects the end-to-end principle.

Here is a charter of rights that all Governments can provide to their own citizens right now to promote the right of privacy:

– IPv6 connectivity with adequate public address space for all!
– At least one DNS Domain Name for every citizen!
– At least one Government signed SSL certificate for every citizen!

If each citizen had a public address space, a domain name and a signed certificate, the leaky pipes of the Internet could be fixed, the surveillance mould would dissipate, and new privacy-respecting applications could flourish!

DEMAND IPv6 NOW!

Anti-Capitalism: The Musical! and The Panel! // Galway, Ireland.

I’ll be in Galway, Ireland, this weekend, to particate in a panel to discuss the themes in “Anti-Capitalism: The Musical!”

The Panel is free to attend. All are welcome. Please pass it on.

Anti-Capitalism: The Musical!

Venue: The Cube, NUI Galway
Áras Na Mac Léinn
NUI Galway National University of Ireland, Galway
University Road
Co. Galway
Phone: +353 91 524411
URL: http://www.nuigalway.ie/

choreographed & composed by Deirdre Murphy

A fairytale of musical theatre, using circus, song, and dance to create a near-future political reality similar to our own. A large scale theatrical narrative with a cast of 15 dancers, singers, circus performers, actors and a live band, it uses acrobatics, plot twists, assassinations, and intrigue to tell the tale of a heroic group of everyday acrobats, aided by the Greek-chorus style narrations of a powerful triumvirate of fairy godmothers.
‘Social criticism wrapped in wit, glee and stunning physical & vocal performance’.

Anti-Capitalism: The Panel!
Sunday, March 30 at 2:30pm

A panel of speakers from various backgrounds lead an open discussion on issues raised by Anti-Capitalism: the Musical!

Since the financial crash of 2008 the world has entered into a new depression. While some in the top 1% have remained comfortable, the .01% of the richest in the population have become much richer. At the same time that the broad mass of the population is struggling to make ends meet. Unemployment, underemployment, emigration and the misuse of internships are serious problems. Climate change and other environmental problems which at one time seemed to be on the agenda have disappeared from public discourse.

We will explore what we can do and how to get involved in doing it.

Chair: Gavin Mendel-Gleason
panel speakers: Deirdre Murphy, Dmytri Kleiner, John Bissett

Information is like Uranium

This post is addressed to my friends and colleagues Prabir Purkayastha, Kiran Chandra and all the others participating in FreedomFest at Acharya Nagarjuna University today and Tomorrow!

Years ago I was visiting my old friend Cory Doctorow in London, and was lucky enough to catch a talk he gave during a CopyFight night in a local pub. He was speaking about an incident in the UK where the government lost a whole lot of personal data about UK residents, collected as part of a effort to create a large identity database.

Cory used a great analogy that has stuck with me, and is a useful way to think about information collection in general; Information is like Uranium.

Uranium, for the most part, is not dangerous at all. It’s a naturally occurring element, it exists everywhere, all around us, even inside of us, distributed far and wide in tiny amounts. There is no problem with uranium per se. Heck, for some plants uranium even appears to be a micronutrient, essential for healthy growth, like other vitamins and minerals.

However, when you have a lot of uranium in one place, when it’s collected in one place, concentrated and refined, when you have giant pile of refined uranium it becomes dangerous, very very dangerous. Doomsday scenario kind of dangerous. **KABOOM** kind of dangerous.

Information is the same, we share information all the time, even personal information, even through insecure channels like a casual conversation with a friend on a park bench, on the telephone, in the office with co-workers, and this is just fine. When information is defuse, casual, fleeting and everywhere it causes no harm. It is the main nutrient for human relationships and action.

However, when you have a lot of information, all in one place, concentrated and refined, tagged and categorized and cross referenced, it becomes very dangerous.

This is true for government databases, for social media platforms, communications systems, ecommerce platforms, even for Bitcoin exchanges. We’ve seen bad things happen many times with personal information being used for crime, surveillance, identity theft, fraud, etc. Over and over again we’ve seen information stockpiles putting peoples lives, finances, privacy and identity at risk.
.
Put a whole bunch of information in one big pile and sooner or later there will be tears

Yet, so many companies and institutions, so many projects and even individuals think nothing of collecting data indiscriminately, after all, it is argued, storage is cheap, and becoming cheaper, so why not just simply collect every bit of data you can grab. What the heck, even if you have no use for the data now, it might be useful later, why not just stockpile it and see what value can be squeezed out of whenever we get around to it. The data is perceived to have potential value, but not potential risk.

This is like arguing, what the heck, plastic bags are cheap, why don’t we just assemble all the Uranium we can get our hands on, snag it all, enrich it and concentrate it, and dump it in the basement, where’s the harm, sure, we may not have any use for enriched Uranium now, but who knows, it could be useful later!

Nobody would stockpile enriched Uranium in their basement just in case it might be valuable later, the risk would be considered too high. Similarly, no one should stockpile personal information without seriously considering the risks involved.

So, to all the amazing activists at FreedomFest, to the great community that I was lucky enough to meet some of in Hydrabad, when thinking about our campaigns against surveillance we must remember, that it’s not just a matter of kooky spies at places like the NSA illicitly collecting piles and piles of data thought deception and trickery, it is all of us, from the biggest abusers, companies like Google and Facebook, to our governments and institutions, to individual users, like all those Bitcoin users who thought that the right place to store a distributed crypto currency was in wallets hosted on giant centralized servers.

Stockpiles of information bear risks, often these risks far out way any “benefits,” since the same benefits can be achieved with secure distributed systems if we put our minds to it, except, or course the benefit to spies, crackers and criminals of having a whole bunch of juicy data all in one place.

So we need you at FreedomFest, our next generations of developers, of entrepreneurs, of activists, our future technologists, to take this knowledge and bring it to society broadly; Information is like Uranium, when it flows freely it is a nutrient, when it is contained and concentrated it is toxic, put too much of it in one place and eventually it goes **KABOOM** and people get hurt.

M-C-LOL: Circuits of value in the Lulz economy.

Neither free software, nor crowd funding will save us from capitalism. We can’t overthrow capitalism by undertaking work merely for the Lulz, we need to create new value circuits that allow is to build new means of survival for the planet, and only then can we do away with capitalism.

In the stages of capitalist production, the Capitalist comes to market twice. The first time as a buyer, the second time as a seller.

Marx described this as M – C – M’

In the first stage the capitalist buys commodities and labour time. In the second stage, the purchased commodities and labour time are put into production. The result is a commodity of more value than that of the elements entering into its production. In the third stage, the capitalist returns to the market as a seller; the new commodities are turned into more money.

As the capitalist winds up with more money as a result of the productive process, the capitalist can purchase more labour time and commodities and repeat the process again, and again.

Investing in production allows the capitalist to reproduce, increase and accumulate capital. This reproduction cycle is what makes capitalism a thriving, dynamic system, that expands.

This very process of capitalist production has many negatives, many of which extend from the inherent exploitation involved in making labour time into a commodity, many others from the practice of allocating productive assets in the interests of profit, instead of social good, still more from the dispossesion and enclosure required to create the social conditions for capitalist production.

Yet, capitalism sustains us. Despite it’s social costs, its factories and institutions provide the means of survival that the world depends on, even while it’s contradictions jeopardize our survival.

In order to transcend capitalism, we need to find ways to provision the means of survival differently. “Ending” Capitalism, before alternative productive strategies for survival are not only conceived, but actually existing on sufficient scale, would more likely lead to collapse and a new dark ages than it would a fairer and more sustainable society.

In order for any such alternative productive strategies to grow to a scale in which they could be a viable alternative to capitalism, they must, like capitalism be thriving, dynamic systems capable of growth. They need to be able to reproduce their productive inputs. Economic alternatives need to have sustainable value circuits to be truly viable.

Free Software as well as the goods financed by Kickstarter and similar sites seem like production, after all stuff is produced. One can use free software, just like one can consume a movie, book, album or novelty gadget funded by Kickstarter.

Yet, the way the creation of these goods is financed can not reproduce its inputs.

In the creation of free software and in the funding of Kickstarter projects, money to sustain the inputs comes from donation, either actual donation of money in the case of crowd funding, or in-kind in the form of free labour in the case of some free software. These donations and in-kind contributions are done voluntarily. Yet such voluntarist production is different from capitalist production.

M – C – LOL

Thus, like capitalists, voluntary producers, come to market twice. Fist time as buyers, the second time for the lulz. However, unlike capitalists their circuit is not completed, because the lulz do not enable them to be buyers again, do not allow for them to acquire the inputs they need to repeat such production.

Yes, in the case of Free Software, major corporations do provide funding, lots of it. This is when the Capitalist is coming to market as a buyer, not a seller. Thus it is capitalist consumption, they don’t need to make a profit from Free Software directly, they use it in their production process and make money when they return to the market with the resulting product, which is distributed for more money, not lulz.

The source of this money is not a new mode of production, but capitalism. It’s simply part of the investment capital must make in its means of production, it is consumption not production.

And yes, recipients of Kickstarter financing can use such financing to make money, but such income does not flow back to those that donated the funds in the first place. The donors, for the most part, need to go back to work to get another paycheck before donating again. Thus the money comes from their Capitalist employers and is spent out of their “disposable income,” in other words, once again it is consumption, not production.

Both free software and crowd funding are simply novel forms of distribution within the capitalist mode of production, and therefor not a new mode of production that could potentially disrupt capitalism.

In order to transform these practices into genuinely revolutionary forms, we must collectively own the means of production so created, so not only must the software be free, but we must collectively own the wealth that results form using the software in production. We must collectively own the products produced by crowd funding, so that we can use the wealth created to reproduce the cycle, again, and again.

So long as our free labour earns only lulz in return, Capitalism has the last laugh.

I’ll be at Cafe Buchhandlung tonight around 9pm or so, come by if you’re in time, hope we have lots of surprise guests still hanging around Berlin after transmediale.