Abstract
Introduction
In the wake of the Snowden revelations, the question of how data is used for security purposes has re-emerged as a political problem. Critical inquiries around data and security are not new: from the production of traceability (e.g. Bonditti, 2008), proactive and pre-emptive management through data analytics (Aradau and van Munster 2011; Bigo, 2008; de Goede, 2012), database management, algorithmic governance or dataveillance in the ‘war on terror’ (Amoore and de Goede, 2005; Raley, 2013), the literature on security practices has analysed the multiple facets of the social and political transformations that the proliferation and increased use of data have entailed. The Snowden revelations and the emergence of ‘Big Data’ have rekindled these debates and prompted new inquiries into how digital practices and Big Data devices are deployed for the purposes of security and with what political effects.
These transformations have so far been analysed as part of the ‘computational turn’ in security governance, with data mining, predictive analytics and algorithmic decision-making playing an important role in the digital transformation of security (Amoore, 2011; Rouvroy, 2012). Big Data is an indicator of the transformations that digital information has brought about by being ‘too big to know’ (Weinberger, 2011). These dynamics have far-reaching implications for privacy and data protection, alongside civil liberties and human rights, which appear to be most at stake (Bauman et al., 2014; Lyon, 2014; Pasquale, 2015). 1 Therefore, critical scholars have concentrated mostly on social, political and legal challenges to Big Data and have paid less attention to the controversies around concepts and devices in computer and information science. The debates about the transformations that are underway have often tended to embrace particular dominant representations of this perceived computational or, more recently, Big Data ‘revolution’ rather than attending to the contestations and controversies about computing and digital knowledge formulated in these disciplines.
This paper investigates how the supposed ‘novelty’ implied by a digital transformation of security practices has been particularly put to use by security and intelligence experts in order to justify an urgent need for novel responses to anticipate and pre-empt the ‘next terrorist attack’. For security professionals, Big Data, in particular, stands for the promise of solutions to contemporary security problems. Here, they are not so different from other institutional actors and strategies, from the development of e-science to new forms of e-government or a renewed focus in commercial organizations on their data assets. Critical scholars have similarly embraced this discourse of novelty in which Big Data is a ‘game changer’ (Hildebrandt, 2013: 8). We contend that these assumptions of novelty of a Big Data ‘revolution’ and game change in science – and by extension in governance practices – limit the potential of critical engagement. In analysing an emerging Big Data-security assemblage which brings together heterogeneous modes of knowledge, devices, institutions and methods, we address the impasses of critical discourses as formulated by civil liberties activists as well as critical scholars of security and Big Data.
Although the methods used by intelligence agencies like the National Security Agency (NSA) and Government Communications Headquarters (GCHQ) are secret, we start from the assumption that it is unlikely that their methods would be widely different from the state of the art in computing and information science, and the practices developed in dealing with Big Data in academic organizations and commercial institutions. In this sense, the paper understands intelligence agencies as Big Data organizations that employ data-driven methods to anticipate future dangers. A collaboration between social and computer scientists, as this paper proposes, can help go beyond the inscrutability of algorithmic methods in security practices.
In so doing, this paper develops a contribution to critical data studies (Kitchin, 2014) and critical approaches to security and surveillance (Amoore, 2014; Bauman et al., 2014; Lyon, 2014). We propose to take seriously the critical knowledge developed in information and computer science and reinterpret these debates to offer a critique of the common representations by security professionals of the digital transformation of their practices. The paper makes three moves that recast existing critical engagements with data-driven security: from data/metadata distinctions to the production of data as a complex epistemic entity; from a ‘computational turn’ in surveillance to the division of labour between humans and computers in socio-technical assemblages; from an underlying logic of algorithms to algorithmic practices and methods in security analytics.
These moves are developed through the analysis of three sites of public controversy about NSA and GCHQ surveillance and the use of Big Data for security governance in the wake of the Snowden revelations. The first controversy concerns the definition of digital data, and particularly the distinction between metadata and content, which has been used repeatedly by security professionals to justify bulk data collection. The second controversy we discuss relates to mass surveillance and the relationship between humans and machines in justifying surveillance practices. Here, a ‘computational turn’ in surveillance is used by security professionals as a reassurance that no privacy is invaded as only computers look at bulk-collected data. A third justification that security professionals have promoted is that they need to collect everything and make data big for algorithms to develop anticipatory knowledge of the ‘next terrorist attack’. Contra these discourses of novelty, we show how debates in computer and information science can be mobilized to challenge the security professionals’ claims to credible knowledge.
(Meta)data and the remaking of security knowledge
Since the Snowden revelations, a new concept has entered the public vocabulary: metadata. Long used by archivists and computer experts, metadata has more recently been at the heart of controversies about security practices. President Obama argued that the NSA programme was not gathering data but metadata, namely how long a call was or where it was made from. He reinforced that metadata collection would then be different from surveillance and proceeded to allay the public’s fears: ‘Nobody is listening to your telephone calls. That’s not what this program is about’ (Obama, 2013).
Metadata is therefore seen as not encroaching upon rights or privacy, as it does not reach content. In the UK, Theresa May has also claimed that privacy concerns only come into effect ‘at the point at which the communication is opened’ (Wheeler, 2014). Metadata is supposed to not convey information about what people say or do in their homes. In that sense, metadata is rendered as the opposite of the telephone tap and the secret agent listening in for revealing clues. It is also the opposite of the camera, the extended CCTV-surveillance that appears to pry into the intimate and intricate details of everyday life. The content/metadata distinction therefore justifies the practices of intelligence agencies, as ultimately they are deemed to be using a qualitatively different form of data.
Yet, this benign reading of metadata has been challenged by critics, who have argued, on the one hand, that metadata is data about content and that the uses of metadata were no more devoid of knowledge than data is, on the other. For instance, Edward Snowden has pointed out that ‘[m]etadata is what allows an actual enumerated understanding, a precise record of all the private activities in all of our lives’ (Plume, 2014). In an
Indeed, this distinction between metadata and content has been contested by critical scholars and civil liberties activists alongside whistle-blowers and computer scientists. However, we contend that there are limitations to the critical arguments that metadata = data = knowledge. In information science, the field of knowledge engineering covers the transformation of content into data that computers can process. For knowledge engineering, content is really anything that can be expressed digitally like any video, software, text, or audio. In our case, content is, for instance, a phone conversation. Knowledge engineering sets out to discover all computer-actionable data in this content. Communications (meta)data such as location and time is an important part of this data. Computers prefer (meta)data to content, as the former is structured and semantically defined for their processing. The distinction between content and (meta)data, which is invoked and justified by the professionals of security and politics, relies on an implicit hierarchy of knowledge production that underlies the field of knowledge engineering. In computing and information science, this hierarchy is generally referred to as data-information-knowledge (DIK) and is widely discussed as one of the foundational issues of these disciplines. The taxonomy of data-information-knowledge starts with ‘raw’ data and systematically builds information and finally knowledge.
For the purposes of our argument, what counts is the fact that professionals of security and politics relegate (meta)data to the bottom of the supposed DIK hierarchy. This has important implications, as it might mean that the collection of data in all its forms is regarded as irrelevant and ‘encourages the mindless and meaningless collection of data in the hope that one day it will ascend to information – pre-emptive acquisition’ (Frické, 2009: 136). In this sense, the DIK hierarchy enables Obama and others to claim that the bulk collection of metadata is not politically significant and only becomes problematic once it has been transformed into information and knowledge. Data becomes a raw and neutral material ready to be mined.
As the hierarchy of DIK is translated into the political realm, security professionals take up what is an essentially messy distinction from computer and information science and transform it into an absolute one in order to justify knowledge production about ‘known and unknown terrorists’. In a declassified opinion from the US Foreign Intelligence Surveillance Court (FISC) in the wake of the Snowden revelations, Judge Claire Eagen noted that ‘it is necessary to obtain the bulk collection of a telephone company's metadata to determine those connections between known and unknown international terrorist operatives …’ (FISC, 2013: 18).
To avoid this move to the irrelevance of (meta)data, computer and information scholars have challenged the strict separation of categories and the DIK hierarchy. This move is not quite the same as the recent re-evaluation and claim by critical scholars that ‘there is no such thing as raw data’ (Gitelman, 2013; Lyon, 2014) or that data is not simply ‘pre-analytical and pre-semantic’ (Markham, 2013). Information scientists have indeed long argued that the imagination of data as ‘discrete objects that can be located in time and space’ or as ‘raw numbers and facts’ (Alavi and Leidner, 2001: 109) is fundamentally flawed. Ilkka Tuomi goes furthest by stating that ‘data is more than knowledge’ (1999). Yet, they have also emphasized that data needs to be understood in the context of more fluid boundaries between data, information and knowledge (Alavi and Leidner, 2001; Frické, 2009; Tuomi, 1999). In the Although the law may try to draw hard and fast distinctions between the two, see, e.g., Smith v. Maryland, 442 U.S. 735, 741 (1979), the reality is far murkier and typically depends on context. A change in technical protocols or standards can cause information traditionally regarded as metadata to be treated as content, and vice-versa.
Data and metadata both refer to particular practices of knowledge production in the context of the field of knowledge engineering, which simultaneously draw boundaries (often through standards) between structured/unstructured data, information and knowledge. These distinctions are compounded by political distinctions between kinds of data and metadata. The UK Parliament’s Intelligence and Security Committee recent report on security and privacy recognises these fluid epistemic and political boundaries: Metadata is a term commonly used in the USA, but it has no legal definition in the RIPA and therefore no bearing on the UK system of interception. For example, in the UK a record of a website visited (e.g. http://www.google.com) is treated as CD [communications data], whereas the full web address, which includes the precise words searched (e.g. http://www.google.co.uk/search?q=ISC), is treated as content. (Intelligence and Security Committee, 2015: 52)
The turn to metadata in the emerging Big Data-security assemblage needs to be understood in the context of an economy of Big Data production where ‘digital sources create data as a by-product’ (Ruppert et al., 2013) and we become ‘walking data generators’ (McAfee and Brynjolfsson, 2012). Data itself is a complex epistemic object, and distinctions between kinds of data are produced depending on how data is actionable by digital devices. In another
Seeing like a computer? Big Data as artificial intelligence
A second argument in the controversies about digital surveillance has been formulated in terms of controlled entries to and views onto data. Critics have drawn attention to the practices of ‘mass surveillance’ and their effects on human rights and democracy (Bauman et al., 2014; De Hert and Gutwirth, 2006; Rouvroy and Poullet, 2009). Security professionals have attempted to justify these practices as non-intrusive by arguing that it is computers that read data first and foremost, while humans only see little of what is otherwise processed by machines. Justificatory discourses of ‘bulk data collection’ and targeted rather than mass surveillance have hinged upon a distinction between humans and machines, computers and analysts. In an op-ed for the shouldn’t I be concerned that F.B.I. agents are trampling my rights, just like the I.R.S. might have trampled the rights of certain organizations seeking tax-exempt status? As it turns out, the answer is no. The raw ‘metadata’ requested will not be directly seen by any F.B.I. agent. (2013)
Shea and other intelligence experts invoke an analogy between NSA bulk data processing and targeted surveillance. Ultimately, the assumption is that there is no surveillance where data in not ‘seen’ by a human being. The human/machine distinction, with humans supposedly only coming in at the end of the data processing, aims to render these practices of data collection and processing legitimate by enacting a strong separation between humans and machines. These justifications do not only deactivate criticisms about mass surveillance, but are mobilized to disable claims before the law. James R Clapper, the Director of US National Intelligence, pursues this strategy to ask for a dismissal in
In the UK, the 2013 Annual Report of the Interception of Communications Commissioner reiterates this logic, where GCHQ data mining is deemed legal given that ‘intrusion in this context into the privacy of innocent persons would require
The justification of the separation between humans and computers is difficult to sustain if we understand the computing involved in terms of socio-technical assemblages, an unstable and contingent collection of heterogeneous elements (Latour, 2005). Yet, this distinction has been much more resistant to critique than the metadata/data distinction. To push our critical vocabularies further, it is not enough to point out that the Big Data-security assemblage is socio-technical. We need to understand not just human–non-human relationality, but also ‘the content of the relationships that hold assemblages in place’ (Allen, 2011: 156). What matters in the Big Data-security assemblage is how the relation between humans and computers gains content, and how the assembling of humans and computers is both an association and a division of labour.
Big Data organizations are characterized by a new division of labour between humans and computers (Blanke, 2014; Brynjolfsson and McAfee, 2011) that is about integrating human and machine reasoning at each step of processing relevance and learning how to distinguish significant and non-significant information from each other. In this division of labour, humans and machines are brought together in the same infrastructures to process the data. Recent advances in information systems have focused on connections between humans and computers, given a fundamental shift in how these systems are designed as artificial intelligence applications. A historical perspective on information systems for artificial intelligence helps shed light on this division of labour.
In social sciences, debates about automation, robotics, and data-driven science also suggest a change in human/non-human assemblages. Much critical work has focused on a shift towards computational decision-making and has downplayed the division of labour between humans and computers. We propose a different reading of human–computer assemblages by revisiting historical debates and transformations in the wake of the so-called ‘winter of artificial intelligence’ (AI). The phrase captures the perceived failure in the late 1980s of attempts to develop analytical capacities that would bring computing machines close to human intelligence. All the enthusiasm and the early promises of creating the thinking machines seemed gone. However, theoretical and practical breakthroughs in artificial intelligence capacities have since put in doubt this supposed ‘winter’ (Kurzweil, 2005: 264), as AI has developed a new type of ‘intelligence’ by combining human and computer reasoning.
In a reply to Noam Chomsky’s critique of artificial intelligence, Peter Norvig (2012), Google’s Director of Research, captures the epistemic transformation that took place in the 1980s. The ‘winter of AI’ was related to an over-reliance on logical models to simulate human reasoning and their subsequent failure. Computing scientists wanted to create an artificial intelligence that replicated human intelligence, but was separate from humans. Norvig (2012) locates an epistemico-material transformation from logical models to statistical models which ‘have achieved a dominant (although not exclusive) position.’ Unlike logical models, statistical models focus on assemblages of humans and machines that can process data. Rather than re-creating a black box that veils how humans reason, artificial intelligence has focused on building models that can solve particular problems. These models are developed within complex workflows of human–computer interaction, starting with large testbeds to tune algorithms so that they simulate human judgement on information relevance or expressed sentiments. In order to perform, algorithms need to be constantly evaluated for their effectiveness.
The epistemic crisis of AI has led to a reconfiguration and integration of human work and intelligence into new human–computer assemblages. Information systems to develop analytical capacities have thus been built in a similar way since the winter of AI. Security systems in all their components share with other data science applications that they are also the result of the winter of AI, when research and funding shifted from ‘wild-eyed dreams’ of creating a human-like machine and started to concentrate on particular applications that were made possible by fostering different connections and division of labour between humans and computers.
In these models, humans do not interact with the data just at the end, but are involved at every stage through evaluation, optimization, training, etc. Yet, this does not mean that computers are simply passive tools, because as much processing as possible needs to be computerized. Computers should learn ‘unsupervised’ or, to put it differently, develop agency in these assemblages. A successful unsupervised learning technique is, for instance, the so-called topic modelling, which auto-summarizes a collection of documents into a number of common topics. In security applications, for instance, topic modelling is used to summarize cyber-threats in web data mining. However, these topic models can be too suggestive and require careful intervention by humans (Schmidt, 2011). Even so-called ‘unsupervised’ techniques are not black boxes but a human–computer assemblage.
Similarly, since 2006 we have witnessed yet another transformation in machine learning techniques through ‘deep learning’ for speech and image recognition. Again, however, as soon as one analyses the components of this new black box, it becomes apparent that even deep learning as the state-of-the-art in unsupervised computational learning requires human participation. Rather than becoming autonomous, computers are still enrolled in a socio-technical assemblage. Two leading machine-learning researchers, Socher and Manning, thus define machine learning as the ‘numerical optimization of weights for human-designed representations and features’ (2013). Security analytics is no different from other domains that enrol computers, artificial intelligence practices and data scientists.
In analysing this emerging assemblage through the transformation of AI, we have built upon critical work that draws attention to the socio-technical character of Big Data (Lyon, 2014). Yet, our analysis also shows how relations in a socio-technical assemblage gain content historically. Debates in AI have reconfigured the human–computer assemblage through an epistemico-material division of labour between humans and computers. This re-reading of Big Data through the history of AI also allows us to challenge the credibility of the distinctions that security professionals attempt to institute between computer-based targeted surveillance and mass surveillance by showing how the ‘winter of AI’ has led to particular modes of assembling computers, techniques of machine learning and (data and security) analysts.
Algorithmic security: Anticipation and probabilities
A third and related site of controversy in the emerging Big Data-security assemblage has focused on the epistemic capabilities of algorithms. Intelligence experts speak about these capacities as finding the ‘needle in a haystack’ and the NSA’s General Alexander as ‘connecting the dots’. In an intervention before the US Congress following the Snowden revelations, former Federal Bureau of Investigation (FBI) Director Robert Mueller noted that ‘If you narrow [the scope of surveillance], you narrow the dots and that might be the dot that prevents the next Boston’ (Roberts, 2013). The shift in reasoning towards the anticipation of possibilities, conjecture and speculation underpins the justifications that security professionals proffer for the necessity of extensive Big Data mining. Critical security and surveillance scholars have analysed the recalibration of security practices through anticipatory knowledge (Amoore, 2014; Aradau and van Munster, 2011; Cooper, 2008; de Goede, 2012; Lyon, 2014). Critical data studies have highlighted to the predictive fallacy of Big Data, the tension between correlation and causation, the ‘return to empiricism’, and the opacity of algorithms (Boyd and Crawford, 2012; Ekbia et al., 2015, Kitchin, 2014; van Dijck, 2014).
As important as these criticisms are, they do not address the perceived necessity to ‘collect it all’, the whole ‘haystack’ of data, in order to enhance the capabilities of algorithms. As Judge Pauley III glosses in No doubt, the bulk telephony metadata collection program vacuums up information about virtually every telephone call to, from, or within the United States. That is by design, as it allows the NSA to detect relationships
Firstly, we take seriously Norvig’s widely quoted claim that Google does not necessarily have better algorithms than everybody else, but more data (quoted in Schutt and O'Neil, 2013). Marissa Mayer, Google’s former VP of Search Products and User Experience, had also noted ‘that having access to large amounts of data is in many instances more important than creating great algorithms’ (quoted in Perez, 2007). Data has become more important than algorithms themselves, because ‘[t]here is no single scientific breakthrough behind Big Data. On the contrary, the methods used have been well known and established for quite some time’ (Lehikoinen and Koistinen, 2014: 39). Secondly, we analyse algorithms through the probabilistic methods that all algorithms deploy independent of their exact design in particular institutions. Discourses of the ‘novelty’ of Big Data and algorithmic capacities need to be located within probabilitistic methods and their limitations, which are the foundation of Big Data analytics as they allow reasoning about uncertainty (Bengio et al., in preparation).
Justificatory discourses of the capabilities of Big Data for security governance activate a particular imagination of the relation between part and whole. The haystack metaphor that security professionals are using is no longer a metaphor for a sample size but for Big Data, where ‘
Shifting from the focus on algorithms to the data that algorithms need and the modes of probabilistic reasoning designed for algorithmic processes allows us to develop a critical vocabulary about the ‘needle in a haystack’. Security professionals have justified this need for an infinitely expanding haystack through the idea that having all the data can algorithmically reveal better knowledge about potential terrorists, which would make pre-emptive action possible. However, this imaginary of Big Data that yields better knowledge has been challenged in debates about Big Data. Bigger data is not better ‘without limit’ (Frické, 2014: 5).
Let us take an example, which has been at the heart of claims about the unreasonable effectiveness of data: Google. Google engineers can rely on the ‘unreasonable effectiveness’ of web page data, because these pages consist of words whose meaning can be derived from the frequencies with which these words appear within web pages. Over time, ‘human language has already evolved words for the important concepts’ (Halevy et al., 2009: 12). Security applications, on the other hand, share with Big Data applications in humanities and social sciences the interest in shifting concepts and ‘minority’ vocabularies, rare words and rare data. While standard algorithmic reasoning can be ‘successful for a lot of things’ that follow regularities, ‘it amounts to a deliberate neglect of rare words’ (White, 2011). Statistical algorithms tend to ignore rare data, especially as the models get more complex in social domains and are tightly fitted to their original training data. In the language of machine learning for Big Data, these models are ‘overfitted’ to the training data and are challenged by new unknown data items. According to the data scientist Janert (2010: 424), the ‘nature of statistical learning’ requires us to add more and more dimensions to our data and move further away from the rare variables.
Another example of the fallacy to ‘collect it all’ for the purposes of data analytics is the analysis by the computer scientist and outspoken critic of NSA data mining Edward Felten (2014), where he shows how in typical data reasoning reduced data sets can lead to better results than large data sets. He argues that if the NSA covers larger numbers of phone connection hops, starting from a phone number they have under surveillance, this will not render the security analytics more effective. As Felten shows using typical algorithms, the fewer the hops involved (i.e. the smaller the sample), the better the results. The NSA drive to collect and mine Big Data understood as ‘all data’ is revealed as a myth.
The drive to collect more and more data is not just based on a Big Data myth but can be counterproductive not just for the fitting of the machine learning algorithms but also for their methods, as it generally leads to many false positives. Computer scientists involved in mining Big Data have argued for a long time that the larger the haystack, the more likely it is that only meaningless events are added. A comparison of random and non-random events will lead to an overestimation of the non-random ones. One of the standard textbook introductions to machine learning makes an important point for our discussion (Leskovec et al., 2011). In offering a standard example of predictive data analytics for counter-terrorism, the authors develop a criticism of the assumption that if the haystack is just big enough and enough data is collected, the ocean of data will automatically also increase the possibilities to find the needle in the data.
Starting with a typical data mining case for counter-terrorism, Leskovec et al. (2011) work with the assumption that two people are considered to be suspicious if they stayed more than once in the same hotel at the same time, which is a typical assumption in security analytics. In their example, if one billion people were to be tracked over a period of 1000 days, there will be a well-defined probability that two random people stay at the same hotel more than once. This probability is quite small and might therefore look promising for identifying potential suspects. The problem is, however, that in these millions of people there are many possible pairs of people who could have stayed at the same hotel. In their calculation, the result was over 250,000 of such ‘suspicious’ pairs (Leskovec et al., 2011). Therefore, as computer scientists know, the error in data mining methods actually rises if more and more data is collected, which is contrary to the belief that an ‘ocean of data’ makes it possible to identify ‘gossamer contacts’ with more certainty.
The production of large numbers of false positives in the ‘ocean of data’ raises another question for Big Data mining – how to distinguish ‘real’ and ‘false’ suspects. According to Felten (2014), rather than predicting new terrorist activities, typical NSA security analytics algorithms are designed for more realistic use cases in order to avoid overfitting and overestimation of false positives. These algorithms are designed to substantiate the suspicion a security analyst already has, rather than predict new suspects or suspect behaviour, as the analogy of the ‘needle in a haystack’ might suggest. These algorithms are thus not good at eliminating suspicion. In general, they do not help to find the needle in the haystack but help confirm whether there is reason to assume that there might be a needle. Big Data mining has never been good at evaluating false positives, which is less of an issue for an Internet search engine where users get used to skipping irrelevant results, but is an issue in a security context where innocent people become suspicious.
Contra the FBI director Robert Mueller’s concern that a narrower focus of surveillance will miss the dots that need to be connected in order to prevent the ‘next Boston’ (Roberts, 2013), computer scientists have shown that it is often the wider focus that has this effect. If data grows big, both the apparent risk of a terrorist attack and the number of suspicious people will vastly increase. In terms of the theory of data mining, this problem can be seen as typical of distributions with heavy tails, which are those distributions that are extremely skewed with a long tail of events that happen sometimes but not very often.
The ‘next terrorist attack’ is a possible event of low probability such as the ‘rare words’ for which Google’s data enthusiasm is less reasonably effective, as discussed supra. Security-relevant activities are generally heavy-tailed, as they are conducted by a small number of people compared to the overall population. The corresponding heavy-tailed distributions require special methods in order to apply algorithmic reasoning (Clauset et al., 2007). General statistical assumptions about what can be reasonably expected as the next event do not work, because these are ‘distributions without expectations’ (Janert, 2010: 201). The mean event, for instance, does not reveal anything about the expected behaviour of the distribution. For heavy-tailed distributions, events outside the tail do not indicate anything about events in the tail where the suspects can be found. Similarly, the random behaviour of those who are not suspicious does not tell us much about the behaviour of suspects and vice versa. Making the data ‘big’ will therefore not reveal anything that can be used to identify suspects, as long as we do not know whether this data belongs to the tail.
In order for data mining to work on heavy-tailed distributions, these are often split up into various subgroups where each of them is dealt with individually. To make these divisions, cut-off points need to be found that analysts consider as indicative of suspicious behaviour in the tail. For instance, all those are suspects who call the same phone number in Maryland. These features generally cannot be just read from the data, but an analyst needs to make a decision as to which features count to identify a suspect (Janert, 2010: 434). Big Data algorithms require specialized theories and domain assumptions like the feature that suspects call the same number in Maryland. Formally speaking, without assumptions about the data, Big Data mining algorithms do not perform better on new information than any random prediction; this is independent of the size of the data set (Wolpert and Macready, 1997).
The metaphor of finding the needle in the haystack is a powerful one used by Big Data professionals and then reused by security professionals, judges and politicians. While it seemingly justifies the huge effort of collecting more and more data in order to cover all possibilities, the metaphor effaces debates in computer science about the epistemic limits of algorithmic practices and methods. Developing a critique of anticipatory algorithmic security entails an immanent critique of algorithmic practices and methods.
Conclusion: Theses on critique
On 7 May 2015, the US Court of Appeals for the Second Circuit overturned the decision in
In order to address the impasses of critique in public and academic controversies, this paper has proposed a series of reformulations of existing critical vocabularies of the Big Data-security assemblage. A collaboration between social sciences, and computer and information sciences, as developed here, can challenge the credibility of the justifications promoted by intelligence agencies such as the NSA and the GCHQ and contribute to research agendas in critical approaches to security and surveillance, on the one hand, and critical data studies, on the other. Rather than assuming that informatics is ‘the discipline of choice of liberal power’ (Bauman et al., 2014, 139), we have interpreted a series of debates in knowledge engineering, artificial intelligence and data mining in order to address the limits and impasses of critical engagements with Big Data in security governance. Methodologically, we have been able to eschew the invocation of secrecy around NSA and GCHQ practices by engaging with the state-of-the-art in computer and information sciences. As employers of computer scientists, mathematicians, and physicists, we have contended that the NSA and the GCHQ are unlikely to have developed technologies and methods beyond the latest research in academic and commercial organizations. Yet, despite the apparent credibility of their discourses being grounded in the scientificity of these disciplines, we have shown that debates in computer and information science challenge this credibility. We have combined this engagement with computer and information science with an analysis of legal cases, public inquiries, parliamentary and media reports, and declassified documents in the wake of the Snowden revelations.
We have developed a two-pronged argument. On the one hand, we have shown that the invocations of privacy, mass surveillance and decisions by algorithms run into political and epistemic impasses as they have been continually challenged by professionals of security and politics. On the other, we have argued for different vocabularies of critical intervention in the sites of controversy in the emerging Big Data-security assemblage: from metadata privacy to the production of data as a complex epistemic entity; from mass surveillance by computers to the division of labour between humans and computers in artificial intelligence; from an underlying algorithmic logic of security to algorithmic practices and probabilistic methods. These interventions suggest several trajectories for developing critical research about data in security governance. By way of conclusion we propose four theses on critique:
Thesis 1
Big Data theories and methods are not as new or revolutionary as the justificatory discourses of security professionals and many critical academic analyses of Big Data suggest. Rather than assuming a ‘Big Data revolution’, the deployment of Big Data for security purposes needs to be understood in terms of what Michel Foucault has called the ‘inflection of the curve’ (1998 [1976]).
Thesis 2
Data is a complex epistemic object. Therefore, data needs to be approached as an object of inquiry rather than subsumed to knowledge. Meanings of data, metadata, and knowledge differ between social sciences and information science. Rather than simply contending that there is no such thing as raw data, critical data studies need to analyse the ‘making up’ of data and the production of kinds of data at the juncture between information and computer science, on the one hand, and politics, on the other.
Thesis 3
The digital age does not mean that decisions are simply transferred from humans to computers and algorithms. We are only beginning to understand the transformation of intelligence agencies (and other security organizations) into Big Data organizations. There is little existing work on the details of this organizational change, which entails understanding histories of Big Data as artificial intelligence, and particularly the reconceptualization of the relation between humans and computers in AI. Analyses of data-security assemblages need to attend not just to the modes of association and heterogeneous elements of an assemblage, but equally to how it gains content through a division of labour between humans and computers.
Thesis 4
There are no ‘unreasonably effective’ algorithms. Critical approaches to security and surveillance need to engage with the methods and routine practices of Big Data-security analytics. Algorithms continue to rely on probabilistic methods, which depend on and are challenged by the increased amounts of data produced today. Unless we analyse the algorithmic practices and methods from information and computing sciences, we might write and critique only the science fiction of security analytics.
