a brave new world - student surveillance in higher education - revisited
TRANSCRIPT
A brave new world: student surveillance in higher education – revisitedBy Paul Prinsloo (University of South Africa)
Presentation at the Association for Institutional Research (AIR) Forum, Denver May 26-29 – “Data and decisions for higher education”
Photograph: Paul Prinsloo
ACKNOWLEDGEMENTS• Except for the photographs on the title and last slides, I don’t own the
copyright of any of the images used and hereby acknowledge their original copyright and licensing regimes. All the images used in this presentation have been sourced from Google and were labeled for non-commercial reuse
• This work (excluding the licencing regimes of the images from Google) is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License
• An earlier version of this presentation won the Best Paper Award at the Southern African Association for Institutional Research (SAAIR) Conference 16-18 October 2014, held in Pretoria, South Africa
When we hear ‘surveillance’ we may think of…
Source credit: http://mashable.com/2013/06/09/edward-snowden/
Source: http://www.pewinternet.org/2015/03/16/Americans-Privacy-Strategies-Post-Snowden/#the-public-has-divided-sentiments-about-the-surveillance-programs
Source: http://www.pewinternet.org/2015/03/16/Americans-Privacy-Strategies-Post-Snowden/#the-public-has-divided-sentiments-about-the-surveillance-programs
Or we may think of…
Image credit: http://commons.wikimedia.org/wiki/File:Penetentiary_Panopticon_Plan.jpg
‘Panopticon’ Jeremy Bentham, 1873
Greek mythology – Argus Panoptes – A giant with 100 eyes
“Secrets are lies”“Sharing is caring”
“Privacy is theft”(Eggers, 2013, p. 303)
And more recently… TruYou – “one account, one identity, one password, one payment system, per person. (…) The devices knew where you were… One button for the rest of your life online… Anytime you wanted to see anything, use anything, comment on anything or buy anything, it was one button, one account, everything tied together and trackable and simple…”
(Eggers, 2013, p. 21)
The eerie resemblance with the way higher education sees the institutional learning management system (LMS) is accidental…
Every breath you takeEvery move you makeEvery bond you breakEvery step you takeI'll be watching you
Every single dayEvery word you say
Every game you playEvery night you stayI'll be watching you
O can't you seeYou belong to me…
Sting – Every breath you take
Image source: https://www.mpiwg-berlin.mpg.de/en/news/features/feature14 Copyright could not be established
• 1749 Jacques Francois Gaullauté proposed “le serre-papiers” – The Paperholder – to King Louis the 15th
• One of the first attempts to articulate a new technology of power – one based on traces and archives (Chamayou, nd)
• The stored documents comprised individual reports on each and every citizen of Paris
The technology will allow the sovereign “…to know every inch of the city as well as his own house, he will know more about ordinary citizens than their own neighbours and the people who see them everyday (…) in their mass, copies of these certificates will provide him with an absolute faithful image of the city” (Chamayou, n.d)
The Paperholder – “le serre papiers” (1749)
Image credit: http://iconicphotos.wordpress.com/2010/07/29/the-great-ivy-league-photo-scandal/
“… a person’s body, measured and analysed, could tell much about intelligence, moral worth, and probably future achievement…
The data accumulated… will eventually lead on to proposals to ‘control and limit the production of inferior and useless organisms’”
(Rosenbaum, 1995)
The great Ivy League photo scandal 1940-
1970
So how do we understand and
critically engage with the issues surrounding
the increasing surveillance of
students in higher education?
Image credit: http://graffitiwatcher.deviantart.com/art/Big-Brother-is-Watching-173890591
“… ‘educational technology’ needs to be understood as a knot of social, political,
economic and cultural agendas that are riddled with complications, contradictions and conflicts”
(Selwyn, 2014, p. 6)
If we accept that
…what are the implications for the collection, analysis and use of student data?
Understanding the collection, analysis and use of student data in the contexts of …
• Broader trends in higher education• From surveillance to sousveillance• The discourses in data and increasingly Big Data
Understanding the collection, analysis and use of student data in the context of some of the broader trends in higher education
1. Changes in funding regimes – funding follows performance rather than preceding it
2. Increasing concerns regarding student retention and dropout3. Ranking systems and the internationalization of higher education4. Higher education as business – the dominant neoliberal paradigm5. The algorithmic turn and quantification fetish in higher education6. The increasing digitization of learning and teaching7. The gospel of technosolutionism in higher education8. The hype, promise and dangers of (Big) data
Higher education is mesmerized/seduced by the potential of the collection, analysis and use of student data
Image credit: http://en.wikipedia.org/wiki/Medusa
(Student) data as Medusa – techno-solutionism in action
Every page you viewEvery click you makeEvery link you follow
Every step you (don’t) takeI'll be watching you
Every single dayEvery word you say
Every game you playEvery night you stayI'll be watching you
O can't you seeYou belong to me…
Adapted from Sting – Every breath you take
As student data are increasingly used to personalise learning and to allocate resources, we are creating a “brave new world” with Delta children wearing khaki, Epsilons wearing black (“they’re too stupid to be able to read or write”), and Gammas and Deltas (Huxley, 2007, p. 22)
The dream/nightmare of personalizing learning…
Understanding the collection, analysis and use of student data in the context of the change from surveillance to sousveillance
Image credit: http://commons.wikimedia.org/wiki/File:SurSousVeillanceByStephanieMannAge6.png
Jennifer Ringely – 1996-2003 – webcam Source: http://onedio.com/haber/tum-zamanlarin-en-etkili-ve-onemli-internet-videolari-36465
If I did not share it on Facebook, did it really happen?
We share more than ever before, we are watched more than ever before and we watch each other more than ever before…
From surveillance to sousveillance…
We are increasingly watched/measured
We increasingly watch/measure each
other
We increasingly watch ourselves
Image credit: http://en.wikipedia.org/wiki/Surveillance#/media/File:Surveillance_video_cameras,_Gdynia.jpeg
Understanding the collection, analysis and use of student data in the context of the discourses on (Big) data
Different sources/variety of quality/ integrity of data
Different role-players with different interests• Individuals• Corporates• Governments• Higher education• Data brokers• Fusion centers
Different methods/types of surveillance, harvesting and analysis
Issues re• Informed consent• Reuse/contextual
integrity/context collapse
• Ethics/privacy/ justice/care
The Trinity of Big Data
Adapted & refined from Prinsloo, P. (2014). A brave new world. Presentation at SAAIR, 16-18 October
http://www.slideshare.net/prinsp/a-brave-new-world-student-surveillance-in-higher-education
Three sources of dataDirected
A digital form of surveillance wherein the “gaze of the technology is focused on a person or place by a human operator”
AutomatedGenerated as “an inherent, automatic function of the device or system and include traces …”
Volunteered“gifted by users and include interactions across social media and the crowdsourcing of data wherein users generate data” (emphasis added)
(Kitchen, 2013, pp. 262—263)
Image credit: http://commons.wikimedia.org/wiki/File:Red_sandstone_Lattice_piercework,_Qutb_Minar_complex.jpg
The Trinity of Big Data results in an “elaborate lattice of information networking” (Solove, 2004, p. 3) where consent and protection of
privacy are and remain fragile (Prinsloo & Slade, 2015)
Big data presents “a paradigm shift in the ways we understand and study our world”
(Eynon, 2013, p. 237)
Image credits: http://commons.wikimedia.org/wiki/File:DARPA_Big_Data.jpg
Amidst the hype and the potential there are also…• “Rule by algorithm? Big data and the threat of algocracy” – Danaher
(2014)• “Big Data’s big unintended consequences” – Wigan & Clarke (2013)• “Six provocations for Big Data” – Boyd & Crawford (2011)• “Critical questions for Big Data” – Boyd & Crawford (2012)• “Judged by the Tin Man: Individual rights in the age of Big Data” –
Tene & Polonetsky (2013)• “Algorithmic accountability” – Diakopoulos (2014)• “The scored society: Due process for automated predictions” –
Citron & Pasquale (2014)• “Big data, data integrity, and the fracturing of the control zone” –
Lagoze (2014)
Shi(f)t happens…• The claim that Big Data is
equavalent to “allness” (Lagoze, 2014) – n=all – providing a complete view of reality
• Big data “lessen our desire for exactitude” (Mayer-Schönberger & Cukier, 2013 in Lagoze, 2014)
• It is no longer necessary to investigate the why things happen… More important is to note what is happening – data speaks for itself…
“Hidden algorithms can make (or ruin) reputations, decide the destiny of entrepreneurs, or even devastate an entire economy. Shrouded in secrecy and complexity, decisions at major Silicon Valley and Wall Street firms were long assumed to be neutral and technical. But leaks, whistleblowers, and legal disputes have shed new light on automated judgment”
Source: http://www.hup.harvard.edu/catalog.php?isbn=9780674368279
Critical questions for big data – boyd & Crawford (2012)
1. Big data changes the definition of knowledge – “Who knows why people do what they do? The point is they do it, and we can track and measure it with unprecedented fidelity. With enough data, the numbers speak for themselves” (Anderson, 2008, in boyd & Crawford, 2012, p. 666)
2. Claims to objectivity and accuracy are misleading – “working with Big Data is still subjective, and what it quantifies does not necessarily have a closer claim on objective truth” (Boyd &
Crawford, 2012, p. 667). Big Data “enables the practice of apophenia: seeing patterns where none actually exist, simply because enormous quantities of data can offer connections that radiate in all directions” (ibid., p. 668)
Critical questions for big data (2) – boyd & Crawford (2012)
3. Bigger data are not always better data
4. Taken out of context, Big Data loses its meaning – leading to context collapse & lack of contextual integrity
5. Just because it is accessible does not make it ethical – the difference in ethical review procedures and overview between research and ‘institutional research’
6. Limited access to Big Data creates new digital divides
Points of departure(Big) data is…
…not an unqualified good (Boyd and Crawford, 2011) and “raw data is an oxymoron” (Gitelman, 2013) – see Kitchen, 2014
Technology and specifically the use of data have been and will always be ideological (Henman, 2004; Selwyn, 2014) and embedded in relations of power (Apple, 2004; Bauman, 2012)
Towards a typology of surveillance in higher education (Knox, 2010)
• Panoptic – Automated/internalised visibility • Rhizomatic – Multidirectional/sousveillance/ Surveillant assemblages – merging of different spheres/multiplying/amplifying
• Predictive
Predictive surveillance
•The (digital) past + (digital) real-time => future behavior … “The past becomes a simulated prologue” (Knows, 2010, p. 6; emphasis added)•Structuring/constraining choices/possibilities – shaping material conditions (Henman, 2004; Napoli, 2013) •Issues of identity/race/gender/class•The end of forgetting (Mayer-Schönberger, 2009; Rosen, 2010)
Preliminary seven dimensions of surveillance (Knox 2010)
1. Automation2. Visibility3. Directionality4. Assemblage5. Temporality6. Sorting7. Structuring
1. Automation
Key questions Dimensional intensity
What is the timing of the collection?
Intermittently/infrequently
Continuous
Locus of control? Human Machine
Can it be turned on and off (and by whom?)
All the monitoring can be turned on/off
None of the monitoring can be turned off
2. VisibilityKey questions Dimensional intensity
Is the surveillance apparent and transparent?
All parts (collection, storage, processing and viewing) are visible
None of the monitoring is visible
Ratio of self-to-surveillant knowledge?
Subject knows everything the surveillant knows
Subject does not know anything that the surveillant knows
3. Directionality
Key questions Dimensional intensity
What is the relative power of surveillant to subject?
Subjects hold all the power
Surveillant holds all the power
Who has access to monitoring/recording/ broadcasting functions?
Subjects Surveillant
4. Assemblage
Key questions Dimensional intensity
Medium of surveillance Single medium (e.g. text)
Multimedia
Are the data stored? No Yes
Who stores the data? Subject or collector
Third party
5. Temporality
Key questions Dimensional intensity
When does the monitoring occur?
Confined to the present
Combines the present with the past
How long is the monitoring frame?
One, isolated, relatively short frame (e.g. test)
Long periods, or indefinitely
Does the system attempt to predict future behavior/outcomes
No – only assessment of the present
Present + past used to predict the future
When are the data available?
All of the data available only after event is completed
Available in real-time and experienced as instantaneous
6. Sorting
Key questions Dimensional intensity
Are subjects’ data compared with other data – other individuals/ groups/ abstract configurations/ state mandates?
None Other data are used as basis for comparison
7. StructuringKey questions Dimensional intensity
Are data used to alter the environment (i.e. treatment, experience, etc.)?
Not used Used to alter the environment of all subjects
Are data used to target the subject for different treatment that they would otherwise receive?
No data are used as basis for differing treatment
Based on data, treatment is prescribed
Do students know/have the right to know… • what data we harvest from them• about the assumptions that guide our algorithms• when we collect data & for what purposes• who will have access to the data (now & later)• how long we will keep the data & for what
purpose & in what format• how will we verify the data & • do they have access to confirm/enrich their
digital profiles…?Adapted from Prinsloo, P., & Slade, S. (2015). Student privacy self-management: implications
for learning analytics. Presentation at LAK15, Poughkkeepsie, NY, 16 March 2015 http://www.slideshare.net/prinsp/lak15-workshop-vulnerability-final
Do they know? Do they have the right to know?
Can they opt out and what are the
implications if they do/don’t?
Adapted from Prinsloo, P., & Slade, S. (2015). Student privacy self-management: implications for learning analytics. Presentation at LAK15, Poughkkeepsie, NY, 16 March 2015
http://www.slideshare.net/prinsp/lak15-workshop-vulnerability-final
What are the implications for the collection, analysis and use of student (digital) data? (Prinsloo & Slade, 2015)1. The duty of reciprocal care
• Make TOCs as accessible and understandable (the latter may mean longer…)
• Make it clear what data is collected, when, for what purpose, for how long it will be kept and who will have access and under what circumstances
• Provide users access to information and data held about them, to verify and/or question the conclusions drawn, and where necessary, provide context
• Provide access to a neutral ombudsperson(Prinsloo & Slade, 2015)
What are the implications …? (2)2. The contextual integrity of privacy and data – ensure the contextual
integrity and lifespan of personal data. Context matters…
3. Student agency and privacy self-management• The fiduciary duty of higher education implies a social contract of
goodwill and ‘do no harm’• The asymmetrical power relationship between institution and
students necessitates transparency, accountability, access and input/collaboration
• Empower students – digital citizenship/care• The costs and benefits of sharing data with the institution should
be clear• Higher education should not accept a non-response as equal to
opting in… (Prinsloo & Slade, 2015)
What are the implications …? (3)4. Future direction and reflection
• Rethink consent and employ nudges – move away from thinking just in terms of a binary of opting in or out – but provide a range of choices in specific contexts or needs
• Develop partial privacy self-management – based on context/need/value
• Adjust privacy’s timing and focus - the downstream use of data, the importance of contextual integrity, the lifespan of data
• Moving toward substance over neutrality – blocking troublesome and immoral practices, but also soft, negotiated spaces of reciprocal care
(Prinsloo & Slade, 2015)
(In)conclusions
“Technology is neither good or bad; nor is it neutral… technology’s interaction with social ecology is such that technical developments frequently have environmental, social, and human consequences that go far beyond the immediate purposes of the technical devices and practices themselves”
Melvin Kranzberg (1986, p. 545 in boyd & Crawford, 2012, p. 1)
(In)conclusions
Learning analytics are a structuring device, not neutral, informed by current beliefs about what counts as knowledge and learning, colored by assumptions about gender/race/class/capital/literacy and in service of and perpetuating existing or new power relations
Welcome to a brave new world…
Last thing I remember, I wasRunning for the door
I had to find the passage backTo the place I was before
"Relax, " said the night man,"We are programmed to receive.
You can check-out any time you like,But you can never leave! "
Eagles – Hotel California
THANK YOUPaul Prinsloo (Prof)Research Professor in Open Distance Learning (ODL)College of Economic and Management Sciences, Office number 3-15, Club 1, Hazelwood, P O Box 392Unisa, 0003, Republic of South Africa
T: +27 (0) 12 433 4719 (office)T: +27 (0) 82 3954 113 (mobile)[email protected] Skype: paul.prinsloo59Personal blog: http://opendistanceteachingandlearning.wordpress.comTwitter profile: @14prinsp
Photograph: Paul Prinsloo
References and additional reading Apple, M.W. (Ed.). (2010). Global crises, social justice, and education. New York, NY: Routledge. Bauman, Z. (2012). On education. conversations with Riccardo Mazzeo. Cambridge, UK: Polity. Booth, M. (2012, July 18). Learning analytics: the new black. EDUCAUSEreview, [online]. Retrieved from http://
www.educause.edu/ero/article/learning-analytics-new-black Boyd, D., & Crawford, K. (2013). Six provocations for Big Data. Retrieved from http://
papers.ssrn.com/sol3/papers.cfm?abstract_id=1926431 Chamayou, G. (n.d.). Every move will be recorded. [Web log post]. Retrieved from https://
www.mpiwg-berlin.mpg.de/en/news/features/feature14 Cirton, D.K., & Pasquale, F. (2014). The scored society: Due process for automated predictions.
http://ssrn.com/abstract=2376209 Danaher, J. (2014, January 6). Rule by algorithm? Big Data and the threat of algocracy.[Web log post].
Retrieved from http://philosophicaldisquisitions.blogspot.com/2014/01/rule-by-algorithm-big-data-and-threat.html
Deleuze. G. (1992). Postscript on the societies of control. October, 59 pp. 3-7. Diakopoulos, N. (2014). Algorithmic accountability. Digital Journalism. DOI: 10.1080/21670811.2014.976411 Gitelman, L. (Ed.). (2013). “Raw data” is an oxymoron. London, UK: MIT Press.Henman, P. (2004). Targeted!: Population segmentation, electronic surveillance and governing the unemployed in Australia. International Sociology, 19, 173-191Gray, J. (2004). Heresies. Against progress and other illusions. London, UK: Granta Books.
References and additional reading (cont.)Kitchen, R. (2013). Big data and human geography: opportunities, challenges and risks. Dialogues in Human
Geography, 3, 262-267. SOI: 10.1177/2043820613513388 Knox, D. (2010). Spies in the ouse of learning: a typology of surveillance in online learning environments. Paper presented at Edge, Memorial University of Newfoundland, Canada, 12-15 October.Kranzberg, M. (1986) Technology and history: Kranzberg's laws’. Technology and Culture, 27(3), 544—560.Lagoze, C. (2014). Big Data, data integrity, and the fracturing of the control zone. Big Data & Society (July-
December), 1-11. Mayer-Schönberger, V. (2009). Delete. The virtue of forgetting in the digital age. Princeton, NJ: Princeton
University Press.Mayer-Schönberger, V., Cukier, K. (2013). Big data. London, UK: Hachette. Morozov, E. (2013a, October 23). The real privacy problem. MIT Technology Review. Retrieved from
http://www.technologyreview.com/featuredstory/520426/the-real-privacy-problem/ Morozov, E. (2013b). To save everything, click here. London, UK: Penguin Books. Morozov, E. (2013). To save everything, click here. London, UK: Penguin Books. Napoli, P. (2013). The algorithm as institution: Toward a theoretical framework for automated media
production and consumption. In Media in Transition Conference (pp. 1–36). DOI: 10.2139/ssrn.2260923Pasquale, F. (2015). The black box society. Harvard Publishing, US.Prinsloo, P. (2009). Modelling throughput at Unisa: The key to the successful implementation of ODL. Retrieved
from http://uir.unisa.ac.za/handle/10500/6035
References and additional reading (cont)Prinsloo, P., & Slade, S. (2014). Educational triage in open distance learning: Walking a moral tightrope. The
International Review of Research in Open and Distributed Learning, 15(4), 306-331. Retrieved from http://www.irrodl.org/index.php/irrodl/article/view/1881/3060
Prinsloo, P., & Slade, S. (2015, March). Student privacy self-management: implications for learning analytics. In Proceedings of the Fifth International Conference on Learning Analytics And Knowledge (pp. 83-92). ACM. Retrieved from http://dl.acm.org/citation.cfm?id=2723585
Rambam, S. (2008). Privacy is dead. Get over it. Retrieved from https://www.youtube.com/watch?v=Vsxxsrn2Tfs&index=1&list=PL8C71542205AA51E5
Rosen, J. (2010, July 21). The web means the end of forgetting. New York Times [Online].Rosenbaum, R. (1995, January 15). The great Ivy League nude posture photo scandal. The New York Times.
Retrieved from http://www.nytimes.com/1995/01/15/magazine/the-great-ivy-league-nude-posture-photo-scandal.html
Serlwyn, N. (2014). Distrusting educational technology. Critical questions for changing times. New York, NY: Routledge
Subotzky, G., & Prinsloo, P. (2011). Turning the tide: A socio-critical model and framework for improving student success in open distance learning at the University of South Africa. Distance Education, 32(2), 177-193.
Tene, O. & Polonetsky, J. (2013). Judged by the Tin Man: Individual rights in the age of Big Data. J. on Telecomm. & High Tech. L., 11, 351.
.
References and additional reading (cont)Totaro, P., & Ninno, D. (2014). The concept of algorithm as an interpretive key of modern rationality. Theory
Culture Society 31, pp. 29—49. DOI: 10.1177/0263276413510051 Therborn, G. (ed.).(2006). Inequalities of the world. New theoretical frameworks, multiple empirical
approaches. London, UK: Verso BooksWagner, E., & Ice, P. (2012, July 18). Data changes everything: delivering on the promise of learning analytics in
higher education. EDUCAUSEreview, [online]. Retrieved from http://www.educause.edu/ero/article/data-changes-everything-delivering-promise-learning-analytics-higher-education
Watters, A. (2013, October 13). Student data is the new oil: MOOCs, metaphor, and money. [Web log post]. Retrieved from http://www.hackeducation.com/2013/10/17/student-data-is-the-new-oil/
Watters, A. (2014). Social justice. [Web log post]. Retrieved from http://hackeducation.com/2014/12/18/top-ed-tech-trends-2014-justice
Wigan, M.R., & Clarke, R. (2013). Big data’s big unintended consequences. Computer,(June), 46-53.