My latest paper, “Situated data analysis: a new method for analysing encoded power relationships in social media“, started out as an analysis of the data visualisations in Strava, but ended up as something more ambitious: a method that I think can be used to analyse all kinds of digital platform using personal data in different contexts. Here is a video teaser explaining the key points of situated data analysis:

This paper has been a long time in the works, and started off as part of the INDVIL project on data visualisations, where I was tasked with thinking about the epistemology of data visualisations. Working through revision after revision of my analyses of data visualisations in Strava I found that what really interested me about Strava was the many different ways that the personal data it collects from runners and cyclists are presented—or, more precisely, how the data are situated. Once I’d analysed the different ways the Strava data was situated, I realised that the same method could be applied to any social media platform, app or digital infrastructure that uses personal data. So I decided to change the focus of the paper so it was about the method, not just about Strava.

Donna Haraway coined the term situated knowledges in 1988 to demonstrate that knowledge can never be objective, that it is impossible to see the world (or anything) from a neutral, external standpoint. Haraway calls this fiction of objectivity “the god trick,” a fantasy that omniscience is possible.

With Facebook and Google Earth and smart homes and smartphones vastly more data is collected about us and our behaviour than when Haraway wrote about situated knowledge. The god trick as it occurs when big data are involved has been given many names by researchers of digital media: Anthony McCosker and Rowan Wilken write about the fantasy of “total knowledge” in data visualisations, José van Dijck warns against an uncritical, almost religious “dataism“: a belief that human behaviour can be quantified, and Lisa Gitelman points out that “Raw Data” is an Oxymoron in her anthology on the digital humanities. There is also an extensive body of work on algorithmic bias analysing how machine learning using immense datasets is not objective but reinforces biases in the data sets and inherent in the code itself (there are heaps of references to this in the paper itself if you’re curious!).

Situated data analysis provides us with a method for analysing how data is always situated, always partial and biased. In my paper I use Strava as an example, but let’s look at a different kind of data: how about selfies?

When you take a selfie and upload it to Facebook or Instagram, you’re creating a representation of yourself. You’ve probably taken selfies before, and you’ve probably learnt which camera angles and facial expressions tend to look best and get you the most likes. Maybe you’re more likely to get likes if you post a selfie of yourself with your friends, or in front of a tourist attraction, or wearing makeup – and probably what works best depends on your particular community or group of friends. You’ve internalised certain disciplinary norms that are reinforced by the likes and comments you get on your selfies. So at this level, there’s a disciplinary power of sorts. This is the first layer, where your selfie is situated as a representation of yourself that you share with friends or a broader audience.

Facebook and Instagram and other platforms will show your selfie to other people for their own purposes of course. Their goal is to earn money from advertisers by showing people content that will make them spend more time on the platform, and also by gathering personal data about users and their behaviour. Here your selfie is situated differently, as a piece of content shown to other users. There is an environmental power happening here – the environment (in this case the social media feed) is being altered in order to change peoples’ behaviour – for instance I might pause to look at your selfie and then happen to notice the ad next to it.

Drawing of Facebook interface

A third level at which the data of your selfie is situated happens when Clearview illegally scrapes your selfie, along with three million other selfies, and uses them as a dataset for training facial recognition algorithms. Next time you are at a protest or rally, the police or an extremist group might use Clearview to identify you. Perhaps in the future you’ll be banned from entering a shop or a concert or a country because your face was identified as being at that protest. Maybe a nightclub or a university will have a facial recognition gate and will only let in people without a history of attending protests. Obviously this was not something you had in mind when you uploaded that selfie – but it’s an example of how placing data in new contexts can make a huge difference to what the data means and what it can be used for. A facial recognition gate that refused entry to people would also be a kind of environmental power – the environment is simply changed so you can no longer act in certain ways.

Although I started this paper in the INDVIL project, I finished it while working on the Machine Vision project, and the selfie example is just one example of how situated data analysis is relevant for machine vision technologies.


Discover more from Jill Walker Rettberg

Subscribe to get the latest posts sent to your email.

Leave A Comment

Recommended Posts

Academics in Norway: Sign this petition asking for research-based discussions of how to use AI in universities

I just signed a petition calling for Norwegian universities to use research expertise on AI when deciding how to implement it, rather than having decisions be made mostly administratively. ,  If you are a researcher in Norway, please read it and sign it if you agree – and share with anyone else who might be interested. The petition was written by three researchers at UiT: Maria Danielsen (a philosopher who completed her PhD in 2025 on AI and ethics, including discussions of art and working life), Knut Ørke (Norwegian as a second language), and Holger Pötzsch (a professor of media studies with many years of research on digital media, video games, disruption, and working life, among other topics).  This is not about preventing researchers from exploring AI methods in their research. It is about not uncritically accepting the hype that everyone must use AI everywhere without critical reflection. It is about not introducing Copilot as the default option in word processors, or training PhD candidates to believe they will fall behind if they do not use AI when writing articles, without proper academic discussion. Changes like these should be knowledge-based and discussed academically, not merely decided administratively, because they alter the epistemological foundations of research. Maria wrote to me a couple of months ago because she had read my opinion piece in Aftenposten in which I called for a strong brake on the use of language models in knowledge work. She was part of a committee tasked with developing UiT’s AI strategy and was concerned because there was so much hype and so few members of the committee with actual expertise in AI. I fully support the petition. There are probably some good uses for AI in research, but the uncritical, hype-driven insistence that we must simply adopt it everywhere is highly risky. There are many researchers in Norway with strong expertise in AI, language, ethics, working life, and culture. We must make use of this expertise. This is also partly about respect for research in the humanities, social sciences, psychology, and law. Introducing AI at universities and university colleges is not merely a technical issue, and perhaps not even primarily a technical one. It concerns much more: philosophy of science, methodological reflection, epistemology, writing, publishing, the working environment, and more. […]

screenshot of Grammarly - main text in the middle, names of experts on the left with reccomendations and on the right more info about the expert review feature
AI and algorithmic culture Teaching

Grammarly generated fake expert reviews “by” real scholars

Grammarly is a full on AI plagiarism machine now, generating text, citations (often irrelevant), “humanizing” the text to avoid AI checkers and so on. If you’re an author or scholar, they also have been impersonating and offering “feedback” in your name. Until yesterday, when they discontinued the Expert Review feature due to a class action lawsuit. Here are screenshots of how it worked.