top of page
Search

Hito Steyerl - Proxy politics and relational photography

  • departmentgalleryh
  • Jan 24, 2018
  • 3 min read

Hito Steyerl

Proxy Politics: Signal and Noise

A while ago I met an extremely interesting software developer who was working on smartphone camera technology. Photography is traditionally thought to represent what is out there by means of technology, ideally via an indexical link. But is this really true anymore? The developer explained to me that the technology for contemporary phone cameras is quite different from traditional cameras: the lenses are tiny and basically crap, which means that about half of the data being captured by the camera sensor is actually noise. The trick, then, is to write the algorithm to clean the noise, or rather to discern the picture from inside the noise.

But how can the camera know how to do this? Very simple: it scans all other pictures stored on the phone or on your social media networks and sifts through your contacts. It analyzes the pictures you already took, or those that are associated with you, and it tries to match faces and shapes to link them back to you. By comparing what you and your network already photographed, the algorithm guesses what you might have wanted to photograph now. It creates the present picture based on earlier pictures, on your/its memory. This new paradigm is being called computational photography.1

The result might be a picture of something that never even existed, but that the algorithm thinks you might like to see. This type of photography is speculative and relational. It is a gamble with probabilities that bets on inertia. It makes seeing unforeseen things more difficult. It will increase the amount of noise just as it will increase the amount of random interpretation.

And that’s not even to mention external interference into what your phone is recording. All sorts of systems are able to remotely shut your camera on or off: companies, governments, the military. It could be disabled in certain places—one could for instance block its recording function close to protests or conversely broadcast whatever it sees. Similarly, a device might be programmed to autopixelate, erase, or block secret, copyrighted, or sexual content. It might be fitted with a so-called dick algorithm to screen out NSFW (Not Suitable/Safe For Work) content, automodify pubic hair, stretch or omit bodies, exchange or collage context, or insert location-targeted advertising, pop-up windows, or live feeds. It might report you or someone from your network to the police, PR agencies, or spammers. It might flag your debt, play your games, broadcast your heartbeat. Computational photography has expanded to cover all of this.

It links control robotics, object recognition, and machine learning technologies. So if you take a picture on a smartphone, the results are not as premeditated as they are premediated. The picture might show something unexpected, because it might have cross-referenced many different databases: traffic control, medical databases, frenemy photo galleries on Facebook, credit card data, maps, and whatever else it wants.

Relational Photography

Computational photography is therefore inherently political—not in content but in form. It is not only relational but also truly social, with countless systems and people potentially interfering with pictures before they even emerge as visible.2 And of course this network is not neutral. It has rules and norms hardwired into its platforms, and they represent a mix of juridical, moral, aesthetic, technological, commercial, and bluntly hidden parameters and effects. You could end up airbrushed, wanted, redirected, taxed, deleted, remodeled, or replaced in your own picture. The camera turns into a social projector rather than a recorder. It shows a superposition of what it thinks you might want to look like plus what others think you should buy or be. But technology rarely does things on its own. Technology is programmed with conflicting goals and by many entities, and politics is a matter of defining how to separate its noise from its information.3

So what are the policies already in place that define the separation of noise from information, or that even define noise and information as such in the first place? Who or what decides what the camera will “see”? How is it being done? By whom or what? And why is this even important?


 
 
 

Comments


Gallery House, Leeds LS1, UK

  • Instagram - Black Circle

©2018 by Department. 

bottom of page