Abstract
Introduction
In this article, I consider how the redesign of vehicle dashboards has restructured car-related data processes. I do so by charting the emergence of two such processes enabled by the redesign of vehicle dashboards. Firstly, with the integration of voice-activated navigation systems into vehicle dashboards, ‘geodata’ becomes ‘navigational data’. Here, this transformation is enabled through the implementation of new addressing and speech protocols that radically change the relationship between driver and vehicle, when performing navigational tasks. Secondly, ‘vehicle data’ becomes ‘driving data’ in the convergence, and customization, of dashboard features and functionality. Here, this transformation is enabled through the spatial, aesthetic and operational integration of typically separate aspects of the driving experience (instrument cluster, navigation and entertainment), re-presenting vehicle-related data in new, and novel, ways. In evaluating these concomitant ‘datafication’ processes, I use Mejias and Couldry's (2019: 3) definition, in which datafication involves ‘the transformation of human life into data through processes of quantification, and the generation of different kinds of value from data’.

An illustration of how the What3word grid system works.

The ‘frustration’ of inputting addresses by touch.

The new MBUX ‘Hyperscreen’ with three visible screen areas, wider than the widescreen cockpit. Source: Daimler (2021).

The Hyperscreen visualizing new kinds of driving data (electric vehicle energy ‘boost’ and ‘recuperation’) using a novel ‘clasp’-style display.
Both transformations are enabled through strategic design decisions, persuading drivers to participate in novel practices they might otherwise not. Firstly, through the strategy of ‘representational transparency’ (Agre, 1995: 186), voice activation is depicted as a seamless, unmediated interface (Bolter and Grusin, 1998) between the normal, natural speech of a driver and the vehicle itself. Here, a voice activation system, a novel addressing system and an infotainment system combine to reorganize the navigational activities of a typical driver. By imposing a new ‘grammar of action’ (Agre, 1994: 107), these technologies work to capture automotive navigation in new ‘datafied’ ways. A significant component of this new navigational grammar, in which drivers not only issue ‘vocalized instructions’ (Hind and Seitz, 2021: 24) as opposed to using physical dials or interfaces, is that localized place names (neighbourhoods, streets and buildings) are replaced with ‘unique, three word strings’ (Hind and Seitz, 2021: 24). The result is what Agre refers to as ‘semantic colonization’ (Agre, 1995: 186), as these new grammars impose themselves on the established navigational practices of drivers.
Secondly, through the strategy of control, the driver is persuaded to believe they have full(er) customizable power within, and critically
Here, I interrogate how such systems transform car-related data from one state (geodata and vehicle data) into another (navigational data and driving data). The systems discussed here are representative of broader efforts within the automotive industry to transform the vehicle itself into ‘mobile spatial media’ (Alvarez Leon, 2019a) or wholesale into a ‘platform’, through which the use of data is integral (Alvarez Leon, 2019b; Wilken and Thomas, 2019).
While the automotive industry is not alone in making use of data streams produced as by-products (Pridmore and Mols, 2020; Thatcher, 2014), there are nonetheless unique challenges to be found in this application, such as interpreting spoken destinations or disambiguating common street names. These provide the possibility of articulating distinct aspects of datafication (Sadowski, 2019; Van Dijk, 2014) within vehicles, and beyond other spaces such as the home (Maalsen and Sadowski, 2019; Pridmore et al., 2019). The effect is manifold: the cultivation of new kinds or streams of data (touchscreen interfaces augmented with voice activation, mirrors replaced by recordable cameras), new ways of representing established kinds of data (vehicle speed or fuel levels) and altered practices in relation to both (entering destinations, accelerating).
The aim of the article is thus threefold: firstly, to map where and how datafication takes place within the car; secondly, to establish the role of vehicle dashboards in enabling this datafication; and thirdly, to identify the strategies that come to shape the nascent datafied driving experience.
In the next two sections, I consider how geodata is (and is not) transformed into navigational data, and how vehicle data is transformed into driving data. In the former, I discuss how some kinds of geographical information escape datafication, while others are subject to a practice I call ‘re-datafication’. In the latter, I discuss how vehicle data is ‘surfaced’ as driving data, generating alternative kinds of value in the process (Mejias and Couldry, 2019), connected, for instance, to the development of electric vehicles. In the subsequent section, I explore how dashboard ‘convergence’ (Hind and Gekker, 2019; Jenkins, 2006) enables these transformations. In the final two sections I discuss two cases: a voice-activated navigation system built on the What3words platform, and a ‘widescreen’ dashboard in a range of Mercedes-Benz vehicles.
Navigational data: ‘Turn-by-turn’
Geodata is data with a geographical, locational or spatial component (Lauriault, 2017; Leszczynski, 2017). Typical examples include coordinates, a house address or a postcode. Geodata may be relatively precise (GPS location) or represent a general geographic area (a state or municipality), with scholars attentive to the spatialities of data, more broadly (Crampton et al., 2013; Shelton, 2017). Furthermore, geodata can be used for various kinds of navigational tasks: to orientate oneself on a hike, to enable the delivery of consumer goods or to arrange a meeting with friends. In this section, I consider first how some kinds of geographical information are
Navigational data is dependent on geodata. Geodata might be
Navigational data is
This is not, however, to say that such geographical information has not been datafied already in other contexts, or that there have not been attempts to ‘datafy’ them in an automotive setting either. The Google Streetview and Google Earth projects have long enabled the capture of buildings, landmarks and street scenes in a visual format, enabling various kinds of datafied experiences, from ‘Dionysian adventures’ (Kingsbury and Jones III, 2009: 502) to geo-based forms of evidencing and intervention (Parks, 2009). Moreover, within the automotive industry, there have been numerous attempts to capture the aspects of the built environment for the purposes of autonomous driving. HERE's HD Live Map (HERE, 2021), Mobileye's Road Experience Management (REM) (Mobileye, 2021) and Waymo's 3D maps (Waymo, 2020) all collect localized, topographic data ‘from the height of the curb’ to the ‘width of an intersection’ (Waymo, 2016). A recent patent filing by Ford details a system capable of reading advertising hoardings to be transmitted into the vehicle itself (Rose, 2021). However, at present, none of these systems constitute a lived driving experience, or are embedded within everyday automotive navigation.
Secondly, geodata may become enrolled into navigational practices, but stubbornly remain as geodata. Typically, such geodata might be found in fixed road signs with place names and numerical distances, or in traffic lights, in which geographical information has already been extracted, coded and displayed. Yet, while enrolled into navigational practices (think of how many times a road sign has been interpreted by a passing driver), this geodata is not transformed into navigational data, as no further quantification has taken place, even though different kinds of value are arguably being generated through its enrolment in many specific, unique navigational tasks each day. Thus, it cannot be said that additional datafication happens in such a case.
As before, however, this has not prevented attempts to turn this kind of geodata into navigational data. The aforementioned mapping systems developed for autonomous vehicles do not only typically capture localized topographic data (curb heights and intersection widths) but also intend to extract geodata displayed in directional signs and traffic lights. In a demonstration of Mobileye's REM system on the roads of Munich, Germany, one is able to see a real-time representation of approaching traffic lights (Mobileye, 2020); data that is collected by vehicles equipped with Mobileye technology. However, once again, this is not a universal driving experience at present. As a result, this geodata, while very much enrolled into automotive navigation, is not typically ‘re-datafied’.
Thirdly, geodata may be enrolled into navigational practices that transform it into navigational data. Ordinarily, this activation occurs through digital devices such as integrated sat-navs, standalone sat-navs, generic or single-purpose map apps (Brown and Laurier, 2012; Chesher, 2012; Hind, 2019; Hind and Gekker, 2014). With each, geodata undergoes a second round of datafication, as pre-existing datapoints (house numbers, postcodes, etc.) are transformed into ‘turn-by-turn’ datapoints in the act of navigating via a digital device, or what Singh et al. (2019: 287) refer to as a ‘turn-taking machine’, thus generating immediate navigational value to the driver. Rather than datafication, per se, this ‘re-datafication’ instead further transforms one kind of data into another. These transformations are necessarily performed in part (or whole) through specific technical relations between devices, apps, platforms and infrastructures, which differ between the examples given above. 2
This final category is of particular interest because of how it usurps these other modes. When geodata is transformed into navigational data, it captures and codifies the navigational experience, ‘turn-by-turn’. Here, this re-datafication is enabled by the imposition of these new infrastructures, which seek to re-route navigational activities through them instead. This is what Couldry refers to as a ‘general project of social re-construction’ (Couldry, 2020: 1140), as ‘life actions previously performed elsewhere’ (Mejias and Couldry, 2019: 3) are carried out through a digital device. However, complicating the abstractive power that datafication offers, these other sources of navigational information are also rather stubborn: they stand in the way – sometimes quite literally – of datafication, limiting the extent to which users might require, or interact with, navigational devices. It is this contestation between reliable, appropriate, and accurate sources of navigational information that I will turn to in the first case study.
In summary, while in the first category (geographical information in navigation) value generation might occur, this is not through datafication. In the second category (geodata as geodata), datafication has already occurred, and value generation does take place, but not through the transformation of geodata into navigational data. In the final category (geodata into navigational data), re-datafication can be said to occur, through which both quantification and value generation take place in the act of navigation. What is important to mention about all three categories is the transformation (or not) of geodata is dependent on certain operational conditions being met. Some sources of geographical information remain untouched by datafication, others have already been extracted and codified, while others still have become more recently exposed to emergent, recurring and persistent re-datafication efforts.
In the next section, I discuss the transformation of another car-related datafication process, the ‘surfacing’ of vehicle data as driving data.
Driving data: Indicator-by-indicator
Vehicle data is data that is generated by the car for the technical operation of the vehicle itself. Ordinarily, the transmission of vehicle data is enabled through a centralized communication system, referred to as a ‘vehicle bus’. Various protocols have been developed over the years to standardize communication procedures, ensuring that components can safely and effectively speak to each other. ‘Electronic control units’ (ECUs) that control a suite of functions within the car are dependent on vehicle data, as well as the vehicle bus that sends such data throughout the vehicle. Examples include the engine control unit (for controlling the engine), the transmission control unit (for controlling gear transmission) or the control unit for anti-lock braking systems. Luxury cars now have between 100 and 150 ECUs (Stoltzfus, 2017; Winning, 2019), despite attempts to consolidate them into multi-functional systems (Intel, 2018).
As vehicle data facilitates machine-to-machine communication, it does not need to be seen by, or made interpretable for, the driver. However, much like geodata is transformed into navigational data, so vehicle data can be transformed into driving data, surfaced through representation in indicators, dials, lights or some other visual (or audio) form. 3 Driving data is vehicle data that is activated in the process of driving a vehicle. While vehicle data lies under the bonnet, (usually) quietly ensuring the vehicle is operating properly, driving data is presented to the driver to aid decision-making. Like navigational data, driving data is brought into being through the various stages or moments in the driving experience.
Vehicle data is surfaced as driving data in multiple ways. Firstly, data can be
Secondly, such data can be
Thirdly, data can be
In summary, in the first category (proximal surfacing) datafication occurs under certain circumstances, with recurring, and familiar, forms of value generated between drivers in the process. This is dependent upon the extent to which the vehicle contains components that at least possess the
Dashboard convergence: From spatial to operational integration
Originally a physical board to prevent material from ‘dashing up’ onto the exposed driver (Mattern, 2015), dashboards in contemporary vehicles display an array of phenomena from operation-critical processes (gear, engine temperature) to multimedia options (radio station, Bluetooth connectivity). While the talk of urban and smart city dashboards is now common (Kitchin et al., 2016; Zook, 2017), the discussion of the evolution of vehicle dashboards is less so. Yet over the years, the dashboard has developed from a device made to prevent any physical hindrances to driving, to a multifunctional aid designed to enhance the driving experience.
In this section, I contend that dashboard ‘convergence’ facilitates datafication, in which the otherwise separate interfaces housed within a vehicle dashboard – typically the instrument cluster, navigational assistance and multimedia – are becoming operationally dependent. In this, I go beyond Alvarez Leon's (2019a: 370) suggestion that cars have become ‘integrated media spaces’, arguing that it is through this convergence that the datafication of wholesale vehicle operations is occurring. More specifically, the generation, and (re-)presentation, of both navigational data and driving data is being enabled through the convergence of dashboard functions.
Dashboard convergence is not necessarily a new phenomenon, with innovation in vehicle dashboard design being a constant since the early 20th century. Yet, as Mattern (2015) discusses, the ‘standard package’ of a Ford Model T in 1908 ‘consisted solely of an ammeter, an instrument that measured electrical current’. While early computers only made use of displays to check for ‘errors’ rather than for ‘complex data output or input’ (Thielmann, 2018: 47), early motor vehicles had more immediate ways to inform users of a problem: ‘[w]ater gushing from the radiator, an indicator you hoped
Following Mattern, then the principle of dashboard design from the 1950s onwards exhibits a kind of rationalizing design tendency, ‘driven primarily by aesthetics’ (Mattern, 2015), with fewer vehicle operations needing to be represented (either indexically or symbolically) to the driver. However, this aesthetic desire has given way to a developmental reality in which additional controls, features and options have increasingly become ‘bolted on’ to existing dashboard layouts. Rather than serving to filter and selectively surface critical performance metrics, vehicle dashboards have evolved to ‘expose’ much more of the vehicle's seeming limitless functionality. This dashboard ‘complexity’ (Everdell, 2015) or ‘clutter’ (Thrift, 2004: 56) has been driven by this counter-tendency, desiring to offer a better driving experience, ordinarily at the expense of driving safety or simplicity.
Thus, the dashboard has become an important space for innovation in recent years as new driving features and data flows are represented to enhance this driving experience. Dashboard convergence is a logical effect of the ‘platformization’ of the car, which has made it connectable, modular and interfaceable (Helmond, 2015). Dashboard convergence presents itself as a fix for ‘clumsy’ (Everdell, 2015) dashboard design, offering to resolve these related aesthetic, experiential and safety questions at once. While urban dashboards provide an illusion that cities can be singularly knowable and measurable through an ‘instrumental rationality’ (Kitchin et al., 2016: 94), vehicle dashboards have typically offered an instrumental plurality driven, in part, by the automotive manufacturer's (increasing) reliance on selling optional add-ons, vehicle configurations and soft/hardware updates (Krings et al., 2017; PWC, 2018).
This dashboard convergence is twofold. Firstly, there is a
Secondly, there is
These two types of convergence – spatial and operational – are integral to re-datafication: the transformation of geodata and vehicle data into navigational data and driving data. This horizontal (spatial) and vertical (operational) integration enables the activation, surfacing and/or extraction of these data types, in ways that were either previously unimaginable, technically impossible, difficult to implement or otherwise ‘siloed’. Through re-datafication, the navigational and driving data streams are made more valuable, both for the driver of the vehicle and – just as critically – for the manufacturer. Through this ‘interoperability’ (Wilmott, 2016) previously separate systems (and their accompanying data types) are made to work with each other.
This push towards greater interoperability, crystallized in the convergence of dashboard functionality, is driven in large part either by the ‘full-stack’ capabilities of big tech companies entering the automotive market, or in their emulation by automotive manufacturers in partnership with ‘tier one’ chip, component and system suppliers. In both cases, access to, and control over, the entirety of data streams generated by, and in, a vehicle is regarded as important for the extraction of value in an industry typically burdened by high production costs. As a result, vehicle manufacturers and their parent groups have typically had to be alert to new profit maximization possibilities, whether through mergers and acquisitions, ‘strategic in-sourcing’, vehicle platform consolidation, labour flexibilization or moving production facilities (Pavlínek, 2008, 2020). Capturing and streamlining navigational and driving data streams have enabled vehicle manufacturers to
In the following two sections: I analyse two innovative vehicle dashboard designs: firstly, a voice-activated navigation system based on the unique addressing platform What3words. This can be seen to have transformed geodata into new forms of navigational data. Secondly, a self-styled ‘widescreen cockpit’ designed by Mercedes-Benz, which has arguably surfaced vehicle data as novel forms of driving data. The result of both is an emerging datafied driving experience. I contend that their comparative successes – as examples of dashboard convergence in which data is transformed within the vehicle – rest on two strategies: representational transparency and customizable control. While reflective of broader developments in the digital economy, these strategies take on specific qualities in an automotive context.
Voice-activated navigation as representational transparency
What3words is a proprietary geocode system that divides the world into 3m2 grids, identifiable by a unique three-word string. In doing so, it algorithmically converts underlying geographic coordinates, enabling users to remember locations such as ‘thread.strollers.bumble’ (somewhere in Germany). What3words claim it is superior to established postal systems, with the 3m2 grids enabling users ‘to specify a precise entrance, unlike a street address which identifies an entire building’ (Macgregor, 2020a), and that ‘unlike street addresses which are often duplicated’, What3words locations are ‘all unique’, available in ‘over 40 languages’ (Macgregor, 2020a) (Figure 1).
In 2018, Daimler integrated What3words into its new infotainment system, ‘Mercedes-Benz User Experience’ (MBUX), aboard the new Mercedes-Benz A-Class (Daimler, 2018a). In doing so, they suggested they had ‘moved one big step closer to [their] goal of making the vehicle into a mobile assistant’ such that ‘[i]nputting locations…makes life easier for our customers and ensures a special experience’ (Daimler, 2018a). In this, What3words could be activated through a touchscreen interface, but also via voice control.
As a promotional video demonstrates (Mercedes-Benz, 2018a), 4 the combination of What3words and LINGUATRONIC (MBUX's voice control system) (Daimler, 2018b) is seen as integral to the navigational experience within the A-Class. Rather than the driver using clunky search boxes, unresponsive knobs and buttons, or external apps or sat-navs, the owner merely issues instructions to the vehicle, with What3words deemed ‘the simplest way to talk about location’ (Mercedes-Benz, 2018a). In this, typical (local) addressing systems are rendered confusing and frustrating. As the manufacturer reminds us, street names are rarely unique (Mercedes-Benz, 2018b), street numbers are difficult to differentiate between, and some towns are entirely unpronounceable to unfamiliar users (What3words, 2019). In short, the integration of What3words into the A-Class is set up to provide an improved navigational experience.
However, rather than simplifying vehicle navigation, the convergence of a novel addressing system, a voice control system and a navigational system yields significant ‘praxeological changes’ (Thielmann, 2018: 50) between driver and vehicle. This operational convergence demands drivers follow, very carefully, a set of new conversational protocols to activate the navigational experience. Rather than offering a ‘transparent’ interface, in which the ‘illusion of representational transparency’ (Agre, 1995: 186) between the driver and vehicle is maintainable, What3words ‘remediates’ (Bolter and Grusin, 1998) this navigational experience, inserting a number of new rules drivers must follow.
As a promotional video shows, ‘Sophie’ receives a message on her smartphone: ‘let's meet at hello.page.brand for brunch’ (Mercedes-Benz, 2018a). As she gets into the car, she utters the words ‘hey Mercedes’, before asking the vehicle to ‘take [her] to What3words hello.page.brand’ (Mercedes-Benz, 2018a). Yet rather than instantly generating a route, Sophie is instead given three choices: her intended destination, but also ‘hello.page.brands’ and ‘hello.page.barn’ (Mercedes-Benz, 2018a). While What3words is designed to remove ambiguity, only a plural form of one word distinguishes two results (brand to brand
The activation of navigational data in the use of a voice-activated navigation system results in a peculiar, if not altogether dangerous, experience. According to the company, What3words is accepted by over 80% of the UK's emergency services, often dependent upon reaching individuals in remote locations, quickly (What3words, 2020). Yet, in the examples above, geodata in the form of a familiar, localized address is rendered mute. In its place is reference to a general geographic area (Bayswater, Cranleigh or Send) and a number of either similar sounding (lanes, lands), or singular/plural forms (brand, brands), of proximally located What3words strings.
This erasure – of localized names and places, idiosyncratic, ambiguous or unpronounceable to outsiders – is common to (digital) capitalism (Nicas, 2018; Rose-Redwood et al., 2019; Sotoudehnia, 2018). 5 The automatic translation, or ‘toponymic reconfiguration’ (Rose-Redwood et al., 2010: 454), of established place names into trivial three-word locations, is the key to the reorganization of navigational practice within the vehicle. With this, the driver is expected to change their navigational habits on two counts: firstly, to shift from using touchscreens/knobs to voice control; secondly, to shift from using postcodes and addresses to using randomized three-word strings.
Regarding the former, it alters how geodata is transformed into navigational data. Rather than street names, whole addresses or postcodes being selected from within an addressing database
6
by the user via a touchscreen or dials, three-word strings are spoken and ‘matched’ to the What3words database by LINGUATRONIC. Thus, entirely
Regarding the latter, it constitutes, what Agre (1995: 186) refers to as ‘semantic colonization’, in which new semantic terms usurp others. Not only is one kind of geodata (postcodes) bypassed, but another kind altogether (three-word strings) created, even acting parasitically on the other to ‘convert’ postcodes into What3word locations. This re-datafication process generates an altogether more problematic form of value rendered not from local road names or through a standardized (and nationalized) postcode system, but from randomized three-word strings. While it is a strategy that purports to offer representational transparency, it only succeeds in adding further layers, as new navigational ‘procedures’ are designed (Garfinkel, 1996; Thielmann, 2019). This semantic colonization is made possible through the proprietary nature of the system itself, under which What3words data (i.e. vocalized search requests), and the conversion software code, is protected by copyright owned, or licensed, by What3words (What3words, 2021c).
The new conversational protocols also arguably remove the noted ‘interpretative flexibility’ (Pridmore and Mols, 2020: 2) of typical household voice-activated devices, considering their specificity of use within an automotive context, and for strictly navigational tasks. Although LINGUATRONIC bears some similarities with generic, or all-encompassing devices that allow users to access information (weather, news, entertainment, etc.) or control a suite of secondary appliances (lights, heating, music systems, etc.), its integration with What3words establishes a far clearer use case: automotive navigation. Unlike users of other household voice-activated devices, drivers of vehicles equipped with both do not necessarily ‘struggle…to determine what these devices are really for’ (Pridmore and Mols, 2020: 3). Instead, the obvious hurdle becomes re-learning established ways of automotive route-planning and wayfinding, from entering postcodes to using a touchscreen sat-nav.
What is also important to note is that while partnerships with external suppliers are typical within the automotive industry, the integration of third-party digital services such as What3words into the in-car experience is less so. This represents a significant development in the automotive industry, as vehicle manufacturers begin to involve companies with expertise in ‘datafying’ user experiences, such as with Polestar's use of Android Automotive OS (Polestar, 2021) and Ford's partnership agreement with Google from 2023 (Korosec, 2021). One can only expect the type, mode, intensity, and granularity of navigational (re-)datafication to therefore increase in subsequent years as the value of such data is realized.
The widescreen cockpit as customizable control
The 2018 A-Class not only integrates both voice control and a unique address system into the navigational experience, but also showcases a new dashboard design referred to as a ‘widescreen cockpit’ (Mercedes-Benz, 2018c). In this new dashboard, previously separate screens for the instrument cluster and the infotainment system (MBUX) are combined into a single entity operated either via touch or voice. The housing for the instrument cluster now extends towards the centre of the dashboard, with the infotainment system stretching across to the driver. In 2021, Mercedes-Benz launched the ‘Hyperscreen’, a 56-inch display stretching the full width of the interior (Hawkins, 2021), enabling even deeper forms of customizable control (Figure 3).
In another promotional video, we see a YouTuber explain the features of the widescreen cockpit (Mercedes-Benz, 2018c). As it begins, he guides viewers through the ‘three screen set-ups available’ to the driver (Mercedes-Benz, 2018c): two 7-inch displays, one 7-inch display and one 10.25-inch display, and two 10.25-inch displays, meaning ‘drivers can now customize their display screens’ (Mercedes-Benz, 2018c). In this, there is a double spatial convergence: firstly, the integration of both instrument cluster and infotainment system into a single entity: the widescreen cockpit; secondly, a variable spatial convergence (the ‘three screen set-up’) in which the new systems embedded within the interface can be resized. Either users can select two smaller screens (less intrusive), enlarge the media screen (for scrolling through songs, etc.) or choose the full, ‘widescreen’ experience.
Next, he describes how drivers can perform this customization: For example, if you wanted to rearrange the order of these apps [on the infotainment system], you just press and hold on one of the apps, in this case the navigation app, slide it across to where you want it, and then tick confirm and it locks it nicely in place. (Mercedes-Benz, 2018c)
While the claim of a ‘zero layer’ display is certainly another example of the strategy of representational transparency, I argue it represents another, novel, tendency in vehicle dashboard design. While media functions within cars have been ‘appified’ (Morris and Murray, 2018) for a while, instrument clusters and driving-related features have typically remained off-limits. Here, app-ecosystems manifesting as ‘permanent excess: excessive downloads, excessive connections…and excessive amounts of information’ (Mellamphy et al., 2015: 2) have been constrained by vehicle safety standards that govern the style, size, illumination and appropriate abbreviations (ABS, ESC, etc.) for display icons. However, as the above, and another later video suggests, there are now ‘many ways to customize your digital dashboard’ (Mercedes-Benz, 2019) that increasingly extend beyond media screens and the integration of apps (Goggin, 2021), and into the representation of vehicle states. That is, to how vehicle data is internally surfaced and transformed into driving data.
This, I argue, constitutes a novel application of Phil Agre's ‘empowerment and measurement regime’ (Agre, 1995: 176), in which drivers are ‘empowered’ to customize their digital dashboards, ‘personalizing’ their own driving experience. This is akin to how Tanya Kant (2015: 34) describes Facebook apps as ‘tools for self performance’, framing and staging the construction of identities. Yet through novel measurement techniques – datafication by another name – vehicle manufacturers can further ‘manage’ the driving experience.
In the 2021 update, this regime is made even clearer and increasingly proactive, as app presentation is ‘supported by artificial intelligence’ through a ‘context-sensitive awareness [that] is constantly optimized by changes in the surroundings and user behaviour’ (Daimler, 2021). Here, the ‘MBUX Hyperscreen continually gets to know the customer better and delivers a tailored, personalized infotainment and operating offering before the occupant even has to click or scroll anywhere’ (Daimler, 2021). What display to show when, and for what purpose, is therefore a kind of situated surfacing in which the MBUX system offers a greater level of customizable control without the need for direct user interaction. What becomes clearer, then, is how the Hyperscreen is intended to function like a recommendation engine, hooking, looping or ‘trapping’ (Seaver, 2018) the driver into certain kinds of driverly attention not dissimilar to those found on other cultural, digital platforms (Bucher, 2012; Pederson et al., 2021; Siles et al., 2020).
The new ‘clasp’-style display also renders the driving experience anew. Designed for forthcoming electric models, it internally surfaces new driving data streams. Rather than a combination of speedometer and tachometer common to petrol/diesel vehicles, the Hyperscreen uses a ‘spatially moving clasp’ to represent the ‘boost’ and ‘recuperation’ stages of electric motor use (Daimler, 2021). Here we witness a re-datafication process, as outlined previously, that generates new data, from a new data source, using novel data metrics (motor ‘boost’ not engine revolutions per minute). In this, the redesign of a critical dashboard indicator is intended to cultivate a sensitive, mindful driver, embodying the fluid, regenerative qualities of the electric motor powering the vehicle (Figure 4).
With these technologies, spatial convergence (both fixed and customizable) leads to an operational convergence, in which both the media display (with app content) and the driving display (with driving data) are presented on the same ontological plane. In this, both are made customizable, not only to the driver's interests, tastes, feelings and passions (Sheller, 2004), but also to their driving
Conclusions
Vehicle dashboards are being radically redesigned to represent data differently within the car. With this, vehicle manufacturers are moving beyond the representation of typical features such as fuel levels and speed, or simply the digitization of previously mechanical indicators such as fingers and dials. As vehicles are becoming platformized, new data streams are being generated, sometimes derived from entirely novel operational states. New techniques are being employed to represent this data, and new strategies to convince drivers of their utility.
In this article, I have considered how the redesign of vehicle dashboards has restructured both navigation and driving through two datafication processes: firstly, in the transformation of geodata into navigational data, and secondly, in the surfacing of vehicle data as driving data. The spatial and operational convergence of navigation, entertainment and driving features within the car has been a critical enabler of these processes. To convince the drivers of the value of these design changes, two strategies have been deployed by manufacturers. Firstly, ‘representational transparency’ (Agre, 1995: 186) in which new vehicle interfaces are sold as fixes to existing systems deemed annoying, confusing or complicated to use. Secondly, ‘customizable control’ in which drivers are afforded greater ability to personalize vehicle displays. These strategies, I have argued, are specifically enabled by re-datafication, in which existing forms of data are transformed into others, actively shaping the experience of driving a car.
To evidence the transformation of geodata into navigational data, I discussed the integration of What3words – a unique addressing system – into a Mercedes-Benz navigation system. Constituting a rather complex operational convergence of multiple systems and functions, navigational capabilities become primarily enabled through voice activation. Here, geodata is activated as a very particular kind of navigational data, posing both novel practical issues for the driver, as well as constituting a kind of ‘semantic colonization’ (Agre, 1995: 186) in which placenames (i.e. geodata) are rendered as randomized three-word strings for the purposes of navigation.
To elucidate the transformation of vehicle data into driving data, I turned to Mercedes-Benz's MBUX system. Here there is an obvious spatial convergence as driving, navigational, and entertainment features are fitted into a single ‘widescreen’ dashboard, or ‘Hyperscreen’. However, the MBUX system also demonstrates another more subtle convergence, in which the different display screens of the vehicle dashboard can be customized, enabling driving data to reflect a driver's interests, tastes or personality. As an imminent new version of the MBUX system demonstrates, this customizable control is to be further enhanced through artificial intelligence, enabling a ‘situated surfacing’ of context-relevant functions. These customizable forms, far from being additional elements or features, subtly yet substantially reconfigure the experience of driving. Indeed, rather than a curious side effect, they constitute intended design effects: techniques to transform the driver and their ways of driving.
Thus, this article has sought to add to emerging work on data and platforms, considering how automobiles are being subjected to datafication processes. In the process of transforming both geodata and vehicle data into navigational data and driving data, respectively, new datafied driving experiences are emerging. These, I argue, are worthy of ongoing investigation, as they generate unique effects that help to better understand how datafication is shaping the world-at-large and on the road.
