The Good, the Bad and the Ugly

Blogging Archaeology

 

I am here snicking in a late (and brief) entry to Doug’s archaeology blogging carnival. This month’s challenge is to set out the best, the worst and the ugly about archaeological blogging. So here goes…

The Good

My own experience is that academic (not necessarily just archaeological) blogging is at its best when it emerges from some real-world collaborative or communal activity. In 2012, I was fortunate enough to spend a couple of weeks participating in IUPUI’s Spatial Narratives Summer Institute, funded by the NEH. There was an official blog for this to which we all contributed, but many of us blogged independently about the experience, and the ideas that we were developing.  It can be no coincidence that my own post composed in this period, ‘Deep Maps in Indy‘ is my most viewed blog article ever. Apart from the fact that we generally referred to each other’s posts, thereby increasing our page views and likes (see below), this contributed to a sense of shared purpose and common cause – and this is especially so when one is in the company of great archaeology/cultural heritage bloggers such as Mia Ridge. The same was true of the CAA2012 session on the ‘Archaeology/Digital Humanities Venn Diagram session, which was subsequently Storified by Graeme Earl. Again, providing a sense of coming together from the real world, and continuity through a variety of different perspectives.

The Bad 

I would go with what many others have said on this subject. One of the worst and most frustrating aspects of academic blogging is low hits/visitor numbers. As Doug says, one feels that one is talking to a brick wall. I have to say I get less hung up on getting low volumes of comments on my posts (although getting any comments, good or bad, is always very welcome). My suspicion is that this aspect of blogging is being eroded by the Twittersphere — if you have something to say about a post, chances are you’ll tweet your reaction rather than using the Comment button. Whether positive or negative, this can actually be a good thing, drawing new readers to your blog and increasing your profile. That is unless, of course, you have some meaty response to make to a posting that could not possibly fit into 140 characters — but is perhaps the increasingly Twitterfied internet drawing blog readers away from that kind of reaction?

The Ugly

Having read some of the hair-raising examples in the Blogging Carnival of the things that can go so badly wrong with blogging and tweeting – for example getting fired for saying the wrong thing about one’s employers – I would have to say that I have, as of today, been spared any such experiences (and hope very much that things continue so).

The most apt sense of ugliness in reference to my own blogging is the literal one. I am not at all enamoured with my rather prosaic WordPress layout, but alas all the time I have to give to blogging is spent writing content rather than on aesthetics. But I know that is not what is really meant by the question.

Why blog?

Blogging Archaeology

 

From @OpenAccessArch over at http://dougsarchaeology.wordpress.com comes a simple, yet important and engaging question:

Why blogging? – Why did you, or if it was a group- the group, start a blog?

This is part of a so-called Blogging Carnival in the run up to next year’s SAA. So here goes…

The immediate answer is straightforward enough. I started my – subsequently rather neglected – blog because I wanted a platform where I could post personal musings and viewpoints, of generally low scholarly import, which I would not expect any peer-reviewed environment to publish, and probably would not want them to. Also, as the years advance and the brain cells retreat, having an archive, however irregular and infrequently accessioned, of what I was thinking about a particular subject at a particular time, is inherently useful. For example, I was having a discussion on Friday last week about the A G Leventis Cyprus Gazetteer project and its software requirements: we touched on the difficulties of defining lines between historic territories for which there were and are no contemporary maps (for example the later prehistoric periods in Cyprus). I remembered that I had blogged on a similar question that arose in the CHALICE project three years ago or so. Simply as an aide memoire of how we approached the problem back then helped a lot.

However, Doug’s question raises a deeper issue about ‘why blog?’, as opposed to ‘why do I blog?’ I was taken back to my first-year garret at Durham University, where – in 1995, before the ubiquity of email, before the e-publishing revolution (has there been one?) and, certainly, before Web 2.0 – I read Paul Bahn’s ‘Bluffer’s Guide to Archaeology’. One thing from this terrific little book that stuck with me was Bahn’s railing against those who dug but did not publish. Excavation was the experiment that destroyed its subject, not to publish your results promptly was (and is) an abdication of a sacred responsibility: Bahn spoke of an (unnamed) professor, a leader of his field, who had published nothing for decades as being ‘the clot that blocks the system’. But since 1995, the whole concept of ‘publishing’, never mind ‘scholarly publishing’ has been transformed out of any possible recognition by the web. While I doubt that anyone attending SAA needs to be told this, the question of ‘why blog’ in archaeology makes me wonder if we have really thought through the issues that change has wrought for our subject in as much detail as they require. In some ways perhaps we have. The thoroughly excellent Journal of Open Archaeological Data has bought us the concept that (digital) data can be published alongside scholarly articles, provided they are Open Access, and in a trusted repository which has a long term sustainability plan, and evidence of durability in the future.

But you only have to look at the mass of links the Doug alone has identified to see that any the age of sourcing archaeological discourse to only ‘professional’ or ‘official’ channels is long past. Intelligent and informed comment and analysis sits alongside the great mass of everything out there. While blogging will never, and should never, try to replace scientific documentation of archaeological site data, its very informality can drive interest in the quirky, the unusual, in subjects neglected by the scholarly discourse. It moves us even beyond the Hodderian notion of ‘multivocaility’, allowing a platform not merely many voices, but the many narratives that those voices tell. This has happened elsewhere in the Digital Humanities – witness Melissa Terras’s work on resource creation via amateur digitisation, which seeks out those topics neglected by official memory institutions. What is archaeological blogging but the ‘amateur’ – in the primary sense of the word – done for love – digitzation of ideas, outside the formal framework of documentation and publication?

Most of us can probably agree that the writing is on the wall (it has been for decades) for the Gibbonesque views of the past, a past told through the elite and literate eyes. In their marvellously entertaining UnRoman Britain (2010, The History Press), Miles Russell and Stuart Laycock note that “[i]t is perhaps the high visibility and obvious distinctiveness of Rome’s archaeological footprint that has caused disproportionate focus upon things that are more ‘Roman’ than the more ‘normal’ aspects that are, to coin a phrase, ‘UnRoman’” (21). Blogging, and the integration of other kinds of social media with the archaeological discourse, provide space for the discussion of such perspectives (there are some fantastic examples of this – eg Rita Roberts’s blog and Bones Don’t Lie), and encourage the process of re-evaluation and scrutiny of established narratives.    

Image

So – blogging in archaeology means we do not have to extend the monolithic method-based systems which have bought us the established narratives by which we have come to know both the near and distant past, rather it allows us to expand (on) them by allowing those who are curious enough to form alternative narratives. Ian Hodder’s dictum that ‘interpretation occurs at the trowel’s edge’ may still hold true, but the changing nature of archaeological data and publishing means it can occur on the keyboard too. And this is what the SAA blog carnival is such a great idea.

Gazetteer of Byzantine Cyprus

Work is getting underway on our Gazetteer of Medieval Cyprus project. By ‘our’ I mean myself and my KCL colleagues Tassos Papacostas and Charlotte Roueche, and various others in the Department of Digital Humanities whose contributions will be kicking in soon. The project as a whole is funded by the A. G. Leventis Foundation. This is a pilot gazetteer to develop a methodology, which will be tested in the online publication of a small body of Byzantine material, already assembled by Tassos. The long-term aim is to provide archaeologists and historians studying Cyprus at any period with a freely available set of tools and skills. The medium-term aim is to provide a well-structured framework for the digital analysis and publication of materials from Byzantine Cyprus, from the end of Late Antiquity to the period of the Crusades (c. AD 650-1200). This will subsequently serve as the basis for other independently funded but interoperable projects to display, interlink and contextualize their data.

Image

The collaborative landscape is fast moving. We have been talking to the excellent Pelagios project, and hope that our GBC will contribute data to the Pelagios federation. The exemplum is based around a Cyprus Gazetteer of places. The aim here is to draw on existing resources in order to build a stable and usable resource; while our work will only focus on ‘Byzantine’ places, the resultant resource will be designed to be  steadily expanded.  We will begin by linking toponym entries in the authoritative Complete Gazetteer of Cyprus / CGC (Christodoulou and Konstantinidis 1987) with their equivalents in GeoNames. This will mean, in the future, that ancient sites will be searchable by the modern administrative areas in which they lie, as well as being linkable via archaeological attributes such as feature type, time period, etc. It will also allow us to build a data structure that will be linkable to other geospatial resources on the web. A basic URI structure has been developed, which will allow the geographical hierarchy in the database to be described, and to be infinitely extensible. We begin by aligning placenames in the CGC with those in geonames, and then treat each archaeological entity identified in the exemplum Architectural Catalogue as a ‘child’ of that unique geographical entity, using the unique GeoNames reference number. The resource we are currently developing will have a user interface which will allow users to add archaeological entity points to this structure.

For me, the key interest of this project lies in the inversion of scale it brings to the digital gazetteer world. Many existing digital gazetteers, such as GeoNames, deal with very large geographic regions but the data associable with them is very ‘thin’ (for more discussion see Linda L. Hill’s Georeferencing: The Geographic Associations of Information, and my 2007 review of it). This, I think, is a natural imperative of the Geospatial Web to expand to wider and wider coverage. Cyprus, however, represents a relatively small geographic area, with a very thick and chronologically complex layer of data, with interconnections across the Aegean and Near East from the Bronze Age to the medieval period. This will present us with exciting opportunities to test how locations are ‘attested’ across many disparate sources, and how those attestations can be most usefully presented and documented.

Reconstruction, visualization and frontier archaeology

Recently on holiday in the North East, I took in two Roman forts of the frontier of Hadrian’s Wall, Segedunum and Arbeia. Both have stories to tell, narratives, about the Roman occupation of Britain, and in the current period both have been curated in various ways. At both, the curating authorities (Tyne and Wear Museums), with ongoing archaeological research being undertaken by the fantastic WallQuest community archaeology project.

The public walkthrough reconstructions of what the buildings and the contents might have been like at both sites pose some interesting questions about the nature of historical/archaeological narratives, and how they can be elaborated. At Segedunum, there is a reconstruction of a bath house. Although the fort itself had such a structure, modern development means that it is not in the same place, nor does the foundations of the reconstruction relate directly to archaeological evidence. The features of the bath house are drawn from composite analysis of bath houses from throughout the Roman Empire. So what we have here is a narrative, but it is a generic narrative: it is stitched together, generalized, a mosaic of hundreds of disparate narratives, but it can only be very loosely constrained by time (a bath house such as that at Segedunum would have had a lifespan of 250-300 years), and not to any one individual. we cannot tell the story of any one Roman officer or auxiliary solider who used it.

Image
Reconstructed bath house at Segedunum

On the other hand at Arbeia, there are three sets of granaries, the visible foundations all nicely curated and accessible to the public. You can see the stone piers and columns that the granary floors were mounted on, to allow air movement to stop the grain rotting. Why three granaries for a fort of no more than 600 occupants? Because in the third century, the Emperor Severus wanted to conquer the nearby Caledonii; and for his push up into Scotland we needed a secure supply base with plenty of grain.

Image
Granaries at Arbeia, reconstructed West Gatehouse in the background

This is an absolute narrative. It is constrained by actual events which are historical and documented. At the same fort is a reconstructed gateway, which is this time situated on actual foundations. This is an inferential narrative, with some of the gateway’s features being reconstructed again from composite evidence from elsewhere (did it have two or three stories? A shingled roof? We don’t know, but we infer). These narratives are supported by annotated scale models in the gateway structure which we, they paying public (actually Arbeia is free), can view and review at our leisure. This speaks to the nature of empirical, inferential and conjectural reconstruction detailed in a forthcoming book chapter by myself and Kirk Woolford (of contributions to the EVA conference, published by Springer).

Narratives are personal, but the can also be generic. In some ways this speaks back to the concept of the Deep Map (see older posts). The walkthrough reconstruction constitutes, I think, half a Deep Map. It provides a full sensory environment, but is not ‘scholarly’ in that it does not elucidate what it would have been like for a first or second century Roman, or auxiliary soldier to experience the environment. Maybe the future of 3D visualization should be to integrate modelling, reconstruction, remediation, and interpretation to bring available (and reputable) knowledge from whatever source about what that original sensory experience would have been – texts, inscriptions, writing tablets, environmental archaeology, experimental archaeology etc. In other words, visualization should no longer be seen as a means of making hypothetical visual representations of what the past might of been, but of integrating knowledge about the experience of the environment derived from all five senses, but using vision as the medium.  It can never be a total representation incorporating all possible experiences under all possible environmental conditions, but then a map can never be a total representation of geography (except, possibly, in the world of Borges’s On the Exactitude of Science).

N=all: Big Data and curation practices

Data curation and digital preservation are often confused, but they are very different things. Terminology is a big problem in this area, especially where common terms from one domain – e.g. ‘curation’ in a musuem or cultural heritage context – are used in another. So can the emerging debate on Big Data help us move forward on a definition of ‘digital curation’?

The current issue of ‘Foreign Affairs’ has a paper by Kenneth Cukier and Viktor Mayer-Schoenberger entitled ‘The Rise of Big Data: How It’s Changing the Way We Think About the World‘. In it they argue that big data represents an epistemic change n how we do statistics, from the model of extrapolating general trends of patterns and populations from small representative random samples, to generalised overviews of entire datasets using data mining. In this world, ‘N=all’. The latter, they argue, are both imperfect and about correlation, rather that causation. I.e. Google claims to be able to track flu outbreaks by correlating certain search terms; but it doesn’t claim to know the actual reason why people made those searches – which would be a ‘traditional’ statistical research question. Recently however, Google’s method dramatically overestimated peak flu levels; a cursory reminder that correlation and causation are very different things.

Cukier and Mayer-Schoenberger argue that big data research means ‘giving up on clean carefully curated data and tolerating some messiness’. They also argue that the process of ‘datafication’ – capturing more and more forms of intangible processes such as friendships (as in Facebook likes) thoughts (Twitter) and professional relationships (LinkedIn) means that this body of data is growing less formal even as it exponentially grows in volume.

For me this raises two questions:

1. What does this mean for a museum-focused definition of ‘curation’. Can we give up on cleanly curated museum and cultural heritage collections and tolerate messiness? If so how, and where does that data come from?

2. By what processes can ‘the museum experience’ be ‘datafied?’ I have an idea forming that this could be to do, at least partly, with removing some of the interaction between audience and collection from being time and space specific. E.g. I don’t have to actually go to the British Museum to encounter all aspects of the experience of the Pompeii exhibition as some of those aspects have been datafied by others (both employees of the BM and other visitors), and I can review them wherever or whenever I like.

The main question is what does ‘big data curation’ actually mean? I am not sure I agree with the definition implied in the Cukier/Mayer-Schoenberger view, where it is precluded. That a curated dataset is necessarily one that is ‘small data’, shaped, presented and processed by a series of well-understood human interventions into a human readable narrative.  However, they also make the very valid point that ‘in a world of big data, it is the most human traits that will need to be fostered – creativity, intuition, and intellectual ambition’. So whereas the present understanding in cultural heritage of what ‘curation’ means – the communication of a story or narrative of a collection of objects for an audience of specialists and/or non-specialists, where N can never = all – in big data terms, it means taking the imperfections of correlation across patterns in big data, and refining these by bridging with communities of experts – experts with the uncomputable human traits to take the broad brushstrokes that software tools are pulling out of our datafied world, and make worldly sense of them.

To crowd-source or not to crowd-source

Shortly before Christmas, I was engaged in discussion with a Swedish-based colleague about crowd-sourcing and the humanities. My colleague – an environmental archaeologist – posited that it could be demonstrated that crowd-sourcing was not an effective methodology for his area. Ask randomly selected members of the public to draw a Viking helmet. You would get a series of not dissimilar depictions – a sort of pointed or semi-conical helmet, with horns on either side. But Viking helmets did not have horns.

Having recently published a report for the AHRC on humanities crowd-sourcing, a research review which looked at around 100 publications, and about the same number of projects, activities, blogs etc, I would say the answer to this apparent fault is: don’t identify Viking helmets by asking the public to draw them. Obvious as this may sound, it is in fact just an obvious example of a complex calculation that needs to be carried out when assessing if crowd-sourcing is appropriate for any particular problem. Too often, we found in our review, crowd-sourcing was used simply because there was a data resource there, or some infrastructure which would enable it, and not because there was a really important or interesting question that could be posed by engaging the public – although we found honourable exceptions to this. Many such projects contributed to the workshop we held last May, which can be found here. To help identify which sorts of problems would be appropriate, we have developed – or rather, since this will undoubtedly involve in the future, I should say we are developing – a four facet typology of humanities crowd-sourcing scenarios. These facets are asset type (the content or data forming the subject of the activity), process type (what is done with that content) task type (how it is done), and the output type (the thing, resource or knowledge produced). What we are now working on is identifying – or trying to identify – examples of how these might fit together to form successful crowd-sourcing workflows.

To put it in the terms of my friend’s challenge: an accurate image of a Viking helmet is not an output which can be generated by setting creative tasks to underpin the process of recording and creating content, and the ephemeral and unanchored public conception of what a Viking helmet looks like is not an appropriate asset to draw from. Obvious as this may sound, it hints that a systematic framework for identifying where crowd-sourcing will, and won’t, work, is methodologically possible. And this could, potentially, be very valuable as the humanities faces increasing interest from well-organized and well-funded citizen science communities such as Zooniverse (which already supports and facilitates several of the early success stories in humanities crowd-sourcing such as Ancient Lives and OldWeather).

This of course raises a host of other issues. How on earth can peer-review structures cope with this, and should they try to? What motivates the public, and indeed academics, to engage with crowd-sourcing? We hint at some answers. Transparency and documentation is essential for the former area, and we found that in the latter, most projects swiftly develop a core community of very dedicated followerswho undertake reams of work, but – possibly like many more conventional collaborations – finding those people, or letting them find you, is not always easy.

The final AHRC report is available: Crowdsourcing-connected-communities.

Visualising the visualisers

The main event to be at in London this month is, of course EVA London 2012, from the 10th to the 12th. There is, apparently, some kind of sporty shindig going on out Stratford way, but I don’t really know very much about it. Maybe it  would have helped if the media had covered it a bit more.

As outgoing editor of EVA’s Proceedings, I’ve prepared some rudimentary visualisations of the conference’s content, which seemed sort of appropriate. The number of papers involved is far too small to have any statistical significance, which means I can’t hope to do anything on the scale of David McCandless’ Information is Beautiful work, or the UCL Digital Humanities Infographic, plus life got in the way of my grand plans to do anything more sophisticated. I nod respectfully in the direction of both however, as the ultimate source of the thought that this might be an interesting exercise. More/better visualisations may well be added in the course of the conference next week, as the inspiration sinks in.

In each case, you can click on the image for a bigger/more legible version.

1. A Wordle of the full text of the EVA London 2012 Proceedings

Image

Wordles can, of course, lead you up erroneous pathways, but herewith the full (unlemmatized) text of the Proceedings (courtesy of www.wordle.net). There is no significance in the layout, but the size of the font is proportional to the frequency of the word.

2. A Wordle of the keywords supplied by accepted EVA authors when they submitted their abstracts

Image

Lemmatized this time, and regularised for spelling variants (e.g. ‘Visualization’ changed to ‘Visualisation’).

3. Topic areas covered (as determined by the EVA London 2012 Programme Committee)

Image

Categorisations made by the EVA Committee when developing the programme back in March. This, and the following two images, were developed using IBM’s Many Eyes software.

4. Institutional affiliations of EVA 2012 authors, by number of instances of authorship or co-authorship

Image

5. Current national affiliations of EVA 2012 authors

Image

Last day in Indiana

It’s my last day in Indianapolis. It’s been hard work and I’ve met some great people. I’ve experienced Indianapolis’s hottest day since 1954, and *really* learned to appreciate good air conditioning. Have we, in the last two weeks, defined what a deep map actually is? In a sense we did, but more importantly than the semantic definition, I reckon we managed to form a set of shared understandings, some fairly intuitive, which articulate (for me at least) how deep mapping differs from other kinds of mapping. It must integrate, and at least some of this integration must involve the linear concepts of what, when and where (but see below). It must reflect experience at the local level as well as data at the macro level, and it must provide a means of scaling between them. It must allow the reader (I hereby renounce the word ‘user’ in relation to deep maps) to navigate the data and derive their own conclusions. Unlike a GIS – ‘so far so Arc’ is a phrase I have co-coined this week – it cannot, and should not attempt to, actualize every possible connection in the data, either implicitly or explicitly. Above all, a deep map must have a topology that enables all these things, and if, in the next six months, the Polis Center can move us towards  a schema underlying that topology, then I think our efforts, and theirs, will have been well rewarded.

The bigger questions for me are what does this really mean for the ‘spatial humanities’; and what the devil are the spatial humanities anyway. They have no Wikipedia entry (so how can they possibly exist?). I have never particularly liked the term ‘spatial turn’, as it implies a setting apart, which I do not think the spatial humanities should be about. The spatial humanities mean nothing if they do not communicate with the rest of the humanities, and beyond. Perhaps – and this is the landscape historian in me talking – it is about the kind of topology that you can extract from objects in the landscape itself. Our group in Week 2 spent a great deal of time thinking about the local and the experiential, and how the latter can be mapped on to the former, in the context of a particular Unitarian ministry in Indianapolis. What are the stories you can get from the landscape, not just tell about it.

Allow me to illustrate the point with war memorials. The city’s primary visitor information site, visitindy.com, states that Indianapolis has more war memorials than any city apart from Washington D.C.. Last Saturday, a crew of us hired a car and visited Columbus IN, an hour and a half’s drive away. In Columbus there is a memorial to most of America’s wars: eight by six Indiana limestone columns, arranged in a close grid formation with free public access from the outside. Engraved on all sides of the columns around the outside, except the outer facing edges, are names of the fallen, their dates, and the war in which they served. On the inner columns– further in, where you have to explore to find them, giving them the mystique of the inner sanctum – are inscribed the full texts of letters written home by fallen servicemen. In most cases, they seem to have been written just days before the dates of death.  The deeply personal natures of these letters provide an emotional connection, and combined with the spatiality of the columns, this connection forms a very specific, and very deliberately told, spatial narrative. It was also a deeply moving experience.

Today, in Indianapolis itself, I was exploring the very lovely canal area, and came across the memorial to the USS Indianapolis. The Indianapolis was a cruiser of the US Navy sunk by Japanese torpedoes in 1945, with heavy loss of life. Particular poignancy is given to the memorial by a narrative of the ship’s history, and the unfolding events leading up to the sinking, inscribed in prose on the monument’s pedestal. I stood there and read it, totally engrossed and as moved by the story as I was by the Columbus memorial.

Image
USS Indianapolis memorial

The point for deep maps: American war memorials tell stories in a very deliberate, designed and methodical way, to deeply powerful effect in the two examples I saw. British war memorials tend not to do this. You get a monument, lists of names of the fallen and the war in question, and perhaps a motto of some sort. An explicit story is not told. This does not make the experience any less moving, but it is based on a shared and implicit communal memory, whose origins are not made explicit in the fabric of the monument. It reflects a subtle difference in how servicemen and women are memorialized, in the formation of the inherently spatial stories that are told in order to remember them.

This is merely one example of subtle differences which run through any built environment of any period in any place, and they become less subtle as you scale in more and more cultures with progressively weaker ties. Britain and America. Europe, Britain and America. Europe, America and Africa, and so on. We scale out and out, and then we get to the point where the approaches to ‘what’ ‘when’ and ‘where’ – the approaches that we worked on in our group – must be recognised not as universal ways of looking at the world, but as products of our British/American/Australian backgrounds, educations and cultural memories. Thus it will be with any deep map.

How do we explain to the shade of Edward Said that by mapping these narratives we are not automatically claiming ownership of them, however much we might want or try not to? How deep will these deep maps need to go…?

Deep maps in Indy

I am here in a very hot and sunny Indianapolis trying to figure out what is meant by deep mapping, with an NEH Summer Institute at UIPUI hosted by the Polis Center here. There follows a very high-level attempt to synthesize some thoughts from the first week.

Deep mapping – we think, although we’ll all probably have changed our minds by next Friday, if not well before  – is about representing (or, as I am increasingly preferring to think, remediating) the things that Ordnance Survey would, quite rightly, run a perfectly projected and triangulated mile from mapping at all. Fuzziness. Experience. Emotion. What it means to move through a landscape at a particular time in a particular way. Or, as Ingold might say, to negotiate a taskscape. Communicating these things meaningfully as stories or arguments. There has been lots of fascinating back and forth about this all week, although – and this is the idea at least – next week we move a beyond the purely abstract and grapple with what it means to actually design one.

If we’re to define the meaning we’re hoping to build here, it’s clear that we need to rethink some pretty basic terms. E.g. we talk instinctively about ‘reading’ maps, but I have always wondered how well that noun and that verb really go together. We assume that ‘deep mapping’ for the humanities – a concept which we assume will be at least partly online – has to stem from GIS, and that a ‘deep map, whatever we might end up calling that, will be some kind of paradigm shift beyond ‘conventional’ computer mapping. But the ’depth’ of a map is surely a function of how much knowledge – knowledge rather than information – is added to the base layer, where that information comes from, and how it is structured. The amazing HGIS projects we’ve seen this week give us the framework we need to think in, but the concept of ‘information’ therein should surely be seen as a starting point. The lack of very basic kinds of such information in popular mapping applications has been highlighted, and perhaps serves to illustrate this point. In 2008, Mary Spence, President of the British Cartographic Society, argued in a lecture:

Corporate cartographers are demolishing thousands of years of history, not to mention Britain’s remarkable geography, at a stroke by not including them on [GPS] maps which millions of us now use every day. We’re in danger of losing what makes maps so unique, giving us a feel for a place even if we’ve never been there.

To put it another way, are ‘thin maps’ really all that ‘thin’, when they are produced and curated properly according to accepted technical and scholarly standards? Maps are objects of emotion, in a way that texts are not (which is not to deny the emotional power of text, it is simply to recognize that it is a different kind of power). Read Mike Parker’s 2009 Map Addict for an affectionate and quirky tour of the emotional power of OS maps (although anyone with archaeological tendencies will have to grit their teeth when he burbles about the mystic power of ley lines and the cosmic significance of the layout of Milton Keynes). According to Spence, a map of somewhere we have never been ties together our own experiences of place, whether absolute (i.e. georeferenced) or abstract, along with our expectations and our needs. If this is true for the lay audiences of, say the Ordnance Survey, isn’t the vision of a deep map articulated this past week some sort of scholarly equivalent? We can use an OS map to make a guess, an inference or an interpretation (much discussion this week has, directly or indirectly, focused on these three things and their role in scholarly approaches). What we cannot do with an OS map is annotate or embed it with any of these. The defining function of a deep map, for me, is an ability to do this, as well as the ability to structure the outputs in a formal way (RDF is looking really quite promising, I think – if you treat different mapped objects in the object-subject-predicate framework, that overcomes a lot of the problems of linearity and scale that we’ve battled with this week). The different levels of ephemerality that this would mean categorising (or, heaven help us, quantifying) is probably a story for another post, but a deep map should be able to convey experience of moving through the landscape being described.

There are other questions which bringing such a map into the unforgiving world of scholarly publication would undoubtedly entail. Must a map be replicable? Must someone else be able to come along and map the same thing in the same way, or at least according to their own subjective experience(s)?  In a live link up the UCLA team behind the Roman Forum project demonstrated their stuff, and argued that the visual is replicable and –relatively easily – publishable, but of course other sensory experiences are not.  We saw, for example, a visualisation of how far an orator’s voice could carry. The visualisation looks wonderful, and the quantitative methodology even more so, but to be meaningful as an instrument in the history of Roman oratory, one would have to consider so many subjective variables – the volume of the orator’s voice (of course), the ambient noise and local weather conditions (especially wind). There are even less knowable functions, such as how well individuals in the crowd could hear, whether they had any hearing impairments etc. This is not to carp –after all, we made (or tried to make) a virtue of addressing and constraining such evidential parameters in the MiPP project, and our outputs certainly looked nothing like as spectacular as UCLA’s virtual Rome – but a deep map must be able to cope with those constraints.

To stand any chance of mapping them, we need to treat such ephemera as objects, and object-orientation seemed to be where our – or at least my – thinking was going last week. And then roll out the RDF…

CAA1 – The Digital Humanities and Archaeology Venn Diagram

The question  ‘what is the digital humanities’ is hardly new; nor is discussion of the various epistemologies of which the digital humanities are made. However, the relationship which archaeology has with the digital humanities – whatever the epistemology of either – has been curiously lacking. Perhaps this is because archaeology has such strong and independent digital traditions, and such a set of well-understood quantitative methods, that the close analysis of of those traditions – familiar to readers of Humanist, say –  seem redundant. However, at the excellent CAA International conference in Southampton last week, there was a dedicated round-table session on the ‘Digital Humanities/Archaeology Venn Diagram’, in which I was a participant. This session highlighted that the situation is far more nuanced and complex that it first seems. As is so often the case with digital humanities.

A Venn Diagram, of course, assumes two or more discrete groups of objects, where some objects contain the attributes of only one group, and others share attributes of multiple groups. So – assuming that one can draw a Venn loop big enough to contain the digital humanities – what objects do they share with archaeology? As I have not been the first to point out, digital humanities is mainly concerned with methods. This, indeed, was the basis of Short and McCarty’s famous diagram. The full title of CAA – Computer Applications and Quantitative Methods in Archaeology – suggests that a methodological focus is one such object shared by both groups. However unlike the digital humanities, archaeology is concerned with a well defined set of questions. Most if not all, of these questions derive from ‘what happened in the past?’. Invariably the answers lie, in turn, in a certain class of material; and indeed we refer to collectively to this class as ‘material culture’.  And digital methods are a means that we use to the end of getting at the knowledge that comes from interpretation of material culture.

The digital humanities have much broader shared heritage which, as well as being methodological, is also primarily textual. This fact is illustrated by the main print publication in the field being called Literary and Linguistic Computing. It is not, I think, insignificant as an indication of how things have moved on that that a much more recently (2007)  founded journal has the less content-specific title Digital Humanities Quarterly. This, I suspect, is related to the reason why digitisation so often falls between the cracks in the priorities of funding agencies: there is a perception that the world of printed text is so vast that trying to add to the corpus incrementally would be like painting the Forth Bridge with a toothbrush (although this doesn’t affect my general view that the biggest enemy of mass digitisation today is not FEC or public spending cuts, but the Mauer im Kopf that form notions of data ownership and IPR). The digital humanities are facing a tension, as they always have, between variable availability of digital material, and the broad access to content that any porting over to the ‘digital’ that the word ‘humanities’ implies. As Stuart Jeffrey’s talk in the session made clear, the questions facing archaeology are more about what data archaeologists throw away: the emergence of Twitter, for example, gives an illusion of ephemerality, but every tweet adds to the increasing cloud of noise on the internet; and those charged with preserving the archaeological record in digital form must decide where where the noise ends and the record begins.

There is also the question of what digital methods *do* to our data. Most scholars who call themselves ‘digital humanists’ would reject the notion that textual analysis, which begins with semantic and/or stylometric mark-up is a purely quantitative exercise; and that qualitative aspects of reading and analysis arise from, and challenge, the additional knowledge which is imparted to a text in the course of encoding by an expert. However, as a baseline, it is exactly the kind of quantitative  reading of primary material which archaeology – going back to the early 1990s – characterized as reductionist and positivist. Outside the shared zone of the Venn diagram, then, must be considered the notions of positivism and reductionism: they present fundamentally different challenges to archaeological material than they do to other kinds of primary resource, certainly including text, but also, I suspect, to other kinds of ‘humanist’ material as well.

A final point which emerged from the session is the disciplinary nature(s) of archaeology and the digital humanities themselves. I would like to pose the question as to why the former is often expressed as a singular noun whereas the latter is a plural. Plurality in ‘the humanities’ is taken implicitly. It conjures up notions of a holistic liberal arts education in the human condition, taking in the fruits of all the arts and sciences in which humankind has excelled over the centuries. But some humanities are surely more digital than others. Some branches of learning, such as corpus linguistics, lend themselves to quantitative analysis of their material. Others tend towards the qualitative, and need to be prefixed by correspondingly different kinds of ‘digital’. Others are still more interpretive, with their practitioners actively resisting ‘number crunching’. Therefore, instead of being satisfied with ‘The Digital Humanities’ as an awkward collective noun, maybe we could look to free ourselves of the restrictions of nomenclature by recognizing that can’t impose homogeneity, and nor should we try to. Maybe we could even extend this logic, and start thinking in terms of ‘digital archaeologies’; of branches of archaeology which require (e.g.) archiving, communication, semantic web, UGC and so on; and some which don’t require any.  I can’t doubt that the richness and variety of the conference last week is the strongest argument possible for this.