I decided to call my inaugural lecture "A Decade in Digital Humanities" for three reasons.
1. The term Digital Humanities has been commonly used to describe the application of computational methods in the arts and humanities for 10 years, since the publication, in 2004, of the Companion to Digital Humanities. "Digital Humanities" was quickly picked up by the academic community as a catch-all, big tent name for a range of activities in computing, the arts, and culture. A decade on from the publication of this text, I thought it would be useful to reflect on the growth, spread, and changes that had occurred in our discipline, and my place within them.
2. This year sees me in my 10th year of being in an academic post. I joined UCL in August 2003, my first academic post after obtaining my doctorate, and since then have worked my way up the ranks from probationary lecturer, to senior lecturer, to reader, and now full professor. The professorial lecture gives me a rare chance to pause and look behind me to see what the body of work built up over this time represents, and what it means to be undertaking research in this area.
3. You'll have to wait for later in the lecture to see the third reason...
Who here would be comfortable defining what is meant by the term Digital Humanities? In this, the week of UCL Festival of the Arts, celebrating all things to do with the Arts and Humanities, let's go back to first principles. In UCLDH and 4Humanities' award winning infographic "The Humanities Matter" we defined the humanities as "academic disciplines that seek to understand and interpret the human experience, from individuals to entire cultures, engaging in the discovery, preservation, and communication of the past and present record to enable a deeper understanding of contemporary society." It stands to reason, then, that the Digital Humanities are computational methods that are trying to understand what it means to be human, in both our past and present society. But it may be easier if I give some brief examples to demonstrate the kind of work we Digital Humanists get up to.
One of the easiest things we can do with computers is count things. For data to be computationally manipulated, it has to be in numeric form. If we can get text into a computational form, we can easily count and manipulate the language, showing trends across time. For example, if we take a million words of conference abstracts from my discipline from the ALLC/ACH conference across various years, we can easily see how mentions of one technology (XML) becomes more popular, while another (SGML) is in decline. Much of the work in DH is in manipulating and processing and analysing text - our iOS app Textal is just part of that trajectory. Much of my work, though, has been in digital images, starting with developing systems to try and read damaged documents from Hadrian's wall, and more recently working on multispectral and 3D manipulation of damaged texts. We've also worked with museums on large scale 3D capture of cultural and heritage objects. The important thing about all of this is that as well as implementation, we're also interested in use and usage of these technologies, and what impact that they have on those working in culture and heritage, and the ability to study the past and present human record. We often innovate new systems, or adopt concepts and apply them to humanities projects, such as the crowdsourcing of Jeremy Bentham's handwriting by volunteers, or working with visitors to the Grant Museum of Zoology at UCL to encourage debate about zoological collections. We build, we test, we reflect back on what using these technologies means for the humanities, giving recommendations which can be useful across the sector. From these projects, its difficult to pin down what Digital Humanities actually is, but that sums up the difficulty of our discpline's title: it encourages thinking about computational methods in the arts and humanities, and then into culture and heritage, in as broad a sense as possible.
What made Digital Humanities spring, fully formed like Athena from the Head of Zeus, as an academic field in 2004? Was it because that was the first time quantifiable methods had been used in the Arts and Humanities? (remember - all computational methods require quantification). Well, of course that is nonsense. When you look back across the history of Humanities scholarship, quantifiable methods were used in the Arts and Humanities since the birth of Universities. If we think of the book as technology, from its inception scholars took it to pieces to see under the hood: concordances and indexes of works were manually created, such as this "Concordance or table made after the order of the alphabet" from 1579 which lists how many times concepts such as "abomination" appear in the New Testament. Or the work of Joseph Scaliger who in the early 1600s plotted the different periods in time in which different civilizations must have existed, through quantifiable methods. Or the work of August Schleicher in the 1850s who showed, by quantifiable methods, that the languages of Europe must have had a common historical root. All of these texts are available from UCL Library, none of which I have to leave my sofa to see because YAY! Digitisation! Changing humanities scholarship! - but the point is that quantifiable methods are part of established methods in the humanities, and have been for as long as the Humanities have existed. So when I undertook my first project at UCL, looking at whether we could use the high performance computing facilities at UCL to analyse historical census data - this is part of an quantifiable humanities academic tradition which harks back 500 years, just at a grander scale.
So what made Digital Humanities spring, fully formed like Athena from the Head of Zeus, as an academic field in 2004? Perhaps in 2004, this was the first time people had used computational techniques in the arts and humanities? But of course, that is nonsense too. When you look back at the history of computing - and not even digital computing, but the very first computer - the very first computer programmer, Ada Lovelace, hints at the possibilities for art, music, and understanding human knowledge and culture in her earliest writings. She understood that there was something more to the mathematical calculations afforded by this machine than science, and they called her a madwoman for it. Well, this madwoman has a (yet unproven) theory that if you look at the history of the first 100 electronic programmable computers in the 1950s, 1960s and 1970s across the world, you will see humanists eyeing them up and asking "how can I use, or develop this tool for use, in my research"? Its certainly true of Father Busa, working with IBM in the 1950s on the concordance of the works of Thomas Aquinas (counting, indexing, and manipulating words, as part of the historical trajectory of humanities methods stretching back 500 years, just a change in scale...) but also of Roy Wisbey, in Cambridge, who set up the Literary and Linguistic Computing Centre there in the 1960s. When the first computers arrived at UCL, the artists from the Slade School of Fine Art were over there like a shot to establish the Experimental and Computing Department. We should also mention Susan Hockey, who led various initiatives in text encoding, text analysis, and digital libraries. Susan, incidentally, gave me my first academic job here at UCL in 2003: UCL had included a Digital Resources in the Humanities module course as part of its MA offering for librarians and archivists in the School of Library, Archive and Information Studies (now the Department of Information Studies) from 2000, under Susan's auspices. But the point is, considering how best to use computing in the arts and humanities is not something which started in the 21st Century, nor 2004, and Humanists have been looking at available tools, and how best to use them, since computation began. So when we undertook one of the latest projects at UCLDH, which came from looking at an iPhone, thinking "how can I use, or develop this tool for use, in my research in the Humanities" and developed an iOS app for text analysis, this is part of a longer trajectory of considering available computational tools, and how they may be appropriated, adopted, and adapted for our means in the humanities, just at a grander scale, as processing technologies increase in speed.
So why Digital Humanities, in 2004? Firstly, the coalescing of interested scholars into an identifiable field is an understandable academic response to societal changes. The speed of computing rises, the price of computing plummets, the information available on the internet (and the possibility to create new information) increases, use and usage of internet technologies has become commonplace. Remember, its up to Humanities scholars to look at the past and present record to enable a deeper understanding of contemporary society: quite frankly, it would be more alarming if an academic movement hadn't emerged looking at what using computational methods could do for our understanding of human society, both past and present, and how best we can grab the technical opportunities which fly by and appropriate them for our means, to inform both ourselves and others about the prospects of using computing in this area. The discipline of Digital Humanities is inevitable, and would have appeared whatever the title it was given.
Secondly, Digital Humanities is a handy, all inclusive, modern title which rebrands all the various work which has gone before it, such as Humanities Computing, Computing and the Humanities, Cultural Heritage Informatics, Humanities Advanced Technology... DH has a ring to is, and boy, what a rebranding it was. We tend to call it "Big Tent Digital Humanities" meaning: roll up! roll up! everyone using any computational method in any aspects of the arts and humanities is welcome! but really, Big Wave Digital Humanities may be more appropriate, as we countenance the sudden swell, dissipation, and speed of the activities of the discipline. Taking a peek at the mention of Digital Humanities on Google Ngrams we can see its sudden growth, and the fact that it is now used as a proper noun, with Capital Letters (although remember that this, counting words, is part of a long tradition of humanities scholarship, Google simply have more books to include in their count). We can see how DH has trended over time, appearing in headlines in the media. Many, many textbooks in DH appear, some of which I am responsible for myself. Journals appear, such as Digital Humanities Quarterly (of which I'm one of the general editors), and the ALLC/ACH conference renames itself Digital Humanities (this year, for my sins, I'm the Program Chair for DH2014 which will be held in Lausanne, Switzerland. We have seen over 700 proposals from more than 2000 vying for a space to present). There are many more DH conference presentations and workshop slots, worldwide, year on year. In 2010, I gathered together all the available evidence I could on DH in an infographic called Quantifying Digital Humanities, showing that there were 114 DH Centres in 24 countries. Today, not even four full years later, there are 195 DH Centers in 27 Countries. Those knowing how long it takes to set up a research centre know that this is phenomenal growth in the university and GLAM sector, and that institutional support must be strong, behind each and everyone of these.
UCL Centre for Digital Humanities is part of those who have joined the recently founded centres. We officially launched four years ago to the week of this lecture, in the same lecture hall where this lecture is being presented. We dont talk about the launch much - its not often I'm part of something at work which ends up featured in the political pages of the newspapers - but you'll have to google that to find out more (YAY! digital media! the internet never forgets!) but in those four years since launch we've undertaken a phenomenal amount of projects, covering many aspects of Humanities and Arts research, and considered Digital Humanities in its broadest sense. This isnt all me - there is an amazing team who are part of the Centre, and we've won various awards for our academic projects and collaborations, published many books, papers, and book chapters, and been part of successful funding bids from research councils worth tens of millions of pounds. One wonders what makes a Digital Humanities Centre attractive to universities that dont have one. Nope, I cant see what makes that level of activity attractive, at all.
So what proportion of Humanities scholars are now digital humanists? Back in 2005, participants in the Summit on Digital Tools in the Humanities at the University of Virginia estimated that "only about six percent of humanist scholars go beyond general purpose information technology and use digital resources and more complex digital tools in their scholarship" (p.4 of this PDF). By 2012, N. Katherine Hayles, in her chapter "How we think: transforming power and digital technologies" in David M. Berry's edited text "Understanding Digital Humanities", estimates that 10 per cent of Humanists are now digital humanists (p.59). Now, in 2014, a forthcoming study from Ithaka S+R (with the working title of Sustaining the Digital Humanities: Institutional Strategies beyond the Start-up Phase) includes surveys of faculty at four American universities. In the departments surveyed at each institution, nearly 50% of faculty members indicated they have "created or managed" digital resources. Granted, the departments were chosen by campus staff (often at the library) who felt there was some significant activity taking place there. The percentage of these "creators" was consistent across all universities (Brown, Columbia, University of Wisconsin, Indiana University), and most of the creators also felt that their creation was intended for public use (not just their own research aims), and would require ongoing development in the future.
50% of humanists are involved in digital activity, are digital humanists. How can this possibly be? And how can we conceptualise what it means to be a digital humanist, amongst this spread of activity and range of available technology: is creating or managing digital resources the same as being a digital humanist? At a time where (nearly) every library catalogue is digitised and available online, and (nearly) every book manuscript written on a work processor, and many historical documents digitised and available for consulting from your own sofa, does that make everybody working in the humanities a digital humanist? How can I begin to conceptualise my contribution, and my place, and where my work sits within Big Wave Digital Humanities?
I find it useful, here to turn to Roger's Innovation Adoption Curve, a sociological model that looks at how technology spreads through society. This is a bell curve, and right at the start of adoption of technology, are a few innovators, experimenting (and developing) new technology. These innovators sometimes persuade a larger number of early adopters to take up the new technology on offer, and only once a sufficient mass of users are achieved, does the technology "cross the chasm" and become used by the majority of individuals in a society (who are split into an early majority, or late majority). Finally, we have adoption by the "laggards", who are slow in taking up technologies, but do so if they have permeated throughout society. (Hard not to think, here, of my elderly grandmother who recently got her first mobile phone). Now, this model is useful as we can plot along it some of the technologies which are available to a humanist. Things like word processing, and searching for references online, and even looking up the digitised texts which I showed at the start of this lecture: even the technologically laggard humanists can do it now, and although these technologies are changing scholarship, its a question of scale (better! faster! more!) rather than of approach or technique, for the main. Technically facilitated tasks like updating websites, using and updating wikis, using social media: even the late majority of humanists can do it now. Online tools are available, such as Voyant, which allow you to do text analysis, and manipulate texts to see the underlying patterns: so the early majority of humanists can use these tools should they want to. But the most difficult, intellectual work of applying technology in the humanities still occurs before the chasm has been crossed, in the phase of innovation, and early adoption, where we are looking at the technologies that cross our path and saying "how can I use, or develop this tool for use, in my research?", much like those in the 1950s or 1960s who were coming across university mainframes and asking how best to apply that in the literary and linguistic arena. It's important to note, of course, that this wave of technology keeps on coming at us, and the place of where technology sits along the curve changes: 20 years ago, had you been making a website for your humanities project, you would have been an innovator, rather than a late majority, and the same holds for word processing 40 years ago. The technology keeps coming: we have to respond to this, innovate, adopt, and see what is useful or useable for, or used by, the majority of people in our discpline.
Now (and this is the most contentious thing I'm going to say in my whole lecture, for those attending who are dyed-in-the-wool Digital Humanists) one of the problems that we have as a movement is that we tend to get caught up and fixated upon a certain technological solution. For example, every DH program I've come across teaches XML, that technology which took over from SGML in the conference abstracts - as the best practice way to encode text. And there's no doubt that XML provides the framework with which we can both explore theoretically what is means to describe texts computationally, in such a way they retain the information in their printed or manuscript form, whilst also the means to build and test prototypes. But XML as a technological standard has been around for 16 years, and technology moves on, but DH doesnt seem to be doing so. In many ways, DH's relationship to XML is similar to the AI community's relationship with LISP: the means of computational expression in the language or format suit the questions which need to be asked by the field, so there is no need to use other technologies which come on stream, which may be more efficient from a computational point of view, as we explore what is means to work with our question in this computational way. And that's ok, but we shouldnt be blind to the fact that, hey! technology is advancing all the time and, also, XML is not a technology that crossed the chasm: it may be in use for technical systems, but its not one that you see a lot of the general populace using. This, in turn, means that DH has permanently hitched its wagon to an aging technology, which is hard to explain to others, including other non-XML humanists, whilst other things are happening in the technological world around us. Just something we have to watch out for, when building teaching programs, or looking at the scope of outputs in our field. We dont want to be left behind as the digital in digital humanities rolls on without us.
I find it useful to plot my research on the Innovation Curve, to see where what I am doing sits. So, the work on counting terms across a corpus - very much sits in the early majority, given the availability of tools to do so. But the work on building an iPhone app to do so - very much innovation: it took a lot of pure programming in a relatively new space to achieve it. The work in image processing I do is either innovation (we are publishing here in pure computer/engineering science venues, as well as in humanities venues, which I'm very proud of), or we adopt technologies our academic colleagues in the engineering sciences have generated and roll them out to a humanities or heritage application. Our work on user studies is something completely different though: here we are generally looking at how the majority of people are using an extant text, or (in the case of something like Transcribe Bentham, or QRator) we are conducting reception studies, where we innovate and build a technology, launch it, and study its uptake across the whole cycle. We can see, then a range of DH activity across the innovation cycle, but the majority of the work I do is certainly at the start of the innovation curve. Is this where DH sits? I like to think so, but more to the point, I'm confident its where I sit best, when doing DH.
I need here to show you another curve, though. This time, the Gartner Hype Cycle, which looks at how technologies are launched, mature, and are applied (so people know when to invest). The premise of this is that when technologies are first triggered, everyone thinks they are going to be the Next Big Thing, and so they reach "the peak of inflated expectations", before crashing down into a "trough of disillusionment" when those adopting them realise they aren't that great at all. Its hard work to get technologies up the "slope of enlightenment" where useful, useable applications are found, and few technologies make it to the "plateau of productivity" where they become profitable. Its a useful curve - this year's predictions show Big Data right at the top of the peak, which chimes in with media coverage of how it will solve everything, for example. So where would I put DH, if I had to as a movement, on this curve?
First, I would argue that we have to understand our lineage: that the current manifestation of DH is a logical progression of qualitative methods used in the humanities for the past 500 years. That the current manifestation of DH is a logical progression of humans wondering what the potential is for applying computational methods to humanities problems, which has been going on in the digital space for the past 60 years. These combined trajectories aren't going away, and despite what funding cuts and media backlash may come at us, it is the role of the digital humanist to understand and investigate how computers can be used to question what it means to be human, and the human record, in both our past and present society. Secure in our mission, we can carry on whatever the storm throws at us.
Second, I would argue we have to ignore naysayers who are unsure about this new Digital Humanities lark (and believe me, there are plenty, even in my own department) and just do good work. The way to demonstrate our worth is to demonstrate our worth through doing good work. We have to keep asking questions about computational methods, computational processes, and the potentials that they offer humanities scholars, as well as the pitfalls, to explore this changing information environment from the humanities viewpoint. Its not just about building websites, or putting information online, its about innovating and adopting, and questioning while we build about the ramifications of doing this, the impact on the humanities, the issues using technology raises, and the answers it provides that you couldn't otherwise generate, to do good work in Digital Humanities. I realise this is very Calvinist of me - you can take the lass out of Scotland - but I do see that we have to be engaging with theories and questions of what is means to be doing this work in this way, as well as updating a website or creating a digital file. A continuation of what it means to be a humanities scholar, in the digital space.
I'm not one for looking back, and despite the title, I deliberately didn't want this inaugural to be a survey of all the projects I have undertaken over the past ten years - then I did this, then I talked to that person, then I visited there - but when I look back over the variety and range of projects, publications, and outputs that I've worked on, either on my own, or as part of a team (there's a lot of teamwork that has gone on here) I'm firstly surprised at how much of it there is and the range of topics we've covered, and the opportunities we've pounced on. I see a body of work which explores various aspects of what it means to be applying digital technologies in the humanities space, and facilitates both those in engineering science and those in the humanities to explore issues which are important to them. I've learn't things along the way about the nature of interdisciplinary work, the nature of teams, the nature of the academic publishing and peer review process, the nature of the grant funding process, but I've written about that elsewhere. There are things, also, that I am proud of that are physical rather than purely digital: over the last few years I'm most proud of building the UCL Multi-Modal digitisation suite, which is a shared space between the UCL Library Services, UCL Faculty of Arts and Humanities, and UCL Faculty of Engineering Science, contributing to the infrastructure of UCL in a collaborative endeavor. But what I see here, as a common thread, is that the work I do tends to sit right at the beginning of the technology adoption cycle, aiding and abetting the application of technology within the arts, humanities, and heritage, and I'm comfortable with that. There's a strength in knowing your place, and your remit, and what you do best.
So the third reason for calling my talk "A Decade in Digital Humanities" is that I didn't say which decade we were talking about, and it is time also to look towards the future, and what the next ten years holds for both DH, as the field turns into a teenager, and for me, as I go into my next decade here at UCL. I'm not one for crystal balls, so I'll keep my scrying brief. I see an inevitable fragmentation of the DH community and DH focus - it was never conceived of as a homogenous entity anyway, and it is the nature of waves and swells that they will dissipate. We'll see (we are already seeing) more focussed groups of scholarly work around, say, Geographical Information Systems and literature, as people specialise and work on specific technologies and specific methods. The technology will keep coming, and its up to individual humanities scholars to respond to what is appropriate to their research question: the effects of DH scholarship will continue to ripple out across the humanities as technologies go along the adoption cycle, and certain aspects of digital research will just become normal for humanities scholars, as time goes on. But I do see that there will always be a place, right at the start of the technology innovation uptake curve, for specialists in Digital Humanities to sit, watching out for these changing and emerging technologies, setting up pilot projects to experiment with different aspects of these technologies, feeding back recommendations and the potential ramifications for other humanities and engineering scholars and those within the wider cultural and heritage sector, and exploring what is means to be doing humanities research in that area. I'm happy to remain there, and I see that this will remain my place working with other humanists, and engineers and computer scientists, over the next decade. I'm delighted to be a co-investigator on the doctoral training centre for Science and Engineering in the Arts Heritage and Archaeology, which is the EPSRC's largest every investment in Heritage Science, and for the next 8 years we'll be training up a range of doctoral students in this cross section of the arts, heritage, humanities, and engineering and conservation science. (Perhaps what I really do is Heritage Science, but that's another talk entirely, and DH has work to do with the Heritage Science community in future). That said, we do have work to do, in keeping an eye to making sure people know about the successes, outputs, and impacts of DH work. Given the expectations foisted upon us, we have to learn to be more vocal about our objectives, our remit, and our results. It's our job to be thinking what it means to use digital technologies in humanities research, and just research, full stop. As a result, our insights can benefit a range of other fields, if we communicate them effectively.
Digital technologies are not going away any time soon: and although DH has had a rapid swell, it will remain essential that we investigate, use, and experiment with technologies over the coming decade. There is a new Companion to Digital Humanities coming out in late 2014, showing how the technologies used in humanities research have developed since the first edition (I'm delighted to have written a chapter on our public engagement work for it), and our see our field, as well as knowing where we have come from, has to understand that the technological wave on which we sail is continually on the move. I hope I've shown here that our uptake of technologies in the humanities is, and will continue to be, a moving target, and that as part of a longer trajectory of investigation into humanities methods, DH is a modern but necessary, and even inevitable, part of the Humanities, and even computational, landscape. I look forward to what adventures the next Decade in Digital Humanities holds. There is so much to do!
Now, that is where I'd normally pause and say thank you for your attention, but hey, its my inaugural, so I'll cry if I want to. I have a few brief thanks to make - its quite a lick to go from probationary lecturer to full prof in ten years, and so I have to thank those who have supported me. Thanks go to my family up in Scotland for all their support, and my family of my own: many of you know that in the past few year's I've had three children, so biggest thanks of all go to my husband Os, aka Expert Sleepers, for his forbearance and baby juggling skillz. I've been blessed with an amazing support network of friends, who have supported my enormously over this period. My first academic supervisor was Professor Seamus Ross, who kick started my interest in this area, and his support and interest at the start of my career really set me up for the work I do today. Likewise, my PhD supervisor Professor Alan Bowman remains a fantastic mentor: thank you, Alan. My other PhD supervisor, Professor Sir Mike Brady, made me promise (when I got my doctorate in engineering) not to go near any nuclear power stations or bridges, a promise I have kept - thanks Mike. I've already mentioned that Professor Susan Hockey gave me my first academic job: but her work remains an inspiration on what is possible in computing in the arts and humanities. I work with an amazing team of people at UCLDH and I thank them for their input both for the centre and on our various projects. Special thanks go to Rudolf Ammann, our designer at large, who helped prepare the graphics for this lecture.
But in this week of UCL's Festival of the Arts and Humanities, its good to pause and see how embedded Digital Humanities research is now throughout college, and how much we work, in the Humanities, with those around us. The projects I've shown, albeit briefly, today, are carried out in league with various other faculties (UCLDH reports to both the Arts and Humanities and Engineering Faculties here). Colleagues come from a range of different departments including not only those across the Arts Faculty, but the Bartlett Centre for Advanced Spatial Analysis (in the UCL Bartlett Faculty of the Built Environment), and across the UCL Faculty of Engineering (I have joint projects with Medical Physics, Computer Science, and Civil, Environmental, and Geomatic Engineering). We are dependent on input from both our colleagues in UCL Library Services, and UCL Museums and Collections, and work very closely with items in all the collections across college. The success of DH at UCL is then dependent on the institutional context we have here. Digital Humanities is now embedded into college life at UCL, and in this week of the Festival of the Arts, my final thanks go to UCL as an community for its institutional support in encouraging us to ride the DH wave: for without being at UCL, my decade in digital humanities would have been completely different.