Archive for the ‘Events’ Category

Institute of Classical Studies Work-in-Progress seminars (London)

Thursday, May 1st, 2008

Digital Classicist Work-in-Progress seminars
Institute of Classical Studies

Fridays at 16:30 in NG16, Senate House, Malet St, London, WC1E 7HU
(June 20th, July 4th-18th seminars in room B3, Stewart House)
(June 27th seminar room 218, Chadwick Bdg, UCL, Gower Street)


6 June (NG16): Elaine Matthews and Sebastian Rahtz (Oxford), The Lexicon of Greek Personal Names and classical web services

13 June (NG16) Brent Seales (University of Kentucky), EDUCE: Non-invasive scanning for classical materials

20 June (STB3) Dot Porter (University of Kentucky), The Son of Suda On Line: a next generation collaborative editing tool

27 June (UCL Chadwick 218) Bruce Fraser (Cambridge), The value and price of information: reflections on e-publishing in the humanities

4 July (STB3) Andrew Bevan (UCL), Computational Approaches to Human and Animal Movement in the Archaeological Record

11 July (STB3) Frances Foster (KCL), A digital presentation of the text of Servius

18 July (STB9) Ryan Bauman (University of Kentucky), Towards the Digital Squeeze: 3-D imaging of inscriptions and curse tablets

25 July (NG16) Charlotte Tupman (KCL), Markup of the epigraphy and archaeology of Roman Libya

1 Aug (NG16) Juan Garcés (British Library), Digitizing the oldest complete Greek Bible: The Codex Sinaiticus project

8 Aug (NG16) Charlotte Roueché (KCL), From Stone to Byte

15 Aug (NG16) Ioannis Doukas (KCL), Towards a digital publication for the Homeric Catalogue of Ships

22 Aug (NG16) Peter Heslin (Durham), Diogenes: Past development and future plans


We are inviting both students and established researchers involved in the application of the digital humanities to the study of the ancient world to come and introduce their work. The focus of this seminar series is the interdisciplinary and collaborative work that results at the interface of expertise in Classics or Archaeology and Computer Science.

The seminar will be followed by wine and refreshments.

Audio recordings and slideshows will be uploaded after each event.

(Sponsored by the Institute of Classical Studies, University of London, and the Centre for Computing in the Humanities, King’s College London.)

For more information please contact or, or visit the seminar website at

EpiDoc Summer School, July 14th-18th, 2008

Wednesday, April 23rd, 2008
The Centre for Computing in the Humanties, Kings College London, is again offering an EpiDoc Summer School, on July 14th-18th, 2008. The training is designed for epigraphers or papyrologists (or related text editors such as numismatists, sigillographers, etc.) who would like to learn the skills and tools required to mark up ancient documents for publication (online or on paper), and interchange with international academic standards.You can learn more about EpiDoc from the EpiDoc home page and the Introduction for Epigraphers; you wil find a recent and user-friendly article on the subject in the Digital Medievalist. (If you want to go further, you can learn about XML and about the principles of the TEI: Text Encoding Initiative.) The Summer School will not expect any technical expertise, and training in basic XML will be provided.

Attendees (who should be familiar with Greek/Latin and the Leiden Conventions) will need to bring a laptop on which has been installed the Oxygen XML editor (available at a reduced academic price, or for a free 30-day demo).

The EpiDoc Summer School is free to participants; we can try to help you find cheap (student) accommodation in London. If any students participating would like to stay on afterwards and acquire some hands-on experience marking up some texts for the Inscriptions of Roman Cyrenaica project, they would be most welcome!

All interested please contact both and as soon as possible. Please pass on this message to anyone who you think might benefit.

Digitization and the Humanities: an RLG Programs Symposium

Thursday, April 17th, 2008

Is anyone here attending this?

As primary source materials move online, in both licensed and freely available form, what will be the impact on scholarship? On teaching and learning practice? On the collecting practices of research libraries? These are questions we are hoping to explore in the third day of our annual meeting (June 4th). This symposium, which we’re calling “Digitization and the Humanities: Impact on Libraries and Special Collections,” will feature perspectives from scholars on how digital collections are impacting both their research and teaching practice. We’ll also have perspectives from university librarians (Paul Courant, University of Michigan and Robin Adams, Trinity College Dublin) on the potential impact on library collecting practices.

The symposium will be held at the Chemical Heritage Foundation, and on Tuesday evening (June 3rd), the Philadelphia Museum of Art will host a reception for attendees. It should be a great event and a thought provoking conversation, and we hope you will join us. RLG Partners may register online.

Report on NEH Workshop “Supporting Digital Scholarly Editions”

Friday, April 4th, 2008

The official report on the NEH Workshop “Supporting Digital Scholarly Editions”, held on January 14, has been released and is available in PDF form:

Attendees included representatives from funding agencies and university presses, historians, just one or two literary scholars, one medievalist, and no classicists. It appears that much of the discussion focused on creating a service provider for scholarly editions, something to work between scholars and university presses to turn scholarship into digital publications.

I’m of two minds about this. On one hand, I know a lot of “traditional scholars” who find the idea of digital publication a little scary, just the idea of having to learn the technology. So it could be a good way to bring digital publication into the mainstream. But on the other hand, this kind of model could be stifling for creativity. One of the exciting things about digital projects is that, at this time, although there are standards there is no single model to follow for publication. There’s a lot of room for experimentation. It’s certainly not either/or – those of us doing more cutting-edge work will continue to do it whether there are mainstream service providers at university presses or not. But it’s interesting that this is being discussed.

Informatique et Egyptologie, I&E 2008

Wednesday, April 2nd, 2008

A date has been set for the next meeting of the International Association of Egyptologists Computer Group (Informatique et Egyptologie, I&E), which last met in Oxford in 2006.

Thanks to the kindness of Dr Wilfried Seipel, the meeting will take place in the Kunsthistorisches Museum, Vienna, Austria, on 8-11 July 2008, with the sessions on 9-10 July.

Further information can be found here

Problems and outcomes in digital philology (session 3: methodologies)

Thursday, March 27th, 2008

The Marriage of Mercury and Philology: Problems and outcomes in digital philology

e-Science Institute, Edinburgh, March 25-27 2008.

(Event website; programme wiki; original call)

I was asked to summarize the third session of papers in the round table discussion this afternoon. My notes (which I hope do not misrepresent anybody’s presentation too brutally) are transcribed below.

Session 3: Methodologies

1. Federico Meschini (De Montfort University) ‘Mercury ain’t what he used to be, but was he ever? Or, do electronic scholarly editions have a mercurial attitude?’ (Tuesday, 1400)

Meschini gave a very useful summary of the issues facing editors or designers of digital critical editions. The issues he raised included:

  • the need for good metadata standards to address the problems of (inevitable and to some extent desirable) incompatibility between different digital editions;
  • the need for a modularized approach that can include many very specialist tools (the “lego bricks” model);
  • the desirability of planning a flexible structure in advance so that the model can grow organically, along with the recognition that no markup language is complete, so all models need to be extensible.

After a brief discussion of the reference models available to the digital library world, he explained that digital critical editions are different from digital libraries, and therefore need different models. A digital edition is not merely a delivery of information, it is an environment with which a “reader” or “user” interacts. We need, therefore, to engage with the question: what are the functional requirements for text editions?

A final summary of some exciting recent movements, technologies, and discussions in online editions served as a useful reminder that far from taking for granted that we know what a digital critical edition should look like, we need to think very carefully about the issues Mechini raises and other discussions of this question.

2. Edward Vanhoutte (Royal Academy of Dutch Language and Literature, Belgium) ‘Electronic editions of two cultures –with apologies to C.P. Snow’ (Tuesday, 1500)

Vanhoutte began with the rhetorical observation that our approach to textual editions is in adequate because the editions are not as intuitive to users, flexible in what they can contain, and extensible in use and function as a household amenity such as the refrigerator. If the edition is an act of communication, an object that mediates between a text and an audience, then it fails if we do not address the “problem of two audiences” (citing Lavagnino). We serve the audience of our peers fairly well–although we should be aware that even this is a more hetereogenous and varied a group than we sometimes recognise–but the “common audience”, the readership who are not text editors themselves, are poorly served by current practice.

After some comments on different types of editions (a maximal edition containing all possible information would be too rich and complex for any one reader, so minimal editions of different kinds can be abstracted from this master, for example), and a summary of Robinson’s “fluid, cooperative, and distributed editions”, Vanhoutte made his own recommendation. We need, in summary, to teach our audience, preferably by example, how to use our editions and tools; how to replicate our work, the textual scholarship and the processes performed on it; how to interact with our editions; and how to contribute to them.

Lively discussion after this paper revolved around the question of what it means to educate your audience: writing a “how to” manual is not the best way to encourage engagement with ones work, but providing multiple interfaces, entry-points, and cross-references that illustrate the richness of the content might be more accessible.

3. Peter Robinson (ITSEE, Birmingham) ‘What we have been doing wrong in making digital editions, and how we could do better?’ (Tuesday, 1630)

Robinson began his provocative and speculative paper by considering a few projects that typify things we do and do not do well: we do not always distribute project output successfully; we do not always achieve the right level of scholarly research value. Most importantly, it is still near-impossible for a good critical scholar to create an online critical edition without technical support, funding for the costs of digitization, and a dedicated centre for the maintenance of a website. All of this means that grant funding is still needed for all digital critical work.

Robinson has a series of recommendations that, he hopes, will help to empower the individual scholar to work without the collaboration of a humanities computing centre to act as advisor, creator, librarian, and publisher:

  1. Make available high-quality images of all our manuscripts (this may need to be funded by a combination of goverment money, grant funding, and individual users paying for access to the results).
  2. Funding bodies should require the base data for all projects they fund to be released under a Creative Commons Attribution-ShareAlike license.
  3. Libraries and not specialist centres should hold the data of published projects.
  4. Commercial projects should be involved in the production of digital editions, bringing their experience of marketing and money-making to help make projects sustainable and self-funding.
  5. Most importantly, he proposes the adoption of common infrastructure, a set of agreed descriptors and protocols for labelling, pointing to, and sharing digital texts. An existing protocol such as the Canonical Text Services might do the job nicely.

4. Manfred Thaller (Cologne) ‘Is it more blessed to give than to receive? On the relationship between Digital Philology, Information Technology and Computer Science’ (Wednesday, 0950)

Thaller gave the last paper, on the morning of the third day of this event, in which he asked (and answered) the over-arching question: Do computer science professionals already provide everything that we need? And underlying this: Do humanists still need to engage with computer science at all? He pointed out two classes of answer to this question:

  • The intellectual response: there are things that we as humanists need and that computer science is not providing. Therefore we need to engage with the specialists to help develop these tools for ourselves.
  • The political response: maybe we are getting what we need already, but we will experience profitable side effects from collaborating with computer scientists, so we should do it anyway.

Thaller demonstrated via several examples that we do not in fact get everything we need from computer scientists. He pointed out that two big questions were identified in his own work twelve years ago: the need for software for dynamic editions, and the need for mass digitization. Since 1996 mass digitization has come a long way in Germany, and many projects are now underway to image millions of pages of manuscripts and incunabula in that country. Dynamic editions, on the other hand, while there has been some valuable work on tools and publications, seem very little closer than they were twelve years ago.

Most importantly, we as humanists need to recognize that any collaboration with computer scientists is a reciprocal arrangement, that we offer skills as well as receive services. One of the most difficult challenges facing computer scientists today, we hear, is to engage with, organise, and add semantic value to the mass of imprecise, ambiguous, incomplete, unstructured, and out-of-control data that is the Web. Humanists have spent the last two hundred years studying imprecise, ambiguous, incomplete, unstructured, and out-of-control materials. If we do not lend our experience and expertise to help the computer scientists solve this problem, than we can not expect free help from them to solve our problems.

DHI Now Known as Office of Digital Humanities (ODH)

Tuesday, March 25th, 2008

Not specifically classics, but this news from the National Endowment for the Humanities should be of interest, at least to those of us in the US: The Digital Humanities Initiative (DHI) has been made permanent, and is now the Office of Digital Humanities (ODH)
From the ODH Webpage:

The Office of Digital Humanities (ODH) is an office within the National Endowment for the Humanities (NEH). Our primary mission is to help coordinate the NEH’s efforts in the area of digital scholarship. As in the sciences, digital technology has changed the way scholars perform their work. It allows new questions to be raised and has radically changed the ways in which materials can be searched, mined, displayed, taught, and analyzed. Technology has also had an enormous impact on how scholarly materials are preserved and accessed, which brings with it many challenging issues related to sustainability, copyright, and authenticity. The ODH works not only with NEH staff and members of the scholarly community, but also facilitates conversations with other funding bodies both in the United States and abroad so that we can work towards meeting these challenges.

Digital Classicist seminars update

Tuesday, March 25th, 2008

To bring you all up to date with what is going on with the Digital Classicist seminar series:

Some papers from the DC seminar series held at the Institute of Classical Studies in London in the summer of 2006 have been published as a special issue of the Digital Medievalist (4:2008).


The dedication reads: In honour of Ross Scaife (1960-2008), without whose fine example of collaborative spirit, scrupulous scholarship, and warm friendship none of the work in this volume would be what it is.

Gabriel and I are putting together a collection of papers from the DC summer series of 2007 and working on the programme for the coming summer (2008). With the continued support of the Institute of Classical Studies (London) and the Centre for Computing in the Humanities, King’s College London it is anticipated that this seminar series will continue to be an annual event.  

Services and Infrastructure for a Million Books (round table)

Monday, March 17th, 2008

Million Books Workshop, Friday, March 14, 2008, Imperial College London.

The second of two round tables in the afternoon of the Million Books Workshop, chaired by Brian Fuchs (Imperial College London), asked a panel of experts what services and infrastructure they would like to see in order to make a Million Book corpus useful.

  1. Stuart Dunn (Arts and Humanities e-Science Support Centre): the kinds of questions that will be asked of the Million Books mean that the structure of this collection needs to be more sophisticated that just a library catalogue
  2. Alistair Dunning (Archaeological Data Service & JISC): powerful services are urgently needed to enable humanists both to find and to use the resources in this new collection
  3. Michael Popham (OULS but formerly director of e-Science Centre): large scale digitization is a way to break down the accidental constraints of time and place that limit access to resources in traditional libraries
  4. David Shotton (Image Bioinformatics Research Group): emphasis is on accessibility and the semantic web. It is clear than manual building of ontologies does not scale to millions of items, therefore data mining and topic modelling are required, possible assisted by crowdsourcing. It is essential to be able to integrate heterogeneous sources in a single, semantic infrastructure
    1. Dunning: citability and replicability of research becomes a concern with open publication on this scale
    2. Dunn: the archaeology world has similar concerns, cf. the recent LEAP project
  5. Paul Walk (UK Office for Library and Information Networking): concerned with what happens to the all-important role of domain expertise in this world of repurposable services: where is the librarian?
    1. Charlotte Roueché (KCL): learned societies need to play a role in assuring quality and trust in open publications
    2. Dunning: institutional repositories also need to play a role in long-term archiving. Licensing is an essential component of preservation—open licenses are required for maximum distribution of archival copies
    3. Thomas Breuel (DFKI): versioning tools and infrastructure for decentralised repositories exist (e.g. Mercurial)
    4. Fuchs: we also need mechanisms for finding, searching, identifying, and enabling data in these massive collections
    5. Walk: we need to be able to inform scholars when new data in their field of interest appears via feeds of some kind

(Disclaimer: this is only one blogger’s partial summary. The workshop organisers will publish an official report on this event.)

What would you do with a million books? (round table)

Sunday, March 16th, 2008

Million Books Workshop, Friday, March 14, 2008, Imperial College London.

In the afternoon, the first of two round table discussions concerned the uses to which massive text digitisation could be put by the curators of various collections.

The panellists were:

  • Dirk Obbink, Oxyrhynchus Papyri project, Oxford
  • Peter Robinson, Institute for Textual Scholarship and Electronic Editing, Birmingham
  • Michael Popham, Oxford University Library Services
  • Charlotte Roueché, EpiDoc and Prosopography of the Byzantine World, King’s College London
  • Keith May, English Heritage

Chaired by Gregory Crane (Perseus Digital Library), who kicked off by asking the question:

If you had all of the texts relevant to your field—scanned as page images and OCRed, but nothing more—what would you want to do with them?

  1. Roueché: analyse the texts in order to compile references toward a history of citation (and therefore a history of education) in later Greek and Latin sources.
  2. Obbink: generate a queriable corpus
  3. Robinson: compare editions and manuscripts for errors, variants, etc.
    1. Crane: machine annotation might achieve results not possible with human annotation (especially at this scale), particularly if learning from a human-edited example
    2. Obbink: identification of text from lost manuscripts and witnesses toward generation of stemmata. Important question: do we also need to preserve apparatus criticus?
  4. May: perform detailed place and time investigations into a site preparatory to performing any new excavations
    1. Crane: data mining and topic modelling could lead to the machine-generation of an automatically annotated gazeteer, prosopography, dictionary, etc.
  5. Popham: metadata on digital texts scanned by Google not always accurate or complete; not to academic standards: the scanning project is for accessibility, not preservation
    1. Roueché: Are we talking about purely academic exploitation, or our duty as public servants to make our research accessible to the wider public?
    2. May: this is where topic analysis can make texts more accessible to the non-specialist audience
    3. Brian Fuchs (ICL): insurance and price comparison sites, Amazon, etc., have sophisticated algorithms for targeting web materials at particular audiences
    4. Obbink: we will also therefore need translations of all of these texts if we are reaching out to non-specialists; will machine translation be able to help with this?
    5. Roueché: and not just translations into English, we need to make these resources available to the whole world.

(Disclaimer: this summary is partial and partisan, reflecting those elements of the discussion that seemed most interesting and relevant to this blogger. The workshop organisers will publish an official report on this event presently.)

Million Books Workshop (brief report)

Saturday, March 15th, 2008

Imperial College London.
Friday, March 14, 2008.

David Smith gave the first paper of the morning on “From Text to Information: Machine Translation”. The discussion included a survey of machine translation techniques (including the automatic discovery of existing translations by language comparison), and some of the value of cross-language searching.

[Please would somebody who did not miss the beginning of the session provide a more complete summary of Smith's paper?]

Thomas Breuel then spoke on “From Image to Text: OCR and Mass Digitisation” (this would have been the first paper in the day, kicking off the developing thread from image to text to information to meaning, but transport problems caused the sequence of presentations to be altered). Breuel discussed the status of professional OCR packages, which are usually not very trainable and have their accuracy constrained by speed requirements, and explained how the Google-sponsored but Open Source OCRopus package intends to improve on this situation. OCRopus is highly extensible and trainable, but currently geared to the needs of the Google Print project (and so while effective at scanning book pages, may be less so for more generic documents). Currently in alpha-release and incorporating the Tesseract OCR engine, this tool currently has a lower error-rate than other Open Source OCR tools (but not the professional tools, which often contain ad hoc code to deal with special cases). A beta release is set for April 2008, which will demo English, German, and Russian language versions, and release 1.0 is scheduled for Fall 2008. Breuel also briefly discussed the hOCR microformat for describing page layouts in a combination of HTML and CSS3.

David Bamman gave the second in the “From Text to Information” sequence of papers, in which he discussed building a dynamic lexicon using automated syntax recognition, identifying the grammatical contexts of words in a digital text. With a training set of some thousands of words of Greek and Latin tree-banked by hand, auto-syntactic parsing currently achieves an accuracy rate something above 50%. While this is still too high a rate of error to make this automated process useful as an end in itself, to deliver syntactic tagging to language students, for example, it is good for testing against a human-edited lexicon, which provides a degree of control. Usage statistics and comparisons of related words and meanings give a good sense of the likely sense of a word or form in a given context.

David Mimno completed the thread with a presentation on “From Information to Meaning: Machine Learning and Classification Techniques”. He discussed automated classification based on typical and statistical features (usually binary indicators: is this email spam or not? Is this play tragedy or comedy?). Sequences of objects allow for a different kind of processing (for example spell-checking), including named entity recognition. Names need to be identified not only by their form but by their context, and machines do a surprisingly good job at identifying coreference and thus disambiguating between homonyms. A more flexible form of automatic classification is provided by topic modelling, which allows mixed classifications and does not require the definition of labels. Topic modelling is the automatic grouping of topics, keywords, components, relationships by the frequency of clusters of words and references. This modelling mechanism is an effective means for organising a library collection by automated topic clusters, for example, rather than by a one-dimensional and rather arbitrary classmark system. Generating multiple connections between publications might be a more effective and more useful way to organise a citation index for Classical Studies than the outdated project that is l’Année Philologique.

Simon Overell gave a short presentation on his doctoral research into the distribution of location references within different language versions of Wikipedia. Using the tagged location links as disambiguators, and using the language cross-reference tags to compare across the collections, he uses the statistics compiled to analyse bias (in a supposedly Neutral Point-Of-View publication) and provide support for placename disambiguation. Overell’s work is in progress, and he is actively seeking collaborators who might have projects that could use his data.

In the afternoon there were two round-table discussions on the subjects of “Collections” and “Systems and Infrastructure” that I may report on later if my notes turn out to be usable.

Changing the Center of Gravity

Tuesday, March 4th, 2008

Changing the Center of Gravity: Transforming Classical Studies Through Cyberinfrastructure

University of Kentucky, 5 October 2007

This is the full audio record of “Changing the Center of Gravity: Transforming Classical Studies Through Cyberinfrastructure”, a workshop funded by the National Science Foundation, sponsored by the Center for Visualization and Virtual Environments at the University of Kentucky, and organized by the Perseus Digital Library at Tufts University.

1) Introduction (05:13)
- Gregory Crane
(download this presentation as an mp3 file – 4.78 MB)

2) Technology, Collaboration, & Undergraduate Research (26:23)
- Christopher Blackwell and Thomas Martin, respondent Kenny Morrell
(download this presentation as an mp3 file – 24.1 MB)

3) Digital Criticism: Editorial Standards for the Homer Multitext (29:02)
- Casey Dué and Mary Ebbott, respondent Anne Mahoney
(download this presentation as an mp3 file – 26.5 MB)

4) Digital Geography and Classics (20:23)
- Tom Elliot, respondent Bruce Robertson
(download this presentation as an mp3 file – 18.6 MB)

5) Computational Linguistics and Classical Lexicography (39:16)
- David Bamman and Gregory Crane, respondent David Smith
(download this presentation as an mp3 file – 35.9 MB)

6) Citation in Classical Studies (38:34)
- Neel Smith, respondent Hugh Cayless
(download this presentation as an mp3 file – 35.3 MB)

7) Exploring Historical RDF with Heml (24:10)
- Bruce Robertson, respondent Tom Elliot
(download this presentation as an mp3 file – 22.1 MB)

8) Approaches to Large Scale Digitization of Early Printed Books (24:38)
- Jeffrey Rydberg-Cox, respondent Gregory Crane
(download this presentation as an mp3 file – 22.5 MB)

9) Tachypaedia Byzantina: The Suda On Line as Collaborative Encyclopedia (20:45)
- Anne Mahoney, respondent Christopher Blackwell
(download this presentation as an mp3 file – 18.9 MB)

10) Epigraphy in 2017 (19:00)
- Hugh Cayless, Charlotte Roueché, Tom Elliot, and Gabriel Bodard, respondent Bruce Robertson
(download this presentation as an mp3 file – 17.3 MB)

11) Directions for the Future (50:04)
- Ross Scaife et al.
(download this presentation as an mp3 file – 45.8 MB)

12) Summary (01:34)
- Gregory Crane
(download this presentation as an mp3 file – 1.44 MB)

CFP: DRHA 2008: New Communities of Knowledge and Practice

Monday, March 3rd, 2008

By way of a long string of reposts, originally to AHESSC:

Date: Fri, 29 Feb 2008 17:37:17 -0000
From: Stuart Dunn


Forthcoming Conference

DRHA 2008: New Communities of Knowledge and Practice

The DRHA (Digital Resources in the Humanities and Arts) conference is held annually at various academic venues throughout the UK. The conference theme this year is to promote discussion around new collaborative environments, collective knowledge and redefining disciplinary boundaries. The conference, hosted by Cambridge with its fantastic choice of conference venues will take place from Sunday 14th September to Wednesday 17th September.

The aim of the conference is to:

  • Establish a site for mutually creative exchanges of knowledge.
  • Promote discussion around new collaborative environments and collective knowledge.
  • Encourage and celebrate the connections and tensions within the liminal spaces that exist between the Arts and Humanities.
  • Redefine disciplinary boundaries.
  • Create a forum for debate around notions of the ‘solitary’ and the collaborative across the Arts and Humanities.
  • Explore the impact of the Arts and Humanities on ICT: design and narrative structures and visa versa.

There will be a variety of sessions concerned with the above but also with a particular emphasis on interdisciplinary collaboration and theorising around practice. There will also be various installations and performances focussing on the same theme. Keynote talks will be given by our plenary speakers who we are pleased to announce are Sher Doruff, Research Fellow (Art, Research and Theory Lectoraat) and Mentor at the Amsterdam School for the Arts, Alan Liu, Professor of English, University of California Santa Barbara and Sally Jane Norman, Director of the Culture Lab, Newcastle University. In addition to this, there will be various round table discussions together with a panel relating to ‘Second Life’ and a special forum ‘Engaging research and performance through pervasive and locative arts projects’ led by Steve Benford, Professor of Collaborative Computing, University of Nottingham. Also planned is the opportunity for a more immediate and informal presentation of work in our ‘Quickfire’ style events. Whether papers, performance or other, all proposals should reflect the critical engagement at the heart of DRHA.

Visit the website for more information and a link to the proposals website.

The Deadline for submissions will be 30 April 2008 and abstracts should be approximately 1000 words.

Cambridge’s venues range from the traditional to the contemporary all situated within walking distance of central departments, museums and galleries. The conference will be based around Cambridge University’s Sedgwick Site, particularly the West Road concert hall, where delegates will have use of a wide range of facilities including a recital room and a ‘black box’ performance space, to cater for this year’s parallel programming and performances.

Sue Broadhurst DRHA Programme Chair

Dr Sue Broadhurst
Reader in Drama and Technology, Head of Drama, School of Arts
Brunel University
West London, UB8 3PH
Direct Line:+44(0)1895 266588 Extension: 66588
Fax: +44(0)1895 269768

Rieger, Preservation in the Age of Large-Scale Digitization

Sunday, March 2nd, 2008

CLIR (the Council on Library and Information Resources in DC) have published in PDF the text of a white paper by Oya Rieger titled ‘Preservation in the Age of Large-Scale Digitization‘. She discusses large-scale digitization initiatives such as Google Books, Microsoft Live, and the Open Content Alliance. This is more of a diplomatic/administrative than a technical discussion, with questions of funding, strategy, and policy rearing higher than issues of technology, standards, or protocols, the tension between depth and scale (all of which were questions raised during our Open Source Critical Editions conversations).

The paper ends with thirteen major recommendations, all of which are important and deserve close reading, and the most important of which is the need for collaboration, sharing of resources, and generally working closely with other institutions and projects involved in digitization, archiving, and preservation.

One comment hit especially close to home:

The recent announcement that the Arts and Humanities Research Council and Joint Information Systems Committee (JISC) will cease funding the Arts and Humanities Data Service (AHDS) gives cause for concern about the long-term viability of even government-funded archiving services. Such uncertainties strengthen the case for libraries taking responsibility for preservation—both from archival and access perspectives.

It is actually a difficult question to decide who should be responsible for long-term archiving of digital resources, but I would argue that this is one place where duplication of labour is not a bad thing. The more copies of our cultural artefacts that exist, in different formats, contexts, and versions, the more likely we are to retain some of our civilisation after the next cataclysm. This is not to say that coordination and collaboration are not desiderata, but that we should expect, plan for, and even strive for redundancy on all fronts.

(Thanks to Dan O’Donnell for the link.)

Registration: 3D Scanning Conference at UCL

Tuesday, February 26th, 2008

Kalliopi Vacharopoulou wrote, via the DigitalClassicist list:

I would like to draw to your attention the fact that registration for the 3D Colour Laser Scanning Conference at UCL on the 27th and 28th of March has now opened.

The first day (27th of March) will include a keynote presentation and papers on the themes of General Applications of 3D Scanning in the Museum and Heritage Sector and of 3D Scanning in Conservation.

The second day (28th of March) will offer a keynote presentation and papers on the themes of 3D Scanning in Display (and Exhibition) and Education and Interpretation. A detailed programme with the papers and the names of the speakers can be found in our website.

If you would like to attend the conference, I would kindly request to fill in the registration form which you can find in this link and return it to me as soon as possible.

There is no fee for participating (or attending the conference) (coffee and lunch are provided free of charge). Please note that attendance is offered on a first-come, first-served basis.

Please feel free to circulate the information about the conference to anyone who you think might be interested.

In the meantime, do not hesitate to contact me with any inquiries.

International Seminar of Digital Philology: Edinburgh, March 25-27, 2008

Saturday, February 23rd, 2008

Seen on the AHeSSC mailing list:

The e-Science Institute Event Announcement

The e-Science Institute is delighted to host the “The Marriage of Mercury and Philology: Problems and Outcomes in Digital Philology”. The conference welcomes both leading scholars and young researchers working on the problems of textual criticism and editorial scholarship in the electronic medium, as well as students, teachers, librarians, archivists, and computing professionals who are interested in representation, access, exchange, management and conservation of texts.

Organiser: Cinzia Pusceddu
Dates and Time: Tuesday 25th March 09.00 – Thursday 27th March 17.00
Place: e-Science Institute
University of Edinburgh
13-15 South College Street

For registration and more details see


Digitizing Early Material Culture (CFP)

Thursday, February 14th, 2008

Posted for Brent Nelson:

Digitizing Early Material Culture: from Antiquity to Modernity
A Seminar to be held in conjunction with
CaSTA (the Canadian Symposium on Text Analysis) 2008:
New Directions in Text Analysis
A Joint Humanities Computing, Computer Science Seminar and Conference at University of Saskatchewan, Saskatoon, 16-18 October 2008
“Digitizing Early Material Culture: from Antiquity to Modernity” seminar will be held at the University of Saskatchewan in Saskatoon 16 October 2008 and will feature guest speakers:

  • Melissa Terras, Lecturer in Electronic Communications in the School of Library, Archive and Information Studies at University College London
  • Lisa Snyder, Associate Director of the Experiential Technologies Centre, University of California Los Angeles

It will be held in conjunction with CaSTA 2008–“New Directions in Text Analysis,” 17-18 August, featuring guest speakers:

  • David Hoover, Professor of English at New York University (keynote)
  • Hoyt Duggan, Professor Emeritus in English at University of Virginia
  • Geoffrey Rockwell, Associate Professor in Humanities Computing at University of Alberta
  • Cara Leitch, PhD candidate in English at University of Victoria


CFP: Open Scholarship: Authority, Community and Sustainability in the Age of Web 2.0

Wednesday, January 23rd, 2008

By way of JISC-Repositories:

The 12th International Conference on Electronic Publishing (25 to 27 June 2008, Toronto, Canada) has just extended its call for papers to 31 January 2008. Full details below …


CFP: Virtual Worlds: Libraries, Education and Museums

Monday, January 21st, 2008

Gabriel Bodard just posted a call for papers for a “virtual worlds” conference, to be held in Second Life on 8 March 2008. You can read the full CFP in the Digital Classicist Archive. I find it unfortunate that the conference organizers (Bodard is not one) have chosen to organize and publicize the conference via a facebook group that requires interested parties to log in just to read about the event.

Humanities GRID Workshop (30-31 Jan; Imperial College London)

Monday, January 21st, 2008

By way of the Digital Classicists List:

Epistemic Networks and GRID + Web 2.0 for Arts and Humanities
30-31 January 2008
Imperial College Internet Centre, Imperial College London

Data driven Science has emerged as a new model which enables researchers to move from experimental, theoretical and computational distributed networks to a new paradigm for scientific discovery based on large scale GRID networks (NSF/JISC Digital Repositories Workshop, AZ 2007). Hundreds of thousands of new digital objects are placed in digital repositories and on the web everyday, supporting and enabling research processes not only in science, but in medicine, education, culture and government.  It is therefore important to build interoperable infra-structures and web-services that will allow for the exploration, data-mining, semantic integration and experimentation of arts and humanities resources on a large scale.  There is a growing consensus that GRID solutions alone are too heavy, and that coupling it with Web 2.0 allows for the development of a more light-weight service oriented architecture (SOA) that can adapt readily to user needs by using on demand utility computing, such as morphological tools, mash-ups, surf clouds, annotation and automated workflows for composing multiple services.  The goal is not just to have fast access to digital resources in the arts and humanities, but to have the capacity to create new digital resources, interrogate data and form hypotheses about its meaning and wider context.  Clearly what needs to emerge is a mixed-model of GRID + Web 2.0 solutions for the arts and humanities which creates an epistemic network that supports a four step iterative process: (i) retrieval, (ii) contextualisation, (iii) narrative and hypothesis building, and (iv) creating contextualised digital resources in semantically integrated knowledge networks.  What is key here is not just managing new data, but the capacity to share, order, and create knowledge networks from existing resources in a semantically accessible form.

To create epistemic networks in the arts and humanities there are core technologies that must be developed.  The aim of this expert METHNET Workshop is to focus on developing a strategy for the implementation of these core technologies on an inter-national scale by bringing together GRID computing specialists with researchers from Classics, Literature and History who have been involved in the creation and use of electronic resources.  The core technologies we will focus on in this two day work-shop are: (i) infrastructure, (ii) named entity, identity and co-reference services, (iii) morphological services and parallel texts, (iv) epistemic networks and virtual research environments.  The idea is to bring together expertise from the UK, US, and European funded projects to agree upon a common strategy for the development of core infra-structure and web-services for the arts and humanities that will enable the use of GRID technologies for advanced research.

DAY ONE- 10:00 – 6:00

SESSION I: GRID + Web 2.0 Infrastructure

SESSION II: Computational and Semantic Services: Named Entity, Identity and Co-reference

  • Paul Watry: Named Entity and Identity Services for the National Archives
  • Greg Crane –  Co-Reference (Perseus)
  • Hamish Cunningham/Kalina Bontcheva: AKT and GATE: GRID-WEB Services AKT/GATE
  • Martin Doerr – Co-Reference and Semantic Services for Grid + Web 2.0 (FORTH)

DAY TWO: 10:00 – 6:00

SESSION I:  Morphological, Parallel Texts and Citation Services

  • Greg Crane – “Latin Depedency Treebank”, Perseus Project
  • Marco Passarotti – “Index Thomisticus” Treebank
  • Notis Toufexis – ‘Neither Ancient, nor Modern:  Challenges for the creation of a Digital Infrastructure for Medieval Greek’
  • Rob Iliffe – Intelligent Tools for Humanities Researchers, The Newton Project

SESSION II: Epistemic Networks and Virtual Research Environments

Registration fee is £60 and places are limited.

Please contact Dolores Iorizzo ( to secure a place or for further information.  Please send registration to Glynn Cunin (

The Imperial College Internet Centre would like to acknowledge generous support from the AHRC METHNET for co-hosting this conference.

CFP: International aerial archaeology conference

Monday, January 21st, 2008

Another interesting call for papers via Jack Sasson’s Agade list:


International aerial archaeology conference (AARG 2008)
Ljubljana, 9 – 11 September 2008

Hosted by the Department of Archaeology, Faculty of Arts, University of Ljubljana

Proposals for sessions, papers and posters are invited

The following sessions have been proposed, for which offers of papers are welcome:

  • Aerial Archaeology in the Mediterranean; New Projects; Postgraduate research;
  • Airborne Thematic Mapping/Airborne Laser Scanning;
  • An archaeology of natural places … from the air;
  • Aerial photography in context – recording landscape and urban areas

11 September Conference Day 3

Field Trip

Note: session titles are provisional and all papers and session proposals are welcome.

Oral papers should usually be 20 minutes duration, and equal weighting is given to poster presentations.

Closing date for abstracts is 31st May 2008.

Address for conference correspondence:

Dave Cowley
16 Bernard Terrace
Edinburgh, EH8 9NX

International School in Archaeology and Cultural Heritage

Friday, January 18th, 2008

By way of Jack Sasson’s Agade list:

We would like to bring your attention to the International School in Archaeology and Cultural Heritage that we’re organizing in May 2008, in Ascona, Switzerland.

It’s a jointly organization between:

The School will face the problem of the modern technologies in the heritage field, giving participants the opportunity to obtain a detailed overview of the main methods and applications to archaeological and conservation research and practice. Furthermore, our School will give the chance to participants to enter in a very short time the kernel of the scientific discussion on 3D technologies — surveying methods, documentation, data management and data interpretation — in the archaeological research and practice.

The School will be open to ca 60 participants at graduate level, to those carrying out doctoral or specialist research, to established research workers, to members of State Archaeology Services and to professionals specializing in the study and documentation, modeling and conservation of the archaeological heritage.

The deadline for the registration is 31st March, 2008.

Grants provided by UNESCO and ISPRS will be available for students with limited budgets and travel possibilities. The deadline for the grant application is 15st February, 2008.

The grant application and registration form are available online [pdf].

The School is to be held in the congress centre Centro Stefano Franscini, Monte Verità, Ascona, Switzerland. The centre is an ETH-affiliated seminar complex located in a superb botanical park on the historic and cultural Monte Verità area, which will also be the residence of the participants with its integrated hotel and restaurant.

We would be grateful if you could also circulate this announcement to all the possible participants.

Don’t hesitate to contact by email the organization if you should have any question.

Thank you and best regards,

Prof. Armin Gruen

Dr. Stefano Campana

Dr. Fabio Remondino

Prof. Maurizio Forte

THATCamp: May 31 – June 1, 2008

Thursday, January 17th, 2008

See further

a BarCamp-style, user-generated “unconference” on digital humanities … organized and hosted by the Center for History and New Media at George Mason University, Digital Campus, and THATPodcast

NEH/IMLS Advancing Knowledge Grant Program Announced

Thursday, January 17th, 2008

From Brett Bobley:

This is a reminder that the National Endowment for the Humanities (NEH) and the Institute of Museum and Library Services (IMLS) invite applications to a digital humanities grant competition sponsored by the two federal agencies. The grant program, “Advancing Knowledge: The IMLS/NEH Digital Partnership,” seeks applications for projects that would explore new ways to share, examine, and interpret humanities collections in a digital environment and develop new uses and audiences for existing digital resources.   Grants are intended to spur innovation and new collaborations; advance the role of cultural repositories in online teaching, learning, and research; and develop collaborative approaches involving the scholarly community and cultural repositories for the creation, management, preservation, and presentation of reusable digital collections and products.

Projects must be collaborative with at least one museum, library, or archive as an integral member of the project team. Awards normally are for two years and typically range from $50,000 to a maximum of $350,000. Nonprofit institutions interested in applying can find guidelines onlineThe deadline for applications is March 18, 2008.  Applicants are encouraged to contact program officers who can offer advice about preparing the proposal and review draft proposals. Draft proposals should be submitted six weeks before the deadline. Questions and drafts may be submitted by e-mail to

NEH/JISC joint event at King’s College, London

Thursday, January 17th, 2008

Monday, 21 January 2008 in room 2B08, Strand Campus, King’s College London: