Abstract
Since the early 1990s, theorizing in the digital humanities has often celebrated
open-endedness and incompletion as inherent qualities of digital work. But a
scholarly publisher undertaking preparation and sale of digital objects cannot
altogether dispense with traditional notions of deadlines and completion if those
publications are to enter the dual marketplaces of peer review and institutional
purchase. The Electronic Imprint of the University of Virginia Press was funded in
2001 with the goal of bringing born-digital scholarly projects under the aegis of the
same review and marketing system that applies to books. In this article I describe
how we defined the criteria for “done-ness” in creating two very different
projects, a born-digital edition of Herman Melville’s Typee manuscript and a conversion of the letterpress Papers of George Washington into a digital edition. Our experience
suggests that it is possible to categorize different genres of digital creations
based on the extent to which intrinsic criteria for “done-ness” can be applied to
them, and that decisions about completeness are always subject to extrinsic factors
as well, such as budgetary constraints and the pressures created by competition and
the evolution of standards.
Some History of Terms
Viewed from the perspective of someone who works for a university press, the
semantics of the term “done” as applied to digital objects is rather curious.
From our point of view, it's generally a good thing for a scholarly publication to be
“done”: review copies can be sent out, books can be shipped to distributors,
and budgets perhaps even met. Traditional publication in the scholarly publishing
world has always meant the implicit guarantee that a work is the end product of a
rigorous process of peer review, revision, copyediting, design, and proofreading
shared institutionally by author, press boards, outside scholars, and in-house staff.
When a book or journal issue is “done” it is a source of pride and satisfaction
for everyone concerned.
The case seems to be different with digital objects. The claim that a digital project
or publication is “done” may be met with suspicion. What do you mean, your
Web-thing is finished? Since it's nonlinear, how do you know where it starts or ends?
Won't there always be more features or links you can add? If your Web-thing is so
much like an old-fashioned codex book that you can call it “done”, does it
really belong online in the first place? This suspicion has a history. Theoretical
discussion of projects in the digital humanities has, since the 1990s, suffered from
semantic slippage between two related but nonidentical pairs of contradictory terms:
on the one hand, “open” versus “closed”; on the other hand, “complete”
versus “incomplete” (or “unfinished” versus “done”, etc.). The
tendency has been to merge these two sets into a single pair, then to valorize the
first pair of terms and to demonize the second.
One of the more polished articles on Wikipedia these days, ironically, is on the
topic “Unfinished Work”; it discusses incomplete works in
various domains from literature and music through architecture to software. On the
article's discussion page, the first thing we find is some amused perplexity about
the label's applicability to the very source it appears in [
'Unfinished Work' 2007]:
It is a familiar conundrum about the nature of digital texts. Obviously, a formally
defined text like a sonnet can be recognized as complete or incomplete; it's the
difference between a well-wrought urn and a pot whose clay is still wet. But can a
nonlinear, extensible, text ever be said to be finished? Is it by definition
unfinished, or is the opposition “finished/unfinished” just plain inapplicable
to open-ended texts?
These are theoretical questions I'm not in a position to answer, but I would submit
that early in the 1990s the postmodern admiration of the “open-ended” at the
expense of the “closed” somehow got turned into a celebration of the
“unfinished” and a suspicion of the “done,” and that this transmutation
may have been one of the things that delayed the entrance of digital scholarship into
the traditional system of peer-reviewed academic publication.
Consider these assertions from George Landow and Paul Delany's 1991 essay “Hypertext, Hypermedia and Literary Studies: The State of the
Art”:
Particularly inapplicable [to hypertext]
are the notions of textual “completion” and of a “finished” product.
Hypertext materials are by definition open-ended, expandable, and incomplete.
If one put a work conventionally considered complete, such as the Encyclopedia Britannica, into a hypertext format, it
would immediately become “incomplete.”
[Landow and Delany 1991, 13]
A clever reader might object that even in print the
Encyclopedia Britannica is always incomplete: like any reference work, it
is constantly being updated and reissued. So when Landow revises this particular
passage for his 1992 book
Hypertext, he makes the claim even more
radical by making a single change to the second sentence to replace the encyclopedia
with a work of literature:
Hypertextual materials, which by
definition are open-ended, expandable, and incomplete call such notions into
question. If one put a work conventionally considered complete, such as Ulysses, into a hypertext format, it would immediately
become “incomplete.”
[Landow 1992, 59]
Landow is now claiming that even a recognizably closed, well-wrought modernist
text becomes both open and unfinished when put online. And he ends his 1992
discussion of completion by citing Derrida to the effect that “a form of textuality that goes beyond print ‘forces us to
extend...the dominant notion of a “text”
’,” so that it “is henceforth no longer a finished
corpus of writing, some content enclosed in a book or its margins but a
differential network, a fabric of traces referring endlessly to something other
than itself, to other differential traces”
[
Landow 1992, 59].
Julia Flanders has observed in a memorable phrase that the digital humanities have
sometimes suffered from “a culture of the perpetual
prototype”
[
Flanders 2007], and identified some plausible economic and institutional causes. To them I
think we can add the theoretical conflation of the digital with
différance. After all, what was the postmodern project if not a
cult of the perpetual prototype?
Rotunda: A Scholarly Digital Imprint
My organization, the Electronic Imprint of the University of Virginia Press, was
established in 2001 to test the proposition that instances of digital scholarship
can be bounded, completed, and presented for review, sale, and
academic consumption in much the way journals and monographs had been for decades. We
were grant funded, with support from the University and the Mellon Foundation awarded
to a proposal co-written by the Press and John Unsworth, who was then head of the
Institute for Advanced
Technology (IATH) at Virginia. We became fully staffed in late 2002, and two
years later released our first publication, a born-digital edition of Dolley
Madison's correspondence, under our new imprint name of “Rotunda.” Since then we
have expanded to a total of seven publications in two separate collections:
nineteenth-century literature and culture, and the American Founding Era. Our main
focus for the next few years will be creating fully-featured digital versions of the
papers of American presidents and other Founding Era figures that began as
multivolume (and often still ongoing) print editions, joining our
Papers of George Washington Digital Edition (
PGWDE), which was released in February 2007.
The underlying data format of all of our Rotunda publications is XML, tagged
according to the guidelines of the Text Encoding Initiative (TEI), plus accompanying
digitized graphical material. Unlike many other university presses with digital
projects, we outsource none of our technical work except for graphic Web design; our
markup specifications, stylesheets for file transformation, and programming for Web
delivery (mostly coded in XQuery using
MarkLogic Server as the back-end platform) are all done in-house by several
programmers and technical editors.
Born-digital Rotunda publications go through the same steps that our books go
through: approval by a Press committee and then the Press Board after reports from
external reviewers; signing of a contract complete with royalty agreements; sharing
of “review copies” (in the form of password access) with librarians and academic
reviewers. Digital editions such as PGWDE that are based
on existing print series are produced in close collaboration with the scholarly
communities (historians and documentary editors) who create and use the letterpress
volumes. Clearly all parties to our process of publication and sale are implicitly
agreeing to bracket the theoretical issue of when or whether a digital work is ever
“done” by applying a socioeconomic definition: it is “done” when the
Press is prepared to offer it for purchase and customers are prepared to buy it.
I turn now to two very dissimilar examples of our publications — PGWDE and an edition of Herman Melville's Typee manuscript — and discuss the decisions we made about what we could
or couldn't include in the finished work; when we counted each as “done” for
initial release; and to what extent we consider the published release genuinely
complete or part of a work still in progress.
Melville's Typee: A Fluid-Text Edition
John Bryant's edition of a portion of Herman Melville's novel Typee was first envisioned and prototyped years before Rotunda came into
existence. Bryant has been editing Melville's texts for two decades, and has long
felt that any critical edition of a text that survives in more than a single version
needs to be faithful to its evolutionary history; it should be what he calls a
“fluid-text” edition. Because a fluid-text edition needs to capture a dynamic
process, a computer-based format is a natural fit, and he began imagining one for
Melville as early as the 1980s (personal communication).
The textual history of
Typee is fairly complicated. The
only surviving manuscript fragment covers about three chapters of the published
novel. It contains a multitude of cancellations, erasures, and additions by Melville,
both in ink from the time of first composition and in pencil from later stages of
revision and proofreading
Figure 3.
We know that Melville made changes in proof before the first English
edition was issued, and that the first American edition contained still more changes,
some requested by Melville, others made by the publisher probably without the
author's assent. Bryant's goal for a digital fluid-text edition was to capture all of
these stages and to allow the reader to follow the sequence of composition and the
editor's narrative reconstruction of that sequence, zooming in and out to any point
in manuscript time and space during the entire period from initial composition
through the published editions.
Development of the Edition
Bryant was not himself a programmer or XML specialist, but he did have ideas about
what a computer edition might look like, and created detailed storyboards before
any actual programming work began. Although these were necessarily static, they
used frame- and button-like boxes to suggest how a screen presentation might
respond dynamically to reader choices (
Figure 4). In
1998 he was named an IATH
Visting Fellow
and received technical assistance to create a first proof-of-concept prototype of
the edition [
Bryant 2000], which translated the storyboards into
standard HTML frames (
Figure 5).
In late 2003 we received Bryant's “manuscript” of the edition, consisting of
Microsoft Word and PageMaker files containing manuscript transcriptions flagged
with hundreds of “revision sites” and for each separate revision site a
“revision sequence” and a “revision narrative.” We licensed from the
New York Public Library the rights to reproduce their full-color photographs of
the entire manuscript. Our goals at this point were (1) to convert all
transcription and commentary to TEI-XML, and (2) to design an environment that
could deliver combinations of text and image to realize as closely as possible the
author's intentions for his edition. Our own finished rendering of the original
concept [
Bryant 2006] would look like this:
Once we had our basic page display working, all that remained was to code a search
page and add the editorial introductions before declaring the edition “done”
and releasing it in March 2006.
Three months later we added an enhancement, our major one to date, an XML-based
version of the entire first British edition of the novel, which the University of
Virginia Library digitized for us from a copy in their holdings. We created for it
a display interface combining a transcription of the text with images (
Figure 9).
Is Our Typee Done?
Neither Bryant nor the Press conceived of the Typee
edition as an open-ended project. The editor's work was done once he had finished
all the manuscript transcription, identification of revision sites, exposition of
revision sequences and narratives, and the introductory editorial essay. Our work
was done once we had translated the editor's vision into a fully functional
edition that coordinated photographic facsimiles with several transcription
formats, and that hyperlinked all “revision sites” with their editorial
expansions. The March 2006 edition was lacking one intended feature, the first
British edition, owing to extrinsic factors (our library's digitization schedule).
Once that was added, Typee was for practical purposes
stable and complete.
Nevertheless, we were aware of the potential for improvements and enhancements,
some more immediately practicable than others:
- We could generate RDF metadata files in the format used by the Collex tool
created by Jerry McGann and his NINES
team. In July 2007 we did this, so that the base view of each manuscript page
exists as an indexed object in Collex, along with the editorial introduction
and the publication home page.
- The full-text search needs improvements to return hits on supplied text and
to properly handle word tokens containing XML tags (for example:
savage<add>ry</add>). The first item is on our
to-do list; the second is on hold until our MarkLogic software adds the ability
to ignore selected elements for the purpose of word tokenization.[1]
More radically still, it is conceivable that all of our underlying XML markup and
presentation might be entirely revised if John Bryant were to incorporate the
proposals for temporal encoding in genetic digital editions that Elena Pierazzo
has advanced [
Pierazzo 2007], as both her tagging strategy and
theoretical approach vary significantly from his own. A revision of that magnitude
would be analogous to issuing a second edition of a book that differs markedly
from the original because it has responded to new evidence and/or arguments. All
scholarly and scientific publications are potentially imperfect and thus
“incomplete” to the extent that later work can call them into question,
but it would be an equivocation to say that they are therefore always unfinished
in a formal sense.
The Papers of George Washington Digital Edition
The
Papers of George Washington Digital Edition
[
Crackel 2007] is a very different project, one initiated in 2004 by
the Press in collaboration with the editorial staff of the
Papers of George Washington (also
based at the University of Virginia), and partially funded by a grant from Mount
Vernon. Our mission was to produce an online version of the fifty-two volumes then in
print of our letterpress
Papers of George Washington
[
Jackson 1976], the authoritative scholarly edition of the documentary
legacy of the first president. Owing to the size and complexity of the letterpress
edition,
[2] its adaptation to a fully-featured online format offered us as many design
and programming challenges as a born-digital project like
Typee. We needed to establish an appropriate XML schema and encoding
specifications, decide on what structural and semantic tagging to do and what
metadata to add, figure out how much regularization of inconsistencies in the
letterpress edition we could accomplish, and design a Web environment for display,
navigation, and searching of the edition usable by advanced scholars and beginning
students alike.
For the editorial staffs of The Papers of George Washington and UVa Press, the
criteria for regarding a letterpress volume as complete have been well established
since the project began in the 1970s:
- all known documents from the period covered by the volume are included or
referenced
- all document transcriptions are complete and have been checked for accuracy
against manuscript facsimiles
- all possible identifications of persons have been made and included in the
endnotes
- all other annotation and editorial introductions are written
- the manuscript has been copyedited
- page proofs have been checked and used to produce a back-of-the-book
index.
But for us, including the full content of the print edition would not be
sufficient. We could consider
PGWDE
“done” only when we had reliably translated textual and scholarly conventions
into an online format that offered as much (to use the inevitable marketing phrase)
added value as possible beyond simply being able to access the publication without
visiting a library.
Goals for the PGWDE
Determining where we could add value to the print edition required a preliminary
analysis of what makes a scholarly edition valuable in the first place. In such an
edition, the basic textual unit is the single document, always accompanied at a
minimum by bibliographical information and usually by editorial annotation, and
sometimes by translations, enclosed documents, or other ancillary material.
(Diaries and journals are a special case: depending on how chronologically
structured they are, the basic textual unit may be the single-day entry, the
single-month entry, or a longer narrative.) Besides the original text and
editorial material, documents contain metadata, cross-references, abbreviations
and other special features that are represented using a variety of editorial and
typographical conventions, as highlighted in the facsimile of the original
letterpress version of a letter from William Livingston to George Washington
(shown
below). Beyond the document level, most
volumes contain scholarly apparatuses (lists of abbreviations and bibliographic
expansions of short-title references), editorial and historical introductions, and
a detailed index of all proper names and hundreds of topic categories. The
translation of all of these print conventions into their TEI-XML equivalents is
what must undergird a digital edition. (Our final XML encoding of the Livingston
letter may be seen in the
appendix.)
Our initial goals for the digital edition were:
- to provide document-by-document display (or, for diaries, month-by-month or
day-by-day, as appropriate) closely resembling that of the letterpress
source;
- to offer a wide variety of means for navigating into the documents: through
full-text search; through a hyperlinked consolidated index based on the
back-of-the-book print indexes; via tables of contents similar to those in the
print edition; and by chronology (in order to collect all documents and diary
entries for a given date, for example);
- to use as much tagged information as possible for display, linking, and
refinement of searching;
- to create a genuinely new edition incorporating corrections to the print
edition submitted by the Papers of George Washington staff, along with
consolidated and regularized lists of names and titles that had varied from
volume to volume in the letterpress edition.
Work on
PGWDE began in fall of 2004; a beta version
for public display was ready by October 2006; and we formally released a published
version for sale in February 2007. Screen captures of the online version of the
Livingston letter illustrate how we realized some of our goals (
Figure 12,
Figure 13).
Compasses are used to navigate the four hierarchies identifed in goal 2. A
“breadcrumb trail” allows quick navigation up to any higher node of the
current tree. Hyperlinks or mouseovers provide dynamic equivalents of their print
counterparts in ways that are familiar to Web users: endnote superscripts connect
to their notes via bidirectional linking; abbreviations and short-title references
(indicated by dotted underlining) are expanded when the user mouses over the
abbreviated text; and cross-references to other documents in
PGWDE are active links.
Along with the document navigation and display, we programmed a search page that
combines full-text search with optional filtering based on author, recipient,
and/or date range; and we added an online version of the consolidated index that
resembles a back-of-the-book index except that document titles and dates replace
page numbers and are, of course, hyperlinked.
We had scheduled official release of
PGWDE for
President's Day — February 19, 2007. By a month or so ahead of this deadline, we
realized that every last cleanup task could not be completed by that date. Online
publication meant we could do a triage: fix first the things that affect the most
documents, or that are most obvious to the average reader; fix afterwards problems
or errors limited to single documents, or ones that would be noticed only by a
specialist (for example, an incorrect birth year for a minor historical figure).
Corrections of bad links and minor formatting glitches continued for about a month
after the February 19 release. Corrections to errors in transcriptions or
annotations, as identified by Papers of George Washington staff, have been
ongoing. So, too, have further regularization and consolidation: since first
publication, PGW staff have provided us with fully normalized lists of names of
all document authors and recipients that we have used to update the document
metadata, and with a corrected and up-to-date list of repository abbreviations and
expansions based on
MARC
codes that we have used to globally update the XML volume files.
Planned Enhancements
So is
PGWDE done? Yes and no. It is a stable release
version with some remaining imperfections, but there's a lot more we plan to do
with it, even apart from adding content as we digitize new volumes after they
appear in print. We've recently met with the PGW staff to agree on a list of
priorities for enhancement. The tasks fall into three general categories:
- Optimization of existing features. Examples: improving search speed and
index retrieval; rewriting the search parser to make it more Google-like and
to include more boolean operators.
- New features. Add an “advanced search” page that will allow users to
search by document features or language, for example. We'll add full-text
searching on the index. Farther down the road may be “keeping up with the
Joneses” enhancements, like enabling the user to save a personal
workspace containing bookmarked documents and search result sets as the Works of Jonathan Edwards
Online at Yale has done.
- Features required by aggregating PGWDE into
the larger Founding Era collection. Over the next year we will be adding
editions of the Adams Family Papers, the Papers of Thomas Jefferson, and The Documentary History of the Ratification of the
Constitution. In order to build an extensible framework, we have
already begun a thorough rewrite of our delivery code and encoding
specifications so that our publication system will scale gracefully as more
publications are added.
But it is impossible for us to project all of the enhancements that may
become desirable or possible in the future. We are in a position not too
different, really, from that of the editors who began planning American
documentary historical editions beginning in the 1950s. They of course knew that
their completed volumes would eventually need to be supplemented and corrected as
new documents and historical information emerged. What they couldn't envision was
a time, our own, when scholarship that exists only in print is increasingly seen
as
ipso facto lacking an essential quality. Likewise, we
can really only guess at what new features advancing technology may make possible.
For instance, we have often wished we had time and funding to add rich XML tagging
for personal and place names (beyond the author/recipient identifications in
existing metadata), and assumed that this would require a major commitment of
human labor. But it's not impossible that advances in automated
named entity
recognition will enable us in a not-too-distant future to pipe
PGWDE through a program that will reliably recognize and
tag those names and link them to, say, genealogical databases or a GIS-based
interface like Google Earth.
So we don't really expect ever to be able to say more than that PGWDE is done — for now.
A Few Conclusions: Generalizing from the Rotunda Experience
Ask any developer out there if a program is ever finished and they'll tell
you, “No, of course not, I still need to...”. But, ask any developer
out there if the program is almost finished and, assuming that
the development cycle has progressed far enough along, their answer will
invariably be, “Yes, all I have to do is...”. They may even quantify
it: “80% complete”. Ask them a couple of days, weeks, months (depending
on the magnitude of the project) later and you will get a similar response,
but with a different percentage, say 90%. And so forth…but never 100%. (“Never Finished; or Zeno's Paradox as an Analog to Software
Development”)
It is entirely possible to define “done-ness” for computer software in such a
way that no instantiation of a software project can ever satisfy the definition. For
example, suppose we stipulate that: “A program is complete and can be distributed
when it (1) satisfies all initial design requirements, (2) is known to run 100%
bug-free under all potential conditions of use, and (3) incorporates
state-of-the-art programming techniques and tools at the time of distribution to
offer the user an optimal mix of powerful features and ease of use.” The second
and third criteria entail that none but the simplest programs could ever count as
complete. To claim “my program is bug-free because I have found no bugs in it”
is argumentum ad ignorantiam; that you haven't found any doesn't
mean they aren't there. And criterion 3 turns done-ness into a Red Queen's race,
since the state of the art is constantly advancing, and at release time most complex
software projects are already “belated” relative to the cutting edge of
technology. To the extent that digital publications count as software projects, they
too would fail ever to count as finished under such a definition.
The adjective “simplest” in the preceding paragraph hints at a way around the
paradox. If my goal is to write a “hello world” program in Perl and I respond
with the one-line program “
print "Hello World!\n";
” I can confidently say I'm done. If my task is to write a new operating system,
it's another matter. As a rule, the more
complex a task is,
the less susceptible it is of being judged finished by any set of formal criteria.
Contrast these two assignments:
- Create a crop circle in the shape of a simple circle with a diameter of 40
meters.
- Create a crop circle representing the coastline of Great Britain at 1/10000
scale.
The first project is done once you've made a single circuit while tromping
down wheat at the end of a 20m tether. The second project is done when you've created
as accurate a representation as your time and skill allow. Tracing a coastline is a
problem in fractal geometry for which completeness will always be relative. In a
sense, it is a formal property of project 2 that it is done only when you decide it
is done. To put it another way, intrinsic criteria are used in both cases to
determine when the project qualifies as finished, but as project 2 is formally
undecidable (embodying the Turing halting problem that Matt Kirschenbaum mentions in
his introduction), extrinsic criteria are also required to make the
determination.
The digital publications that we have worked on in Rotunda have tended to resemble
the fractal project more than the simple circle. With PGWDE, for example, the “coastline” we needed to reproduce was, like
that of Britain, a pre-existing and well-defined object, the fifty-two volumes of the
print edition. To have omitted a volume would have been as clear a sign of
incompleteness as leaving Cornwall out of the crop circle. But decisions about the
richness of our feature set were very much a matter of “how far to trace”, and
in the end were dictated by our available time and skill (and budget). If our
experience is representative, deciding when to call a digital project “done”
usually requires a process of negotiation between intrinsic criteria and external
factors.
Intrinsic Criteria
Intrinsic criteria are formalist: they assume that the completeness of an object
derives from its inner properties alone, without reference to any social or other
external context. In the following table there is a continuum from objects like a
monograph (or a lyric poem) that can be judged to possess organic unity, to ones
like a collaborative virtual world that cannot. It is no accident that the latter
are the ones felt to be characteristically “digital”.
Category |
Object has definable boundaries? |
Object has satisfied its design goals? |
Print world example |
Digital example |
Is it “done”? |
1 |
yes |
yes |
monograph, journal article |
monograph-like object, online article |
yes |
2 |
yes |
no |
preprint, “rough cuts”
|
beta or 0.9 release |
not yet |
3 |
no |
yes (for current stage) |
encyclopedia; any work issued in discrete series |
same as for print world |
yes (for now) |
4 |
no |
no |
? ? |
open-ended wiki, collaborative blog or social space, virtual world,
etc. |
no (by definition) |
Table 1.
“Done” as a function of intrinsic criteria
Category 1 objects are the most familiar in scholarly publishing, hence the most
fully integrated into the tenure-and-review system. Bryant's
Typee essentially falls into this category. Probably owing to the
influence of online publishing, Category 2 objects are becoming more familiar:
online preprints are accredited scholarly communications in a growing number of
disciplines, and cutting-edge book publishers like O'Reilly with their
Rough Cuts series of
early-access PDF are adopting a “versioned” model of publication.
Category 3 objects are also familiar from the print world, where they represent
the one kind of open-endedness that does not upset traditional notions about
scholarly authority. The Oxford English Dictionary is
a good example. It has been supplemented, transformed, and extended many times
since the first fascicles were issued in 1884. Yet each discrete stage of
publication was accepted by the academic community as authoritative for its
moment. It is no accident that this category has translated easily to digital
format: the only essential difference between the print and online OED is that the latter is updated far more often.
Category 4 is the one to which the term “done” seems the most inapplicable.
Its characteristic objects are more like processes than products, and it is
difficult to think of genuine analogues in the print world outside the realm of
experimental literature of the Oulipo variety. A publication like
Wikipedia can perhaps be seen as a special case of Category 3 in
which discrete stages succeed each other with extreme rapidity, but a virtual
world like
Second Life exists in such constant motion that it requires something akin to calculus
for adequate description. Unsurprisingly, Category 4 is the form of digital
creation least amenable to naturalization in the academic reward system or the
scholarly publishing marketplace.
Extrinsic Factors
For a scholarly publisher, intrinsic criteria of done-ness are important but are
often trumped by extrinsic factors. The judgement that a book manuscript is done
and ready for press requires an agreement among author, acquiring editor, external
reviewers, and the manuscript editorial and production departments that is based
largely on its formal content. But completely extrinsic factors such as the desire
to include the book in a particular season's list will often lead a press to veto
an author's wish to continue tinkering with a manuscript. Similarly, an author may
not consider a monograph on Chinese art formally complete without the inclusion of
several dozen full-page color reproductions on glossy inserts, but a publisher may
omit them for the wholly extrinsic reason that the profit-and-loss sheet doesn't
budget for them. Once a book is in print, decisions about its subsequent
“done-ness” (i.e., whether to reprint, revise, issue in paperback, etc.)
are based almost entirely on economic factors. In the case of digital
publications, I will suggest, extrinsic factors become important at an earlier
stage and are proportionately more important at every stage of composition and
publication.
The following list of extrinsic factors is not meant to be exhaustive; they are
the ones that have been most prominent in Rotunda's experience.
-
Economic constraints
Two maxims apply: (1) if a digital publication doesn't sell, it's
“done”; (2) if the projected cost of upgrade exceeds projected
revenue, it's “done”
. (For freely distributed projects, substitute “when no more grant
funding or volunteer time is available, it's done.”)
-
Competition
Maxim: when your competition is adding features to its product, they
can render your finished product “incomplete.”
In the print world, this phenomenon is familiar in textbook and
reference publishing. In the digital world, it is absolutely pervasive. No
online publication, free or for sale, can afford long-term stasis when the
peer publications it is compared with are adding bells and whistles (a list
that would include, as of 2008, things like Ajax-powered form fields, tag
clusters, user reviews and personal workspaces, page previews on mouseover,
selectable themes or “skins” . . .) In the prestige economy as in the
market economy, keeping up with the Joneses is not optional.
-
Standards evolution
Maxim: even absent competition, the evolution of standards can make a
finished project “incomplete.”
This is primarily a matter of adhering to best practices, though not
entirely free from the keeping-up-with-the-Joneses factor. Certainly if your
academic discipline adopts a new format for metadata, or your institution
adds a requirement that Web publications meet accessibility guidelines, your
projects need to be revised for conformity. In other cases, it may be a
matter of pride to demonstrate that a project has upgraded to the latest
standard, for example by converting archival XML from TEI P4 to P5
compliance, or by following the very latest W3C recommendation for XHTML or
CSS.
-
Aggregation
Maxim: a stand-alone publication will probably become
“incomplete” when it is aggregated with other material. In
Rotunda's experience, it is inevitable that the user interface and back-end
coding one develops for a single digital project will need to be
substantially revised once a second project is added and meant to
interoperate with the first. (As a case in point, it would require major
effort to get our first publication, the Dolley Madison Digital Edition, seamlessly integrated with PGWDE, as
the back-end programming and underlying XML data formats of the two
publications are quite different.)
-
Technological change
Maxim: new technology will make your publication
“incomplete”
. This goes almost without saying. The evolution of hardware,
operating systems, programming languages, and Web standards will eventually
make any online publication obsolete. Failure to migrate a digital object
periodically as technical conditions require is the analogue of allowing a
published book to go out of print. (In fact it's worse: it's like printing
the book on high-acid-content paper with ink that fades on exposure to
light, and then letting it go out of print.)
A Necessary Synthesis
Whether you are a publisher or the editor of an open-access publication, allowing
extrinsic factors to influence your decision about whether a digital project is
done is in no way an admission of defeat or an abdication of responsibility. It
is, in fact, the only way to avoid the form of Zeno's paradox whimsically
propounded in the epigraph to this section. The progress of knowledge in the arts
and sciences is continuous, but in order for it to happen at all, scholarly
discourse must be distributed in the form of discrete objects that can be shared,
read or viewed, responded to, assimilated, quoted, disputed, and revised. In the
marketplace of ideas, it's less important how you decide when your piece is done
than that you do decide, label it and put it on display, and prepare
to haggle with others over its value.
Appendix: XML markup of a sample Washington letter
(Formatting has been applied for convenience in reading; carriage returns are not
introduced within mixed-content elements in the original files.)
<div1 xml:id="Rev13d180" type="doc"> <FGEA:mapData
id="GEWN-03-13-02-0189"> <bibl> <title>From William Livingston, 13
January 1778</title> <author>Livingston, William</author> <name
type="recip">GW</name> <date when="1778-01-13"/> </bibl>
<FGEA:Author>Livingston, William</FGEA:Author>
<FGEA:Recipient>GW</FGEA:Recipient> <FGEA:mapDates>
<FGEA:searchRange from="1778-01-13" to="1778-01-13"/> <FGEA:dayRange
from="1778-01-13" to="1778-01-13"/> </FGEA:mapDates> <FGEA:pageRange
from="Rev13p227" to="Rev13p227"/> </FGEA:mapData> <pb n="Rev13p227"/>
<head>From William Livingston</head> <div2 type="docbody">
<opener> <salute>Sir</salute> <dateline>Morris Town [N.J.]
<date when="1778-01-13">13th Jany 177[8]<ptr n="1"
target="Rev13d180n1"/></date></dateline> </opener> <p>Upon
frequent Complaints that Capt. Kennedy's Residence at his Farm was injurious to the
State, & occasioned great Clamours from the People in This Neighbourhood, the
Council ordered his Attendance on the Board — they at the same time desired a
Gentleman near the Spot, to procure what Affidavits he could respecting Captn
Kennedy's Conduct — He sent us by return of the Express three Affidavits with Copies
of which I take the Liberty of troubling you; Capt. Kennedy denies the Accusations
sworne against him, & refers to a Parole he signed to your Excellency in this
Town. The Board would therefore be glad to know the Nature of that Parole (of which
he has no Copy) & whether you consider him as a Prisoner of War, since Your
Excellency has taken Paroles from persons professedly Subjects of this State
& not pretending to any Connextions with Britain, meerly to prevent their
being detrimental to this State as disaffected Subjects — If he is considered as a
Prisoner we suppose him exchangeable & in the mean time it would probably be
best to have him removed at a greater Distance from the Enemy's Lines — If his Parole
was taken only to prevent mischief & in Aid of the magestrate whose Authority
was then very inadequate to suppress Disaffection we shall consider him as altogether
within the Civil Line<ptr n="2" target="Rev13d180n2"/> — I have the Honor to be
With great Respect your Excellys Most Hum: Servt</p> <closer>
<signed>Wil: Livingston</signed> </closer> <ps> <p>P.S.
I am sorry that Troup has been suffered to return to the Enemy after being so clearly
convicted of being a Spy. I have this moment received Intelligence that a party is
engaged to way-lay me between this place & my house, of which I have reason
to think Troup is at the bottom.<ptr n="3" target="Rev13d180n3"/></p>
</ps> </div2> <div2 type="docback"> <note type="source"
xml:id="Rev13d180sn"> <p> <bibl n="docSource"> <rs type="dType">
<abbr>LS</abbr> </rs> <rs type="dWhere"> <ref
target="GWPrep37" type="repository">DLC:GW</ref> </rs> </bibl>
The postscript is in Livingston's writing.</p> </note> <note n="1"
type="fn" xml:id="Rev13d180n1"> <p>Livingston wrote '1777.'</p>
</note> <note n="2" type="fn" xml:id="Rev13d180n2"> <p>The New
Jersey council of safety had agreed on 10 Jan. 'that a Warrant do issue for
apprehending and bringing before the Board Archibald Kennedy Esqr, that an Enquiry be
made into his past conduct, and that the Oaths of Government may be tendered to him.'
On 13 Jan., Kennedy appeared before the council, which after considering 'sundry
Affidavits' resolved that 'a letter be written to Genl Washington respecting the
nature of Captn Kennedy's Parole & that copies of the Affidavits relative to
his conduct be also transmitted with the Same' (<ref target="PGWst1204"
type="short-title"> <hi rend="italic">N.J. Council of Safety
Minutes</hi></ref>, 186–88). Copies of the affidavits of Nathaniel Camp,
Jr., Robert Neil, and Robert Nicholls, all dated 12 Jan., are in <ref
target="GWPrep37" type="repository">DLC:GW</ref>. See also <ref
type="document" target="Rev13d236">GW's letter to Kennedy of 20
January</ref>.</p> </note> <note n="3" type="fn"
xml:id="Rev13d180n3"> <p>See <ref target="Rev13d242"
type="document">GW's first letter to Livingston of 20
January</ref>.</p> </note> </div2> </div1>
Works Cited
Flanders 2007 Flanders, Julia. “Panel presentation during session Coalition of Digital Humanities Centers”. Presented at Digital Humanities 2007 (2007).
Jackson 1976 Jackson, Donald, W.W. Abbot, Dorothy Twohig, Philander D. Chase, Theodore J. Crackel and Edward G. Lengel, eds. The Papers of George Washington. Charlottesville, VA: University of Virginia Press, 1976-.
Landow 1992 Landow, G. P. Hypertext: The Convergence of Contemporary Critical Theory and Technology. Baltimore: The John Hopkins University Press, 1992.
Landow and Delany 1991 Landow, George P., and Paul Delany. “Hypertext, Hypermedia and Literary Studies: The State of the Art”. In George P. Landow and Paul Delany, eds., Hypermedia and Literary Studies. Cambridge, MA: MIT Press, 1991. pp. 3-50.
Pierazzo 2007 Pierazzo, Elena. “The Encoding of Time in Manuscripts Transcription: Toward Genetic Digital
Editions”. Presented at Digital Humanities 2007. Digital Humanities 2007
Conference Abstracts (2007).