Archive for the ‘Tools’ Category

Sunoikisis is a successful national consortium of Classics programs developed by the Harvard’s Center for Hellenic Studies. The goal is to extend Sunoikisis to a global audience and contribute to it with an international consortium of Digital Classics programs (Sunoikisis DC). Sunoikisis DC is based at the Alexander von Humboldt Chair of Digital Humanities at the University of Leipzig. The aim is to offer collaborative courses that foster interdisciplinary paradigms of learning. Master students of both the humanities and computer science are welcome to join the courses and work together by contributing to digital classics projects in a collaborative environment.

According to Denis Diderot and Jean le Rond d’Alembert the purpose of an encyclopedia in the 18th century was ‘to collect knowledge disseminated around the globe; to set forth its general system to the people with whom we live, and transmit it to those who will come after us, so that the work of preceding centuries will not become useless to the centuries to come’. Encyclopedias have existed for around 2,000 years; the oldest is in fact a classical text, Naturalis Historia, written ca 77 CE by Pliny the Elder.

Following the (recent) digitalization of raw data, new, digital forms of encyclopedia have emerged. In our very own, digital era, a Wiki is a wider, electronic encyclopedia that is open to contributions and edits by interesting parties. It contains concept analyses, images, media, and so on, and it is freely available, thus making the creation, recording, and dissemination of knowledge a democratised process, open to everyone who wishes to contribute.

A Sprint for Digital Classicists

For us, Digital Classicists, scholars and students interested in the application of humanities computing to research in the ancient and Byzantine worlds, the Digital Classicist Wiki is composed and edited by a hub for scholars and students. This wiki collects guidelines and suggestions of major technical issues, and catalogues digital projects and tools of relevance to classicists. The wiki also lists events, bibliographies and publications (print and electronic), and other developments in the field. A discussion group serves as grist for a list of FAQs. As members of the community provide answers and other suggestions, some of these may evolve into independent wiki articles providing work-in-progress guidelines and reports. The scope of the Wiki follows the interests and expertise of collaborators, in general, and of the editors, in particular. The Digital Classicist is hosted by the Department of Digital Humanities at King’s College London, and the Stoa Consortium, University of Kentucky.

So how did we end up editing this massive piece of work? On Tuesday July 1, 2014 and around 16:00 GMT (or 17:00 CET) a group of interested parties gathered up in several digital platforms. The idea was that most of the action will take place in the DigiClass chatroom on IRC, our very own channel called #digiclass. Alongside the traditional chat window, there was also a Skype voice call to get us started and discuss approaches before editing. On the side, we had a GoogleDoc where people simultaneously added what they thought should be improved or created. I was very excited to interact with old members and new. It was a fun break during my mini trip to the Netherlands, and as it proved, very focused on the general attitude of the Digital Classicists team; knowledge is open to everyone who wishes to learn and can be the outcome of a joyful collaborative process.

The Technology Factor

As a researcher of digital history, and I suppose most information system scholars would agree, technology is never neutral in the process of ‘making’. The magic of the Wiki consists on the fact that it is a rather simple platform that can be easily tweaked. All users were invited to edit any page to create new pages within the wiki Web site, using only a regular web browser without any extra add-ons. Wiki makes page link creation easy by showing whether an intended target page exists or not. A wiki enables communities to write documents collaboratively, using a simple markup language and a web browser. A single page in a wiki website is referred to as a wiki page, while the entire collection of pages, which are usually well interconnected by hyperlinks, is ‘the wiki’. A wiki is essentially a database for creating, browsing, and searching through information. A wiki allows non-linear, evolving, complex and networked text, argument and interaction. Edits can be made in real time and appear almost instantly online. This can facilitate abuse of the system. Private wiki servers (such as the Digital Classicist one) require user identification to edit pages, thus making the process somewhat mildly controlled. Most importantly, as researchers of the digital we understood in practice that a wiki is not a carefully crafted site for casual visitors. Instead, it seeks to involve the visitor in an ongoing process of creation and collaboration that constantly changes the Web site landscape.

Where Technology Shapes the Future of Humanities

In terms of Human resources some with little involvement in the Digital Classicist community before this, got themselves involved in several tasks including correcting pages, suggesting new projects, adding pages to the wiki, helping others with information and background, approaching project-owners and leaders in order to suggest adding or improving information. Collaboration, a practice usually reserved for science scholars, made the process easier and intellectually stimulating. Moreover, within these overt cyber-spaces of ubiquitous interaction one could identify a strong sense of productive diversity within our own scholarly community; it was visible both in the IRC chat channel as well as over skype. Several different accents and spellings, British, American English, and several continental scholars were gathering up to expand this incredibly fast-pacing process. There was a need to address research projects, categories, and tools found in non-english speaking academic cultures. As a consequence of this multivocal procedure, more interesting questions arose, not lest methodological. ‘What projects are defined as digital, really’, ‘Isn’t everything a database?’ ‘What is a prototype?’. ‘Shouldn’t there be a special category for dissertations, or visualisations?’. The beauty of collaboration in all its glory, plus expanding our horizons with technology! And so much fun!

MediaWiki recorded almost 250 changes made in the 1st of July 2014!

The best news, however is that this, first ever wiki sprint was not the last. In the words of the Organisers, Gabriel Boddard and Simon Mahony,

‘We have recently started a programme of short intensive work-sprints to
improve the content of the Digital Classicist Wiki
(http://wiki.digitalclassicist.org/). A small group of us this week made
about 250 edits in a couple of hours in the afternoon, and added dozens
of new projects, tools, and other information pages.

We would like to invite other members of the Digital Classicist community to
join us for future “sprints” of this kind, which will be held on the
first Tuesday of every month, at 16h00 London time (usually =17:00
Central Europe; =11:00 Eastern US).

The Perseus Catalog is an attempt to provide systematic catalog access to at least one online edition of every major Greek and Latin author (both surviving and fragmentary) from antiquity to 600 CE. Still a work in progress, the catalog currently includes 3,679 individual works (2,522 Greek and 1,247 Latin), with over 11,000 links to online versions of these works (6,419 in Google Books, 5,098 to the Internet Archive, 593 to the Hathi Trust). The Perseus interface now includes links to the Perseus Catalog from the main navigation bar, and also from within the majority of texts in the Greco-Roman collection.

The metadata contained within the catalog has utilized the MODS and MADS standards developed by the Library of Congress as well as the Canonical Text Services and CTS-URN protocols developed by the Homer Multitext Project. The Perseus catalog interface uses the open source Blacklight Project interface and Apache Solr. Stable, linkable canonical URIs have been provided for all textgroups, works, editions and translations in the Catalog for both HTML and ATOM output formats. The ATOM output format provides access to the source CTS, MODS and MADS metadata for the catalog records. Subsequent releases will make all catalog data available as RDF triples.

Other major plans for the future of the catalog include not only the addition of more authors and works as well as links to online versions but also to open up the catalog to contributions from users. Currently the catalog does not include any user contribution or social features other than standard email contact information but the goal is to soon support the creation of user accounts and the contribution of recommendations, corrections and or new metadata.

Abstract: The Humboldt Chair of Digital Humanities at the University of Leipzig sees in the rise of Digital Technologies an opportunity to re-assess and re-establish how the humanities can advance the understanding of the past and to support a dialogue among civilizations. Philology, which uses surviving linguistic sources to understand the past as deeply and broadly as possible, is central to these tasks, because languages, present and historical, are central to human culture. To advance this larger effort, the Humboldt Chair focuses upon enabling Greco-Roman culture to realize the fullest possible role in intellectual life. Greco-Roman culture is particularly significant because it contributed to both Europe and the Islamic world and the study of Greco-Roman culture and its influence thus entails Classical Arabic as well as Ancient Greek and Latin. The Humboldt Chair inaugurates an Open Philology Project with three complementary efforts that produce open philological data, educate a wide audience about historical languages, and integrate open philological data from many sources: the Open Greek and Latin Project organizes content (including translations into Classical Arabic and modern languages); the Historical Language e-Learning Project explores ways to support learning across barriers of language and culture as well as space and time; the Scaife Digital Library focuses on integrating cultural heritage sources available under open licenses.

Details of the project, its components, and rationale are provided in the original post.

The members of the Diccionario Griego-Español project (DGE, CSIC, Madrid) are pleased to announce the release of DGE online (http://dge.cchs.csic.es/xdge/), first digital edition of the published section (α-ἔξαυος) of our Lexicon. Although still in progress, the DGE, written under the direction of Prof. F.R. Adrados, is currently becoming the largest bilingual dictionary of ancient Greek: it already includes about 60,000 entries and 370,000 citations of ancient authors and texts. Simultaneously, we are releasing the edition of LMPG online(http://dge.cchs.csic.es/lmpg/), the digital version of the Lexicon of Magic and Religion in the Greek Magical Papyri, written by Luis Muñoz Delgado (Supplement V of DGE). The digitization of this smaller Lexicon is considered as a successful prototype of this ambitious digitization initiative: further on DGE online will be improved with similar advanced features, such as the implementation of a customized search engine. Any critics and suggestions on that matter will be very welcome. We hope these new open access dictionaries will be of your interest and will become, to some extent, valuable tools for Ancient Greek studies.

According to this press-release from Europeana Professional, the massive European Union-funded project has released 20 million records on cultural heritage items under a Creative Commons Zero (Public Domain) license.

The massive dataset is the descriptive information about Europe’s digitised treasures. For the first time, the metadata is released under the Creative Commons CC0 Public Domain Dedication, meaning that anyone can use the data for any purpose – creative, educational, commercial – with no restrictions. This release, which is by far the largest one-time dedication of cultural data to the public domain using CC0 offers a new boost to the digital economy, providing electronic entrepreneurs with opportunities to create innovative apps and games for tablets and smartphones and to create new web services and portals.

Upon registering for access to the Europeana API, developers can build tools or interfaces on this data, download metadata into new platforms for novel purposes, make money off it, perform new research, create artistic works, or anything.

It’s important to note that it’s only the metadata that is being freely released here: I did a search for some Greek inscriptions, and also photographs and transcriptions are available, these are all fiercely copyrighted to the Greek Ministry of Culture: ” As for all monuments of cultural heritage, permission from the Greek Ministry of Culture is required for the reproduction of photographs of the inscriptions.” (According to this same license statement, even the metadata are licensed: “Copyright for all data in the collection belongs to the Institute for Greek and Roman Antiquity of the National Hellenic Research Foundation. These data may be used freely, provided that there is explicit reference to their provenance. ” Which seems slightly at odds with the CC0 claim of the Europeana site; no doubt closer examination of the legal terms would reveal which claim supercedes in this case.)

(It was lovely to be reminded that inscriptions provided by the Pandektis project [like this funerary monument for Neikagoras] have text made available in EpiDoc XML as well as Leiden-formatted edition.)

It would be really good to hear about any implementations, tools or demos built on top of this data, especially if that had a classics component. Any pointers yet? Or do we need to organize a hackfest to get it started….?

The University of Leipzig will host a hackathon that addresses two basic tasks. On the one hand, we will focus upon the challenges of creating a digital edition for the Greek author Athenaeus, whose work cites more than a thousand earlier sources and is one of the major sources for lost works of Greek poetry and prose. At the same time, we use the case Athenaeus to develop our understanding of to organize a truly born-digital edition, one that not only includes machine actionable citations and variant readings but also collations of multiple print editions, metrical analyses, named entity identification, linguistic features such as morphology, syntax, word sense, and co-reference analysis, and alignment between the Greek original and one or more later translations. (more…)

TextGrid (http://www.textgrid.de) is a platform for scholars in the humanities, which makes possible the collaborative analysis, evaluation and publication of cultural remains (literary sources, images and codices) in a standardized way. The central idea was to bring together instruments for the dealing with texts under a common user interface. The workbench offers a range of tools and services for scholarly editing and linguistic research, which are extensible by open interfaces, such as editors for the linkage between texts or between text sequences and images, tools for musical score edition, for gloss editing, for automatic collation etc.

On the occasion of the official release of TextGrid 2.0 a summit will take place from the 14th to the 15th of May 2012. On the 14th the summit will start with a workshop day on which the participants can get an insight into some of the new tools. For the following day lectures and a discussion group are planned.

Those who have been waiting impatiently for the first stable release of the Text Image Linking Environment (TILE) toolkit need wait no longer: the full program can be downloaded from: <http://mith.umd.edu/tile/>. From that site:

The Text-Image Linking Environment (TILE) is a web-based tool for creating and editing image-based electronic editions and digital archives of humanities texts.

TILE features tools for importing and exporting transcript lines and images of text, an image markup tool, a semi-automated line recognizer that tags regions of text within an image, and plugin architecture to extend the functionality of the software.

I haven’t tried TILE out for myself yet, but I’m looking forward to doing so.

The following is a lightly edited version of a talk that I delivered at the 26th Congress of the International Association of Papyrologists, 19 August 2010, in Geneva (program), posted here upon nudging of G. Bodard.

Colleagues. It is a great honor and a privilege to be able to speak with you today. An honor and a privilege that, I hasten to add, I did not seek, but which a number of our colleagues insisted some months back the members of this research team must try to live up to. If I approach this distinguished body with some trepidation it is perhaps because my training as an epigraphist has conditioned me to a tone less attuned to collegiality than that which informs the papyrologists’ discipline. I should add also that am here not to present my own work, but the fruits of a team whose members are in Heidelberg, London, New York, North Carolina, Alabama, and Kentucky, and who have been working heroically for more than three years now.

I shall aim to speak for no more than 40 minutes so that we may at least start discussions, which I know the rest of the team and I will be more than happy to carry on via email, Skype, phone, and separate face to face meetings. I will add also that, since the matters arising from this talk are highly technical in nature, we shall be more than happy to field questions as a team (I and my colleagues Rodney Ast, James Cowey, Tom Elliott, and Paul Heilporn) and in any of the languages within our competence.

First some background. I don’t need to tell you very much about the history of the Duke Data Bank of Documentary Papyri. It was founded in 1983, as a collaboration between William Willis and John Oates of Duke University, and the Packard Humanities Institute. A decade and a half later, around the time, as it happens, that APIS was also starting, the DDbDP decided to migrate from the old CD platform and to the web. John in particular was committed to making the data available for free, to anyone who wanted access. The Perseus Project, from Tufts University, very kindly agreed to host the new online DDbDP, to develop a search interface, to convert the data from old Beta code to a markup language called SGML–all at no cost to us. The DDbDP added a few thousand texts after switching from the Packard CD ROM to Perseus. But the landscape changed dramatically from this point onward, and the DDbDP began to fall behind. The end of the CD ROM meant the end of regular revenues to support data entry and proofreading. And of course, ongoing development of the search interface was not without cost to Perseus, whose generous efforts on our behalf were, as I mention, unremunerated. Within a few years the DDbDP was behind in data entry and the search interface was not able to grow and mature in the ways that papyrologists wanted.

A slight preview of what you’ll find there and of where this all might go:

The goal is to provide a robust and simple convention for indicating that citations are present. How robust? How simple? At a bare minimum, just wrap a citation in ‘<a class=”citation” href=”http://example.org”>…</a>’. That will distinguish intellectually significant citations from other links (such as to a home page for the hosting project). I cribbed the ‘class=”citation”‘ string from Matteo’s articles cited at the bottom of the wiki page. Please also consider adding a ‘title’ and ‘lang’ attribute as described.

We are also interested in encouraging convergence on best practices for communicating information about the entities being cited and about the nature of the citation itself:

Matteo has begun a page “Citations with CTS and Microformats“. CTS, developed by Neel Smith and Chris Blackwell, is important by way of its potential to provide stable URIs to well-known texts.

Merging these conventions is of ongoing interest. And they do illustrate that one goal is to converge on best practices that are extendable and not in unnecessary conflict with existing work.

While it isn’t represented on the wiki yet, I intend to start a javascript library that will identify citations in a page (e.g. jQuery’s “$(‘.citation’)” ) in order to present information about, along with options for following, a particular citation. Or to list and map all the dc:Location’s cited in a text. Etc.

Closing the loop: this work overlaps with a meeting held by the ISAW Digital Projects Team in NYC last week. The preliminary result is a tool for managing URIs in a shared bibliographic infrastructure. This is one example of an entity that can produce embeddable markup conforming to the ‘class=”citation”‘ convention. Such markup would be consumable by the planned js library. Any project that produces stable URIs can have an “Embed a link to here.” (vel sim) widget that produces conforming html for authors to re-use.

I’m grateful to Gabriel Bodard for letting me use the Digital Classicist wiki to start these pages and for encouraging me to summarize here. The effort is inspired by the observation that a little bit of common documentation, sharing, and tool building can lead to big wins for users and developers, as well as to greater interoperability for our citation practices going forward.

I’m looking specifically at general surveys, studies and discussions about the history of the relationship between classics and computers, a disciplinary field that has recently emerged as Digital Classics.

However, as Tom Elliott pointed me out Alison Babeu (Perseus project) has recently published on CiteULike a much broader bibliography as “as part of an IMLS-funded planning project that’s looking a digital infrastructure needs for Classics (Perseus and CLIR are the co-recipients of the grant)”.

For the time being, in order to allow anyone with any interest in this to contribute I created a group on Zotero called digitalclassics. The group is open (i.e. my authorisation is not needed to join) so please join it and start contributing your entries to the list. I’m thinking in particular of publications that I have unintentionally neglected and/or publications in other languages that I was not aware of.

Currently, the entries in the Zotero Library are divided into two main categories: general studies and applications, where the latter is meant to host publications concerning specific applications Digital Classics-related. More subcategories may be added as long as we go further or members of the list can even add new ones by themselves.As soon as the bibliography will reach a reasonably stable shape I will update the page I have already created on the DigitalClassicist wiki.

I want to thank in advance the DigitalClassicist community for the support they have shown me on the list and for the entries they have started already contributing.

The “Wired Campus” section of the Chronicle of Higher Education is reporting on the uses that humanities scholars have found for the U.S. Department of Energy’s High Performance Computing resources. The short article reports on the efforts of several people who have made use of the resources, including Gregory Crane of the Perseus Project, David Bamman, a computational linguist who has been mining data from classical texts, and David Koller, a researcher with the Digital Sculpture Project, which has developed ways to coalesce numerous images of an object into a high-resolution 3D image. The article reports that, according to Mr. Koller, intermediaries are needed who can help humanities and computer researchers communicate with each other.

Rationale: The meeting will bring technical and editorial researchers participating in, or otherwise engaged with, the IOSPE (Inscriptiones Orae Septentrionalis Ponti Euxini = Ancient Inscriptions of the Northern Black Sea Coast.) project together with researchers in related fields, both historical and computational. Existing projects, such as the Inscriptions of Roman Cyrenaica and Inscriptions of Aphrodisias, have explored the digitization of ancient inscriptions from their regions, and employed the EpiDoc schema as markup. IOSPE plans to expand this sphere of activity, in conjunction with an multi-volume publication of inscription data. This event is a joint workshop funded in part by a Small Research Grant from the British Academy, and in part by the eSI through the Arts and Humanities e-Science theme. The workshop will bring together domain experts in epigraphy, and specialists in digital humanities, and e-science researchers, which will provide a detailed scoping of the research questions, and the research methods needed to investigate them from an historical/epigraphic point of view.

The success of previous projects, and the opportunities identified by the IOSPE research team, raise questions of significant interest for the e-science community. Great interpretive value can be attached to datasets such as these if they are linked, both with each other, and with other relevant datasets. The LaQuaT project at King’s, part of ENGAGE, is addressing this. There is also an important adjunct research area in the field of digital geographic analysis of these datasets: again, this can only be achieved if disparate data collections can be meaningfully cross-walked.

Posted on behalf of Greg Crane. Link to the Treebank, which provides more information, at the very end of the post.

We are currently looking for advanced students of Greek and Latin to contribute syntactic analyses (via a web-based system) to our existing Latin Treebank (described below) and our emerging Greek Treebank as well (for which we have just received funding). We particularly encourage students at various levels to design research projects around this new tool. We are looking in particular for the following:

Get paid to read Greek! We can have a limited number of research assistantships for advanced students of the languages who can work for the project from their home institutions. We particularly encourage students who can use the analyses that they produce to support research projects of their own.

We also encourage classes of Greek and Latin to contribute as well. Creating the syntactic analyses provides a new way to address the traditional task of parsing Greek and Latin. Your class work can then contribute to a foundational new resource for the study of Greek and Latin – both courses as a whole and individual contributors are acknowledged in the published data.

Students and faculty interested in conducting their own original research based on treebank data will have the option to submit their work for editorial review to have it published as part of the emerging Scaife Digital Library.

I would like to draw to your attention the fact that registration for the 3D Colour Laser Scanning Conference at UCL on the 27th and 28th of March has now opened.

The first day (27th of March) will include a keynote presentation and papers on the themes of General Applications of 3D Scanning in the Museum and Heritage Sector and of 3D Scanning in Conservation.

The second day (28th of March) will offer a keynote presentation and papers on the themes of 3D Scanning in Display (and Exhibition) and Education and Interpretation. A detailed programme with the papers and the names of the speakers can be found in our website.

If you would like to attend the conference, I would kindly request to fill in the registration form which you can find in this link and return it to me as soon as possible.

There is no fee for participating (or attending the conference) (coffee and lunch are provided free of charge). Please note that attendance is offered on a first-come, first-served basis.

Please feel free to circulate the information about the conference to anyone who you think might be interested.

Search Pigeon is a collection of Google Co-opTM Custom Search Engines (CSEs) designed to make researching on the web a richer, more rewarding, and more efficient process.

Designed for researchers in the Arts and Humanities, with a decidedly interdisciplinary bent, the objective of Search Pigeon is to provide a tool enabling the productive and trustworthy garnering of scholarly articles through customized searching.

Right now SearchPigeon.org provides CSEs that search hundreds of peer-reviewed and open access online journals, provided they are either English-language journals, or provide a translation of their site into English.

Over the past few years, the idea of tool development as a scholarly activity in the digital humanities has been gaining ground. It has been the subject of numerous articles and conference presentations. There has not been, however, a concerted effort to gather information about the perceived value of tool development, not only as a scholarly activity, but in relation to the tenure and promotion process, as well as for the advancement of the field itself.

Ann Hanlon and myself have compiled such a survey and would be grateful if those of you who are or have been engaged in tool development for the digital humanities would take the time to complete an online Digital Humanities Tool Developers’ Survey.

You will need to fill up a consent form before you begin, and there is an opportunity to provide us with feedback on more than one tool (you simply take the survey again). The survey should not take more than 10-15 minutes. It is our intention to present the results of our survey at Digital Humanities 2008.

With all best wishes,

Susan Schreibman
Assistant Dean
Head of Digital Collections and Research McKeldin Library University of
Maryland College Park

There was an article in New Scientist last week on plans for permanent data preservation for the scientific data. The argument in the sciences seems to be that all data should be preserved, as some of it will be from experiments that are unrepeatable (in particular Earth observations, astronomy, particle accelerators, and other highly expensive projects that can produce petabytes of data). It is a common observation that any problems we have in the humanities, the sciences have in spades and will solve for us, but what is interesting here is that the big funding being thrown at this problem by the likes of the NSF, ESF, and the Alliance for Permanent Access is considered news. This is a recognised problem, and the sciences don’t have the solution yet… Grid and Supercomputing technologies are still developing.

(Interestingly, I have heard the argument made in the humanities that on the contrary, most data is a waste of space and should be thrown away because it will just make it more difficult for future researchers to find the important stuff among all the crap. Even in the context of archaeology, where one would have thought practitioners would be sensitive to the fragile nature of the materials and artefacts that we study, there is a school of thought that says our data–outside of actual publications–are just not important enough to preserve in the long term. Surely in the Googleverse finding what you want in a vast quantity of information is a problem with better solutions than throwing out stuff that you don’t think important and therefore cannot imagine anyone else finding interesting.)

Another important aspect of the preservation article is the observation that:

Even if the raw data survives, it is useless without the background information that gives it meaning.

We have made this argument often in Digital Humanities venues: raw data is not enough, we also need the software, the processing instructions, the script, presentation, search, and/or transformation scenarios that make this data meaningful for our interpretations and publications. This is in technical terms the equivalent of documenting experimental methodology to make sure that research results can be replicated, but it also as essential and providing binary data and documenting the format so that this data can be interpreted as structured text (say).

It’s good to see that this is a documented issue and that large resources are being thrown at it. We shall watch their progress with great interest.

Can Amazon succeed where others have found little traction so far? I already spend so much time reading freely available materials (journal articles, blogs, magazines, reviews) with my trusty Macbook Pro that I feel no need at all for a special-purpose e-text reader.

From Greg Crane comes the much-anticipated word that all of the hopper code and much of the content in Perseus is now officially open sourced:

November 9, 2007: o *Install Perseus 4.0 on your computer*:

All of the source code for the Perseus Java Hopper and much of the content in Perseus is now available under an open source license. You can download the code, compile it, and run it on your own system. This requires more labor and a certain level of expertise for which we can only provide minimal support. However, since it will be running on your own machine, it can be much faster than our website, especially during peak usage times. You also have the option to install only certain collections or texts on your version, making it as specialized as you wish. Also, if you want to use a different system to make the content available, you can do so within the terms of the Creative Commons http://creativecommons.org/licences/by-nc-sa/3.0/us license. This is the first step in open sourcing the code: you can modify the code as much as you want, but at this time, we cannot integrate your changes back into our system. That is our ultimate goal, so keep a look out for that!