Arianna Ciula | 27 Feb 12:41 2015

Cultural Litercay in Europe 2015 - programme out

*** Apologies for cross-posting ***


Cultural Literacy in Europe

Full information is on the project website: 

To book your place: BOOKING CLOSES ON 30 MARCH 2015

This international conference brings together academics interested in the development of Literary and Cultural Studies in Europe and beyond with European policymakers. The language of the conference will be English. Venue: Birkbeck, University of London: Gordon Square WC1H 0PD and Malet Street (Torrington Place) WC1E 7HX. The conference includes five keynote talks, 36 short papers, a policymaker panel, an art installation, a film showing, a book-launch, a poster session and a practical workshop.

Researchers practising literary-and-cultural studies [LCS] examine a vast range of social and cultural objects through the lens of literary thinking – analysing textuality, fictionality, rhetoricity and historicity, and developing ‘cultural literacy’. How can LCS research and cultural literacy contribute to solving the major challenges of Europe today?

Academic fields:

Cultural memory, Translation & migration, Digital textuality, Biopolitics & the body.

Keynote speakers:

Prof Aleida Assmann, Prof Michael Cronin, Prof Lars Elleström, Prof Sir David King & Prof Alexandre Quintanilha


Dr Monica Dietl, Prof Maureen Freely, Dr Philippe Keraudren, Prof Svend Erik Larsen, Prof Wolfgang Mackiewicz, Angela Schindler-Daniels, Prof Milena Žic Fuchs

Queries to

Dr Arianna Ciula

Department of Humanities

University of Roehampton | London | SW15 5PH arianna.ciula <at>  |

Tel: +44 (0) 20 8392 5763

Follow us on Twitter <at> UORHumanities <at> ariciula

Julia Flanders | 25 Feb 19:56 2015

Women Writers Online, free for Women's History Month

Dear all, 

We're happy to announce that Women Writers Online will be freely available during the month of March, in
celebration of Women's History Month. We invite you to explore the collection starting March 1 at:

Women Writers Online now contains more than 350 texts published between 1526 and 1850, including new works
by Aphra Behn, Charlotte Turner Smith, and Mercy Otis Warren. We hope you enjoy the collection!

Best wishes,


Julia Flanders
Director, Women Writers Project
Director, Digital Scholarship Group
Interim Co-director, NULab
Northeastern University
DCMI Announce | 23 Feb 23:43 2015

DC-2015 Professional Program call for participation

***Apologies for cross-posting***

DC-2015 -- DCMI 20th Anniversary International Conference & Annual Meeting
September 1-5, 2015 — São Paulo, Brazil

Conference Theme: 
Metadata and Ubiquitous Access to Culture, Science and Digital Humanities


:: DC-2015 Professional Program ::

Professional Program Objectives

Each of the past 20 years, practitioners and researchers in metadata have gathered for DCMI's annual meeting and international conference. This year, we will be gathering in São Paulo, Brazil. The scope of interests for the Professional Program is broad and inclusive of all aspects of innovative metadata design, implementation, and best practices in all sectors from cultural heritage institutions to governments and commerce. 

While the work of the DCMI progresses throughout the year, the annual meeting and conference provide the opportunity for researchers, newcomers to metadata, students, apprentices, and early career professionals to gather face-to-face to share experiences and knowledge and to learn through Tutorials, Workshops, and Special Sessions tracks.

In addition, the Best Practice Posters & Demonstration tracks offer the practitioner the opportunity to: 
  • showcase metadata work underway or completed in their particular contexts, 
  • seek input from the larger community on metadata matters of concern, 
  • compare notes on innovative practice, and 
  • cast a broader light into their particular metadata work silos. 
Through such a gathering of the metadata communities of practice, DCMI advances its "first goal" of promoting metadata interoperability and harmonization across the metadata ecosystem. 

Submission & Session Languages

Proposals for Professional Program tracks may be in Portuguese or English. Depending on the language of the session presenters, simultaneous English/Portuguese or Portuguese/English translation will be provided. Submissions and presentations by Portuguese speakers are strongly encouraged. 

Important Professional Program Deadlines
  Tutorials, Workshops, and Special Session
  ---Proposal Abstract Deadline: 28 March 2015
  ---Proposer Notification: 25 April 2015
  Best Practice Posters & Demonstrations
  ---Submission Deadline: 14 July 2015
  ---Author Notification: Ongoing

Questions Regarding the Professional Program?

-- Questions in Portuguese: 
     Mariana Curado Malta - mariana(at)iscap(dot)ipp(dot)pt
-- Questions in English: 
     Stuart Sutton - sasutton(at)dublincore(dot)net; or
     Mariana Curado Malta - mariana(at)iscap(dot)ipp(dot)pt

Conference Organizers:
--Universidade Estadual Paulista--São Paulo State University (UNESP)
--Dublin Core Metadata Initiative (DCMI)

Please join us at DC-2015 in São Paulo, Brazil?
Hayim Lapin | 23 Feb 22:33 2015

String-range() pointing

Dear TEI-ers,

I recently encountered a problem and I don't know if it is an encoding error on my part or a validation problem under TEI-all.

I am developing markup of morphological segments from a classical Hebrew text. Since the markup stands off from the source text, I want to use string-range on the element m to designate the segment, thus:
<w xml:id="abc">postprandial</w>
and then, elsewhere,
<m type="prefix" corresp="string-range(#abc,0,4)" ana="...">post</m>
So far so good.

The problem arises because the particular <w> is in another file, and I want to use a prefix definition to abbreviate the path to the file. So, if the path to the file is "../folder/text.xml", and the prefix is "wd", the following is valid,
<m type="prefix" corresp="string-range( ../folder/text.xml#abc,0,4)" ana="...">post</m>
but the following, using the prefix, is not
<m type="prefix" corresp="string-range(wd:abc,0,4)" ana="...">post</m>
Am I doing something wrong?

Please advise!
-- Hayim Lapin Robert H. Smith Professor of Jewish Studies Professor of History University of Maryland College Park, MD 20742 301 405 4296 |
Dubnicek, Ryan C | 23 Feb 20:12 2015

Register Now for HTRC UnCamp - March 30-31, 2015

Apologies for cross posting!

Registration Now for HTRC UnCamp, March 30-31, 2015


HathiTrust Research Center UnCamp

March 30-31, 2015


University of Michigan

Palmer Commons

100 Washtenaw Avenue

Ann Arbor, MI 48109-2218


2015 UnCamp: 

This year’s HathiTrust Research Center UnCamp will be held March 30-31, 2015 at the University of Michigan Palmer Commons. This is the third iteration of the UnCamp—an event that is part hands-on coding and demonstration, part inspirational use-cases, part community building, part informational, all structured in the dynamic setting of an un-conference programming format. It has visionary speakers mixed with boot-camp activities and hands-on sessions with HTRC infrastructure and tools. This year’s keynote speakers are Professor Michelle Alexopoulos, of the University of Toronto Department of Economics and Professor Erez Lieberman Aiden of the Department of Genetics at the Baylor College of Medicine. Read more about Michelle and Erez on the HTRC website.


Registration now at

There are a limited number of spots, and some sessions will be participant driven, so plan to register in advance of the deadline (March 16, 2015)

Additional information, including detailed bios for speakers, introductions for keynotes and the full UnCamp program, will be posted at as it becomes available. The 2012 and 2013 UnCamp programs and presentations are also still available online.

Greta Franzini | 23 Feb 13:20 2015

CfP: Göttingen Dialog in Digital Humanities 2015

**With apologies for cross-posting**

Call for Papers: Göttingen Dialog in Digital Humanities

The Göttingen Dialog in Digital Humanities (GDDH) has established a new forum for the discussion of digital methods applied to all areas of the Humanities, including Classics, Philosophy, History, Literature, Law, Languages, Social Science, Archaeology and more. The initiative is organized by the Göttingen Centre for Digital Humanities (GCDH).

The dialogs will take place every Tuesday at 5pm from late April until early July 2015 in the form of 90 minute seminars. Presentations will be 45 minutes long and delivered in English, followed by 45 minutes of discussion and student participation. Seminar content should be of interest to humanists, digital humanists, librarians and computer scientists.

We invite submissions of complete papers describing research which employs digital methods, resources or technologies in an innovative way in order to enable a better or new understanding of the Humanities, both in the past and present. Themes may include text mining, machine learning, network analysis, time series, sentiment analysis, agent-based modelling, or efficient visualization of big and humanities-relevant data. Papers should be written in English. Successful papers will be submitted for publication as a special issue of Digital Humanities Quarterly (DHQ). Furthermore, the author(s) of the best paper will receive a prize of €500, which will be awarded on the basis of both the quality and the delivery of the paper.

A small budget for travel cost reimbursements is available.

Full papers should be sent by March 20th to gkraft <at> in Word .docx format. There is no limitation in length but the suggested minimum is 5000 words. The full programme, including the venue of the dialogs, will be sent to you by April 1st.

For any questions, do not hesitate to contact gkraft <at>
For further information and updates, visit

GDDH Board (in alphabetical order):

Camilla Di Biase-Dyson (Georg August University Göttingen)
Marco Büchler (Göttingen Centre for Digital Humanities)
Jens Dierkes (Göttingen eResearch Alliance)
Emily Franzini (Göttingen Centre for Digital Humanities)
Greta Franzini (Göttingen Centre for Digital Humanities)
Angelo Mario Del Grosso (ILC-CNR, Pisa, Italy)
Berenike Herrmann (Georg August University Göttingen)
Péter Király (Gesellschaft für wissenschaftliche Datenverarbeitung mbH Göttingen)
Gabriele Kraft (Göttingen Centre for Digital Humanities)
Bärbel Kröger  (Göttingen Academy of Sciences and Humanities)
Maria Moritz (Göttingen Centre for Digital Humanities)
Sarah Bowen Savant (Aga Khan University, London, UK)
Oliver Schmitt (Gesellschaft für wissenschaftliche Datenverarbeitung mbH Göttingen)
Sree Ganesh Thotempudi (Göttingen Centre for Digital Humanities)
Jörg Wettlaufer (Göttingen Centre for Digital Humanities & Göttingen Academy of Sciences and Humanities)
Ulrike Wuttke (Göttingen Academy of Sciences and Humanities)

This event is financially supported by the German Ministry of Education and Research (No. 01UG1509).

-- Greta Franzini Postdoctoral Researcher Göttingen Centre for Digital Humanities Faculty of Mathematics and Computer Science Georg-August-University Göttingen Papendiek 16 D-37073 Göttingen, Germany W: E: gfranzini <at> T: <at> GretaFranzini
Jens Østergaard Petersen | 23 Feb 06:44 2015

Re: temporary page-based XML fragments for curatorial purposes

Hello Peter, 

Have you checked this against <> and <>? I can't very well move around the <pb/>s in the original. The empty <l/> and <sp/> are there because the page breaks occur at the onset and close of these elements – I cannot delete elements just because they are empty. 

Personally, I believe a page break should always occur inside a text node, but I am probably the only one who thinks so. This is a question related to the theory of text behind TEI, not of extracting these XML fragments.



On 23 Feb 2015 at 03:12:27, Robinson, Peter (peter.robinson <at> wrote:

As Jens’ piece of code shows, extracting XML (as opposed to simple data strings) between milestone elements is far from straightforward.  One can see this too in the example from Jens in an earlier posting, for the putative pages 5 and 6, which I paste in below.

A few points:
— why does the <pb/> come after the opening of the <div> etc? why not before the first <div>? so: <body><pb/><div… The point is of some consequence, if you are seeking to identify exactly what text is on the page
— most difficult: both page 5 and 6 contain <div n=“1”>. One might assume that given consecutive pages both beginning and ending <div n=“1”> that the <div> spans the pages. But there are cases where this may not be the case (for example, two versions of the one poem written on consecutive pages).  So how do we identify exactly which elements span across pages, which do not? You can see the problem with the speech by the king which starts on page 5.  It appears that this speech (<sp>) continues on page 6. But what if it does not? How do you mark which <sp>, <div> etc elements run across page boundaries? (by the way — I’m puzzled by the empty <l/> and <sp/> elements at the end of each page.  what are these?)

page 5:
<TEI xmlns="">
    <text xml:lang="eng">
            <div n="1" type="act">
                        <pb facs="tcp:21002:5"/>Was that the warlike Prince of Portingale,</l>
                    <l>That by our Nephew was in triumphled?</l>
                <!-- omitted -->
                    <l>Then by my iudgement thus your strife shall end,</l>
                    <l>You both deserue and both shall haue reward.</l>
                    <l>Nephew, thou tookst his weapon and his horse,</l>

Page 6:
<TEI xmlns="">
    <text xml:lang="eng">
            <div n="1" type="act">
                        <pb facs="tcp:21002:6"/>His weapons and his horse are thy reward.</l>
                    <l><hi>Horatio</hi> thou didst force him first to yeeld,</l>
                    <l>His ransome therefore is thy valours fee:</l>
                    <l>Appoint the sum as you shall both agree.</l>
                    <l>But Nephew thou shalt haue the Prince in guard,</l>
                    <l>For thine estate best fitteth such a guest.</l>
                    <l><hi>Horatios</hi> house were small for all his traine,</l>
                    <l>Yet in regarde thy substance passeth his,</l>
                    <l>And that iust guerdon may befall desert,</l>
                    <l>To him we yeeld the armour of the Prince.</l>
                    <l>How likes. <hi>Don Balthazar</hi> of this deuice?</l>
                <!-- omitted -->
                    <l>My soueraign pardon the Author of ill newes,</l>
                    <l>And Ile bewray the fortune of thy Sonne.</l>

On Feb 22, 2015, at 8:19 PM, Jens Østergaard Petersen <oesterg <at> GMAIL.COM> wrote:

Dear David,

Sorry, you are right: I XML-parsed the string output.Ine

I don't understand why a node-based approach is not used for this. The transform below I cobbled together after reading your mail. It gives the same result as using util:get-fragment-between(). It can easily be generalised.




xquery version "3.0";

declare namespace tei="";

declare function local:get-page-from-pb($element as element(), $start-attribute-value as xs:string,
                                        $end-attribute-value as xs:string)
as element()
    element { node-name($element) } {
        $element/ <at> *,
        for $child in $element/node()
            if ($child instance of element()) then
                if ($child/descendant-or-self::tei:pb[ <at> n eq $start-attribute-value] or
                    $child/descendant::tei:pb[ <at> n eq $end-attribute-value]) then
                    local:get-page-from-pb($child, $start-attribute-value, $end-attribute-value)
                else if ($child/preceding::tei:pb[ <at> n eq $start-attribute-value]/parent::* and
                    $child/following::tei:pb[ <at> n eq $end-attribute-value]/parent::*) then
                    local:get-page-from-pb($child, $start-attribute-value, $end-attribute-value)
            else if ($child instance of text()) then
                if (($child/parent::*/tei:pb and $child/following::tei:pb[ <at> n eq $start-attribute-value]) or
                    ($child/parent::*/tei:pb and $child/preceding::tei:pb[ <at> n eq $end-attribute-value])) then
                else if ($child/preceding::tei:pb[ <at> n eq $start-attribute-value] or $child/following::tei:pb[ <at> n eq
                    $end-attribute-value]) then

let $doc := doc('/db/eebo/A00002/A00002.xml')/tei:TEI
    local:get-page-from-pb($doc, "5", "6")


Gives same result as:

let $doc := doc('/db/apps/sarit-data/data/eebo/A00002/A00002.xml')
let $fragment := util:get-fragment-between($doc//tei:pb[ <at> n eq "5"], $doc//tei:pb[ <at> n eq "6"], true(), true())

On 22 Feb 2015 at 21:18:45, Birnbaum, David J (djbpitt <at> wrote:

Dear Jens (cc TEI-L),

xs:string : the string containing the fragment between the two node/milestone elements.

If it returns an XML fragment, so much the better. Perhaps the documentation is wrong, or perhaps I've misunderstood it, although I just read it again and it still seems to say string. Returning a fragment and fixing up the markup is possible, even if the selection isn't well-formed or well-balanced on its own; that's what a JavaScript Range interface does.



From: Jens Petersen <oesterg <at>>
Date: Sunday, February 22, 2015 at 3:07 PM
To: David Birnbaum <djbpitt <at>>
Cc: TEI-L <TEI-L <at>>
Subject: Re: temporary page-based XML fragments for curatorial purposes

On Sunday, February 22, 2015, Birnbaum, David J <djbpitt <at>> wrote:
Dear Martin (cc TEI-L),

>I note your concern that the re-integration of page fragments may be
>difficult. That may be where this project collapses. But the eXist
>function mentioned by Jens Petersen in his response looks interesting.

I note that this eXist function returns a string, which means that
internal markup is lost.

No, an XML fragment is returned, with full markup. With the path completion, it is actually a mini TEI document. 

I am curious to know how you could think it is a string.

 Whether that's a problem depends on how much
you're extracting, but since individual words may contain internal markup,
it can be challenging even at that rather granular level.



Fiddling with my iPad ...

Fabio Ciotti | 22 Feb 22:36 2015

Fwd: Are "independent headers" a thing of the past?

This use case can be implemented with a server side transformation
producing the stand alone TEIheader before sending the OAI response (I
did it some years ago for biblioteca italiana project hacking the OCLC
oai server code).
There are a lot of possible solutions for this. I do not think this is
a definitive reason to call back to life independent headers.

Il 22/feb/2015 21:47 "Stuart A. Yeates" <syeates <at>> ha scritto:

> I have a use case for an independent header that I've never had time
> to implement (and likely will never get time to implement).
> There's a protocol in library land called OAI-PMH which is used for
> querying remote repositories for their contents. It's used by google
> scholar, for example to learn about the contents of issues of academic
> journals.
> OAI-PMH is essentially parametrised by the format of metadata returned
> for each item. Variants of Dublin Core and MARCXML are common, but
> there's no reason why TEI XML couldn't be exposed alongside these.
> cheers
> stuart
> --
> ...let us be heard from red core to black sky
> On Mon, Feb 23, 2015 at 7:03 AM, Lou Burnard
> <lou.burnard <at>> wrote:
> > Hi Piotr
> >
> > If you do a quick search through the archives, you will find that the
> > decision to abandon the "independent header" was quite deliberate, nothing
> > to do with conceit,  and followed quite a bit of independent research. We
> > seem to be weak on institutional memory, so here's a brief summary of some
> > of the salient events.
> >
> > In P4, there was a chapter which defined an "independent header". It defined
> > one element (<ihs>) which could be used (rather than <TEI.2>)  as the root
> > element of an SGML dtd for documents containing only <teiHeader> elements.
> > It also contained some (lengthy) discussion of mapping between elements of
> > the header and other library standards, notably MARC.
> >
> > Reviewing the components of P4, in April 2004, the Metalanguage workgroup
> > noted that the concept of "auxiliary DTD" is no longer relevant in the P5
> > architecture, and that specifically "The IHS is an artefact to enable a
> > valid document consisting only of headers, which could be accomplished in
> > other ways." [] The
> > "other ways" they had in mind were presumably defining an appropriate ODD or
> > embedding TEI Header elements inside a document in another namespace.
> >
> > The council meeting in July 2004
> > ( notes  concerns
> > that the library-standards-mapping aspects of the old chapter should not be
> > lost, and that its material should be reworked either as a new chapter or
> > (as eventually happened) as a component of the existing header chapter. [1].
> > No-one seems to have argued for retaining an element like the old <ihs> in
> > P5, and no serious use case was proposed for it.
> >
> > Two members of the Libraries SIG, who were also members of the Council, did
> > start work on a revision of the old chapter in a bid to bring its
> > recommendations on best practices for interoperability up to date  and a
> > draft of this document[2] was considered for inclusion in P5. However,  at
> > the April 2007 Council meeting
> > ( where the
> > status of all P5 drafts chapters was reviewed, Council felt the work was
> > incomplete, probably out of scope for the Guidelines, and better handled  by
> > the Libraries community. [3]. Again, there is no record of any proposal for
> > a possible container element for a set of <teiHeader> elements.
> >
> > In February and March 2013, there was a brief discussion of this "ghost"
> > chapter on the Council mailing list, which reiterates (most of) the history
> > I have summarised above, and re-affirms that the best way forward is to
> > incorporate recommendations from the Libraries community into the Header
> > chapter. Kevin Hawkins concluded this by observing  [on 17/03/2013] "To
> > close the loop (you all know how much I like to do that), I now think it
> > would be better to leave this unfinished work as is and focus instead, as
> > we've previously discussed, on incorporating relevant portions of the Best
> > Practices for TEI in Libraries into the Guidelines."
> >
> > Since that date, there have indeed been a few proposals for change to the
> > Header chapter addressing this need. But no-one till now has asked for
> > anything like the old <ihs> to be reinstated. As I said before, it is hard
> > to see what the use case for such an element would be, given the
> > availability of other mechanisms for doing the same job.
> >
> > [1] "Noting the intention to remove the chapter on the Independent Header,
> > P[erry] W[illett] commented that members of the Libraries SIG attached
> > considerable importance to the material presently in that chapter about the
> > relationship between the TEI Header and other metadata schemes, which should
> > be  maintained and updated. Council noted that there was a need for such
> > comparative information, whether or not the independent header module was
> > retained as a distinct schema. It was agreed that the editors should try to
> > draft a discussion of related issues  which might become part of the header
> > chapter or a free standing document.
> > Action [by 1 sep 2004] editors:     draft document discussing relationship
> > of TEI Header to  other meta standards on the basis of material already
> > available, to be handed to the TEI Libraries SIG for  improvement and
> > updating."
> >
> > [2] The draft lingers on in the TEI source tree at
> > P5/Source/Defunct/SH-Other-Metadata-Standards.xml
> >
> > [3] " M[atthew] D[riscoll] recommends chapter be dropped. J[ohn] W[alsh]
> > (co-author of the re-write) concurs. JW was assigned to draft a few
> > paragraphs discussing the relationship between the TEI Header and other
> > standards (including MARC and Dublin Core) in general terms, without
> > detailed mappings.
> > Action 52: JW TRAC . Draft
> > paras on metadata standards 2007-05-05
> > Action 53: SB put JW paras in appropriate place at end of HD 2007-05-12 "
> >
> >
> >
> >
> >
> >
> >  On 21/02/15 20:32, Piotr Bański wrote:
> >>
> >> Hi Peter,
> >>
> >> Thank you for the response! Well, the header I've just prepared has a <p>
> >> under <text> and <body>, which contains a single sentence with a single
> >> <ref> in it. Of course, this can be called, hmm, abuse of the intentions.
> >>
> >> But it's the intentions that I'm interested in. Independent headers used
> >> to be printed on many of the flags that the TEI once waved. Have we got so
> >> conceited as to burn them all? A lot of selling points gone in that smoke,
> >> it seems.
> >>
> >> Best regards,
> >>
> >>   Piotr (rubbing his hands in anticipation...)
> >>
> >> On 21/02/15 21:05, Peter Stadler wrote:
> >>>
> >>> Others will probably be able to tell you more on the historic shift from
> >>> P4 to P5 but to my knowledge a TEI P5 conformant document *must* contain "a
> >>> single teiHeader element followed by a single text element, in that order“
> >>> [1,2] (I’m skipping the teiCorpus stuff and note that we introduced
> >>> model.resourceLike which can replace <text>, but still, there has to come
> >>> something after <teiHeader>)
> >>> So, officially there are no independent headers any more, I’d say!?
> >>>
> >>> Hope that helps
> >>> Peter
> >>>
> >>> [1]
> >>> [2]
> >>>
> >>>> Am 20.02.2015 um 12:02 schrieb Piotr Bański <bansp <at> O2.PL>:
> >>>>
> >>>> Hi all,
> >>>>
> >>>> Are independent/free-standing TEI headers gone with P4?
> >>>>
> >>>> I can't find any mention on them, either in the P5 guidelines or in the
> >>>> wiki.
> >>>>
> >>>> I can find this, though:
> >>>>
> >>>>
> >>>>
> >>>> And a poster from TEI <at> 20, by Michelle Dalmau and Melanie Schlosser,
> >>>> addressing among others "alternatives to the [independent header schema] in
> >>>> light of P5" (which might just indicate a change of the technology for
> >>>> maintaining said schema).
> >>>>
> >>>> I'll be grateful for any pointers and information.
> >>>>
> >>>> Best regards,
> >>>>
> >>>>   Piotr
> >>>
> >>>
> >>>
> >

Martin Mueller | 22 Feb 16:23 2015

temporary page-based XML fragments for curatorial purposes

I am thinkin about the collaborative curation of the 25,000 EEBO-TCP texts
that have moved into the public domain. What are proper tools and methods
for going about this? Whatever is done will involve the consistent
application of fairly granular divide and conquer strategies across a wide
swath of texts. Improvements are made word by word or page by page, and a
document of many pages may be touched by many hands.

The TCP texts exist in XML encoded form, with light structural annotation
and page breaks. For many curatorial tasks, including changes or additions
to the XML encoding at the microlevel of words, sentences, lines of verse
or paragraphs, the page remains the most convenient chunk because it
offers the quickest route to the "source," the digital scan of a microfilm
of the printed page.  Is there a "smart" way of creating page fragments,
such that each page (or double page) is a <div>, with opening or closing
elements automatically provided so as to make the page fragment a
well-formed document.  The process would have to algorithmically
reversible so that the fragments can be re-integrated into the text. And
it would have to scale to thousands of texts, some of which which many
hundreds of pages. 

Below is a very straightforward example, with the elements in capital
letters marking the elements that need to be provided to create the
minimal scaffold. There are probably much more complicated cases.

It seems to me that in principle something like that should be possible. I
don't know how hard it would be to do or whether similar things are being
done all the time. Note that the point of the exercise is not to create
beautiful encoding for ever but to create a temporary scaffold for the
kind of work that can be done page by page. A lot of work can be done that
way, and whatever can be done page by page is best done page by page
because readers know about pages.

 <l xml:id="A04942-e102350"><pb facs="5"/>Was that the warlike Prince of
          <l xml:id="A04942-e102360">That by our Nephew was in triumph
led? </l>
        <sp xml:id="A04942-e102370" who="A04942-general">
          <l xml:id="A04942-e102380">It was my Liege, the Prince of
Portingale. </l>
        <sp xml:id="A04942-e102390" who="A04942-king">
          <l xml:id="A04942-e102400">But what was he that on the other
          <l xml:id="A04942-e102410">Held him by th'arme as partner of the
prize? </l>
<sp xml:id="A04942-e102760" who="A04942-king">
          <l xml:id="A04942-e102770">I Balthazar, if he obserue this


Martin Mueller
Professor emeritus of English and Classics
Northwestern University

Laurie Allen | 20 Feb 22:23 2015

linking or reg?

Hi All,
I'm working with a linguist on encoding some 16th through 18th Century texts. The first text we're working on is a book published in 1578. We're mostly encoding for the structure of the book, and for the languages used in it (it's mainly written in spanish, with Zapotec, latin, and some Nahuatl interspersed).
In addition to showing the book in various ways, we are pulling in additional information when the Zapotec language is used in the book. For Zapotec terms, we join the TEI (in XSLT) to the linguistic analysis of the same zapotec term as exported from the linguists database. 

All good until...

The author of the book, in describing the Zapotec language, often says things that would translate to: (this is a made up example.) 

"They say over or well cooked when meat is done cooking."
However, the sentence would include three languages so that 
---"They" "say" "when meat is done cooking" would be in spanish. 
--- "over" and "well" and "cooked" would be in zapotec. 
--- "or" would be in latin. And it would actually be abbreviated as l.

So, we've been encoding it as:

They say <foreign xml:lang="cvz">over</foreign> <foreign xml:lang="lat"><choice><abbr>l</abbr><expan>vel</expan></choice></foreign> <foreign xml:lang="cvz">well cooked</foreign> when meat is done cooking.

Now, when the linguist enters this into her database, she says that this section includes 
"over-cooked" and "well-cooked" as phrases in Zapotec, which it does, in its meaning. This comes up a lot in the book, and it's important that the TEI include a matching over-cooked somewhere that can be searched for, and matched up with the exported linguistic analysis of over-cooked. 

I suspect I want to be using link and ptr here, but I'm having a difficult time making sense of how that would work. 

Many thanks for any advice.

Coordinator for Digital Scholarship and Services
<at> librlaurie
Haverford College Library
370 Lancaster Ave
Haverford, PA 19041
lallen <at>
Piotr Bański | 20 Feb 13:57 2015

jEdit page at the TEI wiki

Dear all,

Since I can see jEdit mentioned here and there in the context of XML 
editing and TEI, I have started a page with just the basic information 

Those who use / have used this tool are very welcome to expand the page 
(it can be done step by step, 5 minutes at a time -- there is a template 
there waiting to be filled in).

Thanks and best,