Welcome to the Critical Edition of the Hebrew Psalter Project

The first eclectic and digital critical edition of the first 50 Psalms.

 

This Virtual Manuscript Room digital workspace will support the three-year NEH-funded project “I Shall No Longer Want’ (Psalm 23:1): The Critical Edition of the Hebrew Psalter” (CEHP) to produce the first full eclectic and digital critical edition of the Hebrew Psalms 1-50, as well as the printed volume for Psalms 1-50 for the Hebrew Bible: A Critical Edition series. The project is co-directed by Prof. Brent Strawn and Dr. Drew Longacre at Duke University/Divinity School.

The project will move beyond existing scholarly editions that use one manuscript as their basic text and will instead collate and evaluate the best readings of the text of the Hebrew Bible from all available sources and publish those in an eclectic edition with critical apparatus and full textual commentary. Among these sources are the important Psalms manuscripts among the Dead Sea Scrolls that have been missing or inadequately treated in previous editions. The world has waited more than two thousand years for a critical text of one of the greatest classics of religious literature, the book of Psalms. When complete, all who study, translate, or read the Psalms will be able to say ‘I no longer want’—a play on Psalm 23:1—for a critical and reliable edition.

This exciting project would not be possible without the help of many collaborators and especially our volunteer transcribers to record the vast evidence of the ancient and medieval manuscripts in multiple languages. If you would like to participate, please register your interest by filling out this Google form. Note that, when you create transcriptions within this VMR system, they will be jointly owned by you and the project. Your transcription will be connected to your personal account, and you are free to reuse it as you would like. But you also grant the CEHP project rights to use, adapt, and/or publish the transcription in conjunction with the project.

Financial Sponsors

*If you're interested in partnering with us by supporting the research financially, please reach out to Prof. Brent Strawn.

National Endowment for the Humanities Scholarly Editions and Translations grant
*Any views, findings, conclusions, or recommendations expressed in this site, do not necessarily represent those of the National Endowment for the Humanities.

Mr. and Mrs. Alex and Anne Bernhardt

Mr. George Blumenthal

Rev. Dr. Edgardo Colón-Emeric

Rev. Dr. James and Lisa Howell, in memory of Bishop Thomas B. Stockton and in honor of Fr. Roland Murphy

Rev. John H. McMullen, Jr.

Bishop Carlton P. Minnick, Jr.

Mr. John Stubbs

Portrait

Dear Volunteers and Friends of the Critical Edition of the Hebrew Psalter Project:

We write you with heavy hearts. We have just learned last week that our federally funded grant through the National Endowment for the Humanities to produce the Critical Edition of the Hebrew Psalter was summarily terminated without warning by the US federal government. The long and short of this terrible news is that Drew’s position at Duke University must end because there is no longer funding for it, and Brent’s research leave to work on the project is immediately over as well.

This is obviously a major blow to the project and its timely completion. We wanted you to hear it from us, first, before learning of it some other way. We will appeal the decision with the NEH, but we do not expect the appeal to be successful. We are presently considering alternative funding sources, but we have no clear immediate prospects for continuing the research at its previous pace. Drew will need to be seeking regular employment elsewhere, and Brent will be occupied once again full time with teaching and service responsibilities. It seems that the work on the CEHP will, therefore, slow down considerably; indeed it may be on a long hiatus barring some unexpected benevolence.

We are, again, very sorry to report this news. We could never thank each of you enough for all you’ve done for this project and for us. We are forever grateful and we remain committed to bringing this work to completion as opportunity allows. We are convinced it is of great importance for biblical studies, Bible translation, and larger communities of faith, and we also wish to honor the work our volunteers and supporters have done. Indeed, one of the things that makes us most upset about this termination is that it is not just a cancellation of two researchers at a university somewhere. It is a cancellation of the work of some 200 volunteers drawn from across the world—indeed from every continent on the planet except Antarctica. Thankfully, none of that transcription work will be lost, but for the foreseeable future we will be considerably limited in our ability to manage new transcription work and construct the edition itself.

So please accept our heartfelt thanks for your help and service. We welcome your prayers and any suggestions or advice you might offer as we ponder our next steps. If you wish to stay involved with CEHP and help us continue toward our shared goals, we would be happy to know that as well.

Yours sincerely,

Brent and Drew

Portrait For those who are looking for more information about the methodology and goals of the CEHP project, Andrew Case has a nice podcast where we discuss these issues and examples in detail.
Portrait Yonat Shimron has published a nice article highlighting the work of our transcribers on the CEHP project.
Portrait If you want to learn more about our work with the Cairo Genizah Psalms manuscripts, Melonie Schmierer-Lee interviewed me on the Genizah Fragments blog about our work.
Portrait

Now that we have completed our first round of transcription assignments, I wanted to take stock of the progress we have made. Crowdsourcing in digital humanities projects is risky business—often leaving little or nothing in the way of useful results—so we owe it to the field and our supporters to be transparent about our experiment and our results. Thankfully, I am proud to be able to report truly astonishing success! Thanks to the generous contributions from our volunteer transcribers, we have made excellent progress already in this first round.

By way of background, I divided Psalms 1-50 into five sections that are roughly equal in size: Pss 1-15, 16-25, 26-34, 35-40, and 41-50. For each section of each manuscript, we require two independent transcriptions, which makes for 10 assignment slots per manuscript (unless it is lacunose for parts, like the Aleppo Codex [MA]). By tracking hours during our Psalm 22 pilot project, I was able to estimate that each of these assignments should take approximately 20 hours of transcriber time. Factoring in about 5 hours for setup and training, each first assignment should take about 25 hours to complete. The transcribers would have about 1.5 to 2 months to finish the assignments at their own pace.

With project management help from Caleb Punt, we distributed these assignments on our forum optimally based on transcribers’ reported competencies. It was critical to make sure to have enough material from Pss 1-15 to begin processing these psalms immediately, so I built in redundancy by assigning more transcribers than necessary for the assignment slots for Pss 1-15 (and a few others). We also selected a team of experienced scholars to transcribe the very difficult Berlin Codex (MZ)—which presents the Babylonian Masorah and is badly damaged—and planned some redundancy there as well. The response from the transcribers was amazing!

 

Indexing

We were able to complete the full indexing for every one of the manuscripts, identifying which verses are on which pages.

 

Completed Assignments

In all, we distributed a total of 120 first-round assignments to transcribers who had set up accounts in the Virtual Manuscript Room (VMR). 68 (= 57%) of these were completed on time and according to project standards! 50 out of 88 (57%) assignments for Hebrew manuscripts were completed, as were 18 out of 32 (56%) Greek assignments. At the estimated average of 25 hours per assignment (68 x 25 = 1700 volunteer hours), the transcribers completed in total approximately an entire year’s full-time work for an individual researcher over the course of a couple of months in the summer!

Of these completed assignments, only 5 Hebrew and 1 Greek (= 9%) turned out to be redundant. But even these are of value, because reconcilers can select the best two transcriptions to use in cases of redundancy.

 

Manuscript Coverage

For the 11 major manuscripts we chose to include, there were a total of 100 assignment slots required for full coverage, 62 Hebrew and 38 Greek. All of the Hebrew slots were assigned to transcribers, as were 29 out of the 38 required Greek slots. For the Hebrew assignments (which had greater redundancy built in), 45 (73%) distinct slots were filled with completed transcriptions. And the Greek transcribers filled 17 slots (59% of assigned slots; 45% of total slots). This makes for a grand total of 62 completed assignment slots accounting for 62% of the total project need. Not bad for a first round! Significantly, all of the manuscripts except GA had at least two completed assignments for Pss 1-15, which allowed us to begin reconciling transcriptions right away.

 

Quality Control

This progress was indeed remarkable and exceeded even my own optimistic expectations. But it raises the question of quality control. How useful and accurate are the data that were produced? There were several steps we took to ensure accuracy and compliance with project standards:

  1. Training - All transcribers underwent training regarding working with manuscripts and electronic transcription (live sessions were recorded for later viewing).
  2. Transcription manual - I wrote a detailed transcription manual for reference with instructions defining all project standards.
  3. Checking first pages - Transcribers had the option of requesting a more experienced transcriber to check their first transcription page upon completion to provide feedback and guidance for going forward.
  4. Forum - I set up a forum category for posting and answering questions regarding the transcription process, project standards, difficult passages, and website bugs.
  5. Checks and revisions - When transcribers reported their assignments as complete, I checked their work quickly and requested revisions where project standards were not yet met. After revisions, these assignments were marked as complete.
  6. Reconciliation - We have now begun the reconciliation process, which provides another critical quality control measure. A select team of experts have the job of comparing two completed transcriptions using the VMR’s reconciliation tool and adjudicating whenever they disagree. This process is intended to correct any remaining errors in the individuals’ transcriptions and to ensure adherence to project standards. So, unless both transcribers independently made the same mistake, the resulting approved project transcription should reflect the text and layout of the manuscript accurately. These project transcriptions are what will be published on the website and used for the digital edition.
  7. Proofreading and editing - In the future, we hope to have specialists proofread the project transcriptions against the manuscripts to identify any remaining errors and inconsistencies in meeting project standards. Additional problems may be flagged in the collation editing process. In these cases, we can edit existing project transcriptions directly.

In a project of this magnitude, there will inevitably be errors and inconsistencies in the data. But these quality control measures have proven both manageable and effective in ensuring a high-quality data set of manuscript transcriptions.

 

Feedback

In order to further evaluate our process, I sent a Google form to the transcribers soliciting feedback. We received 18 responses from transcribers, all of whom had completed their first assignments. These responses indicated a very high level of volunteer satisfaction. 

 

Overall, the transcribers were very satisfied with the tools and process as well. However, we received useful critical feedback in the following areas:

  1. Timely feedback - Because we struggled to manage the massive influx of volunteers at first, it took longer than expected to get the assignments distributed and to provide feedback to transcribers (e.g., checking first pages, replying to forum posts). We have brought Caleb Punt on board to help with project mangement going forward, and we expect subsequent rounds to be more manageable.
  2. Technical glitches - Some users experienced technical glitches on the website. In rare cases, this required redoing previously completed work. We have invited volunteers with software development skills to join a debugging team to help alleviate these problems in the future.

 

Retention

Of the 68 transcribers who completed their first assignments, 42 (62%) actively volunteered for a second round of assignments. This nicely matches the US national average for volunteer retention of about 65%. It reflects a high level of volunteer satisfaction (evident in the feedback), based on the perceived value of the work as an enjoyable learning experience and as meaningful service to the field. Furthermore, in the two months since the end of the first round of transcription, we have had 13 new volunteers register to do transcriptions. This suggests that our initial push for volunteers did not exhaust the pool of willing volunteers. Furthermore, several repeat transcribers have signed up to do more extensive transcription work for independent study credit, which considerably increases our capacity for further transcription.

 

Next Steps

We have recently released a second round of transcription assignments that aims to fill the remaining gaps in our coverage. Almost all remaining slots are currently assigned, so if the transcribers complete their tasks again this round, then we will essentially have full coverage for all of the originally selected manuscripts.

Our continued volunteer capacity has simultaneously allowed us to expand the corpus of manuscripts included. With help from Vince Beiler, we were recently able to add 8 new model Masoretic codices from the Firkovich collections in St. Petersburg that date from around the 10th-11th centuries. These are very substantial manuscripts that are just as old as the Aleppo and Leningrad codices, but have never before been transcribed or thoroughly studied for purposes of a critical edition. Ben Outhwaite and Kim Phillips have also helped us identify a selection of the most important fragments from the Cairo Genizah that we hope to include in the future. Thus, this crowdsourcing experiment has not only helped us achieve our project goals more quickly and efficiently, but also opened up new possibilities to make the edition even more robust.

In parallel with the ongoing second round of transcriptions, we have already started reconciling transcriptions for Pss 1-15 with the help of a select team of skilled and experienced volunteer transcribers. And I plan to begin editing a preliminary edition of Pss 1-15 in the coming weeks based on these transcriptions. Troy Griffitts and I will report on our progress and tools at an SBL Digital Humanities session on November 25, so please do join us for further updates and reflections.

 

Of course, none of this would have been possible with the support of our amazing volunteers, who have made this crowdsourcing experiment such a success. So I want to conclude this post with a big “Thank you!” and "Congratulations!" to all of you!

RSS (Opens New Window)