FRDB Archives

Freethought & Rationalism Archive

The archives are read only.


Go Back   FRDB Archives > Archives > Religion (Closed) > Biblical Criticism & History
Welcome, Peter Kirby.
You last visited: Yesterday at 03:12 PM

 
 
Thread Tools Search this Thread
Old 01-31-2007, 06:35 PM   #11
Veteran Member
 
Join Date: Jan 2007
Location: Mondcivitan Republic
Posts: 2,550
Default

Quote:
Originally Posted by spin View Post
This stirs up distant memories of texts I read long ago, one of which chopped up the Pauline corpus into sections and analysed them statistically for a collection of linguistic markers, lengths of sentences and paragraphs, to see if any distinctions exist between the sections. I seem to remember the names MacGregor and Morton in such statistical analyses, perhaps for the book of John, but certainly this methodology has no way of recognizing interpolations. It may distinguish between samples of texts, as long as the samples are large enough to make the statistics based on them indicative. I wonder what happened to the means of analysis.

spin
You might check out Anthony Kenney (I hope I didn't get that wrong, my library is all in boxes in the basement fort a while longer). A Q Morton co-wrote a study entitled Paul: the Man and the Myth (or via: amazon.co.uk) which studied mainly sentence length and to a lesser degree use of various grammatical features.

I once actually input his data into a spreadsheet and created graphs of the sentence length distributions for each of the letters of Paul, the general epistles (James, etc), and a few early fathers (Clement of Rome and maybe a couple others).

You are quite correct that it is possible to use stylostatistical analysis when you have a sizeable, unquestionably authentic corpus to use as a control against the unknown sample. You also need a suitably large sample to ensure that the results aren't significant simply by chance.

Unfortunately, the individual letters of the Pauline corpus are just too small to yeild high confidence in an analysis, and besides, there is a methodological problem inherent in using the entire corpus as the control for analysis of a specific book within the corpus.

Dave
DCHindley is offline  
Old 01-31-2007, 07:17 PM   #12
Contributor
 
Join Date: Jun 2000
Location: Los Angeles area
Posts: 40,549
Default

Do you mean this Anthony Kenny - author of The Computation of Style: an introduction to statistics for students of literature and humanities ?

Googling it brings up this Masters Thesis, which I have not read yet:

Quantitative Authorship Attribution: A History and an Evaluation of Techniques

=====

More on Morton:

A Hypertextual History of Humanities Computing: The Pioneers

Quote:
Morton was reported in the New York Times, 7 November 1963 for his claim that a computer study of sentence length and Greek function words 'proved' Paul only wrote four of the letters attributed to him. John Ellison in "Computers and the Testaments" (Computers for the Humanities?, 1965) replied to Morton by using his methods on James Joyce's Ulysses (five authors) and Morton's own essays (several authors). Morton was also heavily criticised for deciding sentence length by modern punctuation. Subsequently Morton steered clear of the bible and remained with less contentious Greek texts.
Toto is offline  
Old 01-31-2007, 08:59 PM   #13
Veteran Member
 
Join Date: Apr 2001
Location: Orions Belt
Posts: 3,911
Default

Quote:
Originally Posted by Malachi151 View Post
Is there computer software that analyzes texts to determine the likelihood that a passage is an original part of the text or from a later hand?
I saw something like this once. But when I tried to read the code, it was all greek to me....

Thank you. Thank you very much.
Kosh is offline  
Old 02-01-2007, 10:25 AM   #14
Veteran Member
 
Join Date: Feb 2004
Location: Washington, DC (formerly Denmark)
Posts: 3,789
Default

There is very little material on the topic of authorship and/or interpolation detection. I have written a bunch of software to do just that and I have blogged quite a bit on some of the methodology (see early entries here: http://julian.textcrit.com) that can be used on words. What I didn't have, and what would help immensely, is a syntactically parsed text. I have since been working on creating just that and am quite close. Once we have syntactic information we add a whole new layer of information to our stylometric representation of a text. There are a number of papers on how all these factors help by using google: http://www.google.com/search?hl=en&s...eek+authorship I am not sure how many of these you will be able to read since most are protected, I have excellent access from my office here at the patent office.

The main problem is that any computer methods work less and less well the less text you have. Most of the time an interpolation is so short as to yield little to no information.

Julian
Julian is offline  
Old 02-01-2007, 10:33 AM   #15
Veteran Member
 
Join Date: Apr 2001
Location: Orions Belt
Posts: 3,911
Default

Quote:
Originally Posted by Julian View Post
I have since been working on creating just that and am quite close.
Hopefully you have been using something like the UNIX utility "lex" (lexigraphical analyzer) to help you write the parser?

I'm sure there's something better than that out there these days.
Kosh is offline  
Old 02-01-2007, 10:51 AM   #16
Veteran Member
 
Join Date: Feb 2004
Location: Washington, DC (formerly Denmark)
Posts: 3,789
Default

Quote:
Originally Posted by Kosh View Post
Hopefully you have been using something like the UNIX utility "lex" (lexigraphical analyzer) to help you write the parser?

I'm sure there's something better than that out there these days.
No, that would not be possible. A simple CFG is not sufficient to parse a free word order language like Greek. Lex, yacc and all those things are very nice but insufficient. I did use a CFG (context-free grammar) and Parse::RecDescent to define my compiler compiler (not a mistake, I did mean to write that twice) but there are no simple rules that can define rules for natural language processing. That's why there are so many approaches like LFG, SFG, X-Bar, CDG, etc, etc... My approach is a hierarchial multi-pass linear sequence analyzer combined with a dependency grammar algorithm and that seems to do a pretty decent job.

Julian
Julian is offline  
Old 02-01-2007, 07:14 PM   #17
Veteran Member
 
Join Date: Jan 2007
Location: Mondcivitan Republic
Posts: 2,550
Default

Quote:
Originally Posted by Toto View Post
Do you mean this Anthony Kenny - author of The Computation of Style: an introduction to statistics for students of literature and humanities ?
Well, actually no.

While Anthony Kenny (not Kenney, my bad) wrote _A Stylometric Study of the New Testament_ (Oxford-New York-Toronto-Sydney-Paris-Frankfurt, 1986), I was actually thinking of Kenneth J. Neumann, The Authenticity of the Pauline Epistles in the Light of Stylostatistical Analysis (or via: amazon.co.uk) (Atlanta, 1990).

I have really got to get my library out of those boxes.

Dave
DCHindley is offline  
Old 02-01-2007, 07:52 PM   #18
Veteran Member
 
Join Date: Jan 2007
Location: Mondcivitan Republic
Posts: 2,550
Default

Quote:
Originally Posted by Julian View Post
There is very little material on the topic of authorship and/or interpolation detection.
I think we should distinguish between the confirmation of possible authorship from detection of interpolations. As I mentioned above, I do not believe it is possible to determine whether any of the 13 books attributed to Paul are *actually* his, due to limitations of the materials that have been preserved.

That being said, I do believe that stylostatictical analysis can help us determine the degree to which a text seems to be written by the same author throughout (homogeneity).

Interpolations, at least as I use the term, refers to glosses or other words and phrases that get added to an original author's text by an editor or redactor, not a copyist making spelling or grammatical corrections or accidentally including a marginal note.

One could mark off text that one thinks is intrusive to the sentence or argumentative structure, or offers an explanation, or is a form of commentary, and see if the remaining text is more homogeneous, but that can be a slippery slope. Some authors are more complex writers than others, so reducing the number of aporia or anomolies may be deceptive.

This is the problem I have been grappling with in my ever so evil and obviously wrong hypothesis that the Pauline letters were originally written by a non-christian who felt that gentiles who trusted in the promises made to Abraham by God were justified in God's sight without having to fully convert to Judaism.

These letters were later adopted and adapted by a radically anti-semite editor/redactor who added all sorts of digressions and glosses to make Paul into a good christian of his own era.

Dave Hindley
DCHindley is offline  
Old 02-02-2007, 11:40 AM   #19
Contributor
 
Join Date: Mar 2003
Location: London UK
Posts: 16,024
Default

Quote:
my ever so evil and obviously wrong hypothesis
The Spruce Goose flew!

(And why no further mention of Ellegard in this thread?)
Clivedurdle is offline  
Old 02-02-2007, 05:04 PM   #20
Contributor
 
Join Date: Jun 2000
Location: Los Angeles area
Posts: 40,549
Default

Interesting synchonicity - this article just popped into my inbox:

Shedding Light on The Dark Tower: A C.S. Lewis mystery is solved.

Quote:
Ten years after her original article, Lindskoog published her formal allegations against Hooper in The C. S. Lewis Hoax. She considered The Dark Tower to be "inferior writing … a blemish on [Lewis's] otherwise stable intellect." .... She concluded that Hooper had forged the manuscript and lied about rescuing papers from a fire.

With the 1994 release of the movie Shadowlands, Lindskoog reissued her book as Light in the Shadowlands, adding two new chapters. In this edition, she reported on a new study by the Rev. A. Q. Morton, which employed cusum (cumulative sum) statistical analysis of the first 23 sentences of chapter one of The Dark Tower, the first 24 sentences of chapter four, and the first 25 sentences of chapter seven, comparing them with similar passages from Out of the Silent Planet and That Hideous Strength. This type of style analysis has been used to prove that Shakespeare did not write his plays, that Paul did not write some epistles attributed to him, and that Jesus did not speak some sayings attributed to him. It assumes that a person's use of language remains constant over one's lifetime and in all situations. Morton concluded that Lewis could not have written chapters one and four, but that he did write chapter seven. Therefore, The Dark Tower was "a composite work."
But - a witness has been found who authenticates the work as Lewis'.
Toto is offline  
 

Thread Tools Search this Thread
Search this Thread:

Advanced Search

Forum Jump


All times are GMT -8. The time now is 01:01 AM.

Top

This custom BB emulates vBulletin® Version 3.8.2
Copyright ©2000 - 2015, Jelsoft Enterprises Ltd.