Textos


Programa de doctorado en Lengua y Literatura, 1999/2000

Nuevas tendencias en la lingüística contemporánea

Eclecticismo en la práctica lingüística

Joseba Abaitua
Departamento de Filología Inglesa
Universidad de Deusto

Programa

Índice

  • Antecedentes (Jesús Mosterín, Yorick Wilks, Alan K. Melby)
  • Sintaxis: Teorías, formalismos, análisis, datos (Stuart Shieber). La jerarquía de Chomsky. Gramáticas de rasgos (Stuart Shieber, Ivan Sag)
  • Pragmática
    • Discurso, resolución de referencias anafóricas
    • Traductología y traductemas
    • Actos de habla y tipología textual
  • Lecturas recomendadas
  • Citas y otra bibliografía útil
  1. Primer día
  2. Segundo día
  3. Tercer día: Unidades de traducción y bitextos

Primer día

  1. Tres manzanas
  2. El maestro y los maestrillos
  3. Virtudes y límites de los formalismos
  4. Alternativas (1990):

References

Formal linguistics

Noam Chomsky. 1963. Formal Properties of Grammars. R. Kyce, R. Bush & E. Galanter (eds.) Handbook of Mathematical Psychology, vol II. Wiley.

Martin Kay. 1985. Unification in Grammar. V. Dahl & P. Saint-Dizier (eds.) Natural Language Understanding and Logic Programming, vol I: 251-278.

Alan K. Melby. 1995. The Possibility of Language. A discussion of the nature of language with implications for human and machine translation. John Benjamins

Jesús Mosterín. 1973. La matemática como lenguaje. Filosofía y ciencia en el pensamiento español contemporáneo (1960-1970): 25-36. Simposio de lógica y filosofía de la ciencia. Tecnos. Madrid.

C. Raymond Perrault. 1984. On the Mathematical Properties of Linguistic Theories. Computational Linguistics 10: 165-176.

Stuart Shieber. 1987. Separating Linguistic Analysis from Linguistic Theories. Whitelock, P., M Wood, H. Somers, R. Johnson, & P. Bennett. (eds.) Linguistic Theory and Computer Applications. Academic Press.

Yorick Wilks. 1976. Philosophy of Language. Eugene Charniak and Yorik Wilks (eds.) Computational Semantics: An introduction to Artificial Intelligence and Natural Language Comprehension North-Holland: 205-233.

Constraint-based grammars

Gerald Gazdar & G. K. Pullum. 1985. Computationally Relevant Properties of Natural Languages and their Grammars. W.J. Savitch et al. (eds.) The Formal Complexity of Natural Language: 387-437

Ronald Kaplan & Joan Bresnan. 1982. Lexical-Functional Grammar: A Formal System for Grammatical Representations. J. Bresnan (ed.) The Mental Representation of Grammatical Relations. MIT Press.: 173-280.

Martin Kay. 1985. Parsing in Functional Unification Grammar. D.R. Dowty, L. Karttunen & A.M. Zwicky (eds.) Natural Language Parsing, Cambridge University Press: 251-278

G. K. Pullum & Gerald Gazdar. 1982. Natural Languages and Context-Free Languages. Linguistics and Philosophy 4: 471-504.

Ivan A. Sag. 1991. Linguistic Theory and Natural Language Processing. Klein & Veltman (eds.) Natural Language and Speech. Springer-Verlag: 69-80

Peter Selles. 1985. Lectures on Contemporary Syntactic Theories. CSLI. Stanford University. [Versión española: 1989. Teorías sintácticas actuales. Editorial Teide, Barcelona].

Stuart Shieber. 1986. An introduction to unification-based approaches to grammar. CSLI. Standford University


Extracts from Mosterín 1973:

"La filosofía -decía Galileo (en Il Saggitario, 1963)- está escrita en ese grandísimo libro que continuamete está abierto ante nuestros ojos (a saber, el universo), pero no puede entenderse si antes no se procura entender su lenguaje y conocer los caracteres en que está escrito. Este libro está escrito en lenguaje matemático, y sus caracteres son triángulos, círculos, ...".

Para Galileo, la naturaleza es un libro escrito en lenguaje matemático. Como él mismo dice, si los hombres no la habían entendido hasta entonces es que "no conocían los caracteres en que estaba escrita". Pero la naturaleza no es un libro ni está escrita en lenguaje alguno. Lo que sí es un "libro" es una teoría física y ésta es la que puede estar escrita en lenguaje matemático.

El éxito de la física moderna que Galileo inaugura se debe -al menos en parte- a que, para hablar de la naturaleza, los físicos modernos dejan de hablar en el lenguaje metafísico en que habían hablado los antiguos y se ponen a hablar en un nuevo lenguaje: el lenguaje matemático.

Lo nuevo no es aquí -como la cita de Galileo parecería implicar- el conocimiento pasivo del lenguaje matemático, sino la aplicación activa de ese lenguaje a la descripción de la naturaleza.

A partir de Galileo, la ciencia deja de hablar del mundo en el lenguaje ordinario para pasar a hablar en el lenguaje de la matemática.

La ciencia física progresó rápidamente durante los siglos XVII, XVIII y XIX y el lenguaje matemático de la época de Galileo muy pronto se le hubiera quedado pequeño. Pero la matemática fue creciendo junto con la física, creando nuevos lenguajes matemáticos en los que formular las nuevas teorías físicas.

Al llegar el siglo XIX, los físicos disponían de una amplia gama de lenguajes matemáticos en que formular sus teorías, pero la estructura de estos lenguajes no estaba nada clara. La gran labor que se propusieron y en gran parte realizaron los matemáticos del siglo XIX fue la clarificación de esos lenguajes.

:26 Y la gran obra de Cantor fue la creación de un nuevo y extraordinariamente expresivo lenguaje: el lenguaje de los conjuntos, la teoría de los conjuntos. Resultaba que todos los otros lenguajes matemáticos eran traducibles al lenguaje conjuntista y que todos los otros conceptos matemáticos eran fácilmente definibles en este lenguaje. Con ello, la tarea de clarificar la estructura de los diversos lenguajes matemáticos de que disponía el físico se había convertido en la tarea de clarificar el lenguaje de la teoría de conjuntos.

Cantor creía haber descubierto los conjuntos que ya desde siempre estaban en el mundo. Pero en realidad no era un descubridor, sino un inventor. Lo que había hecho es inventar un nuevo modo de hablar.

Es una concepción ingenua la de que el mundo está dado de antemano, estructurado en cosas determinadas relacionadas entre sí de modo únivoco, y que el lenguaje viene después, reflejando con mayor o menor perfección la estructura de ese mundo.

Podemos aproximarnos al mundo con distintos lenguajes y habrá tantas estructuraciones distintas del mundo como lenguajes diferentes usemos para describirlo.

:34 La matemática se basa en una gran ilusión: en la ilusión de suponer acabado lo inacabable, ya terminado lo que no se puede terminar. Así se forman totalidades ilusorias, infinitos actuales que a su vez sirven de punto de partida a nuevos procesos semejantes, cada vez más alejados de la realidad.

Sin embargo, estas clases infinitas de las que nos hacemos la ilusión de que hablamos no existen en ningún sentido fuerte de la palabra existir. No están dadas en este mundo, ni pueden estarlo nunca. Postular otro mundo en que estén dadas es hacer mitología.

Las teorías matemáticas -en la medida en que hablen de algo- hablan unas de otras y encuentran sus modelos unas en otras. Y la posición central de la teoría de conjuntos entre ellas se debe a que todas las demás son en ella interpretables. En efecto, la relación semántica de satisfacibilidad de una teoría (algo lingüístico) en un modelo (algo extralingüístico) fácilmente puede ser substituida por la de interpretabilidad de una teoría en otra, adoptando la siguiente definición, adaptada de Tarski y Sxhonfield: [...]

:35 Esta noción de interpretabilidad de una teoría en otra cubre tanto la satisfacibilidad en modelos internos como en modelos externos.

Conclusión

A principios de siglo decía Russell que las matemáticas es una disciplina en la que no sabemos sobre lo que hablamos ni si lo que decimos es verdadero.

Hoy podríamos decir que al menos la teoría de conjuntos es una disciplina en la que sabemos que no hablamos acerca de nada y que lo que decimos no es verdadero ni falso.

¿Es esto una razón para dejar de hacer teoría de conjuntos, o, en general, para abandonar la matemática clásica, como piensan los intuicionistas?

Si el único sentido de una teoría o un lenguaje consistiese en decir verdades acerca del mundo, entonces, evidentemente, habría que abandonar la teoría de conjuntos.

Si, como dice Quine, el uso de variables cuantificadas en la teoría de conjuntos nos comprometiese "ontológicamente" a aceptar la existencia de esos fantasmas que son las totalidades infinitas actuales, y no creemos en ellos, también entonces tendríamos que abandonar la teoría de conjuntos, so pena de inconsecuencia.

Pero ni la única posible función de un lenguaje consiste en hablar acerca del mundo real ni los cuantificadores son amuletos mágicos cuyo uso nos compromete a vivir entre fantasmas.

En definitiva, la teoría de conjuntos es el foco central de ese universo lingüístico en expansión que constituyen las teorías matemáticas. La teoría de conjuntos ofrece un lenguaje común al que traducir los diversos lenguajes matemáticos y un arsenal de nociones y principios indispensables para el desarrollo de múltiples teorías formales.

Muchos de estos lenguajes, muchas de estas teorías -y otras que están aún por surgir- no han tenido hasta hoy aplicación alguna en la ciencia empírica. Pero otras teorías han encontrado interpretaciones empíricas fecundas y han servido para formular y desarrollar las ciencias de la naturaleza y de la sociedad.

En la medida en que el conocimiento y dominio del mundo real en que vivimos nos interese, nos interesarán también las teorías matemáticas que a través de su contribución a la ciencia empírica posibiliten ese conocimiento y ese dominio, y, por tanto, al menos indirectamente, nos interesará la teoría de conjuntos.

Extracts from Wilks 1976:

Richard Montague and Ludwig Wittgenstein represent views which are diametrically opposed on the key issue of formalization: of whether, and in what ways, formalisms to express the content of natural language (NL) should be constructed. "I shall argue, too, that the influence of Wittgenstein has been largely beneficial while that of Montague has been largely malign." The contrast between their views comes down to two issues:

  • Is there a hidden structure to NL?
    • Montague: Yes; clearly he belives there is a simple logic as the hidden structure of NL.
    • Wittgenstein: Yes, but not one to be revealed by simple techniques, like logic. (He refers to "deep grammatical structures", but it is not clear what exactly he ment for them).
  • Is NL itself and its use the final source of judgment or evidence?
    • Montague: No; his choice of examples, far from the concern of ordianry speakers ("Every man loves some woman") suggests that he is more concerd with logic and formal semantics than with NL "properly".
    • Wittgenstein: Yes. The different structures of "Every man loves some woman" in Predicate Calculus interest the logician, but the ordinary speaker rarely, if ever, sees that there is a "second interpretation" of the sentence.

Two very broad trends in the philosophy of language, as regards the role and importance of formalization can be distinguished:

  • One group of philosophers has been for it, and for as much of it as possible. (? Leibniz, Russell'25, Tarski'35, Carnap'37, Montague'70)
  • While the other group has been uncompromisigly against it. (? Popper'45, Wittgenstein'53, Hanson'58, Quine'60, Cohen'62, Dreyfus'72)

The "?" indicates a tentative list of philosophers

Leibniz, in the 17th century, believed that the formalism he proposed was the real structure of ordinary language (and, indeed, of the physical world), though without its awkward ambiguities, vaguenesses, and fallacies. In his most fantastic moments he envisaged the replacement of ordinary language by this "Universal Characteristic", to the general improvement of clarity and precision.

Whitehead and Russell's 1925 Principia Mathematica was a decisive contribution to the definition of propositional and general logic and the investigation of their properties. They were applied to the formalization of the notion of mathematical proof, but already Russel at least was setting out the ways in which this approach to logic was also, for him, a formalization of natural language. Like Leibniz, he wished to clear away what he thought of as the confusions of ordinary language. He was much concerned with the grammatical similarity of sentences like "Tigers are existent" and "Tigers are fierce" and how, in this view, this similarity had led philosophers into the error of thinking that tigers therefore had a property of existence as well as one of fierceness.

Wittgenstein was closely associated with Russell during the period between 1910 and 1920, and was developing what is now thought of as his early philosophy. This was set out in a curious early work called the Tractatus Logico-Philosophicus (1922) where he proposed what is now called the "picture theory of truth", according to which sentences in ordinary language signify or mean because their structures reflect or exhibit the same relation as that which holds between the things mentioned by the sentences. His theory of meaning at this point was more obscure than that of Russell, or Russell's predecessor, Frege.

Frege, in the late 19th century, had proposed a "dualist" theory of meaning in which each word signified in two ways (his famous example of "the Morning Star" and "the Evening Star"):

  1. One way (Bedeutung), referred to some entity, or REFERENT (i.e. Venus)
  2. The other (Sinn) referred to the "sense" of the word. The two phrases of the example mean something different in that they have different SENSES.

These meaning theories are referential, even when they talk of "sense". For, whatever the status or nature of the thing that is "the meaning", it is an entity that is somehow pointed at by the word. In Frege's case the word points in two different ways to two quite different sorts of entity, whereas in Wittgenstein more obscure "picture theory" there was only one kind of pointing.

[The truth of expressions can be established by truth tables and is demostrated as semantic. This is different from the method referred as proof theoretic (that comes from the sequential relation of structures in a proof)]. The word "semantic" is so troublesome that it cannot be introduced into the discussion without a word of warning, because on its ambiguities rests much of the difficulty of our whole subject.[...] It will be written as "L-semantic" when used in connection with formal logic.

Formal logic of the Twenties rests in two fundamental ideas, namely:

  1. Significant combination
  2. L-semantic demostration

Montague and Wittgenstein represent respectively a reaction to, and a extension of that logic. In the hands of Tarski'35 and Carnap'37 the above concepts took on new names and properties:

  1. Significant combination
  2. Logical syntax (replacing "L-semantic demonstration")
  3. Meta-language

Carnap in his "Logical Syntax of Language" developed the notion of ill formed expression (as "Mortality is Socrates"). Carnap distinguished: (Chomsky was many years later a student of Carnap)

  • Rules of formation: determined what were, and were not, well formed expressions in the logical language. Thus, in the Propositional Calculus, (P IMPLIES (OR Q)) was not, while (P OR Q) was a well formed expression.
  • Rules of transformation: operated on those well formed expressions that were also thrue so as to produce theorems in logic.

Carnap was a formalizer in the Leibnizian sense since, to him, these distinctios applied equally well to proper formalization of natural language.

The distinction of language and meta-language was due to Tarski who proposed it to solve an apparently intractable problem of logic that he had inherited from Russell, who had discovered certain logical paradoxes. Tarski's statement of the problem of the paradoxes was usually in terms of the example:

The sentence in this rectangle is false

where any attempts to assign a truth-value to the sentence led to trouble. Tarski thought that this problem would be solved if we only used "true" in a meta-language and never in an object -language. Thus "That John is happy is true" would be a sentence with level confusion, for the proper name would be the meta-language sentence "'John is happy' is true", where the sentence "John is happy" is mentioned in the meta-language, but not used in it.

[Truth theory of meaning: 212]

Tarski's fundamental achievement was a theory of truth and logical consequence for formalized languages (a L-semantic theory).

Montague extended it to sentences in natural languages: a system in which the truth conditions of complex statements could be seen to be composed from the truth conditions of simple ones. Making even stronger claims than Tarski's, because Tarski did not think that such a theory could be constructed for a natural, non-formal, language. Otherwise the object-language and the meta-language would be the same, and Tarski thought that was bound to lead to trouble.

Wilks'76 and others in linguistics and AI do not share the assumption that the stating of truth conditions is explaining the meaning of a sentence. Those in the Tarskian tradition are highly critical of much work in linguistics and AI because meanings are not stated in that way, and it is an interesting open question whether AI work on NL could be so stated in general. Wilks does not criticize formal analysis of NL as such. What is being opposed is pointless formalizations of NL [such as Montague's], and it is suggested that Wittgenstein is on of the few philosophers who can provide insights into what a fruitful approach to NL migh be.

Logical positivists also had a thesis about the dependence of meaning on truth but should not be confused with the "truth-condition" approach followed by Tarski. Their principle was called the Principle of Verification, and it said "The meaning of a statement is the procedures we would carry out to establish its truth of falsehood". However wrong this principle might be (and it is wrong), it was at least serious, in a way that the modern truth condition approach is not.

Blackburn'9? (for one in a large school of logicians) holds general principles concerning the applicability of Tarski-like theories to NL. The thesis that they share is that of the "truth theory of meaning": namely that the meaning of a sentence is determined by its truth conditions, in the Tarskian sense of that phrase. It would be quite possible to reject Montague's detailed "Semantics for NL" and still accept the general tenents of the school about meaning and truth, namely that truth conditions determine the meaning of the senteces of NL in just the way they can be said to do so for the sentences of logic and methematics.

Dreyfus, a full-time opponent of the very possibility of AI, has made much use of Wittgenstein's arguments against formal logic in his own arguments against AI. However, the general structure of this chapter is that AI has real philosophical importance because it has completely changed the old debate between formalists and anti-formalists. For to handle language "formally" on a computer it is in no way necessary to accept the tenets of Tarski, Montague or any other approach based on formal logic. On the contrary, the most fruitful approaches to understanding natural language are precisely those not subservient to a powerful logical or L-semantic theory.

AI has provided a sense to the notion of the precise manipulation of language that is not necessarily open to the attacks of the anti-formalists like Wittgenstein. And a large part of the credit for breaking the old formalist/anti-formalist opposition in a new way must go to Chomsky.

Chomsky's theory of grammar is a precise theory of language: it has the form of a logic, but not the content. Chomsky took the structure of proof-theoretic (what Carnap called transformation rules), namely the repeated derivation of structures from other structures by means of rules of inference, but he let its content be no more than what Carnap had meant by Formation Rules, namely the separation of the well-formed from the ill-formed. Thus, in transformational grammar the inference (transformational) rules were to apply to axioms and theorems, but to produce not new theorems but well-formed English sentences.

Thus, Chomsky has a precise system of handling language, but with no semantic definition of truth, and not even a syntactic, proof-theoretic one either. Thus, Chomsky's was the first concrete proposal to breach the wall between formal and anti-formal approaches.

AI has gone further, and indeed Chomsky's paradigm still shares many of the drawbacks to formal approaches: in particular the rigid "derivational" structure common to logical proofs and to Chomskyan transformational derivations.

[ Wittgenstein'58 ]

  1. Reference. Thesis: words do not in general have meaning in virtue of pointing at objects in the real world (or "conceptual objects" either). We could have a language based on the referential notion, but it would be a language more primitive than what we call NL.
  2. Mini languages and language games. Thesis: we can construct mini-languages obeying any rules we like, and we can think of them as games. The important question is whether these games are sufficiently like the "whole game" of NL.
  3. Family resemblances and boundaries. Thesis: the conventional notion of concept is wrong: namely the view that a concept relates in some way to the qualities or characteristics that all things falling under the concept have. There are not firm boundaries to concepts, nor are there to linguistic usage, nor to the application of linguistic rules.
  4. The linguistic whole and confronting the world. Thesis: a language is a whole and does not confront the world sentence by sentence for the testing of the truth or falsity of each individual part.
  5. Logicians have a false picture of how language is. Thesis: logicians think that language is like their favorite calculus, but they are quite wrong. Moreover, it is language itself and its use that is the standard for testing disputes that arise, not what logicians dictate.
  6. Understanding is not a feeling. Thesis: we have the idea that "understanding" something involves, or is associated with, a special feeling of being right. But the tests of our being right are quite different from the feeling.
  7. Application justifies our structure. Thesis: the significance of a representational structure cannot be divorced from the process of its application to actual language.
  8. Real world knowledge and forms of life. Thesis: language understanding is not independent of very general inductive truths about our human experience

Tipos de gramática

Las reglas se clasifican de mayor a menor complejidad o generalidad (based on Chomsky 1963):

From Shieber 1985:

THE UNDERLYING FRAMEWORK

  1. Grammar formalism describe language themselves:
    1. THE STRING SET, (describe the set of sentences we find in a language)
    2. THEIR SYNTAX, (structural properties of such sentences)
    3. THEIR SEMANTICS, (meaning of the sentences)
  2. Grammar formalism, the same as metalanguage. Metalanguage:
    • Provides a precise tool for the description of natural languages.
    • Delimits the class of passable natural languages.
    • Offers a computer-interpretable characterisation of natural languages.
    The metalanguage determines the following three parameters:
    • Linguistic felicity: (the degree to which descriptions of linguistic can be stated).
    • Expressiveness: (which class of analyses can be stated at all).
    • Computational effectiveness: (whether these exit computational devices for interpreting the grammars expressed in the formalism and, if they do exist, what computational limitations inhere in them.

Goals of grammar formalism:

  • Descriptive tool: (linguistic felicity and expressiveness are most important).
  • Universal linguistic delimitation: (linguistic felicity and lack of expressiveness are of foremost importance, i.e. expressive power must be constrained in a way that it is sufficient for natural languages, but nothing else beyond that).
  • Computer-interpretable linguistic characterisation: (all three criteria are vital).

SOME PARTICULAR DESIGN CHOICES:

Unification-based formalisms tend to make certain specific assumptions about what grammar formalism should do:

  • Surface-based: providing a direct characterisation of the actual surface order of string elements in a sentence.
  • Informational: associating with the strings and informational domain.
  • Inductive: defining the association of strings and informational elements recursively and merging the associated informational elements according to prescribed information-combining operations, and
  • Declarative: defining the association between strings and informational elements in terms of what associations are permissible, not how they are computed.

These informational elements (surface-based, informational, inductive, declarative) are characterizable as:

  • Complex-feature-based: as associations between features and values taken from some well-defined, possibly structured set.

This will provide a useful mathematical abstraction of the notion of informational element, which admits of several combinatorial operations currently used in linguistics.

COVERAGE OF FRAMEWORK

The list below shows the breadth and diversity of linguistic formalisms. Not all are based on unification but can all be modelled to a great extent in this way:

  • Categorial Grammar --> allows a functional application, it uses string concatenation to form constituents . Informational elements may be regarded as having a category -valued functor feature and an argument feature .An (S/NP)/NP category can be encoded by S/NP because we simplify the recursive functor feature NP. This variations are used in PATR-II, head grammar and HPSG. The categorial system of Ades and Steedman includes both functional applications and composition. Montague stated grammars as pairings of strings-combining and denotation-combining rules . The informational elements are comprised as a complex category feature whose value is the denotation of the expression.
  • GPSG --> a context-free base. Only uses concatenation to build up the surface string. Its informational domain is restricted feature/value system, involving simply and complex-valued features.
  • Head Grammars --> Extends GPSG introducing head-wrapping string > operations and removing the restrictions on the feature system that yield GPSG?s context freeness.
  • LFG --> A recursive feature/value system with certain specialised types of features and information.
  • FUG --> Though FUG?s patterns, concatenation is mandated as a constituent-forming operation.
  • DCG --> Terms are basic information-bearing structures. A degenerate case of a feature/value system in which the features correspond to argument positions.

Many paradigms of linguistics description can be encoded directly with generalised feature/value structures of this sort.

In summary, the methodological class outlined above involves:

  1. The association of strings with elements in a system of features and values.
  2. The inductive building of such associations by the simultaneous rule-based combination of substrings as well as of the associated informational elements.

PATR-II

The feature constraints associated with phrase structure rules in PC-PATR consist of a set of unification expressions. Each expression has three parts, in this order:
  1. a feature path, the first element of which is one of the symbols from the phrase structure rule
  2. an equal sign (=)
  3. either a simple value, or another feature path that also starts with a symbol from the phrase structure rule
As an example, consider the following PC-PATR rules:
Rule  S -> NP VP (SubCl)
        <NP head agr>  = <VP head agr>
        <NP head case> = NOM
        <S subj> = <NP>
        <S pred> = <VP>

Rule  NP -> {(Det) (AdjP) N (PrepP)} / PR
        <Det head number> = <N head number>
        <NP head> = <N head>
        <NP head> = <PR head>

Rule  Det -> DT / PR
        <PR head case> = GEN
        <Det head> = <DT head>
        <Det head> = <PR head>

Rule  VP -> VerbalP (NP / AdjP) (AdvP)
        <NP head case>   = ACC
        <NP head verbal> = -
        <VP head> = <VerbalP head>


Rule  VerbalP -> V
        <V head finite> = +
        <VerbalP head>  = <V head>

Rule  VerbalP -> AuxP V
        <V head finite> = -
        <VerbalP head>  = <AuxP head>

Rule  AuxP -> AUX (AuxP_1)
        <AuxP head> = <AUX head>

Rule  PrepP -> PP NP
        <NP head case> = ACC
        <PrepP head> = <PP head>

Rule  AdjP -> (AV) AJ (AdjP_1)

Rule  AdvP -> {AV / PrepP} (AdvP_1)

Rule  SubCl -> CJ S

PC-PATR output with feature structure:

1:
                S 
      __________|__________
     NP                  VP 
   ___|____      _________|__________
  Det     N   VerbalP  NP         AdvP 
   |     man     |      |           |
  DT             V     PR         PrepP 
  the           saw    us      _____|______
                              PP         NP 
                             with     ____|_____
                                     Det       N 
                                      |    telescope 
                                     DT 
                                      a 

1 parse found
[ cat:   S
  pred:    [ cat:   VP
             head:    [ agr:   $1[ 3sg:   + ]
                        finite:+
                        pos:   V
                        tense: PAST
                        vform: ED ] ]
  subj:    [ cat:   NP
             head:    [ agr:   $1
                        case:  NOM
                        number:SG
                        pos:   N
                        proper:-
                        verbal:- ] ] ]

Otras dificultades

Propiedades idiosincráticas vs. genéricas

Subcategorizaciones

estructuras argumentales de los predicados

beber(viviente,líquido,sin cuchara)

sorber(viviente,líquido,con labios)

poner(viviente,objeto,lugar)

Colocaciones

locura (cometer, hacer, *tomar, adoptar)

decisión (tomar, adoptar, *cometer, hacer, coger)

carácter(imprimir, crear, dar, desarrollar, ?forjar, *hacer)

personalidad(forjar, desarrollar, ?imprimir, crear, *hacer)

curso(dar, ofrecer, dictar, ?presentar, hacer)

exposición(hacer, presentar, ofrecer ?dar, dictar)

Confrontamos lo que la sintaxis puede hacer con categorías y subcategorías con la realidad de los textos:

Frases hechas, expresiones

No me refiero a cosas como "a mal tiempo, buena cara", sino a expresiones más flexibles y abiertas, pero que mantienen una

En lo fundamental no he cambiado, sigo siendo la misma persona.

¿Qué haces para mantenerte en forma?

Su {ausencia, recuerdo} ha dejado {un gran vacío, una profunda huella} en {en mi vida, en todos nosotros}

Las apariencias engañan. No te puedes fiar de las apariencias.

Cada uno sabe lo que {debe, le corresponde, le toca} hacer en cada momento.

Cada uno es dueño de hacer {con su vida, con su dinero} lo que {quiera, más le convenga}.

Hay cosas que le {enseñan, obligan} a uno a ser más {comprensivo, generoso} con los demás.

Es en {ella sobre la que, mi sobre quien} recae la {peor parte, responsabilidad}.