عجفت الغور
Language and the Brain (class)
fall 2021 classes
Class 1
- What do linguists do?
- Linguists as part of cognitive science
- There exists some sentences of setnences such that it allows for recursive composition
- So once we setup the idea that English cannot be a FSM, what is the data?
- Langauges that use case markings
- case markings allow it to mtaintain a more consistent structure modulo case marking
- chomsky is unconcerned with the statistical metadata that arises from a corpus
- “internalist” - grammars and structures within a single sentence
- “externalist” - set of sentences within a corpus
- for sentences to make sense, we assign a phoronlogical structure to pronounce sentences
- which then means a sentence makes sense iff:
- syntax (chomsky is here) is correct
- phoronological sufficent (this can be derived from statistics)
- “there exists a faculty of language” - but is there a single faculty of langauge
Class 2 - 9/13
- what is the neo-durkheimian relationship between langauge and thought?
- is language just a set of sentences?
- in the 70’s, they thought that sentences are a sequence of words or sounds
- what does this leave out? -> hidden syntax and semantics
- syntax cannot directly be seen
- semantics sometimes contain a truth value
- oftentimes this forms the \(\{form/sound, meaning\}\)
- chomsky identified this as a triplet of \(\{form, syntax, meaning\}\)
- sentences in english are infinite due to recursion (aka not finite structured)_
- chomsky also distinguished between “lists” and “conjunctions”
- lists as \(\mathbb{N}\)
- however, langauge (or at least english) is “uninterestingly” infinite
- as you push further out and generate more sentences, the sentences themselves become regularized
- for chomsky, grammar is the compression of language, and if you memorize words you can generate an infinite set of sentences
- also must make a distinction between “actual” and “potential”
- “actual” sentences are bound by the maximum speaker time or computation time
- “potential” sentences are the infinite set of all sentences generatable
- what are the grammatical rules for telling us if a word is an actual word?
- “moak” -> is this a word?
- we have grammars to say this is a potential word, but to reject a word, we actually do a semi-exhaustive search through our internal dictionaries to find out
- some words enter the vernacular after it is used
- “jabberwocky”, for example
- how do we distinguish words that are archaeic/low frequency from words that are incorrect?
- what was the theory of cognitive science before the 60’s?
- learning is the intentional modification of the environment in order to produce different future behaviors
- what is actually being “learned” when you acquire a langauge?
- a child does not do a blind word/object association (i.e. this is a ball)
- the faculty of langauge extends beyond this into “teaching” or “learning” a grammar and a generative capacity
- grammars have variance built in
Class 3 - 9/15
- Mendel’s laws - laws of inheritances that require abstract concepts like genes
- organized in such a way that gives rise to laws
- linguistic laws
- grimm’s law - stated in terms of independent sounds - phonemes
- werner’s law
- before these system laws, the idea that sound changes happened consistently/completely across langauges
- when chomsky says linguistics as a cognitive science
- this is not to reject the generalization finding tools
- dan everett - linguist in piraha
- no recursion in this language
- he proposes it comes from culture
- many things are understood about linguistics via numbers and color systems
- how to understand rules?
- through the enforcement of cultural roles
- how are the roles communicated and enforced?
- what happens if you violate it?
- studies about generalizablity of specific groups have never been totally innocent - largely driven by social policy
- does grimm’s law focus on the right things?
- are phonemes the correct unit of generalization?
- the persuasion of a generalization lies in the correlation of its properties
- future is tied up with the concept of “irrealis”
- modals, conditions
- future
- counterfactuals
- future tense does not seem to be a a fundalmental building blocks of language
- present tense is a mess because stateive and others
- grimm’s laws are simple, but the explanation could be complicated
- “laws” of languages often end up generalizing idiosyncratic effects to that languge
- anthropoligists suggest that systematic study of langauge by pieces that exhibit discontinuities reflect culture
- is culture downstream of langauge?
- compositionality
- meaning is composition
- labels can be generated from the hash of meaning
- label is the form/shape/demarker of the word?
Class 5 - 9/22
- Is phornology finite state?
- is it less powerful than finite state?
- why do we care where in the brain something happens
- it matters how the brain recongizes faces, but to do so, we must first learn where
- jakobson
- invariance
/t/
- phoneme, slashes denote an obstruct t
/t/
- goes into \([t^h]\) (as in “top”) and \([t]\) (as in “stop”, this doesn’t actually exist in mandarin)
- the other inviariance is when you have two different phonemmes, but how do we know what the nominative case is?
- Jakobson says it is meaning
- the features are semantic ones
- therefore, it was incorrect to think of this as the level of the phoneme, but rather we should look at it from the meaning
- internal/eternal
- prinicple of contrast
- example: “top” vs “(t^)hop” - your pronouncation of “t” does not change the menaining, therefore the asperation is not used to define meaning
- feature grounding - in terms of articulation, Jakobson thought we should use acoustics (wrong)
- paper fights with sausser
- meaning is distinct because it arises from the rest of the words
- cat vs dog, elephant, etc
- sausser says language is ungrounded and that it arises from scarcity
- it is because meaning can be generated through differences, like a hash function
- what about religion?
- what does the invariance come from?
- jakobson is still a version of structuralism
- are word embeddings structuralist?
- quine - early sentence embeddings -> the context is what matters, the company it keeps
- possiblies of derivational frequencies
- derevational family entropy says there is neural reponse
- why do generalizations happen in suffixes in english?
Class 6 - 9/27
- Standard linguistics metholdogy is the method of contrast principles
- contrast structures that are parallels
- examples:
Toler- |
Teach- |
Clash |
tolerant (adj) |
teach (/) |
Clash (noun) |
tolerance (noun) |
teach |
Clash (verb) |
tolerate (verb) |
|
|
tolerable (adj) |
|
|
- the (/) suffix is somethimes null, which contrasts with the existence sometimes
- sometimes the metholodgy tells you the suffix is null
- when you read toler[ate|ance|able] your brain generates the frequency distribution
- we have entropy/uncertainty over the continuities, called derivational family entropy
- derivational family entropy on mandarin?
- example:
- unflush(/) - no
- unflushable - yes
- why does (un) require a suffix?
- what does linguistics look like with implementation within the brain?
- context free - a sentence consists of a noun phrase and a verb phrase
- case markings - NP-> NPACC/v-
- when a noun phrase is accusative when its next to a verb, which are context sensitive
- A context free grammar cannot be duplicated. Lanugages like Yoruba which exhibits duplication, “buildhouse-buildhouse” -> builder
- Swiss german is the only clear examples of midly context-sensitive
- Context-free is recursive (no memory)
- for some languages, context sensitivy is used only for certain features, like reduplication in Yoruba
- why is this a problem?
- this points to an issue where human langauges have certain mathematical properties that map to the brain
- langauges are not evenly distributed between context free & context sensitive distribution
- maybe we can’t process certain context sensitivity
- finite -> computablably enumerable are rough categories
- pholonogy is theorized to be subregular
- morpohology and syntax are much more complicated
Class 7 - 9/29
- repetition implies context sensitivty
- does english have similar properties?
- np by np construction
- “dog after dog” is different than “little by little”
- little by little is not a noun phrase
- noun phrases are potentially infinite
- heinz and isardi claim that phronology is sub-regular
- constraints on grammar
- generally deal with locality
- grammar trees can show you how things block locality
- regular, natural, and CFG’s circle layering
- stabler
- for each grammar of a specific type (MC) there eixists an efficient parser
- stabler - subregularity of syntax gives you constraints on languages
- some sentences cannot be extracted from subject and adjunct
- called CED by J Huang
- for langauges can do subextraction from subject
- example: “what did pictures of fall on Fred” -> “what” must be extracted from subject
- merge ops
- even if you limit ones to subregular, the merge operations does whatever you want
- noun-phrase is equivilant to a noun-prep phrase, since a noun-phrase is a merge of noun-prep
- let us assume subregular grammar math is the math used for human language -> formal universal
- langauges have nouns and verbs as invariants -> Jakobson
- Stabler sees that we parse strings of words: word by word
- do you
- generate sentences on your own until you find a match?
- inefficient and non-deterministic
- since we do not wait until the end of a sentence before parsing it, model a sentence interpretation as a probability distribution that’s incremental
- lineraly porportional to the number of words
- using the finite set of words and finite set of rules: the cross product of both gives you the search space