The meeting of the Society from Computation in Linguistics is again being organized by Gaja Jarosz and Joe Pater, this year with help from Max Nelson and Brendan O’Connor of Computer Science. Jarosz is also one of the invited speakers, along with Mark Johnson of Macquarie University. This year will also feature an exciting experiment in conference design: there will be two simultaneous locations, with at least some of the talks being held jointly by videoconferencing. SCiL 2019 will again be co-located with the LSA, this time in NYC Jan. 3-6, and it will also take place in Paris over the same dates, at the Laboratoire de linguistique formelle at l’université Paris 7 Diderot. The abstract and paper deadline is Aug. 1: the full call can be found here: http://websites.umass.edu/scil/scil-2019/call-2019/.
Category Archives: Computational linguistics
Summer Dialect Research Project 2018
Four undergraduate students participated in the Summer Dialect Research Project (SDRP) at UMass in June hosted by the Center for the Study of African American Language (CSAAL). The Center, directed by Lisa Green, fosters and integrates research on language in African American speech communities and applications of that research in different realms. Three students, Christian Muxica, Alexander Santos, and Emily Smith, are enrolled at UMass and majoring in linguistics. Janiya Gilbert is a sophomore at North Carolina A&T University in the animal science program and has interests in language-related and social justice fields. The participants gained research experiences in areas in the study of African American English (AAE), a linguistic variety spoken by some African Americans. They worked on their skills in linguistics while also building broader analytical, argumentation, and collaboration skills. They completed group critical review projects and individual research projects that required analysis of data sets from AAE. During the three-week program, participants attended lecture/discussion sessions with UMass faculty, researchers, and graduate students, who covered topics in syntax, phonology, acquisition, psycholinguistics, and natural language processing. Professors Joe Pater and Kristine Yu worked with the participants in interactive sessions on topics related to sound patterns of AAE, such as the production of word final -t/-d and prosodic properties of yes-no questions in the variety. Professors Tom Roeper and Brian Dillon shared research on topics in acquisition and language processing and linked that research to data in AAE. For instance, Professor Dillon made the connection between research on garden path sentences and subject relative constructions in AAE. In other sessions, researchers discussed ways in which work in linguistics relates to other disciplines. Dr. Barbara Pearson, former research associate at UMass, demonstrated how research used to develop assessments in communication disorders for children who speak all varieties English, including AAE, has drawn on linguistics. Computer science graduate student Su Lin Blodgett presented her research on natural language processing and AAE and Twitter. One participant summed up his experience in the program in the following way: “I really enjoyed these three weeks and got a lot out of our work and hope to shape my senior year linguistic work around some of this research and our projects.”
Richard Futrell in CLC/Psycholing Workshop, Friday, April 27
All are welcome at the final CLC event this spring: Richard Futrell (MIT, BCS) will speak at Psycholinguistics Workshop on Friday, April 27th, 10am-11am, in ILC N400. Richard will also be available for individual meetings – please contact Chris Hammerly to set up an appointment. See below for abstract and title.
Memory and Locality in Natural Language
April 27th (Fri), 10am-11am, ILC N400 (Psycholing Workshop)
I explore the hypothesis that the universal properties of human languages can be explained in terms of efficient communication given fixed human information processing constraints. First, I show corpus evidence from 54 languages that word order in grammar and usage is shaped by working memory constraints in the form of dependency locality: a pressure for syntactically linked words to be close to one another in linear order. Next, I develop a new theory of human language processing cost, based on rational inference in a noisy channel, that unifies surprisal and memory effects and goes beyond dependency locality to a new principle of information locality: that words that predict each other should be close. I show corpus evidence for information locality. Finally, I show that the new processing model resolves a long-standing paradox in the psycholinguistic literature, structural forgetting, where the effects of memory on language processing appear to be language-dependent.
Music and language events this week
On Tuesday April 10th 3-4 pm in ILC N458, there will de a discussion of “Harmonic syntax of the 12-bar blues” by UMass Linguistics undergrad alum Jonah Katz. A link and abstract appear below.
On Friday April 13th 2:30 – 3:30 in ILC N400, Stefanie Acevedo (Yale) will present “Explaining expectation entropically: An empirical study of harmony in popular music” (abstract below).
At 3:30 Friday the 13th, David Temperley (Eastman School of Music) will present “A Model of Emotional Expression in Rock”.
All are welcome to all of these events. Please contact Joe Pater if you would like to meet with either Acevedo or Temperley while they are here.
==========
Jonah Katz (2017). Harmonic syntax of the 12-bar blues: a corpus study. Music Perception, 35(2), 165-192. Preprint (LingBuzz). Supplementary materials: data, statistical models, tree graphs, description of modeling.
Abstract. This paper describes the construction and analysis of a corpus of harmonic progressions from 12- bar blues forms included in the jazz repertoire collection The Real Book. A novel method of coding and analyzing such data is developed, using a notion of ‘possible harmonic change’ derived from the corpus and logit mixed-effects regression models describing the difference between actually occurring harmonic changes and possible but non-occurring ones in terms of various sets of theoretical constructs. Models using different sets of constructs are compared using the Bayesian Information Criterion, which assesses the accuracy and complexity of each model. The principal results are that: (1) transitional probabilities are better modeled using root-motion and chord- frequency information than they are using pairs of individual chords; (2) transitional probabilities are better described using a mixture model intermediate in complexity between a bigram and full trigram model; and (3) the difference between occurring and non-occurring chords is more efficiently modeled with a hierarchical, recursive context-free grammar than it is as a Markov chain. The results have implications for theories of harmony, composition, and cognition more generally.
Acevedo abstract: Given a preponderance of common _stock_ progressions in popular music, like the “Doo-Wop” (I-vi-IV-V) or the “Axis” (I-V-vi-IV) progressions, sequences of chords are often taken as a starting point for analysis. These chord sequences contextualize the sometimes _non-functional_ chord usage in popular music. While recent music-theoretical work uses computational methods to analyze harmonic probabilities in musical corpora and model their stylistic norms, it often focuses on analyzing lower-order probabilities such as single chord counts or chord-to-chord transitional probabilities. In this talk, I propose the use of information entropy, a measure of statistical uncertainty, as a way to segment harmonic progressions in a corpus of popular music (the McGill Billboard Corpus). The resultant harmonic segments are classified into prototypical chains based on functional categories that are determined by chord sequences as opposed to individual chords. The results and implications of the project are contextualized within recent research on popular music harmony and implicit learning of musical style.
Temperley abstract. In this talk, I present a framework for the analysis of emotional expression in rock music. The talk surveys some of the material in my new book The Musical Language of Rock (Oxford, 2018).
I begin with a two-dimensional model of emotion, well-established in music psychology, with valence (positive versus negative emotion) on one axis and energy (also known as arousal or activity) on the other. Valence is determined mainly by pitch collection (roughly, major versus minor, though there is more to it than that); energy depends on a variety of cues such as tempo, pitch register, loudness, and textural thickness. I then add a third dimension for complexity, or (in experiential terms) tension. Tension is affected by the density of events and also by their expectedness, with faster rhythms and low-probability events being higher in tension. Low-probability events can arise from such things as surprising harmonies, shifts outside of the currently established scale, irregular phrases, and extreme or unusual syncopations.
I then apply this model to the verse-chorus unit (VCU)—a formal section containing a verse and chorus; this is the core element of conventional rock form. We find consistent trajectories across the VCU in all three expressive dimensions—valence, energy, and tension. The chorus tends to be higher in energy than the verse; in terms of valence, many songs show a “sharp-ward” shift between verse and chorus, reflected not only in simple minor-to-major shifts but also in more subtle ways. With regard to tension, however, the peak tends to be in the middle of the VCU, either in the prechorus (if there is one) or in an extension of the verse. I present a number of examples, showing how the current model sheds light on both normative and exceptional cases.
Charlie O’Hara visiting this week
Charlie O’Hara, a fifth year phonology student at USC (http://dornsife.usc.edu/ohara) , will be visiting our department for the week. He will be presenting to a joint meeting of Gaja Jarosz and Joe Pater’s graduate phonology classes on Tuesday at 11:30 (all are welcome), and will be generally available for discussion, especially about his specialty, modeling soft typology with agent-based learning.

Music and Language CogSci Incubator Tues. April 3rd at 2:30 in ILC N458
The Music and Language CogSci Incubator (https://websites.umass.edu/cogsci/2018/03/10/music-and-language-cogsci-incubator-with-acevedo-and-temperley/) will begin on Tues. April 3rd at 2:30 in ILC N458 with a discussion of David Temperley’s The Musical Language of Rock. Participants are encouraged to bring questions and discussion points that don’t assume everyone has read the book (in other words, please come even if you haven’t even cracked its spine). The book turns out to assume a fair bit of music theory, so clarification questions are very much appropriate (and we might pick something else to discuss the next week).
Michael Becker in Sound Workshop / CLC Monday April 9th at 10 am
UMass alum and current Stony Brook faculty Michael Becker will be presenting on modeling Arabic plurals in Sound Workshop
Monday April 9th from 10am-11am in ILC N451. This is one of the meetings of the Computational Linguistics Community.
Music and Language CogSci Incubator with talks April 13
There is growing interest in the 5 Colleges in music cognition and its relation to language. To build on this, Mara Breen (Mount Holyoke Psychology and Education), Joe Pater (UMass Linguistics) and Christopher White (UMass Music and Dance) have organized the first “CogSci Incubator” event. On April 13th, Stephanie Acevedo of Yale University (https://stefanieacevedo.com) and David Temperley of the Eastman School of Music (http://davidtemperley.com) will present talks from 2:30 – 5 in N400 of the Integrative Learning Center. This will be followed by a reception at the Hangar Pub and Grill. In preparation for Temperley’s visit, there will also be two meetings, on April 3rd and 10th, from 3-4:15, to discuss his new book The Musical Language of Rock (linked below). Details on those meetings, including location, will be forthcoming in the CogSci newsletter (subscribe here).
https://global.oup.com/academic/product/the-musical-language-of-rock-9780190870522?cc=us&lang=en&#
These events are sponsored by the CogSci Initiative, the Department of Linguistics, and the Department of Music and Dance. If you are interested in “Incubating” another emerging UMass CogSci research area, please contact Joe Pater.
Prickett, Traylor and Pater in Sound Workshop and NLP Group
Brandon Prickett, Aaron Traylor and Joe Pater will present their ongoing work on learning reduplicative patterns using modern neural networks in Sound Workshop at 10-11 Monday Feb. 26th in ILC451 and in the NLP group CS 303 Wednesday Feb. 28th from 4-5 (the NLP meeting is part of this semester’s CLIC series). Brandon’s earlier negative results (http://scholarworks.umass.edu/ics_owplinguist/2/) thrilled Gary Marcus, but we suspect that his new positive ones will make Yann LeCun happier.
Spring 2018 Computational Linguistics Community (CLC) Events
See below for our exciting line-up of CLC events this semester. All welcome! Mark your calendars!
- Soroush Vosoughi (MIT) Data Science Seminar talk
- Feb 22nd at 4pm, CS 150/1
- COLING Paper Clinic
- February 28th (Wed), 4-5pm, CS 303 (NLP Reading Group)
- Yulia Tsvetkov (CMU Computer Science)
- March 1st (Thu), 12pm, CS 150/1 (MLFL)
- Yelena Mejova (QCRI) iSchool Seminar talk
- March 6th, at 4pm, CS 150/1
- Brian Dillon (UMass Linguistics) on “Syntactic Frequency Effects in Recognition Memory”
- March
9th30th (Fri), 12:20-1:20, ILC N451 (Experimental Lab)
- March
- Michael Becker (Stony Brook Linguistics) on Modeling Arabic Plurals
- April 9th (Mon), 10am-11am, ILC N451 (Sound Workshop)
- Richard Futrell (MIT BCS) title TBA
- April 27th (Fri), 10am-11am, ILC N400 (Psycholing Workshop)




