Author Archives: Joseph Pater

Musso in Data Science Tuesday Feb. 13 at 4

What: DS Seminar
Date: February 13, 2018
Time: 4:00 – 5:00 P.M.
Location: Computer Science Building, Room 151

Cameron Musco
The Power of Simple Algorithms: From Data Science to Biological Systems 

Abstract:  In recent years, very simple randomized methods, such as stochastic iteration, sampling, and hashing, have become dominant computational tools in large-scale machine learning and data science. In this talk, he will discuss his efforts to understand and harness the remarkable power of these methods.

In particular, he will describe his research on developing simple, but principled, sampling methods for learning, estimation, and optimization. He will present a new class of iterative sampling algorithms, which give state-of-the-art theoretical and empirical performance for regression problems, low-rank matrix approximation, and kernel methods. In many cases, the computational improvement offered by these algorithms is quite surprising. For example, our methods can be used to compute a near-optimal low-rank approximation to any positive semidefinite matrix in sublinear time.

In addition to their power in algorithm design, he will discuss his efforts to understand simple, randomized methods through a different lens: by studying how complex behavior emerges from low-level randomized interactions in biological systems. He will demonstrate how many of the same mathematical tools used to study algorithms in data science can be applied to these systems. As an example, he will highlight his research on noisy estimation and decision making in social insect colonies.

Bio:  Cameron Musco is a fifth year Ph.D. student (graduating spring 2018) in the Theory of Computation Group at MIT.
He is advised by Nancy Lynch and supported by an NSF Graduate Fellowship. He studies algorithms, focusing on applications in data science and machine learning. He often works on randomized methods and algorithms that adapt to streaming and distributed computation. He is also interested in understanding randomized computation and algorithmic robustness by studying computational processes in biological systems.

Before MIT, he studied Computer Science and Applied Mathematics at Yale University and worked as a software developer on the Data Team at Redfin.

Faculty Host: Arya Mazumdar
A reception for attendees will be held at 3:30 P.M. in CS 150.(The back of the presentation room.)

Fornaciai and Park in Cognitive Brown Bag Weds. Feb. 14th at noon

Cognitive Brown Bag, 2/14/18, 12:00-1:15, Tobin 521B

Michele Fornaciai & Joonkoo Park (PBS)

Serial dependence in numerosity perception

Attractive serial dependence represents an adaptive change in the representation of sensory information, whereby current stimuli appear more similar to previous ones. Here, we characterize the behavioral and neural signatures of serial dependence in numerosity perception, demonstrating that the perceived numerosity of dot-array stimuli in different numerical ranges is biased by a preceding irrelevant stimulus (“inducer”) in an attractive way. Using electroencephalogram and a passive-viewing paradigm, we show that a neural signature of attractive serial dependence emerges even in the absence of an explicit task early in the visual stream, suggesting that serial dependence has a clear perceptual origin independently from a decision process. With a series of follow-up experiments, we further characterize serial dependence in visual number perception. First, we show that this effect has a weak spatial specificity and a relatively broad tuning for numerosity, and that it has a clear cortical origin (rather than subcortical). Second, we show that the attractive effect is strongly modulated by attention, suggesting the involvement of higher level modulatory influences. Our results collectively suggest that serial dependence results from a cortical neural computation starting from an early level of perceptual processing, possibly subserving perceptual stability and influencing downstream cognitive stages. However, these findings also suggest that the integration of past and present stimuli is in turn modulated by higher-level processes, and potentially amplified at later processing stages.

Hirsch in Linguistics Friday Feb. 9 at 3:30

“Towards an Inflexible Semantics for Cross-Categorial Operators”
Aron Hirsch (McGill)
It is generally thought that the semantics is flexible in such a way that one operator can compose with different kinds of arguments (e.g. Montague 1973, Partee & Rooth 1983, Rooth 1985, Keenan & Faltz 1987). This flexibility seems to be required for operators such as “and”, which show a broad distribution. In (1), “and” appears to compose with truth-values in (a), quantifiers in (b), and relations in (c).
(1a) [TP John saw every student] and [TP Mary saw every professor]
(1b) John saw [DP every student] and [DP every professor].
(1c) John [V hugged] and [V pet] the dog.
In this talk, however, I argue that the semantics does not allow for this flexibility, and that “and” has a uniform meaning across its distribution: and operates on truth-values, parallel to the ^-connective of propositional logic (e.g. Schein 2017). The central case study will be data such as (1b), where and occurs between object quantifiers. First, I will argue that (1b) has a “Conjunction Reduction” parse as underlying vP conjunction. Since vPs denote truth-values, and can then compose as ^. Second, I will present data which I argue are best understood if “and” does not have the additional option to operate directly on the quantifiers. “And” is interpreted as ^, while surface cross-categoriality is created by the syntax. I will show that this view extends to another cross-categorial operator (“only”), and receives support from operators which could in principle have a cross-categorial distribution but don’t (e.g. “yesterday”).

Hariharan in MLFL Thursday at 11:45

who: Bharath Hariharan, Cornell University
when: 11:45 A.M. 1:15 P.M., Thursday, February 8th
where: Computer Science Building Rm 150

Visual recognition beyond large labeled training sets

Abstract: The performance of recognition systems has grown by leaps and bounds these last 5 years. However, modern recognition systems still require thousands of examples per class to train. Furthermore, expanding the capabilities of the system by introducing new visual concepts again requires collecting thousands of examples for the new concept. In contrast, humans are known to quickly learn new visual concepts from as few as 1 example, and indeed require very little labeled data to build their powerful visual systems from scratch. The requirement for large training sets also makes it infeasible to use current machine vision systems for rare or hard-to-annotate visual concepts or new imaging modalities.

I will talk about some of our work on reducing this need for large labeled training sets. I will describe novel loss functions for training convolutional network-based feature representations so that new concepts can be learned from a few examples, and ways of hallucinating additional examples for data-starved classes. I will also discuss our attempt to learn feature representations without any labeled data by leveraging motion-based grouping cues. I will end with a discussion of where we are and thoughts on the way forward.

Bio:  Bharath Hariharan is an assistant professor at Cornell. Before joining Cornell, he spent two years as a postdoc in Facebook AI Research after obtaining a Ph.D. from UC Berkeley with Jitendra Malik. At Berkeley, he was the recipient of the Microsoft Research fellowship. His interests are in all things visual recognition. Of late, he has become bothered by the reliance on massive labeled datasets and the scalability of such datasets to harder problems such as visual reasoning. His current work is on building recognition systems that learn with less data and/or output a much deeper understanding of images.

Dahlstrom-Hakki in Cognitive Bag Lunch Weds. Feb. 7

Cognitive Brown Bag, 2/7/18, 12:00-1:15.

NOTE:  This week only, the talk will be in Bartlett 119

Ibrahim Dahlstrom-Hakki (Landmark College)

Teaching Students with Disabilities Online: Language-Based Challenges and Cognitive Access

 

Many students with Learning Disabilities (LD), Attention Deficit Hyperactivity Disorder (ADHD), and Autism Spectrum Disorder (ASD) struggle in online learning environments. Online courses tend to place high demands on their language processing and executive function skills. In this NSF funded study (DRL-1420198), we look at some of the barriers facing students with disabilities learning statistics concepts through online discussions. We report on the impact of a Social Presence manipulation on their performance and some of the language-based difficulties involved in assessing their knowledge.

Neubig on “What Can Neural Networks Teach us about Language?” Thurs. Feb 1 at 11:45

Machine Learning and Friends

who: Graham Neubig (CMU)
when: 11:45a – 1:15p, Feb 1st
where: Computer Science Building Rm 150
food: Antonios Pizza
generous sponsor:  ORACLE LABS 

What Can Neural Networks Teach us about Language?

Abstract:

Neural networks have led to large improvements in the accuracy of natural language processing systems. These have mainly been based on supervised learning: we create linguistic annotations for a large amount of training data, and train networks to faithfully reproduce these annotations. But what if we didn’t tell the neural net about explicitly, but instead *asked it what it thought* about language
without injecting our prior biases? Would the neural network be able
to learn from large amounts of data and confirm or discredit our
existing linguistic hypotheses? Would we be able to learn linguistic
information from lower-resourced languages where this information has not been annotated? In this talk, I will discuss methods for
unsupervised learning of linguistic information using neural networks
that attempt to answer these questions. I will also explain briefly
about automatic mini-batching, a computational method (implemented in the DyNet neural network toolkit), which greatly speeds large-scale
experiments with complicated network structures needed for this type
of unsupervised learning.

Bio:

Graham Neubig is an assistant professor at the Language Technologies Institute of Carnegie Mellon University. His work focuses on natural language processing, specifically multi-lingual models that work in many different languages, and natural language interfaces that allow humans to communicate with computers in their own language. Much of this work relies on machine learning to create these systems from data, and he is also active in developing methods and algorithms for machine learning over natural language data. He publishes regularly in the top venues in natural language processing, machine learning, and speech, and his work occasionally wins awards such as best papers at EMNLP, EACL, and WNMT. He is also active in developing open-source software and is the main developer of the DyNet neural network toolkit.

Sadil in Cognitive Brown Bag Weds. Jan. 31 at noon

Cognitive Brown Bag, 1/31/18, 12:00-1:15, Tobin 521B.

Patrick Sadil   (PBS)

Pattern-completion of intra-item associations

Some neurocomputational theories of episodic memory have cast recollection as a pattern-completion process during which information is retrieved. In contrast to the pattern-matchingprocess of familiarity, recollection is thought to be most useful in tasks that require retrieval across episodic-like, arbitrary associations (e.g., retrieval of an item cued by a context). However, it is unclear whether pattern-completion occurs for the retrieval of purely visual, intra-item associations that exist at a lower level of representation in the brain’s processing hierarchy. In this experiment, we asked whether recollection, operationalized as the retrieval of previously studied material that is not provided as part of the cue, can occur for the features of single, everyday objects. Participants studied images of everyday objects. To manipulate the degree to which participants could encode arbitrary associations (e.g., the object with the experimental context) some objects were masked with continuous flash suppression. At test, participants were given a cued-recall test in which a small part of the studied object was presented and the task was to identify the whole object. Immediately after the cued-recall, participants engaged in a target detection go/no-go task where the inverse of the object part (i.e., the entire object save for the part used as a cue) might or might not emerge out of visual noise. The key question was whether the response time distribution was shifted earlier on go trials for objects that were studied under continuous flash suppression (as compared to a not-studied baseline) , in particular when the name of the object was not recalled immediately prior. Participants detected the non-recalled CFS-studied objects appearing from the noise more rapidly, and the speed-up corresponded to a shift in the response time distribution. This suggests that participants had a ‘head-start’ on the go trials, despite an inability to identify the cued object. This head-start suggests that a visual pattern-completion process occurred during cued-recall even when participants only encoded – and so could only retrieve – intra-item associations.

Sue Carey, Friday April 20th at 3:30

The Five College Cognitive Science Speaker Series and the UMass Initiative in Cognitive Science is pleased to announce that Sue Carey, Department of Psychology, Harvard, will speak at UMass on Friday, April 20 at 3:30pm in ILC S131. The title and abstract are below.

—————–

Do Non-Linguistic Creatures have a Fodorian (Logic-Like/Language-Like) Language of Thought?

The adult human conceptual repertoire is a unique phenomenon earth. Human adults build hierarchical representations on the fly, distinguishing “Molecules are made of tiny atoms” (True) from “Atoms are made of tiny molecules” (False). It is unknown whether non-linguistic creatures are capable of representing structured propositions in terms of hierarchical structures formulated over abstract variables, assigning truth values to those propositions, or are capable of abstract relational thought. How abstract knowledge and abstract combinatorial thought is acquired, over both evolutionary and ontogenetic time scales, is one of the outstanding scientific mysteries in the cognitive sciences, and has been debated in the philosophical literature at least since Descartes. Many philosophers, from Descartes through Davidson, have argued that abstract combinatorial thought is absent in creatures who lack natural language; others, such as Fodor, argue that such thought must be widely available to non-linguistic creatures, including human babies and animals at least throughout the vertebrates. A priori arguments will not get us far in settling this issue, which requires both theoretical analysis and empirical work. Theoretically, those who think there is a joint in nature between the kinds of representations that underlie perception and action, on the one hand, and abstract combinatorial thought, on the other, owe us an analysis of the essential differences between the representations and computations involved in each. Empirically, then, we must develop methods that could yield data that bear on the question of whether non-human animals or human infants have representations/computations on the abstract combinatorial thought side of the putative joint in nature. I will illustrate progress on both the theoretical and empirical fronts through two case studies: logical connectives and abstract relations.

Hammerly in Cognitive Brown Bag Weds. Feb. 24 at noon

Cognitive Brown Bag, 1/24/18, 12:00-1:15, Tobin 521B.

Speaker:  Chris Hammerly, UMass Linguistics

TitleResponse bias modulates illusions of (un)grammaticality in subject-verb agreement: A diffusion model account
 
Abstract: Over the past decade, the comprehension of subject-verb agreement has been argued to best be modeled via the dynamics of direct-access, cue-based, memory retrieval. The central area of study has focused on when and how illusions in comprehension occur. For example, readers who are asked to judge sentences often fail to notice the ungrammaticality of sentences like The key to the cabinets are on the table. In sentences of this form, the correct form of the verb should be singular (i.e. is), as the true controller of agreement (key) is singular. However, the presence of a plural distractor noun (cabinets), which correctly covaries with the plural form of the verb (are), results in these sentences being erroneously judged as grammatical. Crucially, it has been shown that there is an asymmetry such that readers systematically judge ungrammatical sentences to be grammatical (i.e. there is an illusion of grammaticality), but not vice versa (there is no illusion of ungrammaticality). This asymmetry has been the empirical keystone in supporting memory-based accounts. In three experiments conducted jointly with Brian Dillon (UMass Linguistics) and Adrian Staub (UMass Psychology), I show that this asymmetry only appears when participants are biased towards making grammatical responses. When response bias is neutralized, both illusions of grammaticality and ungrammaticality appear. This finding calls into question the claim that the dynamics of memory retrieval lead to grammatical illusions in agreement comprehension. I instead model the results using a drift diffusion process, implemented using fast-dm (Voss & Voss, 2007), where grammatical illusions are argued to be a function of the rate at which evidence about the grammaticality of the sentence accumulates during the judgement decision process, and the asymmetry is a function of pre-decision response bias.