Loading Events

Past Events › CSS Seminar

Events Search and Views Navigation

Event Views Navigation

June 2019

Summer Institute in Computational Social Science Panel Presentation

June 21, 2019 @ 2:00 pm - 6:00 pm PDT
Luskin Conference Center Laureate Room

Summer Institute in Computational Social Science Panel Presentation

Friday June 21, 2019 2:00pm – 5:00pm
Reception 5:00pm – 6:00pm
Luskin Conference Center Laureate Room
• 2:00pm – 3:15pm Digital Demography
Prof. Dennis Feehan, UC Berkeley and Prof. Ka-Yuet Liu, UCLA
• 3:30pm – 4:45pm Computational Causal Inference
Prof. Judea Pearl, UCLA and Prof. Sam Pimentel, UC Berkeley

Find out more »

May 2019

Workshop: Getting All Your Research Computing Tools for Summer and Beyond – Hardware and Software

May 22, 2019 @ 12:00 pm - 1:30 pm PDT
4240 Public Affairs Bldg

Title: Getting All Your Research Computing Tools for Summer and Beyond - Hardware and Software Location: May 22, 2019 @ 12:00-1:30 PM 4240 Public Affairs Building CCPR Seminar Room Instructors: Matt Lahmann & Mike Tzen Content: We’ll get CCPR researchers all the computing tools for a productive summer of data science exploration. We'll get you started on computing hardware: personal, terminal, and cluster. We'll get you started on software: R, stata, python, etc. To get the most out of this…

Find out more »

Brandon Stewart, Princeton University

May 8, 2019 @ 12:00 pm - 1:30 pm PDT
CCPR Seminar Room, 4240 Public Affairs Building
Los Angeles, CA 90095 United States

How to Make Causal Inferences Using Texts

Texts are increasingly used to make causal inferences: either with the document serving as the treatment or the outcome. We introduce a new conceptual framework to understand all text-based causal inferences, demonstrate fundamental problems that arise when using manual or computational approaches applied to text for causal inference, and provide solutions to the problems we raise.  We demonstrate that all text-based causal inferences depend upon a latent representation of the text and we provide a framework to learn the latent representation.  Estimating this latent representation, however, creates new risks: we may unintentionally create a dependency across observations or create opportunities to fish for large effects.  To address these risks, we introduce a train/test split framework and apply it to estimate causal effects from an experiment on immigration attitudes and a study on bureaucratic responsiveness.  Our work provides a rigorous foundation for text-based causal inferences, connecting two previously disparate literatures. (Joint Work with Egami, Fong, Grimmer and Roberts)

Co-sponsored with the Center for Social Statistics

Find out more »

Susan Athey, Stanford University

May 1, 2019 @ 12:00 pm - 1:30 pm PDT
CCPR Seminar Room, 4240 Public Affairs Building
Los Angeles, CA 90095 United States

Estimating Heterogeneous Treatment Effects and Optimal Treatment Assignment Policies

This talk will review recently developed methods for estimating conditional average treatment effects and optimal treatment assignment policies in experimental and observational studies, including settings with unconfoundedness or instrumental variables. Multi-armed bandits for learning treatment assignment policies will also be considered.

Co-sponsored with the Center for Social Statistics

Find out more »

March 2019

Lan Liu, University of Minnesota at Twin Cities

March 5, 2019 @ 2:00 pm - 3:30 pm PST

Lan Liu, University of Minnesota at Twin Cities "Parsimonious Regressions for Repeated Measure Analysis"  Abstract: Longitudinal data with repeated measures frequently arises in various disciplines. The standard methods typically impose a mean outcome model as a function of individual features, time and their interactions. However, the validity of the estimators relies on the correct specifications of the time dependency. The envelope method is recently proposed as a sufficient dimension reduction (SDR) method in multivariate regressions. In this paper, we demonstrate…

Find out more »

February 2019

Adeline Lo, Princeton University

February 26, 2019 @ 2:00 pm - 3:00 pm PST
1434A Physics and Astronomy Building

Covariate screening in high dimensional data: applications to forecasting and text data

High dimensional (HD) data, where the number of covariates and/or meaningful covariate interactions might exceed the number of observations, is increasing used in prediction in the social sciences. An important question for the researcher is how to select the most predictive covariates among all the available covariates. Common covariate selection approaches use ad hoc rules to remove noise covariates, or select covariates through the criterion of statistical significance or by using machine learning techniques. These can suffer from lack of objectivity, choosing some but not all predictive covariates, and failing reasonable standards of consistency that are expected to hold in most high-dimensional social science data. The literature is scarce in statistics that can be used to directly evaluate covariate predictivity. We address these issues by proposing a variable screening step prior to traditional statistical modeling, in which we screen covariates for their predictivity. We propose the influence (I) statistic to evaluate covariates in the screening stage, showing that the statistic is directly related to predictivity and can help screen out noisy covariates and discover meaningful covariate interactions. We illustrate how our screening approach can removing noisy phrases from U.S. Congressional speeches and rank important ones to measure partisanship. We also show improvements to out-of-sample forecasting in a state failure application. Our approach is applicable via an open-source software package.

Find out more »

January 2019

Kosuke Imai, Harvard University

January 23, 2019 @ 12:00 pm - 1:30 pm PST
CCPR Seminar Room, 4240 Public Affairs Building
Los Angeles, CA 90095 United States

Matching Methods for Causal Inference with Time-Series Cross-Section Data

Matching methods aim to improve the validity of causal inference in observational studies by reducing model dependence and offering intuitive diagnostics. While they have become a part of standard tool kit for empirical researchers across disciplines, matching methods are rarely used when analyzing time-series cross-section (TSCS) data, which consist of a relatively large number of repeated measurements on the same units.

We develop a methodological framework that enables the application of matching methods to TSCS data. In the proposed approach, we first match each treated observation with control observations from other units in the same time period that have an identical treatment history up to the pre-specified number of lags. We use standard matching and weighting methods to further refine this matched set so that the treated observation has outcome and covariate histories similar to those of its matched control observations. Assessing the quality of matches is done by examining covariate balance. After the refinement, we estimate both short-term and long-term average treatment effects using the difference-in-differences estimator, accounting for a time trend. We also show that the proposed matching estimator can be written as a weighted linear regression estimator with unit and time fixed effects, providing model-based standard errors. We illustrate the proposed methodology by estimating the causal effects of democracy on economic growth, as well as the impact of inter-state war on inheritance tax. The open-source software is available for implementing the proposed matching methods.

Co-sponsored with the Political Science Department, Statistics Department and the Center for Social Statistics

Find out more »

Rocio Titiunik, University of Michigan

January 16, 2019 @ 12:00 pm - 1:30 pm PST
CCPR Seminar Room, 4240 Public Affairs Building
Los Angeles, CA 90095 United States

Internal vs. external validity in studies with incomplete populations

Researchers working with administrative data rarely have access to the entire universe of units they need to estimate effects and make statistical inferences. Examples are varied and come from different disciplines. In social program evaluation, it is common to have data on all households who received the program, but only partial information on the universe of households who applied or could have applied for the program. In studies of voter turnout, information on the total number of citizens who voted is usually complete, but data on the total number of voting-eligible citizens is unavailable at low levels of aggregation. In criminology, information on arrests by race is available, but the overall population that could have potentially been arrested is typically unavailable. And in studies of drug overdose deaths, we lack complete information about the full population of drug users.

In all these cases, a reasonable strategy is to study treatment effects and descriptive statistics using the information that is available. This strategy may lack the generality of a full-population study, but may nonetheless yield valuable information for the included units if it has sufficient internal validity. However, the distinction between internal and external validity is complex when the subpopulation of units for which information is available is not defined according to a reproducible criterion and/or when this subpopulation itself is defined by the treatment of interest. When this happens, a useful approach is to consider the full range of conclusions that would be obtained under different possible scenarios regarding the missing information. I discuss a general strategy based on partial identification ideas that may be helpful to assess sensitivity of the partial-population study under weak (non-parametric) assumptions, when information about the outcome variable is known with certainty for a subset of the units. I discuss extensions such as the inclusion of covariates in the estimation model and different strategies for statistical inference.

Co-sponsored with the Political Science Department, Statistics Department and the Center for Social Statistics 

Find out more »

October 2018

Erin Hartman, University of California Los Angeles

October 17, 2018 @ 12:00 pm - 1:30 pm PDT
CCPR Seminar Room, 4240 Public Affairs Building
Los Angeles, CA 90095 United States

Covariate Selection for Generalizing Experimental Results

Researchers are often interested in generalizing the average treatment effect (ATE) estimated in a randomized experiment to non-experimental target populations. Researchers can estimate the population ATE without bias if they adjust for a set of variables affecting both selection into the experiment and treatment heterogeneity.Although this separating set has simple mathematical representation, it is often unclear how to select this set in applied contexts. In this paper, we propose a data-driven method to estimate a separating set. Our approach has two advantages. First, our algorithm relies only on the experimental data. As long as researchers can collect a rich set of covariates on experimental samples, the proposed method can inform which variables they should adjust for. Second, we can incorporate researcher-specific data constraints. When researchers know certain variables are unmeasurable in the target population, our method can select a separating set subject to such constraints, if one is feasible. We validate our proposed method using simulations, including naturalistic simulations based on real-world data.

Co-Sponsored with The Center for Social Statistics

Find out more »

March 2018

Jake Bowers, University of Illinois at Urbana-Champaign

March 13, 2018 @ 2:00 pm - 3:15 pm PDT
Franz Hall 2258A

"Rules of Engagement in Evidence-Informed Policy: Practices and Norms of Statistical Science in Government"

Abstract: Collaboration between statistical scientists (data scientists, behavioral and social scientists, statisticians) and policy makers promises to improve government and the lives of the public. And the data and design challenges arising from governments offer academics new chances to improve our understanding of both extant methods and behavioral and social science theory. However, the practices that ensure the integrity of statistical work in the academy — such as transparent sharing of data and code — do not translate neatly or directly into work with governmental data and for policy ends. This paper proposes a set of practices and norms that academics and practitioners can agree on before launching a partnership so that science can advance and the public can be protected while policy can be improved. This work is at an early stage. The aim is a checklist or statement of principles or memo of understanding that can be a template for the wide variety of ways that statistical scientists collaborate with governmental actors.

Find out more »

February 2018

Yu Xie, Princeton

February 21, 2018 @ 12:00 pm - 1:30 pm PST
CCPR Seminar Room, 4240 Public Affairs Building
Los Angeles, CA 90095 United States

"Heterogeneous Causal Effects: A Propensity Score Approach "

Abstract: Heterogeneity is ubiquitous in social science. Individuals differ not only in background characteristics, but also in how they respond to a particular treatment. In this presentation, Yu Xie argues that a useful approach to studying heterogeneous causal effects is through the use of the propensity score. He demonstrates the use of the propensity score approach in three scenarios: when ignorability is true, when treatment is randomly assigned, and when ignorability is not true but there are valid instrumental variables.

Find out more »

June 2017

James Robins, Harvard University

June 9, 2017 @ 12:00 pm - 1:00 pm PDT
Room 33-105 CHS Building, 650 Charles E Young Drive South
Los Angeles, CA 90095 United States

The UCLA Departments of Epidemiology, Biostatistics, Statistics and the Center for Social Statistics presents: Causal Methods in Epidemiology: Where has it got us and what can we expect in the future? The principal focus of Dr. Robins’ research has been the development of analytic methods appropriate for drawing causal inferences from complex observational and randomized studies with time-varying exposures or treatments. The new methods are to a large extent based on the estimation of the parameters of a new class…

Find out more »

May 2017

Shahryar Minhas, Duke University

May 24, 2017 @ 12:00 pm - 1:30 pm PDT
CCPR Seminar Room, 4240 Public Affairs Building
Los Angeles, CA 90095 United States

The Center for Social Statistics Presents: Predicting the Evolution of Intrastate Conflict: Evidence from Nigeria url: http://css.stat.ucla.edu/event/shahryar-minhas/ The endogenous nature of civil conflict has limited scholars' abilities to draw clear inferences about the drivers of conflict evolution. We argue that three primary features characterize the complexity of intrastate conflict: (1) the interdependent relationships of conflict between actors; (2) the impact of armed groups on violence as they enter or exit the conflict network; and (3) the ability of civilians to influence…

Find out more »

May 2016

Ilan H. Meyer & Mark S. Handcock, UCLA

May 27, 2016 @ 12:00 pm - 2:00 pm PDT
CCPR Seminar Room, 4240 Public Affairs Building
Los Angeles, CA 90095 United States

"Innovative Sampling Approaches for Hard to Reach Populations: Design of a National Probability Study of Lesbians, Gay Men, Bisexuals, and Transgender Peoples and Network Sampling of Hard to Reach Populations"


Speakers:

Ilan H. Meyer, Williams Distinguished Senior Scholar for Public Policy at the Williams Institute

Mark S. Handcock, Professor of Statistics at UCLA and Director of the Center for Social Statistics


Description:


Come for the exciting seminar then stay for the free lunch and discussion. A seminar led by Ilan H. Meyer followed immediately by a Brown Bag Lunch led by Mark S. Handcock.

Dr. Meyer is Principal Investigator of the Generations and TransPop Surveys. Generations is a survey of a nationally representative sample of 3 generations of lesbians, gay men, and bisexuals. TransPop is the first national probability sample survey of transgender individuals in the United States. Both studies attempt to obtain large nationally representative samples of hard to reach populations. Dr. Meyer will review sampling issues with LGBT populations and speak on the importance of measuring population health of LGBTs and the underlying aspects in designing a national probability survey.

From a contrasting perspective, the field of Survey Methodology is facing many challenges. The general trend of declining response rates is making it harder for survey researchers to reach their intended population of interest using classical survey sampling methods.

In the followup Brown Bag Lunch, led by Mark S. Handcock, participants will discuss statistical challenges and approaches to sampling hard to reach populations. Transgenders, for example, are a rare and stigmatized population. If the transgender community exhibits networked social behavior, then network sampling methods may be useful approaches that compliment classical survey methods.
Participants are encouraged to speak on ideas of statistical methods for surveys.

Find out more »

March 2016

Rick Dale, University of California, Merced

March 31, 2016 @ 2:30 pm - 4:00 pm PDT
314 Royce Hall, 340 Royce Dr
los angeles, CA 90095 United States
+ Google Map

"Quantifying the dynamics of multimodal communication with multimodal data."

*Presented by the Center for Social Statistics

Abstract: Human communication is built upon an array of signals, from body movement to word selection. The sciences of language and communication tend to study these signals individually. However, natural human communication uses all these signals together simultaneously, and in complex social systems of various sizes. It is an open puzzle to uncover how this multimodal communication is structured in time and organized at different scales. Such a puzzle includes analysis of two-person interactions. It also involves an understanding of much larger systems, such as communication over social media at an unprecedentedly massive scale.

Collaborators and I have explored communication across both of these scales, and I will describe examples in the domain of conflict. For example, we've studied conflict communication in two-person interactions using video analysis of body and voice dynamics. At the broader scale, we have also used large-scale social media behavior (Twitter) during a massively shared experience of conflict, the 2012 Presidential Debates. These projects reveal the importance of dynamics. In two-person conflict, for example, signal dynamics (e.g., body, voice) during interaction can reveal the quality of that interaction. In addition, collective behavior on Twitter can be predicted even by simple linear models using debate dynamics between Obama and Romney (e.g., one interrupting the other).

The collection, quantification, and modeling of multitemporal and multivariate datasets hold much promise for new kinds of interdisciplinary collaborations. I will end by discussing how they may guide new theoretical directions for pursuing the organization and temporal structure of multimodality in communication.

Find out more »

Betsy Sinclair, Washington University in St Louis

March 29, 2016 @ 2:30 pm - 4:00 pm PDT
314 Royce Hall, 340 Royce Dr
los angeles, CA 90095 United States

"Electronic Homestyle: Tweeting Ideology"

Abstract: Ideal points are central to the study of political partisanship and an essential component to our understanding of legislative and electoral behavior. We employ automated text analysis on tweets from Members of Congress to estimate their ideal points using Naive Bayes classification and Support Vector Machine classification. We extend these tools to estimate the proportion of partisan speech used in each legislator's tweets. We demonstrate an association between these measurements, existing ideal point measurements, and district ideology.

Find out more »

October 2015

Aude Hofleitner, Facebook

October 15, 2015 @ 12:00 pm - 1:30 pm PDT
CCPR Seminar Room, 4240 Public Affairs Building
Los Angeles, CA 90095 United States
+ Google Map

"Inferring and understanding travel and migration movements at a global scale"

Abstract: Despite extensive work on the dynamics and outcomes of large-scale migrations, timely and accurate estimates of population movements do not exist. While censuses, surveys, and observational data have been used to measure migration, estimates based on these data sources are constrained in their inability to detect unfolding migrations, and lack temporal and demographic detail. In this study, we present a novel approach for generating estimates of migration that can measure movements of particular demographic groups across country lines.

Specifically, we model migration as a function of long-term moves across countries using aggregated Facebook data. We demonstrate that this methodological approach can be used to produce accurate measures of past and ongoing migrations - both short-term patterns and long-term changes in residence. Several case studies confirm the validity of our approach, and highlight the tremendous potential of information obtained from online platforms to enable novel research on human migration events.

Find out more »
+ Export Events