Day 1: Feb 25,  Principles and Conceptual Framework for Selecting CI Methods

08:30 AM

Registration and Breakfast

09:00 AM

 

Welcome and Opening Remarks

Speaker: Yi Zhang, MTPPI

09:10 AM

 

Conference Introduction: the Importance of CI Methods for PCOR/CER

Speaker: Jason Gerson, Associate Director for CER Methods and Infrastructure team, PCORI (Slides)

ModeratorJulia Kim, MD, Assistant Professor, Pediatrics, John Hopkins School of Medicine

09:30 AM

 

Causal Inference as an attempt to emulate a target experiment

Speaker: Miguel Hernan, Harvard School of Public Health. (Slides)

Moderator: Julia Kim, MD, Assistant Professor, Pediatrics, John Hopkins School of Medicine

Abstract: Ideally, questions about comparative effectiveness or safety would be answered by an appropriately designed and conducted randomized experiment. When we cannot conduct the randomized experiment, we analyze observational data. Causal inference from large observational databases can be viewed as an attempt to emulate a randomized experiment---the target experiment or target trial--- that would answer the question of interest. When the goal is to guide decisions among several strategies, causal analyses of observational data need to be evaluated with respect to how well they emulate a particular target trial.

This talk outlines a framework for comparative effectiveness research using observational data that makes the target trial explicit. This framework channels counterfactual theory for comparing the effects of sustained treatment strategies, organizes analytic approaches, provides a structured process for the criticism of observational studies, and helps avoid common methodologic pitfall

10:10 AM

 

Fundamental issues: Understanding differences between observed and unobserved confounding

Speaker: Mike Baiocchi, Stanford University (Slides)

Moderator: Julia Kim, MD, Assistant Professor, Pediatrics, John Hopkins School of Medicine

Abstract: This talk will provide a foundational framework for understanding the primary concerns in estimating causal effects. We will discuss different sources of variation in your data that, if not addressed properly, might lead to erroneous conclusions. We will focus on the difference between observed and unobserved confounding - providing mathematical, intuitive and real-world examples - and then move on to how to pick methods that address problems arising from observed, unobserved or both kinds of confounding. The goal of this talk is to make you comfortable with the terminology and key ideas of confounding, as well as aware of the class of methodologies/study designs that appropriate to address your concerns.

10:50 AM

Tea Break

11:05 AM

Major threats to validity in observational studies

Speaker: Jay Kaufman, McGill University (Slides)

Moderator: Julia Kim, MD, Assistant Professor, Pediatrics, John Hopkins School of Medicine

Abstract: In observational studies in which the goal is etiologic inference, the investigator seeks to estimate the contrast between two potential interventions on some defined target population.  When this is based on observing associations or adjusted associations between exposure and outcome in the data set, the estimate is considered to be valid if a study with an increasing large sample size would converge to the true value of the causal effect.  There are several broad threats to the validity of the estimate -- improprieties that can lead the estimated value of the effect to be too high or too low.  The most common classification of these threats is into 1) confounding biases, 2) selection biases, and 3) information biases.  Each of these classes of potential bias has a unique structure that can be illustrated in a causal graph. 

Confounding bias involves common causes of the exposure and the outcome.  Selection bias, in contrast, involves conditioning in the design or analysis on consequences of the exposure and/or consequences of the outcome. Finally, information biases involve measurement error or misclassification of any of the study variables, including exposure, outcome or covariates.  Beyond these sources of bias, there are other potential errors in the analysis, for example the misspecification of the model or the inappropriate adjustment for factors affected by the exposure.   Taken together, these various challenges contribute to yielding an estimate that can deviate systematically from the true causal effect, which is the value that would be obtained under an actual intervention.  P-values and confidence intervals relate to only one alternative explanation for a study finding:  that it is a chance variation due to sampling variability.  These other sources of bias and error represent other alternative explanations, and also need to be considered and ruled out before one can extend much credibility to the estimated effects of an observational study.

11:45 AM

 

The 'best' or 'best possible' methods: controversies in the field

Speaker: Heejung Bang, University of California, Davis (Slides)

Moderator: Julia Kim, MD, Assistant Professor, Pediatrics, John Hopkins School of Medicine

Abstract: Causal inference is inherently complicated, yet a number of methods are available nowadays. Some methods are popularly adopted in practice, especially, with programming guide/software, while some methods are still considered unwieldy or some users do not feel full comfort and trust in use.  Despite high demand and increased acceptance, causal inference based on non-experimental data is still a controversial topic, including in statistical training/curriculum, where mechanical use of a method with inadequate considerations to qualitative issues may do more harm than good.

Causal inference naturally entails a number of assumptions, some of which are strong or unverifiable with ‘observed’ data.  Moreover, it can be a daunting task for practitioners or even quantitative researchers/data analysts to remember and check all of the assumptions and caveats in each step in implementation.  Also, we find that different methodology camps tend to claim their method is the best. In this talk, we will discuss how to choose best or best feasible methods for a given problem and report the findings. We will further discuss some controversies in causal inference, and see if we can harmonize the practice in wide but different fields

12:25 PM

Lunch Break

01:25 PM

 

Assumptions, sensitivity analyses, and software.

Speaker: Dylan Small, University of Pennsylvania (Slides)

Moderator: Martin Ho, FDA/CDRH/OSB/DBS; 

Abstract: Two fundamental assumptions for many observational study analysis methods are (1) that there is no unmeasured confounding (also called treatment ignorability, exchangeability, selection on observables or treatment exogeneity); (2) positivity -- that for every covariate value, there is a positive probability of being in both the treatment and control group.  We will discuss these assumptions, ways of dealing with violations of these assumptions through sensitivity analysis and limiting the study population, and associated software.

02:05 PM

 

Sensitivity analysis for untestable assumptions: formulation, implementation, interpretation

Speaker: Joseph Hogan: Brown University (Slides)

Moderator: Martin Ho, FDA/CDRH/OSB/DBS

Abstract: A rigorous analysis about causal effects includes some accounting for extra variation attributable to untestable assumptions. To fix ideas, this talk focuses on unmeasured confounding. Using an example from HIV epidemiology, we describe three specific approaches: (i) using bounds instead of point estimates to reflect lack of information about the unmeasured confounder(s); (ii) representing the unmeasured confounder as an unobserved variable that is associated with the outcome and treatment; and (iii) representing unmeasured confounding in terms of differences between potential outcomes distributions. Methods (ii) and (iii) can be used to generate graphical depictions, ‘tipping point’ analyses, and other representations of unmeasured confounding bias. The talk will compare and contrast each approach with focus on usage in practical settings.

02:45 PM

Tea Break

03:00 PM

Considerations on the choice of the statistical methods for PCOR questions

Speaker: Sara Lodi, Harvard School of Public Health (Slides)

Moderator: Martin Ho, FDA/CDRH/OSB/DBS

Abstract: PCOR questions often involve the comparison of complex treatment strategies sustained over long periods of time using data collected from routine clinical practice (or electronic health records). Traditional statistical methods (e.g. Cox or logistic regression analysis) do not always provide correct answers to these types of questions under valid assumptions. Using examples from the medical literature, we will illustrate when and why traditional methods fail. In particular, we will show that in the presence of time-updated treatments and/or confounders, a common situation when using routinely collected clinical data, an alternative class of methods, so-called g-methods become the most appropriate choice.

 

03:40 PM

 

A systematic road map for translating a CER question into a statistical analysis

Speaker: Maya Petersen, University of California, Berkley (Slides)

Moderator: Martin Ho,FDA/CDRH/OSB/DBS

Abstract: This session will describe a systematic road map for translating an applied causal question into a statistical analysis, highlighting the following key points in this process: 1) translation of a scientific question into formal causal query, with an emphasis on how to select a target counterfactual quantity (defined in terms of an ideal experiment) that comes as close as possible to answering the motivating scientific question; 2) translation of a causal query into a statistical estimation problem, with an emphasis on how to choose an appropriate statistical model and estimand (statistical target parameter); and, 3) estimation and inference, with an emphasis on how to choose the estimator that will perform best (in terms of robustness, precision, and reliable inference) in your data.

 

04:20 PM

 

Day 1 Wrap-up ; Day 2 Overview

Speaker: Yi Zhang, MTPPI.

04:30 PM

 

Adjournment

 

Day 2: Feb 26 2016,  Matching the method(s) to the question

08:30 AM

Registration and Breakfast                                                                                          

08:55 AM

 

Day 2 : Introduction

Speaker: Yi Zhang, MTPPI

09:00 AM

 

"Study design" vs "inference"

Speaker: Mike Baiocchi, Stanford University (Slides)

Co-Moderators: Yan Ma, PhD, Associate Professor and Marinella Temprosa, PhD, Assistant Professor, Department of Epidemiology and Biostatistics, The George Washington University

Abstract: Most causal inference analyses have two main phases: (i) the study design phase followed by (ii) statistical inference. In randomized trials these two phases are more obvious (e.g., (i) do we implement a cluster-randomized trial or a uniformly randomized trial? (ii) do we then analyze the data using a permutation test or some sort of regression?). In observational studies many new researchers focus on the "fancy" forms of inference required to analyze observational data, while overlooking the fact that study design is still a very important step in the analysis plan. And, perhaps even more overlooked, it is the case that the researcher has options regarding study design in observational studies. In this talk we will disentangle study design from inference and discuss how to improve your observational studies by investing more thought in the study design phase.

09:40 AM

 

Introduction to various CI Methods

Speaker: Miguel Hernan, Harvard School of Public Health. (Slides)

Co-Moderators: Yan Ma, PhD, Associate Professor and Marinella Temprosa, PhD, Assistant Professor, Department of Epidemiology and Biostatistics, The George Washington University

Abstract: The application of causal inference methods is growing exponentially given the interest in using observational data to address CER/PCOR questions. This session will present the conceptual framework and an accessible overview of various causal inference methods and their applications. With several examples based on recent publications using real-world epidemiologic data as well as a variety of analytical techniques, Dr. Hernán will give a detailed introduction to available CI methods for appropriately comparing non-time-varying treatments and complex long-term dynamic treatment strategies.

10:20 AM

 

Matching/Stratification/Regression

Speaker: Heejung Bang, University of California, Davis (Slides)

Co-Moderators: Yan Ma, PhD, Associate Professor and Marinella Temprosa, PhD, Assistant Professor, Department of Epidemiology and Biostatistics, The George Washington University

Abstract: When we compare different strategies A vs. B (and C) on an outcome, a gold standard approach is randomization. But when randomization is not feasible, which is very common in practice, we use observational data with statistical techniques, which may be the next best thing. ​N​on-experimental observational data and natural Big​Data offer indispensable sources to comparative effectiveness research (CER). Among traditional design and analysis methods, matching, stratification​,​ and regression may be most popular with a long history of usage. Nowadays, more advanced, novel methods based on propensity score (PS; which will be discussed ​in a separate session by Dr. Landsittel​) have been popular, where PS can be used via the traditional framework of matching/stratification/regression. In this talk, we will ​review ​traditional methods​, illustrate their use for causal inference by case examples​ ​, ​​and make us ready for learning the PS. We will also discuss the advantages of disadvantages of BigData

11:00 AM

Tea Break

11:10 AM

 

Practical applications and decisions for using propensity score methods

Speaker: Douglas Landsittel, University of Pittsburgh (Slides)

Co-Moderators: Yan Ma, PhD, Associate Professor and Marinella Temprosa, PhD, Assistant Professor, Department of Epidemiology and Biostatistics, The George Washington University

Abstract: Observational data represent an important and increasingly available tool for patient-centered comparative effectiveness research. Many different methods are available for making causal inferences from observational data, but propensity scores are probably the most common approach. This talk, after noting some important considerations for study design, reviews the main concept behind estimating and applying propensity scores to observational data. We also introduce a project aimed at developing a Decision Tool for Observational Data Analysis Methods in Comparative Effectiveness (DecODe CER) for recommending optimal methods for a given data set. Finally, we review some issues that arise in disseminating these methods in the context of clinical research education.

11:55 AM

Structural Nested Models/G-estimation

Speaker: Dylan Small, University of Pennsylvania (Slides)

Co-Moderators: Yan Ma, PhD, Associate Professor and Marinella Temprosa, PhD, Assistant Professor, Department of Epidemiology and Biostatistics, The George Washington University

Abstract: Structural nested models and the associated method of G-estimation are approaches to modeling and estimating the joint effects of a sequence of treatments or exposures. We will present the approach with an application and discuss advantages over other methods developed for estimating such joint effects.

12:35 PM

Lunch Break

 01:25 PM

 

The parametric g-formula and inverse probability weighting

Speaker: Sara Lodi, Harvard School of Public Health (Slides)

Moderator: Jodi Segal, MD, MPH, Professor of Internal Medicine, John Hopkins University School of Medicine

Abstract: The g-methods are a class of methods that can be used to answer PCOR questions with complex longitudinal data and in presence of time-updated treatment and confounders. We will focus on two g-methods: the parametric g-formula and inverse probability weighting (IPW) of marginal structural models (MSMs). Both methods can be used to answer the same questions, but differ in the set of assumptions about the model specification. We illustrate examples of practical implementation of both methods and will discuss the relative advantages and disadvantages of each.

 02:05 PM

 

Targeted Maximum Likelihood Estimation/Super Learning/Doubly Robust Estimator

Speaker:  Maya Petersen, University of California, Berkeley (Slides)

Moderator: Jodi Segal, MD, MPH, Professor of Internal Medicine, John Hopkins University School of Medicine

Abstract: In most practical applications, we are faced with a large number of candidate adjustment variables and our contextual knowledge is insufficient to correctly a priori specify parametric models for outcome regressions or for propensity scores/inverse probability weights. In such settings, machine-learning approaches can be used to reduce bias due to model misspecification. This session will first review Super Learning, a powerful machine-learning tool based on internal data splitting, which can be used to estimate both outcome regressions and inverse probability weights. The session will next provide an overview of doubly robust Targeted Maximum Likelihood Estimation (TMLE) in the context of causal effect estimation. TMLE naturally incorporates Super Learning and offers several theoretical and practical advantages over alternative statistical methods, including improved robustness to model misspecification and potentially greater precision. Both Super Learning and TMLE are available as open source R software packages; example software code will be provided.

 02:45 PM

Tea Break

 03:00 PM

Instrumental variables and principal stratification for causal inference

Speaker: Joseph Hogan, Brown University (Slides)

Moderator: Jodi Segal, MD, MPH, Professor of Internal Medicine, John Hopkins University School of Medicine

Abstract: This talk will provide an overview of how instrumental variables and principal stratification can be used to estimate causal effects. To motivate the methods in ideal settings, we will first consider the problem of estimating the causal effect of a treatment in a randomized trial where there is noncompliance, and describe conditions needed to estimate average (overall) causal effect, and causal effect among certain subgroups (such as those who would comply with the treatment if offered)

Instrumental variable methods are widely used by economists and social scientists to analyze observational data, especially for policy evaluation. The use of IV and PS is somewhat more complicated for observational data because each method requires the existence of a ‘randomizer’, or a mechanism that is related to receipt of treatment while being independent of the outcome of interest. We will illustrate both methods in an observational cohort study, focusing on conceptual issues that are important for practice

03:40 PM

Summary: connecting the question to the analysis(es)

Speakers: Jay Kaufman, McGill University (Slides)

Moderator: Jodi Segal, MD, MPH, Professor of Internal Medicine, John Hopkins University School of Medicine

Abstract: Previous sessions have covered in-depth discussions on principles for choosing appropriate methods to answer PCOR questions using real-world data. Dr. Kaufman will provide a summary from the multi-faceted presentations of conference speakers regarding fundamentals of CI, study design, threats to valid inference, and how to choose wisely between various methods to mitigate those threats. Attendees are encouraged to ask their own research questions from current work and receive feedback from all speakers and moderators regarding study design and methods.

04:20 PM

 

Look Forward : CIMPOD 2017

Speaker: Yi Zhang, MTPPI.

04:30 PM

 

Adjournment