2016 Charles L. Odoroff Memorial Lecture

In honor of Dr. Charles L. Odoroff, the founding Director of the Division (now Department) of Biostatistics at the University of Rochester, the Department hosts an annual lecture by a distinguished statistician. This lecture series is supported by funds contributed by family and friends of Dr. Odoroff after his untimely death in 1987 at age 49.

**Thursday, May 5, 2016**

**Roderick Little, PhD **

Richard D. Remington Distinguished University Professor

Department of Biostatistics

University of Michigan

## Treatment Discontinuation and Missing Data in Clinical Trials

I briefly review recommendations of a recent National Research Council study on the treatment of missing data in clinical trials. I then discuss two aspects of the analysis of clinical trials when participants prematurely discontinue treatments. First, I formulate treatment discontinuation and “analysis dropout” as different missing data problems: analysis dropout concerns missing data on outcomes, and treatment discontinuation concerns missing data for a covariate defining principal strata for completion under alternative treatments. Second, the choice of estimand is an important factor in considering when follow-up measures after discontinuation are needed to obtain valid measures of treatment effects. Follow-up measures are recommended for the standard choice of intention to treat (ITT) estimand, the average effect of randomization to treatment, but this measure has the disadvantage that it includes the effects of any treatments taken after the assigned treatment is discontinued. I discuss alternative estimands for the ITT population, and argue that one in particular, an on-treatment summary of the effect of treatment prior to discontinuation, should receive more consideration. Ideas are motivated and illustrated by a reanalysis of a past study of inhaled insulin treatments for diabetes, sponsored by Eli Lilly.

## Previous Lectures

### 2015 Charles L. Odoroff Memorial Lecture

**Marie Davidian, PhD **

William Neal Reynolds Professor

Department of Statistics

North Carolina State University

*The Right Treatment for the Right Patient (at the Right Time): Personalized Medicine and Dynamic Treatment Regimes*

With the advent of the 'omics era, achieving the goal "personalizing" treatment to the patient based on his/her genetic/genomic as well as physiological, demographic, and other characteristics and past history has become more promising than ever. One perspective on personalized medicine involves identifying subgroups of patients sharing certain characteristics who are likely to benefit from a specific treatment or to whom a new treatment may be targeted ("the right patient"). Another is based on formalizing how clinicians make treatment decisions in practice, where the goal is to identify the most beneficial treatment to administer to a patient from among the available options given his/her characteristics ("the right treatment"). In chronic diseases and disorders like cancer or substance abuse, a series of treatment decisions must be made, and the objective is to determine the "best" treatment option for a patient at each decision given all information accrued on the patient to that point, including responses to previous treatments, so as to lead to the most beneficial long term outcome. This sequential decision-making introduces many complications; for example, treatments chosen early on may affect how well treatments given later will work ("at the right time").The development of optimal, evidence-based personalized treatment strategies (dynamic treatment regimes) can be formulated as a fascinating statistical problem. I will provide an overview of challenges involved and of study designs and methodological developments that can be used in the quest for personalized medicine.

Thursday, April 30, 2015

### 2014 Charles L. Odoroff Memorial Lecture

**Thomas A. Louis, PhD **

Professor, Department of Biostatistics

Johns Hopkins Bloomberg School of Public Health

*Research at Census, with Links to Biostatistics *

In order to meet the challenges of efficiently obtaining valid demographic, economic, and activity-based information, making it available to the public, and protecting confidentiality, research at the U.S. Census Bureau and other federal statistical agencies, indeed survey research more generally, burgeons. Many research goals and methods are similar to those addressed by and used in Biostatistics or Informatics. To set the scene, I briefly describe the Census Research & Methodology directorate, list major issues and approaches, then provide details on a small subset. Candidate topics include adaptive design (dynamic survey modes, R-factors in the National Survey of College Graduates, timing of mailing hard copy based on K-M curves, challenges of learning from experience), stopping rules, randomized experiments (the effect of interviewer training in the National Crime Victimization Survey), record matching, prediction (of response propensity, of occupancy, of the “fitness for use” of administrative records), imputation, Bayesian methods (design-consistent analyses, post-processed {confidence} intervals, benchmarking), small area/spatio-temporal analysis (estimation of poverty rates, estimating omissions in the master address file), development and use of paradata (in the National Health Interview Survey), double-robustness, dynamic data posting (“OnTheMap” Local Origin-Destination Employment Statistics), disclosure avoidance/limitation, Big Data (opportunities and challenges), micro-simulation (benefits of research in designing the 2020 Census), and IT infrastructure (the Multi-mode Operational Control System). I close with a call for increased collaboration among statistical agencies and academe, building on the NSF-Census Bureau Research Network. Visit (http://www.census.gov/research/) for some background.

Thursday, May 1, 2014

### 2013 Charles L. Odoroff Memorial Lecture

**Jack Kalbfleisch, PhD **

Professor Emeritus of Statistics and Biostatistics

The University of Michigan School of Public Health

**Randomization and re-randomization in clinical trials**

Randomization was a key contribution of Sir Ronald Fisher to the conduct of scientific investigations and statistical methods. Along with the protective aspects of randomization, Fisher also noted that the distribution induced by randomization can form the basis of inference. Indeed, in some instances, the randomization test and related procedures seem to be the only tools available for inference. Several authors have noted the advisability of rerandomizing if, in a particular instance, the observed randomization leads to an unacceptable degree of imbalance in important factors between and among the treatment groups. Morgan and Rubin (2012, Annals of Statistics) provide an excellent discussion and some interesting results. This talk begins with some discussion of randomization and then considers problems arising in the design of relatively small cluster randomized trials, which have been widely used in recent years for evaluation of health-care strategies. The balance match weighted (BMW) design, introduced in Xu and Kalbfleisch (2010, Biometrics), applies propensity score matching ideas to choose a design through a rerandomization approach with the general aim of minimize the mean squared error of the treatment effect estimator. The methods are evaluated by simulation. Extensions of the methods to multiple armed trials are also considered and simply implemented numerical methods are proposed to achieve good matching algorithms that achieve near optimum results. Analysis issues are also discussed. Standard parametric and nonparametric methods are often inappropriate for analysis of designs involving rerandomization, though the distribution generated by the rerandomization approach provides a general framework for analysis. With the matching approach of the BMW design, the use of analysis models that respect the matching are also investigated.

This is based on joint work with Dr. Zhenzhen Xu.

Thursday, April 25, 2013

### 2012 Charles L. Odoroff Memorial Lecture

**Nancy L. Geller, PhD **

Director, Office of Biostatistics Research

National Heart, Lung, and Blood Institute

**Has the time come to give up blinding in clinical trials?**

Should all trials be double blinded, that is, should treatment allocation be concealed from both the subjects and those administering the treatment? In the late 1980's and early1990's trialists advocated strongly for double blinding of clinical trials, yet in the past 15 years, we have seen more and more clinical trials that are unblinded. While it is relatively easy to make a placebo controlled trial of a medication given orally double blinded, reasons for not blinding include that in some situations it is too difficult (or expensive) to blind, in some situations, it may be unethical to blind and in other situations, it is impossible to blind. Complex interventions may make blinding especially difficult. Comparative effectiveness studies also encourage unblinded trials because “blinding is not done in the real world.” We give several examples of recent trials which have not been blinded and examine the consequences.

Thursday, May 3, 2012

### 2011 Charles L. Odoroff Memorial Lecture

**Amita K. Manatunga , Ph.D.**

Professor

Rollins School of Public Health

Emory University

**A Framework for the Assessment of Disease Screening Instruments in Mental Health Studies**

A fundamental objective in biomedical research is to establish valid measurements of the clinical disease of interest. Measures of agreement have been widely used for validating a new instrument by assessing similarity of measurements with an established instrument.

Although the foundation of agreement methodology has been mostly laid out, many important statistical issues have not yet been resolved. In this presentation, I will present our recent work on the following two problems: (1) how to extend the classical framework of agreement to evaluate the capability of interpreting a continuous measurement in an ordinal scale; (2) how to subdivide a continuous scale into ordered categories when there is high correspondence between two scales.

To address the first question, we propose a new concept, called “broad sense agreement", which characterizes the correspondence between a continuous scale and an ordinal scale. We present a natural measure for broad sense agreement. Nonparametric estimation and inference procedures are developed for the proposed measure along with theoretical justifications. To address the second question, we develop a new approach for determination of cut-points in a continuous scale according to an established categorical scale by adopting the idea of optimizing the agreement between the discretized continuous scale and the categorical scale. We also discuss analytic and empirical advantages of our method. Finally, we apply these methods to a mental health study to illustrate their practical utility.

Thursday, May 26, 2011

### 2010 Charles L. Odoroff Memorial Lecture

**Raymond J. Carroll, Ph.D.**

Distinguished Professor of Statistics, Nutrition and Toxicology

Texas A&M University

**Robust Powerful Methods for Understanding Gene-Environment Interactions**

We consider population-based case-control studies of gene-environment interactions using prospective logistic regression models. Data sets like this arise when studying pathways based on haplotypes as well as in multistage genome wide association studies (GWAS). In a typical case-control study, logistic regression is used and there is little power for detecting interactions. However, in many cases it is reasonable to assume that, for example, genotype and environment are independent in the population, possibly conditional on factors to account for population stratification. In such as case, we have developed an extremely statistically powerful semiparametric approach for this problem, showing that it leads to much more efficient estimates of gene-environment interaction parameters and the gene main effect than the standard approach: decreases of standard errors for the former are often by factors of 50% and more. The issue of course that arises is the very assumption of conditional independence, because if that assumption is violated, biases result so that one can announce gene-environment interactions or gene effects even though they do not exist. We will describe a simple, computationally fast approach for gaining robustness without losing statistical power, one based on the idea of Empirical Bayes methodology. Examples to colorectal adenoma studies of the NAT2 gene and prostate cancer in the VDR pathway are described to illustrate the approaches.

Friday, April 16, 2010

### 2009 Charles L. Odoroff Memorial Lecture

**Terry M. Therneau , Ph.D.**

Professor of Biostatistics

Mayo Clinic

*Random Effects Models and Survival Data*

The literature on random effects models for survival data, also known as frailty models, has burgeoned in the last few years. With multiple choices for the distribution, interrelations, and computation of the random effects it has been fruitful soil for theoretical forays. This talk will focus on the practical uses of the models: what software is readily available, what types of problems can we apply it to, and most importantly, what does the approach add to our final clinical or biological understanding.

Thursday, April 2, 2009

### 2008 Charles L. Odoroff Memorial Lecture

**Marvin Zelen, Ph.D.**

Lemuel Shattuck Research Professor of Statistical Science

Department of Biostatistics

Harvard School of Public Health

*Early Detection of Disease and Stochastic Models*

The early detection of disease presents opportunities for using existing technologies to significantly improve patient benefit. The possibility of diagnosing a chronic disease early, while it is asymptomatic, may result in diagnosing the disease in an earlier stage leading to better prognosis. Many cancers, diabetes, tuberculosis, cardiovascular disease, HIV related diseases, etc. may have better prognosis when combined with an effective treatment. However gathering scientific evidence to demonstrate benefit has proved to be difficult. Clinical trials have been arduous to carry out, because of the need to have large numbers of subjects, long follow-up periods and problems of non-compliance. Implementing public health early detection programs have proved to be costly and not based on analytic considerations. Many of these difficulties are a result of not understanding the early disease detection process and the disease natural histories. One way to approach these problems is to model the early detection process. This talk will discuss stochastic models for the early detection of disease. Breast cancer will be used to illustrate some of the ideas. The talk will discuss breast cancer randomized trials, stage shift and benefit, scheduling of examinations, issue of screening younger and older women and the probability of over diagnosis of disease.

Tuesday, April 29, 2008

### 2007 Charles L. Odoroff Memorial Lecture

**Mark van der Laan, Ph.D.**

University of California, Berkeley

*Targeted Maximum Likelihood Learning of Scientific Questions*

The main point of this presentation is that choice of a statistical model and method must be based on careful consideration of the scientific question of interest in order to provide robust tests of the null hypothesis and to minimize bias in the parameter estimate. For this purpose we developed a new generally applicable targeted maximum likelihood estimation methodology.

As an example, I will distinguish between scientific questions concerned with prediction of an outcome based on a set of input variables versus scientific questions in which the goal is to estimate the variable importance or causal effect of one particular variable/treatment. I will show the limitations of fitting regression models for the purpose of learning about a causal effect or variable importance, and present the alternative targeted maximum likelihood approach. Both observational studies and randomized trials will be used to illustrate the advantages of the targeted approach. I will present results from data analyses in which the targeted approach is used to 1) analyze the importance of each of a set of HIV mutations for protease inhibitor resistance and 2) estimate the causal effect of interventions to improve adherence to antiretroviral drugs.

The differences between prediction and causal effect estimation are further highlighted by the additional assumptions needed for the estimation of the causal effect of an intervention in an observational study. Beyond the familiar ''no unmeasured confounding'' assumption, causal effect estimation also requires an experimental treatment assignment assumption, violation of which can cause severe bias and increased variance in a causal effect estimate. To address this problem, I will show that estimation of the causal effect of a "realistic" intervention (similar to the parameter one estimates in an intention-to-treat analysis) provides an important generalization which can always be fully identified from the data. Targeted estimators of this realistic parameter are also available.

Finally, I will discuss the advantages of applying targeted estimation in the context of a randomized trial. Like standard approaches, the targeted approach relies only on the randomization assumption. However, the targeted approach yields an improved estimate of the causal effect of a treatment in a randomized trial relative to the commonly used marginal estimate of the treatment effect.

Thursday, September 20, 2007

### 2006 Charles L. Odoroff Memorial Lecture

**Butch Tsiatis, Ph.D.**

Department of Statistics

North Carolina State University

*Estimating Mean Response as a Function of Treatment Duration in an Observational Study, Where Duration May be Informatively Censored*

In a recent clinical trial "ESPRIT" of patients with coronary heart disease who were scheduled to undergo percutaneous coronary intervention (PCI), patients randomized to receive Integrilin therapy had significantly better outcomes than patients randomized to placebo. The protocol recommended that Integrilin be given as a continuous infusion for 18--24 hours. There was debate among the clinicians on the optimal infusion duration in this 18--24-hour range, and we were asked to study this question statistically. Two issues complicated this analysis: (i) The choice of treatment duration was left to the discretion of the physician and (ii) treatment duration would have to be terminated (censored) if the patient experienced serious complications during the infusion period. To formalize the question, "What is the optimal infusion duration?" in terms of a statistical model, we developed a framework where the problem was cast using ideas developed for adaptive treatment strategies in causal inference. The problem is defined through parameters of the distribution of (unobserved) potential outcomes. We then show how, under some reasonable assumptions, these parameters could be estimated. The methods are illustrated using the data from the ESPRIT trial.

Thursday, May 11, 2006

### 2005 Charles L. Odoroff Memorial Lecture

**Louise Ryan, Ph.D.**

Professor of Biostatistics

Harvard School of Public Health

*Prenatal Methylmercury Exposure and Childhood IQ*

Controversy continues regarding the impact of chronic methylmercury exposures on childhood development. Adverse effects are difficult to quantify at low doses, and conflicting results have been obtained from several well-designed epidemiological studies, one in the Faroe Islands, one in the Seychelles and an older small study in New Zealand. We describe the use of hierarchical modeling techniques to combine data on several endpoints from these three studies. We find convincing evidence of an effect of methylmercury exposure on full scale IQ in children aged 6 to 9 years.

Wednesday, March 16, 2005

### 2004 Charles L. Odoroff Memorial Lecture

**Bruce Levin, Ph.D.**

Professor and Chair, Department of Biostatistics

Columbia University Mailman School of Public Health

*A Generalization of the Levin-Robbins Procedure for Binomial Subset Selection and Recruitment Problems*

We introduce a family of sequential selection and recruitment procedures for the subset identification problem in binomial populations. We demonstrate the general validity of a simple formula providing a lower bound for the probability of correct identification in a version of the family without sequential elimination or recruitment. A similar theorem is conjectured to hold for the more efficient version which employs sequential elimination or recruitment.

Thursday, April 15 2004

### 2003 Charles L. Odoroff Memorial Lecture

**Danyu Lin, Dennis Gillings Professor**

University of North Carolina at Chapel Hill

*Selection and Assessment of Regression Models*

Residuals are informative about the adequacy of regression models. Conventional residual analysis based on the plots of individual residuals is highly subjective, while most numerical goodness-of-fit tests provide little information about the nature of model misspecification. In this talk, we present objective and informative strategies for model selection and assessment based on the cumulative sums of residuals over certain coordinates (e.g., covariates or fitted values) or some related aggregates of residuals (e.g., moving averages and kernel smoothers). The distributions of these stochastic processes under the assumed model can be approximated by the distributions of certain zero-mean Gaussian processes whose realizations can be easily generated by computer simulation. Each observed residual pattern can then be compared, both graphically and numerically, with a number of realizations from the null distribution. Such comparisons enable one to assess objectively whether a specific aspect of the model (e.g., the functional form of a covariate, the link function or the proportional hazards assumption) has been correctly specified. They also provide helpful hints on how to obtain an appropriate model. We apply this approach to a wide variety of statistical models and data structures, and provide illustrations with several clinical and epidemiologic studies. The methods presented in this talk will be featured in the next release of SAS.

Wednesday, May 7 2003

### 2002 Charles L. Odoroff Memorial Lecture

**Professor Mark Espeland**

Wake Forest University

*Heterogeneity of Response*

It is to be expected that responses to treatments, measurement techniques, and diseases may vary among individuals. Characterizing the underlying distributions of heterogeneous responses is often difficult in that it requires statistically removing the confounding influences of measurement error from the observed response distributions. This is an important problem, however, to the extent that there have been calls that drug approval be based on the nature of underlying response distributions, rather than merely average treatment effects. Some general approaches to the problem of estimating response distributions will be discussed. Three examples, from separate studies of postmenopausal hormone therapy, hypertension control, and carotid atherosclerosis, will be used as case studies. In each, multivariate hierarchical measurement error models were fitted with varying success to estimate characteristics of underlying response distributions. The goal of describing heterogeneous responses presents major challenges for study designs. Improved methodology and increased data sharing are needed.

Thursday, April 4 2002

### 2001 Charles L. Odoroff Memorial Lecture

**Professor Scott Zeger**

**Department of Biostatistics, John Hopkins University**

This talk will describe "SQUARE", a novel method for estimating the difference in means between two skewed distributions given one relatively smaller and one larger sample. This problem arises in assessing the medical costs of smoking.

We will give an overview of the statistical problem of determining smoking attributable expenditures which includes estimating the average cost of services for persons with smoking-caused diseases relative to otherwise similar persons without disease. We then introduce an estimator of this difference that relies on estimation of the ratio of the quantile functions for the two distributions. When the degrees of freedom to estimate this function is small, our estimator approximates the log-normal model mle; when they are large, it converges to the difference in sample means.

We illustrate SQUARE with an analysis of the difference in medical costs with persons who suffer lung cancer and chronic obstructive pulmonary disease as compared to otherwise similar people who do not.

Thursday, April 12 2001

### 2000 Charles L. Odoroff Memorial Lecture

**Professor Norman Breslow**

Department of Biostatistics, University of Washington

*Descriptive Statistics and the GeNetics of Wilms Tumor: Lessons From the National Wilms Tumor Study Group*

Wilms tumor is an embryonal tumor of the kidney that affects approximately one child per 10,000 before the age of 15 years. It was almost universally fatal a century ago, but cure rates today approach 90% due largely to the use of modern combination chemotherapy. The National Wilms Tumor Study Group (NWTSG), founded in 1969, has been in the forefront of the recent advances.

Wilms tumor has served as a model of complexity for understanding of the genetic origins of cancer. It was initially believed to follow the two-hit mutational model proposed by Knudson (1972) on the basis of his statistical analysis of data for retinoblastoma, another childhood tumor involving a paired organ. Simple descriptive analyses of the large NWTSG database, however, suggested that the genetics of Wilms tumor were more complex. This suggestion was subsequently confirmed by laboratory studies demonstrating a role for genomic imprinting in the etiology of Wilms tumor, the limited number of cases associated with mutations in the first (and to date only) Wilms tumor gene to be cloned (WT1), and the linkage of familial cases to at least two other distinct loci. This talk will present several examples of how simple descriptive statistical analyses can challenge prevailing genetic models and suggest new avenues for laboratory investigation.

Thursday, March 23, 2000

### 1999 Charles L. Odoroff Memorial Lecture

**Professor David L. DeMets**

**University of Wisconsin**

*Surrogate End Points in Clinical Trials: Are We Being Misled?*

Phase 3 clinical trials, which evaluate the effect that new interventions have on the clinical outcomes of particular relevance to the patient(such as death, loss of vision, or other major symptomatic event), often require many participants to be followed for a long time. There has recently been great interest in using surrogate end points, such as tumor shrinkage or changes in cholesterol level, blood pressure, CD4 cell count, or other laboratory measures, to reduce the cost and duration of clinical trials. In theory, for a surrogate end point to be an effective substitute for the clinical outcome, effects of the intervention on the surrogate must reliably predict the overall effect on the clinical outcome. In practice, this requirement frequently fails. Among several explanations for this failure is the possibility that the disease process could affect the clinical outcome through several causal pathways that are not medated through the surrogate, with the intervention's effect on these pathways differing from its effect on the surrogate. Even more likely, the intervention might also affect the clinical outcome by unintended, unanticipated, and unrecognized mechanisms of action that operate independently of the disease process. Examples from several disease areas illustrate how surrogate end points have been misleading about the actual effects that treatments have on the health of patients. Surrogate end points can be useful in phase 2 screening trials for identifying whether a new intervention is biologically active and for guiding decisions about whether the intervention is promising enough to justify a large definitive trial with clinically meaningful outcomes. In definitive phase 3 trials, except for rare circumstances in which the validity of the surrogate end point has already been rigorously established, the primary end point should be the true clinical outcome.

Thursday, April 8, 1999

### 1998 Charles L. Odoroff Memorial Lecture

**Professor Alan Agresti**

**University of Florida**

*Small Sample Analysis of Categorical Data: Recent Advances and Continuing Controversies*

The Development of methods for "exact" small-sample analyses has been a major advance of the past decade in contingency table analysis. Exact methods guarantee that the size of a test is no greater than some prespecified level and that the coverage probability for a confidence interval is at least the nominal level. A variety of exact methods now exist, both of a conditional and unconditional nature. The great variability in results that can occur with different methods reflects complications due to discreteness. As discreteness increases, exact tests and confidence intervals tend to become overly conservative. We illustrate these issues by studying interval estimation of two basic parameters -- the proportion and the odds ratio. In each case, even for small samples one can argue that "large-sample" solutions are superior to exact ones for many purposes. There will always be an important niche for exact methods, but issues discussed here suggest that statisticians should perhaps reconsider how to evaluate inference procedures.

Thursday, April 16, 1998

### 1997 Charles L. Odoroff Memorial Lecture

**Professor John C. Gower**

**The Open University**

*Algebra, Geometry, and Statistical Graphics*

Thursday, April 24, 1997

### 1996 Charles L. Odoroff Memorial Lecture

**Dr. Mitchell Gail**

**Biostatistics Branch, National Cancer Institute**

*Statistics in Action*

This talk describes two of the most important developments in medical investigation: the adoption of randomization in clinical trials, and the use of statistical ideas, including the case-control design, to establish and investigate the association between smoking and lung cancer. I shall discuss these two developments and two statisticians who contributed enormously to their realization, Austin Bradford Hill and Jerome Cornfield.

Thursday, April 18, 1996

### 1995 Charles L. Odoroff Memorial Lecture

**Professor Nan Laird**

**Department of Biostatistics, Harvard School of Public Health**

*Handing Dropouts in Longitudinal Clinical Trials: Alternative Strategies for Intention-To-Treat Analyses*

Tuesday, April 25, 1995

### 1994 Charles L. Odoroff Memorial Lecture

**Professor Ross Prentice**

**Division of Public Health Sciences, Fred Hutchinson Cancer Research Center**

*A Low Fat Eating Pattern for the Prevention of Cancer and Other Diseases*

Tuesday, April 26, 1994

### 1993 Charles L. Odoroff Memorial Lecture

**Professor Joseph Fleiss**

**Columbia University**

*Clinical Trials Are Not Always Judged On Scientific Groups: The Example of the EC/IC Bypass Study*

Wednesday, March 3, 1993

### 1992 Charles L. Odoroff Memorial Lecture

**Professor Bradley Efron**

**Department of Statistics, Stanford University**

*Bootstrap Confidence Intervals*

Thursday, March 19, 1992

### 1991 Charles Odoroff Memorial Lecture

**Sir David Cox, F.R.S.**

**Warden of Nuffield College, Oxford University**

*Causality: A Review with Statistical and Epidemiological Implications*

A review of various definitions of causality, with special reference to epidemiological applications. Implications for empirical statistical analysis and for clinical trials.

Thursday, May 23, 1991

### 1990 Charles L. Odoroff Memorial Lecture

**Professor Frederick Mosteller**

**School of Public Health, Harvard University**

*Probabilistic Expressions as Quantified by Medical Professionals and Science Writers*

For 20 different studies, we tabulate numerical averages of opinions on quantitative meanings of 52 qualitative probabilistic expressions. Populations with differing occupations, mainly students, physicians, other medical workers, and science writers contributed. In spite of the variety of populations, format of question, instructions, and context, the variation of the averages for most of the expressions was modest.

The paper also reviews studies that show stability of meanings over 20 years, mild effects of translation into other languages, context, small order effects, and effects of scale for reporting on extreme values.

Wednesday, March 28, 1990

### 1989 Charles L. Odoroff Memorial Lecture

**Professor Paul Meier**

**Department of Statistics, University of Chicago, Chicago, Illinois**

*Trials of a Wonder Drug: Does Aspirin Prevent Heart Attacks?*

In the 1960's it was reported that patients who suffered heart attacks were seldom regular takers of aspirin. From that time forward many studies have been devoted to the question of whether there may be a possible therapeutic benefit of aspirin in preventing myocardial infarction. Although the benefit of aspirin prophylactically (i.e., in preventing a new infarction) remains unclear, it has at least been demonstrated that aspirin is beneficial therapeutically (i.e., in acute post-infarct survival).

This example well illustrates the strengths and weaknesses of different styles of clinical evaluation, a topic central to the life and work of Charles Odoroff.

Wednesday, March 8, 1989