JiscMail Logo
Email discussion lists for the UK Education and Research communities

Help for ALLSTAT Archives


ALLSTAT Archives

ALLSTAT Archives


allstat@JISCMAIL.AC.UK


View:

Message:

[

First

|

Previous

|

Next

|

Last

]

By Topic:

[

First

|

Previous

|

Next

|

Last

]

By Author:

[

First

|

Previous

|

Next

|

Last

]

Font:

Proportional Font

LISTSERV Archives

LISTSERV Archives

ALLSTAT Home

ALLSTAT Home

ALLSTAT  2003

ALLSTAT 2003

Options

Subscribe or Unsubscribe

Subscribe or Unsubscribe

Log In

Log In

Get Password

Get Password

Subject:

SHORT COURSE: Intermediate/Advanced Bayesian Hierarchical Modeling

From:

David Draper <[log in to unmask]>

Reply-To:

David Draper <[log in to unmask]>

Date:

Wed, 28 May 2003 12:54:45 -0700

Content-Type:

text/plain

Parts/Attachments:

Parts/Attachments

text/plain (378 lines)

Greetings, and apologies for cross-posting.

If you could please pass this on to anyone you know who may be interested
in it, that would be terrific.

Many thanks and best wishes, David Draper

                              Short Course

Intermediate and Advanced Applications of Bayesian Methods: Bayesian
Hierarchical/Multilevel Modeling

                                Presenter

Professor David Draper
Chair, Department of
  Applied Mathematics and Statistics
Baskin School of Engineering
University of California
1156 High Street
Santa Cruz CA 95064 USA

email [log in to unmask]
web   www.ams.ucsc.edu/~draper
phone US (831) 459 1295, nonUS +1 831 459 1295
fax   US (831) 459 4829, nonUS +1 831 459 4829

                                  Venue

A continuing education (CE) course offered as part of the 2003 Joint
Statistical Meetings (JSM) in San Francisco, CA (6 hours of CE credit
available for participation in the course)

                                  Date

Tuesday 5 Aug 2003, 8.15am - 4.15pm

                                Location

Hotel Nikko San Francisco, 222 Mason Street, San Francisco, CA

                             Course Fee (US$)

Students $200, ASA members $325, non-members $415 (if paid by 18 Jul 2003)

                               To Register

The code for this CE course is CE2003_21C -- you can go to

    www.amstat.org/meetings/jsm/2003/

to register electronically (you don't need to register for the other parts
of the JSM to attend a CE course, but if you're going to the JSM anyway
it's straightforward to include one or more CE courses in the registration
process).

         American Statistical Association (ASA) Contact Information

In case you have difficulties with the electronic registration process, or
if you have any other questions, the ASA may be contacted as follows:

  American Statistical Association
  1429 Duke Street
  Alexandria VA 22314-3415 USA

  phone     703-684-1221
  toll-free 888-231-3473
  fax       703-684-2037
  email     [log in to unmask]

                                  Abstract

This course provides coverage of intermediate and advanced topics arising
in the formulation, fitting, and checking of hierarchical or multilevel
models from the Bayesian point of view.

Hierarchical models (HMs) arise frequently in four main kinds of
applications:

* HMs are common in fields such as health and education, in which
  data -- both outcomes and predictors -- are often gathered in a nested
  or hierarchical fashion: for example, patients within hospitals, or
  students within classrooms within schools.

  HMs are thus also ideally suited to the wide range of applications in
  government and business in which single- or multi-stage cluster samples
  are routinely drawn, and offer a unified approach to the analysis of
  random-effects (variance-components) and mixed models.

* A different kind of nested data arises in meta-analysis in, e.g.,
  medicine and the social sciences.

  In this setting the goal is combining information from a number of
  studies of essentially the same phenomenon, to produce more accurate
  inferences and predictions than those available from any single study.

  Here the data structure is subjects within studies, and as in the
  clustered case above there will often be predictors available at both
  the subject and study levels.

* When individuals -- in medicine, for instance -- are sampled
  cross-sectionally but then studied longitudinally, with outcomes
  observed at multiple time points for each person, a hierarchical data
  structure of the type studied in repeated-measures or growth curve
  analyses arises, with the readings at different time points nested
  within person.

* Hierarchical modeling also provides a natural way to treat issues of
  model selection and model uncertainty with all types of data, not just
  cluster samples or repeated measures outcomes.

  For example, in regression, if the data appear to exhibit residual
  variation that changes with the predictors, you can expand the model
  that assumes constant variation, by embedding it hierarchically in a
  family of models that span a variety of assumptions about residual
  variation.

  In this way, instead of having to choose one of these models and risk
  making the wrong choice, you can work with several models at once,
  weighting them in proportion to their plausibility given the data.

The Bayesian approach is particularly effective in fitting hierarchical
models, because other model-based methods -- principally involving maximum
likelihood -- often do not fully capture all relevant sources of
uncertainty, leading to over-confident decisions and scientific
conclusions.

In this course the basic principles of Bayesian hierarchical modeling will
be reviewed, with emphasis on practical rather than theoretical issues,
and intermediate- and advanced-level ideas will be illustrated with real
data drawn from case studies involving complicated applications of HMs in
cluster sampling, longitudinal data analysis, and mixture modeling.

The course is intended for applied statisticians with an interest in
learning more about intermediate and advanced topics in hierarchical
modeling in general, and the Bayesian analysis of such models in
particular.

An understanding of probability and statistics at the level typically
required for an M.S. degree in statistics provides sufficient mathematical
background.

This course is intended to be a follow-on from an introductory treatment
of Bayesian statistics, so I will assume that participants have background
in Bayesian methods and hierarchical modeling at the level (for instance)
of the first six chapters of the textbook by Gelman et al. (Bayesian Data
Analysis, 1995) or equivalent.

An example of an equivalent background would be for participants to have
completed the basic Bayesian methods short course to be offered by Mike
Escobar at the 2003 JSM, which will be given on Sunday 3 Aug 2003; JSM
participants who want to go into more depth on Bayesian methods could take
the Escobar course first and then my course.

                             Target Audience

The principal target audience includes applied statisticians (1) who work
with data possessing a hierarchical or multilevel character (clustered
and/or longitudinal) on a regular basis, or who wish to do so; (2) who
wish to extend their experience in the contemporary fitting of
random-effects and mixed models, in meta-analysis and other settings; (3)
who wish to improve their skills in formulating and fitting mixture
models; and (4) who wish to learn more about current methods for coping
with problems of model selection and model uncertainty (with all kinds of
data, not just cluster samples).

Application areas in which hierarchical modeling occurs frequently include
policy analysis and other governmental activities, agriculture, medicine
and health, education, and biology.

Others who may be interested in this course include applied and
methodological workers who wish to learn more about (5) comparisons in
complexity and performance between Bayesian and frequentist methods and
(6) Markov Chain Monte Carlo (MCMC) techniques and how to ensure that they
work well in practice.

There are no formal mathematical prerequisites, but a working knowledge of
probability and statistics at the M.S. level (from such books as Hogg and
Craig, Bickel and Doksum, or Casella and Berger) -- particularly the
ability to conceptualize and manipulate conditional probabilities -- will
be desirable, and upper-division undergraduate or M.S.-level study in
linear models will also be helpful.

                       Tentative Syllabus/Outline

8:15 -- 9:00am (review of Bayesian modeling). Probability as
quantification of uncertainty about observables. Prior, posterior, and
predictive distributions for parameters and observables. Comparison with
frequentist modeling. Case studies: Diagnostic screening for HIV,
estimation of a physical constant in a measurement error model.

9:00 -- 9.45am (review of hierarchical models (HMs) for combining
information). Exchangeability as a Bayesian concept parallel to
frequentist independence. Formulating hierarchical models for quantitative
outcomes from scientific context. Appropriate prior distributions for
multilevel models. Case study: Educational meta-analysis of effects of
teacher expectancy on pupil performance.

9.45 -- 10:15am  Break

10:15 -- 11:00am (review of Bayesian computation). Markov Chain Monte
Carlo (MCMC) methods. User-friendly implementation of MCMC via WinBUGS and
MLwiN, the two most widely available Bayesian hierarchical modeling
packages. MCMC diagnostics. Case study: Random-effects Poisson regression
models in a controlled trial of in-home geriatric assessment.

11:00am -- noon (review of diagnostics, model checking, and model
elaboration). Bayesian cross-validation as an approach to diagnostics:
comparing outcomes from omitted cases with their predictive distributions
given the rest of the data. The Deviance Information Criterion (DIC) for
model choice. Expansion of a simple model that does not satisfy all
diagnostic checks, by embedding it in a richer class of models of which
it's a special case. Case study: Continuation of geriatric example.

noon -- 12:45pm  Lunch

12:45 -- 1:45pm (HMs for clustered data). The model-based approach to the
analysis of cluster samples. Random-effects and mixed models from the
Bayesian viewpoint. Case study: Hierarchical profiling of universities by
comparing observed and expected student dropout rates.

1:45 -- 2:45pm (HMs for longitudinal data). Repeated-measures and
growth-curve data analysis. Case study: Tumor growth rates as a function
of covariates in lung cancer investigations.

2:45 -- 3:15pm  Break

3:15 -- 4:15pm (HMs for mixture modeling). Hierarchical modeling with
latent variables as an approach to mixture modeling. Bayesian
nonparametric inference and prediction with Polya trees. Case study: Risk
assessment in nuclear waste disposal.

As time permits, examples from small-area estimation and spatial data
analysis will also be presented.

At several times throughout the day the following themes will also be
emphasized:

Conclusions and caveats; warnings on the unwary use of HMs; several
examples of poor HMs and how to improve them; Bayes does not equal free
lunch; but also: situations in which Bayesian estimates can out-perform
likelihood-based procedures using repeated-sampling criteria.

                Learning Outcomes/Performance Objectives

* Participants will improve their skills at translating scientific and
  decision-making problems involving nested or clustered data into
  appropriate hierarchical models (including random-effects and mixed
  models, repeated-measures models, and mixture models), and will also
  become more proficient -- with any kind of data, not just a cluster
  sample -- at embedding a given model hierarchically in a richer model
  class, as a way to realistically approach issues of model selection and
  model uncertainty;

* Participants will gain a deeper understanding of methods for computing
  posterior and predictive distributions for quantities of interest
  arising in the hierarchical models formulated in (1); and

* Participants will learn new methods for examining the results of the
  model-fitting in (2) for weaknesses and for sensitivity to modeling
  assumptions.

By taking part in this course, participants will develop and/or extend
facility in

* Formulating appropriate hierarchical (random-effects and/or mixed)
  models for nested outcomes (both qualitative and quantitative) in
  meta-analyses, investigations involving cluster samples, longitudinal
  studies, and mixture modeling, and in situations with predictor
  information available at some or all levels of the hierarchy;

* Using Bayesian reasoning and MCMC methods to compute posterior
  distributions for parameters of greatest interest in a given
  hierarchical model;

* Diagnosing problems with a given hierarchical model by looking for
  discrepancies between predictive distributions for observables and the
  actual values the observables take on, and using DIC to guide model
  choice; and

* Hierarchically expanding an existing model (for all kinds of data, not
  just cluster samples) which does not pass all diagnostic checks, by
  embedding it in a richer model class of which it's a special case.

                   Content and Instructional Methods

I will help participants achieve the above learning outcomes by exploring
a variety of case studies -- drawn from education, health policy,
medicine, gerontology, and risk assessment -- with emphasis on the
practical interaction between scientific, decision-making, and statistical
considerations.

Extensive details required for carrying out the analyses will be provided
in the course materials.

I will offer real-time computer demonstrations to participants, using the
two major Bayesian hierarchical modeling packages now most widely
available (WinBUGS and MLwiN), with hardcopy of these real-time sessions
so that participants can recreate them later.

Time permitting, the fitting of hierarchical models in SAS PROC MIXED,
S-PLUS, and Stata will also be addressed.

                           About the Presenter

David Draper is a Professor in, and Chair of, the Department of Applied
Mathematics and Statistics in the Baskin School of Engineering at the
University of California, Santa Cruz (UCSC).

He did his Ph.D. work at the University of California, Berkeley, finishing
in 1981, and he has since taught and done consulting and public policy
research at the University of Chicago (1981-84); the Rand Corporation
(1984-91); the University of California, Los Angeles (1991-93); the
University of Bath, UK (1993-2001); and UCSC (2001-present), with a
sabbatical visit to the University of Washington in 1986.

He is a Fellow of the Royal Statistical Society and has served as
Associate Editor for the Journal of the American Statistical Association
(Theory and Methods, 1988-91; Applications and Case Studies, 1988-94) and
for the Journal of the Royal Statistical Society, Series B (1995-1998).

From 2001 to 2003 he has served and is serving as President-Elect,
President, and Past-President of the International Society for Bayesian
Analysis (ISBA).

His research is in the areas of Bayesian inference and prediction, model
uncertainty and empirical model-building, hierarchical modeling, MCMC
methods, and Bayesian nonparametrics, with applications mainly in health
policy, education, and environmental risk assessment.

He is the author or co-author of 4 books, 4 book chapters and 49 research
papers and invited discussions in leading statistics and applications
journals, including 3 discussion papers on exchangeability, hierarchical
modeling and MCMC in the Journal of Educational and Behavioral Statistics
and the Journal of the Royal Statistical Society (Series A and B).

From 1998 to the present he has served as a statistical advisor to the UK
Higher Education Funding Council for England on the use of hierarchical
models to construct performance indicators (measures of quality) for
British universities, and from 1995 to the present he has served as
principal advisor to 3 Masters students and 5 Ph.D. students whose
dissertations were devoted to aspects of Bayesian hierarchical modeling.

He has been nominated for teaching awards at every university in which he
has taught (e.g., he was the recipient of the Quantrell Award for
Excellence in Undergraduate Teaching at the University of Chicago in
1984), and he has given two extended courses, of six and 18 hours, on
Bayesian statistics and hierarchical modeling at the Universities of Bern
(October 1994) and Neuchatel (June 1995, 1996, 1997) in Switzerland.

In 1997 and 1998 he gave 1- and 2-day courses on Bayesian hierarchical
modeling at the Anaheim and Dallas JSMs; his 1997 course won an ASA
Excellence in Continuing Education award.

He has a particular interest in clearly expositing methodological ideas,
in the context of applied case studies, to diverse audiences.


 =============================================================================

Professor David Draper
Chair, Department of
  Applied Mathematics         web    http://www.ams.ucsc.edu/~draper/
  and Statistics              email  [log in to unmask]
Baskin School of              phone  US (831) 459 1295, nonUS +1 831 459 1295
  Engineering                 fax    US (831) 459 4829, nonUS +1 831 459 4829
University of California
1156 High Street              departmental web pages     www.ams.ucsc.edu
Santa Cruz CA 95064 USA

                 Interesting quotes, number 24 in a series:

     The end is in the beginning; and yet you go on.

       -- Samuel Beckett

 =============================================================================

Top of Message | Previous Page | Permalink

JiscMail Tools


RSS Feeds and Sharing


Advanced Options


Archives

May 2024
April 2024
March 2024
February 2024
January 2024
December 2023
November 2023
October 2023
September 2023
August 2023
July 2023
June 2023
May 2023
April 2023
March 2023
February 2023
January 2023
December 2022
November 2022
October 2022
September 2022
August 2022
July 2022
June 2022
May 2022
April 2022
March 2022
February 2022
January 2022
December 2021
November 2021
October 2021
September 2021
August 2021
July 2021
June 2021
May 2021
April 2021
March 2021
February 2021
January 2021
December 2020
November 2020
October 2020
September 2020
August 2020
July 2020
June 2020
May 2020
April 2020
March 2020
February 2020
January 2020
December 2019
November 2019
October 2019
September 2019
August 2019
July 2019
June 2019
May 2019
April 2019
March 2019
February 2019
January 2019
December 2018
November 2018
October 2018
September 2018
August 2018
July 2018
June 2018
May 2018
April 2018
March 2018
February 2018
January 2018
December 2017
November 2017
October 2017
September 2017
August 2017
July 2017
June 2017
May 2017
April 2017
March 2017
February 2017
January 2017
December 2016
November 2016
October 2016
September 2016
August 2016
July 2016
June 2016
May 2016
April 2016
March 2016
February 2016
January 2016
December 2015
November 2015
October 2015
September 2015
August 2015
July 2015
June 2015
May 2015
April 2015
March 2015
February 2015
January 2015
December 2014
November 2014
October 2014
September 2014
August 2014
July 2014
June 2014
May 2014
April 2014
March 2014
February 2014
January 2014
December 2013
November 2013
October 2013
September 2013
August 2013
July 2013
June 2013
May 2013
April 2013
March 2013
February 2013
January 2013
December 2012
November 2012
October 2012
September 2012
August 2012
July 2012
June 2012
May 2012
April 2012
March 2012
February 2012
January 2012
December 2011
November 2011
October 2011
September 2011
August 2011
July 2011
June 2011
May 2011
April 2011
March 2011
February 2011
January 2011
December 2010
November 2010
October 2010
September 2010
August 2010
July 2010
June 2010
May 2010
April 2010
March 2010
February 2010
January 2010
December 2009
November 2009
October 2009
September 2009
August 2009
July 2009
June 2009
May 2009
April 2009
March 2009
February 2009
January 2009
December 2008
November 2008
October 2008
September 2008
August 2008
July 2008
June 2008
May 2008
April 2008
March 2008
February 2008
January 2008
December 2007
November 2007
October 2007
September 2007
August 2007
July 2007
June 2007
May 2007
April 2007
March 2007
February 2007
January 2007
2006
2005
2004
2003
2002
2001
2000
1999
1998


JiscMail is a Jisc service.

View our service policies at https://www.jiscmail.ac.uk/policyandsecurity/ and Jisc's privacy policy at https://www.jisc.ac.uk/website/privacy-notice

For help and support help@jisc.ac.uk

Secured by F-Secure Anti-Virus CataList Email List Search Powered by the LISTSERV Email List Manager