Skip to Main Content

YSPH Biostatistics Seminar: “Versatile Deep Learning Provider Profiling: A Design-Based Approach”

November 02, 2023
  • 00:00<v Host>Assistant professor</v>
  • 00:01in the Department of Population Health
  • 00:03and in the Department of Medicine at New York University,
  • 00:06Dr. Wu's research synthesizes state-of-the-art methods
  • 00:09from statistics, machine learning, optimization,
  • 00:11and computational science to address critical
  • 00:14and far reaching issues in health services,
  • 00:16research and clinical practice.
  • 00:18Leveraging large scale data
  • 00:20from national disease registries, administrative databases,
  • 00:24electronic health records, and randomized control trials.
  • 00:27Let's give a warm welcome to Dr. Wu.
  • 00:31<v Dr. Wu>Thank you for the nice introduction.</v>
  • 00:34And it's a great honor to be here with all of you.
  • 00:39And so I'm Wenbo, I am from New York.
  • 00:45I joined NYU just a bit over a year ago.
  • 00:53So I think, 'cause we have so many people here,
  • 00:55I think it would be good to run a promotion first.
  • 00:58(Dr. Wu laughs)
  • 00:59So this is our group.
  • 01:01So at NYU we have,
  • 01:03I mean it's a tremendously growing group
  • 01:07and we have like 24 faculty
  • 01:08and we're about to welcome our newest,
  • 01:12like the 25th faculty member into our divisions.
  • 01:16And we have 7 staff.
  • 01:19We have a small PhD program,
  • 01:22we have 20 PhD students and 10 postdocs.
  • 01:26And we have a team of 25 research scientists.
  • 01:30And part of the reason I wanna do this is
  • 01:33because I wanna encourage you guys
  • 01:36to apply to our PhD programs.
  • 01:38So if you're interested,
  • 01:40scan this QR code and you apply, okay?
  • 01:43All right,
  • 01:45so I have been doing things in provider profiling
  • 01:50for the past for five years
  • 01:53and so this is the overview of what it is.
  • 01:59So provider profiling is basically the assessment,
  • 02:03the evaluation of the performance of healthcare providers.
  • 02:09So I listed here,
  • 02:10could be say acute-care hospitals.
  • 02:14(Wu speaks indistinctly)
  • 02:17This acute-care hospitals, kidney dialysis facilities,
  • 02:21I have been working on other evaluations
  • 02:25like organ procurement organizations,
  • 02:27which is a type of organizations
  • 02:31that are responsible for procuring organs
  • 02:34for patients who are in great need
  • 02:36of organ transplant patients.
  • 02:39And the transplant centers, of course, physician, surgeons.
  • 02:42So you can see,
  • 02:44this includes so many different types
  • 02:46of healthcare providers and stakeholders include,
  • 02:50say, insurance companies, regulation, government,
  • 02:55federal agencies.
  • 02:56They're all interested in provider profiling,
  • 02:59I will tell you why.
  • 03:00Providers is basically who are doing profile evaluations
  • 03:06and of course patients.
  • 03:08So because they are interested in the information,
  • 03:11interested in the profiling results
  • 03:14so they can make care seeking decisions.
  • 03:17Okay?
  • 03:18And so I listed here a few outcomes,
  • 03:22like emergency department encounters,
  • 03:25unplanned re-hospitalizations,
  • 03:29which is hospital readmissions.
  • 03:31And I will jump into the details later
  • 03:34and post-discharge deaths and you can,
  • 03:36I mean there are so many different types of outcomes
  • 03:39to consider in provider profiling.
  • 03:41And one of the goals was
  • 03:43to basically identify those providers
  • 03:47with very bad performance in terms
  • 03:49of patient-centered outcomes.
  • 03:50And they can get penalization,
  • 03:55like they can have payment reductions
  • 03:58from government agencies.
  • 04:01Okay?
  • 04:02And as you can see here, this is very important.
  • 04:04This is a very important business,
  • 04:07and profiling can actually help
  • 04:10improve evidence-based accountability
  • 04:13for those providers and how facility targeted interventions
  • 04:18that aimed at improving the care quality.
  • 04:24Alright, so,
  • 04:36so,
  • 04:40this is a slide of a few example papers
  • 04:43that are about evaluating hospitals across the nations.
  • 04:47So they're mostly from the program called,
  • 04:53Hospital Re-admission Reduction Program,
  • 04:55which is a very important national level program
  • 04:59that I will explain later.
  • 05:01But there are just so many papers in this field.
  • 05:05I mean, these are just,
  • 05:07like there are publications in top,
  • 05:10medical journals, analysts of internal medicine,
  • 05:13and New England Journal of Medicine.
  • 05:24Okay?
  • 05:27So, this is another type of profiling stuff.
  • 05:30So it's called physician profiling.
  • 05:31Basically they wanna evaluate physicians.
  • 05:36So this is, as you can see, it's a report,
  • 05:39it's called the health report
  • 05:41from Massachusetts Medical Society,
  • 05:44which is the publisher
  • 05:46of The New England Journal of Medicine.
  • 05:47Okay?
  • 05:48So they prepared this principles
  • 05:50for profiling physician performance, I think many years ago.
  • 05:56So this is a list of exemplar profiling programs
  • 06:01and they are still existing.
  • 06:04So the first one is an interesting state level program
  • 06:08which is arguably one of the first programs.
  • 06:12So it is still administered
  • 06:19by the New York State of Department of Health.
  • 06:22Basically they're interested in evaluating hospitals
  • 06:25that do coronary artery bypass graft surgeries,
  • 06:31and also PCIs and the program have been running
  • 06:35for at least 20 years or so.
  • 06:39And the second one is another important program,
  • 06:41which was launched I think in 2003.
  • 06:46And it is,
  • 06:49I think it is from the one of the Federal Level Act.
  • 06:52And it is currently administered
  • 06:54by the US Centers for Medicare and Medicaid Services.
  • 06:58And their interest in outcomes for, again,
  • 07:0430-day readmissions and mortality for a AMIs
  • 07:09and the heart failure, et cetera.
  • 07:11And the next one is another federal level readmission,
  • 07:17federal level profiling program,
  • 07:19which is also established by Affordable Care Act,
  • 07:24which is Obama care.
  • 07:25You guys probably know that, in 2012.
  • 07:29And so, yeah, they're also interested in,
  • 07:33evaluating hospitals and they will punish those hospitals
  • 07:37with very bad performance in terms of payment reductions.
  • 07:40Okay?
  • 07:41The last one is an interesting program,
  • 07:43which is kind of my focus.
  • 07:46I have been working on evaluating kidney dialysis facilities
  • 07:54for patients with kidney failure.
  • 07:56And there are actually over 7,000 dialysis facilities
  • 08:01across the nation, believe it or not.
  • 08:04But this is the first to pay for performance program
  • 08:09in contrast to other pay for service programs.
  • 08:14Okay.
  • 08:15And the program is called ESRD.
  • 08:17ESRD is short for End Stage Renal Disease.
  • 08:21Basically the patients with kidney failure,
  • 08:24a quality incentive program, okay?
  • 08:26Alright.
  • 08:27So as you can see, there are so many programs,
  • 08:30so many initiatives across the nation about profiling.
  • 08:36And one natural question is about the,
  • 08:41how the landscape of the statistical landscape
  • 08:44of profiling looks like.
  • 08:46And because of the importance of profiling
  • 08:49and here I said,
  • 08:53there are many far reaching implications
  • 08:55because providers can get penalizations
  • 08:58and it's high stakes.
  • 09:02So it's important
  • 09:03that we have principles statistical methods
  • 09:05to evaluate them, right?
  • 09:08So this is like two examples.
  • 09:11The first,
  • 09:13it's a paper published on analysts of internal medicine,
  • 09:17but it is written by two statisticians.
  • 09:21They are calling for the improvement
  • 09:25of statistical approach in this field.
  • 09:28And also the second one,
  • 09:30this one is even more important
  • 09:32because it is a white paper issued
  • 09:35by the Committee of Presidents of Statistical Society.
  • 09:40You probably know about COPS.
  • 09:42So one of the most important words in the statistic field,
  • 09:47it's the COPS presence of work, right?
  • 09:49So this is a white paper by COPS
  • 09:53and also a group of people from the CMS.
  • 09:58So this is also an important work.
  • 10:00It's about the statistical issues
  • 10:02and assessing hospital performance.
  • 10:05So as you can see,
  • 10:07there are many people are interested
  • 10:10in improving the statistical landscape for profiling.
  • 10:14Alright,
  • 10:15so this is a slight briefly introducing the existing methods
  • 10:23of provider profiling.
  • 10:25There are a few.
  • 10:26I grouped them into like roughly four categories.
  • 10:31So the first group,
  • 10:34is hierarchical random-effects models,
  • 10:38there are many papers in this group,
  • 10:42but I just highlighted one paper in,
  • 10:45I think in 1997 was published on Jassa
  • 10:50by Dr. Sharon Lee Norman at Harvard Medical School.
  • 10:54So it's about hierarchical random-effects models
  • 10:58which is still being used in many settings.
  • 11:02Especially, I mean,
  • 11:04not sure whether you guys know
  • 11:05that there is a group at Yale called Yale Core,
  • 11:08I think Center for Outcomes Research and,
  • 11:14Something. <v ->Evaluation.</v>
  • 11:16<v Dr. Wu>Okay, great, thank you.</v>
  • 11:17So they have been using hierarchical random-effects model
  • 11:21for over 30 years, I guess.
  • 11:24And the second stream of approach is fixed-effects models,
  • 11:31as you can tell from the names,
  • 11:36people are using like a fixed effects in the models.
  • 11:40And this is one example paper,
  • 11:44actually was published in 2013 by my advisors.
  • 11:49And the next one is,
  • 11:53I mean these groups of papers,
  • 11:56they're not mutually exclusive because,
  • 12:00for example, this one,
  • 12:01competing risks or semi-competing risks.
  • 12:04I mean there are some papers
  • 12:06that use higher hierarch random-effects model
  • 12:08or they're also papers using fixed-effects models.
  • 12:12But they are just kind of,
  • 12:13they're handling like different types of outcomes.
  • 12:16So I listened here.
  • 12:18And also for recurring events,
  • 12:20if you take a class in survival analysis,
  • 12:23you probably know that, for example,
  • 12:26patient can have multiple hospitalizations in a year.
  • 12:29So they are considered as recurring events.
  • 12:31Okay.
  • 12:32And then the last one is,
  • 12:35some people are using causal inference
  • 12:37and some clustering approaches to handle profiling issues.
  • 12:44But these papers are relatively new,
  • 12:47and this is one paper here.
  • 12:50It was by all statistics, I think.
  • 12:54Alright, so I wanna discuss a few limitations
  • 12:59of the current landscape,
  • 13:01the current statistical in profiling.
  • 13:05So the first limitation is, people have been, I think,
  • 13:10intensely using models with a linear predictor.
  • 13:15So the limitation is this may not be true
  • 13:19when we have very complex outcome
  • 13:23and the factor associations.
  • 13:25So this is an example.
  • 13:28This figure.
  • 13:30This is in my one of my papers.
  • 13:36So the background,
  • 13:38I'll give you a bit of background information.
  • 13:41So this is about, okay,
  • 13:42evaluating the effect of covid
  • 13:46and the outcome is a 30 day unplanned hospital readmissions.
  • 13:51So this, on the left is the surface plot.
  • 13:54On the right is the conquer plot.
  • 13:56As you can see,
  • 13:59we are interested in the variation
  • 14:02of the covid effect across, this might be too small,
  • 14:06but across post discharge time,
  • 14:09post discharge days and also across calendar days
  • 14:13because we used data in 2020.
  • 14:15So we set time zero at, I think mid-March or,
  • 14:21yeah, mid-March.
  • 14:22So this is April the 1st.
  • 14:25And then May 1st until I think mid-October.
  • 14:30So as you can see there's a lot of variation going on here.
  • 14:36So the covid effect is definitely not constant here.
  • 14:38So basically it means that we cannot use the linear model
  • 14:44to do this.
  • 14:44It's just not valid, right?
  • 14:48So the second methodological limitation is existing methods
  • 14:54have been historically driven by cost effective spending.
  • 14:57Like,
  • 15:01I think in the very first program,
  • 15:03in those first early programs,
  • 15:06people are interested in how to reduce costs
  • 15:10by, of course they wanna improve,
  • 15:13they wanna improve care quality
  • 15:15but cost effectiveness is a very important factor.
  • 15:20So,
  • 15:22and these analysis,
  • 15:22they basically combine all racial ethnic groups together
  • 15:25without accounting for their heterogeneity.
  • 15:31So this is an another example.
  • 15:33So we basically look at the performance
  • 15:37of Organ Procurement Organizations, OPOs.
  • 15:42So we are interested
  • 15:44in organization level transplantation rates.
  • 15:49And we have data in 2020.
  • 15:53So these are,
  • 15:55so on the y-axis we have the normalized OPO IDs,
  • 16:03and this is just like a three panels of caterpillar plots.
  • 16:07And if we focus on a certain OPO, then,
  • 16:12for example, in this panel,
  • 16:13this is a panel for white patients.
  • 16:16And if you look at this is,
  • 16:19I know this is a little bit small,
  • 16:21but this is OPO 30 and this,
  • 16:23the conference interval is above the national rate
  • 16:27for white patients.
  • 16:29So it's significantly better than the national average.
  • 16:32But if you look at the this panel,
  • 16:37this is also OPO 30
  • 16:40and we have the confidence interval being lower
  • 16:44than the national average for black patients.
  • 16:46And this is a panel for Asian Americans
  • 16:52and Pacific Islanders.
  • 16:53We also have the same issue going on here for OPO 30.
  • 16:58So as you can see, there's definitely racial disparity here,
  • 17:04but this was never examined in those early programs.
  • 17:11So this is an limitation of course.
  • 17:15And the last one is,
  • 17:17there is a lack of a unifying framework
  • 17:20to accommodate different provider profiling objectives
  • 17:24and the different performance benchmarks.
  • 17:27I will give you like four different examples.
  • 17:31The first one,
  • 17:34I tried to make the notation very easy.
  • 17:37So say we have a random-effects model here.
  • 17:42We just consider a binary outcome.
  • 17:45Y can be zero or one.
  • 17:47Okay?
  • 17:48And we basically use the logistic regression, here.
  • 17:52So this gamma i, it's a sum of two things.
  • 17:56The first one is mu as the mean effect.
  • 17:58And the second one is ID normally distributed,
  • 18:05a random variable, okay?
  • 18:07And we can construct a type of,
  • 18:08we call it standardized measure.
  • 18:10It's Oi divided by Ei,
  • 18:13O is just a sum of all those YIJs.
  • 18:17And the Ei is the,
  • 18:19basically the sig y function transformation
  • 18:23of mu plus beta.
  • 18:25Okay?
  • 18:27So here, if you look at the model,
  • 18:30we have gamma I here,
  • 18:31but when we calculate the expected number of events
  • 18:35or outcomes, we replace this with the mean.
  • 18:40Okay?
  • 18:41So this is the first example
  • 18:44of course using random effects models.
  • 18:46But if we look at the fixed effects model,
  • 18:49we have the similar formulation here,
  • 18:52but here because this is a fixed-effects model,
  • 18:54gamma I is just unknown fixed effect, okay?
  • 18:58And if we define gamma,
  • 19:02start to be the median of gamma, this is a vector actually.
  • 19:05So it's a vector of vault fixed-effects.
  • 19:08Then this is basically the median of vault provider effects
  • 19:12or fixed effects.
  • 19:14And so we can also construct this standardized measure,
  • 19:17but this time, this E is defined as this,
  • 19:22and this is gamma star.
  • 19:26So we basically use the median of all fixed effects
  • 19:30to construct the standardized measure.
  • 19:33Okay?
  • 19:34So now we have two cases.
  • 19:36One is, okay, we use the, oops,
  • 19:39we use mu, which is the mean of all provider effects,
  • 19:44although it's a random effects model.
  • 19:46And,
  • 19:48here we have median of all fixed provider effects, okay?
  • 19:54So these are two cases,
  • 19:55basically two types of models that have been used before.
  • 19:58And next one is, and some causal papers,
  • 20:04they can use a selected set of provider,
  • 20:09it could be a single provider,
  • 20:11let's say, I'm a a hospital administrator,
  • 20:14I wanna see, okay,
  • 20:15whether my hospital is performing better or worse
  • 20:19than another hospital,
  • 20:21then of course I can use my hospital as the benchmark,
  • 20:25as the reference and compare all other hospital
  • 20:29with my hospital, okay?
  • 20:30So this is the first case.
  • 20:32We can just choose a single hospital or provider
  • 20:36as the benchmark.
  • 20:37And the second case is we can group a few providers,
  • 20:42hospitals in the specific geographic region together
  • 20:45and to form a benchmark, this is also doable, okay?
  • 20:49And it is actually used in the paper.
  • 20:53The last one is, we can basically treat all hospitals,
  • 20:58you can group all hospitals together
  • 21:00into a large super hospital, of course,
  • 21:02this is a hypothetical one but we can do that.
  • 21:06And that is kind of like a national average thing, right?
  • 21:10These are all reasonable ways to define a benchmark.
  • 21:17And there is the last one.
  • 21:19So the last one is kind of more like equity driven thing.
  • 21:23So we can form a benchmark such that say,
  • 21:26okay, say,
  • 21:27from the regulator's perspective,
  • 21:29we really wanna push hospitals to improve their performance
  • 21:34for minority patients.
  • 21:36So say, we can set the benchmark to be something like,
  • 21:41okay, for within the minority groups,
  • 21:43we can intentionally select patients with better outcomes.
  • 21:48We can make the proportion to be very large
  • 21:51so that in the benchmark group,
  • 21:54we can have a very good performance for minority patients.
  • 21:59And then black non-Hispanic patients.
  • 22:03So this is kind of a equity driven thing.
  • 22:06So as you can see, I give you like,
  • 22:11at least the four examples.
  • 22:12But these are scattered in the literature
  • 22:15and there is no unifying framework
  • 22:18to accommodate all of these cases.
  • 22:20But we actually can develop a general framework
  • 22:25to accommodate all.
  • 22:26I will give you the details later.
  • 22:30So, all right,
  • 22:34so the framework
  • 22:36that we proposed is what we termed,
  • 22:40a versatile deep learning provider profiling.
  • 22:43So we proposed a versatile or probabilistic framework
  • 22:50based on the, so-called provider comparators,
  • 22:52which is, you can name it as you know, provider comparator,
  • 22:56hypothetical provider performance benchmark
  • 22:58or population norm.
  • 22:59These are all the same interchangeable terms.
  • 23:03Okay?
  • 23:04Here versatile means, okay,
  • 23:06we can use the framework to do a lot of different things.
  • 23:10So they are adaptable to different profiling objectives
  • 23:14and contexts, okay?
  • 23:15It's why we use the term versatile
  • 23:18and here provider comparator,
  • 23:21which is defined to be a hypothetical reference provider
  • 23:28that is corresponding to your profiling objective.
  • 23:30So if you have a certain objective,
  • 23:32of course you can define your own hypothetical provider.
  • 23:37And if you have a different objective,
  • 23:39you can define another one, okay?
  • 23:42And the deep learning thing comes
  • 23:45into play because it is nice that,
  • 23:49generally it relaxed the linearity assumption
  • 23:51in most existing portfolio models
  • 23:55that relies heavily on linear this assumption.
  • 23:58Okay?
  • 24:00Alright, so this is slide of the basic setup
  • 24:07of this new approach.
  • 24:09So let's say we have a ID random sample
  • 24:13with Y as the outcome,
  • 24:17and the Fi star is the provider identifier,
  • 24:22and Zi is simply a vector of variants,
  • 24:27and they are one from a population Y, F star, Z.
  • 24:38And we have the following assumptions
  • 24:40that these two assumptions, one and two,
  • 24:46so F star.
  • 24:47So basically this script F star is the support
  • 24:52of this provider identifier, F star.
  • 24:56Okay?
  • 24:58So we require that this report for any value
  • 25:05that this F star can pay,
  • 25:07we assume that the probability of F star equal
  • 25:11to F is positive,
  • 25:13which means that in the dataset,
  • 25:15you can at least observe one patient from that provider.
  • 25:19Okay?
  • 25:20Say if this is zero, then basically it means,
  • 25:24okay, we do not observe any patient from that provider,
  • 25:27which is useless, right?
  • 25:31So the second assumption is simply,
  • 25:34okay, so this script F star includes all possible providers,
  • 25:41we wanna evaluate.
  • 25:42So basically this F star has to fall
  • 25:45into this set of values, okay?
  • 25:49So that's why it's the probability as equal to one.
  • 25:52Okay?
  • 25:54So we have two important assumptions,
  • 25:58regarding data generating mechanism.
  • 26:00So the first one is basically the distribution
  • 26:03of this F star.
  • 26:05The provider identifier depends on covariate.
  • 26:10And this is like, okay, so for a patient,
  • 26:14say, I'm a patient, I wanna choose my provider,
  • 26:17I wanna choose my hospital,
  • 26:19my decision will largely based on,
  • 26:21okay, what conditions I have,
  • 26:23and what insurance I have, right?
  • 26:27And say what is the possible feasible set
  • 26:31of hospitals I can choose from?
  • 26:34Okay?
  • 26:34So these are all covariates
  • 26:36that we can include in the model.
  • 26:37So basically the F star is the distribution
  • 26:41of a star depends on all those covariates
  • 26:45which is reasonable assumption.
  • 26:48The second one,
  • 26:49the distribution of the outcome Y
  • 26:51as a function of Z and F star,
  • 26:54which means that, okay, the outcome,
  • 26:57if I go to the hospital and say I have a certain disease
  • 27:03and I got a treatment and whether I feel better
  • 27:08or not really depends on, okay,
  • 27:10of course, depends on my conditions,
  • 27:12and also depends on which hospitals I went to, right?
  • 27:17So the distribution is denoted
  • 27:20as pi, y, given Z and F star.
  • 27:25Okay?
  • 27:26So basically these two assumptions gives us the,
  • 27:31basically the basic setting for a patient who is looking
  • 27:35for care to improve their conditions.
  • 27:42So the main idea in this new framework is reclassification.
  • 27:48So basically,
  • 27:49we wanna construct a hypothetical provider comparator
  • 27:54as a performance benchmark
  • 27:56that is corresponding to our specific profiling objective.
  • 28:01Okay?
  • 28:02So reclassification here means that we wanna,
  • 28:06we reclassify subjects from existing providers
  • 28:11into a hypothetical one
  • 28:13following a certain probability distribution.
  • 28:15Okay?
  • 28:16To do this, we introduced a random indicator,
  • 28:19it's just a 0, 1.
  • 28:21Which we termed reclassifier.
  • 28:24This reclassifier is equal to 0.
  • 28:26Here it is kind of different.
  • 28:28So reclassifier is equal to zero.
  • 28:31When the subject is reclassified
  • 28:33into the hypothetical provider,
  • 28:35if it is equal to one, then the subject is not reclassified.
  • 28:39So the patient stays in their original provider, okay?
  • 28:47And with this reclassified redefined, F,
  • 28:51so F is different from F star.
  • 28:54So F is defined as the product of R,
  • 28:57basically R times F star.
  • 28:59And we basically add a singleton to this F script F star.
  • 29:06So now we can see, okay,
  • 29:09so whatever providers we have originally,
  • 29:13now we add a single hypothetical provider
  • 29:16and we provide the provider indicator,
  • 29:21we fix that as zero.
  • 29:23So zero is the hypothetical one.
  • 29:26So now this F can take values,
  • 29:29importantly, it can take whatever values
  • 29:31from the original script F
  • 29:34but now it can also take values
  • 29:37to take the value zero, right?
  • 29:40So basically this R is used
  • 29:43to manipulate a subject's provider membership.
  • 29:46So, a subject from a provider F star equal to F.
  • 29:54So here in this case,
  • 29:55because it's F star, it cannot be equal to zero, right?
  • 29:59So we wanna reclassify patients
  • 30:01from a certain existing real provider
  • 30:04to that hypothetical provider.
  • 30:07You know, this F is equal to zero.
  • 30:10So this is a new provider membership for that patient, okay?
  • 30:14But if R is equal to zero,
  • 30:16then the patient stays in that original hospital.
  • 30:20Okay?
  • 30:22Alright.
  • 30:23We have additional two assumptions
  • 30:25regarding this reclassification thing.
  • 30:29So the first one is for any provider, real provider,
  • 30:35we have this probability, being less than one.
  • 30:38This means that, okay,
  • 30:40so given a set of covariates and given
  • 30:44that the patient is in a certain provider,
  • 30:49then the patient being reclassified
  • 30:53into the new hypothetical provider,
  • 30:56the probability is less than one,
  • 30:58which means that we should keep at least a few patients
  • 31:03in their original provider
  • 31:05so that we can still evaluate the outcome distributions
  • 31:10of the original provider, okay?
  • 31:13And this actually,
  • 31:15if you do some, a simple algebra,
  • 31:20we can show that basically this implies that,
  • 31:23I mean this, we can basically drop this condition
  • 31:26because if you do the sum
  • 31:28of the conditional probability thing,
  • 31:31you can basically drop this condition
  • 31:33and this actually holds.
  • 31:35So it's like, okay, no matter which hospital,
  • 31:38no matter which provider the patient is in currently,
  • 31:42the probability that the patient
  • 31:44will be reclassified is less than one.
  • 31:46So not all patients will be reclassified, right?
  • 31:50And this is the second condition.
  • 31:52So combining these two, basically, okay,
  • 31:58so basically not all patients can be reclassified
  • 32:02or also all patients cannot be living
  • 32:08in their original providers.
  • 32:10Basically we require that, okay, each patient can,
  • 32:15so we should have
  • 32:17at least a few patients who are remaining
  • 32:20in their original hospitals so that we can evaluate
  • 32:22their original outcome distributions.
  • 32:25And also we need a, of course characterize the distribution,
  • 32:28that hypothetical reference provider.
  • 32:31Okay?
  • 32:33Alright.
  • 32:34Then the last assumption is,
  • 32:38this is kind of an interesting setting.
  • 32:40So rather than observing the original data,
  • 32:44Y, F star, Z, we can only observe this set.
  • 32:52So it's R, Y, F, Z, this tuple.
  • 32:59So the big difference between these two is,
  • 33:02for this one, we know exactly for every patient,
  • 33:05we know exactly where they're from,
  • 33:08which provider they are in.
  • 33:11But for the this one, say if R is equal to 0,
  • 33:17F is automatically 0
  • 33:18because F is defined as R times F star.
  • 33:22So for those patients,
  • 33:23we actually don't know where they come from, right?
  • 33:28But here we assume
  • 33:31that we can only observe post-reclassification data.
  • 33:35And this actually is nice,
  • 33:37I mean this is not always necessary in the practice,
  • 33:42but this assumption actually helps,
  • 33:45facilitates the implementation
  • 33:50of some certain privacy preserving protocols
  • 33:53and data security protocols.
  • 33:54If say, okay, we don't want the,
  • 33:57because of certain powerful influential providers
  • 34:01can actually have a strong influence
  • 34:05in policy making.
  • 34:07So, because this is capped like confidential,
  • 34:11so they actually don't know how we design,
  • 34:14how we choose the re-classification scheme.
  • 34:18So it can help reduce some unwarranted inference
  • 34:25from those very powerful stakeholders.
  • 34:29So this is a nice setting,
  • 34:32but it doesn't have to be like this in reality.
  • 34:36Alright, so now we have four assumptions,
  • 34:41important assumptions
  • 34:42to regarding the data generating mechanism
  • 34:44and to regarding the reclassification scheme.
  • 34:48So, the ultimate goals of profiling is
  • 34:54to first to evaluate all providers,
  • 34:57and then we wanna identify goals,
  • 35:00especially with very bad performance
  • 35:03and we can take additional actions
  • 35:07and so we can, you know,
  • 35:10improve their performance in certain way.
  • 35:12Okay?
  • 35:13But yeah,
  • 35:14so this quantitatively or mathematically,
  • 35:20we have the two overarching goals.
  • 35:22The first one is to harness,
  • 35:24to use the post reclassification data,
  • 35:29to contrast the distribution of each existing
  • 35:35or real provider.
  • 35:37F star was the newly defined reference group.
  • 35:42So we wanna compare,
  • 35:44basically, compare the distribution of these two groups.
  • 35:47I mean each of them
  • 35:48because we have so many real providers,
  • 35:51and we only have a single hypothetical provider, okay?
  • 35:54We wanna compare them, we wanna do contrasts.
  • 35:57And of course the second goal is
  • 35:59to identify those providers with very bad performance.
  • 36:06All right,
  • 36:08so, this actually,
  • 36:14because we introduced this hypothetical provider,
  • 36:17this is really nice actually.
  • 36:19But there is a difficult issue here
  • 36:24because we introduced this hypothetical provider,
  • 36:30we actually have to account for
  • 36:32or address reclassification dues to bias.
  • 36:35So the details are in this proposition.
  • 36:39So let's assume that those four assumptions hold
  • 36:43and the distribution of the outcome given Z and this F,
  • 36:50F is the newly defined provider indicator.
  • 36:54We can actually write the outcome distribution,
  • 36:58like in two cases.
  • 36:59So when F is equal to 0,
  • 37:01this is corresponding to the reference,
  • 37:05the hypothetical provider.
  • 37:07So this is actually the average,
  • 37:14you can consider as the distribution of the outcome
  • 37:21basically for all patient.
  • 37:22If you group all patients together into a single group,
  • 37:25this is basically the distribution of that group.
  • 37:28Okay?
  • 37:29But we have this term here,
  • 37:31and this is not necessarily equal to 1,
  • 37:34F is equal to 1 then it's very simple,
  • 37:38but it could be unequal to 1.
  • 37:43And also in the second case when F is not equal to 0,
  • 37:48which means that okay, for those existing providers,
  • 37:53their distribution also changes because you basically,
  • 37:56you move a few patients to the new provider.
  • 37:59So the original distribution changes, right?
  • 38:02And because we cannot observe this by assumption.
  • 38:06So this is basically the observed outcome distribution
  • 38:10for existing providers.
  • 38:11But according, as you can see here,
  • 38:14it's a bias distribution.
  • 38:15It's no longer the original one, right?
  • 38:17Because this ratio, again,
  • 38:18it is not necessarily equal to 1, okay?
  • 38:22Right?
  • 38:23So as I said,
  • 38:26you can consider this as the average distribution,
  • 38:29basically as the outcome distribution
  • 38:31of the whole patient population, okay?
  • 38:33So of course you can write it as a sum of the,
  • 38:39you know, weighted probabilities.
  • 38:42So the weight being the probability provider membership,
  • 38:46and this is basically, okay, within this certain provider,
  • 38:50what does the outcome distribution look like?
  • 38:53Okay. All right.
  • 38:58So a few things.
  • 39:03This proposition basically outlines a,
  • 39:06what we call design based approach
  • 39:08to provider profiling, basically, okay.
  • 39:12So,
  • 39:14I actually, I mentioned this early,
  • 39:17in profiling there are a few different parties.
  • 39:20The first one is regulars
  • 39:23who initiated the profiling process
  • 39:25because they are interested
  • 39:26in the performance of these providers.
  • 39:28And also we have profilers,
  • 39:31which basically evaluates the performance,
  • 39:33but they don't have to be the same as regulators.
  • 39:36And also we have of course,
  • 39:38providers who are the subject of evaluation
  • 39:42and we also have patient who need the information
  • 39:44to make their decision, okay?
  • 39:46So the design-based approach
  • 39:47basically tell us that, okay, so, for regulators,
  • 39:51they can basically lead the development
  • 39:53of a reclassification scene because in this framework,
  • 39:57we never say what the distribution,
  • 39:59say, what this looks like, where, right?
  • 40:02So this is a very general specification
  • 40:05and we only made that four assumptions,
  • 40:08but we don't have any distributional assumption here.
  • 40:12So we can make it very general.
  • 40:15And so in this framework,
  • 40:19regulators will get more involved in this process.
  • 40:24So that's why they can
  • 40:25basically design the reclassification scheme
  • 40:30based on their specific objectives, okay?
  • 40:35Alright.
  • 40:35So, and given a specific reclassification scheme,
  • 40:41of course they can design their own reference group,
  • 40:45their hypothetical providers
  • 40:47and having defined this hypothetical provider,
  • 40:52profilers of course can use post the reclassification data
  • 41:00and also the dependence.
  • 41:01Because here, as you can see here,
  • 41:03this R actually depends on Y,
  • 41:05depends on the outcome covariate
  • 41:06and the provider identification.
  • 41:10So using this information
  • 41:16and also the post reclassification data,
  • 41:20profilers that can actually do the profiling
  • 41:23and we can use the framework
  • 41:26to estimate the probabilities reclassification,
  • 41:29which is also the propensity scores actually.
  • 41:33So the next step would be
  • 41:39to use the estimated propensity scores
  • 41:41to correct for reclassification induced bias.
  • 41:45And then we can basically construct the distribution
  • 41:51of the hypothetical provider with the distribution
  • 41:55of the existing provider, okay?
  • 42:00Alright.
  • 42:01So as sketched in the previous slide,
  • 42:05there are a few important things
  • 42:07or advantages of the design-based approach.
  • 42:11So this approach actually,
  • 42:13in this framework,
  • 42:14providers can be more involved in this framework.
  • 42:21And,
  • 42:24so we can use the profiling result,
  • 42:28from this new approach can be more relevant
  • 42:31to what people are interested
  • 42:33in the care decision making process, okay?
  • 42:38So, I think I'm a bit over time,
  • 42:43but I wanna quickly skim through a few examples.
  • 42:48But these examples are basically,
  • 42:51we need a few assumptions like
  • 42:54whether the reclassifier is depending on the outcome,
  • 43:00so in this example, it's very simple.
  • 43:02Basically the reclassifier is independent of everything.
  • 43:08So,
  • 43:09actually this reduces to the most simple case.
  • 43:12So nothing changes actually after reclassification,
  • 43:15but this is an example about the setting.
  • 43:21And we also have like a few non-dependent settings.
  • 43:27This R can depend on F star and given F star,
  • 43:32it can be independent with Y.
  • 43:35And we also have some examples,
  • 43:37this is called equal rate representation.
  • 43:39We also have singular representation,
  • 43:42basically the setting
  • 43:43where we only choose a single provider
  • 43:47and we also have the case
  • 43:49where R actually depends on Y, the outcome.
  • 43:54So we can basically choose the outcome,
  • 43:56sorry, we can choose patients based on the outcome.
  • 44:02And I also give an example,
  • 44:04this is actually an interesting example,
  • 44:06but seems like we don't have enough time today.
  • 44:10So this is the most general case where R is allowed
  • 44:13to depend on F and also Y.
  • 44:18So we don't have independence anymore,
  • 44:20but unfortunately this case,
  • 44:22we have the unidentifiability issue.
  • 44:26So this case won't work
  • 44:27under the post-reclassification data assumption.
  • 44:31So we actually developed a framework,
  • 44:38we looked at the deep learning methods
  • 44:41and the singular representation case.
  • 44:44And this is a relatively simple framework.
  • 44:47We only consider exponential distribution.
  • 44:50I mean the outcome
  • 44:51involves the exponential family distribution
  • 44:56and we construct a neural network model.
  • 45:01So we have the input layer
  • 45:02and the fully connected hidden layers and the outcome layer,
  • 45:05and we use stratify sampling based optimization algorithm.
  • 45:11Here, I will skip the detail.
  • 45:14And we developed a exact test based outcome distribution,
  • 45:19exact test based approach to identify outlined performers.
  • 45:25Okay?
  • 45:27And this is basically the motivation
  • 45:29why we need deep learning here, because simply speaking,
  • 45:32the covid effect is not constant over calendar time
  • 45:36and we have to easily account for that
  • 45:39while doing profiling,
  • 45:40but the effect itself is not of interest.
  • 45:48Basically a visualization of the profile results.
  • 45:53So here we construct the,
  • 45:57we construct what we call the funnel plot here.
  • 46:01So the benchmark, the reference, the indicator,
  • 46:06we use is again Oi divided by Ei
  • 46:09and Ei and defined where this one is the median.
  • 46:14And this is actually the neural network part.
  • 46:18And we have the funnel plots.
  • 46:22So those dots represent providers, okay?
  • 46:25So because this, I mean,
  • 46:29the higher, the worse the performance,
  • 46:31the lower, the better the performance.
  • 46:32So these blue dots here are actually better performers.
  • 46:38So as you can see, if you add these two supporters up,
  • 46:42this is like over 20%,
  • 46:47what does not make practical sense
  • 46:48because in practice you cannot identify outliers
  • 46:52with over 20%, you know, this is too much.
  • 46:57So we have to somehow account
  • 46:59for provider level unmeasured confounding.
  • 47:03And I didn't include the technical details here.
  • 47:08But after the adjustment,
  • 47:11as you can see the proportion of a better
  • 47:14and the worse performers are much lower than before.
  • 47:18And I think I only have one more slide.
  • 47:23So some takeaways.
  • 47:25So profiling is very important
  • 47:27as a major societal undertaking in the United States.
  • 47:30And we have so many applications,
  • 47:33important implications and important consequences as well.
  • 47:38And the new framework actually
  • 47:41increased the regulators engagement in this process.
  • 47:45And it's called versatile
  • 47:47because we can handle different profiling objectives
  • 47:49and it is compatible
  • 47:50with many different model specifications,
  • 47:53machine learning models, data science models.
  • 47:55And here we use deep learning
  • 47:58because it relaxes the linearity assumption
  • 48:01and it is often a good idea to account
  • 48:05for provider level measure confounding
  • 48:08when we do this profiling stuff.
  • 48:11And that's all for today.
  • 48:15Thank you so much for.
  • 48:20I know we only have like two-
  • 48:22<v Host>Yeah, We have two minutes.</v>
  • 48:23Thank you very much Dr. Wu for your presentation.
  • 48:26Any questions in the audience?
  • 48:36Anyone online?
  • 48:37Just giving everyone a chance.
  • 48:40No, I'll ask a question.
  • 48:41So I think it's really cool to be able
  • 48:45to identify providers who are doing really well
  • 48:48or doing bad.
  • 48:49What do you do with that?
  • 48:51Now that you have that result?
  • 48:52Like do you tell the profiler
  • 48:54or the patient get to give it to say,
  • 48:56"Oh, I don't wanna go to them, they're bad."
  • 48:58<v Dr. Wu>Yeah, that's a good question.</v>
  • 48:59So actually CMS,
  • 49:02they have many programs say, one is for dialysis patients,
  • 49:07they have dialysis facility compare,
  • 49:10which is an online program.
  • 49:12So patient can have access to different types
  • 49:15of information like whether diet facility is good or bad
  • 49:20and many other different fields
  • 49:24of information they have online.
  • 49:26So they can choose their favorite providers.
  • 49:30Yeah, that's possible.
  • 49:32And it's something that is going on, yeah.
  • 49:36<v Host>Oh, I think we have questions.</v>
  • 49:37<v ->Yep.</v> <v ->Just very briefly,</v>
  • 49:39because I know we're out of time but.
  • 49:43To what extent do you feel that,
  • 49:45if this is true, I guess, and doesn't matter,
  • 49:47the patients don't necessarily have meetings.
  • 49:51So for example, like I grew up in a rural county,
  • 49:55we had one hospital, you were going to a hospital,
  • 49:57you were going there.
  • 49:58Even in New Haven,
  • 50:00there are two campuses of Yale New Haven Hospital,
  • 50:02but there's only one hospital in metro area.
  • 50:06So, I mean, choice is kind of not a real thing.
  • 50:11How does that affect?
  • 50:13<v Dr. Wu>Right, that's a very good point, so-</v>
  • 50:17<v Questioner>We are actually in city,</v>
  • 50:18I understand there's more than one.
  • 50:19(Host laughs) Right, there are so many.
  • 50:21<v Dr. Wu>Yeah, but that's a very good point</v>
  • 50:23because we are actually considering another framework
  • 50:27which is also clustering framework,
  • 50:30which basically gives you
  • 50:32under certain conditions you can choose,
  • 50:34there's a feasible set of providers
  • 50:36that you can choose from,
  • 50:37of course, under certain strengths,
  • 50:39say your insurance, your location, many other conditions.
  • 50:45But I mean, in this framework,
  • 50:49maybe we can address that issue
  • 50:53in the set of areas that we included here.
  • 50:57But yeah, I mean, you know, very important issue.
  • 51:06<v Host>Unfortunately, that's time.</v>
  • 51:06So let's thank Dr. Wu again.
  • 51:12If you haven't signed in, please sign in before you speak.
  • 51:15You are registered.
  • 51:17Oh no, it's good, I don't know.
  • 51:19(indistinct chattering)