Skip to Main Content

Yale Psychiatry Grand Rounds: December 16, 2022

December 16, 2022
  • 00:00Um.
  • 00:04Thank you so much Seth and John
  • 00:06for inviting me and hosting me.
  • 00:09Finding out that I'm the last
  • 00:11grand round speaker of the year,
  • 00:12I feel like I'm standing in
  • 00:14between you and your holidays.
  • 00:16Hopefully it doesn't feel that way to you.
  • 00:19But very excited to be here today
  • 00:21and talk about our research.
  • 00:23Let me. Without any.
  • 00:31Further in do we'll get.
  • 00:37We'll get started. It it it is still
  • 00:40early out here on the West Coast,
  • 00:42so you'll have to put up with me being
  • 00:45lit by the glow of my computer monitor.
  • 00:48So. Very excited to be able to
  • 00:50be here and talk about our work,
  • 00:54which is focused on AI broadly.
  • 00:57And we'll unpack exactly what that means
  • 00:59and what we do to try and understand
  • 01:02really at a fundamental level what
  • 01:04happens in counseling and psychotherapy.
  • 01:06How is it that we can capture
  • 01:09some of those active ingredients
  • 01:11using new technology so that we
  • 01:14can understand something about?
  • 01:16What is in the pill of
  • 01:17psychotherapy out in the real world?
  • 01:22Hold on a second.
  • 01:24Before we jump in, I I do want to
  • 01:28acknowledge our research support.
  • 01:30Which has come really in two phases.
  • 01:32There has been foundational
  • 01:34research at the universities.
  • 01:36This is really been about 15 years
  • 01:39of trajectory at this point and that
  • 01:42research is continuing at the startup
  • 01:44that we founded that Seth had mentioned.
  • 01:47And just to disclose,
  • 01:49I am a cofounder and have an
  • 01:52equity stake in that company.
  • 01:55Alright, so where are we headed today?
  • 01:57What am I hoping to cover?
  • 01:59I want to talk a little
  • 02:01bit about the problem.
  • 02:02What is the problem that
  • 02:03we're trying to solve?
  • 02:04And then spend a bit of time at
  • 02:06at at least an intuitive level
  • 02:08trying to provide an understanding
  • 02:11of how the AI technology works.
  • 02:13How is it that we can go from
  • 02:16fundamentally a conversation
  • 02:17to something about the quality
  • 02:20kind of actionable information
  • 02:23fidelity and competence?
  • 02:25Of psychotherapy and counseling.
  • 02:26And then we'll shift and we'll look
  • 02:28at some of the technologies that we
  • 02:30are developing and studying right now.
  • 02:37SO1 slide hitting an issue that we
  • 02:39all are intimately familiar with,
  • 02:41behavioral health problems
  • 02:43are massive and disabling.
  • 02:45This figure of 20% of Americans
  • 02:47is now a couple years old.
  • 02:49I can only imagine,
  • 02:50courtesy of the pandemic,
  • 02:52that that is higher.
  • 02:54If anything,
  • 02:55my colleagues in the Institute for
  • 02:57Health Metrics and evaluation here
  • 02:59at the University of Washington have
  • 03:01been conducting the global burden of
  • 03:03disease study for a number of years.
  • 03:05And one of the findings?
  • 03:06That was provocative when it first came out,
  • 03:09was how large a proportion of psychiatric
  • 03:13and psychological conditions account
  • 03:15for in terms of the global burden of disease.
  • 03:19So again to say something that we all know,
  • 03:22there is a huge need for us to have
  • 03:25effective and widely available treatments.
  • 03:28There are two fundamental problems
  • 03:30that we bump into on a daily basis.
  • 03:34One is access,
  • 03:34and this tends to be front and center,
  • 03:37I think in our minds for good reasons,
  • 03:39this is being.
  • 03:43Exacerbated by the workforce shortage and
  • 03:45how do we train up a new workforce members?
  • 03:50And then the second issue is quality.
  • 03:52And quality can often I think fly under the
  • 03:55radar to a certain extent behind access,
  • 03:58but let me make the case that
  • 04:00it is pretty significant.
  • 04:02That,
  • 04:03as best we can tell,
  • 04:04there's approximately 100 million counseling,
  • 04:07psychotherapy,
  • 04:07behavioral health oriented
  • 04:09intervention conversations each year.
  • 04:12And we basically don't know what happens
  • 04:15in any of them for the most part.
  • 04:18When we do get glimpses into the
  • 04:21actual conversations that are
  • 04:23counseling and psychotherapy,
  • 04:25we can find incredible variability.
  • 04:27And so let me unpack what
  • 04:29we're looking at here.
  • 04:30My colleague John Bayer,
  • 04:32who's at the VA Puget Sound
  • 04:34on several years ago,
  • 04:36did a motivational
  • 04:38interviewing training study,
  • 04:39partnering with a variety
  • 04:41of community agencies,
  • 04:42substance use agencies here
  • 04:44in the Puget Sound region.
  • 04:46As part of that,
  • 04:48we created 200 providers,
  • 04:49got six recordings of
  • 04:52their actual counseling,
  • 04:53and then spent the time effort resources
  • 04:57to actually fidelity code those.
  • 05:00And so this is just the empathy rating
  • 05:02that comes out of the motivational
  • 05:04interviewing treatment integrity system,
  • 05:06their fidelity coding system.
  • 05:07And so each one of those purple
  • 05:10dots is the average empathy
  • 05:12for each one of 200 providers,
  • 05:14most of whom have about 6 sessions.
  • 05:16Going into that,
  • 05:18so based on what we know from this scale,
  • 05:21we have here at the upper end
  • 05:23of this scale some superstars,
  • 05:25these are clinicians that are
  • 05:28out practicing in substance use
  • 05:30settings who are demonstrating
  • 05:31a deep understanding of their
  • 05:33clients worldview and capturing
  • 05:35something about the meaning of what
  • 05:37is being said by their clients.
  • 05:39On the other hand,
  • 05:41down here we have a group of no
  • 05:43stars I guess who are demonstrating
  • 05:45no interest in their clients
  • 05:47worldview and little to no attention
  • 05:50to what the client says.
  • 05:51And if you are like me when
  • 05:53I first saw this thinking,
  • 05:54how is that possible?
  • 05:56If these are actually behavioral health
  • 05:58providers, these sessions tend to
  • 06:00sound a little bit like Doctor Phil.
  • 06:03You know, when are you going to get real?
  • 06:04When are you going to come to grips with
  • 06:06the fact that substance use is ruining
  • 06:08your life and it's causing problems?
  • 06:09And so there really is a sense that
  • 06:11they are not listening to the client.
  • 06:12They have a message that
  • 06:14they're hammering home.
  • 06:15And based on what we know of
  • 06:18the very broad literature on the
  • 06:20association of empathy and outcomes,
  • 06:23this is really toxic treatment.
  • 06:26But the main point that I'm trying
  • 06:28to make here is not that there's
  • 06:30some toxic treatment happening,
  • 06:31but just that there is incredible
  • 06:33variability and we don't have
  • 06:35any line of sight into this.
  • 06:36This is data that has only come out
  • 06:39of a well funded NIH research study.
  • 06:42And so in those well funded
  • 06:44NIH research studies,
  • 06:45we use the traditional evaluation method
  • 06:48for counseling and psychotherapy,
  • 06:50which is behavioral coding or human coding.
  • 06:52We don't usually do it in real
  • 06:55time like this picture.
  • 06:56We record sessions and then have
  • 06:58a team of experts evaluate them.
  • 07:00But it is slow and it is expensive and
  • 07:03it really is not used in the real world
  • 07:06outside of well funded NIH research studies.
  • 07:09If we had AI or or some alternative
  • 07:13rapid means of assessing quality,
  • 07:15there's a variety of uses that we
  • 07:18could use it for performance based
  • 07:22feedback and training for supervision.
  • 07:25One of the best ways we'll we'll
  • 07:27look at an example of using this
  • 07:29in a in a few minutes,
  • 07:30but one of the best ways to learn
  • 07:32is to get specific feedback on
  • 07:34new skills that you're learning.
  • 07:38Similar to the image,
  • 07:40the figure that we just saw.
  • 07:43Having some means of quality and
  • 07:45sure quality assurance or quality
  • 07:47improvement within service delivery.
  • 07:51We think a little more commercially,
  • 07:54payers are writing checks for services
  • 07:56of unknown quality at this point.
  • 07:58So payers could have something to
  • 08:01know what they're paying for and
  • 08:03could potentially be the basis for a
  • 08:05type of value based care arrangement.
  • 08:10Finally, there are in in many ways we
  • 08:14still don't know exactly how it is.
  • 08:17The conversations of of counseling and
  • 08:19psychotherapy lead to behavior change.
  • 08:22And so being able to open the black box
  • 08:25in this sense could help us understand
  • 08:27how is it that those conversations,
  • 08:30those intimate engagements that
  • 08:32we have with our clients lead
  • 08:35to sustained behavior change.
  • 08:39So let, let's tip into.
  • 08:41So over the last 15 years, I have been,
  • 08:45I am a clinical psychologist by background.
  • 08:47I do have an interest in data science,
  • 08:50but this work has really been
  • 08:53enabled by deep and sustained
  • 08:56collaborations with technical
  • 08:58experts across machine learning,
  • 09:01natural language processing,
  • 09:03speech engineering.
  • 09:05And so we'll we'll dip into.
  • 09:08Within this next section,
  • 09:10I'm hoping to give kind of
  • 09:11an intuitive understanding
  • 09:13of how the technologies work,
  • 09:15and if we want to dip into more details,
  • 09:17happy to do that in the Q&A.
  • 09:22So let me start with just thinking
  • 09:24through at at a really basic if we were
  • 09:27to to trying to describe to a layperson,
  • 09:30you know, what are the raw data,
  • 09:32what are the basic building blocks that
  • 09:35go into psychotherapy or counseling.
  • 09:38So first and foremost it's words,
  • 09:40whether that is you know
  • 09:42increasingly now telehealth.
  • 09:43But it could be in person,
  • 09:45it could be telehealth,
  • 09:46could be on a phone, could be video,
  • 09:48could be text or chat based text interaction.
  • 09:51But it is a conversation,
  • 09:53so words are one of the basic ingredients.
  • 09:57For everything except text based chat,
  • 10:02interaction, tone and other types
  • 10:05of paralinguistic information.
  • 10:07So there's tone prosody things like
  • 10:10linguistic disfluencies when someone,
  • 10:12someone,
  • 10:13someone maybe perseverates on a certain word,
  • 10:16and that's indicative of cognitive load.
  • 10:18So all of the types of information
  • 10:21about how something is said versus just
  • 10:24what is said in the words is important.
  • 10:27If it's in person or video,
  • 10:29there's different types of
  • 10:31nonverbal information.
  • 10:31That could be posture,
  • 10:33facial emotion,
  • 10:34gestures.
  • 10:37And then finally, it's not simply
  • 10:39each one of these components,
  • 10:41but the dynamic way in which they
  • 10:43unfold in the interaction itself.
  • 10:46So the we can't just take
  • 10:47a statement such as, wow,
  • 10:49I cannot imagine how difficult
  • 10:51that must have been to lose your
  • 10:53kids and on its own say whether
  • 10:55that is an appropriate or good
  • 10:57high quality intervention.
  • 10:59We need to know what's the context,
  • 11:00how would, where is that being said?
  • 11:04So then let's switch over and think about.
  • 11:06I mean the challenge of measuring
  • 11:09counseling and psychotherapy is
  • 11:11fundamentally that it is a conversation.
  • 11:14It's very unstructured.
  • 11:15We think about other types
  • 11:17of data that we collect,
  • 11:19whether that's lab tests or
  • 11:22questionnaire data like the PHQ 9.
  • 11:25There is inherently nothing
  • 11:27numeric about a conversation.
  • 11:29So how do we actually
  • 11:31quantify this information?
  • 11:35Historically, natural language
  • 11:37processing use what's called ngrams,
  • 11:40and that's really a fancy way of saying.
  • 11:44Quite literally.
  • 11:45They would dummy code create indicator
  • 11:48variables for unique words, for vocabulary,
  • 11:51or for short common phrases.
  • 11:53So two word and three word phrases.
  • 11:57And if you're thinking like,
  • 11:58how is that even possible?
  • 12:00If you have a basic understanding
  • 12:01of a regression model,
  • 12:02wouldn't that mean that there are thousands
  • 12:05and thousands of predictors in these models?
  • 12:07Yes, that that is exactly right.
  • 12:10That is a good intuition.
  • 12:12Increasingly, these models are now
  • 12:14using something called word embeddings,
  • 12:17and this is the idea that
  • 12:19when we see a word there,
  • 12:21it has a certain meaning.
  • 12:23And and these word embeddings are ways of
  • 12:26trying to get at a meaning so that it's
  • 12:29not just a word with an indicator of binary.
  • 12:33Yes, no, this word showed up, but it's
  • 12:35implying something about the meaning.
  • 12:37I won't try to go in to explain that, though.
  • 12:40Happy to go into the weeds if
  • 12:42that were of interest later.
  • 12:43But so there's there.
  • 12:45There are ways of quantifying words.
  • 12:48Similarly,
  • 12:49there are a variety of speech
  • 12:52signal processing methods so that
  • 12:54we can estimate acoustic features.
  • 12:57So these are things like tone,
  • 13:00the vocal arousal that we can hear,
  • 13:03and in someone's voice.
  • 13:04When someone is excited and the pitch
  • 13:07or the tone of their voice goes up,
  • 13:08we can measure that reliably.
  • 13:11And in addition,
  • 13:12we can also measure something called jitter.
  • 13:15This is when someone is really upset
  • 13:16and we say that their voice is shaking.
  • 13:19That is the very extreme form of jitter,
  • 13:21but we can measure that over a broad range.
  • 13:23So again,
  • 13:24point being that there are speech
  • 13:27signal processing methods for
  • 13:30quantifying these types of information.
  • 13:33Similarly,
  • 13:33there is an area of machine learning
  • 13:36called computer vision that is the
  • 13:39reason if you use Google Photos and
  • 13:41you can search Google Photos for
  • 13:43finding your dog or what have you,
  • 13:45that is computer vision is the
  • 13:48AI engine that's enabling that.
  • 13:50Although our team has some expertise in that,
  • 13:52that's not a focus of our current
  • 13:55work as we have felt that language
  • 13:57and words is really the lowest common
  • 14:00denominator across all the different.
  • 14:03Medium of counseling and psychotherapy.
  • 14:07Finally,
  • 14:08there are a variety of techniques
  • 14:10both machine learning and natural
  • 14:12language processing,
  • 14:13but also outside of that dynamic
  • 14:16systems models for understanding how
  • 14:19is it that interactions unfold over time.
  • 14:22So just a couple again,
  • 14:24my goal here is really to provide
  • 14:26some sense for what is this process
  • 14:28and how does this work and so let's
  • 14:30take an example of 1 specific example.
  • 14:32So our work started.
  • 14:35Is motivational interviewing.
  • 14:37In hindsight,
  • 14:37that seemed like an incredibly wise choice.
  • 14:39The reality was, of course,
  • 14:41it was a bit happenstance.
  • 14:43Was collaborating with some
  • 14:45colleagues who were using MRI and
  • 14:48had recordings from some RCT's,
  • 14:50that that was the very first
  • 14:52grant in this work.
  • 14:54But MI is a fantastic place
  • 14:56to start because it
  • 14:58is very linguistic focused.
  • 15:01So am I is interested in things like.
  • 15:05Or is a clinician asking an open-ended
  • 15:08question versus a close ended question,
  • 15:10so really tightly tied to the language itself
  • 15:14in a way that is fundamentally different
  • 15:16than say cognitive behavioral therapy where
  • 15:19they're interested in assessing how well
  • 15:21the clinician set in the agenda. And yes,
  • 15:23there we we can know that from the words,
  • 15:27but it's at a kind of higher level,
  • 15:29it's more of a psychological
  • 15:30construct that's in the words.
  • 15:32So am I was a great place
  • 15:34for us to start this work?
  • 15:36And so we have this brief little
  • 15:38snippet of transcript here.
  • 15:40You know, client says,
  • 15:41I wouldn't mind coming here for treatment,
  • 15:43but I don't want to go to one of
  • 15:44those places where everyone sits
  • 15:46around crying and complaining all day.
  • 15:48The counselor says you don't want to do that,
  • 15:50so you're kind of wondering
  • 15:51what it would be like here.
  • 15:54So this was a an example that we used
  • 15:57in in one of our early research papers,
  • 16:00where the goal was can we
  • 16:03automatically identify when a
  • 16:04therapist is making reflections,
  • 16:07when they're providing a brief summary
  • 16:10and reflecting back to the client whether
  • 16:14they are understanding them correctly?
  • 16:17And so let's just use this to unpack.
  • 16:20How do we actually go from a transcript
  • 16:24of words to a predictive model?
  • 16:27So again, as we talked about, um,
  • 16:29one of the traditional ways is that
  • 16:31we would use what's called Ngram
  • 16:34features where literally it's just
  • 16:36identifying there's particular
  • 16:38words that are in this statement.
  • 16:40And also common two or three word phrases.
  • 16:44So again we would basically be quantifying
  • 16:48those in types of indicator variables.
  • 16:52This is also a conversation,
  • 16:53so something is happening over
  • 16:55time and so the local context,
  • 16:57especially for trying to understand
  • 16:59something like a reflection,
  • 17:00which is inherently something being said
  • 17:03back in response to a previous statement.
  • 17:07So something about the context so
  • 17:09we can look at words in the local
  • 17:12context before or after.
  • 17:13There's also a little bit of metadata,
  • 17:16and when we use metadata in
  • 17:19natural language processing,
  • 17:21it refers to anything that is
  • 17:23not the words themselves.
  • 17:25And in this case,
  • 17:26what we minimally know is that
  • 17:28there are two different speakers
  • 17:30and they have different roles.
  • 17:32So is this the client or is this
  • 17:34the therapist who's speaking?
  • 17:38Finally, we can create other types
  • 17:40of features to include as predictors,
  • 17:43and because a reflection is inherently
  • 17:46capturing something about the,
  • 17:48there should be some similarity
  • 17:49with what the client has just said.
  • 17:51We can identify other types
  • 17:53of similarity features,
  • 17:54whether those are parts of speech
  • 17:56such as are using an adverb or a
  • 17:59direct match in terms of words.
  • 18:01So all of these would be ways
  • 18:04that we could quantify and and,
  • 18:06you know, in a statistic sense.
  • 18:08Create a a set of predictors
  • 18:11for our prediction equation,
  • 18:12trying to identify a reflection.
  • 18:17One of the other things that as
  • 18:20I got into this work and began
  • 18:23to collaborate with computer
  • 18:24scientists and speech engineers is.
  • 18:27I do enjoy data science and there was
  • 18:30a period early in my career where I
  • 18:32was a lot of my time was being spent
  • 18:35really as an applied biostatistician.
  • 18:37As we got into this work,
  • 18:39I quickly realized there's a whole
  • 18:41set of models and methodologies
  • 18:44that I was never exposed to.
  • 18:47And so part of the inter interdisciplinary
  • 18:50work is really that translation,
  • 18:52being able to form a foundation of knowledge,
  • 18:56both clinical knowledge,
  • 18:57so the computer scientist and the speech
  • 19:00engineers needed to learn something
  • 19:02about motivational interviewing,
  • 19:04but also for the rest of the team
  • 19:06to understand something about the
  • 19:07models that are being applied.
  • 19:09And so if you look at these different models,
  • 19:13latent Dirichlet allocation,
  • 19:14conditional random field,
  • 19:16recursive neural networks.
  • 19:18And think I have never heard of any of
  • 19:20those and I at least had some stats.
  • 19:21You are not alone and that's
  • 19:24something about the the work here
  • 19:26able to bridge those gaps.
  • 19:30So the initial phases of our research
  • 19:34focused strongly on this idea of can we
  • 19:37go from the word spoken in a session
  • 19:41to reliably estimating fidelity codes.
  • 19:46And we over the course of about 8 to 10
  • 19:49years, we have probably 25 or 30 publications
  • 19:52capturing different aspects of this,
  • 19:54looking at words themselves,
  • 19:56looking at paralinguistic information
  • 19:58tone itself, combining them,
  • 20:00different types of models.
  • 20:02And and let me,
  • 20:03so let me show you a couple results and let
  • 20:06me tell you what we're looking at here.
  • 20:08So the traditional method for
  • 20:11estimating a fidelity code is to
  • 20:14have a team of raters learn a
  • 20:16well validated clinical system.
  • 20:18So here the motivational interviewing
  • 20:21treatment integrity system or the
  • 20:23motivational interviewing skills code
  • 20:25system and then they make their ratings.
  • 20:29But even well trained humans do not
  • 20:32agree with each other perfectly.
  • 20:34So we call that inter rater reliability
  • 20:37and so that's an important piece.
  • 20:39For training a computer to be
  • 20:41able to do this,
  • 20:43which is that interrater reliability
  • 20:45of functionally sets a ceiling for us.
  • 20:48And so the goal here is really can we
  • 20:51develop an AI algorithm that will be as
  • 20:54accurate as the most accurate human.
  • 20:57And So what we're looking at there on
  • 21:00the X axis each one of these labels,
  • 21:02advice, giving, affirmation,
  • 21:04confront these are specific fidelity
  • 21:06codes within the motivational
  • 21:09interviewing system.
  • 21:10Either things that you should do,
  • 21:12such as asking open questions
  • 21:14and making reflections,
  • 21:15or things that you should not do,
  • 21:18such as confronting your client
  • 21:20or giving up giving them advice.
  • 21:22And that X axis is asking,
  • 21:25out of the reliability of the human raters,
  • 21:28how reliable is the computer estimate,
  • 21:31and so at 100% the computer
  • 21:36is estimating providing.
  • 21:38Fidelity codes that are identical
  • 21:40to our most reliable human coders,
  • 21:44and so we can see that over time
  • 21:46this is not where we started,
  • 21:47but over time we have been able to
  • 21:50develop AI algorithms that would
  • 21:52start with a recording and generate
  • 21:55codes that are highly reliable and
  • 21:57very similar to expert human coders.
  • 22:00The the one other thing that I'll
  • 22:02mention here is that this graph
  • 22:05and all of our results use the.
  • 22:08Additional methods within machine
  • 22:09learning to evaluate models,
  • 22:11which is we take the whole data and
  • 22:13we cut it up into a couple pieces
  • 22:16and there's a set of data that
  • 22:18we use to train models.
  • 22:19And then when we are completely done
  • 22:21with the training of those models,
  • 22:23then there is a separate piece of
  • 22:25data that they never saw in training.
  • 22:27That is our test set or evaluation set.
  • 22:29And so these numbers and every time
  • 22:32we evaluate them and come out of
  • 22:35a test set which and I hope and.
  • 22:38In our Q&A,
  • 22:39we can get into a little bit of
  • 22:41the conversation around what is the
  • 22:43data that trains models and then
  • 22:45where is that model being applied,
  • 22:47because potential AI bias is
  • 22:50inherent in those types of
  • 22:52questions. OK. So that was for
  • 22:55motivational interviewing.
  • 22:56More recently, we have done at a parallel
  • 22:59set of work with cognitive behavioral
  • 23:02therapy that's focused on the CTRS.
  • 23:05And similarly over time,
  • 23:06this was not the where we started,
  • 23:09but over time we've been able to
  • 23:12develop models that reliably replicate
  • 23:14what human experts will do in
  • 23:17terms of generating fidelity codes.
  • 23:22And just to highlight another aspect
  • 23:25of this work, which is I I have
  • 23:28primarily been talking about this
  • 23:31one slice around prediction models,
  • 23:33but the reality is the the entire,
  • 23:36what we call the pipeline starts with
  • 23:40a recording or that spoken language
  • 23:44and there's an incredibly important
  • 23:46and perhaps the most complicated part
  • 23:49of what we do is that a speech signal.
  • 23:52Processing tasks.
  • 23:53So from a recording,
  • 23:55can you tease apart uniquely the
  • 24:00multiple different speakers and can
  • 24:02you identify automatically who the
  • 24:04therapist is and who the client is?
  • 24:06And can you then generate a
  • 24:08highly reliable speech to text,
  • 24:10transcript and and so I never
  • 24:13would have imagined,
  • 24:15as I was getting my PhD
  • 24:17in clinical psychology,
  • 24:18that I would be collaborating
  • 24:20at points on methods for.
  • 24:22Lattice scoring in speech
  • 24:24to text transcription,
  • 24:25but that has been part of the work,
  • 24:28is that to help move this forward I have
  • 24:32needed to move into technical areas.
  • 24:34And my technical colleagues have actually
  • 24:37gone to motivational interviewing
  • 24:38workshops taught by Bill Miller.
  • 24:40And I think that's part of the
  • 24:42magic that has made this work,
  • 24:43is that you have a collaborative
  • 24:45team that's really willing to get
  • 24:47outside of their comfort zone pretty
  • 24:49dramatically in certain cases.
  • 24:53All right. Just to give a snapshot,
  • 24:56so the that AI pipeline that we
  • 24:59were just talking about at the
  • 25:01moment generates around 54 metrics.
  • 25:04And so as we saw for both CBT and
  • 25:06for motivational interviewing,
  • 25:09we generate gold standard fidelity metrics.
  • 25:12These are not systems that we made-up,
  • 25:15but we went to the literature and said,
  • 25:17OK, CBT researchers,
  • 25:18am I clinical developers, what all,
  • 25:21what is the gold standard?
  • 25:23And there's also some other
  • 25:24things that we have kind of baked
  • 25:27into the pipeline over time.
  • 25:29So there are some content codes,
  • 25:31so we can have the goal here was
  • 25:33really to provide a line of sight
  • 25:35into what's this conversation about?
  • 25:37Is it about at a high level,
  • 25:40is it assessment or therapy
  • 25:42or case management?
  • 25:44And then what's the focus
  • 25:45of the conversation?
  • 25:46Is it about mood problems or
  • 25:49trauma or suicide work problems,
  • 25:51intimate partner problems?
  • 25:53So we can capture something about
  • 25:56really what's the conversation
  • 25:58about and then I'll we'll talk a
  • 26:00little bit about this at the end,
  • 26:01but we have some exciting
  • 26:04developments in particular around
  • 26:06suicide risk assessment.
  • 26:07And automatically identifying emotions.
  • 26:12OK. So we, I had started off a couple
  • 26:15minutes ago showing a graph that looks
  • 26:18a lot like this that was courtesy
  • 26:20of my colleague John Bair to give
  • 26:23a sense for the power of the AI.
  • 26:26So that was approximately 900
  • 26:29sessions that took John and his
  • 26:32team about a year to generate.
  • 26:35Now that we have moved this
  • 26:37technology outside of the university,
  • 26:39we have an opportunity to work with.
  • 26:42And partnerships customers and
  • 26:46in partnering with a large
  • 26:50digital telehealth company,
  • 26:52we got access to a a million sessions
  • 26:56on more than 5000 providers.
  • 26:59And so here it's hard to see but
  • 27:01there are actually 5000 purple dots.
  • 27:04And there were summarizing
  • 27:06approximately 1,000,000 sessions
  • 27:07on that same empathy scale.
  • 27:10We see a similar pattern here,
  • 27:12but in addition because we were
  • 27:15partnering with a a real World
  • 27:18Service delivery provider,
  • 27:20we also got access to some KPI's.
  • 27:23And so here we can say not only
  • 27:26something about the empathy but
  • 27:28we can also look at if you saw
  • 27:31providers who are highly empathic.
  • 27:34Turns out that your clients
  • 27:36were much more satisfied.
  • 27:38That's Net Promoter score,
  • 27:40which is functionally a zero to
  • 27:4210 score of how satisfied you are.
  • 27:45You're much more satisfied than if
  • 27:47you saw folks who are less empathic.
  • 27:50It's not rocket science.
  • 27:51If someone is really good at
  • 27:53paying attention to you and trying
  • 27:55to understand your worldview,
  • 27:57it's not surprising that their
  • 27:59clients will be more satisfied.
  • 28:00But we also see this with other effects.
  • 28:05So this what we're calling
  • 28:07active listening here, is,
  • 28:08for those of you who know,
  • 28:10motivational interviewing.
  • 28:11It is a summary of ORS,
  • 28:14how out of all of the language
  • 28:17from the therapist,
  • 28:18how much of that is open-ended questions,
  • 28:22affirmations, reflections and summaries.
  • 28:25And again,
  • 28:27it is those really low level
  • 28:30micro counseling skills.
  • 28:31And what we see is that if you see
  • 28:34a therapist who is really good at
  • 28:37listening at those kind of basic
  • 28:40active listening skills.
  • 28:41You are much more likely to get what
  • 28:43we would consider a full dose of treatment.
  • 28:46Conversely,
  • 28:47if you see someone who does not do
  • 28:49that and is probably giving advice,
  • 28:52giving lots of information and
  • 28:55maybe even confronting,
  • 28:57then you don't stay around
  • 28:58in treatment very long.
  • 29:00So the the key point here is that
  • 29:02both there is a level of validation
  • 29:04that's coming as this gets out
  • 29:07into large real-world data.
  • 29:08It's also demonstrating that these
  • 29:11tools can to a certain extent begin
  • 29:14to open up the the black box that
  • 29:17is therapy in the real world and
  • 29:20provide a line of sight and some
  • 29:22reliable indicators of what's happening.
  • 29:25So let's now kind of shift gears
  • 29:27a bit and we'll look at some
  • 29:30of the specific technologies.
  • 29:32So really everything that I've
  • 29:34talked about thus far is kind of
  • 29:37describing the engine of the car.
  • 29:39And then now let's look at some
  • 29:41different technologies and ways
  • 29:43that it's getting deployed and
  • 29:45some related research as
  • 29:47well.
  • 29:49And so kind of just repeating what I said,
  • 29:52that that university based research
  • 29:54really laid the kind of AI engine,
  • 29:57the AI foundation that now is getting
  • 29:59evaluated in a variety of technologies.
  • 30:05So let's take a look at how we
  • 30:08might use this for training.
  • 30:11One of the things that I love about
  • 30:12Bill Miller and the MIT community is
  • 30:15that they're incredible empiricists.
  • 30:16And so Bill was one of the first people
  • 30:19to do fairly rigorous training studies.
  • 30:22And what he found is that the ways
  • 30:25that we traditionally do training,
  • 30:27which is 1/2 day or a full day
  • 30:31workshop don't actually work very well.
  • 30:33And when when I say they don't work
  • 30:35very well, I mean they don't have
  • 30:38durable effects on provider behavior.
  • 30:40And so he, he is raising this
  • 30:43question of how,
  • 30:44how is it then that we could train to
  • 30:47get broader and more durable effects?
  • 30:50This is particularly critical right
  • 30:52now because we have such a workforce
  • 30:56shortage and so being able to train
  • 30:59rapidly and highly effective ways
  • 31:02it would be incredibly useful.
  • 31:05Fortunately,
  • 31:05we haven't yet really moved past
  • 31:09our traditional methods.
  • 31:11We now might do them online,
  • 31:14but our professional training
  • 31:15often includes a lot of content
  • 31:18and so that's either slides and
  • 31:20written content will have lectures.
  • 31:22You can see some examples which
  • 31:25would be good,
  • 31:26but in terms of practice probably limited
  • 31:29to role plays and that's probably a very,
  • 31:32very small part part of the training.
  • 31:35And in terms of assessment or
  • 31:37demonstrating what has been learned,
  • 31:38we're probably limited to some
  • 31:41some type of knowledge quiz.
  • 31:43And so at the end of traditional training,
  • 31:47oftentimes what we can measure are
  • 31:49things like what content was offered,
  • 31:52how many providers access the content,
  • 31:54to what degree, and again,
  • 31:57maybe something about the demonstration
  • 31:59of of knowledge or attitudes.
  • 32:01But critically,
  • 32:02we have not had good ways of actually
  • 32:06assessing skills or skill development.
  • 32:09If we look at other skills,
  • 32:12and I think we can make a very strong case
  • 32:15that counseling and psychotherapy is a skill,
  • 32:18behavioral skill,
  • 32:19much like other types of
  • 32:21skills that we would learn.
  • 32:22A really key component is practice,
  • 32:25lots of practice, ideally with feedback.
  • 32:29And this is some of the ways in
  • 32:31which I think things break down,
  • 32:33which is if you are learning to play tennis,
  • 32:36you know whether you're shot,
  • 32:37went in or out.
  • 32:38And I don't know that that that there
  • 32:41is a strong parallel at that level.
  • 32:43So feedback often is going to
  • 32:46require a coach or outside input.
  • 32:49The other thing that's interesting and
  • 32:51has motivated some of our training work
  • 32:54is that we practice small components.
  • 32:57So professional musicians
  • 32:58still practice scales.
  • 33:00Professional athletes will practice
  • 33:02their backhand over and over.
  • 33:05And it's it's interesting to to think about,
  • 33:07as far as I know,
  • 33:09we don't do that within
  • 33:10counseling and psychotherapy.
  • 33:11We don't.
  • 33:12Practice.
  • 33:13We might do it a little bit
  • 33:16in our degree based training.
  • 33:19And uh,
  • 33:20but we don't.
  • 33:23We we certainly don't do
  • 33:25that in a professional way.
  • 33:27We don't practice.
  • 33:28I'm going to just practice
  • 33:29making reflections or asking
  • 33:31good open-ended questions.
  • 33:35And sorry, I had a little thing pop up.
  • 33:39And, and this fits with what we know from
  • 33:41the learning sciences community community.
  • 33:44So Danny Kahneman said this in his
  • 33:47book thinking fast and slow that.
  • 33:49The acquisition of skills requires
  • 33:51a regular environment and adequate
  • 33:54opportunity to practice and rapid
  • 33:56and unequivocal feedback about the
  • 33:58correctness of thoughts and actions.
  • 34:01And intuitively,
  • 34:02that makes complete sense for how that
  • 34:04is going to drive good skill acquisition.
  • 34:08And man, those are hard conditions
  • 34:10to meet in terms of training,
  • 34:13counseling, and psychotherapy.
  • 34:16So if we then say OK, if that if
  • 34:19those are the optimal conditions for
  • 34:21skilled development and we wanted
  • 34:24to design A training technology,
  • 34:26these would be some of
  • 34:28the design requirements.
  • 34:29There would be a heavy focus on practice,
  • 34:32many unique practice opportunities.
  • 34:34Varying skills, varying content and feedback.
  • 34:39In particular, immediate feedback
  • 34:41on the correctness of thoughts
  • 34:43and actions according to Kahneman,
  • 34:46and being able to track that overtime.
  • 34:51Excuse me?
  • 34:54So this is these are exactly some of
  • 34:57the design requirements and features
  • 34:59that we have built into several
  • 35:02different training technologies.
  • 35:06And so I want to give you a quick
  • 35:09snapshot of our MIT training platform.
  • 35:13So not surprisingly,
  • 35:14we it's designed around different skills,
  • 35:17so there's different skill modules.
  • 35:20And you know, similar to other
  • 35:22types of training, we are going
  • 35:24to have some expert introduction.
  • 35:26So we'll just hear.
  • 35:28A quick intro from Terry Moyers.
  • 35:31One big reason for burnout in
  • 35:34clinicians is the burden of
  • 35:36trying to convince people to
  • 35:37change when they don't want to.
  • 35:39When you see that as your job,
  • 35:41it can be exhausting to spend all
  • 35:43your time trying to persuade clients
  • 35:45to do what they're fighting against.
  • 35:49So again, this would be similar to
  • 35:51what we find in other trainings.
  • 35:53However, the total video
  • 35:57piece here is 3 minutes long,
  • 35:59so we're going to give you a relatively
  • 36:03light touch in terms of that content.
  • 36:06We're going to give you some
  • 36:08examples of asking good questions.
  • 36:13But then the, the innovation,
  • 36:14the thing that we really
  • 36:16spent the time working on was
  • 36:19developing practice and feedback.
  • 36:21And so in this training tool what we
  • 36:24have are a variety of brief clinical
  • 36:28vignettes from standardized patients
  • 36:30with actors portraying patients.
  • 36:32And so we'll just take a quick listen.
  • 36:35My dad was a drunk and I always
  • 36:37thought I would never be like him.
  • 36:41But lately, all I do is
  • 36:42act like him at his worst.
  • 36:47I get mean when I drink. Just like he did.
  • 36:54And again, this is not a
  • 36:56a long kind of 10 minute,
  • 36:59this is 43 seconds and that's what
  • 37:02these were designed as are like 30
  • 37:04seconds to one minute long and a
  • 37:07variety of them where the point is that
  • 37:10it would give you opportunity to practice.
  • 37:13So we'll look at a couple different
  • 37:16practice examples and the types
  • 37:17of feedback then that is given.
  • 37:19So when you drink a lot it
  • 37:21impairs how your brain works,
  • 37:22it makes your prefrontal cortex.
  • 37:24Part of your brain and the front
  • 37:26of your head work less effectively,
  • 37:28so you end up making really
  • 37:31impulsive decisions like getting
  • 37:32into fights when you don't want to.
  • 37:42And so there's an opportunity to practice.
  • 37:44So you, you. We've just heard our
  • 37:47our trainee listen to Gabriella kind
  • 37:49of talk about her drinking problems.
  • 37:52Our trainee went with a
  • 37:54kind of psychoeducation.
  • 37:56Let me teach you something.
  • 37:58And so the system immediately
  • 38:00transcribes what is said and then
  • 38:02tags it within the motivational
  • 38:05interviewing fidelity coding system.
  • 38:07And so through an MRI fidelity lens,
  • 38:10that was giving information psychoeducation.
  • 38:13And so then we get a brief
  • 38:16encouragement and kind of a redirection.
  • 38:19It looks like you gave information,
  • 38:20let's try asking a question.
  • 38:23And so we'll look at just one
  • 38:24or two more examples of this.
  • 38:29If you want to make better
  • 38:31decisions and get in fewer fights,
  • 38:32you have to consider reducing your drinking.
  • 38:35That's going to help your brain work
  • 38:37better your prefrontal cortex and
  • 38:39allow you to make better decisions.
  • 38:41So you might potentially get in
  • 38:43fewer fights with your partner.
  • 38:45Maybe just talk things out instead?
  • 38:55So again, same idea. Unfortunately,
  • 38:57our trainee has not quite gotten it,
  • 38:59so she's now providing advice.
  • 39:02You know, you should,
  • 39:03you should stop doing this,
  • 39:05you know, kind of hung up on
  • 39:08the prefrontal cortex here, Umm.
  • 39:10And then finally,
  • 39:11example decisions and get in fewer fights,
  • 39:15you have to consider reducing your drinking.
  • 39:18That's going to help your brain work better.
  • 39:21One second. OK.
  • 39:27Have you just considered
  • 39:28trying to drink less?
  • 39:34And we're getting closer,
  • 39:35but the goal here is to ask
  • 39:38an open-ended question,
  • 39:39right one in which our our client
  • 39:41is going to kind of open up and
  • 39:44tell their narrative, their story.
  • 39:47Have you just considered trying to
  • 39:50drink less is a yes or no question and
  • 39:53so that's going to close them down.
  • 39:56So just a couple examples of how we have.
  • 40:01Tried to have you just considered that?
  • 40:03There we go.
  • 40:06A couple of examples of how we're trying
  • 40:09to use AI to provide that practice
  • 40:12opportunities and immediate feedback.
  • 40:14Whereas traditionally if we get
  • 40:17feedback it would be, you know,
  • 40:19having an entire session and then
  • 40:21getting some high level feedback on it.
  • 40:23The idea here is can we provide
  • 40:26small training examples where
  • 40:28practice is coupled with specific
  • 40:30feedback in relatively small bytes?
  • 40:33And then there are dashboards for
  • 40:36providing summative feedback.
  • 40:37How well are you learning the skills
  • 40:40and for then completing additional
  • 40:43practice opportunities over time?
  • 40:47There is a larger RCT of this
  • 40:50training that is underway currently.
  • 40:52The pilot work that laid the
  • 40:55foundation for this showed that
  • 40:58the immediate feedback after each
  • 41:02statement not only led to larger gains.
  • 41:06So it's our red bars down here in our graph,
  • 41:10but at a later testing period
  • 41:13where there was no feedback.
  • 41:16We also saw better retention
  • 41:19and it increases overtime.
  • 41:24So that was a quick snapshot on training.
  • 41:27Let's shift gears and think about
  • 41:30supervision and quality monitoring at scale.
  • 41:35And so here we have been thinking about
  • 41:38that AI pipeline and how we could use it
  • 41:41within a direct clinical service setting.
  • 41:45And so a couple features that we have
  • 41:48built into that technology is that at
  • 41:52a foundational level, we believe that.
  • 41:56Providing a direct contact with the skills
  • 41:59of counseling and psychotherapy are
  • 42:02going to enhance reflection and learning.
  • 42:05So, you know, as opposed to supervision,
  • 42:08that begins with a supervisor asking
  • 42:10what do you want to talk about or
  • 42:13tell me what happened last week?
  • 42:15If we can enhance contact with the
  • 42:17review of the actual skills and practice,
  • 42:20regardless of whether we have any AI
  • 42:24fidelity, that's going to enhance
  • 42:26people's reflection and learning.
  • 42:28And so one piece of it is providing
  • 42:30easy access and easy ways to interact
  • 42:33with the session recording itself.
  • 42:35So there's an automated speech to
  • 42:38text transcript,
  • 42:39it's searchable.
  • 42:40Those little kind of orange bubbles
  • 42:42that you see at the top,
  • 42:43those are our AI identified content
  • 42:47codes kind of what what's happening
  • 42:50in this session.
  • 42:51And then you'll notice that we also provide
  • 42:55a little tags for kind of good good behavior,
  • 43:00if you will.
  • 43:02So particularly empathic moments
  • 43:04within the session get tagged as well.
  • 43:10Each session then gets a
  • 43:12formal fidelity assessment.
  • 43:14So here we're providing kind of a
  • 43:16high level summary of the quality of
  • 43:20motivational interviewing and then some
  • 43:22additional traditional fidelity metrics.
  • 43:25And and so this provides both.
  • 43:28We can drill down really deeply.
  • 43:31Again, MI is is helpful because it
  • 43:34actually will uniquely code every
  • 43:36statement from the therapist in a session
  • 43:39and so our AI will uniquely code every
  • 43:42statement from a therapist in a session.
  • 43:46And then we've also built tools
  • 43:48to try and support really more
  • 43:51of a quality assurance view.
  • 43:53If we now have the ability with that
  • 43:57pipeline to take every session and
  • 44:00generate fidelity metrics or quality metrics,
  • 44:04how then do we interact
  • 44:05with all of that data?
  • 44:07If you are in a large agency,
  • 44:09there's thousands of sessions a month,
  • 44:12if not per week.
  • 44:13And so we have built in this kind of
  • 44:16suite of tools that would allow either
  • 44:19a supervisor or a clinic manager to
  • 44:21be able to have really a population
  • 44:23health view of the quality of services
  • 44:26being provided within their clinic.
  • 44:28And so this,
  • 44:29this GIF is showing that you can select
  • 44:32a particular individual quality metric,
  • 44:34you can select a time frame and then you'll
  • 44:37see summaries of all your providers.
  • 44:40And for Andy,
  • 44:41any individual provider,
  • 44:42you can select them and see something.
  • 44:44Up their caseload and drill down
  • 44:46all the way to individual sessions.
  • 44:52This technology also is getting
  • 44:54evaluated in a current RCT.
  • 44:56Actually I think I've got a
  • 44:58few pieces of information about
  • 45:00that on the next slide in.
  • 45:02In preparation for that work,
  • 45:06we did some user centered design
  • 45:09with publicly funded agencies
  • 45:11in the Philadelphia area.
  • 45:13This is in collaboration with Tori
  • 45:15Creed at the University of Pennsylvania.
  • 45:18And so we spent time talking
  • 45:20with therapists and leadership,
  • 45:22showing them examples of the
  • 45:24technology getting their input both
  • 45:26about implementation feasibility,
  • 45:28how how easy or challenging would
  • 45:30it be to implement these types of
  • 45:33technologies within their setting
  • 45:35and and also measuring kind of
  • 45:37implementation readiness and in
  • 45:39particular after we spent time with
  • 45:41them and showing them the existing
  • 45:44technology all kind of across
  • 45:46the board the acceptability of.
  • 45:48Appropriateness and feasibility increased
  • 45:50for both therapists and leadership.
  • 45:56We are just starting recruitment for a
  • 46:00step wedge design where we will across
  • 46:045 actually due to workforce shortages,
  • 46:07we anticipate it'll be more like
  • 46:097 or 8 clinics to get up to the
  • 46:12number of providers that we need.
  • 46:13But we will randomize clinics to
  • 46:17a sequential essentially turning
  • 46:19on of the technology to support
  • 46:21supervision and quality monitoring in.
  • 46:26Almost 1900 clients and where we
  • 46:29will also be assessing PHQ and GAD
  • 46:32on a weekly basis to assess both
  • 46:35is there an overall effect as well
  • 46:37as what are the particular ways in
  • 46:39which supervisors and clinicians
  • 46:41use the technology that then lead
  • 46:44to improved client outcomes.
  • 46:45So there's a a mediational
  • 46:47hypothesis hypothesis here as well.
  • 46:53All right. I am just got a few more
  • 46:55slides and before wrapping up and so
  • 46:57want to just kind of talk about a couple
  • 47:00things that are in the works right now.
  • 47:05As probably everyone here knows,
  • 47:07there is both a an epidemic of suicide and
  • 47:13suicidality and that the federal government
  • 47:16has been put in has put in place 988,
  • 47:20which is a new crisis care call line.
  • 47:23That is where the idea
  • 47:25is that this mimics 911.
  • 47:27It's a simple 3 digit number
  • 47:30anywhere in the US that can be
  • 47:33used to access crisis services.
  • 47:35As part of this rollout,
  • 47:37which SAMPSA is coordinating,
  • 47:39they are mandating some quality assurance.
  • 47:43I mean, obviously these are some of the most.
  • 47:50Severe interactions within
  • 47:52behavioral healthcare.
  • 47:54As as part of the grant that
  • 47:56we wrote on this topic,
  • 47:57I learned that in crisis
  • 48:00calls in 1% of the calls,
  • 48:02a suicide is taking place during the call.
  • 48:05So these are incredibly important
  • 48:09conversations. And.
  • 48:13But it we're in that same situation
  • 48:15of how do you assess the quality
  • 48:18of these crisis care calls?
  • 48:20Well, the traditional method is you record
  • 48:23them and then they are reviewed manually.
  • 48:26And so we have been working with a partner
  • 48:30to lay the lay the foundation for and
  • 48:33we wrote a grant together to develop an
  • 48:37AI assisted suicide risk assessment.
  • 48:39So these are some of the dimensions.
  • 48:43Of the quality assessment tool that um
  • 48:46Samsa and their partners have put together
  • 48:49and are the focus of our grant work.
  • 48:52So we're we're partnering with protocol
  • 48:55who is provides 988 services and
  • 48:58is also one of the Samsung funded
  • 49:01national backup Centers for 988.
  • 49:04And so as part of our pilot
  • 49:06work with protocol,
  • 49:06we after getting all of the
  • 49:10appropriate IRB in place,
  • 49:12we took some of their existing crisis calls
  • 49:16and put them through our current pipeline.
  • 49:18And one of the things that we
  • 49:21can do right now is identify when
  • 49:23suicide conversations are occurring.
  • 49:25And so as just sort of a proof of concept,
  • 49:27we were able to demonstrate that we
  • 49:30could differentiate both is there
  • 49:31suicide content in this call or not,
  • 49:34as well as how much?
  • 49:35Of a focus,
  • 49:36was it so that X axis is really
  • 49:39kind of what proportion of the call
  • 49:42was focused on suicide?
  • 49:44And then also within that kind of
  • 49:47transcript review phase of the technology,
  • 49:50suicide is one of those items that
  • 49:52gets tagged kind of session content
  • 49:54that gets identified immediately.
  • 49:56And so it's possible to scope into
  • 50:00exactly where in the conversation
  • 50:03suicide is being discussed.
  • 50:04And so as part of what we hope for,
  • 50:07we don't yet have the grant,
  • 50:08but as what we hope for will be our,
  • 50:10our next grant will be developing
  • 50:12some AI technology.
  • 50:14To automatically identify whether
  • 50:16or not suicide risk assessment
  • 50:18is occurring from the call taker
  • 50:20and the quality of that.
  • 50:24Last thing that that I'll mention,
  • 50:26another kind of exciting thing that
  • 50:29is kind of right in the middle of.
  • 50:32You know one of the real world
  • 50:35challenges is implementation and I
  • 50:39think within a training setting it's
  • 50:41you know easier to sell AI based
  • 50:43feedback for learning new skills.
  • 50:45But to you know kind of selling clinicians
  • 50:48in the real world on quality assessment
  • 50:51is a little less straightforward.
  • 50:54And so as we have spent time with clinicians,
  • 50:57I'm talking with them and user
  • 51:00centered design interviews,
  • 51:01you know one of the things that.
  • 51:03We consistently heard from them
  • 51:05was we hate documentation.
  • 51:07If there's any way your technology
  • 51:09could help us with documentation,
  • 51:11we would be more excited about this.
  • 51:14And so we have been in the process
  • 51:18of doing an initial version 1.0 of
  • 51:22a automated clinical documentation.
  • 51:24In particular,
  • 51:25we have done the same basic process
  • 51:28that I described at the beginning,
  • 51:30which is with a clinical partner
  • 51:32we were able to.
  • 51:33Get access to 40,000 clinical notes.
  • 51:38And then we have been training AI,
  • 51:40AI models that well,
  • 51:41here's the recording of the session
  • 51:44and then here is what the human
  • 51:45said is a good summary of it.
  • 51:47And then we can train the AI to
  • 51:50learn that mapping of a recording
  • 51:52to a session summary.
  • 51:54And so this is the type of summary
  • 51:58currently that the system can generate.
  • 52:02So it is if you think about.
  • 52:04Adapt note or a soap note.
  • 52:07It provides that basic initial
  • 52:10discussion summary of what
  • 52:12actually occurred in the session.
  • 52:14And we are still in the process
  • 52:17of evaluating this internally
  • 52:18and with some partners.
  • 52:20But the goal here really is that
  • 52:22it would be a tool that would help
  • 52:25and support providers with not so
  • 52:27much their clinical work per se,
  • 52:29but necessary documentation that
  • 52:31goes along with that clinical work.
  • 52:35And there's some,
  • 52:36some other features that we've
  • 52:38built into it you can include.
  • 52:41The kind of canned phrases that you
  • 52:43might want to include with regularity.
  • 52:47So let me let me wrap up there.
  • 52:48There is a lot of work still to be done.
  • 52:50We have a bunch of these technologies
  • 52:53are currently getting assessed in RCT's.
  • 52:56But it does feel like, you know,
  • 52:59particularly with AI, there can be so,
  • 53:02so much breathless excitement
  • 53:04about AI and what I could do.
  • 53:07And having spent the last
  • 53:0915 years in the trenches,
  • 53:11I feel like we're getting to the
  • 53:13place where there are some practical
  • 53:15tools and we can see some practical
  • 53:17applications of how AI could really
  • 53:20support behavioral healthcare.
  • 53:23So let me wrap up there and happy
  • 53:26to take any and all questions.
  • 53:29Feel free to ask or if they're.
  • 53:33In the chat, I'll try and open the chat.