All Episodes

March 14, 2024 35 mins

What does sickness sound like? Sometimes it’s obvious, like a cough, sniffle, or stuffy nose. But some conditions cause subtle changes that only a trained ear – or AI – can detect. Dr. Yael Bensoussan is a professor of otolaryngology and the director of the Health Voice Center at the University of South Florida. Her problem is this: How do you build a giant, public database of thousands of voice recordings, and use it to train AI tools that can hear when people are getting sick?

See omnystudio.com/listener for privacy information.

Mark as Played
Transcript

Episode Transcript

Available transcripts are automatically generated. Complete accuracy is not guaranteed.
Speaker 1 (00:15):
Pushkin. There are a lot of reasons that I'm excited
about today's show. I'm going to tell you three right now.
Number One, the show is about this whole dimension of
medicine that I essentially didn't know existed, acoustic biomarkers, basically

(00:38):
using a person's voice to assess their health. Second thing
I'm excited about the show is about the intersection of
AI and healthcare, one of my top say five intersections.
Love that intersection. And three, today's guest doctor Yael Bensusan,
gave me what was truly the best excuse that anyone

(01:01):
has ever given me for canceling an interview at the
last minute.

Speaker 2 (01:05):
Yeah, so I'm really sorry for having to cancel on
you yesterday.

Speaker 1 (01:09):
What was the what was the surgery you had to
do yesterday?

Speaker 2 (01:13):
So yesterday we call it airway surgery, where I take
a base to the oar and I have to open
up their windpipe or their trachia because they're a scar
tissue that's blocking them from breathing. So I have to
go with a laser and cut the scar tissue out
and then take a balloon and open up their windpipe
so that they can wake up and breathe better, and

(01:36):
that translates to a different sound when they're breathing.

Speaker 3 (01:38):
So when they're not.

Speaker 2 (01:39):
Breathing because of the scar tissue, we can sound like,
you know, very noisy breathing. We call it the Darth
Vader breathing. And then when wake up they wake up
from surgery and they're done, they have silent breathing, which
means that I know that I did a good job.

Speaker 1 (02:00):
I'm Jacob Goldstein and this is What's your Problem the
show where I talk to people. We're trying to make
technological progress. Doctor Yeah. Eld and Susan run the Health
Voice Center at the University of South Florida, and she
is also leading a team of researchers that's building a
giant database of human voices and breadths and health information.

(02:22):
Her problem is this, how do you record the voices
of thousands of people without violating patient privacy laws while
building a giant public database that could someday allow your
phone to warn you, based solely on your voice that
you may be getting sick. Yeah El told me that
she got into this field in part because she used

(02:43):
to be a singer.

Speaker 2 (02:46):
So I growing up, you know, I always was in
a very musical family. I took singing lessons when I
was a kid, and then I started singing more professionally
around eighteen years old, and I had a short but
exciting singing career. I wrote pop folk music. We had
a bend, and we toured. We had an album out

(03:08):
in two thousand and twelve. Yeah, and I mean it
was a lot of fun. And actually the reason I
was able to have that short and exciting career was
because I met a speech pathologist when I was fifteen.
So I was taking singing classes and one day my
teacher looked at me and she said, there's something wrong
with your voice.

Speaker 3 (03:28):
Go get checked.

Speaker 2 (03:31):
And I met a laryngologist who put his camera down
and she said, you have nodules on your vocal cords
and you might not be able to sing again if
you don't take this seriously. And I went to see
a speech pathologist. I did rehabilitation with my voice for
six months, and I was able to sing again. And
I mean that's what led me to then become a
speech pathologist growing up, and then eventually go to med

(03:54):
school and then decide to become a laryngologist.

Speaker 3 (03:56):
So it was kind of all interconnected.

Speaker 1 (03:59):
So I know that your research now and most of
what I'm really interested to talk with you about is
is around acoustic biomarkers. So just to start, I mean,
what's an acoustic biomarker?

Speaker 2 (04:16):
Very good question. So what is a biomarker? First, A
biomarker is something that indicates the presence of a disease, right,
So if you think about a biomarker for a cancer,
so different cancers have different types of biomarker. For example,
for ovarian cancer, we're looking for a specific thing you know,
called ca in your blood. For different types of cancers,

(04:38):
they could take a blood draw and find a specific biomarker.
It's an indicator of a disease. An acoustic biomarkers is
something that can indicate a presence of a disease, but
that you can hear. So that's the definition of an
acoustic biomarker. So I always say, you know, when you
have people in your family that are not well, you
will always notice first and you'll say you don't sound

(05:00):
good right, or you sound funny. And I have the
luxury to know that because I'm a voice doctor. So
then people will bring me their family members or people
will come saying, I don't know what's wrong with me,
but my wife told me to come because my voice
is not good. And sometimes it's because their vocal cords

(05:20):
are not working, but a lot of times it's because
they can have a neurological issue or a cardiac issue
that is affecting their voice.

Speaker 1 (05:28):
So, more broadly, what's going on with AI and acoustic biomarkers.

Speaker 2 (05:34):
Yeah, so, so many exciting things are going on. I
think that's the first answer. There are so many startups,
so many companies, industry researchers, academic researchers that are working
and looking into voice AI. And the reason is it's
really cheap to collect. Right to think about this, If

(05:54):
you have a phone, it's really cheap to collect Compared
to this.

Speaker 1 (05:57):
You don't have to pick a blood sample. You have
exactly just you've got the phone. You've got the device
literally in your hand already. All you have to do
is talk, and you're talking already.

Speaker 2 (06:07):
And you're talking already, so it's cheap to that's why
pharmaceutical industries are also very interested, and there's a lot
of pharmaceutical projects around it. So there are a lot
of projects that are going on and the state or
the The current landscape is that there's tons of people
working on very similar things and very interesting and various disease.

(06:29):
So I always I kind of categorize them in three
categories of diseases that are being studied. One is the
disease that affects the voice box. Okay, so vocal court paralysis, absolutely,
it's intuitive. There's going to be vocal biomarkers in that
voice box cancer, right, that's easy. Then there's a voice
and speech affecting disorders, so disorders that don't affect the

(06:52):
voice box, but that have an impact on the voice
and the speech. Parkinson is one of them, right, Alzheimer's
is one of them. A stroke somebody having a stroke,
they don't have a problem with their voice box, but
their speech is going to be altered. So these are
voice and speech affecting conditions. So lots of work is
being done in that field. And the third one is
diseases that you don't think would affect speech, but still

(07:15):
people are doing research on that. So there was a
really interesting study on diabetes. They're saying that there was
a group that published that they could diagnose people that
were diabetic versus non diabetics based on their speech and this.

Speaker 1 (07:28):
So this third group is one presumably where there's at
least the potential for AI to detect differences that even
experts like you cannot detect, right, I mean, is that
what's going on there? What?

Speaker 2 (07:45):
So AI is not magical, you know, I think it's
It does a lot of things. But what AI does
that the layperson doesn't do is that it can analyze
a lot more data faster.

Speaker 1 (07:55):
Yeah.

Speaker 2 (07:56):
Right, So AI has the possibility, if you have a
large data set, to then find small differences in these
data sets that we don't have. I mean, I would
have to listen to, you know, thousands and thousands of
voices and compare them statistically.

Speaker 1 (08:11):
It might it might, right. It might also be able
to detect differences that are not even audible.

Speaker 2 (08:17):
It could exactly. I can give it an example. There's
a company looking at atrial fibrillation, and I cannot validate
their data because that's one of the limitations that we're
going to talk about. But obviously their data set is
not public. But they're saying that they can diagnose atrial
fibrillation based on the voice. And their explanation is that
our voice vibrates to the sound of our heartbeats.

Speaker 1 (08:40):
Big if true? Fun if true?

Speaker 2 (08:43):
I mean you know, again, the limitation here is that
it's there's a lot of things you can't validate. But
they say that they've been validating it with EKGs and
that they can see it. They can hear a difference
in the voice between patient patients with a.

Speaker 1 (08:56):
Fib atrial fibrillation. It puts you at risk for a stroke, right,
it can go undiagnosed. So like, if if this works,
that would be very helpful to many people, right, absolutely, absolutely.
So you're mentioning like that's super interesting. It's it's interesting
more generally. So, so you're building a giant database, right,

(09:18):
and I find that interesting for a lot of reasons.
It happens. I don't have you come across the work
of faith A Lee. Absolutely, yeses, So I talked to
faith A Lee for this show not long ago. Wow. Right,
she's like nerd famous, right yeah, And so you know,
as you know, she built this giant database of images

(09:40):
about ten years ago a little more now called image net.
And that was that giant database was what allowed these
early machine learning models AI models to you know, start
recognizing images, right, and so the database was this necessary tool,

(10:02):
necessary thing for the AI to really work, right, And
so are you building the acoustic biomarker version of that?

Speaker 2 (10:13):
So the first the short answer is yes, but I'd
like to start by saying that I am not building
it's our distortion.

Speaker 1 (10:20):
Yes, yes, are you all are?

Speaker 3 (10:23):
Actually, I'll just.

Speaker 2 (10:24):
First start by recognizing here that it's it's a it's
a huge team. So we're the Bridge to Way I
Voice Constortium is a team of fifty investigators across the
US and Canada. We're funded by the NIH through the
Bridge to Way I program and the goal absolutely this
is the first time I hear the analogy to the
image net database.

Speaker 3 (10:43):
I like it.

Speaker 2 (10:43):
I usually give the example of the genomic database, the
Human Genome Project, huge.

Speaker 1 (10:49):
Project, more famous, more famous, they're.

Speaker 3 (10:51):
Both very famous. But I like this analogy.

Speaker 1 (10:53):
Well. Image net is maybe a little bit closer of
an analogy, but maybe less Yeah yeah, sexy, yeah.

Speaker 2 (10:59):
Well, but I mean it's interesting because the genome project
has also very interesting ethical particularities like voice, right, the
image has a little bit less of the ethical constraints.

Speaker 3 (11:08):
For is, when we talk about whole genome.

Speaker 2 (11:10):
Sequencing or genomics data people kind of understand that voice
has similar concerns in terms of process.

Speaker 1 (11:18):
We want to get to the concerns, but I want
to first talk about what you're doing and and then
we can talk about you know, not doing anything wrong. Yeah.
So broadly, if it becomes the thing you hope it
will be, what, what is it going to be? What
is the bridge to AI voice database going to be?

Speaker 2 (11:39):
So it's going to be this large database of thousands
of human voices linked to other health information that are
going to be available to researchers and potentially people other
than researchers as well, to be able to make discoveries, right,
to learn to use a voice AI, to train you know,

(12:00):
the next generation of people on how to learn to
build models on voice AI, to help pharmaceutical companies develop
products or learn even to to develop products, right, And
the other really important thing is to teach people what
type of standards we need right right now, a lot
of different projects, there's really a lack of standards. People

(12:21):
collect voice in different ways. That's why it's really hard
to pull data together. So our dream was really to say, like, hey,
you want to do voice research, here's a manual, my friend, right,
like here is how you collect the voice to make
it accurate. This is the protocols with the task that
we think, based on our studies, give the best biomarkers. Right,

(12:44):
These are the type of biomarkers you can look for
and this is the data you can train, so really
create a manual of operations also for people to be
able to make discoveries, and that's the goal to have
the most impact on patient care.

Speaker 1 (12:58):
So what are the biomarkers? What are you asking people
to do? What are you collecting?

Speaker 2 (13:02):
So I separate things between. So there are respiratory biomarkers,
voice biome markers, speech biomarkers, and linguistics biomarkers, and they're
all different. So let's go about why these are different.
So respiratory is easy, right, So we ask people to breathe,
to cough, to take big breaths in and that has

(13:24):
a lot of information on our pulmonary capacity, on how
our windpipe is shaped. Okay, that's respiratory. Then voice and
speech what's the difference. So voice is really the sound
that we make when our vocal cords come together. So
when we say, like birds can voice, but they can't speak.
If you have a bird that speaks, then you'll be very.

Speaker 1 (13:46):
Rich or you have a parent.

Speaker 2 (13:52):
So when we when we do voice tasks, we ask
patients to say E or.

Speaker 3 (13:57):
Ah or I.

Speaker 1 (13:59):
Get the difference.

Speaker 2 (14:01):
Birds and voice biomarkers will be impacted when our voice
box is changed or our resp is changed. Right, So
somebody with pneumonia probably cannot hold a note for very long,
So that's voice biomarkers. When we talk about speech biomarkers,
then you go into articulation. So some people, for example,

(14:22):
who have neurological deficits or their mouth is not working correctly,
they're going to have trouble articulating. They're going to have
trouble saying some words. So these are biomarkers we can extract.
And then lastly there's linguistic biomarkers. So what type of
words are people using, what type of semantic how fast
do they speak for example? These are all different types

(14:44):
of biomarkers that.

Speaker 3 (14:46):
That we can extract.

Speaker 2 (14:47):
So to give you a very tangible example, I was
reading a paper from a group looking at biomarkers of depression,
and rate of speech was one of the important biomarkers
they found. So people who are sad or depressed will
speak at a slower pace, so words per second is smaller.
So that's simple when you think about it, it's a

(15:08):
simple by marker, right, So that's to give up tangible examples.
So in terms of I think I didn't answer your
question fully, So what are we asking patients? So we
ask people to do all these tasks so coughing, breathing,
a e. Then we make them read those validated passages,
and we also ask open questions. And then when we

(15:31):
ask open questions, we have to ask about questions that
make them emotional and some that don't make them emotional,
because if you trigger emotion, that causes a bias on
how your voice will sound.

Speaker 1 (15:42):
What what question do you ask to make people emotional?

Speaker 3 (15:46):
So it's really interesting.

Speaker 2 (15:47):
So at first we would ask, you know, our first
question was, you know, can you talk to me about
something that makes you sad? It could be somebody that
died in your family or you know, So that was
our prompt. And then our question without emotion was tell
us about your disease and.

Speaker 1 (16:07):
Only a doctor. What'd think that's for that emotional question?

Speaker 3 (16:09):
Exactly?

Speaker 2 (16:10):
I mean, but it's like when you think about it,
like Our consortium is like tons of experts that put
their minds together to develop.

Speaker 1 (16:16):
Tell me about having Parkinson's. That's the unemotional question we're
going to ask.

Speaker 3 (16:19):
And then we I mean, we like, why are you here?

Speaker 2 (16:22):
I think it was not that obvious, but it's like,
tell us about why you're here to see your doctor today.
And then analyzing the data, because we do pilots, right,
we audit our data. We realized that people were starting
to tear up, like we had people crying while talking
about why they were coming to the doctor today, which is.

Speaker 1 (16:38):
Supposed to be the example of unemotional.

Speaker 3 (16:40):
Sure, correct, So we had to change that.

Speaker 1 (16:45):
Yes, interesting, So okay, this is great. So you're getting
a lot of auditory information from every patient. What other
information you're getting from each person? So much?

Speaker 2 (16:58):
So to give you an idea, our full protocol is
about one.

Speaker 1 (17:01):
Hour okay, so of the patient with the patient.

Speaker 2 (17:05):
With an ipassion it's an iPad, So everything is based
on an iPad and there's a helper right now or
research assistant. So we collect data. We collect very extensive
demographics in terms of you know, age, race, geographical location.
We collect language, So what language do you speak? How
many languages is do you speak, what languages do you write?

(17:26):
You know, what part of the world are you from?
That's really important. Then we collect about disabilities. Are you
hearing compared are you visually impaired? Because that makes a
change in your voice, your smoking status, your hydration status,
your fatigue status, because that's so we're we kind of
thought about anything that could affect voice, right, your socio

(17:47):
economical status because if you think about it, that's going
to affect you know, your linguistics as well. And then
so other that extensive demographics, then we collect confounders, so
we think about anything that could change your voice. Do
you have allergies? Do you do you have dental issues?
Do you wear braces? So everybody gets a basic test

(18:09):
about if they are depressed. So no matter what disease
you have, you kind of get the basic tests for
all the other disease to measure if it's possible that
you have concurrent diseases at the same time.

Speaker 1 (18:21):
Because presumably because people are in fact complex, and there
are many people who have depression and Parkinson's and you
want to understand what's going on there.

Speaker 2 (18:30):
I mean, most people are complex, right, It's really rare
to have and people that go to the doctor are
not twenty year old and healthy. Right, most of the
people who will use our technology or will benefit from
these database will be your typical sixty year old chronic
disease patient that comes into the doctor and they're not
they don't have a sterile bill of health.

Speaker 1 (18:50):
How many people do you want to have in the database? Like,
is there a final number you're going for?

Speaker 2 (18:55):
So at the beginning, we were aiming for thirty thousand,
which is extremely it's extremely ambitious, I think to be fair,
I mean, if after four years we get to ten thousand,
I think it'll be a huge success. Okay, And you
know the data collection. I think what we're learning is
that data collection is very resource intensive. To have good

(19:17):
data is very resource intensive.

Speaker 1 (19:20):
So what happened that made you realize that thirty thousand
was maybe harder than you thought?

Speaker 2 (19:28):
So? I think we thought that we wanted to collect
as much data as possible, and our original plan was
to collect a lot shorter protocols, you know, like shorter clips.
But as we started working with patients, we realized that
by getting more data from the same patients, we can
actually have a lot more information and it provides a

(19:51):
lot of interesting you know biomarkers. So we're focusing more
on getting more data from a smaller amount of patients
and really with the right data, kind of right data
with a lot of clinical information attached to it.

Speaker 1 (20:08):
After the break, what the world will look like in
a few years if everything goes well. So this is
a big project that yeah, Elle and her colleagues are
embarked on. It's a four year project. They're about a
year in and there will be interim data releases along

(20:31):
the way. So I asked her, how long will it
take for this project to advance the state of the
science in acoustic biomarkers.

Speaker 2 (20:39):
Yeah, I would say to say at the end of
the four years would be a probably the best answer.
I think at the end of the four years. But
I think that you know, you can just say, oh,
we'll just start training models at the end of the
four year once we have all the data. Right, It's
not just about you know, building one model that I'll
answer your question, is about continuously training models to understand

(21:02):
which biomarkers to extract the then build products that walk.

Speaker 1 (21:06):
So, so, if things go well, what will this world
look like in whatever five years?

Speaker 2 (21:17):
Yes, So, I mean there's there's a few things that
this can help with in general, voice biomarkers. Let's not
talk about just our project. Diagnosis is one thing, right,
early diagnosis, but that's probably the hardest thing, Huh. Screening
is most more important. So when we think about screening,

(21:39):
it means you, let's say you live really far you
don't have access to a doctor, but your doctor has
an iPhone and you can talk into the iPhone and
it can say, hey, something's wrong. You know, you need
a neurological specialist, for example. So to help screen and triage.
I think this probably we're looking at in the next
five years, something definitely possible. The other product that I

(22:00):
think will be very possible within five years is tracking
of diseases. If you want to monitor the evolution of
parkinson or how people respond to drugs. That's why pharmaceutical
companies are very interested.

Speaker 1 (22:13):
Right. So the acoustic biomarker is not just a binary
signal of disease, no disease. It can tell you a
lot about the status of disease. Is it getting better,
is it getting worse?

Speaker 2 (22:24):
Evolution, especially if you train it on your own voice. Right,
it's even easier to detect changes in somebody's voice as
they progress, like your sory for example, or Alexa that
learns listens to your voice. So that's going to be
a really good tool for pharmaceutical companies. That's why they're
investing in it, right, to see how you respond to
a drug, how you respond to a treatment. And when

(22:47):
you think about telehealth at home, right, so more and
more we're going to talk about remote monitoring people. There
were just too many people on this earth to all
be in hospitals when we're sick.

Speaker 1 (22:59):
Well, and if you can stay out of the hospital
when you're sick, that's better, Right, You don't want to
go to the hospital unless you have to do yeah, or.

Speaker 2 (23:07):
That you're you know, your Lexa detects when your voice
starts detailor rating and sends you a nurse before you
need to go to the hospital.

Speaker 1 (23:15):
So there's a more general version of that one, right
that you could imagine, which is you get your whatever,
your iPhone, your Android phone, and you have a choice
when you're setting up your phone, like do you want
to opt into to the phone listening and to tell
you if you need to go talk to your doctor, right,
just like a very broad based thing that you could

(23:39):
opt into, like I would probably opt into that. I mean,
is that a thing that you think about?

Speaker 2 (23:45):
So, I mean yes, I'm sure that you know Apple
is working on that already.

Speaker 3 (23:49):
They are, you know.

Speaker 2 (23:52):
The question is there has to be technology that's being
developed as well to ensure privacy of not only you,
but your environment. Right, because when it's your phone, then
it's your environment as well.

Speaker 1 (24:05):
So you brought up privacy in that context, we can
knock out private see in the context of the database
as well. Here, how could it go wrong? Building a
database of thousands of people's voices with tons of data
about them sort of answers itself.

Speaker 2 (24:26):
Yeah, it can go wrong in many ways. And I
just came out of like two hours of meetings of this.
So add the Bridge to ay I program. We have
a huge group of bioethicists and one of our big
aim as a group is really to ensure patient privacy
and to answer these questions of how do we protect
patient privacy in the context of open data. Right, So,

(24:47):
you are absolutely right, tons of things can go wrong.
People can be potentially reidentified through their voice. So one
of our biggest goals this year is determined what part
of the voice is identifiable and which part is not okay,
And all of this is based on the Hippo law.
Hippo law is from the nineteen nineties.

Speaker 1 (25:08):
Hippola the that governs sharing and security of people's medical information.

Speaker 2 (25:13):
Correct protected health information PHI, we call that, and that
law was made in nineteen nineties, right and back then
they listed a list of things of what they called
PHI or identifiers that cannot be shared openly and that
should stay in the hospital. And voice prints are listed.
When you go into what a definition of a voice

(25:35):
print is, it's very nebulous. It's you know, we don't know.
So because of that nebubularity.

Speaker 1 (25:42):
If I have that word, if that's nebulosity, I'm fred,
I don't know.

Speaker 2 (25:49):
Because it's so nebulous. A lot of institutions, a lot
of hospitals will say, well, you know, voice is not
is not an identifier as long as you don't say hi,
I'm John Doe and I live at four twenty five,
blah blah blah. Other universities will say, no, no, no,
voice is always an identifier. You can never really least

(26:09):
voice data. So what our group is doing right now
is really looking at why the hippo law says this,
what are the actual legal implications of sharing voice? And
we always grade it in terms of risk, Right, if
I talked about all the things that we collect, you
can think that the respiratory sounds are probably very safe

(26:30):
to share versus a speech sample. As we say free speech,
it's probably the most identifying if you have to grade it, right,
And we're kind of looking at, well, where is the balance?
How much can we release? And also we can transform
the data, so for example, we can change the data,

(26:52):
the audio data and what we call visual spectrograms.

Speaker 1 (26:56):
Like a waveform.

Speaker 2 (26:58):
Yeah, it's a sort of waveform that machine learning can use.
We can extract acoustic features, right, like loudness, frequency, stuff like.

Speaker 1 (27:08):
That, and basically trying to figure out how to make
a person be not identifiable based on their voice without
messing up the database. Like that's the balance, right, Like
if you monkey with their voice too much, then your
monkey with the data, the database that we care the
most about. Like that seems like a hard trade off.

(27:28):
So if we go farther out into the future, you
solve all these problems, you build your giant database, the
models get really good. All of these things seem like
things that may well happen. I'm curious about, you know,
AI doing some chunk of what you do now. Right,

(27:48):
we see this happening, say in radiology already. AI is
clearly very good at doing some of the technical work
that radiologists do in diagnosing scans of patients. Right, how
do you think about the future of AI, you know,
using acoustic biomarkers to make diagnosis in a way that
is similar to what you do now as a human being.

Speaker 2 (28:10):
Yeah, So, I mean, I I don't think I'm going
to lose my job yet because I would say that
my primary goal as a doctor is not to necessarily
do that, right, Like, my primary goal is, yes, to diagnose,
but it's to treat patient. So for now, AI is
not going to treat the patient. So I think what
it's going to do is it's going to support a
lot of the workforce. So for example, I'm an academic laryngologist.

(28:35):
I'm a super a super specialist.

Speaker 3 (28:37):
Right.

Speaker 2 (28:37):
For people to get to me, they often see like
five different doctors.

Speaker 1 (28:41):
So instead of going through for specialists who can't figure
it out, you go to your primary care doctor or
even you just talk to your phone, and your phone
says you better talk to your primary care doctor, and
your primary care doctor sends sends the patient directly.

Speaker 2 (28:54):
To you, correct, correct, right to say like hey. Because again,
most of what we do for a very long time
will will need a gold standard right diagnosis. So often
you know it's it's a it's a biopsy, or it's
a it's and imaging.

Speaker 3 (29:09):
You need a gold standard.

Speaker 1 (29:11):
The acoustic biomarker is not a clear enough diagnostic technique.
You need something more reliable.

Speaker 2 (29:18):
So I don't think it's going to you know, no
doctor will say, oh, well, based on this, this is
your diagnostics, start chemotherapy. That's not where we're going. I
wouldn't take chemotherapy based on an acoustic biomarker. But it's
hopefully going to support a lot of primary care and
access to care to get to the right person faster.

Speaker 1 (29:39):
Great anything else we should talk about.

Speaker 2 (29:44):
The one thing we didn't talk about, I guess I
talk about this all day, so sometimes it's hard to
remember what I've said in iowha haven't said. But the
implication for probably all this new telehealth you know, online
world that we live in, a lot of industries are
already integrating tools. So, for example, Canary Speech is a

(30:07):
startup that sold a product. I think they're working with
teams to capture if there's signs in your voice of depression.

Speaker 1 (30:16):
Teams meaning Microsoft teams, like Microsoft's version of Zoom.

Speaker 2 (30:20):
Yeah, yeah, so I think And don't quote me on
the particular. Maybe I'm you know, I'm not giving, but
but I know there's a few startups that are starting
to integrate products in Zoom or in teams to let
employers know that, hey, your employee is not doing well
based on his voice, for example, Right, and.

Speaker 1 (30:36):
What is your view of the efficacy of those?

Speaker 2 (30:41):
So, I mean, I I the easy the quick answer
is it probably works partially. Yeah, But the question is
not if it works full you're not. The question is
does it make a difference? Right, So let's say let's.

Speaker 1 (30:57):
Do what they say it does. Is a question that
matters to me, right, like does it are the claims valid?
Seems like a reasonable starting Yeah.

Speaker 3 (31:05):
I think so.

Speaker 2 (31:05):
So. I just I just reviewed one an article of
one of the startups Fantastic that's looking at like depression,
and I mean their numbers look great. I do think
that's that the results that a lot of these projects
are getting are definitely positive and promising.

Speaker 1 (31:19):
Absolutely, we'll be back in a minute with the light.

Speaker 3 (31:28):
M h.

Speaker 1 (31:37):
Now, as promised, we're back with the lighting around. What
was your band called?

Speaker 2 (31:43):
Ha, My chase stage name was Ella Bence Ella Bence
because my last name is Ben Susan so that was
too long.

Speaker 1 (31:51):
And yeah, al became Ella.

Speaker 3 (31:53):
Yeah, I could say my first name.

Speaker 1 (31:57):
What did you have a hit song in French? What
was it called?

Speaker 2 (32:04):
I wouldn't call it a hit song. It was called
annalis samp means I guess in English, it's like a
one way flight.

Speaker 1 (32:12):
Can you sing a line?

Speaker 3 (32:13):
No, that's my previous life.

Speaker 1 (32:17):
Can you just say a line? Yeah?

Speaker 3 (32:22):
I was Uh, it's in French, though I know it'll
sound great. Yeah, And nalisam.

Speaker 2 (32:33):
Means get me a one way flight for the other
side of the world. I hope people are really happy there.

Speaker 1 (32:39):
Well you're here, now, you're you're you're in Tampa. Now
did it work out as hoped?

Speaker 2 (32:44):
Oh? Yeah, I mean I have I have the best
job in the world, you know. I get my my
mom raised us me and my brother saying you guys
need two jobs, one that make money, makes money and
the other one that makes you really happy. And if
you manage to have both in one job, then you'll
have made it, you know. And I get to be

(33:05):
a surgeon and work with voice and voice professional and
you know, it's been my passion pretty much all my life.

Speaker 3 (33:13):
So yeah, do.

Speaker 1 (33:14):
You work with professional singers as a physician?

Speaker 2 (33:18):
Absolutely? I mean I love treating my professional singers, so yeah,
I love that part.

Speaker 3 (33:22):
Of my job.

Speaker 1 (33:23):
Have you treated anybody famous? Yes?

Speaker 3 (33:26):
But I can't tell.

Speaker 1 (33:28):
What's Taylor Swift really like?

Speaker 3 (33:30):
Oh that I don't know. I wish No.

Speaker 2 (33:32):
She sounds fine, though she probably doesn't need a laryngologist.

Speaker 1 (33:35):
What's the best cure for a sore throat?

Speaker 2 (33:39):
Voice rest and advil? It takes the inflammation away.

Speaker 1 (33:43):
Uh huh? Advil? Just ibuprofen and don't talk and voice rest.

Speaker 3 (33:48):
Yes?

Speaker 1 (33:49):
Okay? Are you just always involuntarily diagnosing people based on
their voice all the time? Okay?

Speaker 3 (33:57):
So funny funny fun fact.

Speaker 2 (33:58):
Two months ago, my girlfriend from residency called me. I
hadn't spoken to her in like a year and a
half and she called me and she said hi, And
I said, you're pregnant?

Speaker 1 (34:09):
Really?

Speaker 2 (34:11):
And I could hear it because pregnancy gives you like this,
you know, you get stuffy in a certain way in
your nose, like we call it rhyanidis of pregnancy.

Speaker 3 (34:21):
And I knew her voice very well.

Speaker 2 (34:22):
She was my girlfriend for a long time, you know,
we studied together, and she just I knew it, you know.
And and I think she says, hey, how are you.
I wanted to talk to you, and I'm like, you're pregnant,
and She's like, how.

Speaker 3 (34:33):
Did you know? So?

Speaker 1 (34:35):
Yes, that's amazing people.

Speaker 2 (34:37):
I mean I was listening to the political debates, you know,
and I'm like, ooh, this guy needs a laryngologist. I
could I'm diagnosing people all the time.

Speaker 1 (34:46):
Well, they should give you a call. Yeah, absolutely, Okay,
lovely to talk with you.

Speaker 2 (34:54):
It was one of the funnest interviews I've done.

Speaker 1 (35:00):
Yeah, Albin Susan runs the Health Voice Center at the
University of South Florida. She's also a principal investigator on
the Bridge to AI Voice Project. Today's show was produced
by Gabriel Hunter Chang, edited by Lydia Jeane Kott, and
engineered by Sarah Bruguer. Just a quick note, We're going
to be taking a break for the next couple of weeks,

(35:20):
but we will have an episode in our feed next
week from our colleagues over at the Happiness Lab that
is timed not coincidentally to World Happiness Day, which I'm
informed is on March twentieth. I'm Jacob Goldstein and we'll
be back soon with more episodes of What's Your Problem
Advertise With Us

Popular Podcasts

Dateline NBC
The Nikki Glaser Podcast

The Nikki Glaser Podcast

Every week comedian and infamous roaster Nikki Glaser provides a fun, fast-paced, and brutally honest look into current pop-culture and her own personal life.

Stuff You Should Know

Stuff You Should Know

If you've ever wanted to know about champagne, satanism, the Stonewall Uprising, chaos theory, LSD, El Nino, true crime and Rosa Parks, then look no further. Josh and Chuck have you covered.

Music, radio and podcasts, all free. Listen online or download the iHeart App.

Connect

© 2024 iHeartMedia, Inc.