Episode 25, May 23, 2018
With 7 billion people on the planet, you might be surprised to learn that approximately a billion of those people experience some form of disability. Enter Principal Researcher and Research Manager, Dr. Meredith Ringel Morris, and the Ability Group at Microsoft Research. They’re working to remove accessibility barriers both to and through technology, empowering people with disabilities to better perform their daily tasks.
Today, Dr. Morris gives us some fascinating insights into the world of “ability,” talks about how technology is augmenting not only sensory and motor abilities, but cognitive and social abilities as well, and shares how Microsoft, through its AI for Accessibility initiative, is committed to extending the capabilities and enhancing the quality of life for every person on the planet.
Meredith Ringel Morris: In the past, there was what we call a “medical model” of disability. And this model proposes that people with disabilities are ill and need to be fixed. More recently, thinking has really evolved toward a more “social model” of disability. And so, the social model emphasizes that many people with disabilities are only disabled by the way that society treats them. So, that philosophy of the social model rather than the medical model of disability is something that the Ability and Enable teams really embrace in our thinking about technology.
Host: You’re listening to the Microsoft Research Podcast, a show that brings you closer to the cutting-edge of technology research and the scientists behind it. I’m your host, Gretchen Huizinga.
Host: With 7 billion people on the planet, you might be surprised to learn that approximately a billion of those people experience some form of disability. Enter Principal Researcher and Research Manager, Dr. Merrie Ringel Morris, and the Ability Group at Microsoft Research. They’re working to remove accessibility barriers both to and through technology, empowering people with disabilities to better perform their daily tasks.
Today, Dr. Morris gives us some fascinating insights into the world of “ability,” talks about how technology is augmenting not only sensory and motor abilities, but cognitive and social abilities as well, and shares how Microsoft, through its AI for Accessibility initiative, is committed to extending the capabilities and the enhancing quality of life for every person on the planet. That and much more on this episode of the Microsoft Research Podcast.
Host: Merrie Ringel Morris. Welcome to the podcast.
Meredith Ringel Morris: Thank you. Glad to be here.
Host: Thanks for coming in. So, your research falls under the general umbrella of Human Computer Interaction, or HCI. But you’re a Principal Researcher and Research Manager of a relatively new group in Microsoft Research. Tell us what it is and what gets you up in the morning?
Meredith Ringel Morris: So, our new team is the Ability team, and it’s a new group that’s focused on harnessing research at the forefront of HCI, Human Computer Interaction. And also, AI, in order to augment the capabilities and extend the abilities of people everywhere, and in particular, the approximately 20 percent of the population that has some form of disability.
Host: So, let’s drill in there a minute. You have referred to some pretty surprising statistics. Give us a picture of the ability spectrum from, you know, um differing levels, categories and lengths, even, of disabilities.
Meredith Ringel Morris: So, the most recent US census indicates that about one in five Americans experience some form of disability in their daily lives. And statistics from the World Health Organization indicate that about 1 billion people worldwide are disabled in some way. And disability refers to a spectrum of different condition, that includes sensory disabilities such as challenges with vision and hearing. That includes motor disability. That can include speech and language disabilities, as well as cognitive disabilities, such as learning challenges, autism, attention challenges, et cetera. And then, of course, in our own research, I think it’s really motivating to think about not only how our research advances can specifically assist those disability constituencies, but actually how advances in research for people with disabilities actually tend to benefit the usability of technology for everyone, because all of us, at some point in our lives, experience temporary, or situational, disabilities. A temporary disability in a motor space might be if someone breaks their leg. So, they might be using crutches or a wheelchair for a short period of time. And, of course, would then benefit from technologies that assist people who have more lifelong motor disabling conditions. But even situational impairments that are much more fleeting affect all of us. If someone is pushing a grocery cart or pushing a baby stroller, just in that moment, they, also, in a sense, have a motor disability.
Host: You know, that’s interesting, because I don’t think most of us think about disability in terms of that it might be temporary and that all of us, at some point, experience some kind of disability, so I think it’s interesting to broaden the vision of what that means, and how technology can help us there. So, talk a bit about how your technologies are aiming to help us.
Meredith Ringel Morris: So, the Ability group has projects focusing in several different areas. In the cognitive space, we’ve been doing some work on how to make search engines easier to use for people with dyslexia. So, dyslexia is a spectrum disorder. It impacts about 15 to 20 percent of English language speakers. The incidence rate is different depending on different languages. But 15 to 20 percent of the population is quite substantial. And people with dyslexia experience a range of challenges with spelling and reading comprehension, and, of course, search engine use, to find information online is, I would argue, one of the most important modern literacy skills today. And so, we’ve been doing some basic studies to understand first, what are the challenges that people with dyslexia experience at different stages of web search? So, choosing their query terms, triaging the information on the search engine results page, actually finding specific nuggets of information within a webpage once they’ve arrived at it. And then thinking about how we might change both the ranking algorithms for search results, themselves, as well as the interface for presenting those results, to make it easier to use for that group. And I think that that’s also another really interesting example of where thinking about a particular disability, such as dyslexia, might yield a user interface that impacts a much larger group of people. Children who are still just learning to read, or people learning English as a second language, might also benefit from many of these same interventions that would benefit users with dyslexia.
Host: Would you look at those as applied to a broad audience? I mean, if you’re using algorithms that are on your search engine, everyone’s going to experience them, right? From the person who’s an early reader to someone who’s a PhD…
Meredith Ringel Morris: Well, I think that’s still a design decision to be made is about whether you would want these interface designs to be applied generally, across all users of the service, or whether it might be another form of personalization. So, today, major search engines generally apply different types of personalization. So, if you and I, both, were to search for “hamburger restaurant,” and I was searching from Seattle, according to the geolocation of my browser, and you were searching from London, we would actually see different search results, because the results are actually being tailored to something about us, in this case, our location. So, you could imagine situations where search results might be tailored to someone’s cognitive profile. Of course, that also raises potential privacy concerns.
Host: Yeah, we’ll talk about that in a bit.
Meredith Ringel Morris: That’s definitely something we’re aware of. I think there are interesting trade-offs in how you might personalize or generalize these kinds of interface designs.
Host: There’s another group within MSR under the umbrella of Human Computer Interaction called Enable that you collaborate with quite often. Why two groups in sort of the same zone, and how would you characterize the research going on in each group?
Meredith Ringel Morris: Absolutely. So, the Ability team and the Enable team work together closely. The Ability group’s mission is in more foundational, academic work, more fundamental, scientific research, that then lays the basis for advanced engineering efforts and shipping more public-facing products by the Enable team. So, for example, the Enable team shipped the hands-free keyboard, which was inspired by their collaborations with Steve Gleason and Team Gleason, which is a foundation for people with ALS or Lou Gehrig’s Disease. And so, the hands-free keyboard is an improvement on eye gaze interaction for typing and then speech synthesis. And so, the Enable team led the engineering efforts on that product. But behind the scenes, they were collaborating with myself and other members of the Ability team, and we did more underlying academic research that we shared broadly with the community that influenced that product. We did research to understand what are some of the limitations of eye trackers for different users and different lighting conditions, and how might interfaces adapt to those changes. We looked at new algorithms for mitigating error in eye typing that can now let users eye type more quickly, because they have to spend less time correcting errors. And we also did some fundamental research about what end users want from an AAC device in terms of how these communication devices can let users more authentically express their personality. So, we were out on the ground doing interviews with people with ALS to learn about their current technology needs. For example, how these devices currently don’t let people express humor, sarcasm, artistic expression, you know, wanting to be able to sing through your device. So, we document those needs, and then we also propose innovations in terms of how these devices can express richer prosody and other channels of feedback that might support these needs in the future. So that’s how the Enable team and the Ability team work together towards shared goals.
Host: So, back to the accessibility technology and the terminology, when we think of disabilities, we most often think of sensory and motor, because those are the most obvious ones. And you refer to some of the work you’re doing with search engines for cognitive challenges as well. Is there anything else happening in the cognitive range that is interesting in this research?
Meredith Ringel Morris: So, another area that our team has been working on is technologies for people on the autism spectrum. So, Microsoft is, I think, a leader in beginning to think more actively about this space, particularly through our inclusive hiring program. So, Microsoft is now part of a consortium of companies that are actively working to change interview and hiring and retention practices to help harness the neuro-diverse talent in the software engineering industry. And so, one thing that the Ability team is involved in, in addition to consulting with that program to help design the inclusive technology program, but also to think about how our software engineering tools and practices, as a tech company, might need to be altered and updated to better support neuro-diverse talent.
Host: That’s the first time I’ve heard that phrase, neuro-diverse.
Meredith Ringel Morris: Neurodiversity. Well, I think there’s some interesting thinking in the accessibility space in the area of disability studies that’s really changed over the past few years. So, in the past, there was what we call a “medical model” of disability. And this model proposes that people with disabilities are ill and need to be fixed, and that medical intervention is the focus. And more recently, thinking in that space has really evolved toward a more “social model” of disability. And so, the social model of disability emphasizes that many people with disabilities are only disabled not by their own state of being, but by the way that society treats them. The onus is really on society to adapt and be more inclusive. And the neurodiversity movement is one manifestation of that. So, I think that philosophy of the social model rather than the medical model of disability is something that the Ability and Enable teams really embrace in our thinking about technology.
Host: Again, back to the when we think of disability, we often think of sensory and motor and so on, and we also think of solutions that are meaningful and important. And so, thinking of something like social media, and how teens like to participate and be part of the group and have their friends share pictures and things like that, tell me a little about what you’re doing for that group of limited vision population in the teen group.
Meredith Ringel Morris: Yes, so, this past summer, we did some really interesting research led by our intern, Cindy Bennett, who is a student nearby at the University of Washington, to understand how teenagers with vision impairment are using photography and social media. And we included, in that study, teens with a range of visual impairment. So, all of the teens who participated were legally blind. But most of them would be classified as having low vision. But the remainder of the group that we interviewed relied, to some degree, on their residual vision. Both of these groups of teenagers, whether they completely lacked vision or had some residual vision, were very interested in using visual social media tools, just as any teen would be, in order to communicate and stay in touch with their friends and participate in pop culture. So, applications like Facebook, Instagram, Snapchat, were all part of these teens’ daily lives. And observing the usability challenges that this group experienced was really interesting for thinking about how we can redesign these kinds of photography capture and editing applications to make them more accessible. It’s really important for all teenagers to be able to take flattering selfies. So. This was particularly challenging for people with low vision, because in order to take a selfie, you need to hold the camera at arm’s length away from your face. That made it impossible for this group to see on the screen whether they were framed in a flattering way on the camera, or, even more challenging, if they were using some apps that tried to overlay different lenses on you while it’s taking the selfie, they couldn’t tell what those looked like or if they were aligned. So, thinking about how you might have better audio feedback in selfie cameras to give people information to help them position…
Host: Like, move to the left, or focus your face?
Meredith Ringel Morris: Right, or “you’re completely in the frame now,” or “you’re out of the frame,” would be really helpful to that audience. And another thing we saw the teens doing that I was really surprised by, was they used filters in an interesting way. What we saw was that teenagers with low vision, in addition to using filters for artistic purposes to kind of fit in with their friends, they were also using filters to actually improve the visual clarity of images for themselves.
Host: So they could see them.
Meredith Ringel Morris: Right. So, people would go through and choose a filter because it best matched their visual abilities to increase their own perception of a photo. And they would even do things like download other people’s photos so that then they could filter them in order to see them better. And I think that suggests a lot of, again, potential for technological innovation in terms of actually designing filters specifically for different visual conditions that might be optimized to improve visual clarity for particular subsets of users, rather than only designing filters for aesthetic purposes.
Host: Is that something that your research team is looking at and studying and…?
Meredith Ringel Morris: It’s on our list of things to tackle going forward. Yeah. So, we have an interesting pipeline in our research. So, we do – Human Computer Interaction more broadly encompasses several different research methods. So, it includes research to understand users’ needs. So this is things like ethnography and interviews and surveys and observations to understand what users really want and need in technology. So, our team does a great deal of that kind of work. And then the next layer is to build on that, right? And now that we understand the needs in technology and the potential for innovation, to actually build new prototypes, new algorithms, new interfaces, that enable these kinds of expression that we see a need for. And that’s also, again, something that our team does. So, we build prototypes, and then actually testing out these prototypes with users to understand what works and what doesn’t and refine them. And then of course the next stage is actually building on that knowledge to engineer, you know, real systems that might be shippable to an end user.
Host: It sounds like you use the full spectrum of research methodology from qualitative to quantitative, and then actual lab testing of the stuff, and then getting it out into the world.
Meredith Ringel Morris: Yeah, so our team has people with backgrounds in different areas. We have one researcher with a background in psychology, others, like myself, with backgrounds in computer science. So, we really have different kinds of skills on the team.
Host: I think it’s important to point out, now, that your team conducts research at, as you call it, the intersection of HCI and AI. And the AI part is super important. Without it, the research would look a lot different, I think. So, tell us how AI is changing the game for innovation in your field in ability, enabling and Human Computer Interaction.
Meredith Ringel Morris: Absolutely. I think there are a lot of AI advances that are going to enable new experiences for people with disabilities. But I think it’s also vitally important that the needs of that user population be considered by people involved in the AI research directly. There’s been an explosion recently, thanks to advances in deep learning, in the ability for computer vision systems to label photographs. And that’s wonderful. And, of course, to me, an obvious application for a system that can automatically label photographs is to caption images for people who are visually impaired. Because when browsing the web right now, a screen reader can only describe an image if the author of a webpage has supplied an alt-text for it.
Meredith Ringel Morris: That’s right. And about 50 percent of images on the web right now, on major websites, lack alt-texts completely. So, a blind user wouldn’t receive any description. So one might think well, we’ll use these new AI technologies to caption the images. But there are some challenges with that. So, most of these technologies are not developed right now with the scenario of use by people who are visually impaired. So, there’s the assumption that if a mistake is made in labeling the image, the cost of that mistake is relatively low, because a sighted user can see that a particular image maybe doesn’t match the retrieval terms and just ignore that mistake. But for someone who is visually impaired, the cost can actually be quite high. And that, I think, is a more fundamental problem in AI research that I think our perspective from HCI and working with end-users can really inform through collaboration.
Host: Right. Let’s talk about collaboration and circle back to the inception of the Enable group. And it began as a partnership with Team Gleason. I’m not sure everyone knows who Steve Gleason is and what the collaboration started as and how it’s advanced and grown.
Meredith Ringel Morris: Sure, so the Enable team, which is led by Rico Malvar, who is the Chief Scientist of Microsoft Research, they grew out of a partnership with Team Gleason as part of Microsoft’s one-week Hackathon. And so, Steve Gleason is a very well-known football player and he was diagnosed at a very young age with ALS, which is also known as Lou Gehrig’s Disease. And as people with ALS progress and they lose muscle control, they need to rely on other tools such as wheelchairs for mobility and AAC devices for communication and speech generation. And Steve was not satisfied with the state of current technologies in that area. Unfortunately, there’s no medical cure right now for ALS, and Steve is very well-known for a quote that I’ll try to reproduce accurately here. So, his approximate quote is that, “While there is no medical cure for ALS, technology can be the cure.” And so, by that, he means that technology can really, in this social model of disability, technology plays an important role in removing barriers to access and making people less disabled in their daily interactions. And so, his team has worked closely with Microsoft on two particular projects, which were the initial projects of the Enable team. One is allowing someone to use eye gaze to drive their own wheelchair, to have more autonomy. And then the other is in improved interfaces for typing with eye gaze. So, typically, the way people with ALS and other serious motor conditions communicate is through using the eyes to type. So, you would stare, with your eyes, at a particular letter on the screen with an eye tracker for a set amount of time until the system’s confident that you’re looking at that letter. And then that letter would be typed, and you do that one at a time. And then, when you’re done typing something, you would hit a button that would speak that out in a computer-generated voice. And, typically, people achieve typing rates that are very slow, about 5 to 10 words a minute. Whereas conversational English speech is closer to 190 words a minute. So that is a huge impediment to participation in daily life when you’re communicating that slowly. So, just letting people type, letter by letter, is not going to get you up to regular rates of speech. You need prediction. And that’s where AI comes in again. And so, the Enable team has been working on better user interfaces and better algorithms behind the scenes, and word prediction in order to improve that kind of speech. And of course, it’s not just the words. So, another thing we’ve been thinking about is, again, how you let people express themselves more quickly and more richly. And so, like they say, a picture is worth 1,000 words. We’ve looked at adding a row of keys to the communication devices that are a different emoji. So, representing the most common human emotions, anger, surprise, sadness, happiness. And if you just add one of these emoji to your sentence as a punctuation, so it’s just one more key press, so low input from the user, since each input is such a great effort. And we use those to not only modify the prosody of the output behind the scenes, but we insert clips of non-speech audio. So, for example, if you add a surprised emoji to a sentence, you might get “Hhhhhhaa!” inserted, and if you add the angry emoji, you might get “Ahhhrrr!” And these communicate such a great amount of emotion and nuance for such a low effort, a single key press. And that’s been really well-received by users in some of our research testing.
Host: You talked about what you do with universities and so on, and I know you’re affiliated with the University of Washington, and you’re involved in a program called DUB. Can you tell us what that is?
Meredith Ringel Morris: Sure.
Host: And what do you do?
Meredith Ringel Morris: DUB is a little bit of a play on words, because of course, University of Washington is abbreviated as UDub. But the DUB Research Group, it’s spelled D-U-B. And it stands for Design, Use, Build. And so, it’s an interdisciplinary consortium of faculty from several different departments, so Computer Science, the School of Information, Human Centered Design and Engineering, Arts and Design, that are all interested in creating technology that is more human-centric for end-users. Myself, and several other researchers from Microsoft Research, are actively involved in collaborations with DUB.
Host: So how did you end up here? Give us a little bit about your background and what brought you to Microsoft Research?
Meredith Ringel Morris: Sure, so my background is that I studied computer science as an undergraduate and I went to Brown University, which at the time, had a very traditional computer science department, so there was no HCI. And I learned about HCI by browsing the web. I found the website of the Stanford Interactive Workspaces Project, which was an early project around ubiquitous computing, so thinking about how we can design spaces where computing is embedded in the environment. And I remember that webpage had a picture of a room that at the time looked very futuristic to me, where there were these large, wall-sized displays, and a table with a display embedded in it and it was really a very futuristic-looking. So I contacted the professor in charge of that project at Stanford, who is Terry Winograd, to ask how I could get involved in this kind of work, and I ended up doing some volunteer-based research projects with him, and I eventually went to graduate school in the computer science department there and that was kind of my route into learning more about this field of HCI. And while I was at Stanford, I was fortunate to get an internship with Microsoft Research, where I worked with Eric Horvitz and Susan Dumais on thinking about how desktop search, which was very new at the time, how you could allow people to use context that they might remember to assist in search. So instead of only searching for something by the file name, you could specify other things that you might remember, like, “Oh, I remember it was that PowerPoint document that I wrote the day after the presidential election.” Or, “Oh, it was that email someone sent me right after my son’s birthday party.” And that you could use these anchors that were more meaningful and memorable to people than file names as a way in to search for information. And so that was a really exciting project that kind of expanded again my knowledge and got me into thinking about information retrieval. But I also learned about the culture of Microsoft Research and then I was excited to come and work here when I graduated.
Host: Given the insights that you get from the work you’re doing – which is both powerfully beneficial for good, but could also have some downsides – is there anything that sort of keeps you up at night?
Meredith Ringel Morris: I think one is in thinking about AI systems that are more inspectable and understandable. Not only inspectable by AI researchers, which is, in and of itself, still a challenge, but inspectable by end users so that they can really understand when to place their trust in a system and what a system is doing. I think there are challenges in developing AI systems that balance augmenting users’ capabilities with the privacy needs of the users themselves and people in the surrounding environment. So, let’s say one might imagine, hypothetically, that someone’s who is visually impaired might benefit from having an outward-facing smart camera that’s a “wearable” that could sense things in the environment using computer vision and describe that to someone. So, that could potentially have a great benefit to the end user but might have a privacy cost to people in the surrounding environment who may not be actively consenting to being captured. And I think thinking carefully about those kinds of ethical challenges is actually a really important part of AI research. And I know Microsoft has now formed a group called AETHER, A-E-T-H-E-R, which is thinking specifically about AI and ethics. The A is for AI, the E is for ethics, and I think that’s one of the areas that will be really interesting to tackle.
Host: Microsoft just announced a new initiative at the annual Build Conference that’s pretty exciting. Tell us what it is, who it will impact, and what it tells us about Microsoft’s broader mission for technology in the 21st century.
Meredith Ringel Morris: Yes, so you’re referring to the AI for Accessibility Initiative. So, this is an exciting new program that is designed to help support grass roots innovation in the accessibility space. So, Microsoft is really interested in encouraging students, entrepreneurs, et cetera, to think about how you can use Microsoft technology to enable important scenarios for people with disabilities. So, for example, how can we use these technologies to allow people to be more productive in their work life or participate more fully in social life outside of work? And so, the AI for Accessibility initiative offers funding opportunities that people can apply, so you describe how your project or app would fit into this vision, and how Microsoft can help support your success in this space, perhaps by donating compute time on our Azure servers, or allowing free API calls to our cognitive services APIs, which offer shortcuts to some of our advances in AI technology, like libraries for computer vision and for natural language processing. So, we’re really excited about this initiative. We’re excited to see what kinds of great project proposals and applications come in. And I think it accentuates Microsoft’s commitment to its mission statement of empowering all users in their lives, and I think empowering all users really refers to all users, including the 1 billion people worldwide with disabilities. So, it’s great to see that being emphasized in this new initiative.
Host: Merrie Ringel Morris, I so enjoyed our conversation today. My eyes were opened, shall we say? Thank you for coming in and sharing the work that you’re doing and the passion behind it.
Meredith Ringel Morris: Great, thank you.
Host: To learn more about Dr. Merrie Ringel Morris and how AI is helping people with disabilities all over the world, visit Microsoft.com/research
You must log in to post a comment.