Player FM - Internet Radio Done Right
Checked 1M ago
Lisätty four vuotta sitten
Sisällön tarjoaa Jaine Schmidt and Lighthouse Guild. Jaine Schmidt and Lighthouse Guild tai sen podcast-alustan kumppani lataa ja toimittaa kaiken podcast-sisällön, mukaan lukien jaksot, grafiikat ja podcast-kuvaukset. Jos uskot jonkun käyttävän tekijänoikeudella suojattua teostasi ilman lupaasi, voit seurata tässä https://fi.player.fm/legal kuvattua prosessia.
Player FM - Podcast-sovellus
Siirry offline-tilaan Player FM avulla!
Siirry offline-tilaan Player FM avulla!
Kuuntelemisen arvoisia podcasteja
SPONSOROITU
“I used to be the largest dairy consumer on the planet. I used to eat so much dairy and meat. The more that I looked into the dairy industry, the more that I saw that it was the singular, most inhumane industry on the planet, that we've all been lied to, including myself, for years. I always believed that the picture on the milk carton, the cow standing next to her calf in the green field with the red barn in the back was true. It’s certainly the complete opposite.” – Richard (Kudo) Couto Richard (Kudo) Couto is the founder of Animal Recovery Mission (ARM), an organization solely dedicated to investigating extreme animal cruelty cases. ARM has led high-risk undercover operations that have resulted in the shutdown of illegal slaughterhouses, animal fighting rings, and horse meat trafficking networks. Recently, they released a damning investigation into two industrial dairy farms outside of Phoenix, Arizona supplying milk to Coca-Cola’s Fairlife brand. What they uncovered was systemic animal abuse, environmental violations, and a devastating betrayal of consumer trust. While Fairlife markets its products as being sourced "humanely," ARM’s footage tells a very different story—one of suffering, abuse, and corporate complicity. Despite the evidence, this story has been largely ignored by mainstream media—likely due to Coca-Cola’s massive influence and advertising dollars.…
On Tech & Vision With Dr. Cal Roberts
Merkitse kaikki (ei-)toistetut ...
Manage series 2934485
Sisällön tarjoaa Jaine Schmidt and Lighthouse Guild. Jaine Schmidt and Lighthouse Guild tai sen podcast-alustan kumppani lataa ja toimittaa kaiken podcast-sisällön, mukaan lukien jaksot, grafiikat ja podcast-kuvaukset. Jos uskot jonkun käyttävän tekijänoikeudella suojattua teostasi ilman lupaasi, voit seurata tässä https://fi.player.fm/legal kuvattua prosessia.
Dr. Cal Roberts, President and CEO of Lighthouse Guild, the leading provider of exceptional services that inspire people who are visually impaired to attain their goals, interviews inventors, developers and entrepreneurs who have innovative tech ideas and solutions to help improve the lives of people with vision loss.
…
continue reading
36 jaksoa
Merkitse kaikki (ei-)toistetut ...
Manage series 2934485
Sisällön tarjoaa Jaine Schmidt and Lighthouse Guild. Jaine Schmidt and Lighthouse Guild tai sen podcast-alustan kumppani lataa ja toimittaa kaiken podcast-sisällön, mukaan lukien jaksot, grafiikat ja podcast-kuvaukset. Jos uskot jonkun käyttävän tekijänoikeudella suojattua teostasi ilman lupaasi, voit seurata tässä https://fi.player.fm/legal kuvattua prosessia.
Dr. Cal Roberts, President and CEO of Lighthouse Guild, the leading provider of exceptional services that inspire people who are visually impaired to attain their goals, interviews inventors, developers and entrepreneurs who have innovative tech ideas and solutions to help improve the lives of people with vision loss.
…
continue reading
36 jaksoa
Kaikki jaksot
×This podcast is about big ideas on how technology is making life better for people with vision loss. Not long ago virtual assistants and VR gaming were the makings of science fiction. But today, as we see developers across the board incorporating these technologies into apps and other tech, including for people who are blind or have low vision, these breakthroughs aren’t just possible, they’re part of the fabric of everyday life. In this episode Dr. Cal talks with Steve Ewell, Executive Director of the Consumer Technology Association Foundation (CTAF). Together they discuss how rapid advances in tech have made previously unimaginable products like self-driving cars, smart homes and virtual reality gaming an attainable, everyday reality for millions of people, including for those who are aging or have low vision. They talk about CES, the CTAF’s annual trade show and the increasing number of consumer companies presenting accessible tech. They also discuss Steve’s work with AARP, and how technology designed for older users is also accelerating… and how those users are keeping up with the pace of new tech. The episode also features an interview with Walt Hickey, the executive editor at Sherwood News and the author of “You Are What You Watch - How Movies and TV Affect Everything.” Walt discusses how scientists and engineers consume the same media as everyone else, and how the ideas from science and science fiction have fed back on one another for millennia to build our modern world. With technology for the blind and low vision community increasingly easy to build and access, companies like Amazon and XBox are making accessibility more of a priority with every new development. In other words, when it comes to accessible technology, the future may have finally arrived. The Big Takeaways The Simplest Tech Can Make a Big Difference: Steve discusses the rise of smart homes, and how something simple like a voice-activated assistant can transform the living space of someone who’s blind or low vision. He talks about the natural evolution from there to doorbell cams, and the possibilities that has opened for people with low vision to maintain their privacy and security while at home. Are Self-Driving Cars Finally Here?: It can seem like self-driving cars have been on the horizon for decades. But with Waymo and other major technology companies recently unveiling breakthroughs, people who are blind or have low vision might finally be able to hop in a car by themselves and go wherever they want. Steve discusses what these breakthroughs mean, not just for activities like visiting the doctor, but for improved opportunities for social interaction. All Fun and Games: Steve talks about his work with AbleGamers and Oak Hill, two organizations dedicated to providing an accessible video game experience for people who are blind or have low vision. Gaming can combat social isolation, he says, but it can do more than that. When paired with wearables, Steve credits accessible gaming (and a healthy dose of competitiveness) with motivating people who are aging or have disabilities to stay active. Old School: It might be surprising to see the words “technology” and “AARP” in the same sentence. But tooday, they are indeed connected. Steve discusses the CTAF’s partnership with AARP, and how recent developments along with a savvier population of seniors has led to increased rates of development and adoption for accessible tech. Tweetables “One of the important things that I've seen that's changed over the last dozen years is the awareness on the company side of the needs of older adults and people with disabilities. That was one of the big challenges, companies recognizing this as a market to engage… and what we've seen is more and more companies are actively engaging this space.” — Steve Ewell, CTAF Executive Director “Technology and movies and pop culture are ways that we express our wants and goals” — Walt Hickey, Author, “You Are What You Watch” “I went to one senior center and a gentleman had come up to me and he said ‘I know you're going to assume I don't use technology, but I want to let you know I'm over a hundred years old and I built all the technology that your technology today is based on.’” — Steve Ewell, CTAF Executive Director Contact Us: Contact us at podcasts@lighthouseguild.org with your innovative new technology ideas for people with vision loss. Pertinent Links Lighthouse Guild CTA Foundation…
O
On Tech & Vision With Dr. Cal Roberts

This podcast is about big ideas on how technology is making life better for people with vision loss. Technology developed specifically for people with low vision has historically been bespoke, specific, and expensive, requiring the creation of new technology for a relatively small population of users. But recently we’re seeing a paradigm shift from building new vision tech from scratch to building on top of existing technology that most people use daily. In this episode, Dr. Cal talks with Karthik Mahadevan, the founder and CEO of Envision, about how his technology company builds on existing tech, like smartphones and smart glasses, to benefit users with low vision. Envision’s software uses AI to recognize important visual cues in a user’s environment and then relays that information to the user with computer-generated speech. Karthik’s breakthrough realization early on that he didn’t need to build hardware when he could just tap into the cameras of smartphones and smart glasses that people already owned, enabling Envision to reach a much wider audience. The episode also features an interview with Troy Otillio, the CEO of Aira. Aira’s platform links users to a human visual interpreter, who can access the camera on their device and guide the user through the task, whether reading a letter, navigating a grocery store, or finding their way around an unfamiliar city. Ultimately, the future of creating vision technology might not be in reinventing the wheel but in improving it and figuring out how to turn the tech people already use every day into easy and accessible solutions for people with low vision. The Big Takeaways Building on Existing Tech: Karthik stresses that for vision tech to be impactful, it has to be accessible. He describes his inspiration as building on top of existing consumer technology to create Envision and thus create a product that millions of people already have the hardware to operate. Later, Dr. Cal discusses the “Taco Bell Phenomenon”, which theorizes that while there may be no such thing as a new idea, the novel combinations of existing ideas are so infinite that they may be limitless. The Importance of a Visual Interpreter: Karthik discusses the limitations of communication formats such as braille and audio announcements in public places for people with vision impairment. While well-meaning, they can never really provide complete access to spaces that are meant to be navigated visually. That’s where Envision comes in, allowing access to the camera on a smartphone or smart glasses to give the user a complete and accurate sense of their environment. Would You Like a Human for That?: To AI or not to AI? Speech-to-text technology is advanced enough that it can excel at reading something simple like a gas bill or a chicken soup recipe. But what if you got a birthday card from your grandchildren? Troy Otillio discusses how Aira recently partnered with Envision to offer their human visual interpreters as part of Envision’s technology — providing camera access to a human agent who can fully understand the nuance of the user’s situation. We Don’t Make Hardware: A small market company like Envision doesn’t have the capacity to build smart glasses (or smartphones) from scratch. The cost would be too high for a company that necessarily has a limited user base. Instead, the future of vision technology is likely to be in creating apps or finding novel uses for hardware that already exists. This will provide accessibility to people with vision impairments via tools they’re already quite familiar with. Tweetables “In the future, these AI are going to interact with other AIs… You know, someone like a Walmart will have their own AI. Someone like a Target will have their own AI. And your AI can speak to that AI and ask for specific information. And that's when it becomes super, super exciting.” — Karthik Mahadevan, Envision Founder & CEO “So for many years, consumer electronics was evolving in one direction, and assistive technology was evolving in a different direction. But what I noticed recently is this merger of the two: that consumer electronics are becoming more accessible.” — Dr. Cal Roberts, Lighthouse Guild President & CEO “You can look at Aira as just a layer, an accessibility layer that exists for when an automated solution, a pure software solution, doesn't work or when there is no solution, right? We're like that, call it a backstop, a more general purpose layer that you can always fall back to.” — Troy Otillio, Aira CEO Contact Us: Contact us at podcasts@lighthouseguild.org with your innovative new technology ideas for people with vision loss. Pertinent Links Lighthouse Guild Envision Aira…
This podcast is about big ideas on how technology is making life better for people with vision loss. Today, self-driving cars are all the buzz when it comes to cutting edge driving technology that could help people with vision impairmentdrive. But there are other, simpler technologies that are already helping people who are blind or visually impaired navigate behind the wheel - some of which draw on technology that’s been around for hundreds of years! In this episode, Dr. Cal talks with Dr. Henry Greene, co-founder and president of Ocutech about their bioptic driving telescopes. These simple devices fit on top of a pair of conventional eyeglasses, and provide a wider field of vision for drivers with vision impairments. This has been shown to increase driver safety and confidence. The episode also features an interview with Dr. Bill Seiple, Lighthouse Guild’s Chief Research Officer. He discusses how other technological advancements that are already in widespread use among all consumers - like GPS - are making it so everyone has to use their vision less while behind the wheel. Ultimately, the ability to pilot a car without relying on our vision isn’t there yet. But with these and other technologies discussed in this episode, the day where none of us need to see to drive may be just around the corner! The Big Takeaways Ancient Tech, Modern Design: Dr. Greene and his team succeeded in adapting telescopic technology that’s existed since Galileo’s time into a bioptic device that fits on top of a conventional pair of eyeglasses. Coupled with other innovations like an auto-focusing device, these bioptic driving assists can vision impaired drivers to see faraway objects with the clarity required for highway driving. Can Something You Use Only 1% of the Time Make You Safer? Ocutech’s bioptic driving telescopes aren’t designed to be used all the time while driving. In fact, Dr. Greene estimates that his patients only use them between one and two percent of the time. This means that while they’re a useful tool for very specific tasks on the road, drivers still need to be competent behind the wheel without bioptics. Per Dr. Seiple, research has shown that drivers with vision impairment tend to be intentional about where and when they drive, adding another layer of safety. GPS Makes Everything Easier: Thanks to 5G technology that can place a user anywhere on Earth within a few feet, GPS is making it so that everyone has to use their vision less when they’re on the road . Is Driving a Visual Thing? As Dr. Seiple points out, visual acuity isn’t of greatest importance behind the wheel. More important is attentiveness, knowing where you’re going, and reaction time once something crosses your way, regardless of how sharply you can see it. This suggests that while cars may soon be able to pilot themselves, getting from one place to another may always require the competent guidance of a human. Tweetables “GPS doesn't tell you that there's a box on the road or a policeman or a pedestrian or an animal or a deviation for construction. GPS won't tell you that stuff. And that's the stuff that you need to be able to anticipate problems and take the appropriate action while you're driving.” — Dr. Henry Greene, Ocutech Co-Founder & President “So, in tongue and cheek, you can say that I don't think driving is a visual thing. Turns out field is much more important than acuity for driving safety.” — Dr. Bill Seiple, Lighthouse Guild Chief Research Officer “These bioptics, they are a game changer for people with low vision… All my life I've actually had to get a ride or wait for someone. And I'm very, very, very independent. I don't like to wait for people..” — Crystal Davis, Ocutech Employee & Bioptic Telescope User Contact Us: Contact us at podcasts@lighthouseguild.org with your innovative new technology ideas for people with vision loss. Pertinent Links Lighthouse Guild Ocutech…
This podcast is about big ideas on how technology is making life better for people with vision loss. When it comes to navigation technology for people who are blind or visually impaired, many apps utilize voice commands, loud tones or beeps, or haptic feedback. In an effort to create a more natural, seamless experience, the team at BenVision has created a different type of system that allows users to navigate using musical cues instead! For this episode, Dr. Cal spoke with BenVision’s CEO and co-founder, Patrick Burton, along with its Technology Leadd, Aaditya Vaze. They shared about the inspiration behind BenVision, how they’re able to create immersive soundscapes that double as navigation aids, and the exciting future applications this technology could offer. The episode also features BenVision’s other co-founder and Audio Director, Soobin Ha. Soobin described her creative process for designing BenVision’s soundscapes, how she harnesses the power of AI, and her bold vision of what’s to come. Lighthouse Guild volunteer Shanell Matos tested BenVision herself and shares her thoughts on the experience. As you’ll hear, this technology is transformative! The Big Takeaways Why Music? Navigation technology that uses voice, tone, or haptics can create an added distraction for some users. But the brain processes music differently. Instead of overloading the senses, for some users music works alongside them, allowing them to single out separate sound cues, or take in the entire environment as a whole. Like how the different instruments correspond to various characters in “Peter and the Wolf,” BenVision assigns unique musical cues to individual objects. User Experience: Shanell Matos appreciated how BenVision blends in more subconsciously, allowing her to navigate a space without having to be as actively engaged with the process. Additional Applications: BenVision began as an augmented reality program, and its creators see a potential for it to grow beyond a navigational tool to expand for use by people who are visually impaired or fully sighted. For example, it could be used to create unique soundscapes for museums, theme parks, and more, augmenting the experience in exciting new ways. The Role of AI: Artificial Intelligence already plays a big role in how BenVision works, and its creators see it being even more important in the future. BenVision already harnesses AI for object detection and its companion app uses AI to provide instant voice support about the immediate surroundings if needed. Moving forward, AI could be used to help instantaneously generate new sound cues or to help users customize their experience at the press of a button. Tweetables “We thought that if the human brain can learn echolocation and we have this amazing technology that’s available to us in the modern day, then why can’t we make echolocation a little bit more intuitive and perhaps a little bit more pleasant.” — Patrick Burton, BenVision CEO & Co-Founder “You can think of it like a bunch of virtual speakers placed at different locations around the user. So like a speaker on a door or a couch or a chair. And then there are sounds coming from all these virtual speakers at the same time.” — Aaditya Vaze, BenVision Technology Lead “I want to gamify this idea so that the user can actually find some interest and joy by using it, rather than just find it only helpful, but also [to create ] some pleasant feeling.” — Soobin Ha, BenVision Audio Director & Co-Founder “So if there’s a lot of people, there’s a lot of conversations happening, a lot of sounds happening, a lot of movement happening. It’s really difficult to keep up with what everything is doing. Whereas with music, it’s not as difficult to pick out layers.” – Shanell Matos, Lighthouse Guild Volunteer Contact Us: Contact us at podcasts@lighthouseguild.org with your innovative new technology ideas for people with vision loss. Pertinent Links Lighthouse Guild BenVision…
We appreciate your support for our show — and now, we need your help nominating the On Tech & Vision podcast for the People’s Choice Podcast Awards! We are participating in these awards so we can showcase On Tech & Vision to a broader audience, gain recognition within the industry, and, most importantly, help spread the message about Lighthouse Guild and the role that technology is playing in tearing down barriers for people who are blind or visually impaired. To help us nominate On Tech & Vision, please go online to www.podcastawards.com , where you can register to vote for On Tech & Vision in both the Technology and Peoples’ Choice Categories. Voting is open until July 31st. Once again, your support is greatly appreciated!…
This podcast is about big ideas on how technology is making life better for people with vision loss. For hundreds of years, health professionals have dreamed of restoring vision for people who are blind or visually impaired. However, doing so, either through transplanting a functioning eye or using technological aids, is an incredibly complex challenge. In fact, many considered it impossible. But thanks to cutting-edge research and programs, the ability to restore vision is getting closer than ever. As a first for this podcast, this episode features an interview with Dr. Cal Roberts himself! Adapting audio from an interview on The Doctors Podcast, Dr. Cal describes his work as a program manager for a project on eye transplantation called Transplantation of Human Eye Allographs (THEA). Funded by a government initiative called ARPA-H, THEA is bringing some of the country’s finest minds together to tackle the complexities of connecting a person’s brain to an eye from a human donor. This episode also features an interview with Dr. Daniel Palanker of Stanford University. Dr. Palanker is working on technology that can artificially restore sight through prosthetic replacement of photoreceptors. Having proved successful in animals, Dr. Palanker and his team are working hard to translate it to humans. And if that can happen, then something once considered impossible could finally be accomplished! The Big Takeaways The Challenges of Eye Transplants: Although eyeball transplants have been done, they’ve only been cosmetic. So far, nobody has been able to successfully connect a donor eyeball to a recipient’s brain. Dr. Roberts’s work with THEA is bringing together multiple teams to tackle the challenges associated with a whole eyeball transplant, from connecting nerves and muscles to ensuring the organ isn’t rejected, and much more. “Artificial” Vision Restoration: Dr. Palanker is working to replace the functions of photoreceptors through technological means. His photovoltaic array is placed underneath the retina and can convert light into an electrical current that activates the cells that send visual information to the brain. While it doesn’t completely restore sight for people with Age-Related Macular Degeneration, this technology shows incredible promise. Decoding “Brain Language”: For both Dr. Roberts and Dr. Palanker, one of the biggest challenges with vision restoration is understanding how the eye and brain communicate. Dr. Roberts likens it to Morse Code — the eye speaks to the brain in “dots and dashes,” which the brain then converts into vision. Right now, the language is still foreign to us, but we’re closer than ever to decoding it. The Evolution of the Brain-Machine Interface: Dr. Palanker imagines incredible possibilities in the interaction between the brain and technology. If we can find a way to truly translate the brain’s signals into information, Dr. Palanker envisions the possibility of direct brain-to-brain communication without verbalization. In a way, this could make people telepathic , able to understand and digest vast amounts of information in an instant. Tweetables: So ideally in medicine, at least the ideal therapy is the restoration of full functionality. If we can grow back photoreceptors and make them reconnect to bipolar cells, undo all the rewiring that right now underwent during degeneration, and restore the full extent of vision, that would be the ideal outcome. — Dr. Daniel Palanker, Professor of Ophthalmology, Stanford University We can think about other aspects of brain-machine interface, which takes you maybe into the realm of capabilities that humans never had. If you enable artificial senses or enable brain-to-brain connectivity so you can communicate without verbalization that would open completely new capabilities that humanity never had. — Dr. Palanker Forty-two years after the implantation of the first mechanical heart, there’s not a single person in the world walking around with a mechanical heart. All that work, all that research, and all that effort to come up with mechanical heart transplants are still state-of-the-art. And so, while I believe that there is a role for a bionic eye or mechanical eye, what I really believe is that everything that we learn from doing an eye transplant will just make it better and easier when we do eventually come up with a bionic or a mechanical eye. — Dr. Calvin Roberts, ARPA-H Health Science Futures Program Manager (President and CEO of Lighthouse Guild and Host of On Tech & Vision !) Contact Us: Contact us at podcasts@lighthouseguild.org with your innovative new technology ideas for people with vision loss. Pertinent Links Lighthouse Guild THEA Program Prima System People’s Choice Podcast Awards We appreciate your support for our show — and now, we need your help nominating the On Tech & Vision podcast for the People’s Choice Podcast Awards! We are participating in these awards so we can showcase On Tech & Vision to a broader audience, gain recognition within the industry, and, most importantly, help spread the message about Lighthouse Guild and the role that technology is playing in tearing down barriers for people who are blind or visually impaired. To help us nominate On Tech & Vision, please go online to www.podcastawards.com , where you can register to vote for On Tech & Vision in both the Technology and Peoples’ Choice Categories. Voting is open until July 31st. Once again, your support is greatly appreciated!…
This podcast is about big ideas on how technology is making life better for people with vision loss. This episode is about how biosensor technology is revolutionizing the field of diagnostic and preventive medicine. Biosensors can take many forms — wearable, implantable, and even ingestible. And they can serve many different functions as well, most notably when it comes to detecting the various pressure levels in our bodies. This episode features interviews with several luminaries working with biosensors. One of them is Doug Adams, a revolutionary entrepreneur who became inspired to create a biosensor that can assist in the treatment of glaucoma patients, initially focusing on a sensor for intraocular pressure. More recently, Doug founded a company called QURA, whose current efforts are focused on a biosensor that detects blood pressure. To elaborate on QURA’s initiatives, this episode also includes insights from its Chief Business Officer, David Hendren. He and Dr. Cal discuss the current state of biosensor technology, the benefits of implantable biosensors, and how they work. Finally, this episode includes a conversation with Max Ostermeier, co-founder and General Manager of Implandata Ophthalmic Products. Max was previously interviewed by Dr. Cal for the episode “Innovations in Intraocular Pressure and Closed Loop Drug Delivery Systems.” This time, Max joins Dr. Cal to discuss the possibilities of biosensor technology and his company’s Eyemate system — which includes biosensor technology for glaucoma patients. All three guests also offer their thoughts on the future of biosensors and their endless possibilities. While it may seem like science fiction, it truly is science reality! The Big Takeaways What Biosensors Do: Currently, biosensors primarily sense the various pressures in the human body. QURA’s current sensor detects blood pressure and assists with hypertension. Meanwhile, Implandata’s Eyemate technology serves glaucoma patients by gathering data on intraocular pressure. The Rapid Shrinking of Biosensors: When Doug Adams first started working on biosensors, the model he saw was the size of a microwave. Now, it’s shrunk to the size of a grain of rice! By making biosensors smaller, they are easier to implant and place in different spots within the body. And by doing so, they can gather more and more data. The Benefits of AI: One drawback of gathering so much data is that it can sometimes be hard to analyze it. However, improvements in AI technology are making it easier to sort through all that data, giving doctors and patients valuable information for medical diagnostics and treatments. The Future of Biosensors: As implantable biosensors become smaller and more sophisticated, all our guests see them becoming a crucial part of healthcare. In addition to gathering data on all sorts of functions within the body, biosensors could provide therapies and treatments with minimal human intervention. Tweetables: So, we are measuring the absolute pressure inside the eye with this kind of technology. It originates from the automotive industry. Tire pressure sensors, where you also have to measure the pressure inside the tire. And so basically we took set technology and advanced it and made it so small that you can also implant this kind of sensor in an eye. — Max Ostermeier, co-founder and General Manager of Implandata Ophthalmic Products So I had a physical a month ago, and along with the physical, they draw blood, and they send that blood off to a lab. I have a feeling in the next decade, that goes away. Why do you have to send a vial of blood to the lab? Because if I had a sensor, not even in an artery, but on top of an artery, I could do a complete analysis of everything in that blood that you’re doing from the lab. — Doug Adams, entrepreneur and founder of QURA The important thing is that you are automatically getting data to the care group that is taking care of these patients, where they are able to see what’s happening. They’re able to see not just a snapshot once in a while, as you’d have from an external pressure cuff, but [get] continuous data longitudinally. — David Hendren, Chief Business Officer of QURA Contact Us: Contact us at podcasts@lighthouseguild.org with your innovative new technology ideas for people with vision loss. Pertinent Links Lighthouse Guild QURA Implandata Ophthalmic Products…
When it comes to emerging technology, there’s no hotter topic than artificial intelligence. Programs like ChatGPT and Midjourney are becoming more popular and are inspiring people to explore the possibilities of what AI can achieve — including when it comes to accessible technology for people who are blind or visually impaired. One of those people is Saqib Shaikh, an engineering manager at Microsoft. Saqib leads the team that developed an app called Seeing AI, which utilizes the latest generation of artificial intelligence, known as generative AI Dr. Cal spoke with Saqib about how Generative AI works, his firsthand experience using an app like Seeing AI, and how it helped improve his daily life. This episode also features Alice Massa, an occupational therapist at Lighthouse Guild. Alice described the many benefits of generative AI, and how it helps her clients better engage in their world. Saqib and Alice also both agreed that the current state of AI is only the beginning of its potential. They shared their visions of what it could achieve in the future — and it doesn’t seem that far off. The Big Takeaways: The Power of Generative AI: Saqib discussed the present condition of artificial intelligence and why generative AI is a massive leap from what came before it. With a deep data pool to draw from, generative AI can do so much more than identify items or come up with an essay prompt. It can understand and interpret the world with startling depth and expediency. Seeing AI: This app can truly put the world in the palm of your hand. It can perform essential tasks like reading a prescription or the sign at a bus stop — and even more than that! It can describe all the colorful details of sea life in a fish tank at the aquarium or help you order dinner off a menu. The app doesn’t just provide people who are blind or visually impaired greater access to the world — it expands it. Embrace Change: There’s understandably a lot of uncertainty about what role AI should play in society. However, Saqib Shaikh and Alice Massa insist that there’s nothing to fear from AI, that the benefits far outweigh any potential drawbacks, and that as long as it’s handled responsibly, there’s a lot AI can do to help improve our lives. Tweetables: “I had a client at the Lighthouse who really was very disinterested in doing anything. The only thing he did on his phone was answer a call from his pastor and call his pastor. And I was able to put Seeing AI on his phone. And his wife said the first time in two years, she saw a smile on his face because now he could read his Bible by himself.” — Alice Massa, Occupational Therapist at Lighthouse Guild “What if AI could understand you as a human? What are your capabilities? What are your limitations at any moment in time? Whether that's due to a disability or your preferences or something else, and understand the environment, the world you're in, or the task you’re doing on the computer or whatever. And then we can use the AI to close that gap and enable everyone to do more and realize their full potential.” — Saqib Shaikh, Engineering Manager at Microsoft “I call my phone my sister because my phone is the person I go to when I’m on the street if I’m walking in Manhattan. The other day I was meeting someone on 47th Street. I wasn’t sure which block I was on. All I did was open Seeing AI short text, hold it up to the street sign, and it told me I was on West 46th Street.” — Alice Massa “Some of the interesting things powered by generative AI is going from taking a photo, say from your photo gallery if you’re reliving memories from your vacation, or even just what’s in front of you right now. It can go from saying it’s a man sitting on a chair in a room to actually giving you maybe a whole paragraph describing what’s in the room, what’s on the shelf, what’s In the background, what’s through the window, even. And it’s just remarkable. I work on this every day. I understand the technology, yet as an end user, I still am surprised and delighted by what this generation of AI is capable of telling me.” — Saqib Shaikh Contact Us: Contact us at podcasts@lighthouseguild.org with your innovative new technology ideas for people with vision loss. Pertinent Links: Lighthouse Guild Together, Tacit Retissa Neoviewer State Tactile Omero Museum Emily Metauten Artist Page (Herminia Blue)…
This podcast is about big ideas on how technology is making life better for people with vision loss. When it comes to art, a common phrase is “look, don’t touch.” Many think of art as a purely visual medium, and that can make it difficult for people who are blind or visually impaired to engage with it. But in recent years, people have begun to reimagine what it means to experience and express art. For this episode, Dr. Cal spoke to El-Deane Naude from Sony Electronics. El-Deane discussed the Retissa NeoViewer, a project developed with QD Laser that projects images taken on a camera directly onto the photographer’s retina. This technology allows people who are visually impaired to see their work much more clearly and with greater ease. Dr. Cal also spoke with Bonnie Collura, a sculptor and professor at Penn State University about her project, “Together, Tacit.” Bonnie and her team developed a haptic glove that allows artists who are blind or visually impaired to sculpt with virtual clay. They work in conjunction with a sighted partner wearing a VR headset, allowing both to engage with each other and gain a new understanding of the artistic process. This episode also includes an interview with Greta Sturm, who works for the State Tactile Omero Museum in Italy. Greta described how the museum’s founders created an experience solely centered around interacting with art through touch. Not only is it accessible for people who are blind or visually impaired, but it allows everyone to engage with the museum’s collection in a fascinating new way. Finally, a painter and makeup artist named Emily Metauten described how useful accessible technology has been for her career. But she also discussed the challenges artists who are blind or visually impaired face when it comes to gaining access to this valuable technology. The Big Takeaways: The Value of Versatility: Many photographers who are visually impaired require the use of large, unwieldy accessories in order to properly capture their work. Sony and QD Laser are determined to solve this problem with the Retissa NeoViewer, which can replace cumbersome accessories like screen magnifiers and optical scopes. Sculpting Virtual Clay: The aim of Together, Tacit, is to “foster creative collaboration between blind, low-vision, and sighted individuals.” A major way this is accomplished is by using the haptic glove to sculpt virtual, rather than physical, clay. Working in VR makes it harder for the sighted partner to unintentionally influence the work of the artist who is blind or visually impaired. As a result, the experience for both users is more authentic and enriching. Reimagining the Museum Experience: The Tactile Omero Museum is much more than an opportunity for people who are blind or visually impaired to interact with art – it’s reimagining how that art is fundamentally experienced. By giving visitors a chance to engage with pieces on a tactile level, the museum allows everyone a chance to reconnect with a vital sense that many take for granted. Expanding Ability to Access Technology: For artists like Emily Metauten who are visually impaired, accessible technology makes it much easier to do their jobs. However, many governmental organizations don’t have the infrastructure to provide this technology to them. Emily wants to raise awareness of how valuable this technology can be, and why providing it to people is so important. Tweetables: “When we’re little kids, we want to touch everything … and then soon after that, we’re told, no, no, no, you shouldn’t touch. You should look and not touch. And so, it becomes the reality and it becomes what you’re supposed to do.” – Greta Sturm, Operator at State Tactile Omero Museum “I carry a Monocular little optical scope. But it becomes extremely difficult when you’re out and about and you’re trying to take a photograph, trying to change your settings. This method, the laser projection, I can actually read, the tiniest little settings.” – El-Deane Naude, Senior Project Manager at Sony Electronics Imaging Division “The VR glasses definitely unlock an ability to see more details more easily for me. Because peripheral vision isn’t designed to see fine details. That's what the central vision is responsible for. So that’s what I have trouble with. But it made what I was already doing easier, and also did give me inspiration. Because we’re trying to unlock the greater things in life, that aren’t just beyond the basics for people with vision loss.” – Emily Metauten, professional painter and makeup artist “I’ve learned through teaching that if a visually impaired or blind person was to use real clay … a sighted person would inevitably start to signify it in terms of what it can be called … And already, immediately, that begins to change the power dynamic on how something is created.” – Bonnie Collura, Professor of Art, Penn State University Contact Us: Contact us at podcasts@lighthouseguild.org with your innovative new technology ideas for people with vision loss. Pertinent Links Lighthouse Guild Together, Tacit Retissa Neoviewer State Tactile Omero Museum Emily Metauten Artist Page (Herminia Blue)…
This podcast is about big ideas on how technology is making life better for people with vision loss. When we buy a product off the shelf, we rarely think about how much work went into getting it there. Between initial conception and going to market, life-changing technology requires a rigorous testing and development process. That is especially true when it comes to accessible technology for people who are blind or visually impaired. For this episode, Dr. Cal spoke to Jay Cormier, the President and CEO of Eyedaptic, a company that specializes in vision-enhancement technology. Their flagship product, the EYE5, provides immense benefits to people with Age-Related Macular Degeneration, Diabetic Retinopathy, and other low-vision diseases. But this product didn’t arrive by magic. It took years of planning, testing, and internal development to bring this technology to market. This episode also features JR Rizzo, who is a professor and researcher of medicine and engineering at NYU — and a medical doctor. JR and his research team are developing a wearable “backpack” navigation system that uses sophisticated camera, computer, and sensor technology. JR discussed both the practical and technological challenges of creating such a sophisticated project, along with the importance of beta testing and feedback. The Big Takeaways: The importance of testing: There’s no straight line between the initial idea and the final product. It’s more of a wheel, that rolls along with the power of testing and feedback. It’s extremely important to have a wide range of beta testers engage with the product. Their experience with it can highlight unexpected blind spots and create opportunities to make something even greater than originally anticipated. Anticipating needs: When it comes to products like the EYE5, developers need to anticipate that its users will have evolving needs as their visual acuity deteriorates. So part of the development process involves anticipating what those needs will be and finding a way to deliver new features as users need them. Changing on the fly: Sometimes, we receive feedback we were never expecting. When JR Rizzo received some surprise reactions to his backpack device, he had to reconsider his approach and re-examine his fundamental design. Future-Casting: When Jay Cormier and his team at Eyedaptic first started designing the EYE5 device, they were already considering what the product would look like in the future, and how it would evolve. To that end, they submitted certain patents many years ahead of when they thought they’d need them — and now, they’re finally being put to use. Tweetables: “I’m no Steve Jobs and I don’t know better than our users. So the best thing to do is give them a choice and see what happens.” — Jay Cormier, President & CEO of Eyedaptic “I started to think a little bit more about … assistive technologies. … And, I thought about trying to build in and integrate other sensory inputs that we may not have natively … to augment our existing capabilities.” — JR Rizzo, NYU Professor of Medicine and Engineering “I think the way we’ve always looked at it is the right way, which is you put the user, the end user, front and center, and they’re really your guide, if you will. And we’ve always done that even in the beginning when we start development of a project.” – Jay Cormier “When we put a 10-pound backpack on some colleagues, they offered some fairly critical feedback that it was way too heavy and they would never wear it. … They were like … it’s a non-starter.” — JR Rizzo Contact Us: Contact us at podcasts@lighthouseguild.org with your innovative new technology ideas for people with vision loss. Pertinent Links Lighthouse Guild Eyedaptic Rizzo Lab…
This podcast is about big ideas on how technology is making life better for people with vision loss. The white cane and guide dogs are long-established foundational tools used by people with vision impairment to navigate. Although it would be difficult to replace the 35,000 years of bonding between humans and dogs, researchers are working on robotic technologies that can replicate many of the same functions of a guide dog. One such project, called LYSA, is being developed by Vix Labs in Brazil. LYSA sits on two wheels and is pushed by the user. It’s capable of identifying obstacles and guiding users to saved destinations. And while hurdles such as outdoor navigation remain, LYSA could someday be a promising alternative for people who either don’t have access to guide dogs or aren’t interested in having one. In a similar vein, Dr. Cang Ye and his team at Virginia Commonwealth University are developing a robotic white cane that augments the familiar white cane experience for people with vision loss. Like the LYSA, the robotic white cane has a sophisticated computer learning system that allows it to identify obstacles and help the user navigate around them, using a roller tip at its base. Although it faces obstacles as well, the robotic guide cane is another incredible example of how robotics can help improve the lives of people who are blind or visually impaired. It may be a while until these technologies are widely available, and guide dogs and traditional canes will always be extremely useful for people who are blind or visually impaired. But with how fast innovations in robotics are happening, it may not be long until viable robotic alternatives are available. The Big Takeaways: Reliability of Biological Guide Dogs: Although guide dogs have only been around for a little over a century, humans and dogs have a relationship dating over 35,000 years. Thomas Panek, the President and CEO of Guiding Eyes for the Blind, points out that there will never be a true replacement for this timeless bond. That being said, he thinks there is a role for robotics to coexist alongside biological guide dogs, and even help augment their abilities. LYSA the Robotic Guide Dog: LYSA may look more like a rolling suitcase than a dog, but its developers at Brazil’s Vix Systems are working on giving it many of the same functions as its biological counterpart. LYSA can identify obstacles and guide its user around them. And for indoor environments that are fully mapped out, it can bring the user to pre-selected destinations as well. The Robotic White Cane: Dr. Cang Ye and his team at Virginia Commonwealth University are developing a Robotic White Cane that can provide more specific guidance than the traditional version. With a sophisticated camera combined with LiDAR technology, it can help its user navigate the world with increased confidence. Challenges of Outdoor Navigation: Both LYSA and the Robotic White Cane are currently better suited for indoor navigation. A major reason for that is the unpredictability of an outdoor environment along with more fast-moving objects, such as cars on the road. Researchers are working hard on overcoming this hurdle, but it still poses a major challenge. The Speed of Innovation: When Dr. Ye began developing the Robotic White Cane a decade ago, the camera his team used cost $500,000 and had image issues. Now, their technology can be run on a smartphone – making the technology much more affordable, and hopefully one day, more accessible if it becomes available to the public. Tweetables: “We’ve had a relationship with dogs for 35,000 years. And a relationship with robots for maybe, you know, 50 years. So the ability of a robot to take over that task is a way off. But technology is moving quickly.” — Thomas Panek, President and CEO of Guiding Eyes for the Blind “Outdoor navigation is a whole new world because if you go on the streets, it could be dangerous. You have to be very careful because you are driving a person, driving a human being.” — Kaio Ribeiro, Researcher at Vix Systems “The first … camera we used, it's close to 500 grand. … But now … the iPhone’s LiDAR … works outdoors. … And it just took … a … bit more than 10 years.” — Dr. Cang Ye, Prof. of Comp. Sci. at Virginia Commonwealth University and Program Director, National Science Foundation “It’s not the traditional … robot … that’s stiff. … We have to move into soft robotics … to accomplish the … activity … a dog can accomplish. … It’s a way off. … If … an engineering student … wants to get into soft robotics, … that’s where it will be.” — Thomas Panek Pertinent Links: Lighthouse Guild Guiding Eyes for the Blind LYSA Robot Guide Robotic White Cane…
O
On Tech & Vision With Dr. Cal Roberts

1 Smart Cities and Autonomous Driving: How Technology is Providing Greater Freedom of Movement for People with Vision Loss 39:13
This podcast is about big ideas on how technology is making life better for people with vision loss. Navigating the world can be difficult for anyone, whether or not they have vision loss. Tasks like driving safely through a city, navigating a busy airport, or finding the right bus stop all provide unique challenges. Thankfully, advances in technology are giving people more freedom of movement than ever before, allowing them to get where they want, when they want, safely. Smart Cities are putting data collection to work in a healthy way by providing information to make busy intersections more secure, sidewalks more accessible, and navigation more accurate. They’re providing assistance for all aspects of travel, from the front door to the so-called “last hundred feet,” while using automated technology to make life easier every step of the way. And although fully autonomous vehicles are still on the horizon, the technology being used to develop them is being applied to improve other aspects of life in incredible ways. These applications are making the world more accessible, safer, and better for everyone, including people who are blind or visually impaired. One example of this is Dan Parker, the “World’s Fastest Blind Man,” who has developed sophisticated guidance systems for his racing vehicles, as well as a semi-autonomous bicycle that could give people with vision loss a new way to navigate the world safely and independently. The Big Takeaways: Smart Cities. Greg McGuire and his team at MCity in Ann Arbor, Michigan are working on the concept of Smart Cities, which focus on using data to improve the everyday lives of their citizens. That means improving traffic intersection safety, greater accessibility options, providing detailed “last hundred feet” guidance, and much more. Autonomous Driving. In a perfect world, self-driving cars will provide ease of transportation for everyone, and create safer, less congested roads. That technology isn’t there yet – but it’s being worked on by talented researchers like John Dolan, the Principal Systems Scientist at Carnegie Mellon’s Autonomous Driving Vehicle Research Center. Sophisticated sensors and advanced robot-human interfaces are being developed to make self-driving cars possible. Application of Technology. Even though the technologies behind Smart Cities and autonomous vehicles are still being developed, they can still be applied to everyday life in exciting ways. Things like miniature delivery robots that can deliver goods, AI-powered suitcases that can help you navigate busy airports, or semi-autonomous bicycles are already here – and there’s more on the way. The World’s Fastest Blind Man. When professional race car driver Dan Parker lost his vision in an accident, he felt lost. But a moment of inspiration led him and his business partner Patrick Johnson to develop a sophisticated guidance system that let him continue racing without human assistance. Thanks to this revolutionary technology, Dan became the “World’s Fastest Blind Man” when he set a land-speed record of 211.043 miles an hour in his customized Corvette. Tweetables: “One of the key pillars of MCity is accessibility. The four areas we think about are safety, efficiency, equity, and accessibility. … Accessibility is that we can make transportation systems available to as many of us as possible.” – Greg McGuire, Managing Director of MCity “I became the first blind man to race Bonneville, with an average speed of 55.331 mph. And I returned in 2014 and set my official FIM class record … at 62.05 mph. … I’m the only blind land speed racer … with no human assistance.”– Dan Parker, the “World’s Fastest Blind Man” “There are chairs, there are tables. ... We know we don’t want to run into them, but we do want to walk in the walkable space. … A car wants to drive in the drivable space.” – John Dolan, Principal Systems Scientist at Carnegie Mellon’s Autonomous Driving Vehicle Research Center “Because we know autonomous technology is increasing every day and it’s coming, you know, a hundred percent it’s coming. You know, transportation is freedom and that’s exactly what that would bring us. Freedom.” – Dan Parker Contact Us: Contact us at podcasts@lighthouseguild.org with your innovative new technology ideas for people with vision loss. Pertinent Links: Lighthouse Guild MCity Carnegie Mellon Autonomous Driving Vehicle Research Center Dan Parker…
O
On Tech & Vision With Dr. Cal Roberts

1 Leveling Up Accessible Video Game Features: How New Technology is Making Gaming More Immersive and Inclusive for People with Vision Loss 31:06
This podcast is about big ideas on how technology is making life better for people with vision loss. For decades, people with vision loss had limited options when it came to accessing video games. Aside from screen magnification and text-to-voice tools, gamers who are blind or visually impaired didn’t have many ways to play their favorite titles. But in recent years, the same cutting-edge technology used to create games has been used to also make them more accessible for people with vision impairment. These advances include more visibility options, the implementation of 3D audio, haptic feedback, and customizable controllers for gamers with vision impairment. Furthermore, 3D audio technologies being developed in live sports may soon make their way to online multiplayer video games. The implementation and improvement of these technologies mean that everyone will be able to play together, regardless of their visual acuity. The Big Takeaways: Leap in Accessible Gaming Options. In the past, the lack of accessibility features has made it much harder for gamers like Elaine Abdalla to access her favorite titles. But as gaming technology and accessibility improve, gamers like Elaine who are visually impaired are now able to more actively participate in their favorite titles. For example, The Last of Us: Part 2 boasts over 60 accessibility options including sophisticated visibility adjustments and inclusive difficulty settings. Participating in the Process. Robin Spinks is the Head of Inclusive Design at the UK’s Royal National Institute of Blind People (RNIB). He helps gaming companies implement accessibility options as part of the design process from the very beginning. By ensuring people with vision loss are part of the process from the jump, Robin is helping make games more immersive than ever — for everyone. Xbox Accessibility Team. Kaitlyn Jones first became passionate about gaming when she helped her father launch a nonprofit foundation that built customized setups for gamers with disabilities. She’s now the program manager of Xbox’s Gaming Accessibility Team. They develop guidelines for all kinds of accessible gaming options, including for people with vision loss. Thanks to her team’s efforts, new titles include options like spatial audio cues, which make it easier for gamers who are blind to navigate complex dungeons and unlock achievements. Action Audio. At the 2022 Australian Open, a new technology called Action Audio became available for tennis fans with vision loss. Through cutting-edge camera tracking technology, Action Audio creates an information-rich 3D audio experience that allows tennis fans with vision loss to experience every thrilling rally as it happened, and to share in a communal experience. Spatial Audio in Gaming. Action Audio’s developers hope to make it a universal technology that can also be used in gaming. Whether it’s working with a team in the latest first-person shooter or dribbling around opponents on a virtual basketball court, 3D (or spatial) audio technology is positioned to help people with vision loss to more equitably participate in the gaming community. Tweetables: “The motivation here is to remove barriers, to form partnerships, and to collaborate so that going forward it becomes a much more possible area of life for people. ’Cause gaming's fantastic, right? And why should you be stopped from enjoying it just because you’re blind or partially sighted?” — Robin Spinks, Head of Inclusive Design at RNIB “When something happens on the court, it’s captured quite fast with Hawkeye within hundreds of milliseconds, and we’re able to take that data relatively quickly and generate the audio and then send it to the broadcast before it’s received in the real world.” — Tim Devine, AKQA Executive Director of Innovation “It’s like that Japanese principle of continuous improvement, kaizen , where you’re kind of constantly looking to do things better. That’s what we want to see in the gaming world. — Robin Spinks, Head of Inclusive Design at RNIB “We’ve always really prioritized accessibility along the way. But I think in terms of the journey of, even where we started when I first joined the team a few years ago versus now, I think the bar honestly just keeps getting higher and higher.” — Kaitlyn Jones, Program Manager: Xbox Accessibility Team “I’m just happy to see the disabled community is finally getting the assistance they need. It’s taken a while and it’s still gonna take a while, but we’re going in the right direction.” — Elaine Abdalla, creator of BlindGamerChick YouTube Channel Contact Us: Contact us at podcasts@lighthouseguild.org with your innovative new technology ideas for people with vision loss. Pertinent Links: Lighthouse Guild RNIB Accessible Video Games Page Xbox Accessibility Guidelines BlindGamerChick YouTube Channel On Tech & Vision: Training the Brain: Sensory Substitution…
O
On Tech & Vision With Dr. Cal Roberts

1 A Celebration of Sound and Song: Music Tech Shines the Spotlight on Musicians with Vision Loss 34:40
This podcast is about big ideas on how technology is making life better for people with vision loss. Marcus Roberts, Stevie Wonder, Ray Charles, and even Louis Braille (who invented the Braille Music Notation system still used today) prove that musicians who are blind or visually impaired have made profound impacts on our musical landscape. However, to get their work to us, musicians who are blind have had to structure complex workarounds, like relying on sighted musicians to demonstrate complex scores; memorizing long pieces; or only performing when they can have a Braille score in front of them, shutting them out from opportunities that fall to those who can sight read, since Braille scores have often been time-consuming and expensive to produce. However, new technologies in music composition and production are making composition, nuanced scoring, and Braille printing easier than ever, bringing musicians and composers who are blind to centerstage to share their sound and song. The Big Takeaways: “Lullay and Lament” by James Risdon. The recorder — pivotal in music from the Renaissance and Baroque periods — only lately has emerged from a long period of obsolescence. James Risdon, a passionate player who lost his vision due to Leiber’s congenital amaurosis as a child, has written the original song for recorder “Lullay and Lament” for his album Echoes of Arcadia, which marries the early recorder with contemporary recorder music. To make this album, he relied on new musical technologies like Dancing Dots. Dancing Dots with Bill McCann. Bill McCann is the founder and president of Dancing Dots Braille Music Notation software. Dancing Dots is a suite of software — plus educational resources and training — that helps people who are blind to read, write and record their music. McCann founded Dancing Dots in 1992. Chris Cooke and PlayHymns.com. When Chris Cooke got the Dancing Dots software in 2016, her creativity exploded. She was able to arrange a song and bring it to church and play a duet with a member of the congregation, something she hadn’t been able to do before, given the former time and cost of translating music into Braille notation. What is Braille Music? Louis Braille, a noted musician, created the Braille musical notation system. Being able to translate music easily between Braille Musical Notation and Western musical notation, and to easily print either of these, helps musicians who are blind share music with other musicians, both sighted and blind, and play music together with ease. Musical Instrument Digital Interface and MusicXML. MIDI has made it possible for musicians to play music into their instruments and have it automatically translated into digital musical notation. MusicXML has made universal the file type for a score and allows musicians to share scores across popular music notation applications like Finale or Sebelius. The question of parity. James and Chris agree that while Dancing Dots technology has enabled them to take advantage of new musical opportunities, no technology exists that offers them complete parity with sighted musicians because musicians who are blind need additional lead time to get the music scanned correctly or to memorize pieces. Chris adds that preparing the music in a timely fashion and on a budget would help. The MIDI-to-brain connection. Bill McCann has explored using the Brainport, a technology from WICAB (which we profiled in an episode from September 2021, on “ Training the Brain: Sensory Substitution ”) to allow musicians who are blind to read music on their tongues. This is important if someone needs to read music live in a performance in order to play any instrument that also requires their hands. Early trials showed signs of success. He posits that maybe someday, maybe soon, people could think new music into notation. Tweetables: “I said, I will never put myself in this position again. If I write something and I am asking other people to play it, and they ask me questions or there’s something I am going to know before we meet exactly what I want and what I have.” — Bill McCann, founder and president of Dancing Dots “So the duet that we played was hot off my printer and went with me and we played it. And it was great to be able to share music in that way because of the technology in the Dancing Dots program.” — Chris Cooke, musician and music arranger, creator of Playhymns.com “As a blind person, I can say this for myself, often we end up following sighted people, or following somebody. Braille music gives blind musicians the chance to become leaders.” — Bill McCann, founder and president of Dancing Dots “I've set aside 2023 as a year. I'd really like to kind of develop some more expertise in the area and also come to grips with some of the technology that would help the process.” — James Risdon, musician and recorder player “Someone sitting there and getting inspired is what we call the MIDI-to-Brain connection. We’re not there yet, but you could … think the music in your head … at a computer and … music materializes in the form of a score.” — Bill McCann, founder and president of Dancing Dots Contact Us: Contact us at podcasts@lighthouseguild.org with your innovative new technology ideas for people with vision loss. Pertinent Links: Lighthouse Guild Dancing Dots James Risdon Chris Cooke and Playhymns.com…
O
On Tech & Vision With Dr. Cal Roberts

This podcast is about big ideas on how technology is making life better for people with vision loss. Lots of people have voice-controlled smart home assistants like Siri, Google, or Alexa in their homes…. to listen to the news or to set timers. But they can do so much more! David Frerichs, Principal Engineer, Alexa Experience at Amazon on the aging and accessibility team, shares his design philosophy for making voice assistants more inclusive, and the preferred mode of engagement for every user. He also shares that the next stage of smart home assistants will be ambient computing, where your devices will intuit your needs without you speaking them. We talk with Lighthouse Guild client Aaron Vasquez, who has outfitted his home with smart home technology, and with Matthew Cho, a client who traveled to the Johnson Space Center in Houston to speak to the unmanned Orion Spacecraft via the Amazon Alexa on board, demonstrating that voice assistant technology can bring inclusivity and accessibility to many jobs and industries and are not just for the home anymore. The Big Takeaways: Alexa Onboard the Orion Spacecraft. NASA partnered with Amazon to use the Alexa voice-controlled smart assistant onboard the unmanned Orion spacecraft so that engineers could guide the spacecraft from Mission Control. This project tested the possibility of Alexa for space travel while demonstrating that voice-controlled smart assistants have uses beyond the home. Matthew Cho, an 18-year-old student and client of Lighthouse Guild had the opportunity to travel to Houston to be one of the volunteers to give voice commands to the spacecraft via an Amazon Alexa device while it hurtled through space. Accessibility and Preferences . David Frerichs, Principal Engineer, Alexa Experience at Amazon, and someone who works on the aging and accessibility team , has spent his career developing technology that adapts to the ever-changing needs of the user and has cultivated a design philosophy that makes clear that design choices (like voice control) that enable inclusivity for people who are blind can also become the preferred way that most users engage with a device or a tool. Curb cuts are a historical example. David often thinks in terms of “Hot tub safe computing.” What can a person do to engage with the device from their hot tub? Ambient Computing. David mentions ambient computing, the next phase of smart home technology, in which a network of devices in the built environment intuit and respond to a user's needs without the user even needing to speak a command. Smart Homes Today, Smart Industries Tomorrow. Aaron Vasquez is a smart home user. He uses Google Echo to power two smart lamps, operate his smart TV, and control a pet camera to oversee his rambunctious kitties when he’s not at home. As a person who is visually impaired, Aaron prefers voice command for running his home in this way. This episode asks how the smart home’s tools can be integrated into offices and industries to make these more accessible and inclusive spaces. Tweetables: “To be able to have a rocket be dependent on an AI without anybody having to control the spacecraft is, it is really amazing, and I feel that later on that they’ll be able to use it for much more things aside from space.” – Matthew Cho, student and client of Lighthouse Guild “So their goal was to eventually get different people to be able to go into space … They were trying to see if Alexa would work properly with all sorts of people, normal people. Not just astronauts, like regular, ordinary, everyday people.” – Matthew Cho, student and client of Lighthouse Guild “You have … permanent need, temporary need, situational need, and preferential need that really can inform us on … how we can address [a] barrier for the particular core use case. But if we do it well, it will serve a much broader community.” – David Frerichs, Principal Engineer, Alexa Experience at Amazon “We’re moving toward … ambient computing. That is … that the system should be able to respond to the needs of the customer, even if the customer doesn’t say anything. ... That’s … where the boundaries are and where it’s gonna continue to be pushed.” – David Frerichs, Principal Engineer, Alexa Experience at Amazon “We had seen this pack of smart bulbs and they were relatively cheap, and I was like, huh, that’s, that’s kind of cool. So we were like, you know what? Let’s get it. So we got ‘em, we hooked them up and that kind of is what started everything.” – Aaron Vasquez, smart home user and client of Lighthouse Guild “Honestly, it’s so much easier if I can ask Google a question and she can come up with the answer, then I’m better off that way instead of actually trying to look it up myself.” – Aaron Vasquez, smart home user and client of Lighthouse Guild Contact Us: Contact us at podcasts@lighthouseguild.org with your innovative new technology ideas for people with vision loss. Pertinent Links: Lighthouse Guild Alexa in Space David Frerichs…
Tervetuloa Player FM:n!
Player FM skannaa verkkoa löytääkseen korkealaatuisia podcasteja, joista voit nauttia juuri nyt. Se on paras podcast-sovellus ja toimii Androidilla, iPhonela, ja verkossa. Rekisteröidy sykronoidaksesi tilaukset laitteiden välillä.