In response to Global Accessibility Awareness Day on May 5, Apple launchedMany accessibility featuresLater, Google also launched theMultiple accessibility features on Android and Chrome.
Among them, Google has integrated Gemini artificial intelligence technology into the Android screen reader "TalkBack" function. Even if the content displayed on the screen does not have specific text content, it can still generate image descriptions through artificial intelligence. It can even further ask questions based on the description content, such as answering the details of the person's clothing and accessories in the image. In addition, this function can also be used to inquire about the material of the products presented by the shopping service and whether there are additional discounts.
The new "Expressive Captions" subtitle function further uses artificial intelligence technology to analyze the tone of voice of characters in the video, and allows the generated subtitles to interpret more "emotions" through additional descriptive methods. For example, when the game commentator excitedly expresses the goal and shouts "wonderful score", the generated subtitles will add punctuation marks such as "~~~~!" or relevant sound tags to present different commentary emotions, allowing more hearing-impaired people to feel the commentator's excitement at the moment through subtitles.
However, this feature will only be available in English-speaking regions, including the United States, the United Kingdom, Canada, and Australia, and can be used on devices running Android 15 or above.
The "Euphonia" project, launched in 2019, allows non-native speakers to use speech recognition smoothly. For example, non-English speakers can use voice commands using unskilled English pronunciation and still have the speech recognition function work properly. The "Euphonia" project is now available as open source code on GitHub, allowing more people to create customized applications and services or train different speech recognition effects.
In addition, Google also assisted University College London in establishing the Center for Digital Language Inclusion (CDLI) through its Google.org unit to improve speech recognition technology for non-English speakers in Africa. This included establishing open source datasets in 10 African languages and building new speech recognition models so that they can be applied to more speech recognition application services.
Other updates include allowing people with disabilities to operate Chromebooks through facial expressions, adding more customization options to reading mode, and being able to view on-screen content through the screen reader "TalkBack" function or watch videos with subtitles. OCR optical character recognition technology makes it easier for the Chrome browser to recognize PDF file content, and even further select, copy or search text content in PDF files. In addition, the automatic magnification effect allows visually impaired people to more easily view web content through the Chrome browser.





