Sorry, but there's really nothing to see here right now. Our server is in maintenance mode – and our
little bot is trying to sort the error messages with its last bit of energy.
Whether it was a data leak, a power outage in the cluster, or just a grumpy bit – we don't know yet.
But what we do know is:
The data drama escalated into a server meltdown. But don't worry: We're on it.
What now?
Everything is surely running smoothly over at our colleagues from inside
digital – why not stop by!
Or you could pass the time by visiting our social media channels – for example Instagram, YouTube or TikTok. At least there are no
server problems there – promised.
Thanks for your patience. Our bot is already rebooting at full speed. 🔧
NewsTabletApple Adds Live Speech and AI-Based Voice Generation in iOS 17
Apple is previewing several iOS 17 features that are focused on enhancing speech, vision, and cognitive accessibility on iPhones and iPads. Among the improvements are text-to-speech and AI-based personal voice generation, which are expected to roll out alongside Assistive Access later this year.
Apple continues to build on existing accessibility features and introduced new capabilities simultaneously. Both Live Speech and Assistive Access are entirely new functions while Point and Speak is an update to Apple’s built-in Magnifier app that also received Door Detection in iOS 16 last year.
The Live Speech feature is the iPhone manufacturer’s version of text-to-speech that will be integrated in FaceTime and regular phone calls. Apple claims users can type in text and the feature will read it aloud to the person or group they’re in the call with. Apple claimed it will arrive on iPhones, iPad, Macs, and MacBooks.
Meanwhile, the other speech-related functionality is Personal Voice, which relies on machine learning on the particular device. It lets users create or synthesize their voice by reading prompted texts for 15 minutes. This will come in handy for those who are at risk of losing their voice due to conditions like ALS or amyotrophic lateral sclerosis. The technology behind it resembles Amazon’s AI voice generation which was announced last year.
Assistive Access simplifies the Apple app experience
Apple’s new Assistive Access provides a customized app experience. Rather than expanding controls and input options to apps, the feature simplifies the use of these apps for senior citizens and people who have cognitive and visual disabilities by offering enlarged buttons and text. It supports apps like the main camera, messages, and phone as well as Apple Music and Photos.
In addition to detecting people, doors, and reading out image descriptions, the Magnifier on iPhones can now identify more objects via Point and Speak. This feature should provide better help to visually impaired individuals. It works with short labels that are found in appliances. For example, a user can point at a button on a microwave and the app will read the said text out.
There are also other accessibility-related updates and features. For those who have motor disabilities, Apple’s Switch Control will turn a switch into a virtual game controller. Furthermore, Apple mentioned that hearing aid accessories made for iPhones will now be compatible with Mac devices. Both Voice Control and Voice Over via Siri will receive phonetic suggestions and speed customizations, respectively.
Apple didn’t confirm whether all of these improvements will debut in iOS 17 or separately as a minor update within the next software iteration. Likewise, we want to know which features do you find to be most helpful?
We mark partner links with this symbol. If you click on one of these links or buttons–or make a purchase through them–we may receive a small commission from the retailer. This doesn’t affect the price you pay, but it helps us keep nextpit free for everyone. Thanks for your support!
0 comments