in reply to ondrosik

@ondrosik I think this is a difference how #friendica generates the home timeline feed as compared to mastodon and other common #fediverse servers. Friendica also includes repies in the home timeline feed. There are apps that can filter out posts that are replies to other posts client side however TW Blue may not be able to do that I think.

Do you know that you can use Subtitle edit to transcribe audio? It has a relatively accessible guy so you can use Purfwiev's faster whisper xxl, cpp, cpp cublas, const-me. Longer post how to use it follows:

Installing Subtitle Edit


Download the program from the developer’s website. Navigate to the level 2 heading labeled “Files.”
If you want to install Subtitle Edit normally, download the first file, labeled setup.zip.
There is also a portable version available, labeled SE_version_number.zip.

If you decide to use the portable version, extract it and move on to the next section of this article. The installation itself is standard and straightforward.

A Note on Accessibility


NVDA cannot automatically obtain focus in lists.
To find out which item in the list is currently selected, move down with the arrow key to change the item, then press NVDA+TAB to hear which one is focused.

Initial Setup


  • In the menu bar, go to Video and activate Audio to text (Whisper).
  • When using this feature for the first time, the program may ask whether you want to download FFMPEG. This library allows Subtitle Edit to open many audio and video files, so confirm the download by pressing Yes.
  • Subtitle Edit will confirm that FFMPEG has been downloaded and then ask whether you want to download Purfwiev’s Faster Whisper – XXL. This is the interface for the Whisper model that we’ll use for transcription, so again confirm by pressing Yes.
  • The download will take a little while.
  • Once it’s complete, you’ll see the settings window. Press Tab until you reach the Languages and models section. In the list, select the language of your recording.
  • Press Tab to move to the Select model option, and then again to an unlabeled button.
  • After activating it, choose which model you want to use. Several models are available:
    • Small models require less processing power but are less accurate.
    • Large models take longer to transcribe, need more performance and disk space, but are more accurate.
      I recommend choosing Large-V3 at this step.


  • Wait again for the model to finish downloading.


Transcribing Your First Recording


  • Navigate to the Add button and press Space to activate it.
  • A standard file selection dialog will open. Change the file type to Audio files, find your audio file on the disk, and confirm.
  • Activate the Generate button.
  • Now, simply wait. The Subtitle Edit window doesn’t provide much feedback, but you can tell it’s working by the slower performance of your computer—or, if you’re on a laptop, by the increased fan noise.
  • When the transcription is done, Subtitle Edit will display a new window with an OK button.


We Got Subtitles, So One More Step


In the folder containing your original file, you’ll now find a new file with the .srt extension.
This is a subtitle file—it contains both the text and the timing information. Since we usually don’t need timestamps for transcription, we’ll remove them in Subtitle Edit as follows:

  • Press Ctrl+O (or go to File → Open) to bring up the standard open file dialog. Select the .srt file you just got.
  • In the menu bar, open File → Export → Plain text.
  • Choose Merge all lines, and leave Show line numbers and Show timecode unchecked.
  • Press Save as and save the file normally.

If you’re transcribing multiple recordings, it’s a good idea to close the current subtitle file by starting a new project using Ctrl+N or by choosing File → New.

Conclusion


Downloaded models can, of course, be reused, so future transcriptions will go faster.
In this example, I used Purfwiev’s Faster Whisper. If you want to use a different model, you can select it from the model list, and Subtitle Edit will automatically ask whether you’d like to download it.

Peter Vágner reshared this.

During last 3 months I am using VDO ninja for all my remote interwiev and podcast recordings. here is my article about it from the blind perspective, focused on accessibility and audio.

Have You Ever Wanted to Record an Interview or Podcast Online? You’ve probably faced a few challenges:
How to transmit audio in the highest possible quality?
How to connect in a way that doesn’t burden your guest with installing software?
And how to record everything, ideally into separate tracks?

The solution to these problems is offered by the open-source tool VDO Ninja.

What Is VDO Ninja


It’s an open-source web application that uses WebRTC technology. It allows you to create a P2P connection between participants in an audio or video call and gives you control over various transmission parameters.
You can decide whether the room will include video, what and when will be recorded, and much more.

In terms of accessibility, the interface is fairly easy to get used to — and all parameters can be adjusted directly in the URL address when joining.
All you need is a web browser, either on a computer or smartphone.

Getting Started


The basic principle is similar to using MS Teams, Google Meet, and similar services.
All participants join the same room via a link.
However, VDO Ninja distinguishes between two main types of participants: Guests and the Director.
While the guest has limited control, the director can, for example, change the guest’s input audio device (the change still must be confirmed by the guest).

A Few Words About Browsers


VDO Ninja works in most browsers, but I’ve found Google Chrome to be the most reliable.
Firefox, for some reason, doesn’t display all available audio devices, and when recording multiple tracks, it refuses to download several files simultaneously.

Let’s Record a Podcast


Let’s imagine we’re going to record our podcast, for example, Blindrevue.
We can connect using a link like this:

https://vdo.ninja/?director=Blindrevue&novideo=1&proaudio=1&label=Ondro&autostart=1&videomute=1&showdirector=1&autorecord&sm=0&beep

Looking at the URL more closely, we can see that it contains some useful instructions:
  • director – Defines that we are the director of the room, giving us more control. The value after the equals sign is the room name.
  • novideo – Prevents video from being transmitted from participants. This parameter is optional but useful when recording podcasts to save bandwidth.
  • proaudio – Disables effects like noise reduction, echo cancellation, automatic gain control, compression, etc., and enables stereo transmission.
    Be aware that with this setting, you should use headphones, as echo cancellation is disabled, and otherwise, participants will hear themselves.
  • label=Ondro – Automatically assigns me the nickname “Ondro.”
  • autostart – Starts streaming immediately after joining, skipping the initial setup dialog.
  • videomute – Automatically disables the webcam.
  • showdirector – Displays our own input control panel (useful if we want to record ourselves).
  • autorecord – Automatically starts recording for each participant as they join.
  • sm=0 – Ensures that we automatically hear every new participant without manually unmuting them.
  • beep – Plays a sound and sends system notification when new participants join (requires notification permissions).

For guests, we can send a link like this:

https://vdo.ninja/?room=Blindrevue&novideo=1&proaudio=1&label&autostart=1&videomute=1&webcam

Notice the differences:
  • We replaced director with room. The value must remain the same, otherwise the guest will end up in a different room.
  • We left label empty — this makes VDO Ninja ask the guest for a nickname upon joining.
    Alternatively, you can send personalized links, e.g., label=Peter or label=Marek.
  • The webcam parameter tells VDO Ninja to immediately stream audio from the guest’s microphone; otherwise, they’d need to click “Start streaming” or “Share screen.”


How to Join


Simply open the link in a browser.
In our case, the director automatically streams audio to everyone else.
Participants also join by opening their link in a browser.
If a nickname was predefined, they’ll only be asked for permission to access their microphone and camera.
Otherwise, they’ll also be prompted to enter their name.

Usually, the browser will display a permission warning.
Press F6 to focus on it, then Tab through available options and allow access.

Controls


The page contains several useful buttons:

  • Text chat – Toggles the text chat panel, also allows sending files.
  • Mute speaker output – Mutes local playback (others can still hear you).
  • Mute microphone – Mutes your mic.
  • Mute camera – Turns off your camera (enabled by default in our example).
  • Share screen / Share website – Allows screen or site sharing.
  • Room settings menu (director only) – Shows room configuration options.
  • Settings menu – Lets you configure input/output devices.
  • Stop publishing audio and video (director only) – Stops sending audio/video but still receives others.


Adjusting Input and Output Devices


To change your audio devices:

  1. Activate Settings menu.
  2. Press C to jump to the camera list — skip this for audio-only.
  3. Open Audio sources to pick a microphone.
  4. In Audio output destination, select your playback device. Press test button to test it.
  5. Close settings when done.


Director Options


Each guest appears as a separate landmark on the page.
You can navigate between them quickly (e.g., using D with NVDA).

Useful controls include:

  • Volume slider – Adjusts how loud each participant sounds (locally only).
  • Mute – Silences a guest for everyone.
  • Hangup – Disconnects a participant.
  • Audio settings – Adjusts their audio input/output remotely.


Adjusting Guest Audio


Under Audio settings, you can:

  • Enable/disable filters (noise gate, compressor, auto-gain, etc.).
  • View and change the guest’s input device — if you change it, a Request button appears, prompting the guest to confirm the change.
  • Change the output device, useful for switching between speaker and earpiece on mobile devices.


Recording


Our URL parameters define automatic recording for all participants.
Recordings are saved in your Downloads folder, and progress can be checked with Ctrl+J.

Each participant’s recording is a separate file.
For editing, import them into separate tracks in your DAW and synchronize them manually.
VDO Ninja doesn’t support single-track recording, but you can use Reaper or APP2Clap with a virtual audio device.

To simplify synchronization:

  1. Join as director, but remove autorecord.
  2. Wait for everyone to join and check audio.
  3. When ready, press Alt+D to edit the address bar.
  4. Add &autorecord, reload the page, and confirm rejoining.
  5. Recording now starts simultaneously for everyone.
  6. Verify this in your downloads.


Manual Recording


To start recording manually:

  1. Open Room settings menu.
  2. Go to the Room settings heading.
  3. Click Local record – start all.
  4. Check PCM recording (saves WAV uncompressed).
  5. Check Audio only (records sound without video).
  6. Click Start recording.


Important Recording Notes


  • Always verify that all guest streams are recording.
  • To end recordings safely, click Hangup for each guest or let them leave.
  • You can also toggle recording for each guest under More options → Record.
  • Files are saved as WEBM containers. If your editor doesn’t support it, you can convert them using the official converter.
  • Reaper can open WEBM files but may have editing issues — I prefer importing the OPUS audio file instead.


Recommended Reading


In this article, I’ve covered only a few features and URL parameters.
For more details, check the VDO Ninja Documentation.

1 / 2: Did you know @GNOME Files aka #nautilus has a nifty feature where it can batch rename files? Advanced features include adding sequential numbering, using placeholders and doing search and replace on the names of selected files. #ScreenReader #a11y is preserved.
In order to use it just select multiple files and find Rename item in the shift+F10 popup menu or simply press F2. Also... Don't be shy to press the add button in the batch rename dialog.

Yeah, I've updated my @Arch Linux to @GNOME 49.
There are some nifty #a11y related tweaks such as better labelling for gnome shell menus, refreshed settings UI, I like how presentation of various lists e.g. List of wireless networks is presented with screen reader including signal strength.

Thanks to everyone involved for the improvements.

Hello @GrapheneOS screen-reader users and other #a11y friends,

There was an interesting debate going on at the end of may where screen reader users were asking for #tts engine included with GrapheneOS base system.
grapheneos.social/@GrapheneOS/…

I understand this is very unlikely to change in the near future as I am not aware of a TTS system that is open-source and modern enough to be included.
@Accessible Android has a list of TTS engines sorted by language at this page: accessibleandroid.com/list-of-…
Except of eSpeak-ng and RHVoice there is another opensource app called SherpaTTS that can use Piper TTS and Coqui based voices at: github.com/woheller69/ttsEngin…
Including eSpeak-ng, RHVoice, SherpaTTS and the list of TTS engines mentioned by accessible android, is there a viable TTS engine or at least one that is close enough to be viable to get included in the foreseable future?

Another approach I have been thinking about is to add / inject the TTS app or any other app I'd like as a part of the install process. It turns out I am not the only one speculating about that idea and it's not practical and feasible either as it's also breaking the security model.
It's been discussed recently at: discuss.grapheneos.org/d/25899…

Another way on how to install an app on an android device would be using adb install from a computer. I am not definatelly sure on this but GrapheneOS does not allow enabling ADB on production builds. In order to instal a TTS app over ADB we'd need to find a way on how to install GrapheneOS with ADB preenabled on first run. This is a huge security hole as well.

There might be a way to build my own flavour of GrapheneOS, but that's too involved, I'd need very powerfull machine for the actual build process and I would again compromise security by either disabling or handling future updates on my own building each new release on my own.

So given the current state I am afraid we screen reader users are out of luck and there is no way to get this thing running on my own with no help from someone else.

The end result is that I'll either get security or I can look elsewhere to get accessibility.

Please am I getting it right or might I have overlooked something that might help me to install GrapheneOS on my own?

Thanks for reading to the end

LunaticStrayDog reshared this.

in reply to Peter Vágner

One of our full time developers is actively working on building our own text-to-speech and speech-to-text integration. It's where all of their effort is currently going. None of the available apps are suitable for inclusion. None are modern enough aside from Sherpa and it has issues including high latency making it unsuitable for use with TalkBack. Our own implementation is going to be significantly better.

ADB works fine on GrapheneOS but you'd have to enable it.

in reply to boredsquirrel

@Rhababerbarbar We're making our own implementation for inclusion in GrapheneOS. It will be similar in design to Sherpa but faster. It will initially just be English. People can still install Sherpa and other TTS implementations if they want them. We just need something available out-of-the-box for blind users to install GrapheneOS and also basic usability. It's fine if people need to install other TTS implementations for other languages, etc. but we can add that too.
in reply to GrapheneOS

I have asked a friend and @GrapheneOS community chat members for the help with initial setup and now I am fully configured with RHVoice as my current TTS of choice.
Except of one GPS navigation app I am used to everything is working fine for me including proprietary stuff for my work like Microsoft Teams, banking apps including Poštová banka, George and Revolut and the other apps I like such as Bitwarden as a password manager, Arcanechat, Conversations, ElementX, FairEmail, Open Key Chain for chatting and emailing, Antennapod, BubbleUPNP, Foobar 2000, Kore, Voice, NewPipe, ytdlnis for podcasts, music, audiobooks and videos, , Catima for lojalty cards and tickets, some other apps. For downloading apps I am mainly using F-droid and Aurora store. I am not signed into the google account but I am using play services for push notifications and other compatibility reasons for apps which need it.

Thanks for everything you are doing, it's fantastic and I like it verry much.

in reply to Peter Vágner

All location-based apps should work, but some may expect network location to be available which it isn't by default. You can enable Network location and Wi-Fi scanning in Settings > Location > Location services if you want network location without needing to use Google Play for location. See grapheneos.org/features#networ….

If you installed apps before sandboxed Google Play and they depend on it, you may need to reinstall the apps depending on it so they detect it properly.

I've just noticed piped.video can still be used for playing videos. It's just that the public instance at piped.video and some other instances require registration.
@Archos and friends, please have you explored ways to eventually host it at @Oscloud ?
I'd host it my-self but I don't have a spare machine at the location with suitable ipv6 range for being somewhat resilient to youtube throttling attempts.

Thanks for considering

Geeky stuff aside, let's enjoy the hottest friday party ever. @Marek Macko is on air again with his awesome show called #playgroundLive. #dance, #trance, a bit of #hardStyle, #eurodance, #90s and other party genres in an incredible live mix performance lasting a few hours spiced up with some random chat messages of fellow listeners and friends.

Fun project for friday evening. I am hosting some unmaintained web sites for a few friends and they help me to pay the hosting costs in return.
Now I have found out I need php 7.4 for some of these and it's not readily available so building php 7.4.33 on up to date arch linux.

I am afraid this will no longer be possible in the future. How do you deal with this? Can I run PHP in some kind of lightweight container?

More and more I am looking through @Delta Chat apps and resources I believe this should become number one messenger of choice for screen reader users.
The developers are constantly improving its #a11y. It's secure from the start of using.
Additionally the desktop chat has under gone an #accessibility audit and accessibility issues are clearly documented in public.
I am not sure other messenger style app on the planet has such dedicated commitment to accessibility ever.

github.com/deltachat/deltachat…

in reply to Paweł Masarczyk

@Paweł Masarczyk While testing I have used gmail or my own classic email account. It worked but I have understood using #chatmail servers is what makes it most attractive. I think I have read a blogpost from someone a few months ago explaining this very well that inspired me, however I can't find that article in my browser history right now. CC @Cleverson @Delta Chat
in reply to Delta Chat

@Paweł Masarczyk @Cleverson I know you are looking for a way to stay in contact with people you are already connected to using traditional email. Still I would recommend creating a #chatmail account on your relay of choice just for testing so you can start with an empty profile and you'll get to experience the @Delta Chat the way it has been meant to. Then as an exercise continue with other more advanced scenarios such as classic email login.
The #chatmail based onboarding is really very simple, there is nothing to do wrong.

Playing with various technologies tonight. Managed to setup a brand new debian 12 systemd-nspawn container. Now looking through various documentation and other internet sources to combine it to a working #chatmail relay server. The thing is I only have a single IP address on the host system so I need to adapt it so the nginx running on the host either redirects or proxies to the container. Plus I guess I need to share the certificate in some way so postfix and dovecot can use it too.

I'd like to create #wireguard tunnel between two natted machines. I'm fine with using an intermediate third machine for IP addresses and ports discovery.
The idea outlined in this article sounds verry appealing to me.
However I'm wondering is there something that is developed further beyond a proof of concept I might look at instead? I know there is #tailscale. That sounds too much for me unfortunatelly as I do need to register with them and similar. Thus I'm exploring if there is something I can self host.

jordanwhited.com/posts/wiregua…

Dear friends, #screenreader users, #blind users and #accessibility professionals in particular.
Are you using chat over traditional email as provided by #DeltaChat / #DeltaLab / #ArcaneChat?
I like it uses traditional email infrastructure, is federated, supports end to end encryption, rich content including audio / voice messages. I have even recognized @adb has implemented screen reader accessibility specific features into the android app.
However I have quickly tried electron based app on linux with orca screen reader, exchanged a few messages back and fort between two of my personal accounts using the DeltaChat app and thunderbird and I haven't found the user experience verry appealing. I can't understand how to effectivelly navigate in the list of conversations and list of messages.
Therefore I would be interested to hear a few comments from people knowing this platform better. How accessible is it on different platforms?
On the desktop is the electron based app a prefered choice?
Also some other questions. For the best experience, do I need a new email address or can I use my existing self-hosted one I have already configured in thunderbird?
Ffeel free to point me to a FAQ, some up to date introductory documentation if you think my questions don't make much sense please.
in reply to WofWca

@WofWca Looking more it looks to me you are working on some amazing accessibility related improvements. While I was testing a few weeks ago I had issues navigating the lists. hmm, perhaps I need to figure out how to build with this PR and see for my-self if it might be related. github.com/deltachat/deltachat…

Edit: oh, there are more PRs resolving keyboard navigation and accessibility related issues. It's really right about perfect time I have discovered this thing.

Thanks and keep up the great work please

in reply to Peter Vágner

@WofWca So I have finally updated to the latest and greatest #deltachat desktop. And I must say I am really impressed.
Both lists, list of conversations and list of messages are now keyboard navigable.
The hint for the screen reader users is that not all of the screen readers may switch into focus mode automatically so once you do this manually you can enjoy the user experience the way it has been designed.
Also there are nice keyboard shortcuts e.g. ctrl+n for jumping into the input box, ctrl+k for search and more.
Now I do really need to explore more i.e. what about the encryption. Can I use my existing GPG key or add a subkey or is that handled transparently on the background? I need to figure out if I should create new email for chatting or use my primary email.
Still from the accessibility point of view we are really getting something usefull I think.
in reply to Peter Vágner

Note to my-self and other people like me:
Read the #deltachat help at delta.chat/en/help it's all perfectly explained there in easy to follow language. It's even translated to various different languages already.
The number one answer for me is that there are #deltachat specific so called #chatmail servers suitable for anonymous instant messaging over email.
So eventhough I like to self host my emails, I think I'll go with existing #chatmail server at least initially.

reshared this

in reply to devSJR

@devSJR If you have mostly or only "green-checkmarked" chats you can reconfigure your e-mail address to another server and then send a message to your contact which should update them. This will not migrate unencrypted or opportunistically encrypted chats, though, which is why we don't prominently recommend it. Another way is to create a new profile, and then add it to all relevant chats and remove the old profile. More cumbersome but also pretty failsafe.
in reply to Matthias

It is possible for #chatmail users to communicate with classic email users who have published their public key.

You just have to do manual chatmail registration, save your login details and private key securely, and use it with something that supports #pgp like #Thunderbird or #Mailvelope.

This entry was edited (9 months ago)
in reply to WofWca

@WofWca Perhaps making sure parent of those list items has an attribute role="list" or if you are sure everything inside a particular DOM subtree has logical keyboard focus handling perhaps role="application" on encompassing parent element would do the ttrick. However it really has to be considered visely as it makes it more difficult for screen reader users do read parts of the UI that should behave like a document content by lines, by words and similar.

If you have ever wondered how S.M.A.R.T. output of a failing hard drive looks like
# smartctl -l selftest /dev/sdb
smartctl 7.4 2023-08-01 r5530 [x86_64-linux-6.12.4-arch1-1] (local build)
Copyright (C) 2002-23, Bruce Allen, Christian Franke, www.smartmontools.org

=== START OF READ SMART DATA SECTION ===
SMART Self-test log structure revision number 1
Num  Test_Description    Status                  Remaining  LifeTime(hours)  LBA_of_first_error
# 1  Extended offline    Completed: read failure       90%     29550         3432497913
# 2  Conveyance offline  Completed: read failure       70%     29550         3432490825

Nice article in czech describing latest advancemend in developing czech and slovak voices for #RHVoice portal-pelion.cz/rodina-zdarma…

Hello #accessibility experts and #screenReader users. I am trying to find out a wordpress addon that might provide self-hostable and screen reader accessible reservation system. The features I am looking for is ability to integrate this with google calendar or any other caldav based self-hostable calendaring solution. Email notifications for both visitors and administrators is a plus. I need this to be accessible to both visitors as well as administrators.
Also recently I think I have seen a mention of croutfunding campaign where the aim is to build open-source accessible booking system. I'd be happy to look at it too if you can remind me of it.

Pred pár dňami sme vďaka @Zvonimir Stanecic a ďalším dobrovoľníkom zverejnili prvý slovenský ženský hlas pre #rhvoice. Hlas dostal aj pekné netradičné meno Jasietka. K dispozícii sú hlasy pre #Windows #nvdasr #android aj #linux . Aktualizácie sa zároveň dočkala aj celková podpora pre slovenčinu, vrátane už skôr zverejneného slovenského hlasu Ondro. Ak potrebujete #tts #textToSpeech k čítaču obrazovky, na čítanie kníh, inštrukcie pre GPS navigáciu, pozrite si prosím podrobnosti na jednoduchom webe.

hlas.ondrosik.sk/

reshared this

Quite number of things have happened since I have mentioned my occassional usage of #emacS:
* @Termux is trying to update its compatibility with recent #android versions returning into the Google play in the process.
* speechd-el the app that acts as a built-in screen reader for emacs got some polishing and fixes recently.
* Touch and speech have recently released an update to Corvus, suite of assistive apps for #android that also feature great liblouis powered braille keyboard that can emulate alt, ctrl and shift keys allowing me to use emacs keybindings on the braille touch keyboard.
* I'm on holiday these days and I can benefit from having a screenreader accessible linux terminal with powerfull tools inside my pocket.

So here is a short article I wrote on the matter last week

pvagner.sk/2024/emacs-a11y-tip…

in reply to Oskar

@Oskar @Projekt Define @Paweł Masarczyk Can you please try to describe how it looks like? Is it built into the smartphone case? Is there a place to order it from? What's the price for an individual user? Does it use liblouis for the back translation from braille? I don't seem to be able to find out these details on the website or should I be looking at something else?
in reply to Peter Vágner

@define @Piciok Metabraille is built like a house with a roof. 8 buttons in the front and two buttons on the roof instead of windows. You hold the house with both hands between your palms and type with the fingers on the front and with your thumbs on the roof. It is like holding a keyboard upside down in the air with your palms and typing on the backside. 12 cm long, 10 cm width, 2.6 cm depth. it is a standalone device, not built in the smartphone case.

Dear !Friendica Support I'm sometimes reading my #friendica timeline from the @Tusky app on #android.
I am having an issue that some parts of the timeline are not loaded into the @Tusky app at all.
I have figured out when replies are added into a thread, the post is being repositioned in the list of posts returned by the #Mastodon compatible API. @Tusky is not fetching older posts when the one that has already been requested is found while syncing new posts. That is causing gaps in my timeline I can read on the phone.
It really sounds complicated. Do you understand what's my issue?

Winter blue tardis reshared this.

A friend just reminded me today that #braille watches for the #blind inspired by the soviet braille watch tradition are still being produced nowadays. I have found an eshop with world wide shipping meranom.com/en/vostok/vostok-t…
Also there is a nice article on the history of these watches: mroatman.wixsite.com/watches-o…

reshared this

Today I have found out when enabling subtitles on #youtube the area where the subtitle track is playing is no longer configured as so called live region so youtube subtitles are no longer read automatically with a #screenReader if it's running. I've also looked at #invidious which has aria-live set to off. This is unfortunate as I was expecting this to just work. It can be tested on this video if you like youtu.be/watch?v=3i7gkN-1sAI yewtu.be/watch?v=3i7gkN-1sAI

Dnes si dovolím odkázať na veľmi vydarený český hlas pre hlasový výstup #TTS #RHVoice. Sám autor o tom píše tu: groups.io/g/Blind-android/mess… . O použití pre #Windows #Linux a #Android sa dočítate aj na jednoduchej komunitnej stránke hlas.ondrosik.sk/ . Gro používateľov sú ťažko zrakovo postihnutí používatelia, ktorí si bez kvalitného hlasového výstupu a čítača obrazovky nedokážu svoj digitálny život predstaviť, možno ale aj vám by sa mohol hodiť takýto hlas pre váš počítač či smartfón nezávislí od obrovských korporácií. Mohli by ste ho napr. použiť na čítanie kníh alebo počas navigácii cez GPS.