Peter Vágner reshared this.

TIL: There's a W3C candidate recommendation draft for a CSS markup to transfer different properties of text and controls on the web via audio cues and changes to the TTS volume, speech rate, tone, prosody and pronunciation, kind of like the attributed strings in iOS apps and it's called CSS Speech. w3.org/TR/css-speech-1/ #Accessibility #A11y #Blind

reshared this

in reply to Paweł Masarczyk

There are people who seem to feel really strongly about this being a good thing for screen reader users, and I must admit to being bewildered about why. Websites changing aspects of screen reader output may be equitable, if we compare it with the way webpages can alter visual presentation through fonts and other aspects. But to me it feels entirely inappropriate to cross that boundary between the browser as the user agent and accessibility software in order to interfere with very personal settings.

Meanwhile on iOS, the related accessibility attributes are being used to achieve outcomes nobody wants or needs, like spaces between all the digits of a credit card number. @miki @prism

in reply to James Scholes

I can see the point for e.g. text-to-speech APIs built into the browser, maybe even read-aloud features. But the case for screen reader compatibility seems to be built on the foundational assertion that SR output is monotonous and can't be "livened up" by brands.

As assertions go, I think that is both true and exactly how it should be. I don't use a screen reader for entertainment. I can think of few things more obnoxious than a marketing person thinking that my screen reader should "shout this bit."

Many web authors can't even label stuff correctly. Why on earth would we expect them to treat this sort of feature with informed respect? @miki @prism

in reply to Drew Mochak

@prism I think without ARIA or an equivalent (like more things built into the web platform), the web would've continued galloping forward with all the same UI widgets and design patterns but with no way to make them even halfway accessible, and we'd be left even more behind than we are now.

By contrast, I don't think the inability for a website to change the pitch of NVDA is a legitimate blocker to anything worthwhile. @Piciok @miki

in reply to James Scholes

@jscholes I have felt for a while that only having TTS for everything is pretty limitting. So, you know, I use unspoken. Problem solved. I haven't really thought to myself, self, it would be great if the website author could script some nonverbal feedback for me instead of what I am currently hearing, or anything like that. So this may well be a solution in search of a problem.
@Piciok @miki
in reply to Drew Mochak

@prism @jscholes @miki I don't see the point because everyone has different ways they like to hear things. People choose the verbosity and speech options that work for them and to have something override that would be irritating. I also feel that this is part of a larger conversation about the perceived need for sighted people to feel like our experience of the web is vastly different. This is why we have a lot of unnecessary context already and here is another example.
in reply to miki

@miki I think it's a trap to suggest that such problems should currently be solved only through speech properties and auditory cues within individual apps. Expressive semantics on the web have only been explored at a surface level so far, and it's a complete stretch to go from "We don't have the ARIA properties to convey complex information," to "Let's have every application implement its own beeps and boops."

Imagine having to learn the sound scheme for Gmail, then Outlook, then Thunderbird. Then going over to Slack where they also have unread state albeit for chat messages rather than emails, but they use an entirely different approach again.

All the while, braille users are getting nothing, and people who struggle to process sounds alongside speech are becoming more and more frustrated. Even if we assume that this is being worked on in conjunction with improvements to ARIA and the like, how many teams have the bandwidth and willingness to implement more than one affordance?

We've already seen this in practice: ARIA has braille properties, but how many web apps use them? Practically none, because getting speech half right and giving braille users an even more subpar experience is easier. Your own example highlights how few apps currently let you control things like verbosity and ordering of information.

CSS Speech could turn out even worse. A product team might opt to implement it instead of semantics because the two blind people they spoke to said it would work for them, and never mind the other few million for whom it doesn't. They'll be the people complaining that there's no alternative to the accessibility feature a team spent a month on and thought was the bee's knees.

@silverleaf57 @prism @Piciok

in reply to James Scholes

@jscholes @silverleaf57 @prism Efficiency, not equity.

Words are a precious resource, far more precious than even screen real estate. After all, you can only get a fairly limited amount of them through a speaker in a second. We should conserve this resource as much as we can. That means as many other "side channels" as we can get, sounds, pitch changes, audio effects, stereo panning (when available) and much more.

Icon fatigue is real. "me English bad, me no know what delete is to mean" is also real, and icons, pictograms and other kinds of pictures is how you solve that problem in sighted land.

Obviously removing all labels and replacing it with pictograms is a bad idea. Removing all icons and replacing them with text... is how you get glorified DOS UIs with mouse support, and nobody uses these.

in reply to miki

@jscholes @silverleaf57 @prism Everything said above also applies to braille, Braille cells are even more precious than words in a speaker. It's a schame we can abbreviate "main landmark heading level 2" to something more sensible, but we can't abbreviate "unread pinned has attachment overdue" if those labels are not "blessed" by some OS accessibility API.
in reply to James Scholes

@miki Note that I'm specifically responding to your proposed use case here. You want beeps and boops, and I think you should have them. But:

1. I think you should have them in a centralised place that you control, made possible via relevant semantics.

2. I don't think the fact that some people like beeps and boops is a good reason to prioritise incorporating beeps and boops into the web stack in a way that can't be represented via any other modality.

@silverleaf57 @prism @Piciok

This entry was edited (1 month ago)
in reply to James Scholes

@jscholes @silverleaf57 @prism Centralized beeps and boops don't make much sense to me. Each app needs a different set, let's just consider important items on a list. That can mean "overdue", "signature required", "has unresolved complaints", "student not present", "compliance certification not granted" or something entirely different. We can't expect screen readers to have styles for all of these, just as we can't expect browsers to ship icons for all of these.
in reply to miki

@miki Sure. Or it can just mean "important" in a domain-specific way that's shared across apps in that domain. We should be taking advantage of that to make information presentation and processing more streamlined, before inventing an entirely new layer and interaction paradigm that hasn't been user tested and will require text alternatives anyway. @silverleaf57 @prism @Piciok
in reply to James Scholes

@miki As noted, I think people who can process a more efficient stream of information should have it available to them. That could be through a combination of normalised/centralised semantics, support for specialised custom cases, and multi-modal output.

My main concern remains CSS Speech being positioned as the only solution to information processing bottlenecks, which I think is a particularly narrow view and will make things less accessible for many users rather than more.

Good discussion, thanks for chatting through it. @silverleaf57 @prism @Piciok

in reply to James Scholes

@jscholes At the same time, I think the chances that CSSSpeech completely takes over the industry and we all stop doing text role assignments is quite low.
explainxkcd.com/wiki/index.php…

So I am decidedly meh about this. It could help but probably won't.
@miki @silverleaf57 @Piciok

in reply to James Scholes

@jscholes @prism @miki @silverleaf57 I found the concept intriguing and am myself in two minds about it. On one hand, I wouldn't mind having the speech experience augmented by things that aren't words. I could imagine browsing a product's details page and reading upon all of it's features with tiny earcons indicating whether certain feature is supported or not rather than hearing "Yes" and "No" every time. This could even be played at the same time as the readout begins. To be fair, I also don't mind having the pronunciation of tricky words that are important for proper understanding and functioning in a domain, predefined just so I could learn it. Character and number processing might come in handy too - recently there was an issue on the NVDA Github opened against a feature to read combinations of capital letters and digits as separate entities for the benefit of ham radio operators and their call signs. Some kinds of numbers I also find easier to remember when they come digit-by-digit etc. The ability to define the spatial location of voice on the stereo sound spectrum could be useful for presenting those spatial relationships in some advanced web apps (thinking scientific contexts, design, web text and code editors etc.. As you say, however, I wouldn't expect this being widely adopted by web devs who already struggle with the proper use of ARIA. Also the trade-offs could be significant, especially if this becomes the sole way of conveying information. Blind users with a profound hearing impairment who will miss out on crucial information because it was read out too quietly, too fast and with a pitch that takes away some of the frequencies they can't discern any more; neurodivergent people confused by sudden changes and unfamiliar sounds on top of exotic keyboard shortcut choices they already have to remember etc. This could create a situation similar to WCAG SC 1.4.1 where the colour is used as the only way of conveying information.
in reply to Paweł Masarczyk

This already exists though, as a screenreader feature. Kind of. NVDA has an add-on called unspoken that will replace the announcement of roles with various sounds, there's a different one for checked vs. unchecked boxes for instance. JAWS did (does?) something similar with the shareable schemes in the speech and sounds manager. Granted, not a lot of people do this, but the ability is there if people want it. VO, TB and cvox also have earcons--they're not used for this purpose, but they could be. Having this under the user's control rather than the author's control does seem better. It prevents for instance a developer deciding to be super abtrusive with ads. I do see the potential for it to be good, the author would be able to convey more nuanced concepts being the author of the content... it just feels like a thing most people wouldn't use, and most of the people who'd try would end up being obnoxious about it.

@jscholes @miki @silverleaf57

in reply to Drew Mochak

@prism @jscholes @miki @silverleaf57 Yes, this is what I'm thinking too. Also, the addons are great - I experiment with Earcons and Speech Rules which is another addon with tons of customization. Bringing it on as a core feature would signal it as industry standard though and from that it would be possible to explore whether any external API's could augment it in any way.
in reply to James Scholes

@jscholes @prism @miki @silverleaf57 As for this being widely adopted, I expect some CSS properties could be mapped to the aural cues on a browser lever just like some HTML elements carry implicit ARIA properties with them by default. This would have to be carefully considered. Regarding sound cues: this would have to be based on some kind of familiarity principle where the sounds are those most users will already know or they resemble the action they are supposed to represent, think emptying the recycle bin on Windows. I really like the approach of JAWS representing heading levels through piano notes in C major - it sounds logical but on the other hand not everyone is able to recognize musical notes at random. I'm not convinced about the marketing value of this - I mean creating brand voices etc. It sounds fun but no more than that, at least in the screen reader context. I guess inclusion in advertising is another can of worms that might derail the discussion. I'm looking forward to when NVDA finally incorporates some kind of sound scheme system because we will then be able to talk about some kind of standard given that JAWS and to some extent VoiceOver and Talkback make use of that already. I guess then the discussion could evolve around this being complementary to something like aria-roledescription or aria-brailleroledescription, assigning familiar sounds and speech patterns to custom-built controls.
in reply to James Scholes

@jscholes @prism @miki @silverleaf57 I think inviting @tink and @pixelate into the discussion is a great idea as they might have valuable insights on this. On a related note: something that's been running around my head is how many Emojis could be faithfully represented by sounds.
in reply to Paweł Masarczyk

@jscholes @prism @miki @silverleaf57 @tink So, I generally like beeps and boops. All shiny and stuff. But the web is made by sighted people, and they will get things wrong. I'd rather we have our own tools, like NVDA'S earcons addon, and maybe have earcon packs for it to, for example, add aural highlighting for VS Code, or make-gmail-shiny, stuff like that.
Peter Vágner reshared this.

Researchers pointed a satellite dish at the sky for 3 years and monitored what unencrypted data it picked up. The results were shocking: They obtained thousands of T-Mobile users' phone calls and texts, military and law enforcement secrets, much more: 🧵👇wired.com/story/satellites-are…
This entry was edited (1 month ago)

reshared this

Peter Vágner reshared this.

During last 3 months I am using VDO ninja for all my remote interwiev and podcast recordings. here is my article about it from the blind perspective, focused on accessibility and audio.

Have You Ever Wanted to Record an Interview or Podcast Online? You’ve probably faced a few challenges:
How to transmit audio in the highest possible quality?
How to connect in a way that doesn’t burden your guest with installing software?
And how to record everything, ideally into separate tracks?

The solution to these problems is offered by the open-source tool VDO Ninja.

What Is VDO Ninja


It’s an open-source web application that uses WebRTC technology. It allows you to create a P2P connection between participants in an audio or video call and gives you control over various transmission parameters.
You can decide whether the room will include video, what and when will be recorded, and much more.

In terms of accessibility, the interface is fairly easy to get used to — and all parameters can be adjusted directly in the URL address when joining.
All you need is a web browser, either on a computer or smartphone.

Getting Started


The basic principle is similar to using MS Teams, Google Meet, and similar services.
All participants join the same room via a link.
However, VDO Ninja distinguishes between two main types of participants: Guests and the Director.
While the guest has limited control, the director can, for example, change the guest’s input audio device (the change still must be confirmed by the guest).

A Few Words About Browsers


VDO Ninja works in most browsers, but I’ve found Google Chrome to be the most reliable.
Firefox, for some reason, doesn’t display all available audio devices, and when recording multiple tracks, it refuses to download several files simultaneously.

Let’s Record a Podcast


Let’s imagine we’re going to record our podcast, for example, Blindrevue.
We can connect using a link like this:

https://vdo.ninja/?director=Blindrevue&novideo=1&proaudio=1&label=Ondro&autostart=1&videomute=1&showdirector=1&autorecord&sm=0&beep

Looking at the URL more closely, we can see that it contains some useful instructions:
  • director – Defines that we are the director of the room, giving us more control. The value after the equals sign is the room name.
  • novideo – Prevents video from being transmitted from participants. This parameter is optional but useful when recording podcasts to save bandwidth.
  • proaudio – Disables effects like noise reduction, echo cancellation, automatic gain control, compression, etc., and enables stereo transmission.
    Be aware that with this setting, you should use headphones, as echo cancellation is disabled, and otherwise, participants will hear themselves.
  • label=Ondro – Automatically assigns me the nickname “Ondro.”
  • autostart – Starts streaming immediately after joining, skipping the initial setup dialog.
  • videomute – Automatically disables the webcam.
  • showdirector – Displays our own input control panel (useful if we want to record ourselves).
  • autorecord – Automatically starts recording for each participant as they join.
  • sm=0 – Ensures that we automatically hear every new participant without manually unmuting them.
  • beep – Plays a sound and sends system notification when new participants join (requires notification permissions).

For guests, we can send a link like this:

https://vdo.ninja/?room=Blindrevue&novideo=1&proaudio=1&label&autostart=1&videomute=1&webcam

Notice the differences:
  • We replaced director with room. The value must remain the same, otherwise the guest will end up in a different room.
  • We left label empty — this makes VDO Ninja ask the guest for a nickname upon joining.
    Alternatively, you can send personalized links, e.g., label=Peter or label=Marek.
  • The webcam parameter tells VDO Ninja to immediately stream audio from the guest’s microphone; otherwise, they’d need to click “Start streaming” or “Share screen.”


How to Join


Simply open the link in a browser.
In our case, the director automatically streams audio to everyone else.
Participants also join by opening their link in a browser.
If a nickname was predefined, they’ll only be asked for permission to access their microphone and camera.
Otherwise, they’ll also be prompted to enter their name.

Usually, the browser will display a permission warning.
Press F6 to focus on it, then Tab through available options and allow access.

Controls


The page contains several useful buttons:

  • Text chat – Toggles the text chat panel, also allows sending files.
  • Mute speaker output – Mutes local playback (others can still hear you).
  • Mute microphone – Mutes your mic.
  • Mute camera – Turns off your camera (enabled by default in our example).
  • Share screen / Share website – Allows screen or site sharing.
  • Room settings menu (director only) – Shows room configuration options.
  • Settings menu – Lets you configure input/output devices.
  • Stop publishing audio and video (director only) – Stops sending audio/video but still receives others.


Adjusting Input and Output Devices


To change your audio devices:

  1. Activate Settings menu.
  2. Press C to jump to the camera list — skip this for audio-only.
  3. Open Audio sources to pick a microphone.
  4. In Audio output destination, select your playback device. Press test button to test it.
  5. Close settings when done.


Director Options


Each guest appears as a separate landmark on the page.
You can navigate between them quickly (e.g., using D with NVDA).

Useful controls include:

  • Volume slider – Adjusts how loud each participant sounds (locally only).
  • Mute – Silences a guest for everyone.
  • Hangup – Disconnects a participant.
  • Audio settings – Adjusts their audio input/output remotely.


Adjusting Guest Audio


Under Audio settings, you can:

  • Enable/disable filters (noise gate, compressor, auto-gain, etc.).
  • View and change the guest’s input device — if you change it, a Request button appears, prompting the guest to confirm the change.
  • Change the output device, useful for switching between speaker and earpiece on mobile devices.


Recording


Our URL parameters define automatic recording for all participants.
Recordings are saved in your Downloads folder, and progress can be checked with Ctrl+J.

Each participant’s recording is a separate file.
For editing, import them into separate tracks in your DAW and synchronize them manually.
VDO Ninja doesn’t support single-track recording, but you can use Reaper or APP2Clap with a virtual audio device.

To simplify synchronization:

  1. Join as director, but remove autorecord.
  2. Wait for everyone to join and check audio.
  3. When ready, press Alt+D to edit the address bar.
  4. Add &autorecord, reload the page, and confirm rejoining.
  5. Recording now starts simultaneously for everyone.
  6. Verify this in your downloads.


Manual Recording


To start recording manually:

  1. Open Room settings menu.
  2. Go to the Room settings heading.
  3. Click Local record – start all.
  4. Check PCM recording (saves WAV uncompressed).
  5. Check Audio only (records sound without video).
  6. Click Start recording.


Important Recording Notes


  • Always verify that all guest streams are recording.
  • To end recordings safely, click Hangup for each guest or let them leave.
  • You can also toggle recording for each guest under More options → Record.
  • Files are saved as WEBM containers. If your editor doesn’t support it, you can convert them using the official converter.
  • Reaper can open WEBM files but may have editing issues — I prefer importing the OPUS audio file instead.


Recommended Reading


In this article, I’ve covered only a few features and URL parameters.
For more details, check the VDO Ninja Documentation.

Peter Vágner reshared this.

Announcing AudioCapture. A win32 application to capture audio from a process and save it to an audio file. Full disclosure: This was written with Claude Code. Why? Because I'm not an experienced c++ programmer, however I saw an idea for an app and no one else was going to write it, so I did it myself this way. The full code is available, so if you wish to contribute, feel free. github.com/masonasons/AudioCap…
This entry was edited (1 month ago)

reshared this

in reply to Andre Louis

Revisiting this thread. For both of you on Windows 10, could you try the latest release of AudioCapture and see if process capture works for you? I've discovered a bug with windows 10 where the process loopback capture doesn't release properly, and unfortunately this means that the process can only be captured one time. If you uncapture the process, you either have to restart the process, or AudioCapture itself. github.com/masonasons/AudioCap…
This entry was edited (1 month ago)
in reply to Andre Louis

Is the executable signed? I have a command I run on the executable before I release it so defender apparently doesn’t flag it. Don’t know if it works though. Install windows SDK, just in case someone needs it, here is the command. powershell:
& 'C:\Program Files (x86)\Windows Kits\10\bin\10.0.26100.0\x64\signtool.exe' sign /fd SHA256 /tr timestamp.digicert.com /td SHA256 /a '.\iptvclient.exe'
If your in CMD:
"C:\Program Files (x86)\Windows Kits\10\bin\10.0.26100.0\x64\signtool.exe" sign /fd SHA256 /tr timestamp.digicert.com /td SHA256 /a ".\iptvclient.exe"
This entry was edited (1 month ago)
Peter Vágner reshared this.

Do you know that you can use Subtitle edit to transcribe audio? It has a relatively accessible guy so you can use Purfwiev's faster whisper xxl, cpp, cpp cublas, const-me. Longer post how to use it follows:

Installing Subtitle Edit


Download the program from the developer’s website. Navigate to the level 2 heading labeled “Files.”
If you want to install Subtitle Edit normally, download the first file, labeled setup.zip.
There is also a portable version available, labeled SE_version_number.zip.

If you decide to use the portable version, extract it and move on to the next section of this article. The installation itself is standard and straightforward.

A Note on Accessibility


NVDA cannot automatically obtain focus in lists.
To find out which item in the list is currently selected, move down with the arrow key to change the item, then press NVDA+TAB to hear which one is focused.

Initial Setup


  • In the menu bar, go to Video and activate Audio to text (Whisper).
  • When using this feature for the first time, the program may ask whether you want to download FFMPEG. This library allows Subtitle Edit to open many audio and video files, so confirm the download by pressing Yes.
  • Subtitle Edit will confirm that FFMPEG has been downloaded and then ask whether you want to download Purfwiev’s Faster Whisper – XXL. This is the interface for the Whisper model that we’ll use for transcription, so again confirm by pressing Yes.
  • The download will take a little while.
  • Once it’s complete, you’ll see the settings window. Press Tab until you reach the Languages and models section. In the list, select the language of your recording.
  • Press Tab to move to the Select model option, and then again to an unlabeled button.
  • After activating it, choose which model you want to use. Several models are available:
    • Small models require less processing power but are less accurate.
    • Large models take longer to transcribe, need more performance and disk space, but are more accurate.
      I recommend choosing Large-V3 at this step.


  • Wait again for the model to finish downloading.


Transcribing Your First Recording


  • Navigate to the Add button and press Space to activate it.
  • A standard file selection dialog will open. Change the file type to Audio files, find your audio file on the disk, and confirm.
  • Activate the Generate button.
  • Now, simply wait. The Subtitle Edit window doesn’t provide much feedback, but you can tell it’s working by the slower performance of your computer—or, if you’re on a laptop, by the increased fan noise.
  • When the transcription is done, Subtitle Edit will display a new window with an OK button.


We Got Subtitles, So One More Step


In the folder containing your original file, you’ll now find a new file with the .srt extension.
This is a subtitle file—it contains both the text and the timing information. Since we usually don’t need timestamps for transcription, we’ll remove them in Subtitle Edit as follows:

  • Press Ctrl+O (or go to File → Open) to bring up the standard open file dialog. Select the .srt file you just got.
  • In the menu bar, open File → Export → Plain text.
  • Choose Merge all lines, and leave Show line numbers and Show timecode unchecked.
  • Press Save as and save the file normally.

If you’re transcribing multiple recordings, it’s a good idea to close the current subtitle file by starting a new project using Ctrl+N or by choosing File → New.

Conclusion


Downloaded models can, of course, be reused, so future transcriptions will go faster.
In this example, I used Purfwiev’s Faster Whisper. If you want to use a different model, you can select it from the model list, and Subtitle Edit will automatically ask whether you’d like to download it.

Peter Vágner reshared this.

Peter Vágner reshared this.

I decided to write a post where I talk about my experiences finding work as a blind person and attempt to give some general advice to blind people who are either looking for work or looking for a position that better aligns with their goals or values. I'm not sure why the strange URL; hopefully it doesn't cause problems. mikegorse.substack.com/p/4834h…

Peter Vágner reshared this.

Peter Vágner reshared this.

On my AMD Ryzen 7 8845HS mini PC, NVDA is a bit sluggish in some cases in Firefox; e.g. cursoring through messages in Gmail folders. For reasons I don't fully understand, setting the processor affinity to a single CPU core and setting the process priority to "above normal" helps significantly, even when the CPU is nearly idle. I don't currently have the time/energy to debug the root cause for this or write a proper add-on, but I wrote an NVDA global plugin to make the change for me automatically when NVDA starts. If it breaks something, you get to keep all the pieces.
```
import ctypes

import globalPluginHandler

class GlobalPlugin(globalPluginHandler.GlobalPlugin):
def __init__(self, *args, **kwargs):
super().__init__(*args, **kwargs)
p = ctypes.c_void_p(ctypes.windll.kernel32.GetCurrentProcess())
ctypes.windll.kernel32.SetProcessAffinityMask(p, ctypes.c_void_p(1))
ctypes.windll.kernel32.SetPriorityClass(p, ctypes.wintypes.DWORD(0x00008000))
```
#nvdasr

This entry was edited (1 month ago)

reshared this

Peter Vágner reshared this.

Sending and receiving money has just become a whole lot easier!

Instant payments are now available to everyone in the eurozone.

⚡ Instant transfers 24/7, no waiting days for your money
💰 No extra fees, same price as regular payments
🔍 Free payee verification, ensuring IBAN and name match before sending
🛡️ Safer payments with daily checks to help prevent fraud and sanctions risks
🏦 More access, not just for banks but also fintechs and e-money institutions

For faster, safer payments than ever!

reshared this

Peter Vágner reshared this.

I'm using Paperback more and more these days. Unlike QRead, which I still need for Bookshare DAISY files, Paperback removes all the unnecessary blank lines between paragraphs and pages, and you can't imagine how nice that feels! It also loads books incredibly fast, no matter how large your EPUB or PDF files are. Huge kudos to the developer, and here's hoping DAISY support gets added soon so I can fully switch over.
github.com/trypsynth/paperback
@TheQuinbox
@Quin

Peter Vágner reshared this.

Peter Vágner reshared this.

To jsem na sebe zase upletl bič. Přihlásil jsem si přednášku na #LinuxDays2025. Je už zítra. Doufám, že dcerka dnes bude spinkací, jinak budou slajdy zejtra dost děravé.
Jinak kdybyste mě chtěli slyšet koktat něco o tom co selhostinguju a jak to jde, můžete se stavit ve 13h do posluchárny 105 na FIT ČVUT v Dejvicích. 😬

Peter Vágner reshared this.

Peter Vágner reshared this.

If Google is killing sideloading, then Android is just iOS with ads and spyware. Why the hell would anyone choose that?
fireborn.mataroa.blog/blog/why…
#Android #Google #Sideloading #FOSS #Privacy #accessibility

reshared this

Peter Vágner reshared this.

For the past 3 years, I’ve been part of the Google Accessibility Trusted Tester program, where people with disabilities test unreleased products and give direct feedback to engineers. In my article, I share how the program works, how I’ve maximized participation, the experience I’ve gained, and why more companies should follow Google’s lead in compensating disabled testers.
linkedin.com/pulse/my-experien…

reshared this

Peter Vágner reshared this.

I just pushed a new edit.video release. 🔥

- Faster video processing on modern browsers with Mediabunny
- New output quality controls
- Improvements to video trimming controls
- Optionally load demo video
- Drag n' drop to edit video files
- Sticky fullscreen mode

100% Free

Peter Vágner reshared this.

Peter Vágner reshared this.

This is a concern for anyone using F-Droid, including users of Privacy Browaer Android.

F-Droid and Google's Developer Registration Decree | F-Droid - Free and Open Source Android App Repository – f-droid.org/en/2025/09/29/goog…

Peter Vágner reshared this.

in reply to Soren Stoutner

Здраствуйте,мне кажется они всегда хотели больше контроля,и хорошо что есть #opensource приложение,я не могу пожертвовать по личным причинам, но я занимаюсь продвижением #deltachat и возможно ещё буду #mastodon
трекеров всё больше и больше,в одной простой игре может быть 30+ трекера...😔
Peter Vágner reshared this.

#introduction:

Welcome to #StroongeCast, a husband and wife team consisting of Andre and Kirsten Louis who live in London. On this podcast, we explore anything that makes us question the world—from relationships and parenting to school memories and beyond. Join our family chats for lively discussions, fun stories, and plenty of curious moments.

Going forward we will post all new episode links to this account before any others.

Subscribe: onj.me/stroongecast

Feel free to follow our main accounts here on the fediverse as well.
Andre: @FreakyFwoof
Kirsten: @MoonCat

This entry was edited (1 month ago)

Peter Vágner reshared this.

in reply to StroongeCast

Did you know: There are many completely *free* sample libraries these days, some could help you end up in writing for your next film, documentary, podcast or TV show?

Andre wrote this piece of intro music for the podcast using purely *free* libraries and nothing else.

#InspiredBySound - Musical Breakdown - #StroongeCast Podcast Music: youtu.be/Ra7prQ6d9Pw

This entry was edited (1 month ago)

reshared this

in reply to StroongeCast

We record #StroongeCast episodes on Thursdays consistently now, meaning that youtube members can listen early if they so wish.
If we can find a way to do that for everyone else, we will. Until then however, if you want your fix a day early, please become a youtube member at any level.
youtube.com/@TheOnjLouis/join

StroongeCast reshared this.

1 / 2: Did you know @GNOME Files aka #nautilus has a nifty feature where it can batch rename files? Advanced features include adding sequential numbering, using placeholders and doing search and replace on the names of selected files. #ScreenReader #a11y is preserved.
In order to use it just select multiple files and find Rename item in the shift+F10 popup menu or simply press F2. Also... Don't be shy to press the add button in the batch rename dialog.
Peter Vágner reshared this.

Gauging user interest! Over the last few weeks I have been working on a website that can auto describe YouTube videos. I am aware that the solution has already been created for windows, however, I don’t want to leave mobile users and non-Windows users out in the cold. I have already experimented with a desktop app and it works quite well. Would this be of interest to the community?
This entry was edited (1 month ago)

reshared this

Peter Vágner reshared this.

libxml2's sole maintainer Nick Wellnhofer steps down, meaning libxml2 is now no longer maintained.

discourse.gnome.org/t/stepping…

It's hard to estimate just how many companies depend on this software and critical security updates to the library, so I'm certain many will quickly step up and offer sponsorship to ensure a fundamental dependency doesn't just deteriorate without proper support.

Any day now.

reshared this

Peter Vágner reshared this.

It turns out that you can already run GUI Linux programs in the new Linux terminal app on Android 16, before Google releases the official GUI support. First I switched the audio system to pipewire in the VM by installing the pipewire-audio package, then I installed xrdp (an RDP server for X11), and pipewire-modules-xrdp, for audio support. Then I installed mate-desktop-environment and orca, enabled accessibility in Mate with "gsettings set org.mate.interface accessibility true", and enabled Orca to start automatically with "gsettings set org.gnome.desktop.a11y.applications screen-reader-enabled true". Then I set the password for the default "droid" user with "sudo passwd droid", and created ~/.xinitrc with "#!/bin/sh" and "mate-session", and made it executable with "chmod +x ~/.xinitrc"" After doing all of this, I pressed the third unlabeled button in the Terminal app to open its menu, went to "Port control" and enabled port 3389. Then I installed Windows App from the Play Store and I added a PC with hostname 127.0.0.1, and added a user with the name "droid" and the password I set. When I connected to it, Orca started speaking, and after turning TalkBack off by holding the volume keys, I could control the Linux system with my Bluetooth keyboard, including using the Control and Alt keys, and after putting Orca in laptop mode (by running "orca -s" to open the preferences dialog), I could perform Orca commands with the caps lock key, although sometimes it types a letter instead and it toggles Android's caps lock state (which is separate from Linux's), but pressing caps lock once toggles it off again.

Peter Vágner reshared this.

Peter Vágner reshared this.

The XMPP Interop Testing project helps ensure XMPP servers and clients play nicely together by providing specification test automation.

New update:
✔️ Option to fail runs if some tests were "impossible" to execute.
✔️ Flexible account provisioning

Details on the blog: xmpp-interop-testing.github.io…

The development journey that @fishbowler and I have been taking was made possible by a grant from @nlnet 🙏. The grant has now concluded, and we’re deeply thankful for their support!

#XMPP #interop #testing

reshared this

Peter Vágner reshared this.

We had to start charging the electric car from the grid this week due to bad weather and shorter days, but between this week and mid-April, when we got the car, we charged it exclusively from solar power. Six months and six thousand kilometers. Not bad.

#ElectricVehicle #electromobility #renewables #photovoltaic #renewableenergy

Peter Vágner reshared this.

in reply to Schmaker

@schmaker before getting an electric car we paid 2.5-3k CZK for gasoline per month. The average petrol consumption cost was 3 CZK/km for the petrol car. When I charge from the solar system I only count the cost of missed opportunity for not being able to sell the energy to the grid. It's 1.25 CZK/kWh and with the average EV consumption of 15 kWh/100 km the cost per km is 0.2 CZK. 15 times lower.
Peter Vágner reshared this.

Anyone using #Tammy as a #Matrix messenger client?

What do you think about it? Why do you use it? How much do you use it? On which device type is it installed? What features do you miss?

  • Yes. (15%, 3 votes)
  • No, but already tried it. (5%, 1 vote)
  • No, but heard about it. (15%, 3 votes)
  • No, never heard about it. (65%, 13 votes)
20 voters. Poll end: 1 month ago

Peter Vágner reshared this.

in reply to Benedict

First of all: Thank you for your work!
I use it on Android alongside with other #Matrix Clients.
So far I remember, the onboarding was good. Generelly it feels fast (after) sync.
The chatlist items, for me, could have a little bit more padding. The gui with the bubble style in the Chats I don't like much.
It's nice to have some settings for gui. The accent color could have more options.
Peter Vágner reshared this.

After a short break, I’m returning to accessibility training services.

I provide one-on-one training for blind and visually impaired users across multiple platforms. My teaching is practical and goal-driven: not just commands, but confidence, independence, and efficient workflows that carry into daily life, study, and work.

I cover:
iOS: VoiceOver gestures, rotor navigation, Braille displays, Safari, text editing, Mail and Calendars, Shortcuts, and making the most of iOS apps for productivity, communication, and entertainment.
macOS: VoiceOver from basics to advanced, Trackpad Commander, Safari and Mail, iWork and Microsoft Office, file management, Terminal, audio tools, and system upkeep.
Windows: NVDA and JAWS from beginner to advanced. Training includes Microsoft Office, Outlook, Teams, Zoom, web browsing, customizing screen readers, handling less accessible apps, and scripting basics.
Android: TalkBack gestures, the built-in Braille keyboard and Braille display support, text editing, app accessibility, privacy and security settings, and everyday phone and tablet use.
Linux: Orca and Speakup, console navigation, package management, distro setup, customizing desktops, and accessibility under Wayland.

Concrete goals I can help you achieve:
Set up a new phone, tablet, or computer
Send and manage email independently
Browse the web safely and efficiently
Work with documents, spreadsheets, and presentations
Manage files and cloud storage
Use social media accessibly
Work with Braille displays and keyboards
Install and configure accessible software across platforms
Troubleshoot accessibility issues and build reliable workflows
Make the most of AI in a useful, productive way
Grow from beginner skills to advanced, efficient daily use

I bring years of lived experience as a blind user of these systems. I teach not only what manuals say, but the real-world shortcuts, workarounds, and problem-solving skills that make technology practical and enjoyable.

Remote training is available worldwide.

Pricing: fair and flexible — contact me for a quote. Discounts available for multi-session packages and ongoing weekly training.

Contact:
UK: 07447 931232
US: 772-766-7331
If these don’t work for you, email me at aaron.graham.hewitt@gmail.com

If you, or someone you know, could benefit from personalized accessibility training, I’d be glad to help.

#Accessibility #Blind #VisuallyImpaired #ScreenReaders #JAWS #NVDA #VoiceOver #TalkBack #Braille #AssistiveTechnology #DigitalInclusion #InclusiveTech #LinuxAccessibility #WindowsAccessibility #iOSAccessibility #AndroidAccessibility #MacAccessibility #Orca #ATTraining #TechTraining #AccessibleTech

reshared this

Peter Vágner reshared this.

Yt-dlp: Upcoming new requirements for YouTube downloads

Link: github.com/yt-dlp/yt-dlp/issue…
Discussion: news.ycombinator.com/item?id=4…

#youtube

reshared this

Peter Vágner reshared this.

lue is a TUI ebook reader with text to speech (TTS) support.

It can read EPUB / DOCX / PDF / TXT / files, supports 100+ languages, highlights words in sync, saves your progress, has themes and more.

Starry Eyes (superstarryeyes on GitHub) made lue using Rich, a Python library by @textualize and is Terminal Tool of the Week! ⭐️

terminaltrove.com/lue/

reshared this

Peter Vágner reshared this.

Here it is, my new, self-hosted home in the #Fediverse running on #GoToSocial! This is really exciting stuff, now I’m truly living the Fedi spirit by supporting decentralization. If you're reading this and don't mind, I'd greatly appreciate a boost of this post to help my tiny new instance discover more servers. Thanks! #NewFedi #FediAdmin #Selfhosting

reshared this

in reply to Talon

Thanks so much for clarifying and for the explanations! I was under the impression that boosting would indeed result in my server pulling in more posts from other instances, but I guess that was a misconception on my part having seen others doing this after spinning up a new server. I did follow many of the people who boosted though as a way to say thank-you, so perhaps this actually did help extend my server's reach by extension :-). GTS does not support relays yet sadly, although it is on their roadmap, I only realized it after I'd already configured most things so there was no turning back. That being said though, I do feel GTS is absolutely the right choice for my single-user scenario, setting up Mastodon or one of its forks would likely be overkill for my small Fedi home. That being said, I did find this project which seems interesting, maybe I should look into setting this up: codeberg.org/tante/hypebot
Peter Vágner reshared this.

Letos volím poštou

xn--ondej-kcb.v.nizozemsku.nl/2025/09/21/Letos_volim_postou.html

Zatímco v Česku vrcholí horká fáze kampaně, já už mám odvoleno. Spolu s více než dvaceti tisíci krajany jsem se rozhodl letos poprvé vyzkoušet korespondenční hlasování. To se dostalo do novely zákona o správě voleb a umožňuje Čechům žijícím v zahraničí…

reshared this

Yeah, I've updated my @Arch Linux to @GNOME 49.
There are some nifty #a11y related tweaks such as better labelling for gnome shell menus, refreshed settings UI, I like how presentation of various lists e.g. List of wireless networks is presented with screen reader including signal strength.

Thanks to everyone involved for the improvements.

Peter Vágner reshared this.

Hi, I'm micr0, the creator of Altbot.

Almost a year ago, your incredible generosity helped us raise the funds to buy the server that Altbot runs on today (locally and privately) It's been operating from my home ever since, and I'm so grateful for the support that made that possible.

But now, the situation has become unsustainable. My home network is under a sustained, targeted DDoS attack aimed at taking Altbot offline. And unfortunatly this isn't just a threat to the bot, it's a serious security and privacy concern for my family.

A lot of people are probabaly going to be asking the same question I did: "Who is doing this?"
but the honest answer is: I don't know, and I likely never will. These attacks are launched through botnets and proxies designed specifically to hide the source. Figuring out the "who" is nearly impossible. The only thing I can do is focus on the "how to stop it."

Running this critical service from a residential address is no longer viable. To protect Altbot and my family, we need to move the server to a professional data center with proper, enterprise DDoS mitigation.

The Goal: $2,880 to cover 12 months of secure colocation.

This will provide a secure, stable home for Altbot with:

  • Enterprise-grade DDoS protection
  • 99.95%+ uptime guarantee
  • 24/7 monitoring and security
  • Separation from my personal home network

Donations can be made via:

This isn't just about maintaining a service. It's about ensuring that an important accessibility tool remains available for everyone who depends on it, while also protecting my family's privacy and safety.

Please consider supporting if you can. If you're unable to donate, boosts are incredibly valuable for raising awareness.

Thank you for your support and for believing in Altbot's mission.

#Altbot #Accessibility

reshared this