One Week with NVDA: A JAWS User’s Immersion Journey

What started as a seven-day experiment ended with a new primary screen reader.

I’ll be honest: I didn’t expect this to go the way it did. On February 14th, 2026, I set myself a challenge — use NVDA exclusively on my personal computer for one full week, switching back to JAWS only if my work required it. I’ve been a longtime JAWS user, and NVDA has always been on my radar as the powerful, free, open-source alternative. But radar is different from reality. So I dove in.

One week later — and several days beyond that — I’m still running NVDA. It has become my primary Windows screen reader. I won’t be abandoning JAWS entirely; both tools have their place. But if you’ve been on the fence about giving NVDA a serious try, read on. Here’s everything that happened.

Day 1 (February 14): First Impressions and the Punctuation Problem

The very first thing that tripped me up was punctuation. NVDA defaults to “some” punctuation, while I was accustomed to “most” in JAWS. The practical effect: symbols like the underscore were being silently skipped. I switched to “most” punctuation right away, and that helped — but it opened its own can of worms.

In “most” mode, NVDA announces the underscore as “line.” I found that maddening. The colon inside timestamps (insert+F12 for the time) was also being spoken aloud, which felt odd. These were small things, but they added up quickly.

I also explored the NVDA Addon Store. It’s a great concept, but I found the execution a bit rough — many addons lack solid documentation, and reading user reviews means navigating away to an external website. There’s room to grow here.

One more early grievance: common commands like Control+C and Control+S are completely silent in NVDA. You press copy or save and hear… nothing. The option to speak command keys does exist, but it makes everything chatty — tab, arrows, all of it. That’s not what I wanted either.

Day 2 (February 15): Muscle Memory Wars and Customization Overload

Day two was the most turbulent. My JAWS muscle memory fought me at every turn, and I spent a significant portion of the day not doing productive work but rather reconfiguring NVDA to survive.

Browse Mode and Focus Mode were a constant source of confusion. In JAWS, Semi Auto Forms Mode handles a lot of this context-switching behind the scenes. With NVDA, I found myself stuck in the wrong mode repeatedly. A simple example: after submitting a prompt to Gemini and hearing its reply, I pressed H to navigate to the heading where the response started. NVDA just said “h” and sat there. I was still in Focus Mode. Insert+Space toggled Browse Mode on and then everything worked — but I had to consciously remember to do that. This will likely get easier with time, but on day two, it was genuinely frustrating.

I remapped a fistful of commands to save my sanity. The NVDA find command in Browse Mode is Control+NVDA+F — not Control+F — which felt deeply wrong. I added Control+F, F3, and Shift+F3 under Preferences > Input Gestures. I also kept repeatedly bumping into Insert+Q being the command to exit NVDA rather than announcing the active application, which nearly gave me a heart attack the first time it happened. I enabled exit confirmation in Preferences > General, then later reassigned Insert+Q to announce the focused app and moved the exit command to Insert+F4.

The underscore-as-“line” issue got its resolution today. The fix wasn’t in NVDA’s speech dictionaries as I first expected — it was in Preferences > Punctuation/Pronunciation. Problem solved. I also tackled the exclamation mark, which sits in the “all” punctuation tier rather than “most.” I mapped it to announce as “bang” when it appears mid-sentence.

There was also a frustrating addon conflict: the NVDA+Shift+V keystroke, officially assigned to announce an app’s version number, was instead being intercepted by the Vision Assistant Pro addon to open its command layer. Addon keystrokes can silently override core NVDA functionality — something worth knowing. I ended up assigning Control+NVDA+V to get version info.

One gap I noticed that NVDA doesn’t yet fill: quickly reading the current page’s URL without shifting focus to the address bar. JAWS handles this with Insert+A. NVDA doesn’t have an equivalent. Alt+D works, but it moves focus, which isn’t always what I want.

Day 3 (February 16): The Good, The Annoying, and a Genuine Win

By day three — President’s Day — I was settling into something like a rhythm, though NVDA was still throwing surprises at me.

One thing I couldn’t crack was typing echo. In JAWS, I run character-level echo at a much higher speech rate than everything else. This gives me fast, confident confirmation of each keystroke without slowing down general speech. NVDA doesn’t appear to support different speech rates per context, so typed characters come through at the same rate as everything else. I know I can’t be the only person who relies on this, so I kept digging — but no solution yet.

I also noticed a recurring issue: NVDA going silent after focus changes. Closing Excel or Word and returning to File Explorer? Silence. Switching browser tabs with Control+Tab? Sometimes silence. This felt like potential bug territory.

PDFs were another pain point. I work with many poorly tagged PDFs, and NVDA with Adobe Reader exposes every formatting flaw without mercy. JAWS has historically done more smoothing and pre-processing before those errors reach the user. I’m withholding final judgment here — there are third-party PDF tools that work well with NVDA, and I planned to test them.

I experimented briefly with turning off automatic say-all on page load to reduce repetitive speech on websites. Bad idea. After toggling an action, nothing was announced — I had to manually navigate just to figure out where I had ended up. I turned it back on immediately.

The genuine win of the day: the Vision Assistant Pro addon. While working on a freelance project that required a visual description of a web page’s layout, I pressed NVDA+Alt+V then O for an on-screen description. Within seconds I had exactly what I needed. A follow-up question was answered just as quickly. Cross-checking with other tools confirmed the accuracy. This was an impressive moment and a real argument for NVDA’s addon ecosystem.

Day 4 (February 17): The 32-Bit Revelation and Eloquence Arrives

I learned something on day four that genuinely surprised me: NVDA 2025.3.3, the current stable release, is 32-bit. I had assumed for years that I was running a 64-bit screen reader. This discovery came about through an unexpected path.

I came across a link to a 64-bit version of the Eloquence speech synthesizer built for NVDA. Excited, I installed it and restarted — only to find NVDA using Windows OneCore voices with no trace of Eloquence. After posting about it on Mastodon, the community quickly pointed out the 32-bit issue. The 64-bit Eloquence addon requires a 64-bit NVDA, which only exists in the 2026 beta builds. I grabbed the beta, installed everything, and was finally running Eloquence on NVDA. The 64-bit upgrade is coming in the official 2026.1 release — well worth watching for.

I also continued searching for an NVDA equivalent to JAWS’s Shift+Insert+F1, which gives a detailed browser-level view of an element’s tags, attributes, roles, and IDs. This is invaluable for accessibility work. I hadn’t found a satisfying answer by end of day.

Day 5 (February 18): Discovering NVDA in Microsoft Word

I don’t often think of Browse Mode as a Word feature, so I was pleasantly surprised to learn — after reading some documentation — that NVDA supports a version of it in Word, allowing quick navigation by headings using the H key. This made my document work much more manageable.

I also received another update to 64-bit Eloquence, which fixed bugs I hadn’t even noticed. As for the work computer, I decided against installing the NVDA beta there — my employer deserves results from the stable release. That upgrade will wait for the official 2026.1 launch.

Day 6 (February 19): The Quiet Day

Day six was uneventful in the best possible way. I used my computer heavily and NVDA just worked. No major incidents, no emergency remappings. I noticed I was reaching for JAWS less and less in my thoughts. That felt significant.

Day 7 (February 20): Amateur Radio and a Happy Ending

The final day of the official challenge coincided with the start of the ARRL International DX CW (Morse Code) contest — one of the bigger amateur radio events of the year. I was curious how N3FJP’s contest logging software would hold up with NVDA, since this is specialized, legacy-adjacent software that doesn’t rely on standard accessibility APIs.

The answer: it worked great — and actually felt snappier than with JAWS. The one wrinkle was reviewing the call log. The standard screen review commands on the numpad didn’t yield useful information at first. The solution was object navigation. By pressing NVDA+Numpad 8 to climb to the parent object (“call window”), I found that each column in the log is its own object. Navigating with NVDA+Numpad 4, 5, and 6 moved between objects at the same level, announcing “Rec Window,” “PWR Window,” “Country Window,” “Call Window,” and so on. From there, Numpad 9 and 7 moved through the log in reverse chronological order. Once I understood the structure, it worked beautifully.

My two radio control apps — JJRadio and Kenwood’s ARCP software — also worked flawlessly. Just when I was expecting NVDA to hit its limits, it didn’t.

What NVDA Does Really Well

After a week of intensive use, here’s what impressed me most:

  • Speed and responsiveness. NVDA frequently felt faster than JAWS, especially in applications like the N3FJP logging software.
  • Deep customizability. The Input Gestures system makes it relatively easy to remap commands. Preferences > Punctuation/Pronunciation gives granular control.
  • The addon ecosystem. Despite rough edges, the Vision Assistant Pro addon alone demonstrated real power. The 64-bit Eloquence support is also a significant upgrade.
  • Object navigation. Once I understood NVDA’s object model, navigating legacy and non-standard interfaces became genuinely manageable.
  • Cost. NVDA is free, actively developed, and open source. The value proposition is extraordinary.

Where NVDA Still Has Room to Grow

  • Silent focus changes. NVDA going quiet after closing apps or switching tabs is disorienting and may be a bug worth filing.
  • PDF handling. Poorly tagged PDFs hit differently with NVDA than with JAWS, which smooths many errors before they reach the user.
  • Typing echo speech rate. The inability to set a faster speech rate specifically for typed characters is a real productivity gap for fast typists.
  • Element inspection. JAWS’s Shift+Insert+F1 for examining element attributes has no obvious NVDA equivalent, which matters for accessibility work when I just need to start with a quick-and-dirty answer before digging deeper into the code.
  • URL reporting without focus change. A read-only way to hear the current page address — without moving focus to the address bar — is missing.
  • Addon documentation and conflict resolution. Keystroke conflicts between addons and core NVDA aren’t surfaced clearly enough.

The Verdict: One Week Became the New Normal

I went in expecting to survive a week and then gratefully return to JAWS. Instead, I’m writing this article as an NVDA user. The first two days were genuinely hard — partly NVDA’s rough edges, partly years of JAWS muscle memory fighting back. But by day six, NVDA was simply humming along, and I wasn’t thinking about JAWS at all.

For experienced JAWS users considering a serious NVDA trial, my main advice is this: budget real time for reconfiguration in the first two days. The defaults won’t feel right. But the tools to make NVDA feel right are mostly there — they just require some digging. Preferences > Punctuation/Pronunciation and Input Gestures will be your best friends.

JAWS isn’t going anywhere in my toolkit. For professional accessibility auditing, PDF work, and certain specialized contexts, it remains the gold standard. But for day-to-day use on my personal computer? NVDA has earned the top spot.

The 2026.1 release — bringing official 64-bit support — is going to be a milestone worth watching. If you’ve been waiting for a good moment to give NVDA a real chance, that moment is here, now.

Sources

This article is primarily a firsthand account based on my direct experience. The following resources document or corroborate the specific factual claims made in the article.

  • NV Access: NVDA 2025.3.3 Released — Official release announcement for the stable version of NVDA tested throughout this article, confirming it is a 32-bit build.
  • NV Access: In-Process, 10th February 2026 — NV Access’s own blog post confirming that NVDA 2026.1 is the first 64-bit release, and discussing the scope of that transition.
  • NV Access: NVDA 2026.1 Beta 3 Available for Testing — The beta release announcement for the 64-bit version of NVDA referenced in the Day 4 entry.
  • NVDA 2025.3.3 User Guide — The official NVDA documentation covering Browse Mode, Focus Mode, Input Gestures, object navigation, Punctuation/Pronunciation settings, and the Add-on Store — all features discussed throughout the article.
  • Switching from JAWS to NVDA — A community-maintained transition guide for experienced JAWS users switching to the free, open-source NVDA screen reader, covering key differences in keyboard commands, terminology, cursors, navigation, synthesizers, settings, add-ons, and common troubleshooting scenarios.
  • N3FJP’s ARRL International DX Contest Log — The official page for the N3FJP contest logging software tested with NVDA on Day 7.
  • ARRL International DX Contest — The American Radio Relay League’s official page for the ARRL International DX CW contest referenced in the Day 7 entry.

Blind Access Journal Launches Community Effort to Improve WSJT-X Accessibility for Aging and Disabled Amateur Radio Operators

FOR IMMEDIATE RELEASE

Blind Access Journal Launches Community Effort to Improve WSJT-X Accessibility for Aging and Disabled Amateur Radio Operators

Peoria, Arizona — December 20, 2025 — Darrell Hilliker, NU7I, a totally blind Amateur Radio operator and accessibility professional, is spearheading a community initiative to improve the accessibility of WSJT-X (and WSJT-X Improved) for blind, low-vision, and mobility-impaired hams. The work is being organized and documented through Blind Access Journal, the blog Hilliker publishes to advance practical accessibility and inclusion in technology.

Digital weak-signal protocols like FT8 have become a core part of modern Amateur Radio. Yet many hams—especially those who are aging or who acquire disabilities—are finding it harder to participate fully when widely used software lacks accessible user interface foundations.

“A month doesn’t go by where I don’t hear at least one conversation on the bands where an older ham is contemplating giving up or curtailing their activities due to a physical disability like arthritis or a visual impairment,” said Hilliker. “We can do better as a community—and we can do it together.”

Recognizing Existing Innovation and Building an Inclusive Future

This initiative is not a critique of existing community solutions, nor is it intended to replace them. Blind Access Journal recognizes and commends the developers of alternative tools such as QLog, whose efforts have helped many operators. Instead, Hilliker’s project aims to broaden inclusion by improving accessibility in the widely adopted WSJT-X ecosystem so that more hams can participate using the tools their clubs, friends, and on-air communities already rely on.

“The entire Amateur Radio community benefits from all efforts to adapt,” Hilliker added, “especially in situations where disabled hams are not fully included from the beginning.”

Goal: Full and Equitable Access to Digital Operating

The initiative’s objective is nothing less than full and equitable access to Amateur Radio digital communication protocols and the software that enables them. Key accessibility goals include:

  • Expected keyboard navigation throughout the application
  • Strong compatibility with screen readers such as JAWS and NVDA (NonVisual Desktop Access)
  • UI that can reflow and resize for operators using magnification
  • Support for dark mode, high contrast, and other visual accommodations that many aging hams depend on

Highest Priority Technical Need

The most critical improvement—especially for blind screen-reader users—centers on the Band Activity and Rx Frequency tables. Today, these areas are widely experienced as inaccessible because the data is effectively “painted” to the screen or presented as unstructured text, rather than implemented using the underlying Qt5 UI structures that expose information to accessibility interfaces.

The initiative seeks a redesign and implementation approach that ensures these tables are true, semantically structured UI components—so assistive technologies can reliably read, navigate, and interact with them.

Call for Volunteer Developers

Blind Access Journal is calling on a small group of experienced Amateur Radio software builders and tinkerers—especially those who:

  • Have deep experience with Qt5 user interfaces
  • Can build and compile WSJT-X or WSJT-X Improved from source with confidence
  • Are willing to collaborate with disabled hams in an open, test-driven, user-centered process

Familiarity with accessibility design and standards such as WCAG (Web Content Accessibility Guidelines) is welcome but not required. Disabled hams involved in the effort are prepared to lead the process, define needs, perform testing, write documentation, and support the work in every way outside of the core design and coding tasks.

Volunteers will gain the satisfaction of delivering long-sought, meaningful accessibility improvements to a widely used mainstream Amateur Radio application—work that can make a real difference for thousands of fellow hams.

Looking Toward 2026

Blind Access Journal thanks the Amateur Radio community for its time, creativity, and tradition of public service. The initiative’s organizers hope to make 2026 a year of digital accessibility and inclusion for all radio amateurs.

To volunteer or learn more:
Email editor@blindaccessjournal.com and follow updates via Blind Access Journal.

Media Contact

Darrell Hilliker, NU7I
Blind Access Journal
Email: editor@blindaccessjournal.com

Using Apple’s Built-In Accessibility Features to Reduce Screen Exposure During Severe Headaches

Summary

Some people experience severe headaches or migraines that make screen use difficult—especially when light sensitivity (photophobia) and flicker or refresh effects are major triggers. While display adjustments can help, there are days when the most effective strategy is to reduce visual reliance as much as possible.

If you use an iPhone and Mac, Apple includes several built-in accessibility tools that can support a “low-screen” or even “no-screen” workflow—particularly for everyday tasks like reading and replying to email.

This article focuses on the built-in Mail app and outlines a practical approach using:
VoiceOver (screen reader),
Voice Control (hands-free voice operation),
and Dictation (speech-to-text composition).


Why VoiceOver and Voice Control can help when light and flicker are triggers

VoiceOver reads on-screen content aloud and provides a structured navigation model that does not require visually scanning the interface. Instead of looking for buttons or reading text, users move through content sequentially and receive spoken feedback.

Voice Control complements this by allowing users to operate their device through spoken commands. Actions such as opening Mail, scrolling, replying, and sending messages can often be completed without touching or looking closely at the screen.

For people whose primary headache triggers include light sensitivity and flicker, combining these tools can significantly reduce both the duration and intensity of screen exposure.


iPhone: Building a low-screen Mail workflow on iOS

Turn on VoiceOver

VoiceOver can be enabled from Settings > Accessibility > VoiceOver. Apple provides a built-in practice experience that introduces the gesture model and basic navigation concepts.

Learn a minimal set of VoiceOver gestures

It is not necessary to learn every gesture. Starting with a small core set allows users to begin working quickly and add complexity later.

  • Swipe right: move to the next item.
  • Swipe left: move to the previous item.
  • Double-tap: activate the selected item.
  • Two-finger swipe up: read the entire screen from the top.
  • Two-finger tap: pause or resume speech.
  • Four-finger tap near the top: jump to the first item.
  • Four-finger tap near the bottom: jump to the last item.

Use Screen Curtain to eliminate display light

When VoiceOver is enabled, the screen itself can be turned off while the device remains fully usable. This feature, called Screen Curtain, allows users to rely entirely on audio output while avoiding light exposure.

  • Three-finger triple-tap: toggle Screen Curtain on or off.
  • If both Zoom and VoiceOver are enabled, a three-finger quadruple-tap may be required.

Adding Voice Control for hands-free interaction

Voice Control allows users to interact with on-screen elements using spoken commands. This can be particularly helpful when precise touch input or visual targeting is uncomfortable.

Common Voice Control commands

  • Open Mail
  • Scroll down / Scroll up
  • Go home
  • Show names (labels interface elements)
  • Show numbers (adds numbered overlays)

When an on-screen control is difficult to activate, VoiceOver can be used to identify the control’s name, and Voice Control can then activate it using that spoken label.


Reading and replying to Mail on iPhone using audio

  1. Open the Mail app using Voice Control or VoiceOver navigation.
  2. Move through the message list using swipe left and swipe right.
  3. Open a message with a double-tap.
  4. Listen to the message using a two-finger swipe up.
  5. Reply using Voice Control or VoiceOver navigation.
  6. Compose the reply using Dictation, speaking punctuation as needed.
  7. Send the message using a spoken command or VoiceOver activation.
  8. Enable Screen Curtain when light sensitivity is a concern.

Mac: Reducing visual load with VoiceOver

On macOS, VoiceOver enables spoken feedback and keyboard-based navigation across apps, including Mail. This allows users to work with less reliance on visual scanning.

Turn VoiceOver on or off

  • Command + F5: toggle VoiceOver.

Core VoiceOver navigation concepts

The VoiceOver cursor moves independently of the system focus and determines what is spoken. Navigation is performed using the VoiceOver modifier keys (often Control + Option).

  • VO + Arrow keys: move between items.

Quick Nav for streamlined navigation

Quick Nav can simplify navigation by allowing arrow keys or single keys to move through content without holding modifier keys. This can be especially useful once the basics feel comfortable.

  • VO + Q: toggle single-key Quick Nav.
  • VO + Shift + Q: toggle arrow-key Quick Nav.

Pacing and learning considerations

When screen exposure can trigger symptoms quickly, it helps to approach learning incrementally.

  • Practice in short sessions (5–10 minutes).
  • Focus first on listening and basic navigation.
  • Add Screen Curtain early if light sensitivity is significant.
  • Introduce Voice Control gradually for common actions.

Sources

Demonstration: Guide Accessifies the Addition of Components to Salesforce Experience Cloud Site Pages

At the intersection of the Salesforce ecosystem and the accessibility community, it has been long known that Experience Builder contains task-blocking accessibility issues that hold many disabled people back from being able to perform important job duties including site administration and content management. While the company continues efforts to improve the accessibility of Experience Builder, disabled administrators, content managers and site developers who rely on keyboard-only navigation and screen readers are finding ways to work around barriers thanks to new tools based on artificial intelligence (AI).

Read more

Finding Balance: Easter in the Time of the Coronavirus

In this approximately 24-minute podcast, Allison, Allyssa, Arabella and Darrell Hilliker celebrate Easter in the time of the Coronavirus while putting together the Alex Active Monkey Kids Toddler Balance Board.

Download: Finding Balance: Easter in the Time of the Coronavirus

Please tell us how you and your family are handling social distancing, feeding yourselves and generally getting along, especially from a blind perspective, in the time of the Coronavirus. Please send an audio recording or a written message to darrell (at) blindaccessjournal (dot) com or tell us about it on our social media channels.

AccessiLife Consulting, Blind Access Journal, and the Hilliker family, must frequently rely on sighted assistance in order to get important, inaccessible tasks done. In most cases, we have chosen Aira as our visual interpreter. If you are ready to become an Aira Explorer, and you feel it in your heart to pass along a small gift to the journal or our family, we ask that you use our referral link. Your first month of Aira service will be free of charge, we will receive a discount on our bill and we will thank you for supporting the important work we do here at Blind Access Journal.

We love hearing from our listeners! Please feel free to talk with us in the comments. What do you like? How could we make the show better? What topics would you like us to cover on future shows?

If you use Twitter, let’s get connected! Please follow Allison (@AlliTalk) and Darrell (@darrell).

Exploring the World with Aira: A Candid Discussion with Suman Kanuganti

In this approximately 50-minute informal podcast, Allison, Allyssa and Darrell Hilliker talk with Aira CEO Suman Kanuganti at the 2017 National Federation of the Blind Convention about this promising, new disruptive service that aims to give blind people “instant access to information” through the use of augmented reality and wearable technology.

We love hearing from our listeners! Please feel free to talk with us in the comments. What do you like? How could we make the show better? What topics would you like us to cover on future shows?

If you use Twitter, let’s get connected! Please follow Allison (@AlliTalk) and Darrell (@darrell).

Who Will Be Your Eyeballs Today? Exploring Remote Sighted Assistance with Be My Eyes for iOS

In this approximately 23-minute podcast, Allison Hilliker and Darrell Shandrow demonstrate the new Be My Eyes app developed to help blind people remotely acquire the sighted help they need to perform many daily visual tasks.

Resources

We love hearing from our listeners! Please feel free to talk with us in the comments. What do you like? How could we make the show better? What topics would you like us to cover on future shows?

If you use Twitter, let’s get connected! Please follow Allison (@AlliTalk) and Darrell (@darrell).

ATIA: Making Friends with Eye-Pal Scanning and Reading Products

In this approximately 14-minute podcast from the 2013 Assistive Technology Industry Association conference in Orlando, Allison Hilliker talks with Leon Reznik, president of ABiSee, about his company’s Eye-Pal line of scanning and reading solutions for blind and low-vision people.

If you use Twitter, let’s get connected! Please follow Allison (@AlliTalk) and Darrell (@darrell).

Join Us for a Special Q&A Session with Fleksy

You’ve got burning questions? We’ve got the answers!

Join us on Monday, July 23 at 6:15 p.m. Pacific, 9:15 p.m. Eastern time in the Out-Of-Sight Presents room on Out-Of-Sight as we host a special Q&A session with the developers of the hot new Fleksy typing app for iOS devices.

You can even ask your question before the live presentation! Please visit Fleksy’s Happy Typing website to ask the Fleksy team anything. We will ask the pre-submitted questions first, then we will move to live audience participation.

Please register as soon as possible for a free Out-Of-Sight membership so you can participate and contact us before the event if you would like assistance.

Freedom Scientific Acquires Blind Access Journal

In a move that is sure to surprise the blindness assistive technology industry, Freedom Scientific has just announced that it has purchased the popular accessibility evangelism website BlindAccessJournal.com for $100,000 from its publisher, Darrell Shandrow.

Freedom Scientific representatives said this move puts the company on a par with other assistive technology companies like Serotek that also have recognized media outlets in the blind community.

“We’re excited to have Blind Access Journal on board with us,” said Jonathan Mosen, Freedom Scientific’s vice president of hardware development. “We know Darrell will put his passion for accessibility evangelism into his new role as our director of public relations, where he will be an excellent advocate for the use of our products as a means to make accessibility happen.”

“It was a very hard sell! After intensive negotiations concerning the content that will be permitted on the site, I must say I am honored to become part of the company in the industry with the largest screen-reader market share,” said the journal’s publisher, Darrell Shandrow. “I know that only JAWS can truly provide the level of accessibility we as blind people need now and well into the future.”