Pro Tip: Fixing a Disappearing Interface in FastSM for Windows

FastSM is a popular accessible client for Bluesky and Mastodon, and for the most part it works beautifully with screen readers like JAWS and NVDA. However, some users have encountered a frustrating issue: shortly after FastSM starts, performing any action — such as trying to create a new post — causes the entire user interface to vanish. When this happens, the app is still running in the background but is completely unresponsive, and the usual fix of pressing Control+Windows+W does not bring it back. The only way out seems to be manually ending the FastSM process in Task Manager.

If this has happened to you, don’t worry. There are three solutions below, starting with the simplest and working up to a more thorough reset. Try them in order and stop as soon as one works.

Solution 1: Restore the Window with a Keyboard Shortcut

Before trying anything else, press Control+Windows+W. This shortcut is designed to restore FastSM’s standard visible interface if it has been accidentally hidden. If this brings your interface back, you’re done! If it doesn’t work, move on to Solution 2.

Solution 2: Turn Off the Invisible Interface in Settings

FastSM has a feature called the Invisible Interface, which, if accidentally enabled, can cause the window to disappear when you interact with the app. Here’s how to check and disable it:

  1. Start FastSM and, before doing anything else, press Control+Comma to open Global Options.
  2. Press Control+Tab five times to navigate to the Invisible Interface tab.
  3. Listen for whether “Enable Invisible Interface” is announced as checked. If it is checked, press Spacebar to uncheck it.
  4. Press Enter to save your changes and close the dialog.

FastSM should now behave normally. If the problem persists, Solution 3 below will give you a fresh start.

Solution 3: Reset FastSM by Removing Its Settings Folder

This is the most thorough fix. It completely removes FastSM’s stored settings so that the app starts fresh, just as it did the first time you installed it. Important: after completing these steps, you will need to sign back into all of your Bluesky and Mastodon accounts. If you are not sure whether you have administrator access to your computer, please ask your IT department or another trusted person for help before proceeding.

These steps have been tested on Windows 11 Workstation (version 25H2) with the latest versions of both JAWS and NVDA as of April 2026.

Step 1: Close FastSM Using Task Manager

  1. Press Control+Shift+Escape to open Task Manager.
  2. Press F repeatedly until you hear something like “FastSM collapsed level 1.”
  3. Press Alt+E to end the FastSM process.
  4. Press Alt+F4 to close Task Manager.

Step 2: Open Your Home Folder in File Explorer

  1. Press Windows+R to open the Run dialog.
  2. Type a single period (.) and press Enter. This opens your personal home folder in File Explorer.

Step 3: Show Hidden Files and Folders

  1. Press Alt by itself to focus the File Explorer ribbon.
  2. Press Right Arrow until you hear “View button collapsed” and press Enter to open the menu.
  3. Press Up Arrow until you hear “Show collapsed,” then press Enter to open the submenu.
  4. Press Up Arrow until you hear “Hidden Items.”
  5. If “Hidden Items” is not announced as checked, press Spacebar to enable it. If it is already checked, press Escape to leave this menu without making changes.

Step 4: Switch to Details View for Easier Navigation

  1. Press Alt by itself, then press Right Arrow five times, then press Enter to reopen the View menu.
  2. Press Down Arrow until you hear “Details,” then press Enter. Note: even if “checked” is announced, pressing Enter here will not change anything — it simply confirms the Details view.

Step 5: Navigate to and Delete the FastSM Settings Folder

  1. In the file list, press Down Arrow until you hear “AppData,” then press Enter to open it.
  2. Press Down Arrow until you hear “Roaming,” then press Enter to open it.
  3. Press Down Arrow until you hear “FastSM,” then press Delete to remove the folder.
  4. If a dialog appears asking for permission, press Tab until you hear “Continue” or “Yes,” then press Enter to confirm.

Step 6: Restart FastSM and Sign Back In

  1. Press Windows+D to go to the desktop.
  2. Press F until you hear “FastSM,” then press Enter to launch the app.
  3. FastSM will start as if it has been installed for the first time. Sign back into your Bluesky and Mastodon accounts as you did during the original setup.

Wrapping Up

A disappearing interface can be alarming, but as you’ve seen, there are reliable ways to get FastSM back on track without losing access to your social feeds for long. In most cases, the quick keyboard shortcut or a single settings change in Global Options will solve the problem in under a minute. The full reset, while a bit more involved, is a dependable last resort that leaves you with a clean, working installation.

FastSM continues to be one of the most screen-reader-friendly ways to enjoy Bluesky and Mastodon on Windows, and hopefully this guide helps you spend less time troubleshooting and more time connecting with your communities.

The AI Design Gap: A Student’s Journey in Accessifying Visual Layouts

As a Certified Professional in Web Accessibility (CPWA), I spend my days ensuring the web works for everyone. But as a student currently enrolled in a design course, I recently hit a wall that even my expertise combined with advanced artificial intelligence couldn’t easily scale.

The assignment was straightforward for most: Review a series of design samples and identify the visual layout being used—specifically, patterns like Z-shape, Grid of Cards, or Multi-column. For a blind student, however, this wasn’t just a design quiz; it was an accessibility challenge in its own right.

The Assignment

I was working through a module on Understanding Website Layouts. While the course platform itself was technically navigable, the “Design” samples provided were purely visual. To complete the assignment and select the corresponding layout buttons, I needed to understand the spatial arrangement of elements I couldn’t see.

I turned to a powerful ally: the March 2026 release of JAWS and its Page Explorer feature. By pressing Insert + Shift + E, I invoked Vispero’s AI-driven summary to “accessify” the assignment’s visual content.

The Experiment (and the “Failure”)

For the first sample, Page Explorer described the main area as “divided into two large colored panels side-by-side or stacked.” Based on this, I guessed Grid of Cards.

Incorrect. The system informed me a grid features a series of cards providing previews of more detailed content.

I tried again with the next sample. This time, I asked the AI specifically to describe the layout from a “design perspective.” It responded with details about a “white rounded rectangular card with a subtle shadow” and “prominent headings.” It sounded exactly like a Grid of Cards.

Incorrect again. The correct answer was a Z-shape layout, which encourages users to skim from left to right, then diagonally.

The Lesson Learned

This experiment was a “failure” in terms of getting the points on my assignment, but a massive success in highlighting where we are in the evolution of Assistive Technology:

  • Identification vs. Synthesis: The AI is getting incredibly good at identifying objects (buttons, shadows, panels). However, it hasn’t quite mastered the synthesis of those objects into cohesive design patterns like “Z-shape.”
  • The Subjectivity of Layout: Design patterns are often about the intended eye-path, a concept that is still a “work-in-progress” for even the most advanced generative models.

A Hopeful Future for Blind Designers

Despite the frustration of getting those “Incorrect” marks on my coursework, I’m deeply hopeful. The very fact that I can now have a “conversation” with my screen reader about the “subtle shadows” and “colored panels” of a design sample is a massive leap forward.

We are standing at the threshold of a new era. As AI models are trained more specifically on design heuristics and visual hierarchy, they will eventually move beyond simple description. They will become the “visual eyes” for blind designers, developers, and students, allowing us to not only participate in design courses but to master the visual language that has long been a barrier.

The experiment didn’t help me pass this specific assignment, but it proved that the tools are coming. We’re just a few iterations away from turning these “impossible” design hurdles into accessible milestones.


Video Demonstration

To see exactly how this played out in real-time, you can watch my screen recording below. In this video, I walk through the attempt to use JAWS Page Explorer to identify the layouts, showing both the AI’s descriptive output and the trial-and-error process of the assignment.

Not a Panacea: Why AI Browser Agents Haven’t Solved the Inaccessible Web—and What Comes Next

When Google launched Auto Browse for Gemini in Chrome in January 2026, a few of us in the blind and low-vision community felt a familiar surge of hope. Could this be the moment when the inaccessible web finally met its match? Could an AI that reasons about web pages—rather than merely reading their code—become the accessibility bridge we’d been waiting for? Microsoft’s Copilot Actions in Edge was already generating similar excitement. For the first time, it seemed like mainstream browser vendors were building tools with the potential to help us navigate software that had never been designed with us in mind.

The reality, as many of us have now discovered, is more complicated. Auto Browse and Copilot Actions are genuine advances—but they are not the panacea we had hoped for. Understanding why matters, both so we can use these tools wisely and so we can advocate effectively for the deeper changes our community needs.

How These Tools Work—and Why They Sometimes Don’t

Both Auto Browse and Copilot Actions belong to a new category called agentic AI browsers. Rather than simply reading out what is on a page, these tools attempt to reason about what you want to accomplish and then take action on your behalf—clicking buttons, filling in forms, navigating menus, even comparing prices across tabs.

Google’s Auto Browse uses Gemini 3, a multimodal model, running within a protected Chrome profile. It can “see” a page through a combination of the page’s underlying code and actual visual images of what the page looks like on screen. Microsoft’s Copilot in Edge takes periodic screenshots and uses those to understand and interact with the page. On a well-structured, accessible website, these approaches can be genuinely impressive.

On a good day, Gemini can select from a combobox that has no accessibility markup at all—because it can see the visual “shape” of the dropdown even when the code offers no semantic clues.

But the web we actually live on is not always well-structured. Enterprise applications like Salesforce Experience Cloud use complex architectural patterns—what developers call Shadow DOM, iframes, and dynamic rendering—that create serious obstacles for these AI tools. Shadow DOM, in particular, hides a component’s internal structure from outside scripts, which means the agent’s map of the page becomes fragmented and incomplete. When the agent tries to interact with a nested component inside such a structure, it may simply not be able to find it.

Drag-and-drop interactions present another profound challenge. A click is a discrete event: the agent identifies a target, fires a command, done. Dragging is a continuous conversation between the agent, the page, and the browser over time. The agent must hold a real-time, high-fidelity picture of the page’s geometry while issuing a rapid sequence of commands—press, move, release—in exactly the right rhythm. Most vision-based agents process a screenshot, wait one to two seconds for the AI model to interpret it, then send a command. By the time that command arrives, the drag event on the page may already have timed out. The result is the “hit-and-miss” experience many of us have encountered: sometimes it works, sometimes it doesn’t, and it’s often impossible to know which you’ll get before you try.

Security: The Wall We Keep Running Into

There is another reason these tools fall short on complex applications, and it has nothing to do with AI capability: security. Both Copilot and Auto Browse operate within the browser’s strict security model, which is designed to prevent one website from accessing or manipulating data from another.

Copilot in Edge operates in three modes—Light, Balanced, and Strict—that govern how freely it can act on a given site. In the recommended Balanced mode, it will ask for your approval on sites it doesn’t recognise, and it is outright blocked from certain sensitive interactions in enterprise applications. If a site isn’t on Microsoft’s curated trusted list, the agent may simply refuse to act, citing security concerns.

These restrictions are not arbitrary. A critical vulnerability discovered in 2026, catalogued as CVE-2026-0628, demonstrated that malicious browser extensions could hijack Gemini’s privileged interface to access a user’s camera, microphone, and local files. In response, browser vendors have tightened the controls on what their AI agents can do—particularly in authenticated enterprise sessions where the stakes of a mistake are high. The same protective walls that prevent attackers from abusing these agents also prevent the agents from helping us with the complex, authenticated workflows where we most need assistance.
The precautions taken to keep attackers out also keep our AI helpers from doing their job.

Enter Guide: A Different Approach

While the browser-native agents struggle with these constraints, a different kind of tool has been quietly demonstrating what’s possible when you step outside the browser sandbox entirely. Guide is a Windows desktop application built specifically for blind and low-vision users. Instead of working within the browser’s security model, Guide takes a screenshot of your entire computer screen and uses AI—powered by Claude—to understand what’s visible. It then acts by simulating physical mouse movements and keystrokes at the operating system level, exactly as a sighted colleague sitting at your keyboard would.

This seemingly simple difference has profound consequences. Because Guide operates at the OS level rather than inside the browser, it is not subject to the Same-Origin Policy restrictions that stop Copilot and Gemini in their tracks. There are no cross-origin security alarms triggered, no curated allow-lists to consult. If a human hand could drag a component onto a canvas in Salesforce Experience Builder, Guide can do it too—and it has been demonstrated doing exactly that.

Guide also does something that matters deeply for users who want to build their own competence: it narrates the steps it is taking. Rather than operating as an opaque black box that either succeeds or fails mysteriously, Guide shows its reasoning, which means users can learn the workflow, understand what went wrong when something fails, and even record successful interaction patterns for later reuse.

It is worth being clear about what Guide is not. It is not a general-purpose browser agent designed for everyone. It is a specialist tool, built with our specific needs in mind, for situations where conventional assistive technology runs aground on inaccessible interfaces. That focus is, in many ways, its greatest strength.

Why the Underlying Problem Remains

Guide, Auto Browse, Copilot Actions, and other agentic tools represent genuine progress. But it is worth naming honestly what none of them actually solve: the inaccessible web itself.

When a screen reader user cannot navigate a Salesforce Experience Builder page, the root cause is not a shortage of clever AI workarounds. The root cause is that the page was not designed with accessibility in mind. The Shadow DOM obscures its structure not because Shadow DOM is inherently inaccessible, but because the developers who implemented it did not expose the semantic information that assistive technologies need. The drag-and-drop interface offers no keyboard alternative because whoever built it did not consider keyboard users.
Layering an AI agent on top of a broken foundation is a workaround, not a solution. It can help in many situations—and we are grateful for any help we can get—but it introduces its own fragility. The agent’s success depends on the visual layout remaining stable, on the AI model making accurate inferences, on security policies remaining permissive enough to allow action. Any of these can change, and when they do, a workaround that worked yesterday may stop working today.

Research is increasingly clear that blind users often find it less effective to patch an inaccessible UI with an AI layer than to address the underlying semantic issues in the code. The global assistive technology market is projected to reach twelve billion dollars by 2030, and yet the fundamental problem—developers building interfaces that exclude us from the start—remains stubbornly persistent.

Reasons for Real Hope

It would be easy to read all of this as a counsel of despair, but that is not what the evidence suggests. There are genuine reasons for optimism, grounded in both technological development and regulatory change.

The Regulatory Landscape Is Shifting
The European Accessibility Act came into force in June 2025, requiring a wide range of digital products and services—including enterprise SaaS platforms—to meet accessibility standards. This is not a minor guideline; it carries legal weight that organisations cannot ignore. As companies face real accountability for inaccessible software, the economic calculus changes. Fixing the foundation becomes cheaper than defending against legal action or building ever-more-elaborate AI patches.

The Technical Path Forward Is Clear
The research community and the web standards world have identified what better AI-assisted accessibility should look like. The Accessibility Object Model—a richer, semantically meaningful representation of web pages designed specifically for assistive technologies—offers a stable foundation that could allow future AI agents to navigate complex applications far more reliably than today’s tools.

Emerging “semantic geometry” approaches map the visual elements a user can see back to the specific, interactable code nodes behind them, eliminating the coordinate-guessing that causes today’s agents to miss by a few crucial pixels. Multi-agent architectures, where a navigation specialist, an execution agent, and a supervisory agent work in concert, promise more robust handling of complex multi-step tasks.

AI as a Last Resort, Not a First Line

Perhaps most importantly, the accessibility community and technologists are beginning to articulate a clearer vision: accessibility designed in from the start, with agentic AI reserved for the small number of genuinely intractable cases where no amount of good design can fully bridge the gap.

This vision has the right shape. It says: we will build the web so that blind and low-vision users can navigate it independently, with their existing assistive technologies, without needing AI intervention for every task. And for the edge cases—legacy systems that cannot be rebuilt, proprietary enterprise software with decades of accumulated inaccessibility, niche tools that will never attract enough attention to be fixed—we will have capable, transparent, OS-level AI assistants like Guide ready to step in.

Accessibility by design. AI as a safety net. That is a future worth working toward.

The supplementary tools we have today—including Auto Browse, Copilot Actions, and Guide—are imperfect instruments for an imperfect web. They will sometimes help us do things that were previously impossible, and they will sometimes frustrate us by failing at what seems like should be simple. Using them wisely means understanding their limitations and knowing which tool to reach for in which situation.

But the story does not end here. The regulatory momentum, the technical research, and the growing awareness among designers and developers that accessibility is not optional are all pointing in the right direction. A web that is built for everyone, with AI available for the hard cases, is not a utopian fantasy. It is an achievable goal, and we are, slowly, getting closer.

Sources

All sources used in the Blind Access Journal article “Not a Panacea: Why AI Browser Agents Haven’t Solved the Inaccessible Web—and What Comes Next” (March 28, 2026).

Primary research document:
Technical Analysis of Agentic AI Efficacy in Navigating Complex Web Architectures for Accessibility Remediation

AI Browser Agents – Auto Browse and Copilot Actions

Security and Vulnerabilities

Salesforce and Complex Web Architecture

Guide – Specialist Accessibility Application

AI Agent Architecture and Failure Modes

Accessibility Research and Standards

Regulatory and Market Context

Alternative Agentic Tools

The Guitarist Who Teaches the World to Navigate the Screen

There’s a particular kind of patience that only musicians develop — the ability to sit with a student through a hundred failed attempts at a chord, knowing that the hundred-and-first try will produce something beautiful. Tony Gebhard has been cultivating that patience since he was seven years old, when he first picked up a guitar and started to sing.

Decades later, he’s applying it to a very different kind of instruction.

Tony is the creator of NVDA Coach, a free add-on for the NVDA screen reader that is quietly changing how blind people take their first steps into independent digital life. With 35 structured lessons covering everything from basic navigation to reading and customization, the tool is designed for users who aren’t ready — or willing — to sit down with a trainer, wade through a dense manual, or admit they need help.

“Think about someone like Sally,” Tony says, describing a hypothetical but achingly familiar user: a 62-year-old woman who has recently lost her sight and needs to check her bank balance, send an email to her daughter, or order something online. “She doesn’t need a classroom. She needs something patient. Something that won’t rush her.”

That sensitivity to the learner’s emotional experience isn’t accidental. Tony has spent seven years as an assistive technology specialist, working for a non-profit in Anchorage, Alaska, coordinating transitions for youth with disabilities, and eventually landing a role with a state government agency in Oregon. A graduate of the Assistive Technology Instructor Career Training Program from World Services for the Blind in Little Rock, Arkansas, he holds and recommends fellow professionals pursue Certified Assistive Technology Instructional Specialist for Individuals with Visual Impairments (CATIS) certification.

But long before any of those credentials, there was the music.

Tony started writing original songs at age 12. Today, he has published 12 albums — primarily in metal and progressive genres — available on Spotify, Apple Music, and Amazon. His music, like his teaching, reflects an appetite for complexity and a refusal to take the obvious path. Progressive metal is not a genre for the faint-hearted; it rewards listeners who are willing to be challenged, who trust that the dissonance will eventually resolve into something extraordinary.

He brings that same philosophy to assistive technology. NVDA Coach doesn’t dumb things down. It meets users where they are.

The add-on is now available on Tony’s GitHub repository. Version 1.0.0 marked its official debut, and Tony is already planning ahead: version 1.2 or 1.3 will introduce language translations, extending the tool’s reach to users in developing regions around the world where access to screen reader training is scarce or nonexistent.

“This isn’t just about the United States,” Tony says. “There are people everywhere who need this.”

From a seven-year-old learning his first guitar chord in a living room somewhere, to a 62-year-old learning to navigate a screen reader alone in her kitchen — Tony understands that the best teachers don’t just transfer knowledge. They make people believe they can do it themselves.

Now at Version 1.2.0 as of this publication date, NVDA Coach is available for download from Tony’s GitHub page.

Not yet convinced? Play the demo video and see for yourself why the connected, online blind community is all abuzz about NVDA Coach!

Slack Accessibility Bug in Simplified Layout Mode Confirmed Fixed

Great news for screen reader users who rely on Slack’s Simplified Layout Mode: the accessibility regression we reported last month has been resolved.

In our February 14 article, we documented a critical, task-blocking bug introduced in Slack version 4.47.69 in which the Activity tab became completely inaccessible to JAWS users when Simplified Layout Mode was enabled. Focus was trapped on toolbar elements, and the screen reader would report only “Loading” or “Blank” rather than the expected list of notifications, mentions, and direct messages.

We are pleased to confirm that the bug has been fixed.

Confirmation from Slack

On March 12, a member of the Slack team reached out by email with the following update:

“The bug was fixed and pushed to production on the 4th of March. That’s an impressive turnaround time for such detailed and helpful bug reports! Our team also mentioned the screen recording made a huge difference in getting it resolved so quickly.”

We want to express our sincere appreciation to the Slack accessibility and product teams. Throughout this process, Slack worked diligently with members of the online disability community, asking for feedback and providing timely updates. That kind of responsive, collaborative engagement is exactly what the assistive technology community needs from software developers, and it made a real difference.

What We Learned About the Bug

In early March, additional details emerged that shed more light on the root cause. The screen-reader accessibility failure in the Activity tab occurred specifically when two conditions were met simultaneously: Simplified Layout Mode was enabled and the Slack window was not maximized. In our own testing, we were able to confirm this — when Slack was maximized, the bug was not observed.

This is a useful piece of information for anyone who may have been experiencing the issue and wondered why it appeared intermittently. If you had Slack in a smaller or restored window, the bug would surface; if you had it maximized, it would not.

A Reminder for Screen Reader Users

This bug serves as a timely reminder of a best practice that benefits all screen reader users across virtually any application: in general, it is a good idea to keep the currently active and focused window — the one where you are currently working — maximized or in full-screen mode at all times.

Here’s how to keep it maxed.

If you’re not sure whether or not the current window is maximized, follow these steps to make it so:

  1. Press alt+spacebar to open the System Menu for the focused application window. “Restore”, “System Menu” or something similar will be announced.
  2. Press the letter x to maximize the window. The window’s title will be announced and you’re brought right back to the task you were working on before maximizing.

Many accessibility-related rendering and focus issues are window-size dependent, and maximizing your working window can work around an entire category of potential problems before they get in the way of achieving your hopes and dreams. It’s a simple habit that can meaningfully improve your day-to-day experience with any application.

Next Steps

If you are running a version of Slack released on or after March 4, 2026, you should have the fix. If you are still experiencing issues navigating the Activity tab with a screen reader and Simplified Layout Mode enabled, we encourage you to reach out to Slack support right away.

Thank you to everyone in the community who tested, reported, and helped document this issue. Your participation — and especially those screen recordings — made this resolution possible.

Beyond the Screen Reader: Can Gemini’s AI Agent “Accessify” the Web?


AI as an Accessibility Bridge: Testing Gemini’s Auto Browse

For blind and low-vision users, the modern web is a minefield of good intentions gone wrong. Developers build visually polished interfaces — date pickers, multi-step dialogs, dynamic dropdowns — but the underlying code often fails to communicate with assistive technology. Screen readers like JAWS and NVDA rely on semantic structure and proper focus management to guide users through a page. When that structure breaks down, so does access.

That gap is exactly what I set out to probe in a recent demonstration of Auto Browse, an agentic AI feature built into the Gemini for Chrome side panel. My test case was deliberately unglamorous: a Salesforce “Add Work” form on the Trailblazer platform, featuring a date picker that routinely defeats standard keyboard navigation. The question wasn’t whether the interface looked functional. It was whether an AI agent could step in and make it work.

The Problem with Date Pickers (and Why It Matters)

Custom date pickers represent one of the most persistent accessibility failures on the web. Unlike native HTML <input type="date"> elements, which browsers render with built-in keyboard support, custom-built widgets frequently rely on mouse interaction, non-semantic markup, or JavaScript behavior that strips focus away from the user mid-task.

In my demo, the Salesforce dialog presents a “start date” selector with separate Month and Year dropdowns. For a sighted mouse user, this is trivial. For a screen reader user navigating by keyboard, it becomes a trap — the list receives focus but refuses to respond to arrow keys or selection commands, leaving the user stuck with no clear path forward.

This is not a niche problem. Date pickers appear in job applications, medical intake forms, financial dashboards, and e-commerce checkouts. When they break, they don’t just create friction — they create exclusion.

Letting the AI Take the Wheel

My approach was straightforward: rather than fighting the inaccessible interface, I delegated the task entirely. With the Gemini side panel open (activated via Alt+G), I issued a plain-language command: “Please set the start date to December 2004.”

What followed was notable not just for what the AI did, but for how it communicated while doing it. Auto Browse autonomously interacted with the form elements — opening the Year dropdown, scrolling to 2004, selecting it — while simultaneously providing real-time status updates in the side panel. Critically, those updates (“Updating the start year to 2004”) were announced by the screen reader, keeping me informed throughout the process without requiring me to shift focus manually.

A “Take Over Task” button remained visible at the top of the browser at all times, ensuring that AI autonomy didn’t come at the cost of user control — a design principle that will resonate with anyone familiar with WCAG’s emphasis on predictability and user agency.

Where It Still Falls Short

I want to be candid about the rough edges, because that honesty is part of what makes this worth examining closely.

During the interaction, the dialog closed unexpectedly at one point, requiring a page reload before I could restart the task. For sighted users, this is a minor inconvenience. For screen reader users, an unexpected context shift — a dialog closing, focus jumping to an unrelated part of the DOM, a dynamic content update that goes unannounced — can be deeply disorienting. Recovery depends on knowing where you are, and that knowledge is precisely what gets lost.

This points to a fundamental challenge for agentic AI in accessibility contexts: it isn’t enough to complete the task correctly; the AI must also maintain a coherent focus environment throughout. If a script refreshes a page region mid-task, the virtual cursor needs to land somewhere intentional. If a dialog closes, the user needs to know what replaced it. These aren’t edge cases — they’re the everyday texture of dynamic web applications, and they’ll need to be handled reliably before tools like Auto Browse can be genuinely depended upon.

A Glimpse of What’s Possible

Despite those caveats, I came away from this demonstration genuinely encouraged. Gemini successfully populated both fields with the correct date, confirmed by the screen reader’s final readout. More importantly, it did so through natural language — no custom scripts, no manual DOM inspection, no workarounds requiring technical knowledge that most users don’t have and shouldn’t need.

The implications extend well beyond date pickers. Agentic AI that can interpret intent and act on a user’s behalf has the potential to make complex web interfaces navigable for people who have been effectively locked out of them. Not by fixing the underlying code — though that remains the gold standard — but by providing a capable, responsive intermediary that can bridge the gap in real time.

The web has always required remediation to be accessible. What’s new is who, or what, might be doing the remediating.

Visual Descriptions (Alt-Text for Video Keyframes)

To ensure this post is as accessible as the technology it discusses, here are descriptions of the critical visual moments in the video:

Frame 1: The Accessibility Barrier
A screenshot of the Salesforce “Add Work” dialog box. The “Month” and “Year” drop-down menus are highlighted, showing the visual interface that I am unable to navigate using standard screen reader commands.
Frame 2: The Gemini Interface
The Chrome browser split-screen view. On the left is the Trailblazer site; on the right is the Gemini side panel where I have typed my request. The AI is showing a progress spinner labeled “Task started.”
Frame 3: Agentic Interaction
The video shows the “Year” drop-down menu on the webpage opening and scrolling automatically as the Gemini agent selects “2004” without any manual mouse movement or keyboard input from the user.
Frame 4: Success Confirmation
The final state of the form showing “December” and “2004” successfully populated in the fields. The Gemini side panel displays a “Task done” message with a summary of the actions performed.

I am a CPWA-certified digital accessibility specialist. When I’m not testing the latest in AI or keeping up with my family, you can find me on the amateur radio bands under the call sign NU7I.

Using Apple’s Built-In Accessibility Features to Reduce Screen Exposure During Severe Headaches

Summary

Some people experience severe headaches or migraines that make screen use difficult—especially when light sensitivity (photophobia) and flicker or refresh effects are major triggers. While display adjustments can help, there are days when the most effective strategy is to reduce visual reliance as much as possible.

If you use an iPhone and Mac, Apple includes several built-in accessibility tools that can support a “low-screen” or even “no-screen” workflow—particularly for everyday tasks like reading and replying to email.

This article focuses on the built-in Mail app and outlines a practical approach using:
VoiceOver (screen reader),
Voice Control (hands-free voice operation),
and Dictation (speech-to-text composition).


Why VoiceOver and Voice Control can help when light and flicker are triggers

VoiceOver reads on-screen content aloud and provides a structured navigation model that does not require visually scanning the interface. Instead of looking for buttons or reading text, users move through content sequentially and receive spoken feedback.

Voice Control complements this by allowing users to operate their device through spoken commands. Actions such as opening Mail, scrolling, replying, and sending messages can often be completed without touching or looking closely at the screen.

For people whose primary headache triggers include light sensitivity and flicker, combining these tools can significantly reduce both the duration and intensity of screen exposure.


iPhone: Building a low-screen Mail workflow on iOS

Turn on VoiceOver

VoiceOver can be enabled from Settings > Accessibility > VoiceOver. Apple provides a built-in practice experience that introduces the gesture model and basic navigation concepts.

Learn a minimal set of VoiceOver gestures

It is not necessary to learn every gesture. Starting with a small core set allows users to begin working quickly and add complexity later.

  • Swipe right: move to the next item.
  • Swipe left: move to the previous item.
  • Double-tap: activate the selected item.
  • Two-finger swipe up: read the entire screen from the top.
  • Two-finger tap: pause or resume speech.
  • Four-finger tap near the top: jump to the first item.
  • Four-finger tap near the bottom: jump to the last item.

Use Screen Curtain to eliminate display light

When VoiceOver is enabled, the screen itself can be turned off while the device remains fully usable. This feature, called Screen Curtain, allows users to rely entirely on audio output while avoiding light exposure.

  • Three-finger triple-tap: toggle Screen Curtain on or off.
  • If both Zoom and VoiceOver are enabled, a three-finger quadruple-tap may be required.

Adding Voice Control for hands-free interaction

Voice Control allows users to interact with on-screen elements using spoken commands. This can be particularly helpful when precise touch input or visual targeting is uncomfortable.

Common Voice Control commands

  • Open Mail
  • Scroll down / Scroll up
  • Go home
  • Show names (labels interface elements)
  • Show numbers (adds numbered overlays)

When an on-screen control is difficult to activate, VoiceOver can be used to identify the control’s name, and Voice Control can then activate it using that spoken label.


Reading and replying to Mail on iPhone using audio

  1. Open the Mail app using Voice Control or VoiceOver navigation.
  2. Move through the message list using swipe left and swipe right.
  3. Open a message with a double-tap.
  4. Listen to the message using a two-finger swipe up.
  5. Reply using Voice Control or VoiceOver navigation.
  6. Compose the reply using Dictation, speaking punctuation as needed.
  7. Send the message using a spoken command or VoiceOver activation.
  8. Enable Screen Curtain when light sensitivity is a concern.

Mac: Reducing visual load with VoiceOver

On macOS, VoiceOver enables spoken feedback and keyboard-based navigation across apps, including Mail. This allows users to work with less reliance on visual scanning.

Turn VoiceOver on or off

  • Command + F5: toggle VoiceOver.

Core VoiceOver navigation concepts

The VoiceOver cursor moves independently of the system focus and determines what is spoken. Navigation is performed using the VoiceOver modifier keys (often Control + Option).

  • VO + Arrow keys: move between items.

Quick Nav for streamlined navigation

Quick Nav can simplify navigation by allowing arrow keys or single keys to move through content without holding modifier keys. This can be especially useful once the basics feel comfortable.

  • VO + Q: toggle single-key Quick Nav.
  • VO + Shift + Q: toggle arrow-key Quick Nav.

Pacing and learning considerations

When screen exposure can trigger symptoms quickly, it helps to approach learning incrementally.

  • Practice in short sessions (5–10 minutes).
  • Focus first on listening and basic navigation.
  • Add Screen Curtain early if light sensitivity is significant.
  • Introduce Voice Control gradually for common actions.

Sources

When Download Links Aren’t Links: A Critical Accessibility Failure in AI Tools Blind People Depend On

Introduction

Artificial intelligence has the potential to dramatically level the playing field for blind and visually impaired people. Every day, blind professionals use tools like ChatGPT to create and export documents needed for jobs, education, and community participation: resumes, legal forms, code, classroom materials, and more.

But a recent shift in how ChatGPT delivers generated files has created a new accessibility barrier — one that directly harms the very users who could benefit most from the technology.

Not a Feature Gap — a Civil Rights Issue

When sighted users see a clickable download link, blind users encounter only this:

sandbox:/mnt/data/filename.zip

JAWS or NVDA reads it aloud like text.
It doesn’t register as a link.
Pressing Enter does nothing.

The file — often essential content — becomes completely inaccessible.

And the consequences are not theoretical:

  • A blind job seeker can’t download the resume they just generated.
  • A blind accessibility engineer can’t retrieve screenshots or audit reports.
  • A blind student can’t access generated study materials.
  • A blind parent can’t obtain forms needed for family programs.

This is not a mere inconvenience. It is a functional blocker to employment, education, and independence.

A Growing Problem in the Tech Industry

Too often, companies “secure” content at the expense of accessibility — and assume the tradeoff is justified. But security and accessibility must coexist. When they don’t, developers have simply chosen the wrong priorities.

One blind accessibility tester put it directly:

“I’m locked out of my own work. The AI wrote me a document — but I can’t download it.”

Another blind user shared:

“If it’s not accessible from the start, it’s not innovation. It’s segregation.”

The Human Impact of a Missing <a> Tag

What looks like a minor UI oversight is actually a critical, task-blocking WCAG 2.2 conformance failure in at least four different success criteria, including keyboard accessibility and name/role/value semantics.

But beyond compliance…

If a blind user cannot access a file — it does not exist for them.

We should not have to rely on workarounds, Base64 hacks, sighted assistance, or manual extraction to download content we requested and created.

This Is Fixable — Today

The solution is simple: make sure every file intended for download is represented as a real hyperlink:

  • Keyboard-focusable using tab and shift+tab navigation
  • Screen-reader announceable
  • Actionable without a mouse
  • Secure and accessible

This is not a feature enhancement — it is a restoration of equal access.

Blind Users Belong in the Future of AI

OpenAI has expressed a strong commitment to accessibility — and I believe the company will resolve this issue. But this situation reminds us of something bigger:

Accessibility must be built into every step of development — not patched later.

When disabled people ask for accessibility, we are asking for inclusion, dignity, and independence.

We are asking to belong.

Call to Action

  • Developers: Test with JAWS, NVDA, VoiceOver and other assistive technologies before shipping.
  • Accessibility leaders: Add file interaction to automated regression tests.
  • Companies building AI tools: Welcome us in — or risk leaving us behind.
  • Disabled people, friends, relatives and others who care about us: Please reach out to the OpenAI Help Center asking them to fix the current accessibility issue and to publicly recommit to at least WCAG 2.2 conformance as a definition of done that must be achieved before shipping new or updated products.

Blind users contribute, create, and advocate every day.
We deserve access to the results of our own work.

— Written by a blind accessibility professional, community advocate, and lifelong champion of equal access to information and technology.


About the Author

Darrell Hilliker, NU7I, CPWA, Salesforce Certified Platform User Experience Designer, is a Principal Accessibility Test Engineer and publisher of Blind Access Journal. He advocates for equal access to information and technology for blind and visually impaired people worldwide.

Demonstration: Guide Accessifies the Addition of Components to Salesforce Experience Cloud Site Pages

At the intersection of the Salesforce ecosystem and the accessibility community, it has been long known that Experience Builder contains task-blocking accessibility issues that hold many disabled people back from being able to perform important job duties including site administration and content management. While the company continues efforts to improve the accessibility of Experience Builder, disabled administrators, content managers and site developers who rely on keyboard-only navigation and screen readers are finding ways to work around barriers thanks to new tools based on artificial intelligence (AI).

Read more

Uncovering the Accessibility of Tabs in Google Docs

Starting all the way back in April of 2024, Google announced a new tabs feature for Google Docs, providing another way of organizing information in documents similar to that already found in spreadsheets. Soon after that, as the new feature rolled out over the next six months, a support article entitled Use document tabs in Google Docs was posted with all the descriptions and instructions necessary for sighted, non-disabled users to avail themselves of the new capabilities. As blind and other disabled people started to encounter documents containing tabs, we wondered how we would be afforded equitable consideration. It turns out that, in large part, we were considered, even if that fact was not documented. If you’re still reading, then, please stay tuned, as the rest of this article will weave together information from several sources to describe how keyboard-only and screen-reader users can choose, create and rename tabs using keyboard shortcuts and menu selections.

Let’s start with listing the useful keyboard shortcuts, then move in to specific, step-by-step instructions for each significant task.

Please Note: These commands assume that a Windows PC is being used with the latest publicly available version of the Google Chrome browser. They may be slightly different on other browsers and operating systems.

  • Choose the previous tab: control+shift+page up. Note: Though the contents of the newly chosen tab will be available, screen readers cannot announce its label.
  • Choose the next tab: control+shift+page down. Note: Though the contents of the newly chosen tab will be available, screen readers cannot announce its label.
  • Show all available document outlines and tabs in a list: control+alt+a immediately followed by control+alt+h. Note: It is absolutely critical that you either hold down both control and alt while typing a and h, or that you enter each separate command rapidly, as control+alt+h by itself enables and disables Braille support. If you hear “Braille support disabled,” simply press control+alt+h again to turn it back on.
  • Create a new tab: shift+f11. Note: Screen readers will announce “tab added.”

Now that we know the available keyboard shortcuts, let’s dive in to some of the most essential tab management tasks.

Choosing A Tab

There are two ways to choose an existing tab: directly using a single keyboard shortcut or selecting an option from a menu.

Choosing A Tab Using a Keyboard Shortcut

  1. Open a Google Doc that contains two or more tabs.
  2. Press control+shift+page down to move to the next tab after the one currently chosen. Note: Although the contents of the new tab will be available, its name is not provided for screen readers to announce.
  3. Press control+shift+page up to move to the previous tab. Note: Once again, its name is not provided for screen readers to announce.

Using Show Tabs & Outlines to Determine the Current Tab or Choose a Different Tab

Although there’s no way to determine the currently chosen tab using a single keyboard shortcut, there is a way to get this information through a menu, which also represents another way to choose tabs.

Determining the Currently Chosen Tab

  1. Open or create a Google Doc that includes two or more tabs.
  2. Press control+alt+a immediately followed by control+alt+h to open the Tabs & outlines menu. Note: Keep in mind that, if you hear “Braille support disabled,” you will want to press control+alt+h by itself to reenable Braille support.
  3. If the screen reader announces the “Show Tabs & Outlines” button followed by the number of tabs, press enter to open the menu. If this button does not appear first, then you will be immediately taken to the menu.
  4. Press Escape to leave everything alone and stay on the currently chosen tab, or see below for choosing another tab using this menu.

Choosing A Tab Using the Show Tabs & Outlines Menu

  1. Open or create a Google Doc that includes two or more tabs.
  2. Press control+alt+a immediately followed by control+alt+h to open the Tabs & outlines menu. Note: Keep in mind that, if you hear “Braille support disabled,” you will want to press control+alt+h by itself to reenable Braille support.
  3. If the screen reader announces the “Show Tabs & Outlines” button followed by the number of tabs, press enter to open the menu. If this button does not appear first, then you will be immediately taken to the menu.
  4. Press the up arrow and down arrow keys to focus and hear all the available tabs.
  5. Press enter on the tab you wish to choose.

Renaming A Tab

  1. Open or create a Google Doc that includes two or more tabs.
  2. Press control+alt+a immediately followed by control+alt+h to open the Tabs & outlines menu. Note: Keep in mind that, if you hear “Braille support disabled,” you will want to press control+alt+h by itself to reenable Braille support.
  3. If the screen reader announces the “Show Tabs & Outlines” button followed by the number of tabs, press enter to open the menu. If this button does not appear first, then you will be immediately taken to the menu.
  4. Press the up arrow and down arrow keys to focus and hear all the available tabs.
  5. Once you have found the tab you wish to rename, press the tab key to move to the “Tab options” button menu and press the space bar to open it.
  6. Press down arrow until Rename is selected, then press enter to choose this option.
  7. Enter or edit the tab’s name and press enter to make the change.
  8. Press Escape to close the Tab options menu.

Adding A New Tab

When adding a new tab to a document, it is created at the end of the existing tabs regardless of where you are editing. This means that, if a document already has four tabs, a new tab would be labeled “Tab5” which would be the last option in the Show tabs & outlines menu and the last tab visually displayed.

  1. Create or open a Google Doc that has at least one tab defined. In most cases, this will be true of all documents as of the June 2025 date this article was originally published.
  2. Press shift+f11 (as described on a Windows PC running Google Chrome). Observe that the screen reader will announce “tab added” and you will return to the place where you were editing.

There are other features in the Show tabs & outlines and Tab options menus, such as adding, duplicating and deleting tabs, which work in exactly the same way as everything that has already been documented, so they will not be covered in this article.

While there are accessible ways to manage tabs in Google Docs, it would be very nice to see Google documenting them as they have done many other capabilities, including docs and editors themselves. It would also be very nice if they enabled the screen-reader announcement of the currently chosen tab after the control+shift+page up or control+shift+page down commands were pressed. If you agree, please be sure to Contact the Google Disability Support Team to directly request these critical positive changes.

Citations

Please Note: While I am including the accessibility-specific citations for the sake of completeness, they do not document tabs functionality as of the writing of this article in June 2025.