-
-
Notifications
You must be signed in to change notification settings - Fork 662
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Refactor speech to support callbacks, speech commands in control/format fields, priority output, etc. #4877
Comments
Comment 2 by camlorn on 2015-02-04 17:37 |
Comment 3 by leonarddr on 2015-02-06 16:35 |
Comment 4 by camlorn on 2015-02-24 17:53 |
Comment 5 by jteh (in reply to comment 4) on 2015-02-24 22:51
What you're suggesting (sample range tagging, etc.) requires that the speech framework has intimate knowledge of audio output. The biggest problem with this is that not all synths output audio through NVDA, so this actually isn't possible. One approach which should cover at least some of your needs is that we allow a callback to specify that it wants to suspend further utterances until it is complete, at which point it will request that speech resume. One major problem with this is that a broken add-on can very easily break speech quite badly (more easily than it can already), so maybe we need to allow a maximum timeout for this or something. |
Comment 6 by camlorn on 2015-02-25 03:06 |
Comment 7 by jteh on 2015-02-25 05:34 Aside from that, the stuff I'm working on relates to how the speech framework processes utterances, passes them to synths and calls callbacks. The synths still generate the samples after they receive the utterance. Therefore, I don't see how you could do sample tagging at this level anyway. You can have the synth fire callbacks when it outputs an index, which is what I plan to do. |
Comment 8 by camlorn on 2015-02-26 22:08 |
Comment 11 by camlorn on 2015-03-18 13:48 |
Comment 12 by Q (in reply to comment 4) on 2015-04-03 19:41
|
I thought of another possible use case, though I'm not sure how useful it would be: implement an option to monitor things in the background. I'm thinking at least aria live regions and the controller API, but it might also be useful to tag terminals for watching. |
Got one more. On the web, we currently don't read the title attribute when moving by arrows. I think this also applies to aria-label and maybe some other things. With this, it would be possible to indicate these attributes efficiently, so that a user could know to check it in one way or another. As it stands, we would have to say something like "has title" or just read it every time. Since this can be used on things like abbreviations, this is possibly less than useful. There's already an issue about them not reading, but I don't remember which one at the moment. Just that it has a weird title and that I'm on the thread. |
I posted on bugzilla about an issue with NVDA (and JAWS) on Firefox where the 'Remember Password' popup interrupts other aria alerts on the page (https://bugzilla.mozilla.org/show_bug.cgi?id=1323070). It was determined that this was a problem with the screen readers not being able to handle multiple alerts at the same time. Would the changes proposed here fix this issue? |
It wouldn't, I don't think. It would allow cool experimental things like saying them all at the same time in different voices and maybe panning them across the sound field or something, but the core issue is probably about queuing things to be spoken, and we should already be able to do that. |
List of tickets which may be related to, blocked by or dependent on this ticket: |
@michaelDCurran I propose to let this issue open because here are many issues referenced which contain use cases for this feature. After solving and closing them, I propose to close this as well. |
I've looked at the initial description of this issue, and I think it is fully covered by #7599. Leaving this issue open until all the mentioned issues are solved as @Adriani90 suggested, will probably mean that this will be open for ever. Furthermore, there's nothing holding you back from referring to this issue to find the use cases for the framework grouped together. Closing it. |
Reported by jteh on 2015-02-03 00:56
A lot of potentially nice functionality is not possible (or is at least ridiculously painful) with our current speech framework. Speech needs a pretty big refactor to allow for such things. It needs to support:
Unfortunately, some of this is going to break backwards compatibility, but I think it's worth it in this case.
Blocking #905, #3188, #3286, #3736, #4089, #4233, #4433, #4874, #4966, #5026, #5104
The text was updated successfully, but these errors were encountered: