KnowBrainer Speech Recognition
Decrease font size
Increase font size
Topic Title: Speech command set design for Markup Languages
Topic Summary: Like Markdown, AsciiDoc, MediaWiki, ...
Created On: 03/16/2021 02:14 PM
Status: Post and Reply
Linear : Threading : Single : Branch
 Speech command set design for Markup Languages   - Ag - 03/16/2021 02:14 PM  
 Speech command set design for Markup Languages   - kkkwj - 03/16/2021 10:33 PM  
 Speech command set design for Markup Languages   - Ag - 03/16/2021 10:51 PM  
 Speech command set design for Markup Languages   - Ag - 03/16/2021 10:59 PM  
 Speech command set design for Markup Languages   - wristofdoom - 03/18/2021 12:37 PM  
 Speech command set design for Markup Languages   - kkkwj - 03/18/2021 06:30 PM  
 Speech command set design for Markup Languages   - benTalks - 12/22/2021 08:18 AM  
 Speech command set design for Markup Languages   - wristofdoom - 12/22/2021 11:53 AM  
 Speech command set design for Markup Languages   - dilligence - 12/22/2021 12:07 PM  
 Speech command set design for Markup Languages   - benTalks - 12/22/2021 08:19 AM  
 Speech command set design for Markup Languages   - kkkwj - 12/23/2021 10:30 PM  
Keyword
 03/16/2021 02:14 PM
User is offline View Users Profile Print this message

Author Icon
Ag
Top-Tier Member

Posts: 775
Joined: 07/08/2019

---+ BRIEF: Does anyone have a set of speech commands/shortcuts for markup languages?

 

I am most interested in the design principles. It seems sensible to have similar speech command set patterns for different markup languages.  The actual commands are fairly straightforward.

In particular, I frequently switch between different markup languages. E.g. on GitHub I will use all of AsciiDoc and MediaWiki and Markdown - and there are a few other markup languages supported by the GitHub wiki. Sometimes different pages in the same wiki are written in different markup languages. I have not yet figured out a good way of inferring Or controlling which markup to use from my speech commands.

 

---+ DETAIL: 

 

Sigh. Every few weeks I need to start adding a new class of applications for speech commands.

 

Today: markup languages like Markdown, AsciiDoc, MediaWiki, ...

 

My job requires me to maintain documents in AsciiDoc, and to interact with people on wikis using markup languages like Markdown, MediaWiki, and a few others.  Sometimes on GitHub, sometimes elsewhere.

 

One quickly gets tired of saying "equal sign; equal sign; Heading #2; equal sign; equal sign" in MediaWiki, and "sharp sign; sharp sign; Heading #2" in Markdown.

 

Obviously, I will have use a set of commands like "Heading <Level> <Dictation>", Similar syntax across markup languages, different implementations. 

 

One part of my question is asking about patterns for the syntax of such a common Markdown speech command set.

 

A second part of my question is how to implement inferring or controlling the class of Markdown.   Application and window/webpage name do not help.  I believe I need persistence.

 

---+ Inferring and controlling markup language to use

So far the main problem is detecting when to use which Markdown language.  

 

I and my coworker/collaborators frequently switch between different markup languages. E.g. on GitHub I will use all of AsciiDoc and MediaWiki and Markdown - and there are a few other markup languages supported by the GitHub wiki. Sometimes different pages in the same wiki are written in different markup languages.

 

I have not yet figured out a good way of inferring which markup language should be used. On GitHub, all of the markups use the same webpage text box editing system. Though I cannot use application name or window name to figure out which command set is appropriate.

 

As I am writing this, I realize one thing that might help: I can make a local clone of the GitHub wiki and repo, and edit the files locally in my favorite text editor. Emacs, in my case. Editing locally the markup language suffix like .md or .asciidoc or .adoc is already known.   There are already emacs modes for many of these markup languages. I've already begun generic interfacing of speech commands to emacs commands. In fact, overall, could leave most of the intelligence in emacs, leaving it out of the speech commands.

 

But I would still like to be able to do quick edits on the fly using the webpage editor for wikis and other markup systems

 

If I cannot use application or window name context and the speech commands, perhaps I need persistence - a command that says "use MediaWiki syntax from here on".   But that requires persistent state.

 

I have asked about persistent state before on this forum.  Edgar's recommendation is to use files. Which I think will amount to opening a file every time I bounce to a new wiki webpage editing tab or window.  I have not tried it yet, but that sounds expensive.

 

Mostly the sort of persistence that I need is "state that persists from one speech command to another".  Probably not across reboot or restart of the threads that are maintaining the persistence. 

 

I already have an unsatisfactory implementation of persistence, sending hotkeys to a persistent AHK script.  

 

I have seen AHK scripts that send text messages to each other using Windows messaging interprocess communication APIs.  It should be straightforward to have a transient speech command, whether in AHK or Dragon/KnowBrainer basic, send such a message.

 

More generically, using standard networking APIs like Berkeley sockets.  This would have the advantage that it could communicate from speech commands on Windows to demand servers in different operating system environments like Linux, whether in different virtual machines on the same PC, or across the network. With my usual paranoia about security.

 

Does anyone have better or alternative ideas?  Here, example code is always appreciated

 

 



-------------------------

DPG15.6 (also DPI 15.3) + KB, Sennheiser MB Pro 1 UC ML, BTD 800 dongle, Windows 10 Pro, MS Surface Book 3, Intel Core i7-1065G7 CPU @ 1.3/1.5GHz (4 cores, 8 logical, GPU=NVIDIA Quadro RTX 3000 with Max-Q Design.

Statistics
32286 users are registered to the KnowBrainer Speech Recognition forum.
There are currently 3 users logged in.
The most users ever online was 12124 on 09/09/2020 at 04:59 AM.
There are currently 523 guests browsing this forum, which makes a total of 526 users using this forum.

FuseTalk Standard Edition v4.0 - © 1999-2022 FuseTalk™ Inc. All rights reserved.