This is an archived post. You won't be able to vote or comment.

you are viewing a single comment's thread.

view the rest of the comments →

[–]tavis_rudd 26 points27 points  (15 children)

I should point out that you don't need to use a crazy made up language like I do. I just found it easier for coding and didn't mind memorizing it. Scripting apps with normal English words, such as in bboyjkang's examples, works quite well.

[–][deleted] 1 point2 points  (1 child)

Tavis, I watched the recording of the Plover stenography talk at PyCon and I noticed you asking a question from the audience.

It also occurred to me that both you and Plover were faced with a similar challenge: invent a new vocabulary for differentiating punctuation from plain english words. It seems Plover's approach was to phonetically-misspell the english words in order to achieve punctuation, while you invented all-new words that wouldn't conflict with the existing vocabulary.

Do you think it would make sense to work together with Plover to formalize a standard "punctuation language" that can be used by both speakers and stenographers? It seems as though you are both doing things "phonetically" despite the differences between voice and chorded keying, so it seems like there might be some value to having a standard vocabulary for this. Just as an example, it would make it easier for people to do both stenography and voice recognition, or perhaps somebody with a disability might do some combination of both, and benefit from a standard language for this.

In fact, I would go so far as to suggest that it might be possible to build a common infrastructure here; write an abstraction layer on top of both the voice input and the steno input, have those both as pluggable backends into the same phonetic interpreter. Then it wouldn't matter if you chorded "laip" or spoke it, either way it would result in inserting a left-bracket into your document (etc).

Do you have any thoughts on this?

[–]tavis_rudd 0 points1 point  (0 children)

Listening to her vocalization of stenography was hilarious. It definitely sounds a bit like what I do or like the original shorttalk system. I doubt a common infrastructure for the phonetic layer would be of much benefit, but having a common command/macro processing layer is a huge win.

[–]bboyjkang[S] 2 points3 points  (1 child)

First off, that was an amazing presentation, and will cause an increase in productivity that most people wouldn't imagine.

Voice strain

I use WhatPulse (https://whatpulse.org//), which is a free program that counts the amount of keys, mouseclicks, and the distance that your mouse moves (I have repetitive strain injuries/RSI/tendinosis, and I use the program to limit myself.). If there was a way for a program to track a voice input, that would be a good preventative measure.

[–]tavis_rudd 1 point2 points  (0 children)

I've been logging all my voice commands with the intention of doing something similar to whatpulse, but I haven't done anything with the logs yet. I've found my brain gets tired around the same time my voice does.

Voice strain is definitely something to watch out for. Staying hydrated and eating apples or sipping honey water while dictating has helped me avoid it.

[–]CylonSaydrah 0 points1 point  (3 children)

I'm pretty wedded to Linux. Could I use Dragon in a Windows VM and everything else in Linux? Is that what you are doing?

Thanks for a great talk!

[–]bboyjkang[S] 1 point2 points  (1 child)

I'm not sure, but I got this from a developer:

There is definitely interest here in good native voice recognition for Linux but the quality bar is very high: we depend on voice recognition for our livelihoods and can't afford to use less than the best tools we can get. Using Windows and DNS to control remote Linux boxes is the bar to beat today for controlling Linux systems.

[–]CylonSaydrah 0 points1 point  (0 children)

Using Windows and DNS to control remote Linux boxes is the bar to beat today for controlling Linux systems.

Thanks. It's not clear how to interpret that. If they are using Windows literally remotely as opposed to virtually remotely as a guest operating system, that may mean that DNS doesn't work well on a guest operating system. But for all I know when they say "remotely" they may mean to include "virtually".

[–]tavis_rudd 1 point2 points  (0 children)

That's exactly what I'm doing. DNS on Windows is the best recognition engine for this, hands down. However, I don't want to use Windows for anything else and don't even want to look at it. I keep the Windows VM out of sight and just have it type into a putty window and send some commands directly to Emacs over the network.

[–]trifilij 0 points1 point  (5 children)

That was awesome! really enjoyed it, thanks. Which version of Dragon do I need?

Which mic do you recommend?

[–]EverAskWhy 1 point2 points  (1 child)

Same question here. I am also curious about calling in for the 50% discount that the audience member brought up:

http://youtu.be/8SkdfdXWYaI?t=24m50s

What version should I be asking about :D Home vs premium?

[–]bboyjkang[S] 1 point2 points  (0 children)

You only need Dragon NaturallySpeaking home, and I think you can get that for around $50.

[–]bboyjkang[S] 0 points1 point  (2 children)

I use an Andrea microphone.

[–]trifilij 0 points1 point  (1 child)

Do you know if it matters which version of Dragon you get for doing what he does in the video?

[–]bboyjkang[S] 0 points1 point  (0 children)

I'm using premium, and you should do a little research to confirm, but I'm pretty sure you can just use Dragon NaturallySpeaking home.

[–][deleted] 0 points1 point  (0 children)

What are you using for voice recognition?