Recognition engines aren't going to be able to discern echoes. If you were calling an airline, and said 'Boston,' the phone couldn't send the signal three times. The server would hear 'Boston Boston Boston,' and think it's a new word every time.
We felt that last time there was a lot of misunderstanding. Each side was claiming one thing and the other side another thing. We felt that if it was open, then everything would be clear. We would not get into squabbles about who said what.
Everybody does their best at what they can do, and we try to accommodate whatever skills they do have. What they bring us, we improve as much as we can, whether they're quite skilled or it's the first time they've ever seen a round ball.
A typical database might contain 100,000 words, maybe 50 or 60M bytes. I know of some twice as large ... people aren't going to wait 30 seconds or a minute to download a vocabulary database when they can just call up the agency and ask the question.
If you have a device with our technology talking to a server with our technology, the device sends a much cleaner signal to the server. Your mouth is so much closer to the microphone. The signal is clearer.
Certainly, the loss of revenue for clubs and other groups is an issue that we're going to be concerned about. But do we do the right thing, or do we hide it under the guise of economics?