How to use Speech Recognition inside the iOS SDK? [closed]

I know that there is no public API for the SIRI-Services, but is there an API for simple Speech-Recognition? So if I have a textfield and the user taps onto that textfield, a keyboard with the typically microphone button appears and if he pressed it the speech get recognized and transformed into a string object? Or is this button maybe presented by default?


Solution 1:

There are many libraries availble. You can use any of them.

  1. openears // This is the best library

  2. VocalKit (Deprecated for open ears)

  3. TTS
  4. ispeech (Not free)

Hope it helps you.

NOTE:

if you download openears (which contains a sample project called "OpenEarsSampleApp") @efimovD mentions this

Check the code in view controller and you will see an array with possible commands. This thing detects commands! Not some talk. It listens and tries to compare what you've said with the words from array

Solution 2:

You can use the MindMeld iOS SDK provided by Expect Labs. It supports continuous (streaming) speech recognition on iOS. It also supports natural language understanding and content discovery for any custom content domain.

Try it out at: https://expectlabs.com/docs/sdks/ios/gettingStarted