Is there a way to invoke google assistant by executing a command (from a linux console) instead of saying "Hey google"?
I found a way, using nodejs, to send text to my home google mini to reproduce text, but I didn't find the way to enable the listening of it how I said "Hey google".
Related
In the Google Cloud Platform Console, upon clicking "Activate Cloud Shell," the bar for the command line shows up and displays "Provisioning your Google Cloud Shell Machine" and "Establishing your connection to Google Cloud Shell," but after that the entire box for the command line goes blank. It still has the option bar that displays the project name, option to open the text editor, etc. I tried going to the text editor hoping that the command line would show itself there, but the whole box remained blank.
Even worse, I started one of my virtual machines and used the SSH option from the console to connect to it, and the SSH window similarly went totally blank. Any ideas what's going wrong here? I'm using Firefox on Windows 10 and haven't had any issues using the Cloud Shell for months.
This was a browser compatibility problem between FF64 and the Google Cloud Shell terminal emulator library. The SSH-in-the browser feature inside Cloud Console is using the same terminal emulator, so it was affected as well. Both have now been fixed.
I had a similar issue. Google cloud shell doesn't work with Firefox.
I waited for a whole thirty minutes and nothing happened. Thirty! I switched to Chrome and the issue was sorted out immediately.
I created a Dialogflow project and have tested via Integration->Web Demo and Google Actions->Simulator; everything works correctly. I now want to test it on the actual device. My account is enabled for testing on the device - see screenshot. However, I can't locate it when I open the Google Assistant app on my iPhone search for it, I don't find it. What else do I have do to get in on my device? Thanks for the help!
Many Thanks
It won't show up in the Assistant Directory until you publish it, however you can start it the same way you start it in the simulator - by saying "Hey Google, talk to action name"
I'm looking for the best approach to implement voice command in a Xamarin app.
Here are my requirements:
I don't need to launch my app by voice. Instead, my users will launch the app through touch (so, when the app is not running, no voice recognition is needed by my app)
My app is a client/server app and it will work always on (the backend will run on azure)
My app will be used primarily by car (so consider environment noise)
My app will work in many languages, such as Italian, Spanish, French and English
My app should be developed with xamarin (and eventually mvvmcross or similar)
In my app there will be two kinds of voice commands:
to select an item from a short list: app will show a list of items, such as "apple, kiwi, banana and strawberry" and user will have to say one of those words.
to change current view. Typically these voice commands will be something like "cancel", "confirm", "more" and stuff like these
The typical interaction between user, app and server should be this:
user says one of the available commands in current view/activity/page
suppose here that the user perfectly knows which commands he/she can use, it does no matter now how he/she knows these commands (he/she just knows them)
user could put before the commands some special words, such as "hey 'appname'", to have a command like "hey 'appname', confirm"
Note: the "hey 'appname'" part of the voice command has the only purpose to allow the app to know when the command starts. The app can be always in listening mode, but has to avoid to send the audio stream continuously to the server to recognize commands
best case is if app would recognize these commands locally, without involve the remote server, since the voice commands are predefined and well-known in each view. Anyway, app can send the audio wave to the server which will return a string (in this example the text returned will be "confirm", since the audio was "hey 'appname', confirm")
app will map the text recognized with the available commands, and will invoke the right one
user will receive a feedback by the app. The feedback could be:
voice feedback (text-to-speech)
visual feedback (something on the screen)
both above
I was looking for azure-cognitive-services, but in this case, as far as I've understood, there is no way to recognize the start of the command locally (everything works on server side through REST api or clients). So the user would have to press a button before every voice command, and I need to avoid this kind of interaction.
Since the app is running, my user has him/her hands on the steering wheel, and he/she can't touch everytime the display. Isn't it?
Moreover, I was looking for cortana-skills-kit and botframework, but:
It seems that Cortana Skills are available in English only
Actually, I don't need to involve Cortana to launch my app
I don't have experiences on these topics, so, hope that my question is clear and, generally speaking, that can be useful for other newbie users as well.
* UPDATE 1 *
The Speech Recognition with the Voice Command Definition (VCD) file is really close to what I'd need, because:
it has a way to activate the command through a command name shortcut
It works in foreground (and background as well, even if in my case I don't need the background)
Unfortunately, this service works only on Windows, since it uses the local API. Maybe the right approach could be based on the following considerations:
Every platform exposes a local speech recognition api (Cortana, Siri, Google Now)
Xamarin exposes Siri and Google Now apis and make them available through C#
It would be useful to create a facade component to expose the three different local speech api through a common interface
I'm wondering if there is something other solution to this. Cortana, as personal assistant, is available on Windows, iOS and Android. Since Cortana works both with local api and with remote service (Cortana Skills), is Cortana the right approach? Has Cortana the support for many languages (or, at least, has the support a road map)?
So, just some thoughts here. If you have some other ideas, or suggestions, please add here. Thanks
Question is simple but answer is not present anywhere!
Is there possibility to use Cortana to start my app with Command "Hey MyApp" instead of "Hey Cortana"?
I don’t mean to run Cortana then say run my app via voice command.
Than you for any information.
No that is not possible. "Hey Cortana" is the only thing that triggers the voice recognition, this is built into the Windows core.
If they would not have done that it would cause a lot of processing power to constantly having to listen and evaluate what one is saying. And on the other hand you could easily trigger all kind of actions and apps by just having your phone next to you and talking to another person.
Therefore, the start point for any voice command will be "Hey Cortana" on Windows, "Ok Google" on Android and "Hey Siri" on iOS.
You can implement certain 'skills' into Cortana, with which you can trigger actions within your app.
To get started, head over to the Cortana Dev Center.
I code in REALbasic.
Whenever I receive an IM via iChat, iChat runs an applescript and sends to it the last message received. I need to somehow forward this message as a string to the app I'm writing in REALbasic. REALbasic is able to respond to AppleEvents but I don't now how to implement this or if I'm even in the right ball park.
Any advice?
Unfortunately I haven´t used RB in ages, but maybe the following link (just a quick google) might be a starting point?
http://www.106automation.com/page2/RB/RB.html
(the link to download the example files is actually the title of the page, so click "REALBasic & Applescript")