Learn how to integrate a voice recognition into your VR/AR application
Step 1: Setup. Configure the Phonal Conversation
Step 2: Launch. Edit and Run Actions or Conversations
Step 3: Empower. Add Contexts to Phonal Conversation
To set things up, please add to your Scene these prefabs:
Copy Assets/PhonalTechnologies/StreamingAssets folder to root of your Assets folder.
If you already have StreamingAssets folder inside your Assets folder - just copy content from Assets/PhonalTechnologies/StreamingAssets to Assets/StreamingAssets
Speech recognition module that listens for a user to talk, detects speech, and pass speech recognition results to the dialogue system.
Add Google Cloud Speech API key inside to enable recognition.
If you don’t have API Key, you can get it from
A module to manage dialogues and their output to UI.
You can have as many managers as UIs to output them.
A basic dialogue entity.
You can have as many Dialogues as you want to. Consists of Intents. Stores all the dialogue flows. Edit dialogue in Dialogue Editor to add new Intents and routes.
After all the prefabs are set up, configure UI to output the results.
Create UI elements with the ability to display text and assign a special script to them:
This script creates an interface for Phonal Conversation to set the content inside. You can extend this class to output things wherever you want.
After UI elements are configured, add them to DialogueManager for Phonal Conversation to know where to display the output.
Drag and drop elements with assigned InteractableUi script to related DialogueManager fields.
Please refer the sample scene to see how all things set up:
GUI to create, edit and configure conversations.
Within Dialogue Editor you can:
Add input samples to intents
Add output samples to intents
Connect intents to make a complex dialogue tree
Add suggestion texts