How to use a service
The social interaction cloud has many components available for you to speed up creating social interactions with the robots.
Service |
Command |
Source |
Demo files |
Install |
Notes |
|---|---|---|---|---|---|
Dialogflow for creating conversational agents using google’s framework. This provides a flow chart like dialog management and speech recognition. |
|
demo_desktop_microphone_dialogflow.py or demo_nao_dialogflow |
|
||
Face detection using OpenCV’s cascading classifier, which is very fast and can run on a laptop CPU |
|
None, no extra dependencies are needed |
|||
DNN Face detection using a YOLOv7 neural network for accurate detection, and detection of small faces. |
|
|
The model file used in this example can be found here: yolov7-face.pt |
||
DNN Face recognition using a resnet50 network to extract face embeddings and assign an id based on automatic clustering. |
|
|
The cascade classifier file used in this example can be found here: haarcascade_frontalface_default.xml. The model file can be found here: resnet50_ft_weight.pt |
||
OpenAI ChatGPT a text based large language model that provides a very capable dialog agent. Requires a credit card. |
``run-gpt `` |
|
An openai api key can be created here: https://platform.openai.com/api-keys |
||
OpenAI Whisper a powerful speech to text model, capable of running both local and in the cloud. Cloud usage requires a credit card. Start and end recognition is performed using python’s SpeechRecognition |
|
|
An openai api key can be created here: https://platform.openai.com/api-keys |
||
Google Text to speech using google cloud API. Requires a credit card. |
|
|
A credential keyfile has to be configured: See https://console.cloud.google.com/apis/api/texttospeech.googleapis.com/. A credit card is required. |
||
Natural language understanding (NLU), a joint learning model of intent and slot classification with BERT. |
a simple demo with ASR+NLU pipeline demo_desktop_asr_nlu.py. |
git clone & checkout the development branch nlu_component |
|||
LLM, A CLI utility and Python library for interacting with Large Language Models, both via remote APIs and models that can be installed and run on your own local machine. llm |
git clone & checkout the development branch nlu_component |
You can use both free local LLMs and remote LLMs with your own API keys. |
|||
Templates for creating your own components |