Home Automation powered by Openwhisk, Raspberry PI, Twilio, and IBM Watson

Kalonji Bankole
Kalonji Bankole
Published in
10 min readApr 7, 2017

Over the past few years, we’ve seen a significant rise in popularity for intelligent personal assistants, such as Apple’s Siri, Amazon Alexa, and Google Assistant. Though they initially appeared to be little more than a novelty, they’ve evolved to become rather useful as a convenient interface to interact with service APIs and IoT connected devices.

This series will guide users through setting up their own starter home automation hub using a Raspberry PI. This first blog post provides a step by step tutorial to create a RF circuit that’ll enable the Raspberry PI to turn power outlets off and on. Once the circuit and software dependencies are installed and configured properly, users will also be able to leverage Watson’s language services to control the power outlets via voice and/or text commands. Furthermore, we’ll show how Openwhisk serverless functions can be leveraged to trigger these sockets based on a timed schedule, changes to the weather, motion sensors being activated, etc. We’ll assume that the reader has a basic understanding of Linux and electronic circuits.

This tutorial requires the following components:

Install software dependencies

Login to Raspberry PI and install prerequisites for the wiringPi library. This library enables applications to read/control the Raspberry Pi’s GPIO pins.

sudo apt-get -y updatesudo apt-get -y install git-coregit clone git://git.drogon.net/wiringPigit pull origin master./wiringPi/build

Ensure the wiringPi library is installed properly by running the following command.

gpio readall
Pi GPIO output

Next, install 433Utils, which will call the wiringPi library to transmit and receive messages via the 433MHz frequency. In our case, each outlet has a unique RF code to turn power on and off. We’ll use one of the wiringPi utilities, titled “RFSniffer” to essentially register each of these unique codes. The 433MHz frequency is standard among many common devices such as garage door openers, thermostats, window/door sensors, car keys, etc. So this initial setup is not limited to only controlling power outlets. This library can be installed by running the following commands on the Raspberry Pi.

sudo apt-get install build-essentialgit clone git://github.com/ninjablocks/433Utils.gitcd 433Utils/RPi_utilsmake

Setup RF Circuit

Arrange the hardware components to complete the following circuit.

Now we will determine which RF codes correspond with the Etekcity outlets. Start by running

sudo /var/www/rfoutlet/RFSniffer

This will listen on the RF receiver for incoming signals, and write them to stdout. As the on/off buttons are pressed on the Etekcity remote, the Raspberry Pi should show the following output if the circuit is wired correctly.

pi@raspberrypi:~ $ sudo /var/www/rfoutlet/RFSnifferReceived 5528835Received pulse 190Received 5528844Received pulse 191

After determining the on/off signal for the RF sockets, place the captured signals into the /etc/environment file like so.

RF_PLUG_ON_1=5528835
RF_PLUG_ON_PULSE_1=190
RF_PLUG_OFF_1=5528844
RF_PLUG_OFF_PULSE_1=191

Now, plug in the associated socket, and run the following command to ensure the Raspberry Pi can turn the socket on and off. This command simply sends the RF code at the requested pulse length, which is to be provided as the -l parameter.

/var/www/rfoutlet/codesend ${RF_PLUG_ON_1} -l ${RF_PLUG_ON_PULSE_1}/var/www/rfoutlet/codesend ${RF_PLUG_OFF_1} -l ${RF_PLUG_OFF_PULSE_1}

Now that we can control the sockets manually via cli, we’ll move forward and experiment with different ways to control them in an automated fashion. Rather than writing and executing pipelines and complex automation logic on the Raspberry Pi, we’ll utilize a serverless, event driven platform called Openwhisk. In this implementation, Openwhisk actions communicate with the Raspberry Pi via MQTT messages.

A Bluemix account is required to set up Openwhisk and the accompanying Watson services.

Provision Watson Services

Moving forward, login to Bluemix and provision the following services

  • Speech To Text — This service transcribes voice input from the Raspberry Pi microphone
  • Conversation — This service analyzes the transcribed text and determines the intent behind the user’s command. Each execution results in an intent, which determines what the user would like the system to do, and a set of entities, which provide context for the action (device type, location, etc). The included training data is set to interpret the requested state (on or off) and device type (TV, light)
  • Watson IoT Platform — This serves as a MQTT broker. The credentials to connect to the broker can be generated by following these instructions. Once the credentials have been generated, set the following values in the /etc/environment file; IOT_ORG, IOT_API_KEY, IOT_AUTH_TOKEN, IOT_DEVICE_TYPE.

Openwhisk

Serverless architectures are taking Cloud and IoT industries by storm for several reasons. First, these solutions allow for developers/makers to offload operational tasks such as networking, server configuration and management, disk failure, etc. Second, these architectures provide a very effective billing model where functions are executed only on demand, and users are only charged for the total amount of time their code is running down to the millisecond. These benefits allow developers to avoid being weighed down with operational tasks and focus purely on their app code and business logic.

If you’re unfamiliar with Openwhisk, it might be best to run through the introductory documentation, which is available here.

To get started, login to Bluemix.

Select the menu icon at the upper left corner, and navigate to the “Openwhisk” section.

Openwhisk dashboard

Follow the instructions to download and install the Openwhisk CLI. Run wsk action list to ensure your credentials and api endpoint are set properly

Continue on by cloning the home automation github repository. This repository contains a set of Openwhisk actions and training data for the Watson Conversation service

git clone github.com/IBM/serverless-home-automation /opt/

Openwhisk allows for multiple code snippets, or “Actions” to be chained together as a sequence. To get started, we will create a sequence that consists of three actions. The first action will transcribe an audio payload to text. The second action will analyze the transcribed text result using Watson’s Conversation service. This analysis will extract the intent behind the spoken message, and determine what the user would like the Raspberry Pi to do. So, for example, if the user says something along the line of “Turn on the light” or “Flip the switch”, the Conversation service will be able to interpret that. Finally, the third action will send a MQTT message that’ll notify the Raspberry Pi to switch the socket on/off.

Architecture flow

  1. User says a command into the microphone, or sends a text to the Twilio SMS number
  2. User input is captured and embedded in an HTTP POST request triggering an IBM Cloud Functions sequence
  3. The first IBM Cloud Functions action in the sequence forwards the audio to Speech to Text service, and waits for the response
  4. Transcription is forwarded to the second IBM Cloud Functions action
  5. IBM Cloud Functions action 2 calls the Conversation service to analyze the user’s text input, again waits for the response
  6. Conversation service result is forwarded to final IBM Cloud Functions action
  7. Final IBM Cloud Functions action publishes a entity/intent pair (fan/turnon for example) to the IoT MQTT broker
  8. MQTT client subscribed on Raspberry Pi receives and interprets result
  9. Raspberry Pi transmits corresponding RF signal to adjust outlet state

The speech to text action is already built in to Openwhisk as a public package, so we’ll just need to supply our credentials for that service. Moving forward, we can create actions to call the Conversation and Watson IoT services with the following commands.

cd /opt/serverless-home-automation/iot_gateway/whisk_actionswsk action create conversation conversation.jswsk action create iot-pub.py iot-pub.py

The additional actions are simply creating a JS promise that makes a request to a given service and returns the results when the service call is complete. For example, here is a snippet of the action responsible for calling the conversation service

Once the actions are successfully created, we can set default service credentials for each of the actions. Otherwise we’d have to pass in the service credentials every time we’d like our actions to call the Watson services. To obtain these credentials, click each provisioned service in the Bluemix dashboard, and then select the “View credentials” dropdown.

Then insert the corresponding credentials when running the commands below.

wsk action update conversation -p username ${conversation_username} -p password ${conversation_password} -p workspace_id ${conversation_workspace_id}wsk action update iot-pub -p org ${iot_org_id} -p device_id ${device_id} -p api_token ${api_token}wsk package bind /whisk.system/watson-speechToText myWatsonSpeechToText -p username ${stt_username} -p password ${stt_password}

Next, we can arrange the actions into a sequence

wsk action create homeSequence --sequence /myWatsonSpeechToText/speechToText,conversation,iot-pub

For the sequence to be able to return the result to the Raspberry Pi, a MQTT client will need to be listening to the Watson IoT service. If the proper values have been set in the /etc/environment file, you should just have to run the following commands to create and enable a systemd service, which will automatically start on boot.

sudo cp /opt/serverless-home-automation/iot-gateway/node-mqtt.service /etc/systemd/system/sudo systemctl enable node-mqttsudo systemctl start node-mqttsudo systemctl status node-mqtt

To test the sequence, plugin the USB microphone to the Raspberry Pi, and run rec sample.wav . Record a command, saying something like “Turn on the light” or “Turn off the socket” . Then use the commands below to write the audio file’s binary output to a json file, and execute the sequence.

echo "{\"payload\": \"$(cat sample.wav | base64)\"}" > parameters.jsonwsk action invoke homeSequence --blocking --result --params-file parameters.json

Finally, we can setup a hotword using the Snowboy detection toolkit. A “hotword” allows a device to listen passively in the background, and “wake up” once a specific phrase is detected. So in this case, when the hotword is spoken by the user, the Raspberry Pi will begin recording a voice command, and forward the audio to the Openwhisk sequence when the recording is complete.

Twilio

As an alternative to using a microphone, we can also control the device outlets using a phone by leveraging Twilio, which is a communications platform that enables developers to integrate SMS and VoIP capabilities into their application. Texts or Phone calls can be made from a registered phone number via a simple HTTP call or library client like so

#!/usr/bin/python
from twilio.rest import Client
client = Client(account_sid, auth_token)
# Send text message
message = client.messages.create(
to="+15558675309",
from_="+15017250604",
body="Hello from Python!"
)
# Make phone call
# 'url' points to a XML document that dictates what will be said on the outbound call, how to respond to user input, etc. docs
call = client.calls.create(
to="+14155551212",
from_="+15017250604",
url="http://demo.twilio.com/docs/voice.xml"
)

In addition to making outbound calls and texts, we can configure the Twilio platform to take action in response to incoming calls and texts. So in our case, we’d like to be able to text something like “turn on the light” or “switch off the fan”, and have the message contents be forwarded to an Openwhisk sequence. In our case this can be done by configuring our active Twilio number’s “Messaging” settings to react to incoming SMS messages by triggering the webhook associated with our Openwhisk sequence.

The exact value for the webhook url can be found in the Openwhisk “Develop” dashboard by selecting the “View Action Details” button and then the “Enable as Web Action” checkbox.

Now that we have our Twilio number configured to trigger the Openwhisk sequence, we’ll need to update our Openwhisk action to extract and forward the relevant SMS information (sender number, message body, time) to our services. As we can see in the Twilio webhook documentation, the information from the incoming text message is forwarded as a JSON response. We can see the contents of the incoming JSON body by creating a action titled “printparams” which simply prints all parameters forwarded from the webhook request.

So here we see that the Twilio platform received a text titled “Turn on the fan” from a phone number ending with 7799. The text message value can be accessed via the “params.Body” variable, so we’ll simply add an “or” statement to use the text value if set, and short-circuit to the “params.data” value if not.

Also, since a text message can be sent from any phone, the “params.From” value can be used to add some level of security. This can possibly be achieved by adjusting the action to only respond to certain phone numbers.

References:

--

--

Kalonji Bankole
Kalonji Bankole

Kalonji Bankole is a developer advocate for IBMs emerging technology team. Day to day, he works with open technologies such as Ansible, MQTT, and Openwhisk