node-red-contrib-spark 2.0.5

Node-RED Nodes to integrate with the Cisco Spark API

npm install node-red-contrib-spark

Node-RED Nodes to integrate with the Cisco Spark API.

Version 2.0.0 (changelog)

Nodes

  • api - This Node receives queries on the input that are then sent to the Spark API. The results are provided at the output.
  • webhook - This Node creates, removes, and manages the Webhook features of the Spark API. Once deployed, and subsequently triggered, the contents of the notification will be sent to the output of this Node.
  • parser - This is a utility Node that accepts input from the output of either the API or Webhook Node.
  • auth - This is a Config Node that holds the credential Bearer Token for the Spark API. Once initially defined, either under the Api or Webhook Node, it will be available to select on all other Api and Webhook Nodes that are created. You can define multiple Auth profiles so as to work with different Spark Accounts or Bots within the same flow.

Install

The simplest installation method is done via Node-RED itself. In the "Manage Palette" menu dropdown, search for this module by name. Alternate methods are outlined below.

Via the node-red-admin CLI tool:

# requires node-red to be running

# install node-red-admin if needed
npm install -g node-red-admin

# authenticate if your Node-RED administration has been secured
node-red-admin login

# install the module
node-red-admin install node-red-contrib-spark

Via source:

# clone repo
git clone https://github.com/cumberlandgroup/node-red-contrib-spark
cd node-red-contrib-spark

# get submodule
git submodule init
git submodule update

# from your node-red user directory
cd $HOME/.node-red
npm install /path/to/node-red-contrib-spark

Via NPM repository:

# from your node-red user directory
cd $HOME/.node-red
npm install node-red-contrib-spark

API Node

The Spark API Node sends REST queries via messages received by the input connector in the msg.payload object. Results of the API call are provided at the output in the msg.payload object. If multiple records are returned from the Spark API Call, these are passed to the output as individual sequential messages. The msg.parts property is set appropriately for use with the join node if a single array payload is preferred.

Module Input

The Spark API request is passed as a JSON object in msg.payload. The msg.payload object can be an empty object {} or contain optional path, query-string, or body variables provided using "key": "value" object parameters. Depending on the particular API method, the "key": "value" properties are defined in either the msg.payload object or the msg.payload.body object.

Module Output

By convention, the output from the Spark API call will have a msg.payload property. This contains the results of the API call in JSON format. The format of this JSON object will be the same as documented at developer.ciscospark.com for the responses from the API call.

Additionally the following are defined as part of the msg object:

  • msg.status : http return code
  • msg.headers - response headers object

Multiple Results

If multiple records are returned from the Spark API Call, these are passed to the output as individual sequential messages. The msg.parts property is set appropriately for use with the join node if a single array payload is preferred.

Example: Get Person by Email

The following object would be sent in the msg.payload input to a Spark API Node setup for People.getPeople:

{
  "email": "[email protected]"
}

Example: Get Rooms by Type

The following object would be sent in the msg.payload input to a Spark API Node setup for Rooms.getRooms:

{
  "body": {
    "type": "group"
  }
}

Example: Update a Room Title

The following object would be sent in the msg.payload input to a Spark API Node setup for Rooms.updateRoom:

{
  "roomId": "someSparkRoomIdString",
  "body": {
    "title": "My Renamed Room"
  }
}

Example: Create a New Message

The following object would be sent in the msg.payload input to a Spark API Node setup for Messages.createMessage:

{
  "body": {
    "roomId": "someSparkRoomIdString",
    "text": "Hello World!"
  }
}

Example: Add Person by Email to a Room

The following object would be sent in the msg.payload input to a Spark API Node setup for Memberships.createMembership:

{
  "body": {
    "roomId": "someSparkRoomIdString",
    "personEmail": "[email protected]"
  }
}

Configuration Options

  • Profile - The Spark credential profile to use with this Node.
  • Resource - The Spark resources for the API action.
  • Method - The specific method available to the selected Spark resource.

Webhook Node

The Spark Webhook Node is triggered when a resource event is matched. When the Node is deployed, it automatically creates the associated Cisco Spark Webhook. When the Node is removed, the Webhook reference is automatically removed in the Spark API.

Example Output : msg.payload

{
    "id": "Y2lzY29zcGFyazovL3VzL1dFQkhPT0svNzJkYzlhNTctYmY4MC00OTdjLWFhM2MtNjMyYzUyOThkMTFk",
    "name": "Test",
    "targetUrl": "http://myhost.com:3000/node_red_contrib_spark_388d9fffca49b8",
    "resource": "messages",
    "event": "created",
    "ownedBy": "creator",
    "created": "2016-10-23T19:50:23.484Z",
    "data": {
        "id": "Y2lzY29zcGFyazovL3VzL01FU1NBR0UvZjgyOGM3YTAtOTk1OS0xMWU2LTk5ODYtYzc4MTAwYzIyYTJm",
        "roomId": "Y2lzY29zcGFyazovL3VzL1JPT00vNjNhYzQ3MzAtOTUzYy0xMWU2LWEwZmQtNDcxNWExOWY2ZDJi",
        "roomType": "group",
        "personId": "Y2lzY29zcGFyazovL3VzL1BFT1BMRS8zYzNlZmYxOS04Njg1LTQ2OTEtODViOS1lZjRmMTViZDk2ZDQ",
        "personEmail": "[email protected]",
        "created": "2016-10-23T19:50:37.594Z"
    }
}

Configuration Options

  • Profile - The Spark credential profile to use with this Node.
  • Resource - The Spark resource to bind a Webhook to.
  • Event - The specific event of the resource selected.
  • Host - The base URL that is used to build the Webhook in the Spark API. This should be reachable from the internet and follow the format of http(s)://domain.tld:<port>. The Webhook Node will dynamicly publish webroutes under this URL as /node_red_contrib_spark_<node-uuid>. Note that the Webhook is automatically created in the Spark API after deploying and automatically removed if the Node is deleted and the flow re-deployed. If you have defined the httpNodeRoot setting in Node-Red, the webhook target URL will be created under that path. This should be mostly transparent as this node will create and remove webhooks from the Spark API as the nodes are added and removed from the workspace. If you shutdown Node-Red with a webhook deployed on the workspace, this webhook will persist in the Spark API and will need to be manually removed.

Parser Node

The Spark Parse Node allows parsing of specific properties from the JSON msg.payload received from either the "Spark Webhook Node" or the "Spark API Node".

Configuration Options

Define each "property" to parse from the inbound msg.payload. Optionally, define a "name" to remap the property name used in the oubound JSON object. This allows remapping of properties such as msg.payload.id to msg.payload.roomId. If "name" is left undefined, it will use the same value as the original property. Add additional rows to define multiple properties and names to construct the output JSON object.

License

MIT License Copyright (c) 2016 Nicholas Marus

Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions:

The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software.

THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.

Node Info

Version: 2.0.5
Updated 8 months, 2 weeks ago
License: MIT

Downloads

2 in the last day
17 in the last week
136 in the last month

Nodes

  • Spark Payload Parser
  • Spark Authentication

Keywords

  • node-red
  • cisco
  • spark
  • client

Maintainers

  • cumberlandgroup
  • nmarus