New Pipelines Data Destination: LightDB State

A new Pipelines data destination for LightDB State is now generally available for Golioth users. LightDB State is a database that is built-in to the Golioth platform, enabling capture and synchronization of device state. It is frequently used as the foundation for digital twin and device shadow services.

How It Works

The lightdb_state data destination differs from many other destinations in that delivered data updates a device’s LightDB State JSON document, rather than appending timeseries data. It supports specifying a path parameter for nesting data in the JSON document, or defaults to the stream path used for filtering if one is not specified. This makes it possible to easily separate data delivered to LightDB State directly from devices from data that is routed through Pipelines.

The lightdb_state data destination only accepts data formatted as JSON, though data can easily converted to other formats prior to delivery using transformers.

filter:
  path: "*"
  content_type: application/json
steps:
  - name: step0
    destination:
      type: lightdb-state
      version: v1
      parameters:
        path: /latest/stream

Click here to use this pipeline in your Golioth project!

For more details on the lightdb_state data destination, go to the documentation.

What’s Next

LightDB State is unique because data routed to it can subsequently be fetched by devices. Combining the lightdb_state data destination with external data sources, such as those accessed via the webhook transformer, enables powerful new capabilities. Imagine a piece of data sent from a device, out through a webhook transformer to something like an AI service, and then returned to LightDB State without any 3rd party tools required. This path opens up a plethora of external compute scenarios.

We’ll be showing off some of these in the coming the days, so make sure to keep an eye on the blog. If you have any questions feel free to reach out to us on the forum!

Dan Mangum
Dan Mangum
Dan is an experienced engineering leader, having built products and teams at both large companies and small startups. He has a history of leadership in open source communities, and has worked across many layers of the technical stack, giving him unique insight into the constraints faced by Golioth’s customers and the requirements of a platform that enables their success.

Post Comments

No comments yet! Start the discussion at forum.golioth.io

More from this author

Related posts

spot_img

Latest posts

Golioth Joins the Google Cloud Partner Program

Golioth joins the Google Cloud Partner program, making your IT and OT integrations easier than ever.

Zephyr: How to Monitor Heap and Locate Memory Leaks

Dynamic memory allocation in C is a powerful tool but also a potential source of hard to replicate bugs. This post goes over tools that are built into Zephyr that monitor consumption over time.

Advanced IoT Data Uploads Using the Golioth Blockwise Stream

Golioth makes it easy to upload large data payloads from your IoT devices to the cloud. We do this using the blockwise Stream service, which includes a versatile set of transform and destination options so the data goes where you want, in the format that you need. In this post we extend the capabilities of the Blockwise stream to show how async makes a large upload even more reliable over a connection to Golioth.

Want to stay up to date with the latest news?

Subscribe to our newsletter and get updates every 2 weeks. Follow the latest blogs and industry trends.