Changelog

Use the following buttons to filter by change type:

added

Builtin models moved to modules

Nov 2024

The following resource models have moved to modules.

ResourceModel
boardcustomlinux
boardjetson
boardpca9685
boardodroid
boardti
boardpi
boardorange-pi
boardupboard
motortmc5072
motor28byj-48
encoderams-as5048
movement sensoradxl345
movement sensordual-gps-rtk
movement sensorgps-nmea-rtk-pmtk
movement sensorgps-nmea-rtk-serial
movement sensorgps-nmea
movement sensorimu-wit
movement sensorimu-wit-hwt905
movement sensormpu6050
power sensorina219
power sensorina226
sensorbme280
sensorsensirion-sht3xd
sensorpi
ML modelTFLite CPU

The following models were removed:

ResourceModel
grippersoftrobotics
motorencoded-motor
motorgpiostepper
motorroboclaw
sensords18b20
added

MoveThroughJointPositions to arm interface

Nov 2024

The arm interface now includes a MoveThroughJointPositions method that moves an arm through an ordered array of joint positions.

added

Set data retention policies

Oct 2024

You can now set how long data collected by a component should remain stored in the Viam Cloud in the component’s data capture configuration. For more information, see Data management service.

added

Pi models moved to module

Sep 2024

The Raspberry Pi 4, 3, and Zero 2 W boards are now supported by viam:raspberry-pi:rpi.

added

ESP32 cameras

Aug 2024

viam-micro-server now supports cameras on ESP32s. For more information, see Configure an esp32-camera.

changed

Micro-RDK now called viam-micro-server

Aug 2024

The lightweight version of viam-server that is built from the micro-RDK is now referred to as viam-micro-server. For more information, see viam-micro-server.

added

Provisioning

Aug 2024

You can now configure provisioning for machines with the Viam Agent. For more information, see Configure provisioning with viam-agent.

added

Data capture for vision

Aug 2024

Data capture is now possible for the vision service. For more information, see Supported components and services.

added

Create custom training scripts

Aug 2024

You can now upload custom training scripts to the Viam Registry and use them to train machine learning models. For more information, see Create custom training scripts.

changed

Operators can now view data

Jul 2024

The operator role now has view permissions for the data in the respective resource a user has access to. For more information, see Data and machine learning permissions.

changed

Python get_robot_part_logs parameters

Jun 2024

The errors_only parameter has been removed from get_robot_part_logs() and replaced with log_levels.

changed

Return type of analog Read

May 2024

The board analog API Read() method now returns an AnalogValue struct instead of a single int. The struct contains an int representing the value of the reading, min and max range of values, and the precision of the reading.

added

CaptureAllFromCamera and GetProperties to vision API

May 2024

The vision service now supports two new methods: CaptureAllFromCamera and GetProperties.

changed

Renamed GeoObstacle to GeoGeometry

May 2024

The motion service API parameter GeoObstacle has been renamed to GeoGeometry. This affects users of the MoveOnGlobe() method.

changed

Return type of GetImage

May 2024

The Python SDK introduced a new image container class called ViamImage. The camera component’s GetImage() method now returns a ViamImage type, and the vision service’s GetDetections() and GetClassifications() methods take in ViamImage as a parameter.

You can use the helper functions viam_to_pil_image and pil_to_viam_image provided by the Python SDK to convert the ViamImage into a PIL Image and vice versa.

Click for an example of using the ViamImage -> PIL Image helper functions.
from viam.media.utils.pil import pil_to_viam_image, viam_to_pil_image

# Get the ViamImage from your camera.
frame = await my_camera.get_image()

# Convert "frame" to a PIL Image representation.
pil_frame = viam_to_pil_image(frame)

# Use methods from the PIL Image class to get size.
x, y = pil_frame.size[0], pil_frame.size[1]
# Crop image to get only the left two fifths of the original image.
cropped_pil_frame = pil_frame.crop((0, 0, x / 2.5, y))

# Convert back to ViamImage.
cropped_frame = pil_to_viam_image(cropped_pil_frame)

# Get detections from your vision service.
detections = await detector.get_detections(cropped_frame)
removed

WriteAnalog from Go SDK

May 2024

The WriteAnalog() method has been removed from the Go SDK. Use AnalogByName() followed by Write() instead.

changed

Python SDK data retrieval behavior

Apr 2024

tabular_data_by_filter() and binary_data_by_filter() now return paginated data.

changed

Renamed AnalogReader to Analog

Apr 2024

AnalogReader has been renamed to Analog. The functionality remains the same, but code that uses analogs must be updated. AnalogReaderByName() and AnalogReaderNames() have become AnalogByName() and AnalogNames() (since deprecated), respectively.

added

Part online and part offline triggers

Apr 2024

You can now configure triggers to execute actions when a machine part comes online or goes offline.

removed

Status from Board API

Apr 2024

Viam has removed support for the following board API methods: Status(), AnalogStatus(), DigitalInterruptStatus(), Close(), Tick(), AddCallback(), and RemoveCallback().

removed

Removed and replaced camera models

Apr 2024

Viam has removed support for following builtin camera models: single_stream, dual_stream, align_color_depth_extrinsics, and align_color_depth_homography.

changed

Updated GetCloudMetadata response

Apr 2024

In addition to the existing returned metadata, the GetCloudMetadata method now returns machine_id and machine_part_id as well.

improved

Viam app interface

Apr 2024

the Viam app machine page UI has been updated significantly. In addition to other improvements, your component, service, and other resource config cards are all displayed on one page instead of in separate tabs.

added

Additional ML models

Mar 2024

Viam has added support for the TensorFlow, PyTorch, and ONNX ML model frameworks, expanding upon the existing support for TensorFlow Lite models. You can now upload your own ML model(/registry/ml-models/) using any of these frameworks for use with the Vision service.

added

Ultrasonic sensor for `viam-micro-server`

Mar 2024

You can now use the ultrasonic sensor component with viam-micro-server to integrate an HC-S204 ultrasonic distance sensor into a machine running viam-micro-server.

added

Edit a machine configuration that uses a fragment

Mar 2024

You can now edit the configuration of an existing machine that has been configured with a fragment by using the fragment_mods object in your configuration. You can use the fragment_mods objects to be able to deploy a fragment to a fleet of machines, but still be able to make additional per-machine edits as needed.

added

Dual GPS movement sensor

Mar 2024

You can now use the dual GPS movement sensor component to integrate a movement sensor that employs two GPS sensors into your machine. The dual GPS movement sensor calculates a compass heading from both GPS sensors, and returns the midpoint position between the two sensors as its position.

added

Viam Agent

Mar 2024

You can now use the Viam Agent to provision your machine or fleet of machines during deployment. The Viam Agent is a software provisioning manager that you can install on your machine which manages your viam-server installation, including installation and ongoing updates, as well as providing flexible deployment configuration options, such as pre-configured WiFi network credentials.

added

Generic service

Feb 2024

You can now use the generic service to define new, unique types of services that do not already have an appropriate API defined for them.

added

ML models in the registry

Feb 2024

You can now upload machine learning (ML) models to the Viam Registry, in addition to modules. You may upload models you have trained yourself using the Viam app, or models you have trained outside of the App. When uploading, you have the option to make your model available to the general public for reuse.

added

Sensor-controlled base

Jan 2024

Viam has added a sensor-controlled base component model, which supports a robotic base that receives feedback control from a movement sensor.

added

Visualize captured data

Jan 2024

You can now visualize your data using many popular third-party visualization tools, including Grafana, Tableau, Google’s Looker Studio, and more. You can visualize any data, such as sensor readings, that you have synced to the Viam app from your machine.

See Visualize data with Grafana for a full walkthrough focused on Grafana specifically.

added

Use triggers to trigger actions

Jan 2024

You can now configure triggers (previously called webhooks) to execute actions when certain types of data are sent from your machine to the cloud.

added

Filtered camera module

Dec 2023

Viam has added a filtered-camera module that selectively captures and syncs only the images that match the detections of an ML model. For example, you could train an ML model that is focused on sports cars, and only capture images from the camera feed when a sports car is detected in the frame.

Check out this guide for more information.

added

Raspberry Pi 5 Support

Dec 2023

You can now run viam-server on a Raspberry Pi 5 with the new board model pi5.

added

Role-based access control

Dec 2023

Users can now have access to different fleet management capabilities depending on whether they are an owner or an operator of a given organization, location, or machine.

added

Authenticate with location API key

Nov 2023

You can now use API keys for authentication. API keys allow you to assign the minimum required permissions for usage. Location secrets, the previous method of authentication, is deprecated and will be removed in a future release.

added

Queryable sensor data

Nov 2023

Once you have added the data management service and synced data, such as sensor readings, to the Viam app, you can now run queries against both captured data as well as its metadata using either SQL or MQL.

For more information, see Query Data with SQL or MQL.

changed

Model training from datasets

Nov 2023

To make it easier to iterate while training machine learning models from image data, you now train models from datasets.

improved

Manage users access

Nov 2023

You can now manage users access to machines, locations, and organizations. For more information, see Access Control

added

Test an ML model in browser

Oct 2023

After you upload and train a machine learning model, you can test its results in the Data tab.

This allows you to refine models by iteratively tagging more images for training based on observed performance.

For more information, see Test classification models with existing images in the cloud.

To use this update, the classifier must have been trained or uploaded after September 19, 2023. The current version of this feature exclusively supports classification models.

added

PLC support

Oct 2023

The Viam platform now supports the Revolution Pi line of PLCs from KUNBUS in the form of a module. This collaboration allows you to leverage the Raspberry Pi-based Revolution Pi, which runs on Linux and has a specially designed I/O modules for streamlined interaction with industrial controls, eliminating the need for additional components.

Read the Viam PLC Support blog post for a step-by-step guide on using a PLC with Viam.

improved

SLAM map creation

Oct 2023

The Cartographer-module now runs in Viam’s cloud for creating or updating maps. This enhancement allows you to:

  • Generate larger maps without encountering session timeouts
  • Provide IMU input to improve map quality
  • Save maps to the SLAM library
  • Create or update maps using previously captured LiDAR and IMU data
  • Deploy maps to machines
added

Modular registry

Sep 2023

The Modular Registry enables you to use, create, and share custom modules, extending the capabilities of Viam beyond the components and services that are natively supported.

You can:

  • Publish modules on the registry
  • Add modules to any machine’s configuration with a few clicks
  • Select the desired module version for deployment, make changes at your convenience, and deploy the updates to a single machine or an entire fleet.
added

Mobile app

Sep 2023

You can use a mobile application, available for download now in the Apple and Google Play app stores, to connect to and control your Viam-powered machines directly from your mobile device.

added

Power sensor component

Sep 2023

You now have the capability to use a power sensor component to monitor the voltage, current, and power consumption within your machine’s system.

added

Filter component’s data before the cloud

Sep 2023

Viam has written a module that allows you to filter data based on specific criteria before syncing it to Viam’s cloud. It equips machines to:

  • Remove data that is not of interest
  • Facilitate high-interval captures while saving data based on your defined metrics
  • Prevent the upload of unnecessary data

To learn more, see this tutorial on creating and configuring a data filtration module.

added

Configure a custom Linux board

Aug 2023

You can now use boards like the Mediatek Genio 500 Pumpkin that run Linux operating systems with the customlinux board model.

improved

Image inspection for ML training

Aug 2023

This update enables you to get a closer examination of your image and streamline your image annotation experience by making it easier to add bounding boxes and labels in the Data tab.

With the latest improvements, you can now:

  • Navigate between images using the arrow keys in the main image view
  • Expand images for a more detailed inspection by clicking the expand button on the right image panel
  • Move between full-screen images effortlessly with the <> arrow buttons or arrow keys
  • Return to the standard view by using the escape key or collapse button
added

Duplicate component button

Aug 2023

You now have the ability to duplicate any config component, service, module, remote, or process.

To use this feature:

  • Click on the duplicate component icon at the top right of any resource
  • Optionally, you can modify the component name to distinguish it
  • Adjust any attributes, such as motor pin numbers
added

Apple SSO authentication

Jul 2023

Viam now supports sign-up/log-in through Apple Single Sign-On.

Note that currently, accounts from different SSO providers are treated separately, with no account merging functionality.

improved

Arm component API

Jul 2023

Arm models now support the GetKinematics method in the arm API, allowing you to request and receive kinematic information.

added

View sensor data within Viam

Jun 2023

You can now view your sensor data directly in the Viam app to verify data creation and accuracy. If you depend on sensor data to plan and control machine operations, this feature increases access to data and supports a more efficient workflow.

added

Session management in the Python SDK

Jun 2023

The Python SDK now includes sessions, a safety feature that automatically cancels operations if the client loses connection to your machine.

Session management helps you to ensure safer operation of your machine when dealing with actuating controls. Sessions are enabled by default, with the option to disable sessions.

added

Connect an ODrive motor controller as a Viam module

Jun 2023

You can integrate and control ODrive motor controllers with Viam using the odrive module from the Viam Registry.

See the Odrive module readme to learn how to connect and use an ODrive motor controller with Viam, and view the sample configurations.

added

Implement custom robotic arms as Viam modules

Jun 2023

When prototyping a robotic arm, you can now facilitate movement without creating your own motion planning. This update enables you to implement custom models of an arm component as a modular resource by coding three endpoints of the Arm API:

  • getJointPositions
  • movetoJointPositions
  • GetKinematics

Then, use the motion planning service to specify poses, and Viam handles the rest.

For more information, see this tutorial on creating a custom arm.

improved

Gantry component

Jun 2023

To better control gantries with Viam, you can now:

  • Specify speed values when calling the MovetoPosition method on Gantry components. This allows you to define the speed at which each axis moves to the desired position, providing enhanced precision and control over the gantry’s movement.
  • Set a home position for Gantry components to facilitate position resetting or maintain consistent starting points.
improved

Optimized Viam-trained object detection models

Jun 2023

This update for object detection models trained with the machine learning service brings significant improvements, including:

  • 76% faster model inference for camera streams
  • 64% quicker model training for object detection
  • 46% reduction in compressed model size
added

TypeScript SDK beta release

May 2023

The beta release of the TypeScript SDK allows you to create a web interface to work with your machine, as well as create custom components and services.

added

Train object detection ML models

May 2023

You now have the capability to directly train object detection models in addition to image classification models from within the Viam app.

This update allows you to:

  • Add labels by drawing bounding boxes around specific objects in your images or a single image.
  • Create a curated subset of data for training by filtering images based on labels or tags.
added

Permissions for organizations in Viam

May 2023

Now when you invite collaborators to join your organization, you can assign permissions to members by setting one of these roles:

  • Owner: These members can see and edit every tab on the machine page, as well as manage users in the app. This role is best for those on your team who are actively engineering and building machines.

  • Operator: These members can only see and use the remote control tab. This role is best for those on your team who are teleoperating or remotely controlling machines.

For more information about assigning permissions and collaborating with others on Viam, see Fleet Management.

improved

Control RoboClaw motor controllers with the driver

May 2023

When using a RoboClaw motor controller without encoders connected to your motors, you now have more direct control over the RoboClaw’s functionality within Viam or through the motor API.

For example, in the Viam app, you can now set Go For values for these motors, utilizing a time-based estimation for the number of revolutions.

improved

Camera webcam names and setting framerates

May 2023

The updates to the camera component have improved the process of connecting to and using cameras with your machines.

The latest updates enable you to:

  • View readable webcam names in the video path of your camera component.
  • Specify your preferred framerate by selecting the desired value in the newly added framerate field on the CONFIGURE tab.
improved

Additions to code samples in the Viam app

May 2023

The updated code samples now includes:

  • Options for C++ and TypeScript
  • The ability to hide or display your machines’ secrets

Access these samples in the Code sample tab on your machine’s page to connect to your machine in various languages.

improved

Delete data in bulk in the Viam app

May 2023

You can manage the data synced to Viam’s cloud with the new capability for bulk data deletion on the Data tab.

changed

Vision service

Apr 2023

Use individual vision service instances

You need to create an individual vision service instance for each detector, classifier, and segmenter model. You can no longer be able to create one vision service and register all of your detectors, classifiers, and segmenters within it.

Click for details on how to migrate your code.

API calls

Change your existing API calls to get the new vision service instance for your detector, classifier, or segmenter model directly from the VisionClient:

Change your existing API calls to get the new vision service instance for your detector, classifier, or segmenter model directly from the VisionClient:

my_object_detector = VisionClient.from_robot(robot, "find_objects")
img = await cam.get_image()
detections = await my_object_detector.get_detections(img)
vision = VisionServiceClient.from_robot(robot)
img = await cam.get_image()
detections = await vision.get_detections(img, "find_objects")

Color detector configurations

You can replace existing color detectors by configuring new ones in the UI or you can update the JSON configuration of your machines:

"services": [
    {
        "name": "blue_square",
        "type": "vision",
        "model": "color_detector",
        "attributes": {
            "segment_size_px": 100,
            "detect_color": "#1C4599",
            "hue_tolerance_pct": 0.07,
            "value_cutoff_pct": 0.15
        }
    },
    {
        "name": "green_triangle",
        "type": "vision",
        "model": "color_detector",
        "attributes": {
            "segment_size_px": 200,
            "detect_color": "#62963F",
            "hue_tolerance_pct": 0.05,
            "value_cutoff_pct": 0.20
        }
    },
    ... // other services
]
"services": [
    {
        "name": "vision",
        "type": "vision",
        "attributes": {
            "register_models": [
            {
                "parameters": {
                    "segment_size_px": 100,
                    "detect_color": "#1C4599",
                    "hue_tolerance_pct": 0.07,
                    "value_cutoff_pct": 0.15
                },
                "name": "blue_square",
                "type": "color_detector"
            },
            {
                "parameters": {
                    "segment_size_px": 200,
                    "detect_color": "#62963F",
                    "hue_tolerance_pct": 0.05,
                    "value_cutoff_pct": 0.20
                },
                "name": "green_triangle",
                "type": "color_detector"
            }
            ]
        }
    },
    ... // other services
]

TFLite detector configurations

You can replace existing TFLite detectors by configuring new ones in the UI or you can update the JSON configuration of your machines:

"services": [
    {
        "name": "person_detector",
        "type": "mlmodel",
        "model": "tflite_cpu",
        "attributes": {
            "model_path": "/path/to/file.tflite",
            "label_path": "/path/to/labels.tflite",
            "num_threads": 1
        }
    },
    {
        "name": "person_detector",
        "type": "vision",
        "model": "mlmodel",
        "attributes": {
            "mlmodel_name": "person_detector"
        }
    },
    ... // other services
]
"services": [
    {
        "name": "vision",
        "type": "vision",
        "attributes": {
            "register_models": [
            {
                "parameters": {
                    "model_path": "/path/to/file.tflite",
                    "label_path": "/path/to/labels.tflite",
                    "num_threads": 1
                },
                "name": "person_detector",
                "type": "tflite_detector"
            }
            ]
        }
    },
    ... // other services
]

TFLite Classifier configurations

You can replace existing TFLite classifiers by configuring new ones in the UI or you can update the JSON configuration of your machines:

"services": [
    {
        "name": "fruit_classifier",
        "type": "mlmodel",
        "model": "tflite_cpu",
        "attributes": {
            "model_path": "/path/to/classifier_file.tflite",
            "label_path": "/path/to/classifier_labels.txt",
            "num_threads": 1
        }
    },
    {
        "name": "fruit_classifier",
        "type": "vision",
        "model": "mlmodel",
        "attributes": {
            "mlmodel_name": "fruit_classifier"
        }
    },
    ... // other services
]
"services": [
    {
        "name": "vision",
        "type": "vision",
        "attributes": {
            "register_models": [
            {
                "parameters": {
                    "model_path": "/path/to/classifier_file.tflite",
                    "label_path": "/path/to/classifier_labels.txt",
                    "num_threads": 1
                },
                "name": "fruit_classifier",
                "type": "tflite_classifier"
            }
            ]
        }
    },
    ... // other services
]

Radius Clustering 3D segmenter configurations

You can replace existing Radius Clustering 3D segmenters by configuring new ones in the UI or you can update the JSON configuration of your machines:

"services": [
    {
        "name": "rc_segmenter",
        "type": "vision",
        "model": "obstacles_pointcloud"
        "attributes": {
            "min_points_in_plane": 1000,
            "min_points_in_segment": 50,
            "clustering_radius_mm": 3.2,
            "mean_k_filtering": 10
        }
    },
    ... // other services
]
"services": [
    {
        "name": "vision",
        "type": "vision",
        "attributes": {
            "register_models": [
            {
                "parameters": {
                    "min_points_in_plane": 1000,
                    "min_points_in_segment": 50,
                    "clustering_radius_mm": 3.2,
                    "mean_k_filtering": 10
                },
                "name": "rc_segmenter",
                "type": "radius_clustering_segmenter"
            }
            ]
        }
    },
    ... // other services
]

Detector to 3D segmenter configurations

You can replace existing Radius Clustering 3D segmenters by configuring new ones in the UI or you can update the JSON configuration of your machines:

"services": [
    {
        "name": "my_segmenter",
        "type": "vision",
        "model": "detector_3d_segmenter"
        "attributes": {
            "detector_name": "my_detector",
            "confidence_threshold_pct": 0.5,
            "mean_k": 50,
            "sigma": 2.0
        }
    },
    ... // other services
]
"services": [
    {
        "name": "vision",
        "type": "vision",
        "attributes": {
            "register_models": [
            {
                "parameters": {
                    "detector_name": "my_detector",
                    "confidence_threshold_pct": 0.5,
                    "mean_k": 50,
                    "sigma": 2.0
                },
                "name": "my_segmenter",
                "type": "detector_segmenter"
            }
            ]
        }
    },
    ... // other services
]

Add and remove models using the machine config

You must add and remove models using the machine config. You will no longer be able to add or remove models using the SDKs.

Add machine learning vision models to a vision service

The way to add machine learning vision models is changing. You will need to first register the machine learning model file with the ML model service and then add that registered model to a vision service.

added

Machine learning for image classification models

Mar 2023

You can now train and deploy image classification models with the data management service and use your machine’s image data directly within Viam. Additionally, you can upload and use existing machine learning models with your machines. For more information on using data synced to the cloud to train machine learning models, read Train a model.

added

Motion planning with new `constraint` parameter

Mar 2023

A new parameter, constraint, has been added to the Motion service API, allowing you to define restrictions on the machine’s movement. The constraint system also provides flexibility to specify that obstacles should only impact specific frames of a machine.

added

Fragments in machine configuration

Mar 2023

You can now access fragments in your machine configuration. The configurations you added will now show up automatically in the Builder view on your machine’s CONFIGURE tab. This makes it easier to monitor what fragments you’ve added to your machine and how they’re configured.

For more information, see Fragments.

improved

Sticky GPS keys

Mar 2023

GPS keys you enter are now saved in your local storage. This ensures that when you reload the page, your GPS keys remain accessible.

improved

More reliable camera streams

Mar 2023

The camera component’s streams are smoother and more reliable with recent improvements.

Additionally, camera streams automatically restart if you momentarily lose internet connection.

improved

UI updates to Logs and History

Mar 2023

The latest UI updates enable you to:

  • Load a previous configuration for reverting changes made in the past
  • Search logs by filtering keywords or log levels such as info or error messages
  • Change your timestamp format to ISO or Local depending on your preference.
added

Rover reuse in Try Viam

Feb 2023

You now have the option to reuse a machine config from a previous Try Viam session.

added

Dynamic code samples

Feb 2023

The Viam app Code sample tab now dynamically updates as you add resources to your machine’s config. The code samples instantiate each resource and include examples of how to call a Get method on it.

added

TypeScript SDK

Feb 2023

Find more information in the TypeScript SDK docs.

added

Frame system visualizer

Feb 2023

When adding frames to your machine’s config in the Viam app, you can now use the Frame System subtab of the CONFIGURE tab to more easily visualize the relative positions of frames.

added

Support for microcontrollers

Feb 2023

viam-micro-server is a lightweight version of viam-server that can run on an ESP32. Find more information in the viam-micro-server installation docs.

added

Remote control power input

Jan 2023

On your machine’s CONTROL tab on the Viam app, you can now set the power of a base. The base control UI previously always sent 100% power to the base’s motors.

added

New encoder model: AMS AS5048

Jan 2023

The AMS AS5048 is now supported.

added

GetLinearAcceleration method

Jan 2023

The movement sensor API now includes a GetLinearAcceleration method.

added

Support for capsule geometry

Jan 2023

The motion service now supports capsule geometries.

The UR5 arm model has been improved using this new geometry type.

added

Modular resources

Dec 2022

You can now implement your own custom resources as modular resources.

added

URDF kinematic file support

Dec 2022

You can now supply kinematic information using URDF files when implementing your own arm models.

added

New movement sensor models

Dec 2022

There are two new movement sensor models:

  • ADXL345: A 3-axis accelerometer
  • MPU-6050: A 6-axis accelerometer and gyroscope
improved

Camera performance and reliability

Dec 2022
  • Improved server-side logic to choose a mime type based on the camera image type, unless a specified mime type is supplied in the request. The default mime type for color cameras is now JPEG, which improves the streaming rate across every SDK.
  • Added discoverability when a camera reconnects without changing video paths. This now triggers the camera discovery process, where previously users would need to manually restart the RDK to reconnect to the camera.
improved

Motion planning with remote components

Dec 2022

The motion service is now agnostic to the networking topology of a machine.

  • Kinematic information is now transferred over the robot API. This means that the motion service is able to get kinematic information for every component on the machine, regardless of whether it is on a main or remote viam-server.
  • Arms are now an input to the motion service. This means that the motion service can plan for a machine that has an arm component regardless of whether the arm is connected to a main or remote-part instance of viam-server.
improved

Motion planning path smoothing

Dec 2022
  • RRT* paths now undergo rudimentary smoothing, resulting in improvements to path quality with negligible change to planning performance.
  • Plan manager now performs direct interpolation for any solution within some factor of the best score, instead of only in the case where the best inverse kinematics solution could be interpolated.
improved

Data synchronization reliability

Dec 2022

Previously, data synchronization used bidirectional streaming. Now is uses a simpler unary approach that is more performant on batched unary calls, is easier to load balance, and maintains ordered captures.

changed

Camera configuration

Nov 2022

Changed the configuration schemes for the following camera models:

  • Webcam
  • FFmpeg
  • Transform
  • Join pointclouds

For information on configuring any camera model, see Camera Component.

added

New servo model

Nov 2022

A new servo model called gpio supports servos connected to non-Raspberry Pi boards.

added

RTT indicator in the app

Nov 2022

A badge in the Viam app now displays RTT (round trip time) of a request from your client to the machine. Find this indicator of the time to complete one request/response cycle on your machine’s CONTROL tab, in the Operations & Sessions card.

added

Python 3.8 support

Nov 2022

The Python SDK now supports Python 3.8, in addition to 3.9 and 3.10.

added

New parameter: `extra`

Nov 2022

A new API method parameter, extra, allows you to extend modular resource functionality by implementing the new field according to whatever logic you choose. extra has been added to the following APIs: arm, data management, gripper, input controller, motion, movement sensor, navigation, pose tracker, sensor, SLAM, vision.

added

Service dependencies

Nov 2022

viam-server now initializes and configures resources in the correct order. For example, if the SLAM service depends on a LiDAR, it will always initialize the LiDAR before the SLAM service.

removed

Width and height fields from camera API

Nov 2022

Removed width and height from the response of the GetImage method in the camera API. This does not impact any existing camera models. If you write a custom camera model, you no longer need to implement the width and height fields.