Capture Data and Train a Model

In this tutorial, you will use three Viam services together to enable your machine to recognize specific objects in the world around it:

  • The data management service, to capture images from a camera on your machine and sync them to the cloud.
  • The ML model service, to manage and deploy a machine learning (ML) model based on these images, once you have added tags to the images matching the objects you want to detect.
  • The vision service, to enable your machine’s camera to detect objects defined in the ML model on its own.

With all three services working together, your machine will be able to analyze its camera feed for the presence of specific shapes, such as a red star or blue circle. When it detects a likely match, it will overlay a confidence score onto the camera feed alongside the name of the detected shape, indicating how closely the shape in the camera frame matches a shape it has seen before.

Prerequisites

Before following this tutorial, ensure you have:

  • A machine running viam-server, the software that runs your machine and connects it to the Viam app.

    • If you are using your own machine:

      Add a new machine in the Viam app. Then follow the setup instructions to install viam-server on the computer you’re using for your project and connect to the Viam app. Wait until your machine has successfully connected.

    • If you are using a Viam Rover through Try Viam, no further action is needed.

  • A configured camera component.

    • If you are using your own machine, add a camera component, such as a webcam, to your machine.
    • If you are using a Viam Rover through Try Viam, a camera is already configured for you.

The data management service

You can manage how your machine works with data files and images by using the data management service.

The data management service has two parts: data capture and cloud sync.

  • Data capture allows you to capture data locally from specific components on your machine running Viam. You can choose the components, corresponding methods, and the frequency of the data capture from the Viam app.

  • Cloud sync runs in the background and uploads your machine’s captured data to the Viam app at a defined frequency. Cloud sync is designed to be resilient and to preserve your data even during a network outage or if your machine has low network bandwidth. With cloud sync enabled for a component, data captured locally to your machine is automatically deleted after a successful sync. Data synced between your machine and the Viam app is encrypted in transit (over the wire) and when stored in the cloud (at rest).

Data capture and data sync are frequently used together, and are both enabled by default when you add the data management service to your machine. However, if you want to manage your machine’s captured data yourself, you can enable data capture but disable data sync.

To capture data from your machine and sync to the Viam app, add the data management service and configure data capture for at least one component.

Add the data management service

First, add the data management service to your machine to be able capture and sync data:

  1. On your machine’s CONFIGURE page in the Viam app, navigate to the Services tab.

  2. Click the Create service button at the bottom of the page, and select Data Management.

  3. Use the suggested name for your service or give it a name, like viam-data-manager, then click Create Service.

  4. On the panel that appears, you can manage the capturing and syncing functions individually. By default, the data management service captures data locally to the ~/.viam/capture directory, and syncs captured data files to the Viam app every 6 seconds (0.1 minutes in the configuration). Leave the default settings as they are, and click Save Config at the bottom of the screen to save your changes.

    The data management service configuration pane with default settings shown for both capturing and syncing

For more information, see Add the data management service.

Configure data capture for a component

Once you have added the data management service, you can configure data capture for specific components on your machine. For this tutorial, you will configure data capture for images from a camera component, but other data types such as sensor data or SLAM map data from other types of components can be captured as well.

To enable image data capture for a camera component:

  1. Navigate to your machine’s CONFIGURE page in the Viam app.

  2. In the configuration pane for your configured camera component, find the Data capture section, and click the Add method button to configure data capture for this camera.

    • Set the Method to ReadImage and the Frequency to 0.333. This will capture an image from the camera roughly once every 3 seconds. You can adjust the capture frequency if you want the camera to capture more or less image data, but avoid configuring data capture to higher rates than your hardware can handle, as this could lead to performance degradation.

    • Set the Mime type to image/jpeg to configure image data capture.

      The camera component configuration pane with data capture configuration enabled using type ReadImage and a capture frequency of 0.333
    • Toggle the On/Off switch to On.

  3. Click Save at the top right of the window to save your changes.

For more information see Configure data capture and Configure cloud sync.

View and filter captured data

Now that you have configured data capture on your camera component, you can view the resulting data files in the Viam app.

Click on the menu icon on the camera configuration pane and select View captured data.

Resource menu with the options Rename, Duplicate, View captured data, and Delete

Here you can view the images captured so far from the camera on your machine. New images should appear roughly every six seconds as cloud sync uploads them from your machine.

You can use the filters on the left side of the screen to filter your images by machine, component, date range, and more.

If you have a lot of images, filter them by limiting the displayed images to a specific date and time range:

The data tab displaying images filtered by date and time range

For more information see View and filter data.

The ML model service

Once your machine is capturing and syncing images to the Viam app, you are ready to train a machine learning (ML) model using those images. You can use an ML model to help your machine adapt its behavior to the world around it.

For this tutorial, you will train an ML model to be able to recognize specific shapes (for example, red and blue stars), and then deploy that model to your machine using the ML (machine learning) model service. With a model deployed to your machine, you can use the ML model service together with the vision service to analyze newly-detected objects for a possible match to a known shape.

To train a model from your captured data, first tag your images with appropriate labels and add them to a dataset. Then train a model based on your dataset and labels and deploy the model to your machine.

Tag images and create a dataset

  1. From the DATA page in the Viam app, select an image captured from your machine that you would like to tag and add to your dataset.

  2. In the Tags field on the right-hand side, enter a new tag describing the object in the image that you want your machine to be able to identify, then click Add as new tag. Tag names support alphanumeric characters, underscores, and hyphens.

    For this tutorial, you will use the objects shown in the Try Viam test octagon, which include shapes like a red star and an orange triangle. The image below shows the red_star tag being added to an image:

    The tags and labels configuration pane of a selected image, with the text red_star entered as the tag

    If you are not using a Try Viam rover, you can choose objects in your machine’s surroundings instead.

    Then use the Datasets dropdown to create a new dataset and assign the image to it. We called our dataset shapes.

  3. Repeat this process for other images that contain objects you want your model to be able to identify. Once you have added a tag to an image and created a dataset, you can select the tag from the Tags dropdown menu and the dataset from the Datasets dropdown for other images.

Feel free to return to your machine’s Control tab to position your camera (and rover) to capture additional images from a variety of different angles, or with different lighting or background compositions. Generally, the more different perspectives of a given object you tag, the more likely your model will be able to identify it, even under differing conditions. The following is an example of a good selection of images containing the blue_star tag, taken from a variety of angles:

The data tab showing images that contain blue stars from various angles, with a search filter for blue_star applied to only show images tagged with that label

If you want to remove a tag, click the X icon to the right of the tag name below the Tags field.

View your dataset

Upon completion of tagging your data set, you can view the data in your dataset by clicking on your dataset’s name on the image sidebar or on the DATASETS subtab.

The shapes dataset.

Train a model on a dataset

To train a model:

  1. From the dataset view, click on Train model.

  2. Give your model a name, like my-classifier-model, and select the Model type:

    • Use Single label classification if you only added one tag per image.
    • Use Multi label classification if you added more than one tag for some images.
  3. Select the tags you want to train your model on from the Labels for training dropdown, then click Train model. Unselected tags will be ignored, and will not be part of the resulting model. If you do not see a tag you expected to see in the Labels for training dropdown, ensure your filtered images contain images with the tag.

    The data tab showing the train a model pane with five tags filtered

Your model will begin training on the images you have tagged, and should be ready after a short time. You can view your model’s training progress from the Models subtab under the DATA page.

The models tab on the data page showing a model named my-classifier-model being trained

Models that are still being trained appear under Training, while models that have completed training and are ready for use appear under Models.

The models tab on the data page showing a completed model named my-classifier-model ready for deployment

For more information, see Train a model.

Deploy a model

Once your model has finished training, add the ML model service and deploy your model to your machine to be able to use it to classify newly-captured images.

To deploy a model to your machine:

  1. Navigate to the CONFIGURE tab of your machine’s page in the Viam app.

  2. Click the + icon next to your machine part in the left-hand menu and select Service.

  3. Select the ML model type, then select the TFLite CPU model.

  4. Enter a name or use the suggested name, like my-mlmodel-service, for your service and click Create.

  5. In the resulting ML Model service configuration pane, select Deploy model on machine, then select the model you just trained from the Select model dropdown menu.

    The ML model service configuration pane showing the required settings to deploy the my-classifier-model.
  6. Click Save at the top right of the window to save your changes.

The vision service

Now that you have deployed an ML model to your machine using the ML model service, you are ready to configure a vision service which can detect the objects you’ve tagged in images from in a camera feed.

To create a camera that identifies the objects you’ve tagged and shows the identifications in its camera feed, first add the vision service, then add a transform camera. You can then see the tagged objects on the Control tab.

Add the vision service

  1. Click the + icon next to your machine part in the left-hand menu and select Service.

  2. Select the vision type, then select the ML model model.

  3. Enter a name, like my-vision-service, or use the suggested name for your service and click Create.

  4. In the resulting vision service configuration pane, select the ML model service you just added from the ML Model dropdown menu.

    The vision service configuration pane showing the ML model service my-mlmodel-service added
  5. Click Save Config at the bottom of the window to save your changes.

Add a transform camera

  1. On your machine’s CONFIGURE page in the Viam app, navigate to the Components tab.

  2. Click the Create components button at the bottom of the page, and select Camera, then select Transform camera.

  3. Give the camera a name, like my-transform-cam, then click Create.

  4. In the resulting camera components configuration pane, enter the following into the attributes field for the transform camera:

    {
      "pipeline": [
        {
          "type": "classifications",
          "attributes": {
            "classifier_name": "my-vision-service",
            "confidence_threshold": 0.5
          }
        }
      ],
      "source": "cam"
    }
    

    If you are not using a Try Viam rover, replace cam with the name of the configured camera on your machine. The confidence_threshold controls how confident the model must be in order to present a matching object tag, on a scale of 0.0 - 1.0, with 1.0 representing a 100% match requirement. The more and varied images you have captured and tagged, the more confidently your model can identify the objects you have tagged.

    The transform camera component configuration pane showing the required attributes entered
  5. Click Save Config at the bottom of the window to save your changes.

Test object detection

Your machine is now ready to detect the objects you’ve tagged.

  1. On your machine’s Control page in the Viam app, find your configured camera component. If you are using a Viam Rover, use the viam_base panel instead, which presents both the camera and the transform camera together.

  2. Enable both the camera and the transform camera, to show both on the right-hand side. On the Viam Rover, and using the transform camera name from earlier in this tutorial, these are cam and my-transform-cam.

  3. Move your machine to a position where your camera can see an object that you have tagged in your ML model, and watch your machine identify it!

    The control tab for a rover showing a live camera feed with the transform camera overlay indicating an object match for the blue_star tag The control tab for a rover showing a live camera feed with the transform camera overlay indicating an object match for the red_star tag

That’s it! Your machine is now smarter and better able to understand the world around it.

Use your model with code

Once you know your model is working, you can write code to change your machine’s behavior based on what it detects. For an example see the following tutorial:

Troubleshooting

If you are using Try Viam and your session expires, you can re-use a configuration from a previous session to keep your configuration changes. You can also extend your existing session while it’s still running, if it hasn’t expired yet.

If your machine isn’t capturing data and syncing it to the Viam app, ensure that both the data management service (named viam-data-manager in this tutorial) and the Data capture configuration for your camera (cam on the Try Viam rover) are enabled.

If your transform camera is not matching objects you have tagged, try lowering the confidence_threshold, or adding and tagging more images.

Next steps

In this tutorial, you learned:

  • how to use the data management service to capture images from your machine’s camera and sync them to the Viam app
  • how to filter and tag your synced images according to the objects you wanted to detect
  • how to use the ML model service to train an ML model based on those images and deploy that model to your machine
  • how to use the vision service to detect objects defined in an ML model from a live camera feed

From here, you could do anything! Try one of the following:

  • Capture images of your hand making specific gestures, and train a model on that data to teach your machine to recognize certain hand gestures, and respond accordingly. For example, you might train it to stop or start based on your hand gesture, to turn in a specific direction, or adjust its speed.
  • Teach your machine to recognize specific pets, and dispense treats appropriately.
  • Teach your machine to recognize specific toys, and to sing a specific song about each.
  • Try creating an object detection model to be able to identify parts of an image specifically with a bounding box.

For more ideas, check out our other tutorials.

You can also ask questions in the Community Discord and we will be happy to help.