Detect Objects Using Deep Learning (Map Viewer)

The Detect Objects Using Deep Learning tool uses a deep learning model to identify and locate objects in an imagery layer.

The output is a hosted feature layer.

Examples

The Detect Objects Using Deep Learning tool can be used to identify building footprints to upgrade property tax data for a local government or regional emergency response group. The output layer from the tool is a feature layer that can identify the buildings in an area. The feature layer created can be used to match existing property records to record the current building footprint of the property.

The Detect Objects Using Deep Learning tool can be used to identify cars in a parking lot to count attendance and prepare traffic surveys. The feature layer created can be used in the Classify Objects Using Deep Learning tool to classify the type of car detected.

Usage notes

The Detect Objects Using Deep Learning tool includes configurations for input layer, model settings, and result layer.

Input layer

The Input layer group includes the following parameters:

  • Input imagery layer or feature layer is used to select the imagery layer or feature layer with attachments that will be used to detect the objects identified in the deep learning model. The imagery layer selected should be based on the requirements for the deep learning model that will be used to classify the pixels.
  • Processing mode describes how the raster items in the imagery layer will be processed. Processing mode includes the following options:
    • Process as mosaicked image—All raster items in the mosaic dataset or image service will be mosaicked together and processed. This is the default.
    • Process all raster items separately—All raster items in the mosaic dataset or image service will be processed as separate images.

Model settings

The Model settings group includes the following parameters:

  • Model for object detection specifies the deep learning model used to detect the objects. The deep learning model needs to be located on ArcGIS Online to be selected in the tool. You can select your own model, publicly available in ArcGIS Online, or from ArcGIS Living Atlas of the World.
  • Model arguments lists additional deep learning parameters and arguments for refinement not defined in the Python raster function by the input model, such as a confidence threshold for adjusting sensitivity. The names of the arguments are populated by the tool from reading the Python module.
  • Non maximum suppression (NMS) determines whether to perform non-maximum suppression to remove duplicate objects that are identified based on confidence values.
  • Confidence score field specifies the field name that will record the confidence scores that are created as output by the object detection method. This parameter is available when Non maximum suppression (NMS) is enabled.
  • Class value field is the field in the output feature layer that will contain the value from the input imagery layer. If not specified, the tool will use the standard class value fields Classvalue and Value. If these fields do not exist, all features will be treated as the same object class. This parameter is available when Non maximum suppression (NMS) is enabled.
  • Maximum overlap ratio defines the ratio of intersection area over the union area for two overlapping features. The default value is 0.This parameter is available when Non maximum suppression (NMS) is enabled.

Result layer

The Result layer group includes the following parameters:

  • Output name determines the name of the layer that is created and added to the map. The name must be unique. If a layer with the same name already exists in your organization, the tool will fail and you will be prompted to use a different name.
  • Save in folder specifies the name of a folder in My content where the result will be saved.

Environments

Analysis environment settings are additional parameters that affect a tool's results. You can access the tool's analysis environment settings from the Environment settings parameter group.

This tool honors the following analysis environments:

Outputs

The output is a feature layer with each detected object as an individual feature with the class value and confidence fields added.

Usage requirements

This tool requires the following licensing and configurations:

Resources

Use the following resources to learn more: