Toolbar

../../_images/map_toolbar.png

Share

Allow users to share maps for other users by link.

Step 1: Create map

Step 2: Click Share button

../../_images/map_share_1.png

Step 3: Create link

Type link and description

../../_images/map_share_2.png

Step 4: Click Submit button

After submitting a link, the link will be created at tab Links. Click tab Links to check.

Step 5: Copy link and send to other users

Click icon to copy link:

../../_images/map_share_3.png

Print

Step 1: Click Print button:

../../_images/map_print_1.png

Step 2: Type input before print:

../../_images/map_print_2.png

Step 3: Click print icon

Enter the parameters:

../../_images/map_print_3.png

Step 4: Click Save button

Capture

Allow users to take screenshots of the map at that time.

Click Capture button:

../../_images/map_capture_1.png

Tools

ARD Toolkit

Click the link ARD Toolkit to see details.

Vector Toolkit

Click the link Vector Toolkit to see details.

Result Toolkit

Includes the tools for analysing results.

../../_images/result_toolkit.png

Crop Scout

Step 1: Type input

  1. Type Name

  2. Select destination folder to store result

  3. Select raster which need to crop scout

../../_images/crop_scout_1'.png

Step 2: Click Tool setting button -> Select raster band (and select vector):

../../_images/crop_scout_2'.png

Step 3: Click Confirm button

The Task will be created in Tasks. Wait for task success, check result.

Reclassification

Reclassifies (or changes) the values in a raster.

Step 1: Type input

  1. Type Name

  2. Select destination folder to store result

  3. Select raster which need to reclass

../../_images/reclassification_1'.png

Step 2: Click “Tool setting” button

  1. Click color box to change color for each class

  2. Change name for each class

  3. Change value for each class

../../_images/reclassification_2.png

Step 3: Click Apply button. The Task will be created in Tasks. Wait for task success, check result.

Morphology

This help users in post processing to eliminate the small area and fill the hole.

Step 1: Type input

  1. Type Name

  2. Select destination folder to store result

  3. Select rater which need to morphology

../../_images/morphology_1'.png

Step 2: Click “Tool setting” button -> Type epsilon

../../_images/morphology_2.png

Step 3: Click Confirm button. The Task will be created in Tasks. Wait for task success, check result.

Vectorization

This help users convert raster to vector.

Step 1: Type input

  1. Type Name

  2. Select destination folder to store result

  3. Select raster which need to vectorize

../../_images/vectorization_1'.png

Step 2: Click “Tool setting” button -> Type epsilon

../../_images/vectorization_2.png

Step 3: Click Confirm button. The Task will be created in Tasks. Wait for task success, check result.

Unsupervised

../../_images/unsupervised.png

K-Means Clustering

K-Means Clustering is a method of vector quantization, originally from signal processing, that aims to partition n observations into k clusters in which each observation belongs to the cluster with the nearest mean (cluster centers or cluster centroid), serving as a prototype of the cluster. This results in a partitioning of the data space into Voronoi cells. k-means clustering minimizes within-cluster variances (squared Euclidean distances), but not regular Euclidean distances, which would be the more difficult Weber problem: the mean optimizes squared errors, whereas only the geometric median minimizes Euclidean distances.

../../_images/unsupervised_kmean_1.png ../../_images/unsupervised_kmean_2.png
  1. Type Name

  2. Select destination folder to store result

  3. Select input image

  4. Type Number of clusters: The number of clusters to form as well as the number of centroids to generate.

  5. Select Init method: Method for initialization (K-means++ - selects initial cluster centers for k-mean clustering in a smart way to speed up convergence. See section Notes in k_init for more details.;**Random** - choose n_clusters observations (rows) at random from data for the initial centroids.)

  6. Type Number of init: Number of time the k-means algorithm will be run with different centroid seeds. The final results will be the best output of n_init consecutive runs in terms of inertia.

  7. Type Maximum iterations:Maximum number of iterations of the k-means algorithm for a single run.

  8. Type Relative tolerance: Relative tolerance with regards to Frobenius norm of the difference in the cluster centers of two consecutive iterations to declare convergence.

  9. Type Verbosity mode

  10. Type Random state: Determines random number generation for centroid initialization. Use an int to make the randomness deterministic.

  11. Checkbox Copy x: If Copy x is True (default), then the original data is not modified

  12. Select Algorithm (Auto, Full, Elkan): The classical EM-style algorithm is “full”. The “elkan” variation is more efficient by using the triangle inequality, but currently doesn’t support sparse data. “auto” chooses “elkan” for dense data and “full” for sparse data.

Isodata Clustering

ISODATA unsupervised classification calculates class means evenly distributed in the data space then iteratively clusters the remaining pixels using minimum distance techniques. Each iteration recalculates means and reclassifies pixels with respect to the new means. Iterative class splitting, merging, and deleting is done based on input threshold parameters. All pixels are classified to the nearest class unless a standard deviation or distance threshold is specified, in which case some pixels may be unclassified if they do not meet the selected criteria. This process continues until the number of pixels in each class changes by less than the selected pixel change threshold or the maximum number of iterations is reached.

../../_images/unsupervised_iso_1.png ../../_images/unsupervised_iso_2.png
  1. Type Name

  2. Select destination folder to store result

  3. Select image

  4. Type Number of clusters desired: Specifies the number of clusters (classes) desired. Note that this is only an estimate; the final number of clusters may vary. You can, however, limit the variation by setting the maximum (MAXCLUS) and minimum (MINCLUS) cluster parameters.

  5. Type Minimum number of clusters: Specifies the minimum number of clusters allowed. This parameter limits the total number of clusters allowed during lumping.

  6. Type Maximum number of clusters: Specifies the maximum number of clusters allowed. This parameter limits the total number of clusters allowed during splitting.

  7. Type Maximum number of iterations: Specifies the maximum number of iterations for calculating the cluster mean positions.

  8. Type Movement threshold: Specifies the movement threshold, in percentage of cluster means.

  9. Type Minimum sample threshold: Specifies the minimum sample threshold. The number of samples in a cluster domain is compared to the value of this parameter.

  10. Type Standard deviation: Specifies the standard deviation. If a cluster has a standard deviation greater than that specified in this parameter, splitting may occur. The default value of 10.0 is reasonable.

  11. Type Lumping parameter: Specifies the lumping parameter. If the distance between two cluster centers is less than LUMP, the total number of clusters is greater than MINCLUS, and the number of lumped pairs is less than MAXPAIR, lumping of clusters will occur.

  12. Type Maximum number of lumping pairs: Specifies the maximum number of pairs of cluster centers that can be lumped during each iteration.

  13. Type Background gray-level value: Optionally specifies a background gray-level value to be ignored during classification. If this parameter is specified, pixels with the specified value will be assigned class 0 (null class).

Fuzzy K-Means Clustering

Fuzzy K-Means is a clustering method that allows each data point to belong to multiple clusters with varying degrees of membership.

../../_images/unsupervised_fuzzy.png
  1. Type Name

  2. Select destination folder to store result

  3. Select image

  4. Type Number of cluster centers: Specifies the number of clusters (classes) centers.

  5. Type Maximum number of iterations: Specifies the maximum number of iterations for calculating the cluster mean positions.

  6. Type Movement threshold: Specifies the movement threshold, in percentage of cluster means.

  7. Type Background gray-level value: Optionally specifies a background gray-level value to be ignored during classification. If this parameter is specified, pixels with the specified value will be assigned class 0 (null class).

Table Toolkit

Click the link Table Toolkit to see details.

Identify

Allow to view attribute value for a feature or pixel value for each band of image.

Click Identify button:

../../_images/map_identify_1.png

Click on a feature or point image to show value.

Training

  • Step 1: Click “Training” button in Maps

../../_images/map_training_1.png
  • Step 2: Type input

  1. Training model name

  2. Select 1 of 4 types:

  • Classification (SK-CL-01)

  • Classification SK-CL-02-SAR)

  • Edge detection (SK-ED-01)

  • Object detection (SK-OD-01)

  • Multi Object Detection (SK-MD-01)

  1. Each type will have the config to adjust the model with data. Changing the config click on setting icon

  • Classification (SK-CL-01): There are 7 functions which can change: Epoch, Training size, N Filters, Batch normalization, Dropout, Optimizer, Loss Function, Metrics.

    • Epoch: a hyperparameter that defines the number times that the learning algorithm will work through the entire training dataset

    • Training size: the size of images which are cropped from the original images.

    • N Filters: is the number of kernel

    • Batch Normalization: this method normalize the features (the output of each layer after going through the activations) to a zero-means state with a standard deviation of 1

    • Dropout: A simple and powerful regularization technique

    • Optimizer: Optimizers are algorithms or methods used to change the attributes of your neural network such as weights and learning rate in order to reduce the losses.

      • Adam optimizer: An algorithm for first-order gradient-based optimization of stochastic objective functions, based on adaptive estimates of lower-order moments

      • Adadelta optimizer: is a more robust extension of Adagrad that adapts learning rates based on a moving window of gradient updates,instead of accumulating all past gradients

      • Stochastic gradient descent optimizer: An iterative method for optimizing an objective function with suitable smoothness properties

      • Adagrad optimizer: an optimizer with parameter-specific learning rates, which are adapted relative to how frequently a parameter gets updated during training

    • Loss Function: Neural networks are trained using stochastic gradient descent and require that you choose a loss function when designing and configuring your model.

      • Categorical cross-entropy: is used for single label categorization. This is when only one category is applicable for each data point. In other words, an example can belong to one class only.

      • Binary cross-entropy: is used on yes/no decisions, e.g., multi-label classification. The loss tells you how wrong your model’s predictions are. For instance, in multi-label problems, where an example can belong to multiple classes at the same time, the model tries to decide for each class whether the example belongs to that class or not.

      • Mean squared error: is the most commonly used loss function for regression. The loss is the mean overseen data of the squared differences between true and predicted values, or writing it as a formula.

    • Metrics: metric for evaluating classification models

      • Accuracy: is the fraction of predictions our model got right. the percentage of correct predictions in total number predictions

      • Binary Accuracy: can also be calculated in terms of positives and negatives. It calculates the percentage of predicted values that match with actual values for binary labels.

      • Categorical Accuracy: calculates the percentage of predicted values that match with actual values for one-hot labels.

  • Classification (SK-CL-02-SAR): There are 2 functions: Epoch and Training size are same above.

  • Edge detection (SK-ED-01): the config has only one option:Epoch

  • Object detection (SK-OD-01): There are 10 hyperparameter to change: Training size, RPN anchor scale, Epoch 1, Epoch 2, Epoch 3, Step per epoch, Validation steps, Max ground truth instances, ROI positive ratio, RPN NMS ratio

    • Training size: the size of images which are cropped from the original images.

    • RPN anchor scale: Length of square anchor side in pixels

    • Epoch 1, 2, 3: the number times that the learning algorithm will work through the entire training dataset

    • Step per epoch: Number of training steps per epoch. This doesn’t need to match the size of the training set. Tensorboard updates are saved at the end of each epoch, so setting this to a smaller number means getting more frequent TensorBoard updates. Validation stats are also calculated at each epoch end and they might take a while, so don’t set this too small to avoid spending a lot of time on validation stats.

    8 Validation steps: Number of validation steps to run at the end of every training epoch. A bigger number improves accuracy of validation stats, but slows down the training. * Max ground truth instances: Maximum number of ground truth instances to use in one cut-image (with size equal training size) * ROI positive ratio: Percent of positive ROIs - Region of interest- used to train classifier/mask heads. * RPN NMS ratio: Non-max suppression threshold to filter RPN proposals.

  • Multi Object Detection (SK-MD-01): There are 11 hyperparameter: Epoch 1, Epoch 2, Anchor box scales, Anchor box ratio, RPN stride, Number RPN predict, RPN min overlap, RPN max overlap, Classifier min overlap, Classifier max overlap, Pooling regions.

    • Epoch 1, Epoch 2: a hyperparameter that defines the number times that the learning algorithm will work through the entire training dataset

    • Anchor box scales: the length of box’s side

    • Anchor box ratio: the ratio between length and width.

    • RPN stride: modifies the amount of movement’s anchor box over the image

    • Number RPN predict: number of boxes will be taken in each stride step.

    • RPN min overlap: is threshold to select the negatives

    • RPN max overlap:is threshold to select the positives.

    • Classifier min overlap:is threshold to select the negatives

    • Classifier max overlap: is threshold to select the positives.

    • Pooling regions: Region of interest pooling to convert the features inside any valid region of interest into a small feature map with a fixed spatial extent

  1. Click Continue button

  2. Select image: select image which you want to train

Type name into search input to search image -> Click image. Click location icon to add image to map.

../../_images/map_training_2.png
  1. Click Continue button

  2. Training area layer: select exist layer or draw new one

../../_images/map_training_3.png ../../_images/map_training_4.png ../../_images/map_training_5.png ../../_images/map_training_6.png ../../_images/map_training_7.png
  • Step 3: Click Run button

../../_images/map_training_8.png

System creates tasks that train models. When the Task is successful, you will see your model in the Model Menu

Detection

  • Step 1: Click Detection button in Maps

../../_images/map_detection_1.png
  • Step 2: Select input

  1. Select image you want to detect

  2. Select model you want to use

../../_images/map_detection_2.png
  • Step 3: Click Continue button

  • Step 4: Type name

  • Step 5: Click Run button

../../_images/map_detection_3.png

System creates tasks that detect objects. When a Task is successful, you will see your result in the Result Menu.

Dataset

Allow users to create training dataset.

For Object detection: only select an annotation. As for Classification detection: select many annotations.

Step 1: Click Dataset button

../../_images/map_dataset_1.png

Step 2: Type input

../../_images/map_dataset_2.png

Step 3: Click Submit button

Accuracy

Allow users to check accuracy of model

  • Step 1: Click Accuracy button

../../_images/map_accuracy_1.png
  • Step 2: Type accurate report input

  1. Type report name

  2. Select model

  3. Select image

../../_images/map_accuracy_2.png
  • Step 3: Click Continue button

  • Step 4: Validate data: Select exist layer or create new one

../../_images/map_accuracy_3.png
  • Step 5: Click Run button

After clicking, a task will be created. When a task is successful, check the result in the menu Model. * Step 6: Check result in Models: Model detail of selected model -> Tab Accurate Report