Dioptra Documentation
  • What is KatiML ?
  • Overview
    • πŸƒGetting Started
    • 🌊KatiML
      • Quick start
      • Ingestion basics
      • Ingestion SDK
      • Query basics
      • Query SDK
      • Dataset basics
      • Dataset SDK
      • Supported fields
      • Matching local data with Kati ML IDs
      • Managing Datapoints with Tags
      • Configuring Object Stores (optional)
    • 🧠Active Learning
      • πŸ“–Miners basics
      • ⛏️Miners SDK
      • πŸš—[Experimental] Mining on the edge
    • πŸ€–PyTorch and Tensorflow integrations
      • Tensorflow
      • PyTorch
  • 😬Enough docs, show me some code !
  • πŸ“‘Case studies
  • Definitions
Powered by GitBook
On this page
  • PreviousCase studiesπŸ“š Datapoints, Predictions, and Ground Truths
  • πŸ“š Machine Learning Calculations

Was this helpful?

Definitions

This section is for to clarify some terms that are used throughout the documentation.

PreviousCase studies

Last updated 2 years ago

Was this helpful?

πŸ“š Datapoints, Predictions, and Ground Truths

Datapoints are the inputs to your model. They are the data that you use to make predictions. For example, if you are building a model to classify images, the datapoints are the images themselves.

Predictions are the outputs of your model. They are the results of your model's predictions. For example, if you are building a model to classify images, the predictions are the labels that your model assigns to the images.

Ground truths are the labels for your datapoints. For example, if you are building a model to classify images, the ground truths are the labels that you know to be correct for the images.

πŸ“š Machine Learning Calculations

Entropy is a measure of uncertainty. The higher the entropy, the more uncertain the model is about its prediction. The lower the entropy, the more certain the model is about its prediction. For our calculations of entropy we utilize . This differs from metric entropy, which is a normalized version of Shannon entropy.

Variance is a measure of the spread of a distribution. The higher the variance, the more spread out the distribution is. The lower the variance, the more concentrated the distribution is. This metric is used with models that have dropout layers. The specific equation for variance we use is the one for sample variance. You can read more about it .

PreviousCase studies
Shannon entropy
here