Movatterモバイル変換


[0]ホーム

URL:


Skip to main content

This browser is no longer supported.

Upgrade to Microsoft Edge to take advantage of the latest features, security updates, and technical support.

Download Microsoft EdgeMore info about Internet Explorer and Microsoft Edge
Table of contentsExit focus mode

Machine learning glossary of important terms

  • 2021-09-15
Feedback

In this article

The following list is a compilation of important machine learning terms that are useful as you build your custom models in ML.NET.

Accuracy

Inclassification, accuracy is the number of correctly classified items divided by the total number of items in the test set. Ranges from 0 (least accurate) to 1 (most accurate). Accuracy is one of evaluation metrics of the model performance. Consider it in conjunction withprecision,recall, andF-score.

Area under the curve (AUC)

Inbinary classification, an evaluation metric that is the value of the area under the curve that plots the true positives rate (on the y-axis) against the false positives rate (on the x-axis). Ranges from 0.5 (worst) to 1 (best). Also known as the area under the ROC curve, i.e., receiver operating characteristic curve. For more information, see theReceiver operating characteristic article on Wikipedia.

Binary classification

Aclassification case where thelabel is only one out of two classes. For more information, see theBinary classification section of theMachine learning tasks topic.

Calibration

Calibration is the process of mapping a raw score onto a class membership, for binary and multiclass classification. Some ML.NET trainers have aNonCalibrated suffix. These algorithms produce a raw score that then must be mapped to a class probability.

Catalog

In ML.NET, a catalog is a collection of extension functions, grouped by a common purpose.

For example, each machine learning task (binary classification, regression, ranking etc) has a catalog of available machine learning algorithms (trainers). The catalog for the binary classification trainers is:BinaryClassificationCatalog.BinaryClassificationTrainers.

Classification

When the data is used to predict a category,supervised machine learning task is called classification.Binary classification refers to predicting only two categories (for example, classifying an image as a picture of either a 'cat' or a 'dog').Multiclass classification refers to predicting multiple categories (for example, when classifying an image as a picture of a specific breed of dog).

Coefficient of determination

Inregression, an evaluation metric that indicates how well data fits a model. Ranges from 0 to 1. A value of 0 means that the data is random or otherwise cannot be fit to the model. A value of 1 means that the model exactly matches the data. This is often referred to as r2, R2, or r-squared.

Data

Data is central to any machine learning application. In ML.NET data is represented byIDataView objects. Data view objects:

  • are made up of columns and rows
  • are lazily evaluated, that is they only load data when an operation calls for it
  • contain a schema that defines the type, format and length of each column

Estimator

A class in ML.NET that implements theIEstimator<TTransformer> interface.

An estimator is a specification of a transformation (both data preparation transformation and machine learning model training transformation). Estimators can be chained together into a pipeline of transformations. The parameters of an estimator or pipeline of estimators are learned whenFit is called. The result ofFit is aTransformer.

Extension method

A .NET method that is part of a class but is defined outside of the class. The first parameter of an extension method is a staticthis reference to the class to which the extension method belongs.

Extension methods are used extensively in ML.NET to construct instances ofestimators.

Feature

A measurable property of the phenomenon being measured, typically a numeric (double) value. Multiple features are referred to as aFeature vector and typically stored asdouble[]. Features define the important characteristics of the phenomenon being measured. For more information, see theFeature article on Wikipedia.

Feature engineering

Feature engineering is the process that involves defining a set offeatures and developing software that produces feature vectors from available phenomenon data, i.e., feature extraction. For more information, see theFeature engineering article on Wikipedia.

F-score

Inclassification, an evaluation metric that balancesprecision andrecall.

Hyperparameter

A parameter of a machine learning algorithm. Examples include the number of trees to learn in a decision forest or the step size in a gradient descent algorithm. Values ofHyperparameters are set before training the model and govern the process of finding the parameters of the prediction function, for example, the comparison points in a decision tree or the weights in a linear regression model. For more information, see theHyperparameter article on Wikipedia.

Label

The element to be predicted with the machine learning model. For example, the breed of dog or a future stock price.

Log loss

Inclassification, an evaluation metric that characterizes the accuracy of a classifier. The smaller log loss is, the more accurate a classifier is.

Loss function

A loss function is the difference between the training label values and the prediction made by the model. The parameters of the model are estimated by minimizing the loss function.

Different trainers can be configured with different loss functions.

Mean absolute error (MAE)

Inregression, an evaluation metric that is the average of all the model errors, where model error is the distance between the predictedlabel value and the correct label value.

Model

Traditionally, the parameters for the prediction function. For example, the weights in a linear regression model or the split points in a decision tree. In ML.NET, a model contains all the information necessary to predict thelabel of a domain object (for example, image or text). This means that ML.NET models include the featurization steps necessary as well as the parameters for the prediction function.

Multiclass classification

Aclassification case where thelabel is one out of three or more classes. For more information, see theMulticlass classification section of theMachine learning tasks topic.

N-gram

A feature extraction scheme for text data: any sequence of N words turns into afeature value.

Normalization

Normalization is the process of scaling floating point data to values between 0 and 1. Many of the training algorithms used in ML.NET require input feature data to be normalized. ML.NET provides a series oftransforms for normalization.

Numerical feature vector

Afeature vector consisting only of numerical values. This is similar todouble[].

Pipeline

All of the operations needed to fit a model to a data set. A pipeline consists of data import, transformation, featurization, and learning steps. Once a pipeline is trained, it turns into a model.

Precision

Inclassification, the precision for a class is the number of items correctly predicted as belonging to that class divided by the total number of items predicted as belonging to the class.

Recall

Inclassification, the recall for a class is the number of items correctly predicted as belonging to that class divided by the total number of items that actually belong to the class.

Regularization

Regularization penalizes a linear model for being too complicated. There are two types of regularization:

  • $L_1$ regularization zeros weights for insignificant features. The size of the saved model might become smaller after this type of regularization.
  • $L_2$ regularization minimizes weight range for insignificant features. This is a more general process and is less sensitive to outliers.

Regression

Asupervised machine learning task where the output is a real value, for example, double. Examples include predicting stock prices. For more information, see theRegression section of theMachine learning tasks topic.

Relative absolute error

Inregression, an evaluation metric that is the sum of all absolute errors divided by the sum of distances between correctlabel values and the average of all correct label values.

Relative squared error

Inregression, an evaluation metric that is the sum of all squared absolute errors divided by the sum of squared distances between correctlabel values and the average of all correct label values.

Root of mean squared error (RMSE)

Inregression, an evaluation metric that is the square root of the average of the squares of the errors.

Scoring

Scoring is the process of applying new data to a trained machine learning model, and generating predictions. Scoring is also known as inferencing. Depending on the type of model, the score can be a raw value, a probability, or a category.

Supervised machine learning

A subclass of machine learning in which a desired model predicts the label for yet-unseen data. Examples include classification, regression, and structured prediction. For more information, see theSupervised learning article on Wikipedia.

Training

The process of identifying amodel for a given training data set. For a linear model, this means finding the weights. For a tree, it involves identifying the split points.

Transformer

An ML.NET class that implements theITransformer interface.

A transformer transforms oneIDataView into another. A transformer is created by training anestimator, or an estimator pipeline.

Unsupervised machine learning

A subclass of machine learning in which a desired model finds hidden (or latent) structure in data. Examples include clustering, topic modeling, and dimensionality reduction. For more information, see theUnsupervised learning article on Wikipedia.

Collaborate with us on GitHub
The source for this content can be found on GitHub, where you can also create and review issues and pull requests. For more information, seeour contributor guide.

Feedback

Was this page helpful?

YesNo

In this article

Was this page helpful?

YesNo