Command Palette
Search for a command to run...
Wiki
Machine Learning Glossary: Explore definitions and explanations of key AI and ML concepts
Search for a command to run...
Machine Learning Glossary: Explore definitions and explanations of key AI and ML concepts
Search for a command to run...
Machine Learning Glossary: Explore definitions and explanations of key AI and ML concepts
Ensemble learning is the idea of combining multiple models into a high-precision model. It is mainly used in the field of machine learning. It is not a single machine learning algorithm, but a learning task completed by building and combining multiple learners. Ensemble learning can be used for classification problems, regression problems, feature selection, outlier detection, etc. It can be said that all machine learning […]
The error-correcting output coding method ECOC can convert multi-class problems into multiple two-class problems, and the error-correcting output code itself has error correction capabilities, which can improve the prediction accuracy of supervised learning algorithms. Output category encoding can convert multi-class problems into two categories, that is, each category corresponds to a binary bit string of length n, forming a total of m codewords, which […]
Empirical risk shows the model's ability to predict training samples. It is obtained by calculating the loss function once for all training samples and then accumulating and averaging. The loss function is the basis of expected risk, empirical risk, and structural risk. The loss function is for a single specific sample and represents the gap between the model's predicted value and the true value. […]
K-means clustering is a vector quantization method that was used in signal processing in the early days. It is currently mainly used as a clustering analysis method in the field of data mining. The purpose of k-means clustering is to divide n points into k clusters so that each point belongs to the cluster corresponding to the nearest mean, and use this as the clustering criterion. This type of problem […]
Margin theory is a concept in support vector machines, where the margin refers to the minimum distance between two types of samples divided by a hyperplane. Margin theory can be used to explain that when the training error of the AdaBoost algorithm is 0, continuing training can further improve the generalization performance of the model. Let x and y represent the input and […]
The perceptron is a binary linear classification model that can be regarded as the simplest form of a feedforward neural network, a model invented by Frank Rosenblatt in 1957. Its input is the feature vector of an instance, and its output is the category of the instance.
Normalization is to map data to a specified range to remove the dimensions and dimensional units of data of different dimensions to improve the comparability between different data indicators.
The proximal gradient method (PGD) is a special gradient descent method, which is mainly used to solve optimization problems with non-differentiable objective functions.
Post-pruning refers to the pruning operation performed after the decision tree is generated.
A probabilistic graphical model is a probabilistic model that uses a graph structure to express the relationship between variables.
Regression is a supervised learning algorithm for predicting and modeling numerical continuous random variables.
Rule learning is to learn a set of IF-THEN rules consisting of atomic propositions from training data. It is a type of unsupervised learning and is often classified as a type of classification.
The root node is the first node in the tree data structure. A normal node may have a parent node and child nodes, but since the root note is the first node, it only has child nodes.
Particle Swarm Optimization (PSO), also known as particle swarm optimization, is an optimization algorithm based on swarm intelligence theory. The particles in the swarm complete the optimization process of the problem in each iterative search process.
The rule engine evolved from the inference engine and is a component embedded in the application. It separates business decisions from application code and writes business decisions using predefined semantic modules.
The nuclear norm is the sum of the singular values of a matrix and is used to constrain the low rank of the matrix.
Association analysis is to find frequent patterns, associations, correlations or causal structures between items or object sets in transaction data, relational data or other information carriers. Association analysis method Apriori algorithm Apriori algorithm is a basic algorithm for mining frequent item sets required to generate Boolean association rules. It makes […]
An individual learner is a relative concept. It is the learner before integration in ensemble learning. According to the generation method of individual learners, ensemble learning methods can be divided into the following two categories: There is a strong dependency relationship, and the serialization method must be generated serially, such as Boosting; There is no strong dependency relationship, and the learners can be generated simultaneously and in parallel.
Induction is a process of reasoning that generalizes a general principle from a series of specific facts. Mathematical induction refers to a way of thinking that generalizes general concepts, principles or conclusions from multiple things. Induction can be divided into complete induction and incomplete induction: Complete induction: includes all objects of this type, thus making a conclusion about this type of object.
Inductive learning is a method of machine learning, which is usually used for symbolic learning. It mainly summarizes a concept description by giving a series of known positive and negative examples about a concept. Inductive learning can obtain new concepts, create new rules, and discover new theories. Its general operation is generalization and specialization, where generalization refers to the expansion of […]
The cumulative error back propagation algorithm (ABP algorithm) is a variant of the standard back propagation (BP) algorithm. If an update rule based on minimizing the cumulative error is derived, the cumulative error back propagation algorithm is obtained.
The hinge loss function is shaped like a hinge, which is why it is named. This loss function mainly exists in support vector machines. It requires not only correct classification, but also high confidence that the loss is 0, that is, the hinge loss function has higher requirements for learning. The formula of the hinge loss function is L ( y ( w * x […]
Hybrid computing is a kind of integrated computing, which includes traditional hard computing and emerging soft computing. This computing method can take the advantages of each to overcome the limitations. The main characteristics of hard computing: It is easy to establish a standard mathematical model of the problem; The established mathematical model is easy to solve and can achieve a high degree of accuracy; It has good stability. Soft […]
Gaussian kernel function is a commonly used kernel function that can map finite-dimensional data to high-dimensional space. The Gaussian kernel function is defined as follows: $latex {k{ \left( {x,x\text{'}} \right) }\text{ }=\text{ }e\m […]
Ensemble learning is the idea of combining multiple models into a high-precision model. It is mainly used in the field of machine learning. It is not a single machine learning algorithm, but a learning task completed by building and combining multiple learners. Ensemble learning can be used for classification problems, regression problems, feature selection, outlier detection, etc. It can be said that all machine learning […]
The error-correcting output coding method ECOC can convert multi-class problems into multiple two-class problems, and the error-correcting output code itself has error correction capabilities, which can improve the prediction accuracy of supervised learning algorithms. Output category encoding can convert multi-class problems into two categories, that is, each category corresponds to a binary bit string of length n, forming a total of m codewords, which […]
Empirical risk shows the model's ability to predict training samples. It is obtained by calculating the loss function once for all training samples and then accumulating and averaging. The loss function is the basis of expected risk, empirical risk, and structural risk. The loss function is for a single specific sample and represents the gap between the model's predicted value and the true value. […]
K-means clustering is a vector quantization method that was used in signal processing in the early days. It is currently mainly used as a clustering analysis method in the field of data mining. The purpose of k-means clustering is to divide n points into k clusters so that each point belongs to the cluster corresponding to the nearest mean, and use this as the clustering criterion. This type of problem […]
Margin theory is a concept in support vector machines, where the margin refers to the minimum distance between two types of samples divided by a hyperplane. Margin theory can be used to explain that when the training error of the AdaBoost algorithm is 0, continuing training can further improve the generalization performance of the model. Let x and y represent the input and […]
The perceptron is a binary linear classification model that can be regarded as the simplest form of a feedforward neural network, a model invented by Frank Rosenblatt in 1957. Its input is the feature vector of an instance, and its output is the category of the instance.
Normalization is to map data to a specified range to remove the dimensions and dimensional units of data of different dimensions to improve the comparability between different data indicators.
The proximal gradient method (PGD) is a special gradient descent method, which is mainly used to solve optimization problems with non-differentiable objective functions.
Post-pruning refers to the pruning operation performed after the decision tree is generated.
A probabilistic graphical model is a probabilistic model that uses a graph structure to express the relationship between variables.
Regression is a supervised learning algorithm for predicting and modeling numerical continuous random variables.
Rule learning is to learn a set of IF-THEN rules consisting of atomic propositions from training data. It is a type of unsupervised learning and is often classified as a type of classification.
The root node is the first node in the tree data structure. A normal node may have a parent node and child nodes, but since the root note is the first node, it only has child nodes.
Particle Swarm Optimization (PSO), also known as particle swarm optimization, is an optimization algorithm based on swarm intelligence theory. The particles in the swarm complete the optimization process of the problem in each iterative search process.
The rule engine evolved from the inference engine and is a component embedded in the application. It separates business decisions from application code and writes business decisions using predefined semantic modules.
The nuclear norm is the sum of the singular values of a matrix and is used to constrain the low rank of the matrix.
Association analysis is to find frequent patterns, associations, correlations or causal structures between items or object sets in transaction data, relational data or other information carriers. Association analysis method Apriori algorithm Apriori algorithm is a basic algorithm for mining frequent item sets required to generate Boolean association rules. It makes […]
An individual learner is a relative concept. It is the learner before integration in ensemble learning. According to the generation method of individual learners, ensemble learning methods can be divided into the following two categories: There is a strong dependency relationship, and the serialization method must be generated serially, such as Boosting; There is no strong dependency relationship, and the learners can be generated simultaneously and in parallel.
Induction is a process of reasoning that generalizes a general principle from a series of specific facts. Mathematical induction refers to a way of thinking that generalizes general concepts, principles or conclusions from multiple things. Induction can be divided into complete induction and incomplete induction: Complete induction: includes all objects of this type, thus making a conclusion about this type of object.
Inductive learning is a method of machine learning, which is usually used for symbolic learning. It mainly summarizes a concept description by giving a series of known positive and negative examples about a concept. Inductive learning can obtain new concepts, create new rules, and discover new theories. Its general operation is generalization and specialization, where generalization refers to the expansion of […]
The cumulative error back propagation algorithm (ABP algorithm) is a variant of the standard back propagation (BP) algorithm. If an update rule based on minimizing the cumulative error is derived, the cumulative error back propagation algorithm is obtained.
The hinge loss function is shaped like a hinge, which is why it is named. This loss function mainly exists in support vector machines. It requires not only correct classification, but also high confidence that the loss is 0, that is, the hinge loss function has higher requirements for learning. The formula of the hinge loss function is L ( y ( w * x […]
Hybrid computing is a kind of integrated computing, which includes traditional hard computing and emerging soft computing. This computing method can take the advantages of each to overcome the limitations. The main characteristics of hard computing: It is easy to establish a standard mathematical model of the problem; The established mathematical model is easy to solve and can achieve a high degree of accuracy; It has good stability. Soft […]
Gaussian kernel function is a commonly used kernel function that can map finite-dimensional data to high-dimensional space. The Gaussian kernel function is defined as follows: $latex {k{ \left( {x,x\text{'}} \right) }\text{ }=\text{ }e\m […]