text
stringlengths
0
3.77k
Feature engineering
Attention mechanism
Singular value decomposition
Softmax function
Adversarial example
Adaboost
Activation function
Neural network
Multi-task learning
Loss function
Image segmentation
Inference
Unsupervised learning
Adversarial attack
Cross-validation
Convolutional neural network
Bias-variance tradeoff
Semi-supervised learning
Hyperparameter
Transfer learning
Ensemble learning
Deep learning
Instance-based learning
Alpha
Data augmentation
Weight initialization
Support vector machine
Evolutionary algorithm
Learning rate
Bag of words
Precision and recall
Zero-shot learning
Autoencoder
Backpropagation
Reinforcement learning
Active learning
Feedforward neural network
Gradient descent
Bayesian optimization
Label
Linear regression
Overfitting
Variational autoencoder
Embedding
Artificial intelligence
K-nearest neighbors
Capsule network
Federated learning
Principal component analysis
Time series analysis
Dropout
Clustering
Fine-tuning
Decision tree
Regression analysis
Validation
Sampling
Machine learning
Data preprocessing
Feature engineering or feature extraction or feature discovery is the process of extracting features (characteristics, properties, attributes) from raw data
Machine learning-based attention is a mechanism mimicking cognitive attention
In linear algebra, the singular value decomposition (SVD) is a factorization of a real or complex matrix
The softmax function, also known as softargmax or normalized exponential function, converts a vector of K real numbers into a probability distribution of K possible outcomes
Adversarial machine learning is the study of the attacks on machine learning algorithms, and of the defenses against such attacks
AdaBoost, short for Adaptive Boosting, is a statistical classification meta-algorithm formulated by Yoav Freund and Robert Schapire in 1995, who won the 2003 G del Prize for their work
Activation function of a node in an artificial neural network is a function that calculates the output of the node (based on its inputs and the weights on individual inputs)
Artificial neural networks (ANNs, also shortened to neural networks (NNs) or neural nets) are a branch of machine learning models that are built using principles of neuronal organization discovered by connectionism in the biological neural networks constituting animal brains.An ANN is based on a collection of connected units or nodes called artificial neurons, which loosely model the neurons in a biological brain
Multi-task learning (MTL) is a subfield of machine learning in which multiple learning tasks are solved at the same time, while exploiting commonalities and differences across tasks
In mathematical optimization and decision theory, a loss function or cost function (sometimes also called an error function) is a function that maps an event or values of one or more variables onto a real number intuitively representing some "cost" associated with the event
In digital image processing and computer vision, image segmentation is the process of partitioning a digital image into multiple image segments, also known as image regions or image objects (sets of pixels)
A reference is a relationship between objects in which one object designates, or acts as a means by which to connect to or link to, another object
Supervised learning (SL) is a paradigm in machine learning where input objects (for example, a vector of predictor variables) and a desired output value (also known as human-labeled supervisory signal) train a model
Adversarial machine learning is the study of the attacks on machine learning algorithms, and of the defenses against such attacks
Cross-validation, sometimes called rotation estimation or out-of-sample testing, is any of various similar model validation techniques for assessing how the results of a statistical analysis will generalize to an independent data set
Convolutional neural network (CNN) is a regularized type of feed-forward neural network that learns feature engineering by itself via filters (or kernel) optimization
In statistics and machine learning, the bias variance tradeoff describes the relationship between a model's complexity, the accuracy of its predictions, and how well it can make predictions on previously unseen data that were not used to train the model
Weak supervision, also called semi-supervised learning, is a paradigm in machine learning, the relevance and notability of which increased with the advent of large language models due to large amount of data required to train them
In machine learning, hyperparameter optimization or tuning is the problem of choosing a set of optimal hyperparameters for a learning algorithm
Transfer learning (TL) is a technique in machine learning (ML) in which knowledge learned from a task is re-used in order to boost performance on a related task
In statistics and machine learning, ensemble methods use multiple learning algorithms to obtain better predictive performance than could be obtained from any of the constituent learning algorithms alone
Deep learning is the subset of machine learning methods which are based on artificial neural networks with representation learning
In machine learning, instance-based learning (sometimes called memory-based learning) is a family of learning algorithms that, instead of performing explicit generalization, compare new problem instances with instances seen in training, which have been stored in memory
Alpha (uppercase , lowercase ; Ancient Greek: , lpha, or Greek: , romanized: lfa) is the first letter of the Greek alphabet
Data augmentation is a technique in machine learning used to reduce overfitting when training a machine learning model, by training models on several slightly-modified copies of existing data
In machine learning, the vanishing gradient problem is encountered when training artificial neural networks with gradient-based learning methods and backpropagation
In machine learning, support vector machines (SVMs, also support vector networks) are supervised learning models with associated learning algorithms that analyze data for classification and regression analysis
In computational intelligence (CI), an evolutionary algorithm (EA) is a subset of evolutionary computation, a generic population-based metaheuristic optimization algorithm
In machine learning and statistics, the learning rate is a tuning parameter in an optimization algorithm that determines the step size at each iteration while moving toward a minimum of a loss function
The bag-of-words model is a model of text represented as an unordered collection of words
In pattern recognition, information retrieval, object detection and classification (machine learning), precision and recall are performance metrics that apply to data retrieved from a collection, corpus or sample space
Zero-shot learning (ZSL) is a problem setup in deep learning where, at test time, a learner observes samples from classes which were not observed during training, and needs to predict the class that they belong to
An autoencoder is a type of artificial neural network used to learn efficient codings of unlabeled data (unsupervised learning)
As a machine-learning algorithm, backpropagation performs a backward pass to adjust a neural network model's parameters, aiming to minimize error
Reinforcement learning (RL) is an area of machine learning concerned with how intelligent agents ought to take actions in an environment in order to maximize the notion of cumulative reward
Active learning is "a method of learning in which students are actively or experientially involved in the learning process and where there are different levels of active learning, depending on student involvement." Bonwell & Eison (1991) states that "students participate [in active learning] when they are doing something besides passively listening." According to Hanson and Moser (2003) using active teaching techniques in the classroom can create better academic outcomes for students
A feedforward neural network (FNN) is one of the two broad types of artificial neural network, characterized by direction of the flow of information between its layers
In mathematics, gradient descent (also often called steepest descent) is a first-order iterative optimization algorithm for finding a local minimum of a differentiable function
Bayesian optimization is a sequential design strategy for global optimization of black-box functions that does not assume any functional forms
A label (as distinct from signage) is a piece of paper, plastic film, cloth, metal, or other material affixed to a container or product, on which is written or printed information or symbols about the product or item
In statistics, linear regression is a linear approach for modelling the relationship between a scalar response and one or more explanatory variables (also known as dependent and independent variables)

No dataset card yet

Downloads last month
39