standardization

In machine learning, standardization is a feature engineering technique by which the dataset features are re-scaled to achieve zero-mean value (μ=0) and unit standard deviation value (σ=1). Each x value in the dataset gets a corresponding x' standardized value, which is calculated as follows. , where μ is the x variable mean and σ is ... Read more

stratified cross validation

Stratified cross validation is a data validation technique used when splitting the ML dataset into k subsets, of which k-1 subsets are used as training subsets (folds) and one (1) is used as the test subset (fold). This process is repeated k times. Stratified cross validation uses stratified sampling in the dataset, in order to ... Read more

Tensor flow

Tensor flow TensorFlow is an open-source machine learning (ML) software library. Tensorflow was originally created by the Google Brain Team in 2015. The main purpose of Tensorflow is building and training neural networks, by using a variety of supported languages, the primary one being Python. To discover the Tensorflow customizable pre-built models, visit its official ... Read more

Transformer machine learning model

Transformer machine learning model A transformer is a deep learning model. Transformer models are mainly used for natural language processing (NLP) and computer vision (CV). Transformers are the evolution of RNN models. A recent example of tranformer-type models in artificial intelligence (AI) are the dp-tranformer models developed by Microsoft research. More details can be found ... Read more