Talk:Semi-supervised learning

From Wikipedia, the free encyclopedia

What is

What is the difference between ‘transductive learning’ and ‘semi-supervised learning’?

Both use a mix of labeled and unlabeled examples, but the performance of the former is measured only on the unlabeled examples; it is a finite task. The performance of the latter is measure in the same way as in supervised learning, that is the expected performance wrt the distribution of the population the training set is sampled from. See also https://mitpress.mit.edu/sites/default/files/titles/content/9780262033589_sch_0001.pdf 1.2.4. The Wikipedia entry takes the alternative view that semisupervised includes transductive and inductive learning. The latter is clearly wrong as fully supervised learning is also inductive (aiming for expected good performance over a potentially infinite population. Personally I see transductive as a special case of inductive, where the full (finite) population is specified by enumeration, and a subset thereof is labelled. There is nothing in inductive learning that requires populations to be infinite, it's just a very common case.  Preceding unsigned comment added by Piccolbo (talkcontribs) 03:10, 31 August 2017 (UTC)

Self-taught learning

The paper "Self-taught learning: transfer learning from unlabeled data" by Raina et al. presents self-taught learning. It uses unlabeled data to improve predictions made via supervised learning. However, in contrary to other semi-supervised methods, it does not make the assumption that the unlabeled data's actual classes correspond to the ones given in the labeled data set. Instead, only higher-level features are extracted from the unlabeled data.

Would this approach still be considered semi-supervised learning? — Preceding unsigned comment added by 188.74.81.25 (talk) 00:10, 19 May 2012 (UTC)

Generative models

Merge with Weak supervision?

Move discussion in progress

Related Articles

Wikiwand AI