Some less popular concepts in CMU 10601

less than 1 minute read

  • Semi-supervised learning: mixing labeled and unlabeled data as training data (no query for labels during training); based on the belief that data has the same structure.
    • Semi-supervised SVM: find a labeling of the unlabeled sample and \(w\) s.t. \(w\) separates both labeled and unlabeled data with maximum margin.
    • Co-training: different features are consistently representing one label. Train multiple classifiers. Use the labeling result of one classifier to help the training of others.
    • Graph-based methods: construct a graph with edges between very similar examples. Examples with strong edges should have same labels.
  • Active learning:
    • use fewer labeled examples
    • algorithm chooses which examples to be labeled
    • can have sampling bias, need to be careful when choosing query strategy: like version space method.
  • Boosting:
    • Combine several weak learners to produce an overall high-accuracy predictor.
    • Steps: apply 1st weak learner -> find the most easily classified examples for the 2nd rule of thumb -> take (weighted) majority vote of all learners
    • Adaptive boosting: find the weight for the majority vote
      • Training error drops exponentially in T(number of rounds/weak learners)
      • In experiments, test error does not increase with training rounds(in constrast with Occam’s razor). We need to also consider the classification confidence.