Machine Learning笔记 第10周

第十周根本没时间上课,只能利用第11周的春假补全。

This week: going over Feature Transformation this week, and starting on Information Theory.

Machine Learning笔记 第10周_第1张图片
Defination of Feature Ransformation
  • Feature selection is a subset of feature transformation
  • Transformation operator is linear combinations of original features

Why do Feature Transformation

Machine Learning笔记 第10周_第2张图片
Example of words
  • XOR, Kernel methods, Neural networks already do FT.
  • ad hoc Information Retrieval Problem: finding documents within
    a corpus that are relevant to an information need specified using a query. (Query is unknown)
  • Problems of Information Retrieval:
    • Polysemy: e.g. a word have multiple meanings; cause false positive problem
    • Synonymy: e.g. a meaning can be expressed by multiple words. can cause false negatives problems.

PCA

This paper does a fantastic job building the intuition and implementation behind PCA

An eigenproblem is a computational problem that can be solved by finding the eigenvalues and/or eigenvectors of a matrix. In PCA, we are analyzing the covariance matrix (see the paper for details)

Machine Learning笔记 第10周_第3张图片
PCA

PCA Features

  • maximize variance

  • mutually orthogonal (every components are perpendicular to each other)

  • Global algorithm: the resulted components have a global constraint which is that they must be orthogonal

  • it gives best reconstruction

  • EigenValue monotonically not increasing and 0 eigenvalue = ignorable (irrelevant, maybe not useful).

  • It's well studied and fast to run.

  • it's like a classification. and using a filtering method to select dimensions to use.

  • PCA is about finding

ICA

ICA has also been applied to the information retrieval problem, in a paper written by Charles himself

Machine Learning笔记 第10周_第4张图片
ICA
Machine Learning笔记 第10周_第5张图片
  • find components that are statistically independent from each other using mutual information.
  • Designed to solve the blind source separation problem.
  • Model: given observables, find hidden variables.
Machine Learning笔记 第10周_第6张图片
quize 1: defining features for PAC and ICA
Machine Learning笔记 第10周_第7张图片
More PCA vs ICA
  • ICA is more suitable for BSS problems and is directional.
  • Eg,
    • PCA on faces will separate image based on brightness and average faces. ICA will get features such as nose, mouth etc, which are basic components of a face.

Alternatives:

Machine Learning笔记 第10周_第8张图片
RCA

Random components Analysis: generates random directions

  • Can project to smaller dimensions (m << n)but in practice often have more dimensions than PCA.
  • Can project to higher dimensions (m > n)
  • It works and works very fast.
Machine Learning笔记 第10周_第9张图片
LDA
  • Linear Discriminant analysis: find a projection that discriminates based on the label

wrap up

Machine Learning笔记 第10周_第10张图片
Wrap up

This excellent paper is a great resource for the Feature Transformation methods from this course, and beyond

2016-03-17 初稿
2016-03-26 补完

你可能感兴趣的:(Machine Learning笔记 第10周)