Semi supervised learning by ddd with a sharing base function
This presentation is the property of its rightful owner.
Sponsored Links
1 / 6

Semi-supervised learning by DDD with a sharing base function PowerPoint PPT Presentation


  • 50 Views
  • Uploaded on
  • Presentation posted in: General

Semi-supervised learning by DDD with a sharing base function. - preliminary result on WDBC data. From the DDD formula. Consider all the Dirichlet distribution share a common base function (similar to what Dunson did), . Where . Affinity matrix. We choose .

Download Presentation

Semi-supervised learning by DDD with a sharing base function

An Image/Link below is provided (as is) to download presentation

Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author.While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server.


- - - - - - - - - - - - - - - - - - - - - - - - - - E N D - - - - - - - - - - - - - - - - - - - - - - - - - -

Presentation Transcript


Semi supervised learning by ddd with a sharing base function

Semi-supervised learning by DDD with a sharing base function

- preliminary result on WDBC data


Semi supervised learning by ddd with a sharing base function

From the DDD formula

Consider all the Dirichlet distribution share a common base function (similar to what Dunson did),

Where

Affinity matrix

We choose


Semi supervised learning by ddd with a sharing base function

Semi-supervised learning (transductive way)

For those data , this is the conditional posterior for .

By performing MCMC, we can get the histograms of their full posteriors given the labeled data set.


Semi supervised learning by ddd with a sharing base function

Apply the DDD to one benchmark data set – WDBC

569 data with dimensionality 32

Randomly choose a portion of data as labeled data and calculate the area under ROC (AUR) for each trial.

The number of labeled data: [20, 40, 60, 80].

For each case: 20 random trials.

MCMC iteration: 2000; Burn-in: 500


Semi supervised learning by ddd with a sharing base function

Qiuhua’s results

Note: the “accuracy” is different from “AUR”.


  • Login