search query: @keyword dimensionality reduction / total: 9
reference: 5 / 9
« previous | next »
Author:Amid, Ehsan
Title:Application of ⍺-Divergence for Stochastic Neighbor Embedding in Data Visualization
Publication type:Master's thesis
Publication year:2014
Pages:66 + 2      Language:   eng
Department/School:Perustieteiden korkeakoulu
Main subject:Machine Learning and Data Mining   (SCI3015)
Supervisor:Oja, Erkki
Instructor:Dikmen, Onur
Electronic version URL: http://urn.fi/URN:NBN:fi:aalto-201408292560
OEVS:
Electronic archive copy is available via Aalto Thesis Database.
Instructions

Reading digital theses in the closed network of the Aalto University Harald Herlin Learning Centre

In the closed network of Learning Centre you can read digital and digitized theses not available in the open network.

The Learning Centre contact details and opening hours: https://learningcentre.aalto.fi/en/harald-herlin-learning-centre/

You can read theses on the Learning Centre customer computers, which are available on all floors.

Logging on to the customer computers

  • Aalto University staff members log on to the customer computer using the Aalto username and password.
  • Other customers log on using a shared username and password.

Opening a thesis

  • On the desktop of the customer computers, you will find an icon titled:

    Aalto Thesis Database

  • Click on the icon to search for and open the thesis you are looking for from Aaltodoc database. You can find the thesis file by clicking the link on the OEV or OEVS field.

Reading the thesis

  • You can either print the thesis or read it on the customer computer screen.
  • You cannot save the thesis file on a flash drive or email it.
  • You cannot copy text or images from the file.
  • You cannot edit the file.

Printing the thesis

  • You can print the thesis for your personal study or research use.
  • Aalto University students and staff members may print black-and-white prints on the PrintingPoint devices when using the computer with personal Aalto username and password. Color printing is possible using the printer u90203-psc3, which is located near the customer service. Color printing is subject to a charge to Aalto University students and staff members.
  • Other customers can use the printer u90203-psc3. All printing is subject to a charge to non-University members.
Location:P1 Ark Aalto  1739   | Archive
Keywords:dimensionality reduction
information visualization
stochastic neighbor embedding
alpha-divergence
exponential divergence with augmentation
Abstract (eng):Dimensionality reduction and information visualization are fundamental steps in data processing, information extraction and reasoning.
In real-world applications, the number of measurements or variables per a single observation is so large that handling the raw data in a specific problem such as regression or classification becomes infeasible or even impractical.
Moreover, in many applications, a faithful representation of the data for a first step analysis and hypothesis development becomes crucial.
Recently, the SNE method has become tremendously popular for data visualization and feature extraction.
The more recent algorithms such as t-SNE and HSSNE extend the basic SNE algorithm by considering general heavy-tailed distributions in the low-dimensional space, while the others, such as NeRV, consider different parameterized cost functions to achieve the desired embedding by tuning the parameter.
In this thesis, we provide another extension to the SNE method by investigating the properties of alpha-divergence for neighbor embedding, focusing our attention on a particular range of alpha values.
We show that alpha-divergence, with a proper selection of the alpha parameter effectively eliminates the crowding problem associated with the early methods.
However, we also provide the extensions of our method to distributions having heavier tail than Gaussian.
Contrary to some earlier methods like HSSNE and NeRV, no hand-tuning is needed, but we can rigorously estimate the optimal value of alpha for given input data.
For this, we provide a statistical framework using a novel distribution called Exponential Divergence with Augmentation.
This is an approximate generalization of Tweedie distribution and enables alpha-optimization after a nonlinear transformation.
We evaluate the performance of our proposed method by considering two sets of experiments: first, we provide a number of visualizations using our method and its extensions and compare the results with the earlier methods.
Second, we conduct a set of experiments to confirm the effectiveness of our alpha-optimization method for finding the optimal alpha for the data distribution, and its consistency with standard quality measures of dimensionality reduction.
ED:2014-08-31
INSSI record number: 49687
+ add basket
« previous | next »
INSSI