Distributed representation of syntactic structure with tensor product representation and non-linear compression

張量乘積表述及非線性壓縮的句法結構分佈式表述

Student thesis: Master's Thesis

View graph of relations

Author(s)

  • Hoi Tik YEUNG

Related Research Unit(s)

Detail(s)

Awarding Institution
Supervisors/Advisors
Award date15 Jul 2005

Abstract

Representing lexicons and sentences with the sub-symbolic approach (using techniques such as Self Organizing Map (SOM) or Artificial Neural Network (ANN) is a relatively new but important research area in natural language processing. The performance of this approach however, is highly dependent on whether representations are well formed so that members within each cluster are corresponding to sentences or phrases of similar meaning. Despite the moderate success and the rapid advancement of contemporary computing power, it is still difficult to establish an efficient learning method so that natural language can be represented in a way close to the benchmark exhibited by human beings. One of the major problems is due to the general lack of effective method(s) to encapsulate semantic information into quantitative expressions or structures. In this thesis, we propose to alleviate this problem with a novel technique based on Tensor Product Representation and Non-linear Compression. This approach is capable of encoding sentences into distributed representations that are closely associated with the semantic contents, being more comprehensible and analyzable from the perspective of human intelligence.

    Research areas

  • Natural language processing (Computer science)