Sentence Generating Function using Neural Network
A language is composed of words and sequences of words. When we listen
to a talk or read an article, we get information not only from
individual words but also from the sequence of these words. If we want to use
neural networks to process a language, a network with memory which can
process temporal sequences, would be a better choice.
Therefore, we use DFA(Discrete Finite Automata) networks to process and model a
language. By taking advantage of DFA's ability of predicting the next word, we can
input the user's query sentence to the trained DFA network and do a similarity comparison for
semantic search with the query sentence plus the next word in the sentence.
We also include the results by NMF(Non-negative Matrix Factorization) and SOM(Self-Organizaing Map).
We use 16 works of Mark Twain as our training
corpus.
- The Innocents Abroad.
- Roughing it.
- The Adventures of TOM Sawyer.
- A Tramp Abroad.
- The Prince and the Pauper.
- Life on the Mississippi.
- The Adventures of Huckleberry Finn.
- A Connecticut Yankee in King Arthur's Court.
- A Horse's Tale
- Christian Science
- Extract From Captain Stormfield's Visit to Heaven
- Is Shakespeare Dead From My Autobiography
- The American Claimant
- The Mysterious Stranger
- The Tragedy of Pudd'n'head Wilson
- Tom Sawyer Abroad
We use these three networks to
obtain semantic codes, and use these codes to accomplish semantic
analysis.
Download Semantic Indexing
on Mark Twain's works Demo
Program
Usage
Tutorials
Any questions, please
contact with cyliou@csie.ntu.edu.tw

All Rights Reserved by Cheng-Yuan,
Liou,
Department of Computer Science and Information Engineering,
National Taiwan University.
|