html5-img
1 / 21

Text Summarization via Semantic Representation

Text Summarization via Semantic Representation. 吳旻誠 2014/07/16. Gist-content Question. Ask the main idea of the talk. The correct answers describe the closest overall theme of the content, and the distractors refer to only small portions of the content. Gist-content Question.

wyatt-stone
Télécharger la présentation

Text Summarization via Semantic Representation

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Text Summarization via Semantic Representation 吳旻誠 2014/07/16

  2. Gist-content Question • Ask the main idea of the talk. • The correct answers describe the closest overall theme of the content, and the distractors refer to only small portions of the content.

  3. Gist-content Question • Q. Which of the following is closest to the main idea of this talk? • (A)We've had three explanations for why we might sleep. • (B)When you're tired, and you lack sleep, you have poor memory, poor creativity, increased impulsiveness, and overall poor judgment. • (C)If you have good sleep, it increases your concentration, attention, decision-making, creativity, social skills, health. • (D)You do not do anything much while you're asleep.

  4. Gist-content question generation • Seem the most important sentence as the main idea of the talk. • Use LexRank to Measure the importance of sentences.

  5. LexRank • Measure the importance of sentences. • Graph-based Model(Undirected). • The nodes represent the sentences. • The edges are the cosine similarity between nodes.

  6. LexRank

  7. LexRank

  8. Conditions that should be satisfied • Stochastic matrix. • Irreducible. • Aperiodic.

  9. LexRank

  10. Similarity Between Sentences • But…what’s the similarity between the following sentences? • I will fully support you. • I'll back you up all the way. 

  11. Deep Learning • A set of algorithms in Machine Learning area. • Learning representations of data. • Has been applied to fields like computer vision , automatic speech recognition, natural language processing.

  12. Reduce the Dimensionality of Data with Neural Networks

  13. Word2Vec • A Google open source tool. • Compute vector representations of words. • Provides an efficient implementation of • Continuous Bag-of-Words(CBOW)architecture. • Skip-gram architecture.

  14. Continuous Bag-of-Words Model(CBOW)

  15. Skip-gram Model

  16. Softmax

  17. Hierarchical Softmax • Uses a binary tree representation of the output layer with the W words as its leaves. • Each word w can be reached by an appropriate path from the root of the tree.

  18. Sentence Representations Now we have the representations of words , but how can we represent sentences by these representations? A recursive Deep Learning model was proposed by Stanford Natural Language Processing Group.

  19. Recursive Autoencoder

  20. Dynamic Pooling

  21. Paraphrase_Identification_(State_of_the_art)

More Related