1 / 7

Long Short-Term Memory in Machine Learning

https://nixustechnologies.com/long-short-term-memory-in-machine-learning/

Sudhanshi
Télécharger la présentation

Long Short-Term Memory in Machine Learning

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Long Short-Term Memory in Machine Learning Art And Design '50 Year 2023

  2. Long Short-Term Memory in Machine Learning Konoadoh University What does LSTM stand for? You must be asking yourself first. The word “LSTM” refers to long short-term memory networks, which are employed in deep learning. It is a kind of recurrent neural network (RNNs) which are capable of learning long-term relationships, particularly in sequence prediction tasks. Except for a single data point, like a picture, the LSTM contains feedback connections, which means it can process the full series of data. Machine translation and other areas benefit from this. A unique class of RNN known as LSTM performs remarkably well across a wide range of issues. It is a specific variety of RNNs that can address the vanishing gradient issue that RNNs encounter. Hochreiter and Schmidhuber created LSTM to address the issue with conventional RNNs and ML techniques. This article will go over all the fundamentals of the LSTM. Let’s begin with the definition of LSTM.

  3. LSTM meaning Konoadoh University In the area of deep learning, the artificial RNN architecture utilized is called LSTM. LSTMs, in contrast to conventional RNNs, have “memory cells” that are capable of retaining data for extended periods of time. Additionally, three gates—the input gate, the forget gate, and the output gate—control the movement of information within and outside of the memory cells. The way LSTMs process information over time distinguishes them from other neural network types. With conventional neural networks, information is processed in a “feedforward” manner, which means that input is received at a one-time step, and an output is generated at the following time step.

  4. LSTM working Konoadoh University The LSTM model makes an effort to get around the issue of RNN (short-term memory) by saving a portion of its data to long-term memory. Known as the Cell State, this long-term memory is kept there. Additionally, there is the hidden state, which is sim-in-machine-learningilar to regular neural networks and where short-term data from earlier computation stages are retained. Short-term memory is the model’s hidden state. Long Short-Term Networks are also explained by this.

  5. Konoadoh University Here, in the above figure, Xt= input ht=hidden state ct=cell state f=forget gate g=memory cell i=input gate o=output gate As we have already seen that there are three gates (i.e. input gate, forget gate, and output gate). Now, let’s look at the role of each one of these. Note that each computation uses the current input ( say x(t)), the prior state of short-term memory (say c(t-1)), and the prior state of hidden state ( say h(t-1)).

  6. LSTM applications Konoadoh University LSTM Networks were the most effective tool for NLP because they could keep the context of a sentence “in memory” for a sizable amount of time. This kind of neural network is used in the following real-world applications: LSTM for Language modelling automated translation identification of handwriting captions for images answering inquiries text conversion from video LSTM for polymorphic music modelling Creating images using attention models Word-by-word text generation

  7. LSTM limitations Konoadoh University Due to their additional parameters and operations, they are more computationally demanding and demand more memory and training time. They are more likely to overfit, which calls for regularization strategies ( like dropout, weight decay, or early termination). Because they contain more hidden layers and states than ordinary RNN cells, they are more difficult to analyze and explain. Comparing LSTM cells to straightforward RNN cells has some disadvantages. Conclusion We have studied that LSTM models are a form of RNN. Although RNNs can accomplish anything, LSTM networks can do it with far more elegance; therefore, they are definitely an upgrade over RNNs. We have also learnt about the mechanism of LSTM, areas where it is being successfully implemented and some of its limitations.

More Related