[PyTorch] Use nn.Embedding() To Load Gensim Pre-trained Model Weight
If you are using PyTorch to processing NLP tasks, you must be familiar with nn.Embedding()
in PyTorch.
If you are using PyTorch to processing NLP tasks, you must be familiar with nn.Embedding()
in PyTorch.
When we using PyTorch to build the model for deep learning tasks, sometimes we need to define more and more model layer.
It is so irritating. No one wants to keep pasting similar code over and over again.
Read More »[PyTorch] Use “ModuleList” To Reduce The Line Of Code That Define The ModelAutoEncoder is often referred to as AE for short. It is a neural network for unsupervised learning, in other words, it does not require labaled data.
Read More »[Machine Learning] Introduction To AutoEncoder (With PyTorch Code)This is a strange error message:
RuntimeError: Attempting to deserialize object on CUDA
device 3 but torch.cuda.device_count() is 1.
Read More »[Solved][PyTorch] RuntimeError: Attempting to deserialize object on CUDA device 3 but torch.cuda.device_count() is 1.Today I want to record a common problem, its solution is very rarely. Simple to put, the error message as follow:
RuntimeError: CUDA out of memory. Tried to allocate 2.0 GiB.
Read More »[Solved][PyTorch] RuntimeError: CUDA out of memory. Tried to allocate 2.0 GiBWhen I use PyTorch to build a model, I often feel at a loss as to how to add the data to the end of the sequence when processing the data.
Read More »[PyTorch] Use torch.cat() To Replace The append() Operation In The List Data When Processing torch TensorIt is very convenient for building a model using the PyTorch framework. Today I want to introduce how to print out the model architecture and extract the model layer weights.
Read More »[PyTorch] How To Print Model Architecture And Extract Model WeightsToday when I using nn.Embedding
in PyTorch, because I used it by mistake, the following error occurred:
GRU (Gate Recurrent Unit), I am not sure whether there is a Chinese translate of its name. Like LSTM, it is a variant of RNN (Recurrent Neural Network), and like LSTM, it aims to solve the gradient problem in RNN.
Read More »Graphical Introduction Note About GRULSTM (Long Short-Term Memory), is a type of Recurrent Neural Network (RNN). The paper about LSTM was published in 1997, which is a very important and easy-to-use model layer in natural language processing.
Read More »[PyTorch] LSTM Principle and Input and Output Format Record