all(408)
-
[DL Wizard] Recurrent Neural Network with PyTorch 번역 및 정리
https://www.deeplearningwizard.com/deep_learning/practical_pytorch/pytorch_recurrent_neuralnetwork/ Recurrent Neural Networks (RNN) - Deep Learning Wizard Recurrent Neural Network with PyTorch About Recurrent Neural Network Feedforward Neural Networks Transition to 1 Layer Recurrent Neural Networks (RNN) RNN is essentially an FNN but with a hidden layer (non-linear output) that passes on informa..
2020.02.08 -
그림으로 보는 RNN/LSTM과 GRU 번역 및 정리
https://towardsdatascience.com/illustrated-guide-to-recurrent-neural-networks-79e5eb8049c9 Illustrated Guide to Recurrent Neural Networks Understanding the Intuition towardsdatascience.com RNN과 은닉상태 초기화 → input 단어와 은닉상태를 RNN에 집어 넣기 → 그 단어의 output과 새로운 은닉상태가 출력됨 → 이 출력을 다시 RNN에 집어넣기 → 단어 없을 때까지 반복 → 출력을 FeedForward 레이어에 넣기 → 최종결과(prediction) 출력 은닉상태가 RNN의 기억장치 (저번 출력을 다음 입력으로 넘겨줌) tanh 함수는 출력을 [-..
2020.02.07 -
[DL Wizard] Convolutional Neural Network with PyTorch 번역 및 정리
https://www.deeplearningwizard.com/deep_learning/practical_pytorch/pytorch_convolutional_neuralnetwork/ Convolutional Neural Networks (CNN) - Deep Learning Wizard Convolutional Neural Network with PyTorch About Convolutional Neural Network Transition From Feedforward Neural Network Hidden Layer Feedforward Neural Network Recap of FNN So let's do a recap of what we covered in the Feedforward Neur..
2020.02.06 -
CNN for NLP 번역 및 정리
http://www.wildml.com/2015/11/understanding-convolutional-neural-networks-for-nlp/ Understanding Convolutional Neural Networks for NLP When we hear about Convolutional Neural Network (CNNs), we typically think of Computer Vision. CNNs were responsible for major breakthroughs in Image Classification and are the core of most Compute… www.wildml.com NLP에서 사용되는 CNN - 입력 데이터는 word embedding, 각 행은 한 단..
2020.02.06 -
CNN : Convolutional Neural Network 정리
(사진 출처) 밑바닥부터 시작하는 딥러닝 1권 http://taewan.kim/post/cnn/#fn:1 CNN, Convolutional Neural Network 요약 Convolutional Neural Network, CNN을 정리합니다. taewan.kim 완전연결 계층(Affine)으로 이루어진 네트워크 (ex) input → Affine → ReLU → Affine → ReLU → Affine → ReLU → Affine → Softmax CNN으로 이루어진 네트워크 (ex) input → Conv →ReLU → Pooling → Conv → ReLU → Pooling → Conv → ReLU → Affine → ReLU → Affine → Softmax = (1) 이미지의 특징을 추출하는 ..
2020.02.06 -
[DL Wizard] Weight Initializations & Activation Functions 번역 및 정리
https://www.deeplearningwizard.com/deep_learning/boosting_models_pytorch/weight_initialization_activation_functions/ Weight Initialization and Activation Functions - Deep Learning Wizard Weight Initializations & Activation Functions Recap of Logistic Regression Recap of Feedforward Neural Network Activation Function Sigmoid (Logistic) \sigma(x) = \frac{1}{1 + e^{-x}} Input number \rightarrow [0,..
2020.02.05