more...In June 2018, OpenAI presented the GPT neural network model, which immediately showed the best results in a number of language tests. GDP-2 appeared in 2019, and GPT-3 was presented in May 2020. These models demonstrated the ability of the neural network to generate related text. Additional experiments concerned the ability to generate music and images. The main disadvantage of such models is connected with the computing resources they involve. It took a month to train the first GPT on a machine with 8 GPUs. This disadvantage can be partially compensated by the possibility of using pre-trained models to solve new problems. But considerable resources are required to maintain the model functioning considering its size.
---------------------
- Neural networks made easy - MT5
- Neural networks made easy (Part 2): Network training and testing - MT5
- Neural networks made easy (Part 3): Convolutional networks - MT5
- Neural networks made easy (Part 4): Recurrent networks - MT5
- Neural networks made easy (Part 5): Multithreaded calculations in OpenCL - MT5
- Neural networks made easy (Part 6): Experimenting with the neural network learning rate - MT5
- Neural networks made easy (Part 7): Adaptive optimization methods - MT5
- Neural networks made easy (Part 8): Attention mechanisms - MT5
- Neural networks made easy (Part 9): Documenting the work - MT5
- Neural networks made easy (Part 10): Multi-Head Attention - MT5
- Neural networks made easy (Part 11): A take on GPT - MT5
- Neural networks made easy (Part 12): Dropout - MT5
- Neural networks made easy (Part 13): Batch Normalization - MT5
Bookmarks