Abstract
In recent years, deep neural networks have made substantial progress in object recognition. However, one issue with deep learning is that it is currently unclear which proposed framework is exaggerated for a specific hitch. As a result, distinct dispositions are attempt before one that produces satisfactory results is discovered. This paper described a distributed supervised learning method for finding the best network architecture by modifying specifications for a perceived task dynamically. In the case of the MNIST information gathering, it is shown that asynchronous supervised learning can agree on a solution space. Setting several hyperparameters can be time-consuming when constructing neural networks. In this post, we'll provide you with some tips and instructions for better organizing your hyperparameter tuning process, which should help you find a good setting for the hyperparameters much faster.
Original language | English |
---|---|
Title of host publication | Proceedings - 2nd International Conference on Smart Electronics and Communication, ICOSEC 2021 |
Publisher | Institute of Electrical and Electronics Engineers Inc. |
Pages | 1587-1593 |
Number of pages | 7 |
ISBN (Electronic) | 9781665433686 |
DOIs | |
State | Published - 2021 |
Externally published | Yes |
Event | 2nd International Conference on Smart Electronics and Communication, ICOSEC 2021 - Trichy, India Duration: 7 Sep 2021 → 9 Sep 2021 |
Publication series
Name | Proceedings - 2nd International Conference on Smart Electronics and Communication, ICOSEC 2021 |
---|
Conference
Conference | 2nd International Conference on Smart Electronics and Communication, ICOSEC 2021 |
---|---|
Country/Territory | India |
City | Trichy |
Period | 7/09/21 → 9/09/21 |
Bibliographical note
Publisher Copyright:© 2021 IEEE.
Keywords
- Artificial intelligence
- Convolutional neural network
- Deep learning
- Hyper-parameter
- Hyperparameter tuning
- Machine learning
- Neural networks
- Orthogonal array