dc.contributor.advisor | Joshi, Manjunath V. | |
dc.contributor.author | Bhaisaheb, Shabbirhussain Hamid | |
dc.date.accessioned | 2020-09-22T14:41:19Z | |
dc.date.available | 2023-02-16T14:41:19Z | |
dc.date.issued | 2020 | |
dc.identifier.citation | Bhaisaheb, Shabbirhussain Hamid (2020). Optimization techniques for deep neural networks. Dhirubhai Ambani Institute of Information and Communication Technology. vii, 44 p. (Acc.No: T00848) | |
dc.identifier.uri | http://drsr.daiict.ac.in//handle/123456789/926 | |
dc.description.abstract | Optimization techniques are used to make the neural networks to converge to an optimal solution. Different techniques such as gradient-based approaches are utilized to obtain the parametric solution having objective function estimation to be minimum. The most commonly used method is Gradient Descent (GD) for this purpose, but there are certain drawbacks to this method such as there is no guarantee that it will converge to the global minima of the objective function, the convergence rate depends on the learning rate α (alpha) and wrong selection of which may lead to divergence. The solution depends on the parameter initialization and may differ with different initial values of the parameters, and also, the underlying assumption is that function should be differentiable, which may not always be possible. There are other non-gradient based optimization techniques such as genetic algorithms, graph cut method, particle swarm optimization, simulated annealing. Each of these techniques has their respective drawbacks and advantages. For the current research work, simulated annealing was kept in focus given it guarantees global optimization theoretically (unlike other methods). The gradient-based optimization methods and simulated annealing (SA) algorithm are used on different neural net architectures to check the running time of the algorithm and optimized parameter values indicating convergence of objective function. | |
dc.subject | Optimization | |
dc.subject | Gradient-based method | |
dc.subject | Reinforcement learning | |
dc.subject | Simulated annealing | |
dc.classification.ddc | 006.32 BHA | |
dc.title | Optimization techniques for deep neural networks | |
dc.type | Dissertation | |
dc.degree | M. Tech | |
dc.student.id | 201811017 | |
dc.accession.number | T00848 | |