Please use this identifier to cite or link to this item: http://localhost:8081/jspui/handle/123456789/16971
Full metadata record
DC FieldValueLanguage
dc.contributor.authorPradhan, Devesh-
dc.date.accessioned2025-06-23T12:10:08Z-
dc.date.available2025-06-23T12:10:08Z-
dc.date.issued2015-05-
dc.identifier.urihttp://localhost:8081/jspui/handle/123456789/16971-
dc.description.abstractThe task of object identification for large scale Image classification has seen rapid advances in the last few years due to advances in architecture of convolutional networks for automated -' feature learning. There is increased interest since these networks have been proven to outperform hand engineered features for the task of Image classification when the number of classes is very large. As the complexity of models increases along with the sizes of datasets, significant development efforts and resources are required to train networks of the scale of billions of parameters. With increasing efforts being expended on multi-GPU training, spreading out the computation across multiple OPUs in cluster requires considerable effort in division of work across the available resources. In this dissertation, we have designed and developed a framework for distributed computation of convolutional neural networks over multiple OPUs, which aims to separate the network definition and the details of distribution from the code that is to be run on the system. The framework is designed to be extensible, allowing for future additions of computation layers- that are the building blocks of neural networks, and the update rules- that determine the network convergence behavior. Further, the support of distributed training allows one to leverage more resources for faster training or for training larger networks. The framework supports this distribution by providing a distributed parameter store for parameter sharing between models, and shared memory abstraction for data sharing between processes, using a combination of which allows for both model and data parallelism as we demonstrate. We discuss the architecture of the framework we built for this purpose, along with its components and the network definition format, demonstrating both data parallel and model parallel execution of computation for a sample network.en_US
dc.description.sponsorshipINDIAN INSTITUTE OF TECHNOLOGY ROORKEEen_US
dc.language.isoenen_US
dc.publisherIIT ROORKEEen_US
dc.subjectObject Identificationen_US
dc.subjectLarge Scale Image Classificationen_US
dc.subjectArchitecture Of Convolutional Networksen_US
dc.subjectMulti-GPU Trainingen_US
dc.titleA FRAMEWORK FOR DISTRIBUTED LEARNING OF CONVOLUTIONAL NEURAL NETWORKSen_US
dc.typeOtheren_US
Appears in Collections:MASTERS' THESES (E & C)

Files in This Item:
File Description SizeFormat 
G25083.pdf10.36 MBAdobe PDFView/Open


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.