Page 91 - sjsi
P. 91

Research Article: Alsibai & Heydari                                                             91


            that  SGD  generally  performs  better  on  image     trained weights frozen except for its final Linear
            classification  tasks.  The  optimizer  and  the      layer.  This  method  is  recommended  in  the
            learning  rate  are  closely  related,  as  the       paper   (19)   for  models  trained  on  ImageNet
            optimizer uses the learning rate to determine         architectures.   Exceptional   results   were
            the  step  size  when  making  updates  to  the       achieved with 100% accuracy on the train set
            model parameters. The learning rate opted for         and 99.83% on the test set. The following two
            is 0.01.                                              figures  show  the  loss  and  accuracy  during
                                                                  training  in  addition  to  the  confusion  matrix
            Results                                               obtained  by  the  results  of  the  trained  model
            After  conducting  the  above  steps  and  setting    predicting on the test set. Since the results are
            the learning rate to 0.01, the optimizer is set to    suspiciously perfect, we decided to investigate
            SGD and training the model is initiated for 15        more using another dataset.
            epochs.  Moreover,  the  model  had  its  pre-

































                                     Figure 5: Results of loss and accuracy during training epochs
























                                             Figure 6: Confusion Matrix for Dataset A

                      SJSI – 2023: VOLUME 1-1
   86   87   88   89   90   91   92   93   94   95   96