찾으시는 취미가 있나요?

I'm going to upload the mission result. (For the fair comparison of the two loss functions, the structure of the artificial neural network was left intact.) Compared to using KL-div as the loss function, it seems to be a relatively smaller value from the loss function of the first epoch. Thus, the same number of epochs were taught, but the loss function value of the last epoch was smaller with CrossEntropyLoss. After verification with the test set, the loss function is slightly larger than the training set (0.003) and the accuracy is about 2% lower (compared to the KL-div), so it seems to have been relatively slightly (?) understudied in the training set (compared to the KL-div. (Is this how you interpret the results...? Or should we consider it a proper learning...?) Even if we increase the number of epochs by one more. I'm going to change the structure of the phytosis... And I have a question. I'd like to check how the weight matrix was determined at the end of the optimization process for each batch. I wonder how I can check it out. When I printed out the print (model.lin3), I couldn't find any specific matrix shapes. + Comparing the codes with the above, they were all the same from the training stage. After running the case of KL-div, we carefully guess if the previously optimized weight matrix was used in the mission because the model was not initialized. (It's almost the same as my code, but the results are different, so I thought about it a little bit.) Thank you for reading the long article!
원본 보기
♡1
0

댓글0

클래스101은 모든 사람이 사랑하는 일을 하며
살 수 있도록 세상을 바꾸고자 합니다.
크리에이터
정규 클래스 지원하기
도움말
고객센터
오전 10시 ~ 오후 6시 (주말, 공휴일 제외)
주식회사 클래스101 | 대표 공대선 | 서울특별시 강남구 테헤란로 415, 4층(삼성동, 엘7강남타워) | ask@101.inc | 전화번호: 1800-2109 | 클라우드 호스팅: Amazon Web Services Korea LLC | 사업자등록번호 : 457-81-00277 | 통신판매업신고 : 2022-서울강남-02525 | 클래스101은 통신판매중개자로서 중개하는 거래에 대하여 책임을 부담하지 않습니다.