Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

您好,请问为什么在class lstm1的forward里用hn_o+hn_1作为relu层的输入,而不直接用hn_1或output? #3

Open
mmm656 opened this issue Aug 5, 2024 · 2 comments

Comments

@mmm656
Copy link

mmm656 commented Aug 5, 2024

          您好,请问为什么在class lstm1的forward里用hn_o+hn_1作为relu层的输入,而不直接用hn_1或output?

屏幕截图 2024-08-05 165646

Originally posted by @mmm656 in #2 (comment)

@jiaxiang-cheng
Copy link
Owner

          您好,请问为什么在class lstm1的forward里用hn_o+hn_1作为relu层的输入,而不直接用hn_1或output?

屏幕截图 2024-08-05 165646

Originally posted by @mmm656 in #2 (comment)

你可以选择直接用hn_1或者output作为输出

@mmm656
Copy link
Author

mmm656 commented Aug 6, 2024

          您好,请问为什么在class lstm1的forward里用hn_o+hn_1作为relu层的输入,而不直接用hn_1或output?

屏幕截图 2024-08-05 165646
Originally posted by @mmm656 in #2 (comment)

你可以选择直接用hn_1或者output作为输出

但是直接用output作为输出会导致误差很大,可能是陷入了局部最小值,不知道您当初是否基于这样的考虑才换成hn_o+hn_1,请问这样做有没有理论支撑,或者只是出于经验?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants