-
Notifications
You must be signed in to change notification settings - Fork 1
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
NLP Transformer Model for Depression Detection #6
base: main
Are you sure you want to change the base?
Conversation
ok everyone, I just reorganized the repository so that our changes are separated into different branches and so main isn't flooded with random experiment files.
|
The latest updates on your projects. Learn more about Vercel for Git ↗︎
|
aw hell nah
…On Sat, Jul 20, 2024 at 12:38 PM vercel[bot] ***@***.***> wrote:
*The latest updates on your projects*. Learn more about Vercel for Git ↗︎
<https://vercel.link/github-learn-more>
Name Status Preview Comments Updated (UTC)
*utd-summer-ai* ✅ Ready (Inspect
<https://vercel.com/james-projects-d63f44ae/utd-summer-ai/EXE3jfhsTeN5nkK4ayPt2oMJdnpk>
) Visit Preview
<https://vercel.live/open-feedback/utd-summer-ai-git-model-james-projects-d63f44ae.vercel.app?via=pr-comment-visit-preview-link&passThrough=1> 💬
*Add feedback*
<https://vercel.live/open-feedback/utd-summer-ai-git-model-james-projects-d63f44ae.vercel.app?via=pr-comment-feedback-link> Jul
20, 2024 5:38pm
—
Reply to this email directly, view it on GitHub
<#6 (comment)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/A5667L34Y6STZY7ZSUAD5CTZNKOBDAVCNFSM6AAAAABLFC6Y32VHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMZDENBRGI2DAMBTGA>
.
You are receiving this because you were assigned.Message ID:
***@***.***>
|
Note this uses the "WEBEATTHEPAPER" model so it might have to be changed if needed
i can't even checkout this branch on CLI because the LFS limit is full... |
observation: amount of posts significantly impacts depression score. i noticed this when fixing the bug where mastodon would only give me max of 40 posts instead of 60 |
#1
2024-07-18:
We began experimenting with different NLP methods and datasets for depression detection.
Found some nice datasets we experimented on:
Experiments:
proj.ipynb
andproj.py
Its_transforming_time_.ipynb
bert_test.ipynb
andbert_test2.ipynb
2024-07-20
Experimented with larger datasets
bert_test3.ipynb
2024-07-22
#11
The next few days are about improving accuracy on the big dataset.
2024-07-23
Sigma Bert.ipynb
averageUserTweet.ipynb
2024-07-24
layer_freezing_test.ipynb
and more regularization with no significant results (although layer freezing was faster to train, its accuracy dropped by ~1-2%). Also expanded the MAX_LEN of tokens per string in a successful attempt to combat overfitting through increase in data (BERT is very complex and can easily overfit if given less data) - this gave ~+5-6% accuracy with a final test accuracy of 79% inbigger_data.ipynb
. Tomorrow, we should plan on increasing the MAX_LEN even more, including testing on one of the older BERTs with no combining of text from the same user.2024-07-25
After significant sacrifices in mental health, Daniel, Kyle and James worked on debugging the LSTM + BERT. This included fixes in the encoder, optimizations in the forward function, and improving the training loop with gradient accumalation. This led to a massive reduction in iteration speed from roughly 50 seconds a batch yesterday to 3 this evening. Work still needs to be done to ensure accuracy, and increase the amount of data that can be used.
2024-07-26
Kyle + James: The error with the model was solved, it had to do with the wrong order of unpacking output tuples(... help me). We finally achieved consistent learning and accuracy, reaching 80%.
improved.ipynb
Then Daniel and Tad worked on a new variation of the LSTM+BERT nicknamed "tolBERT" instead of having one large LSTM with a thousand layers, it splits the output of the BERT models into groupings and uses multiple different LSTMs which are combined into a single MLP. This has the advantage of minimizing vanishing gradients, and more parameters to potentially tune. This achieved an astonishing 82.6% testing accuracy on the MDDL dataset, approaching our target accuracy of 85%. It is the belief of everyone working on the model, that with proper optimisation, both tolBERT and the LSTM+BERT should be able to achieve target accuracy. There is even discussion of an ensemble model in the future combining both. This is a huge step in our research, and we are very happy with the current results.
2024-07-27
everyone did more research on optimizing model accuracy
2024-07-28
Daniel: improved tolBERT to 84.4% by increasing LSTM layer amount to 5
James: improved tolBERT to 84.8% with 20 epochs, 6 LSTMs, and 6 LSTM layers
2024-07-29
we hope to finalize the models today, trying to get the files in this branch organized and ready to merge #23
Daniel: making a .py file for backend to load our model and run it
2024-07-30
fixing an issue where the saved torch model was giving us junk accuracy. we suspect it was because we were not saving the list of LSTMs properly. trying to fix with storing them in
torch.nn.ModuleList
instead of storing in regular Python list