-
-
Notifications
You must be signed in to change notification settings - Fork 4.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
CoherenceModel does not finish with computing #3368
Comments
@silviatti could you check this one please? #3197 was the only change in CoherenceModel, although I don't see how it's related. @PrimozGodec could you interrupt your stuck computation with ctrl-c and post the traceback? Thanks. |
Thank you for your fast response. Here is the traceback.
|
This comment was marked as abuse.
This comment was marked as abuse.
I also had the issue that the cm.get_coherence() call would not terminate for larger lists of texts. Here's how I "fixed" it: The base issue was not actually due to Gensim, but a problem on my end. It's just that (I presume) due to multiprocessing Gensim did not properly raise an error but simply never terminated. You can find out if you have the same problem using the following steps:
For me it was an IndexError: There was a bug in my upstream code and some empty texts, i.e. empty lists, snuck into the final I hope this helps! Though I'm not sure if @PrimozGodec and @nadiaelen are facing the same issue, the root cause might still lie somewhere with multiprocessing. |
@felixrech thank you for the suggestion. When switching to processes=1 I fond the error. |
Problem description
When computing coherence scores, it newer finishes with computing on a bit bigger dataset. Run the code below (with the provided dataset) to reproduce.
Steps/code/corpus to reproduce
coherence-bug.pkl.zip
Versions
The bug appears on Gensim version 4.2, but it does not happen on 4.1.2
macOS-10.16-x86_64-i386-64bit
Python 3.8.12 (default, Oct 12 2021, 06:23:56)
[Clang 10.0.0 ]
Bits 64
NumPy 1.22.3
SciPy 1.8.1
gensim 4.2.1.dev0
FAST_VERSION 0
The text was updated successfully, but these errors were encountered: