-
Notifications
You must be signed in to change notification settings - Fork 102
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add lead time to each labelled query and avg lead time for a case #1003
Comments
@DmitryKey Have you looked at the Books infrastructure yet? YOu can merge various books together into new books, and then use that to populate a Case... There might be some nice operational things in there. In terms of lead time, I'm wondering if the existing update_time and create_time fields that we create for the |
@DmitryKey I am going to be Europe from May 1 to May 8th, so we could pair on this analysis together if you want. Right now on the home page we have this messaging: What if we could predict how long till all ratings are done? |
Hey @epugh ! I think update_time and create_time fields are a good start. Some thoughts:
Predicting ETA for ratings - is a fantastic feature to have. It could start with a simple retrospective prediction ("so far it took this long per query / case, so we can make a prediction"). However, this may vary per annotator, but should probably converge on average across all annotators. |
I was thinking about looking more into this.... Had some discussions about "Workflow" in Quepid, and routing different types of docs to specific judges etc... |
Is your feature request related to a problem? Please describe.
Usually the process in our labelling projects starts like this:
Then, we proceed with scaling out the labelling process by involving a larger set of people (who might be domain, but not search, experts).
At this point, knowing what is the avg lead time per unit of work (a query), we know, how much of workforce to request to reach a specific deadline.
Describe the solution you'd like
A lead time is recorded on query level and is rolled up to the case level.
The lead time can be accessed via Notebook feature to perform analytics, like distribution of lead times per annotator.
Describe alternatives you've considered
Recording this manually, but this is not accurate, plus increases the complexity of labelling.
Additional context
Label Studio (https://github.com/HumanSignal/label-studio) offers a lead time feature on task level. The UI is written in react (I believe), so probably, there is a way to adopt and adapt the UI logic for Quepid.
The text was updated successfully, but these errors were encountered: