Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Request for Release of Original Crowdsourcing Data for Annotator-Level Analysis #10

Open
hhf-hd opened this issue Oct 9, 2024 · 0 comments

Comments

@hhf-hd
Copy link

hhf-hd commented Oct 9, 2024

Hi there,

First of all, I’d like to sincerely thank you for your outstanding work on constructing and open-sourcing the dataset. The effort you put into creating it through crowdsourcing is truly appreciated, and it is a valuable contribution to the research community.

As part of my research, I am interested in analyzing the consistency between individual annotators. However, I noticed that while the labels have been made available, the original crowdsourcing data—such as the annotator IDs, the specific samples each annotator labeled, the provided labels, and time information (both task acceptance and submission times)—has not been released. Access to this information would be extremely helpful for calculating inter-annotator agreement and performing a more detailed analysis of annotator behavior.

Would it be possible to release the original crowdsourcing files with the mentioned details included? Having access to this data would greatly enhance reproducibility and allow for deeper insights into label quality and consistency.

Thank you again for your excellent work, and I look forward to your response!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant