-
Notifications
You must be signed in to change notification settings - Fork 12
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[duplicate prompts found]Clarification on Prompt Diversity and Action Generation #45
Comments
Dear Phongsiri, Thank you for your thoughtful review and for bringing up the concerns regarding the similarity in the dataset prompts, specifically related to the action "dancing". Your observations are insightful. To address your concerns: The dataset intentionally includes prompts with minimal variations to test the model's sensitivity to subtle linguistic cues that might influence the generated motions. For instance, terms like "gracefully" or "breakdancing" indicate different styles and energies of dancing, which are meant to prompt slight variations in the generated motions. This is part of our effort to refine the model's ability to discern and react to nuanced differences in human interaction descriptions. Although the textual annotations are similar (since the semantic category of these motions are "dance"), each motion captured is unique, which does not limit but rather enhances the diversity. For example, the diffusion model inherently has the capability to model such diversity effectively. Hence, similar annotation in this context is not a problem but an opportunity to refine the model's ability to generate nuanced variations of similar actions. Thank you once again for your interests in our work and the detailed review. Best regards, |
Dear Han, Thank you for your earlier response. While I appreciate the augmentation methods mentioned in the paper, they appear to be tied primarily to the evaluation process rather than addressing redundancy in the raw dataset. As I began training for over 20,000 epochs, I noticed some concerning patterns:
This raises two key questions:
These issues impact the diversity and usability of the dataset during real-world training. I’d greatly appreciate clarification on how these potential redundancies and anomalies are handled in the dataset preparation and evaluation stages. Best regards, Phongsiri |
hi, to address your questions:
|
Dear Intergen,
I have been reviewing the dataset for the InterGen project and noticed that many prompts for specific actions, such as "dancing," are highly similar, with minimal variations in wording. Below are 50 examples related to dancing from the dataset:
Dancing : 50 sequences
taichi : 17 sequences
sparring : 28 sequences
rock-paper-scissors : 4 sequences
Given this level of similarity, could you clarify how the model is expected to generate distinct and meaningful actions based on such closely related prompts? Additionally, do these similar tokenized inputs limit the diversity of generated actions, and if so, how does the system address this?
I appreciate your time in clarifying this matter.
Best regards,
Phongsiri
The text was updated successfully, but these errors were encountered: