You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Here is the time embedding code in the TimestepEmbedder() class in model.py ## at the top of Dit class embedding = torch.cat([torch.cos(args), torch.sin(args)], dim=-1)
It seems to connect two tensors in the dimension of half: [cos0, cos1,..., cos(half), sin0, sin1, ..., sin(half)]
However, the formula is alternately inserted: [cos0, sin0, cos1, sin1, ..., cos(half), sin(half)]
Experiments showed that the model works well in A. But I wonder that will this have any impact? Or In what kind of task it will lead to hazard?
Another question is why this part used an MLP for another representation for t? I mean, the representation/understanding could be processed in Transformer.
The text was updated successfully, but these errors were encountered:
Here is the time embedding code in the TimestepEmbedder() class in model.py ## at the top of Dit class
embedding = torch.cat([torch.cos(args), torch.sin(args)], dim=-1)
It seems to connect two tensors in the dimension of half: [cos0, cos1,..., cos(half), sin0, sin1, ..., sin(half)]
However, the formula is alternately inserted: [cos0, sin0, cos1, sin1, ..., cos(half), sin(half)]
Experiments showed that the model works well in A. But I wonder that will this have any impact? Or In what kind of task it will lead to hazard?
Another question is why this part used an MLP for another representation for t? I mean, the representation/understanding could be processed in Transformer.
The text was updated successfully, but these errors were encountered: