You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I am trying to reproduce the rec performance on INSPIRED dataset.
I use the hyperparameters you recommend and the "best" model as prompt-encoder. Unfortunately, I was not able to reproduce the performance on the paper.
---- Here I attached the loss and recall@1 on testset for prompt pre-training, conversational training, and recommendation training steps:
prompt pre-training
conversational training
recommendation training (as you can see, the best recall@1 I got is around 0.04, far from 0.09)
---- and here are the configuration I use for prompt pre-training, conversational training, and recommendation training steps:
Sorry for the late reply! The problem comes from the pre-training stage. You should observe a continuous increase in performance and drop in loss since the answer is actually provided in the response.
Sorry for the late reply! The problem comes from the pre-training stage. You should observe a continuous increase in performance and drop in loss since the answer is actually provided in the response.
loss
Recall
Though my pretraining results is as good as above, the recommendation result is still lower too much compared to paper results.
pretraining:
recommendation:
But I can easily reproduce the UniCRS results in CFCRS code, what's their difference?
@invoker-LL Hi, your problem is interesting. The metric test/loss is abnormal, since it keeps increasing. You can compare the code with UniCRS for more detail. According to my impression, the main difference is that the length of soft token prompt is set to 0 in CFCRS
Dear Author,
I am trying to reproduce the rec performance on INSPIRED dataset.
I use the hyperparameters you recommend and the "best" model as prompt-encoder. Unfortunately, I was not able to reproduce the performance on the paper.
---- Here I attached the loss and recall@1 on testset for prompt pre-training, conversational training, and recommendation training steps:
![image](https://private-user-images.githubusercontent.com/60718476/262405983-20d5f89c-9ba0-4f42-b965-9db7b4e539b4.png?jwt=eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJpc3MiOiJnaXRodWIuY29tIiwiYXVkIjoicmF3LmdpdGh1YnVzZXJjb250ZW50LmNvbSIsImtleSI6ImtleTUiLCJleHAiOjE3MzkyNTkwODksIm5iZiI6MTczOTI1ODc4OSwicGF0aCI6Ii82MDcxODQ3Ni8yNjI0MDU5ODMtMjBkNWY4OWMtOWJhMC00ZjQyLWI5NjUtOWRiN2I0ZTUzOWI0LnBuZz9YLUFtei1BbGdvcml0aG09QVdTNC1ITUFDLVNIQTI1NiZYLUFtei1DcmVkZW50aWFsPUFLSUFWQ09EWUxTQTUzUFFLNFpBJTJGMjAyNTAyMTElMkZ1cy1lYXN0LTElMkZzMyUyRmF3czRfcmVxdWVzdCZYLUFtei1EYXRlPTIwMjUwMjExVDA3MjYyOVomWC1BbXotRXhwaXJlcz0zMDAmWC1BbXotU2lnbmF0dXJlPTZmMThjNjY5MTc5NTBkY2FlZWFlMTMxMDUyN2I3YzgzYTFhMGM4MWJjMDQwOGY1NTgzODA0MDgxYjFkMWM3MDEmWC1BbXotU2lnbmVkSGVhZGVycz1ob3N0In0.soWHq2o3_-dtCB9Fs4l6YtxcZMteHNfpV04hBtGnHCQ)
![image](https://private-user-images.githubusercontent.com/60718476/262416690-dcf74f86-fa4f-4d6f-99e7-999e883b8cc7.png?jwt=eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJpc3MiOiJnaXRodWIuY29tIiwiYXVkIjoicmF3LmdpdGh1YnVzZXJjb250ZW50LmNvbSIsImtleSI6ImtleTUiLCJleHAiOjE3MzkyNTkwODksIm5iZiI6MTczOTI1ODc4OSwicGF0aCI6Ii82MDcxODQ3Ni8yNjI0MTY2OTAtZGNmNzRmODYtZmE0Zi00ZDZmLTk5ZTctOTk5ZTg4M2I4Y2M3LnBuZz9YLUFtei1BbGdvcml0aG09QVdTNC1ITUFDLVNIQTI1NiZYLUFtei1DcmVkZW50aWFsPUFLSUFWQ09EWUxTQTUzUFFLNFpBJTJGMjAyNTAyMTElMkZ1cy1lYXN0LTElMkZzMyUyRmF3czRfcmVxdWVzdCZYLUFtei1EYXRlPTIwMjUwMjExVDA3MjYyOVomWC1BbXotRXhwaXJlcz0zMDAmWC1BbXotU2lnbmF0dXJlPWNkYzM2NWVmMWQ1ODdkMzRlNDQyOWRiZjE3MWU2NWM5NTYzOWNkNWI2NDI2MWRmZDRkNGFmNmQ0ODFiZDYwMzImWC1BbXotU2lnbmVkSGVhZGVycz1ob3N0In0._5D099g4z0pk14m1Gkmzj7-mCE9o_NBUbhjj8vDtdUY)
prompt pre-training
conversational training
recommendation training (as you can see, the best recall@1 I got is around 0.04, far from 0.09)
---- and here are the configuration I use for prompt pre-training, conversational training, and recommendation training steps:
prompt pre-training
conv training
conv infer
rec training
Thank you!
The text was updated successfully, but these errors were encountered: