Replies: 1 comment 12 replies
-
You either need a dataset with emotion annotations and use these annotations as extra inputs to the model or if the dataset has no annotation but the speech is expressive then we can use sentiment classification to create these pseudo labels and train on them. Curious to hear that others say. So maybe there are more alternatives. |
Beta Was this translation helpful? Give feedback.
12 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I'm interested in adding emotion_id in addition to speaker_id so that while inference, I can choose which speaker as well as which emotion. The dataset would have to include an additional parameter (emotion_id) as well. This will enable me to do inference based on a specific speaker and specific emotion (angry, sad, happy, etc.)
I'm wondering how to go about doing this? Has something similar been done? Is this useful to anyone else?
Beta Was this translation helpful? Give feedback.
All reactions