@ltzheng
When manually updating the audio_emotion variable in inference.py (line 142) to set it to a specific emotion (e.g., disgusted with value 1), the output does not reflect the updated emotion.
Modify inference.py at line 142:
audio_emotion = torch.full((audio_emb.shape[0],), 1, dtype=torch.int32, device=device)
num_emotion_classes = 9
This sets the audio_emotion to 1 (disgusted) for all embeddings.
Run the inference pipeline.
Expected Behavior
The output should correctly reflect the manually set audio_emotion value.
Observed Behavior
Observe that the output does not reflect the manually assigned emotion (disgusted).
@ltzheng
When manually updating the audio_emotion variable in inference.py (line 142) to set it to a specific emotion (e.g., disgusted with value 1), the output does not reflect the updated emotion.
Modify inference.py at line 142:
This sets the audio_emotion to 1 (disgusted) for all embeddings.
Run the inference pipeline.
Expected Behavior
The output should correctly reflect the manually set audio_emotion value.
Observed Behavior
Observe that the output does not reflect the manually assigned emotion (disgusted).