
Our team made significant progress in model development this week. We tested the accuracies of Recurrent Neural Networks and Support Vector Classification for the audio modality and set up EmotionCLIP for accuracy evaluation. We initiated the FG-Net system for Facial Action Units training and testing. We also established our intermediate layer fusion code, which we tested using off-the-shelf ResNet and MobileNet models.
Next week, we’ll complete the initial testing of individual components like CLIP, EmotionCLIP, spectrogram CNN, and spectrogram ConvLSTM to ensure smooth integration into our fusion model. We’ll finalize our QRB1 presentation, share testing results with our coach and liaison, and integrate selected models into our fusion implementation. Our project continues to progress systematically toward our goals.