-
Updated
Nov 10, 2021
multimodal-machine-learning
Here are 6 public repositories matching this topic...
Reading list for Multimodal Large Language Models
-
Updated
Aug 17, 2023
MIntRec2.0 is the first large-scale dataset for multimodal intent recognition and out-of-scope detection in multi-party conversations (ICLR 2024)
-
Updated
Jun 28, 2024 - Python
Multimodal datasets for Machine-Learning
-
Updated
Jan 9, 2025 - Julia
This repository contains the code, dataset, and model outputs for the ICMI 2024 paper Multimodal User Enjoyment Detection in Human-Robot Conversation: The Power of Large Language Models. It includes scripts for prompting LLMs, training supervised models, and evaluating multimodal enjoyment detection.
-
Updated
Feb 6, 2025 - Python
This repository is cloned from https://github.com/HLR/LatentAlignmentProcedural. This is a potential baseline explored for the textual_cloze task on the RecipeQA Dataset - https://hucvl.github.io/recipeqa/
-
Updated
Feb 11, 2025 - Jupyter Notebook
Improve this page
Add a description, image, and links to the multimodal-machine-learning topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with the multimodal-machine-learning topic, visit your repo's landing page and select "manage topics."