×
cross-modal model between audio(MFCC) and text(KoBERT) - audiotext-transformer/datasets.py at master · Donghwa-KIM/audiotext-transformer.
Missing: carat url? q= https://
pytorch cross-modal Transformer using two features: MFCC from audio signal (1-channel); BERT last layer fine-tuned by multi-sentiment dataset. Requirements.
Missing: carat url? q= blob/
Multimodal Transformer for Korean Sentiment Analysis with Audio and Text Features - youngbin-ro/audiotext-transformer.
Missing: carat url? q= blob/
cross-modal model between audio(MFCC) and text(KoBERT) - audiotext-transformer/utils.py at master · Donghwa-KIM/audiotext-transformer.
Missing: carat audio/ url? q= https://
cross-modal model between audio(MFCC) and text(KoBERT) - audiotext-transformer/modules.py at master · Donghwa-KIM/audiotext-transformer.
Missing: carat audio/ url? https:// datasets.
The Code Repository for "HTS-AT: A Hierarchical Token-Semantic Audio Transformer for Sound Classification and Detection", in ICASSP 2022.
Missing: carat q= Donghwa- KIM/ audiotext- blob/ master/
This volume contains the papers selected for presentation at the 26th International. Symposium on Methodologies for Intelligent Systems (ISMIS 2022), ...
All the experiments were carried out on AudioCaps dataset, which is sourced from AudioSet. Our download version contains 49274/49837 audio clips in training set ...
Missing: carat q= https:// Donghwa- KIM/ audiotext- blob/ master/
Multimodal Transformer for Korean Sentiment Analysis with Audio and Text Features - audiotext-transformer/model.py at master ...
Missing: carat audio/ url? q= https:// Donghwa- KIM/ datasets.
In order to show you the most relevant results, we have omitted some entries very similar to the 9 already displayed. If you like, you can repeat the search with the omitted results included.