pytorch cross-modal Transformer using two features: MFCC from audio signal (1-channel); BERT last layer fine-tuned by multi-sentiment dataset. Requirements.
Missing: carat url? q=
STEP1: Convert input audio to text using Google ASR API; STEP2: Extract MFCC feature from input audio; STEP3: Conduct MLM on KoBERT through colloquial ...
Missing: carat url? q=
cross-modal model between audio(MFCC) and text(KoBERT) - audiotext-transformer/datasets.py at master · Donghwa-KIM/audiotext-transformer.
Missing: carat url? q= https:// md
cross-modal model between audio(MFCC) and text(KoBERT) - audiotext-transformer/utils.py at master · Donghwa-KIM/audiotext-transformer.
Missing: carat audio/ url? q= https:// md
cross-modal model between audio(MFCC) and text(KoBERT) - audiotext-transformer/modules.py at master · Donghwa-KIM/audiotext-transformer.
Missing: carat audio/ url? https:// md
The modelStudio package automates the explanatory analysis of machine learning predictive models. It generates advanced interactive model explanations in ...
Missing: carat audio/ q= Donghwa- KIM/ audiotext-
Illustration of AST. This repository contains the official implementation (in PyTorch) of the Self-Supervised Audio Spectrogram Transformer (SSAST) proposed in ...
Missing: carat q= Donghwa- KIM/ audiotext-
An app to make it easier to explore and curate output from a Music Transformer - listen-to-transformer/README.md at master · magenta/listen-to-transformer.
Missing: carat audio/ url? q= Donghwa- KIM/ audiotext-
In order to show you the most relevant results, we have omitted some entries very similar to the 8 already displayed. If you like, you can repeat the search with the omitted results included.