×
pytorch cross-modal Transformer using two features: MFCC from audio signal (1-channel); BERT last layer fine-tuned by multi-sentiment dataset. Requirements.
Missing: carat url? q=
STEP1: Convert input audio to text using Google ASR API; STEP2: Extract MFCC feature from input audio; STEP3: Conduct MLM on KoBERT through colloquial ...
Missing: carat url? q=
cross-modal model between audio(MFCC) and text(KoBERT) - audiotext-transformer/datasets.py at master · Donghwa-KIM/audiotext-transformer.
Missing: carat url? q= https:// md
cross-modal model between audio(MFCC) and text(KoBERT) - audiotext-transformer/utils.py at master · Donghwa-KIM/audiotext-transformer.
Missing: carat audio/ url? q= https:// md
This project has not set up a SECURITY.md file yet. There aren't any published security advisories. Footer. © 2024 GitHub, Inc. Footer navigation. Terms ...
Missing: carat audio/ url? q= https:// blob/ master/
Data Scientist. Donghwa-KIM has 80 repositories available. Follow their code on GitHub.
Missing: carat url? q= blob/ master/
The modelStudio package automates the explanatory analysis of machine learning predictive models. It generates advanced interactive model explanations in ...
Missing: carat audio/ q= Donghwa- KIM/ audiotext-
Illustration of AST. This repository contains the official implementation (in PyTorch) of the Self-Supervised Audio Spectrogram Transformer (SSAST) proposed in ...
Missing: carat q= Donghwa- KIM/ audiotext-
In order to show you the most relevant results, we have omitted some entries very similar to the 8 already displayed. If you like, you can repeat the search with the omitted results included.