Movatterモバイル変換


[0]ホーム

URL:


Skip to content

Navigation Menu

Sign in
Appearance settings

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Sign up
Appearance settings
@v-iashin
v-iashin
Follow
View v-iashin's full-sized avatar
👨‍💻

Vladimir Iashin v-iashin

👨‍💻
Postdoc in VGG at University of Oxford. Researcher of multi-modal machine learning

Block or report v-iashin

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more aboutblocking users.

You must be logged in to block users.

Maximum 250 characters. Please don't include any personal information such as legal names or email addresses. Markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more aboutreporting abuse.

Report abuse

PinnedLoading

  1. video_featuresvideo_featuresPublic

    Extract video features from raw videos using multiple GPUs. We support RAFT flow frames as well as S3D, I3D, R(2+1)D, VGGish, CLIP, and TIMM models.

    Python 624 102

  2. SpecVQGANSpecVQGANPublic

    Source code for "Taming Visually Guided Sound Generation" (Oral at the BMVC 2021)

    Jupyter Notebook 367 39

  3. BMTBMTPublic

    Source code for "Bi-modal Transformer for Dense Video Captioning" (BMVC 2020)

    Jupyter Notebook 228 56

  4. MDVCMDVCPublic

    PyTorch implementation of Multi-modal Dense Video Captioning (CVPR 2020 Workshops)

    Python 143 20

  5. SynchformerSynchformerPublic

    Source code for "Synchformer: Efficient Synchronization from Sparse Cues" (ICASSP 2024)

    Python 94 9

  6. SparseSyncSparseSyncPublic

    Source code for "Sparse in Space and Time: Audio-visual Synchronisation with Trainable Selectors." (Spotlight at the BMVC 2022)

    Python 53 10


[8]ページ先頭

©2009-2025 Movatter.jp