A collection of research on knowledge graphs
-
Updated
Oct 7, 2022 - JavaScript
A collection of research on knowledge graphs
🪩 Create Disco Diffusion artworks in one line
Represent, send, store and search multimodal data
PyTorch source code for "Stacked Cross Attention for Image-Text Matching" (ECCV 2018)
Analyze the unstructured data with Towhee, such as reverse image search, reverse video search, audio classification, question and answer systems, molecular search, etc.
Effective prompting for Large Multimodal Models like GPT-4 Vision, LLaVA or CogVLM. 🔥
A curated list of different papers and datasets in various areas of audio-visual processing
The implementation of AAAI-17 paper "Collective Deep Quantization of Efficient Cross-modal Retrieval"
Remote Sensing Sar-Optical Land-use Classfication Pytorch Pytorch高分辨率遥感语义分割/地物分割/地物分类
Weakly Supervised 3D Object Detection from Point Clouds (VS3D), ACM MM 2020
Unofficial Implementation of Google Deepmind's paper `Objects that Sound`
[CVPR 2023] Referring Image Matting
Code for journal paper "Learning Dual Semantic Relations with Graph Attention for Image-Text Matching", TCSVT, 2020.
[NAACL 2022]Mobile Text-to-Image search powered by multimodal semantic representation models(e.g., OpenAI's CLIP)
Official PyTorch implementation of our CVPR 2022 paper: Beyond a Pre-Trained Object Detector: Cross-Modal Textual and Visual Context for Image Captioning
Cross-modal convolutional neural networks
[Paper][AAAI 2023] DUET: Cross-modal Semantic Grounding for Contrastive Zero-shot Learning
Unleash the Potential of Image Branch for Cross-modal 3D Object Detection [NeurIPS2023]
This repository provides a comprehensive collection of research papers focused on multimodal representation learning, all of which have been cited and discussed in the survey just accepted https://dl.acm.org/doi/abs/10.1145/3617833 .
Code for paper "direct speech-to-image translation"
Add a description, image, and links to the cross-modal topic page so that developers can more easily learn about it.
To associate your repository with the cross-modal topic, visit your repo's landing page and select "manage topics."