Skip to content

Latest commit

 

History

History
85 lines (55 loc) · 5.23 KB

File metadata and controls

85 lines (55 loc) · 5.23 KB

Image Captioning With MobileNet-LLaMA 3

colab
architecture MobileNet V3 + LLaMA 3 architecture.

Image captioning is one of the problems in computer vision, constituting two kinds of modalities, i.e., image and text. Given a particular image, a caption regarding it is automatically generated. One can easily leverage a CNN-based architecture to draw the numerical representation out of the image. When interacting with the text, the long-range dependencies method has to be employed. Uplifted by the recent success of LLaMA 3, this project utilizes its computational block called the LLaMA 3 Transformer block. This block comprises RMSNorm, Grouped Multi-Query Attention, Feed Forward SwiGLU, and Rotary Position Embedding. Anyhow, in the original implementation, the Transformer block was only used as the decoder. In this project, the Transformer block is used as both the encoder and the decoder. In the encoder, before image data is funneled into the architecture, a CNN-based architecture, MobileNet-V3, is leveraged, acting similarly to the text embedding. Therefore, this architecture is dubbed MobileNet-LLaMA 3. To get knowledge on the performance of the model, the Flickr-8k dataset is used. The dataset is separated into the train, validation, and test sets in the 80-10-10 rule. Quantitatively, the performance of the model is measured via the ROUGE score, to be precise, the ROUGE-1 F-measure.

Experiment

Proceed to this notebook to vacate and answer your confusion and questions about this project by contemplating each line of code.

Result

Quantitative Result

The MobileNet-LLaMA3 performance on the test set is quantitatively displayed by the following table.

Test Metric Score
ROUGE-1 F-measure 36.69%

Loss Curve

loss_curve
Loss curves of the MobileNet-LLaMA 3 model on the train and validation sets.

Qualitative Result

The following image shows the qualitative results of MobileNet-LLaMA 3 on the test set.

qualitative
The image-caption pairs yielded from MobileNet-LLaMA 3.

The MobileNet-LLaMA 3 model is also assessed in the wild.

qualitative
The result of MobileNet-LLaMA 3 in the wild.

Citation

Feel free to cite this repository:

@misc{mobilenet-llama3,
   title = {Image Captioning With MobileNet-LLaMA 3},
   url = {https://github.com/reshalfahsi/image-captioning-mobilenet-llama3},
   author = {Resha Dwika Hefni Al-Fahsi},
}

Credit