Official repository for EarthMarker.
Authors: Wei Zhang*, Miaoxin Cai*, Tong Zhang, Yin Zhuang, and Xuerui Mao
- The authors contributed equally to this work.
- The dataset, model, code, and demo are coming soon! 🚀
- [2024.07.19]: The paper for EarthMarker is released arxiv. 🔥🔥
The first visual prompting MLLM named EarthMarker is proposed. EarthMarker can interpret RS imagery in the multi-turn conversation at different granularity, including image, region, and point levels, significantly catering to the fine-grained interpretation needs for RS imagery. EarthMarker is capable of various RS visual tasks including scene classification, referring object classification, captioning, and relationship analyses, beneficial to making informed decisions in real-world applications.
@article{zhang2024earthmarker,
title={EarthMarker: A Visual Prompting Multi-modal Large Language Model for Remote Sensing},
author={Zhang, Wei and Cai, Miaoxin and Zhang, Tong and Li, Jun and Zhuang, Yin and Mao, Xuerui},
journal={arXiv preprint arXiv:2407.13596},
year={2024}
}
This paper benefits from llama. Thanks for their wonderful work.