This repo provides official implementations for LCMP & TAME:
LCMP is the first benchmark for evaluating Long-Context MLLM Personalization, which is more practical and conceptually different from existing static, single-turn visual-identification-oriented evaluation.
We will release the complete LCMP benchmark in data/concept, which includes personalized pets, people, and objects, along with their associated dialogues, questions, and personalized images.
We also provide code to reproduce LCMP by adding more personalized concepts.
-
Sample an image from the COCO dataset to
dataset_maker/concept_pet/$concept_id/base.png. -
Run the following code:
# generate concept and history
python dataset_maker/generate_profile_history_question.py --concept_id $concept_id
# generate personalized image
python dataset_maker/generate_prompt_images.py --concept_id $concept_idTAME is a training-free and state-aware personalized MLLM assistant powered by double memories.
-
Move the constructed concepts to
data/concept -
Run the following code:
# build memory on all concepts using InternVL3-8B
python method/main.py build --model internvl
# run TAME on all concepts using InternVL3-8B
python method/main.py qa --model internvl# run evaluation on generated answers
python evaluator/evaluator.py $input_file $output_filedataset_maker/: tools to construct/extend LCMP conceptsmethod/: TAME pipeline (memory building + QA)evaluator/: evaluation scriptsdata/: benchmark concepts
@inproceedings{hong2026tameing,
author = {Hong, Rongpei and Lang, Jian and Zhong, Ting and Wang, Yong and Zhou, Fan},
booktitle = {ACM SIGKDD Conference on Knowledge Discovery and Data Mining (KDD)},
year = {2026},
title = {TAMEing Long Contexts in Personalization: Towards Training-Free and State-Aware MLLM Personalized Assistant},
}
