From 44fb2414f24298e88c16acae6b147f79bd416d9d Mon Sep 17 00:00:00 2001 From: Junhao Liang <43094337+darkliang@users.noreply.github.com> Date: Sun, 29 Oct 2023 21:16:09 +0800 Subject: [PATCH] fix up --- datasets/GREW/README.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/datasets/GREW/README.md b/datasets/GREW/README.md index 6ebf4a4..9c20272 100644 --- a/datasets/GREW/README.md +++ b/datasets/GREW/README.md @@ -83,4 +83,4 @@ CUDA_VISIBLE_DEVICES=0,1,2,3 python -m torch.distributed.launch --nproc_per_node The result will be generated in your working directory, you must rename and compress it as the requirements before submitting. ## Evaluation locally -While the original grew treat both seq_01 and seq_02 as gallery, but there is no ground truth for probe. Therefore, it is nessesary to upload the submission file on grew competitation. We seperate test set to: seq_01 as gallery, seq_02 as probe. Then you can modify `eval_func` in the `./config/baseline/baseline_GREW.yaml` to `identification_real_scene`, you can obtain result localy like setting of OUMVLP. +While the original grew treat both seq_01 and seq_02 as gallery, but there is no ground truth for probe. Therefore, it is nessesary to upload the submission file on grew competitation. We seperate test set to: seq_01 as gallery, seq_02 as probe. Then you can modify `eval_func` in the `./config/baseline/baseline_GREW.yaml` to `evaluate_real_scene`, you can obtain result localy like setting of OUMVLP.