qwmqza commited on
Commit
c192db9
·
1 Parent(s): 1ac9cff
__pycache__/app.cpython-310.pyc ADDED
Binary file (1.19 kB). View file
 
llava/__pycache__/__init__.cpython-310.pyc CHANGED
Binary files a/llava/__pycache__/__init__.cpython-310.pyc and b/llava/__pycache__/__init__.cpython-310.pyc differ
 
llava/__pycache__/constants.cpython-310.pyc CHANGED
Binary files a/llava/__pycache__/constants.cpython-310.pyc and b/llava/__pycache__/constants.cpython-310.pyc differ
 
llava/__pycache__/conversation.cpython-310.pyc CHANGED
Binary files a/llava/__pycache__/conversation.cpython-310.pyc and b/llava/__pycache__/conversation.cpython-310.pyc differ
 
llava/__pycache__/mm_utils.cpython-310.pyc CHANGED
Binary files a/llava/__pycache__/mm_utils.cpython-310.pyc and b/llava/__pycache__/mm_utils.cpython-310.pyc differ
 
llava/__pycache__/utils.cpython-310.pyc CHANGED
Binary files a/llava/__pycache__/utils.cpython-310.pyc and b/llava/__pycache__/utils.cpython-310.pyc differ
 
llava/eval/__pycache__/model_vqa_demo.cpython-310.pyc CHANGED
Binary files a/llava/eval/__pycache__/model_vqa_demo.cpython-310.pyc and b/llava/eval/__pycache__/model_vqa_demo.cpython-310.pyc differ
 
llava/eval/model_vqa_demo.py CHANGED
@@ -31,10 +31,8 @@ def get_chunk(lst, n, k):
31
  def eval_model(image, prompt):
32
  parser = argparse.ArgumentParser()
33
  args = parser.parse_args()
34
- #args.model_path = '/mnt/disk1/zian/LLaVA/checkpoints/llava-v1.5-7b-my_merged-finetune-mixed-lora-test-0628-1'
35
- #args.model_path = None
36
- args.model_base = None
37
- args.model_path = 'teowu/llava_v1.5_7b_qinstruct_preview_v0.1'
38
  args.image = image
39
  args.prompt = prompt
40
  args.conv_mode = "llava_v1"
@@ -95,4 +93,4 @@ if __name__ == "__main__":
95
  prompt = 'Please describe this image.'
96
 
97
  outputs = eval_model(image, prompt)
98
- print("processed.")
 
31
  def eval_model(image, prompt):
32
  parser = argparse.ArgumentParser()
33
  args = parser.parse_args()
34
+ args.model_path = 'qwmqza/VLC_llava_v1.5_cot_lora_v0'
35
+ args.model_base = 'teowu/llava_v1.5_7b_qinstruct_preview_v0.1'
 
 
36
  args.image = image
37
  args.prompt = prompt
38
  args.conv_mode = "llava_v1"
 
93
  prompt = 'Please describe this image.'
94
 
95
  outputs = eval_model(image, prompt)
96
+ print(outputs)
llava/eval/model_vqa_vlc.py ADDED
@@ -0,0 +1,101 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import argparse
2
+ import torch
3
+ torch.backends.cuda.enable_mem_efficient_sdp(False)
4
+ torch.backends.cuda.enable_flash_sdp(False)
5
+ import os
6
+ import json
7
+ from tqdm import tqdm
8
+ import shortuuid
9
+
10
+ from llava.constants import IMAGE_TOKEN_INDEX, DEFAULT_IMAGE_TOKEN, DEFAULT_IM_START_TOKEN, DEFAULT_IM_END_TOKEN
11
+ from llava.conversation import conv_templates, SeparatorStyle
12
+ from llava.model.builder import load_pretrained_model
13
+ from llava.utils import disable_torch_init
14
+ from llava.mm_utils import tokenizer_image_token, process_images, get_model_name_from_path
15
+
16
+ from PIL import Image
17
+ import math
18
+
19
+
20
+ def split_list(lst, n):
21
+ """Split a list into n (roughly) equal-sized chunks"""
22
+ chunk_size = math.ceil(len(lst) / n) # integer division
23
+ return [lst[i:i+chunk_size] for i in range(0, len(lst), chunk_size)]
24
+
25
+
26
+ def get_chunk(lst, n, k):
27
+ chunks = split_list(lst, n)
28
+ return chunks[k]
29
+
30
+
31
+ def eval_model(image, prompt, args):
32
+
33
+ args.model_base = 'teowu/llava_v1.5_7b_qinstruct_preview_v0.1'
34
+ args.image = image
35
+ args.prompt = prompt
36
+ args.conv_mode = "llava_v1"
37
+ args.num_chunks = 1
38
+ args.chunk_idx = 0
39
+ args.temperature = 0.2
40
+ args.top_p = None
41
+ args.num_beams = 1
42
+ # Model
43
+ disable_torch_init()
44
+ model_path = os.path.expanduser(args.model_path)
45
+ model_name = get_model_name_from_path(model_path)
46
+ tokenizer, model, image_processor, context_len = load_pretrained_model(model_path, args.model_base, model_name)
47
+ model.to(dtype=torch.bfloat16)
48
+
49
+ idx = 'test'
50
+ image_file = args.image
51
+ qs = args.prompt
52
+ cur_prompt = qs
53
+ if model.config.mm_use_im_start_end:
54
+ qs = DEFAULT_IM_START_TOKEN + DEFAULT_IMAGE_TOKEN + DEFAULT_IM_END_TOKEN + '\n' + qs
55
+ else:
56
+ qs = DEFAULT_IMAGE_TOKEN + '\n' + qs
57
+
58
+ conv = conv_templates[args.conv_mode].copy()
59
+ conv.append_message(conv.roles[0], qs)
60
+ conv.append_message(conv.roles[1], None)
61
+ prompt = conv.get_prompt()
62
+
63
+ input_ids = tokenizer_image_token(prompt, tokenizer, IMAGE_TOKEN_INDEX, return_tensors='pt').unsqueeze(0).cuda()
64
+
65
+ image = Image.open(image_file).convert('RGB')
66
+ image_tensor = process_images([image], image_processor, model.config)[0]
67
+ image_tensor = image_tensor.to(model.device, dtype=torch.bfloat16)
68
+ print(qs)
69
+ print(torch.version.cuda)
70
+
71
+ with torch.inference_mode():
72
+ output_ids = model.generate(
73
+ input_ids,
74
+ images=image_tensor.unsqueeze(0).cuda(),
75
+ image_sizes=[image.size],
76
+ do_sample=True if args.temperature > 0 else False,
77
+ temperature=args.temperature,
78
+ top_p=args.top_p,
79
+ num_beams=args.num_beams,
80
+ # no_repeat_ngram_size=3,
81
+ max_new_tokens=1024,
82
+ use_cache=True)
83
+
84
+ outputs = tokenizer.batch_decode(output_ids, skip_special_tokens=True)[0].strip()
85
+ print(outputs)
86
+ return outputs
87
+
88
+
89
+ if __name__ == "__main__":
90
+ parser = argparse.ArgumentParser()
91
+ args = parser.parse_args()
92
+ parser.add_argument("--model-path", type=str, default="./checkpoints/VLC_llava_v1.5_cot_lora_v0")
93
+ parser.add_argument("--model-base", type=str, default=None)
94
+ parser.add_argument("--image-folder", type=str, default="")
95
+ parser.add_argument("--question-file", type=str, default="tables/question.jsonl")
96
+ image = '/mnt/disk1/zian/LLaVA/my_test_image/blur_image1.jpg'
97
+ prompt = 'Please describe this image.'
98
+ args.model_path = '/mnt/disk1/zian/LLaVA/checkpoints/qinstruct_llava_v1.5_cot_lora_v5'
99
+
100
+ outputs = eval_model(image, prompt, args)
101
+ print(outputs)
llava/model/__pycache__/__init__.cpython-310.pyc CHANGED
Binary files a/llava/model/__pycache__/__init__.cpython-310.pyc and b/llava/model/__pycache__/__init__.cpython-310.pyc differ
 
llava/model/__pycache__/builder.cpython-310.pyc CHANGED
Binary files a/llava/model/__pycache__/builder.cpython-310.pyc and b/llava/model/__pycache__/builder.cpython-310.pyc differ
 
llava/model/__pycache__/llava_arch.cpython-310.pyc CHANGED
Binary files a/llava/model/__pycache__/llava_arch.cpython-310.pyc and b/llava/model/__pycache__/llava_arch.cpython-310.pyc differ
 
llava/model/language_model/__pycache__/llava_llama.cpython-310.pyc CHANGED
Binary files a/llava/model/language_model/__pycache__/llava_llama.cpython-310.pyc and b/llava/model/language_model/__pycache__/llava_llama.cpython-310.pyc differ
 
llava/model/language_model/__pycache__/llava_mistral.cpython-310.pyc CHANGED
Binary files a/llava/model/language_model/__pycache__/llava_mistral.cpython-310.pyc and b/llava/model/language_model/__pycache__/llava_mistral.cpython-310.pyc differ
 
llava/model/language_model/__pycache__/llava_mpt.cpython-310.pyc CHANGED
Binary files a/llava/model/language_model/__pycache__/llava_mpt.cpython-310.pyc and b/llava/model/language_model/__pycache__/llava_mpt.cpython-310.pyc differ
 
llava/model/multimodal_encoder/__pycache__/builder.cpython-310.pyc CHANGED
Binary files a/llava/model/multimodal_encoder/__pycache__/builder.cpython-310.pyc and b/llava/model/multimodal_encoder/__pycache__/builder.cpython-310.pyc differ
 
llava/model/multimodal_encoder/__pycache__/clip_encoder.cpython-310.pyc CHANGED
Binary files a/llava/model/multimodal_encoder/__pycache__/clip_encoder.cpython-310.pyc and b/llava/model/multimodal_encoder/__pycache__/clip_encoder.cpython-310.pyc differ
 
llava/model/multimodal_projector/__pycache__/builder.cpython-310.pyc CHANGED
Binary files a/llava/model/multimodal_projector/__pycache__/builder.cpython-310.pyc and b/llava/model/multimodal_projector/__pycache__/builder.cpython-310.pyc differ