LLM->多模态理论+案例讲解 Video-LLaVA 运行视频例子出错 #561
Labels
No Label
bug
duplicate
enhancement
help wanted
invalid
question
wontfix
No Milestone
No project
No Assignees
2 Participants
Notifications
Due Date
No due date set.
Dependencies
No dependencies set.
Reference: HswOAuth/llm_course#561
Loading…
Reference in New Issue
Block a user
No description provided.
Delete Branch "%!s()"
Deleting a branch is permanent. Although the deleted branch may continue to exist for a short time before it actually gets removed, it CANNOT be undone in most cases. Continue?
运行图像例子可以成功,运行视频例子会出现下面错误。
/root/autodl-tmp/Video-LLaVA/videollava/serve/gradio_web_server.py:175: GradioUnusedKwargWarning: You have unused kwarg parameters in Chatbot, please remove them: {'bubble_full_width': True}
chatbot = gr.Chatbot(label="Video-LLaVA", bubble_full_width=True).style(height=750)
/root/autodl-tmp/Video-LLaVA/videollava/serve/gradio_web_server.py:175: GradioDeprecationWarning: The
style
method is deprecated. Please set these arguments in the constructor instead.chatbot = gr.Chatbot(label="Video-LLaVA", bubble_full_width=True).style(height=750)
Running on local URL: http://127.0.0.1:7860
IMPORTANT: You are using gradio version 3.37.0, however version 4.44.1 is available, please upgrade.
Running on public URL: https://6977134dd9decf779f.gradio.live
This share link expires in 72 hours. For free permanent hosting and GPU upgrades, run
gradio deploy
from Terminal to deploy to Spaces (https://huggingface.co/spaces)A chat between a curious human and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the human's questions. USER:
Can you identify any safety hazards in this video? ASSISTANT: tensor([[ 1, 319, 13563, 1546, 263, 12758, 5199, 322, 385, 23116,
21082, 20255, 29889, 450, 20255, 4076, 8444, 29892, 13173, 29892,
322, 1248, 568, 6089, 304, 278, 5199, 29915, 29879, 5155,
29889, 3148, 1001, 29901, 29871, -200, -200, -200, -200, -200,
-200, -200, -200, 29871, 13, 6028, 366, 12439, 738, 15332,
447, 29920, 3163, 297, 445, 4863, 29973, 319, 1799, 9047,
13566, 29901]], device='cuda:0') 1 torch.Size([3, 8, 224, 224])
../aten/src/ATen/native/cuda/IndexKernel.cu:92: operator(): block: [176,0,0], thread: [0,0,0] Assertion
index >= -sizes[i] && index < sizes[i] && "index out of bounds"
failed.../aten/src/ATen/native/cuda/IndexKernel.cu:92: operator(): block: [176,0,0], thread: [1,0,0] Assertion
index >= -sizes[i] && index < sizes[i] && "index out of bounds"
failed.../aten/src/ATen/native/cuda/IndexKernel.cu:92: operator(): block: [176,0,0], thread: [2,0,0] Assertion
index >= -sizes[i] && index < sizes[i] && "index out of bounds"
failed.../aten/src/ATen/native/cuda/IndexKernel.cu:92: operator(): block: [176,0,0], thread: [3,0,0] Assertion
index >= -sizes[i] && index < sizes[i] && "index out of bounds"
failed.../aten/src/ATen/native/cuda/IndexKernel.cu:92: operator(): block: [176,0,0], thread: [4,0,0] Assertion
index >= -sizes[i] && index < sizes[i] && "index out of bounds"
failed.../aten/src/ATen/native/cuda/IndexKernel.cu:92: operator(): block: [176,0,0], thread: [5,0,0] Assertion
index >= -sizes[i] && index < sizes[i] && "index out of bounds"
failed.../aten/src/ATen/native/cuda/IndexKernel.cu:92: operator(): block: [176,0,0], thread: [6,0,0] Assertion
index >= -sizes[i] && index < sizes[i] && "index out of bounds"
failed.../aten/src/ATen/native/cuda/IndexKernel.cu:92: operator(): block: [176,0,0], thread: [7,0,0] Assertion
index >= -sizes[i] && index < sizes[i] && "index out of bounds"
failed.../aten/src/ATen/native/cuda/IndexKernel.cu:92: operator(): block: [176,0,0], thread: [8,0,0] Assertion
index >= -sizes[i] && index < sizes[i] && "index out of bounds"
failed.../aten/src/ATen/native/cuda/IndexKernel.cu:92: operator(): block: [176,0,0], thread: [9,0,0] Assertion
index >= -sizes[i] && index < sizes[i] && "index out of bounds"
failed.../aten/src/ATen/native/cuda/IndexKernel.cu:92: operator(): block: [176,0,0], thread: [10,0,0] Assertion
index >= -sizes[i] && index < sizes[i] && "index out of bounds"
failed.../aten/src/ATen/native/cuda/IndexKernel.cu:92: operator(): block: [176,0,0], thread: [11,0,0] Assertion
index >= -sizes[i] && index < sizes[i] && "index out of bounds"
failed.../aten/src/ATen/native/cuda/IndexKernel.cu:92: operator(): block: [176,0,0], thread: [12,0,0] Assertion
index >= -sizes[i] && index < sizes[i] && "index out of bounds"
failed.../aten/src/ATen/native/cuda/IndexKernel.cu:92: operator(): block: [176,0,0], thread: [13,0,0] Assertion
index >= -sizes[i] && index < sizes[i] && "index out of bounds"
failed.../aten/src/ATen/native/cuda/IndexKernel.cu:92: operator(): block: [176,0,0], thread: [14,0,0] Assertion
index >= -sizes[i] && index < sizes[i] && "index out of bounds"
failed.../aten/src/ATen/native/cuda/IndexKernel.cu:92: operator(): block: [176,0,0], thread: [15,0,0] Assertion
index >= -sizes[i] && index < sizes[i] && "index out of bounds"
failed.../aten/src/ATen/native/cuda/IndexKernel.cu:92: operator(): block: [176,0,0], thread: [16,0,0] Assertion
index >= -sizes[i] && index < sizes[i] && "index out of bounds"
failed.../aten/src/ATen/native/cuda/IndexKernel.cu:92: operator(): block: [176,0,0], thread: [17,0,0] Assertion
index >= -sizes[i] && index < sizes[i] && "index out of bounds"
failed.../aten/src/ATen/native/cuda/IndexKernel.cu:92: operator(): block: [176,0,0], thread: [18,0,0] Assertion
index >= -sizes[i] && index < sizes[i] && "index out of bounds"
failed.../aten/src/ATen/native/cuda/IndexKernel.cu:92: operator(): block: [176,0,0], thread: [19,0,0] Assertion
index >= -sizes[i] && index < sizes[i] && "index out of bounds"
failed.../aten/src/ATen/native/cuda/IndexKernel.cu:92: operator(): block: [176,0,0], thread: [20,0,0] Assertion
index >= -sizes[i] && index < sizes[i] && "index out of bounds"
failed.../aten/src/ATen/native/cuda/IndexKernel.cu:92: operator(): block: [176,0,0], thread: [21,0,0] Assertion
index >= -sizes[i] && index < sizes[i] && "index out of bounds"
failed.../aten/src/ATen/native/cuda/IndexKernel.cu:92: operator(): block: [176,0,0], thread: [22,0,0] Assertion
index >= -sizes[i] && index < sizes[i] && "index out of bounds"
failed.../aten/src/ATen/native/cuda/IndexKernel.cu:92: operator(): block: [176,0,0], thread: [23,0,0] Assertion
index >= -sizes[i] && index < sizes[i] && "index out of bounds"
failed.../aten/src/ATen/native/cuda/IndexKernel.cu:92: operator(): block: [176,0,0], thread: [24,0,0] Assertion
index >= -sizes[i] && index < sizes[i] && "index out of bounds"
failed.../aten/src/ATen/native/cuda/IndexKernel.cu:92: operator(): block: [176,0,0], thread: [25,0,0] Assertion
index >= -sizes[i] && index < sizes[i] && "index out of bounds"
failed.../aten/src/ATen/native/cuda/IndexKernel.cu:92: operator(): block: [176,0,0], thread: [26,0,0] Assertion
index >= -sizes[i] && index < sizes[i] && "index out of bounds"
failed.../aten/src/ATen/native/cuda/IndexKernel.cu:92: operator(): block: [176,0,0], thread: [27,0,0] Assertion
index >= -sizes[i] && index < sizes[i] && "index out of bounds"
failed.../aten/src/ATen/native/cuda/IndexKernel.cu:92: operator(): block: [176,0,0], thread: [28,0,0] Assertion
index >= -sizes[i] && index < sizes[i] && "index out of bounds"
failed.../aten/src/ATen/native/cuda/IndexKernel.cu:92: operator(): block: [176,0,0], thread: [29,0,0] Assertion
index >= -sizes[i] && index < sizes[i] && "index out of bounds"
failed.../aten/src/ATen/native/cuda/IndexKernel.cu:92: operator(): block: [176,0,0], thread: [30,0,0] Assertion
index >= -sizes[i] && index < sizes[i] && "index out of bounds"
failed.../aten/src/ATen/native/cuda/IndexKernel.cu:92: operator(): block: [176,0,0], thread: [31,0,0] Assertion
index >= -sizes[i] && index < sizes[i] && "index out of bounds"
failed.Traceback (most recent call last):
File "/root/miniconda3/envs/videollava/lib/python3.10/site-packages/gradio/routes.py", line 439, in run_predict
output = await app.get_blocks().process_api(
File "/root/miniconda3/envs/videollava/lib/python3.10/site-packages/gradio/blocks.py", line 1389, in process_api
result = await self.call_function(
File "/root/miniconda3/envs/videollava/lib/python3.10/site-packages/gradio/blocks.py", line 1094, in call_function
prediction = await anyio.to_thread.run_sync(
File "/root/miniconda3/envs/videollava/lib/python3.10/site-packages/anyio/to_thread.py", line 56, in run_sync
return await get_async_backend().run_sync_in_worker_thread(
File "/root/miniconda3/envs/videollava/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 2357, in run_sync_in_worker_thread
return await future
File "/root/miniconda3/envs/videollava/lib/python3.10/site-packages/anyio/backends/asyncio.py", line 864, in run
result = context.run(func, *args)
File "/root/miniconda3/envs/videollava/lib/python3.10/site-packages/gradio/utils.py", line 704, in wrapper
response = f(*args, **kwargs)
File "/root/autodl-tmp/Video-LLaVA/videollava/serve/gradio_web_server.py", line 87, in generate
text_en_out, state = handler.generate(images_tensor, text_en_in, first_run=first_run, state=state)
File "/root/miniconda3/envs/videollava/lib/python3.10/site-packages/torch/utils/_contextlib.py", line 115, in decorate_context
return func(*args, **kwargs)
File "/root/autodl-tmp/Video-LLaVA/videollava/serve/gradio_utils.py", line 92, in generate
output_ids = model.generate(
File "/root/miniconda3/envs/videollava/lib/python3.10/site-packages/torch/utils/_contextlib.py", line 115, in decorate_context
return func(*args, **kwargs)
File "/root/miniconda3/envs/videollava/lib/python3.10/site-packages/transformers/generation/utils.py", line 1588, in generate
return self.sample(
File "/root/miniconda3/envs/videollava/lib/python3.10/site-packages/transformers/generation/utils.py", line 2642, in sample
outputs = self(
File "/root/miniconda3/envs/videollava/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1501, in _call_impl
return forward_call(*args, **kwargs)
File "/root/miniconda3/envs/videollava/lib/python3.10/site-packages/accelerate/hooks.py", line 165, in new_forward
output = old_forward(*args, **kwargs)
File "/root/autodl-tmp/Video-LLaVA/videollava/model/language_model/llava_llama.py", line 79, in forward
) = self.prepare_inputs_labels_for_multimodal(
File "/root/autodl-tmp/Video-LLaVA/videollava/model/llava_arch.py", line 207, in prepare_inputs_labels_for_multimodal
video_features_minibatch = self.encode_videos(videos_minibatch) # fake list [mini_b, t, l, c]
File "/root/autodl-tmp/Video-LLaVA/videollava/model/llava_arch.py", line 145, in encode_videos
video_features = self.get_model().mm_projector(video_features)
File "/root/miniconda3/envs/videollava/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1501, in _call_impl
return forward_call(*args, **kwargs)
File "/root/miniconda3/envs/videollava/lib/python3.10/site-packages/accelerate/hooks.py", line 165, in new_forward
output = old_forward(*args, **kwargs)
File "/root/miniconda3/envs/videollava/lib/python3.10/site-packages/torch/nn/modules/container.py", line 217, in forward
input = module(input)
File "/root/miniconda3/envs/videollava/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1501, in _call_impl
return forward_call(*args, **kwargs)
File "/root/miniconda3/envs/videollava/lib/python3.10/site-packages/accelerate/hooks.py", line 165, in new_forward
output = old_forward(*args, **kwargs)
File "/root/miniconda3/envs/videollava/lib/python3.10/site-packages/bitsandbytes/nn/modules.py", line 441, in forward
out = bnb.matmul(x, self.weight, bias=self.bias, state=self.state)
File "/root/miniconda3/envs/videollava/lib/python3.10/site-packages/bitsandbytes/autograd/_functions.py", line 563, in matmul
return MatMul8bitLt.apply(A, B, out, bias, state)
File "/root/miniconda3/envs/videollava/lib/python3.10/site-packages/torch/autograd/function.py", line 506, in apply
return super().apply(*args, **kwargs) # type: ignore[misc]
File "/root/miniconda3/envs/videollava/lib/python3.10/site-packages/bitsandbytes/autograd/_functions.py", line 388, in forward
CAt[:, state.idx.long()] = 0
RuntimeError: CUDA error: device-side assert triggered
CUDA kernel errors might be asynchronously reported at some other API call, so the stacktrace below might be incorrect.
For debugging consider passing CUDA_LAUNCH_BLOCKING=1.
Compile with
TORCH_USE_CUDA_DSA
to enable device-side assertions.RuntimeError: CUDA error: device-side assert triggered
CUDA kernel errors might be asynchronously reported at some other API call, so the stacktrace below might be incorrect.
For debugging consider passing CUDA_LAUNCH_BLOCKING=1.
Compile with TORCH_USE_CUDA_DSA to enable device-side assertions.
根据这个报错,可能是cuda版本和安装的pytorch的不匹配;也可以重启下看看。