linux做网站配置,小程序加盟代理平台,东莞住建局电话是多少,怎么做自动下单网站目录
生成视频代码#xff1a;
维度报错#xff1a;
解决方法#xff0c;修改代码#xff1a; 已开源#xff1a; 视频生成模型 Zeroscope开源 免费无水印 视频生成模型 Zeroscope_v2_576w 开源 - 腾讯云开发者社区-腾讯云 生成视频代码#xff1a;
import torch
fro…目录
生成视频代码
维度报错
解决方法修改代码 已开源 视频生成模型 Zeroscope开源 免费无水印 视频生成模型 Zeroscope_v2_576w 开源 - 腾讯云开发者社区-腾讯云 生成视频代码
import torch
from diffusers import DiffusionPipeline, DPMSolverMultistepScheduler
from diffusers.utils import export_to_video
import os
# os.environ[HTTP_PROXY] http://127.0.0.1:7890os.environ[HF_TOKEN] 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
# os.environ[HTTPS_PROXY] https://127.0.0.1:7890# pipe DiffusionPipeline.from_pretrained(rD:\360安全浏览器下载, torch_dtypetorch.float16)
pipe DiffusionPipeline.from_pretrained(cerspense/zeroscope_v2_576w, torch_dtypetorch.float16,use_auth_tokenos.environ[HF_TOKEN])
pipe.scheduler DPMSolverMultistepScheduler.from_config(pipe.scheduler.config)
pipe.enable_model_cpu_offload()prompt Darth Vader is surfing on waves
video_frames pipe(prompt, num_inference_steps40, height320, width576, num_frames24).frames
video_path export_to_video(video_frames)
print(video_path)维度报错
Traceback (most recent call last):File E:\project\jijia\aaa.py, line 18, in modulevideo_path export_to_video(video_frames)File D:\ProgramData\miniconda3\envs\pysd\lib\site-packages\diffusers\utils\export_utils.py, line 135, in export_to_videoh, w, c video_frames[0].shape
ValueError: too many values to unpack (expected 3) 解决方法修改代码 def export_to_video(video_frames: Union[List[np.ndarray], List[PIL.Image.Image]], output_video_path: str None, fps: int 10
) - str:if is_opencv_available():import cv2else:raise ImportError(BACKENDS_MAPPING[opencv][1].format(export_to_video))if output_video_path is None:output_video_path tempfile.NamedTemporaryFile(suffix.mp4).name# Convert PIL images to numpy arrays if neededif isinstance(video_frames[0], PIL.Image.Image):video_frames [np.array(frame) for frame in video_frames]# Ensure the frames are in the correct formatif isinstance(video_frames[0], np.ndarray):# Check if frames are 4-dimensional and handle accordinglyif len(video_frames[0].shape) 4:video_frames [frame[0] for frame in video_frames]# Convert frames to uint8video_frames [(frame * 255).astype(np.uint8) for frame in video_frames]# Ensure all frames are in (height, width, channels) formath, w, c video_frames[0].shapefourcc cv2.VideoWriter_fourcc(*mp4v)video_writer cv2.VideoWriter(output_video_path, fourcc, fpsfps, frameSize(w, h))for frame in video_frames:img cv2.cvtColor(frame, cv2.COLOR_RGB2BGR)video_writer.write(img)video_writer.release()return output_video_pathdef export_to_video_o(video_frames: Union[List[np.ndarray], List[PIL.Image.Image]], output_video_path: str None, fps: int 10
) - str:if is_opencv_available():import cv2else:raise ImportError(BACKENDS_MAPPING[opencv][1].format(export_to_video))if output_video_path is None:output_video_path tempfile.NamedTemporaryFile(suffix.mp4).nameif isinstance(video_frames[0], np.ndarray):video_frames [(frame * 255).astype(np.uint8) for frame in video_frames]elif isinstance(video_frames[0], PIL.Image.Image):video_frames [np.array(frame) for frame in video_frames]fourcc cv2.VideoWriter_fourcc(*mp4v)h, w, c video_frames[0].shapevideo_writer cv2.VideoWriter(output_video_path, fourcc, fpsfps, frameSize(w, h))for i in range(len(video_frames)):img cv2.cvtColor(video_frames[i], cv2.COLOR_RGB2BGR)video_writer.write(img)return output_video_path