jasonaidm / ai_webui Goto Github PK
View Code? Open in Web Editor NEWAI-WEBUI: A universal web interface for AI creation, 一款好用的图像、音频、视频AI处理工具
License: MIT License
AI-WEBUI: A universal web interface for AI creation, 一款好用的图像、音频、视频AI处理工具
License: MIT License
home/parallels/anaconda3/envs/aiwebui/lib/python3.11/site-packages/gradio_client/documentation.py:103: UserWarning: Could not get documentation group for <class 'gradio.mix.Parallel'>: No known documentation group for module 'gradio.mix'
warnings.warn(f"Could not get documentation group for {cls}: {exc}")
/home/parallels/anaconda3/envs/aiwebui/lib/python3.11/site-packages/gradio_client/documentation.py:103: UserWarning: Could not get documentation group for <class 'gradio.mix.Series'>: No known documentation group for module 'gradio.mix'
warnings.warn(f"Could not get documentation group for {cls}: {exc}")
Caching examples at: '/home/parallels/Downloads/ai_webui-main/gradio_cached_examples/108'
Caching example 1/8
Ultralytics YOLOv8.0.120 🚀 Python-3.11.9 torch-2.1.1
Traceback (most recent call last):
File "/home/parallels/Downloads/ai_webui-main/webui.py", line 85, in
launch_webui(**opt.dict)
File "/home/parallels/Downloads/ai_webui-main/webui.py", line 52, in launch_webui
sam_tab(segmentation_args, ai_handler)
File "/home/parallels/Downloads/ai_webui-main/utils/gradio_tabs/image_tabs.py", line 50, in sam_tab
gr.Examples(examples=examples,
File "/home/parallels/anaconda3/envs/aiwebui/lib/python3.11/site-packages/gradio/helpers.py", line 74, in create_examples
client_utils.synchronize_async(examples_obj.create)
File "/home/parallels/anaconda3/envs/aiwebui/lib/python3.11/site-packages/gradio_client/utils.py", line 855, in synchronize_async
return fsspec.asyn.sync(fsspec.asyn.get_loop(), func, *args, **kwargs) # type: ignore
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/parallels/anaconda3/envs/aiwebui/lib/python3.11/site-packages/fsspec/asyn.py", line 103, in sync
raise return_result
File "/home/parallels/anaconda3/envs/aiwebui/lib/python3.11/site-packages/fsspec/asyn.py", line 56, in _runner
result[0] = await coro
^^^^^^^^^^
File "/home/parallels/anaconda3/envs/aiwebui/lib/python3.11/site-packages/gradio/helpers.py", line 276, in create
await self.cache()
File "/home/parallels/anaconda3/envs/aiwebui/lib/python3.11/site-packages/gradio/helpers.py", line 332, in cache
prediction = await Context.root_block.process_api(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/parallels/anaconda3/envs/aiwebui/lib/python3.11/site-packages/gradio/blocks.py", line 1392, in process_api
result = await self.call_function(
^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/parallels/anaconda3/envs/aiwebui/lib/python3.11/site-packages/gradio/blocks.py", line 1097, in call_function
prediction = await anyio.to_thread.run_sync(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/parallels/anaconda3/envs/aiwebui/lib/python3.11/site-packages/anyio/to_thread.py", line 56, in run_sync
return await get_async_backend().run_sync_in_worker_thread(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/parallels/anaconda3/envs/aiwebui/lib/python3.11/site-packages/anyio/_backends/_asyncio.py", line 2144, in run_sync_in_worker_thread
return await future
^^^^^^^^^^^^
File "/home/parallels/anaconda3/envs/aiwebui/lib/python3.11/site-packages/anyio/_backends/_asyncio.py", line 851, in run
result = context.run(func, *args)
^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/parallels/anaconda3/envs/aiwebui/lib/python3.11/site-packages/gradio/utils.py", line 703, in wrapper
response = f(*args, **kwargs)
^^^^^^^^^^^^^^^^^^
File "/home/parallels/Downloads/ai_webui-main/tools/fastsam_handler.py", line 43, in segment_everything
results = self.model(input,
^^^^^^^^^^^^^^^^^
File "/home/parallels/anaconda3/envs/aiwebui/lib/python3.11/site-packages/ultralytics/yolo/engine/model.py", line 111, in call
return self.predict(source, stream, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/parallels/anaconda3/envs/aiwebui/lib/python3.11/site-packages/torch/utils/_contextlib.py", line 115, in decorate_context
return func(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^
File "/home/parallels/anaconda3/envs/aiwebui/lib/python3.11/site-packages/ultralytics/yolo/engine/model.py", line 250, in predict
self.predictor.setup_model(model=self.model, verbose=is_cli)
File "/home/parallels/anaconda3/envs/aiwebui/lib/python3.11/site-packages/ultralytics/yolo/engine/predictor.py", line 297, in setup_model
device = select_device(self.args.device, verbose=verbose)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/parallels/anaconda3/envs/aiwebui/lib/python3.11/site-packages/ultralytics/yolo/utils/torch_utils.py", line 75, in select_device
raise ValueError(f"Invalid CUDA 'device={device}' requested."
ValueError: Invalid CUDA 'device=0' requested. Use 'device=cpu' or pass valid CUDA device(s) if available, i.e. 'device=0' or 'device=0,1,2,3' for Multi-GPU.
torch.cuda.is_available(): False
torch.cuda.device_count(): 0
os.environ['CUDA_VISIBLE_DEVICES']: None
See https://pytorch.org/get-started/locally/ for up-to-date torch install instructions if no CUDA devices are seen by torch.
IMPORTANT: You are using gradio version 3.39.0, however version 4.29.0 is available, please upgrade.
依赖下载不下来怎么办?
Traceback (most recent call last): File "D:\anaconda3\envs\aiwebui\Lib\site-packages\gradio\routes.py", line 442, in run_predict output = await app.get_blocks().process_api( ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\anaconda3\envs\aiwebui\Lib\site-packages\gradio\blocks.py", line 1392, in process_api result = await self.call_function( ^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\anaconda3\envs\aiwebui\Lib\site-packages\gradio\blocks.py", line 1097, in call_function prediction = await anyio.to_thread.run_sync( ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\anaconda3\envs\aiwebui\Lib\site-packages\anyio\to_thread.py", line 56, in run_sync return await get_async_backend().run_sync_in_worker_thread( ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\anaconda3\envs\aiwebui\Lib\site-packages\anyio\_backends\_asyncio.py", line 2134, in run_sync_in_worker_thread return await future ^^^^^^^^^^^^ File "D:\anaconda3\envs\aiwebui\Lib\site-packages\anyio\_backends\_asyncio.py", line 851, in run result = context.run(func, *args) ^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\anaconda3\envs\aiwebui\Lib\site-packages\gradio\utils.py", line 703, in wrapper response = f(*args, **kwargs) ^^^^^^^^^^^^^^^^^^ File "F:\AI\webui\tools\ai_wrapper.py", line 76, in clip_video asr_result= self.whisper_handler.infer(audio_stream_file2) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "F:\AI\webui\tools\whisper_handler.py", line 16, in infer self.init_model() File "F:\AI\webui\tools\whisper_handler.py", line 54, in init_model self.model = whisper.load_model(name=self.model_name, device=self.device, download_root=self.model_dir) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\anaconda3\envs\aiwebui\Lib\site-packages\whisper\__init__.py", line 146, in load_model checkpoint = torch.load(fp, map_location=device) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\anaconda3\envs\aiwebui\Lib\site-packages\torch\serialization.py", line 1014, in load raise pickle.UnpicklingError(UNSAFE_MESSAGE + str(e)) from None File "D:\anaconda3\envs\aiwebui\Lib\site-packages\torch\serialization.py", line 1422, in _load unpickler.persistent_load = persistent_load ^^^^^^^^^^^^^^^^ File "D:\anaconda3\envs\aiwebui\Lib\site-packages\torch\serialization.py", line 1392, in persistent_load nbytes = numel * torch._utils._element_size(dtype) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\anaconda3\envs\aiwebui\Lib\site-packages\torch\serialization.py", line 1366, in load_tensor typed_storage = torch.storage.TypedStorage( ^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\anaconda3\envs\aiwebui\Lib\site-packages\torch\serialization.py", line 1296, in restore_location def restore_location(storage, location): ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\anaconda3\envs\aiwebui\Lib\site-packages\torch\serialization.py", line 381, in default_restore_location result = fn(storage, location) ^^^^^^^^^^^^^^^^^^^^^ File "D:\anaconda3\envs\aiwebui\Lib\site-packages\torch\serialization.py", line 274, in _cuda_deserialize device = validate_cuda_device(location) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\anaconda3\envs\aiwebui\Lib\site-packages\torch\serialization.py", line 265, in validate_cuda_device raise RuntimeError('Attempting to deserialize object on CUDA device ' RuntimeError: Attempting to deserialize object on CUDA device 1 but torch.cuda.device_count() is 1. Please use torch.load with map_location to map your storages to an existing device.
When converting video, it prompts "Your platform doesn't support hardware accelerated AV1 decoding."
Ubuntu22.04 system, ffmpeg has been upgraded to the latest version and still the same
add new feature for supporting openai api
国产工具不容易
~/ai_webui$ python webui.py -c ./configs/webui_configs.yml
Traceback (most recent call last):
File "/home6/wutuo/ai_webui/webui.py", line 4, in
from tools import AIWrapper
File "/home6/wutuo/ai_webui/tools/init.py", line 6, in
from .ai_wrapper import AIWrapper
File "/home6/wutuo/ai_webui/tools/ai_wrapper.py", line 3, in
from .visualchat_handler import VisualChatHandler
ModuleNotFoundError: No module named 'tools.visualchat_handler'
A declarative, efficient, and flexible JavaScript library for building user interfaces.
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
An Open Source Machine Learning Framework for Everyone
The Web framework for perfectionists with deadlines.
A PHP framework for web artisans
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
Some thing interesting about web. New door for the world.
A server is a program made to process requests and deliver data to clients.
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
Some thing interesting about visualization, use data art
Some thing interesting about game, make everyone happy.
We are working to build community through open source technology. NB: members must have two-factor auth.
Open source projects and samples from Microsoft.
Google ❤️ Open Source for everyone.
Alibaba Open Source for everyone
Data-Driven Documents codes.
China tencent open source team.