Giter Site home page Giter Site logo

uminosachi / open-llm-webui Goto Github PK

View Code? Open in Web Editor NEW
30.0 3.0 4.0 1.22 MB

This repository contains a web application designed to execute relatively compact, locally-operated Large Language Models (LLMs).

License: Apache License 2.0

Python 100.00%
language-model llm huggingface transformers gradio nlp llama chatbot llama2 llama3

open-llm-webui's Introduction

Open LLM WebUI

This repository contains a web application designed to execute relatively compact, locally-operated Large Language Models (LLMs).

Installation

Please follow these steps to install the software:

  • Create a new conda environment:
conda create -n ollm python=3.10
conda activate ollm
  • Clone the software repository:
git clone https://github.com/Uminosachi/open-llm-webui.git
cd open-llm-webui

Python Package Installation

General Instructions

  • Install the necessary Python packages by executing:
    pip install -r requirements.txt

Installation for Flash Attention

  • To enable Flash Attention in some models, if CUDA is available, install Flash Attention:
    pip install packaging ninja
    pip install flash-attn --no-build-isolation

Platform-Specific Instructions

  • For Windows (with CUDA support):

    • Install Visual Studio:
      • โš ๏ธ Important: Make sure to select Desktop development with C++ during the installation process.
    • Copy MSBuild extensions for CUDA as an administrator (adjust the CUDA version as necessary):
      xcopy /e "C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v11.8\extras\visual_studio_integration\MSBuildExtensions" "C:\Program Files\Microsoft Visual Studio\2022\Community\MSBuild\Microsoft\VC\v170\BuildCustomizations"
    • Configure the required environment variables for the build (adjust the CUDA version as necessary):
      set PATH=C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v11.8\bin;%PATH%
      "C:\Program Files\Microsoft Visual Studio\2022\Community\VC\Auxiliary\Build\vcvars64.bat"
      set FORCE_CMAKE=1 && set CMAKE_ARGS="-DLLAMA_CUDA=on"
    • Install the necessary Python packages (this process may take some time):
      pip install ninja cmake scikit-build-core[pyproject]
      pip install --force-reinstall --no-cache-dir llama-cpp-python
      pip install -r requirements.txt
  • For Linux (with CUDA support):

    • Configure the required environment variables for the build (if not already set):
      export PATH=/usr/local/cuda/bin:${PATH}
    • Install the necessary Python packages:
      CMAKE_ARGS="-DLLAMA_CUDA=on" pip install --force-reinstall --no-cache-dir llama-cpp-python
      pip install -r requirements.txt
  • For Mac OS (without CUDA support):

    • Install the necessary Python packages:
      BUILD_CUDA_EXT=0 pip install -r requirements.txt
    • Rebuild the bitsandbytes package with the CPU option.
      pip uninstall bitsandbytes
      git clone https://github.com/TimDettmers/bitsandbytes.git
      cd bitsandbytes
      cmake -DCOMPUTE_BACKEND=cpu -S .
      make
      pip install .
    • Known Issue: Running the LLaVA model on Mac results in an error.

Running the application

python ollm_app.py

Downloading the Model

To download the model:

  • Launch this application.
  • Click on the "Download model" button next to the LLM model ID.
  • Wait for the download to complete.

๐Ÿ“œ Model List (transformers)

Provider Model Names
Microsoft Phi-3-mini-4k-instruct
Google gemma-2-9b-it, gemma-1.1-2b-it, gemma-1.1-7b-it
NVIDIA Llama3-ChatQA-1.5-8B
Qwen Qwen2-7B-Instruct
Mistral AI Mistral-7B-Instruct-v0.3
Rakuten RakutenAI-7B-chat, RakutenAI-7B-instruct
rinna youri-7b-chat
TheBloke Llama-2-7b-Chat-GPTQ, Kunoichi-7B-GPTQ
  • ๐Ÿ“‹ Note: By adding the repository paths of models to model_manager/add_tfs_models.txt, they will be included in the list of Model IDs and displayed in the UI.
  • ๐Ÿ” Note: The downloaded model file will be stored in the .cache/huggingface/hub directory of your home directory.

Access and Download Models

Login to Hugging Face

  • Before downloading any models, please log in via the command line using:
    huggingface-cli login
    

๐Ÿฆ™ Model List (llama.cpp)

Provider Model Names
Microsoft Phi-3-mini-4k-instruct-q4.gguf, Phi-3-mini-4k-instruct-fp16.gguf
TheBloke llama-2-7b-chat.Q4_K_M.gguf
QuantFactory Meta-Llama-3-8B-Instruct.Q4_K_M.gguf

Using any GGUF file

  • ๐Ÿ” File Placement: Place files with the .gguf extension in the models directory within the open-llm-webui folder. These files will then appear in the model list on the llama.cpp tab of the web UI and can be used accordingly.
  • ๐Ÿ“ Metadata Usage: If the metadata of a GGUF model includes tokenizer.chat_template, this template will be used to create the prompts.

๐Ÿ–ผ๏ธ Model List (Multimodal LLaVA)

Provider Model Names
Microsoft Phi-3-vision-128k-instruct
llava-hf llava-v1.6-mistral-7b-hf, llava-v1.6-vicuna-7b-hf, llava-1.5-7b-hf
tinyllava TinyLLaVA-Phi-2-SigLIP-3.1B
openbmb MiniCPM-Llama3-V-2_5-int4, MiniCPM-Llama3-V-2_5
SakanaAI EvoVLM-JP-v1-7B
xtuner llava-llama-3-8b-v1_1-transformers
cyberagent llava-calm2-siglip

Usage

  • Enter your message into the "Input text" box. Adjust the slider for "Max new tokens" as needed.
  • Under "Advanced options" adjust the settings for "Temperature", "Top k", "Top p", and "Repetition Penalty" as needed.
  • If replacing the system message of the prompt, under "Advanced options" enable the checkbox and enter text.
  • Press "Enter" on your keyboard or click the "Generate" button.
    • โš ๏ธ Note: If the cloud-based model has been updated, it may be downloaded upon execution.
  • If you click the "Clear chat" button, the chat history will be cleared.

transformers tab

  • By enabling the CPU execution checkbox, the model will use the argument device_map="cpu".
  • Some of the transformers models are loaded with the following 4-bit or 8-bit settings using the bitsandbytes package.

llama.cpp tab

  • Use the radio buttons in the Default chat template to select the template that will be used if the GGUF model lacks a chat_template.

LLaVA tab

  • You can upload an image to the LLaVA Image area of this tab and input a prompt related to the image.
  • Some of the LLaVA models are loaded with the following 4-bit or 8-bit settings using the bitsandbytes package.

options

  • When you enable the Translate (ja->en/en->ja) checkbox:
    • Any input in Japanese will be automatically translated to English, and responses in English will be automatically translated back into Japanese.
    • โš ๏ธ Note: Downloading the translation model for the first time may take some time.

UI image

Model Credit

Developer Model License
Microsoft Phi-3 The MIT License
Google Gemma Gemma Terms of Use
NVIDIA Llama3-ChatQA Llama 3 Community License
Alibaba Group Qwen2-7B-Instruct Apache License 2.0
Mistral AI Mistral-7B-Instruct Apache License 2.0
Rakuten RakutenAI Apache License 2.0
rinna Youri Llama 2 Community License
Meta AI Llama 2 Llama 2 Community License
Sanji Watsuki Kunoichi-7B CC-BY-NC-4.0
Developer Model License
Hugging Face llava-v1.6-mistral-7b-hf Apache License 2.0
Hugging Face llava-v1.6-vicuna-7b-hf, llava-1.5-7b-hf Llama 2 Community License
TinyLLaVA TinyLLaVA-Phi-2-SigLIP-3.1B Apache License 2.0
OpenBMB MiniCPM MiniCPM Model License
Sakana AI EvoVLM-JP-v1-7B Apache License 2.0
XTuner llava-llama-3-8b-v1_1-hf Llama 3 Community License
CyberAgent llava-calm2-siglip Apache License 2.0

open-llm-webui's People

Contributors

uminosachi avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar

open-llm-webui's Issues

build failed on Mac / M1

Hello,
I follow the instruction and run pip install -r requirememts_mac.txt. and it shows the error

Building wheels for collected packages: xformers
Building wheel for xformers (setup.py) ... error
error: subprocess-exited-with-error

ร— python setup.py bdist_wheel did not run successfully.
โ”‚ exit code: 1
โ•ฐโ”€> [216 lines of output]
running bdist_wheel
/Users/development/opt/anaconda3/envs/ollm/lib/python3.10/site-packages/torch/utils/cpp_extension.py:476: UserWarning: Attempted to use ninja as the BuildExtension backend but we could not find ninja.. Falling back to using the slow distutils backend.
warnings.warn(msg.format('we could not find ninja.'))
running build
running build_py
creating build
creating build/lib.macosx-10.9-x86_64-cpython-310
creating build/lib.macosx-10.9-x86_64-cpython-310/xformers
copying xformers/version.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers
copying xformers/init.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers
copying xformers/test.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers
copying xformers/utils.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers
copying xformers/_cpp_lib.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers
copying xformers/info.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers
creating build/lib.macosx-10.9-x86_64-cpython-310/xformers/triton
copying xformers/triton/fused_linear_layer.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/triton
copying xformers/triton/sum_strided.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/triton
copying xformers/triton/k_activations.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/triton
copying xformers/triton/init.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/triton
copying xformers/triton/k_layer_norm.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/triton
copying xformers/triton/k_sum.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/triton
copying xformers/triton/utils.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/triton
copying xformers/triton/k_fused_matmul_fw.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/triton
copying xformers/triton/dropout.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/triton
copying xformers/triton/k_dropout.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/triton
copying xformers/triton/softmax.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/triton
copying xformers/triton/layer_norm.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/triton
copying xformers/triton/k_fused_matmul_bw.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/triton
copying xformers/triton/k_softmax.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/triton
creating build/lib.macosx-10.9-x86_64-cpython-310/xformers/components
copying xformers/components/simplicial_embedding.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/components
copying xformers/components/residual.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/components
copying xformers/components/reversible.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/components
copying xformers/components/activations.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/components
copying xformers/components/multi_head_dispatch.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/components
copying xformers/components/init.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/components
copying xformers/components/input_projection.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/components
copying xformers/components/patch_embedding.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/components
creating build/lib.macosx-10.9-x86_64-cpython-310/xformers/benchmarks
copying xformers/benchmarks/benchmark_mem_eff_attention.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/benchmarks
copying xformers/benchmarks/benchmark_nvfuser.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/benchmarks
copying xformers/benchmarks/benchmark_indexing.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/benchmarks
copying xformers/benchmarks/benchmark_mlp.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/benchmarks
copying xformers/benchmarks/benchmark_triton_stride_sum.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/benchmarks
copying xformers/benchmarks/benchmark_blocksparse_transformers.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/benchmarks
copying xformers/benchmarks/benchmark_transformer.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/benchmarks
copying xformers/benchmarks/benchmark_revnet.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/benchmarks
copying xformers/benchmarks/benchmark_swiglu.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/benchmarks
copying xformers/benchmarks/benchmark_triton_layernorm.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/benchmarks
copying xformers/benchmarks/benchmark_causal_blocksparse.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/benchmarks
copying xformers/benchmarks/benchmark_triton_fused_linear.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/benchmarks
copying xformers/benchmarks/init.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/benchmarks
copying xformers/benchmarks/benchmark_triton_blocksparse.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/benchmarks
copying xformers/benchmarks/benchmark_triton_softmax.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/benchmarks
copying xformers/benchmarks/utils.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/benchmarks
copying xformers/benchmarks/benchmark_encoder.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/benchmarks
copying xformers/benchmarks/benchmark_triton_dropout.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/benchmarks
copying xformers/benchmarks/benchmark_pytorch_transformer.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/benchmarks
copying xformers/benchmarks/benchmark_nystrom_utils.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/benchmarks
copying xformers/benchmarks/benchmark_multi_head_dispatch.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/benchmarks
copying xformers/benchmarks/benchmark_sddmm.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/benchmarks
copying xformers/benchmarks/benchmark_core.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/benchmarks
creating build/lib.macosx-10.9-x86_64-cpython-310/xformers/ops
copying xformers/ops/swiglu_op.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/ops
copying xformers/ops/unbind.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/ops
copying xformers/ops/init.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/ops
copying xformers/ops/common.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/ops
copying xformers/ops/indexing.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/ops
creating build/lib.macosx-10.9-x86_64-cpython-310/xformers/profiler
copying xformers/profiler/device_limits.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/profiler
copying xformers/profiler/init.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/profiler
copying xformers/profiler/api.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/profiler
copying xformers/profiler/slow_ops_profiler.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/profiler
copying xformers/profiler/profiler.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/profiler
creating build/lib.macosx-10.9-x86_64-cpython-310/xformers/sparse
copying xformers/sparse/_csr_ops.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/sparse
copying xformers/sparse/init.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/sparse
copying xformers/sparse/utils.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/sparse
copying xformers/sparse/blocksparse_tensor.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/sparse
copying xformers/sparse/csr_tensor.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/sparse
creating build/lib.macosx-10.9-x86_64-cpython-310/xformers/helpers
copying xformers/helpers/test_utils.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/helpers
copying xformers/helpers/hierarchical_configs.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/helpers
copying xformers/helpers/init.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/helpers
copying xformers/helpers/timm_sparse_attention.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/helpers
creating build/lib.macosx-10.9-x86_64-cpython-310/xformers/_flash_attn
copying xformers/_flash_attn/fused_softmax.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/_flash_attn
copying xformers/_flash_attn/flash_blocksparse_attn_interface.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/_flash_attn
copying xformers/_flash_attn/flash_blocksparse_attention.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/_flash_attn
copying xformers/_flash_attn/bert_padding.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/_flash_attn
copying xformers/_flash_attn/init.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/_flash_attn
copying xformers/_flash_attn/flash_attn_triton_og.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/_flash_attn
copying xformers/_flash_attn/flash_attention.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/_flash_attn
copying xformers/_flash_attn/flash_attn_triton.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/_flash_attn
copying xformers/_flash_attn/flash_attn_interface.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/_flash_attn
creating build/lib.macosx-10.9-x86_64-cpython-310/xformers/factory
copying xformers/factory/init.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/factory
copying xformers/factory/hydra_helper.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/factory
copying xformers/factory/block_factory.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/factory
copying xformers/factory/model_factory.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/factory
copying xformers/factory/block_configs.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/factory
copying xformers/factory/weight_init.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/factory
creating build/lib.macosx-10.9-x86_64-cpython-310/xformers/components/attention
copying xformers/components/attention/global_tokens.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/components/attention
copying xformers/components/attention/ortho.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/components/attention
copying xformers/components/attention/blocksparse.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/components/attention
copying xformers/components/attention/local.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/components/attention
copying xformers/components/attention/compositional.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/components/attention
copying xformers/components/attention/pooling.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/components/attention
copying xformers/components/attention/init.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/components/attention
copying xformers/components/attention/_sputnik_sparse.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/components/attention
copying xformers/components/attention/core.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/components/attention
copying xformers/components/attention/lambda_layer.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/components/attention
copying xformers/components/attention/random.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/components/attention
copying xformers/components/attention/fourier_mix.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/components/attention
copying xformers/components/attention/scaled_dot_product.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/components/attention
copying xformers/components/attention/utils.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/components/attention
copying xformers/components/attention/attention_mask.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/components/attention
copying xformers/components/attention/linformer.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/components/attention
copying xformers/components/attention/attention_patterns.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/components/attention
copying xformers/components/attention/visual.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/components/attention
copying xformers/components/attention/sparsity_config.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/components/attention
copying xformers/components/attention/nystrom.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/components/attention
copying xformers/components/attention/favor.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/components/attention
copying xformers/components/attention/base.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/components/attention
creating build/lib.macosx-10.9-x86_64-cpython-310/xformers/components/feedforward
copying xformers/components/feedforward/init.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/components/feedforward
copying xformers/components/feedforward/mixture_of_experts.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/components/feedforward
copying xformers/components/feedforward/mlp.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/components/feedforward
copying xformers/components/feedforward/conv_mlp.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/components/feedforward
copying xformers/components/feedforward/fused_mlp.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/components/feedforward
copying xformers/components/feedforward/base.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/components/feedforward
creating build/lib.macosx-10.9-x86_64-cpython-310/xformers/components/positional_embedding
copying xformers/components/positional_embedding/vocab.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/components/positional_embedding
copying xformers/components/positional_embedding/init.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/components/positional_embedding
copying xformers/components/positional_embedding/param.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/components/positional_embedding
copying xformers/components/positional_embedding/sine.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/components/positional_embedding
copying xformers/components/positional_embedding/rotary.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/components/positional_embedding
copying xformers/components/positional_embedding/base.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/components/positional_embedding
creating build/lib.macosx-10.9-x86_64-cpython-310/xformers/components/nvfuser
copying xformers/components/nvfuser/bias_dropout_res_layernorm.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/components/nvfuser
copying xformers/components/nvfuser/bias_act_dropout.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/components/nvfuser
copying xformers/components/nvfuser/init.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/components/nvfuser
copying xformers/components/nvfuser/utils.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/components/nvfuser
copying xformers/components/nvfuser/bias_dropout_res.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/components/nvfuser
creating build/lib.macosx-10.9-x86_64-cpython-310/xformers/components/attention/feature_maps
copying xformers/components/attention/feature_maps/init.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/components/attention/feature_maps
copying xformers/components/attention/feature_maps/softmax.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/components/attention/feature_maps
copying xformers/components/attention/feature_maps/base.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/components/attention/feature_maps
creating build/lib.macosx-10.9-x86_64-cpython-310/xformers/benchmarks/LRA
copying xformers/benchmarks/LRA/batch_submit.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/benchmarks/LRA
copying xformers/benchmarks/LRA/batch_fetch_results.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/benchmarks/LRA
copying xformers/benchmarks/LRA/run_with_submitit.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/benchmarks/LRA
copying xformers/benchmarks/LRA/init.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/benchmarks/LRA
copying xformers/benchmarks/LRA/run_tasks.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/benchmarks/LRA
copying xformers/benchmarks/LRA/run_grid_search.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/benchmarks/LRA
creating build/lib.macosx-10.9-x86_64-cpython-310/xformers/benchmarks/LRA/code
copying xformers/benchmarks/LRA/code/init.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/benchmarks/LRA/code
copying xformers/benchmarks/LRA/code/model_wrapper.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/benchmarks/LRA/code
copying xformers/benchmarks/LRA/code/dataset.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/benchmarks/LRA/code
creating build/lib.macosx-10.9-x86_64-cpython-310/xformers/ops/fmha
copying xformers/ops/fmha/triton.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/ops/fmha
copying xformers/ops/fmha/dispatch.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/ops/fmha
copying xformers/ops/fmha/init.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/ops/fmha
copying xformers/ops/fmha/attn_bias.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/ops/fmha
copying xformers/ops/fmha/common.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/ops/fmha
copying xformers/ops/fmha/flash.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/ops/fmha
copying xformers/ops/fmha/small_k.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/ops/fmha
copying xformers/ops/fmha/cutlass.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/ops/fmha
creating build/lib.macosx-10.9-x86_64-cpython-310/xformers/_flash_attn/losses
copying xformers/_flash_attn/losses/cross_entropy.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/_flash_attn/losses
copying xformers/_flash_attn/losses/init.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/_flash_attn/losses
creating build/lib.macosx-10.9-x86_64-cpython-310/xformers/_flash_attn/layers
copying xformers/_flash_attn/layers/init.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/_flash_attn/layers
copying xformers/_flash_attn/layers/patch_embed.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/_flash_attn/layers
copying xformers/_flash_attn/layers/rotary.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/_flash_attn/layers
creating build/lib.macosx-10.9-x86_64-cpython-310/xformers/_flash_attn/utils
copying xformers/_flash_attn/utils/pretrained.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/_flash_attn/utils
copying xformers/_flash_attn/utils/generation.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/_flash_attn/utils
copying xformers/_flash_attn/utils/benchmark.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/_flash_attn/utils
copying xformers/_flash_attn/utils/init.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/_flash_attn/utils
copying xformers/_flash_attn/utils/distributed.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/_flash_attn/utils
creating build/lib.macosx-10.9-x86_64-cpython-310/xformers/_flash_attn/models
copying xformers/_flash_attn/models/gptj.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/_flash_attn/models
copying xformers/_flash_attn/models/init.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/_flash_attn/models
copying xformers/_flash_attn/models/opt.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/_flash_attn/models
copying xformers/_flash_attn/models/llama.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/_flash_attn/models
copying xformers/_flash_attn/models/vit.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/_flash_attn/models
copying xformers/_flash_attn/models/bert.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/_flash_attn/models
copying xformers/_flash_attn/models/gpt_neox.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/_flash_attn/models
copying xformers/_flash_attn/models/gpt.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/_flash_attn/models
creating build/lib.macosx-10.9-x86_64-cpython-310/xformers/_flash_attn/ops
copying xformers/_flash_attn/ops/activations.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/_flash_attn/ops
copying xformers/_flash_attn/ops/init.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/_flash_attn/ops
copying xformers/_flash_attn/ops/fused_dense.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/_flash_attn/ops
copying xformers/_flash_attn/ops/rms_norm.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/_flash_attn/ops
copying xformers/_flash_attn/ops/layer_norm.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/_flash_attn/ops
creating build/lib.macosx-10.9-x86_64-cpython-310/xformers/_flash_attn/modules
copying xformers/_flash_attn/modules/embedding.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/_flash_attn/modules
copying xformers/_flash_attn/modules/init.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/_flash_attn/modules
copying xformers/_flash_attn/modules/mlp.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/_flash_attn/modules
copying xformers/_flash_attn/modules/block.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/_flash_attn/modules
copying xformers/_flash_attn/modules/mha.py -> build/lib.macosx-10.9-x86_64-cpython-310/xformers/_flash_attn/modules
running build_ext
building 'xformers._C' extension
creating build/temp.macosx-10.9-x86_64-cpython-310
creating build/temp.macosx-10.9-x86_64-cpython-310/xformers
creating build/temp.macosx-10.9-x86_64-cpython-310/xformers/csrc
creating build/temp.macosx-10.9-x86_64-cpython-310/xformers/csrc/attention
creating build/temp.macosx-10.9-x86_64-cpython-310/xformers/csrc/attention/autograd
creating build/temp.macosx-10.9-x86_64-cpython-310/xformers/csrc/attention/cpu
creating build/temp.macosx-10.9-x86_64-cpython-310/xformers/csrc/indexing
creating build/temp.macosx-10.9-x86_64-cpython-310/xformers/csrc/swiglu
clang -Wno-unused-result -Wsign-compare -Wunreachable-code -DNDEBUG -fwrapv -O2 -Wall -fPIC -O2 -isystem /Users/development/opt/anaconda3/envs/ollm/include -fPIC -O2 -isystem /Users/development/opt/anaconda3/envs/ollm/include -I/private/var/folders/0m/fj0bcx4n16dd2snclpw414080000gn/T/pip-install-9pmf8iz7/xformers_fec6780c30d34980ba5c4acf372654b4/xformers/csrc -I/Users/development/opt/anaconda3/envs/ollm/lib/python3.10/site-packages/torch/include -I/Users/development/opt/anaconda3/envs/ollm/lib/python3.10/site-packages/torch/include/torch/csrc/api/include -I/Users/development/opt/anaconda3/envs/ollm/lib/python3.10/site-packages/torch/include/TH -I/Users/development/opt/anaconda3/envs/ollm/lib/python3.10/site-packages/torch/include/THC -I/Users/development/opt/anaconda3/envs/ollm/include/python3.10 -c xformers/csrc/attention/attention.cpp -o build/temp.macosx-10.9-x86_64-cpython-310/xformers/csrc/attention/attention.o -O3 -fopenmp -DTORCH_API_INCLUDE_EXTENSION_H -DPYBIND11_COMPILER_TYPE="_clang" -DPYBIND11_STDLIB="_libcpp" -DPYBIND11_BUILD_ABI="_cxxabi1002" -DTORCH_EXTENSION_NAME=_C -D_GLIBCXX_USE_CXX11_ABI=0 -std=c++17
clang: error: unsupported option '-fopenmp'
error: command '/usr/bin/clang' failed with exit code 1
[end of output]

note: This error originates from a subprocess, and is likely not a problem with pip.
ERROR: Failed building wheel for xformers
Running setup.py clean for xformers
Failed to build xformers
ERROR: Could not build wheels for xformers, which is required to install pyproject.toml-based projects

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.