Comments (7)
You can try to increase the value of environment variable XINFERENCE_HEALTH_CHECK_INTERVAL
https://inference.readthedocs.io/en/latest/getting_started/environments.html#xinference-health-check-interval.
from inference.
Hi @aresnow1 that won't work because it's not because healthecheck starts too soon, but it's trying to connect to 0.0.0.0:port, which isn't supported on windows (WinError 1214).
from inference.
xinference version 0.8.5, windows 11, python 3.11
同样的问题。
from inference.
@fffonion @hao203
if you use xinference on win, the xinference site's doc : launch on local, use command xinference-local --host 0.0.0.0 --port 9997
will have error like uper's description, you can solve the problem simply just run command xinference-local
, then xinference will launch on http://127.0.0.1:9997/
like below:
from inference.
Hi @aresnow1 that won't work because it's not because healthecheck starts too soon, but it's trying to connect to 0.0.0.0:port, which isn't supported on windows (WinError 1214).
Indeed, the windows does not support 0.0.0.0, do you have any idea, maybe raise an error or try to get IP for user?
from inference.
Same issue as #1053
from inference.
@qinxuye yeah we probably need to find an available local IP on any interface, and use that for healthcheck. If we just are
just healthchecking the service but not the port connectivity, we can also do some hack to listen both 0.0.0.0 and 127.0.0.1 at the same time if user specified 0.0.0.0, and healthcheck on 127.0.0.1 only.
from inference.
Related Issues (20)
- Support Optional Configurations for Embedding models HOT 3
- BUG 20G以上模型导致机器重启 HOT 6
- BUG When I reasoning the model Qwen-VL-Chat-Int4 and Yi-VL-6B, the Model Engine cannot be found HOT 13
- Suggestions about transcription enhancement HOT 2
- 多卡运行模型启动报错,单卡运行正常 HOT 5
- 自定义模型qwen2-0.5在dify中默认为React,无法更改 HOT 1
- 请问跑minicpm-llama3-v-2_5(int4)支持并发调用接口么?2个及以上并发调用就报错了 HOT 3
- BUG Failed to download model 'glm-4v' (size: 9, format: pytorch) after multiple retries
- 能否支持开机启动加载指定模型?
- 支持 openai messages 指定 type 接口 格式 HOT 1
- QUESTION how to load mini cpmV 2.5 gguf model which has two model files with xinference? HOT 1
- BUG:UI端关闭模型,资源没释放 HOT 2
- BUG
- BUG新的问题进来时会中断前一个问题回复的过程
- It seems that this preserves the ReAct "thought" process, but this is not user-friendly. Is it possible to use the method of the official Qwen function call? HOT 6
- [FER] Intel NPU support HOT 2
- BUG-运行Qwen2-MoE-14B-GPTQ4 模型出错 HOT 4
- [Suggestion]Image and audio models should download necessary files only.
- 支持昇腾处理器 HOT 2
- Failed to build chatglm-cpp llama-cpp-python pynini HOT 4
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from inference.