Comments (5)
如果没有开启dump也出现了dump的文件的话,就要检查一下是否有内存越界把config改掉了
之前有多的图片反映到还有别的地方报错,恐怕是真的越界了
建议追踪一下enable_tensor_debug所在的内存有没有被修改到
from ppl.nn.
出现这个错误可能是因为这些模型的Tensor名字和大小都相同,Dump接口同时写入这个文件导致冲突
Dump文件的命名规则https://github.com/openppl-public/ppl.nn/blob/master/src/ppl/nn/engines/x86/kernel.cc#L79
DumpDebugTensor一般只建议在单模型debug中间tensor的时候使用
- 先排查出输出错误的模型
- dump出模型相应的输入输出数据
- 使用DumpDebugTensor检查具体在那一层出错
如果要在多模型的情况下使用,建议通过Configure选项x86::ENGINE_CONF_DEBUG_DATA_DIR使不同模型的Tensor输出到不同的文件夹中(记得先新建文件夹)
不然模型之间的Tensor文件名字很容易冲突
from ppl.nn.
感谢回答!我的所有模型的Tensor名字和大小确实是相同的
DumpDebugTensor这个功能是通过 ENGINE_CONF_TENSOR_DEBUG 来控制的吗?
我没有手动开启这个选项,在某个地方的文档看这个功能是默认关闭的
不过我看了下代码
https://github.com/openppl-public/ppl.nn/blob/master/include/ppl/nn/engines/x86/options.h#L43
我手动关掉这个设置试试
from ppl.nn.
感谢回答!我的所有模型的Tensor名字和大小确实是相同的 DumpDebugTensor这个功能是通过 ENGINE_CONF_TENSOR_DEBUG 来控制的吗? 我没有手动开启这个选项,在某个地方的文档看这个功能是默认关闭的
不过我看了下代码 https://github.com/openppl-public/ppl.nn/blob/master/include/ppl/nn/engines/x86/options.h#L43
我手动关掉这个设置试试
这里的1指的使CONF的编号,并不是开关
默认关闭是在这里https://github.com/openppl-public/ppl.nn/blob/master/src/ppl/nn/engines/x86/engine_config.h#L27
from ppl.nn.
确实默认关闭了,眼滑看错了...
之前贴的另一张图是挂在ppt::common的 BufferInfo了。 我就提到pplcommon的仓库里了
那个错误应该是我没保证ppl::nn::engine的生命周期导致的,已经fix了
目前看fix后, core在 ppl::nn::x86::X86Kernel::DumpOutputTensors 的问题也没有继续出现了。
感谢解答!
from ppl.nn.
Related Issues (20)
- Floating point exception (core dumped) occured when using cuda engine HOT 1
- Floating point exception (core dumped) HOT 1
- Onnx run error HOT 2
- 请问支持int8在高通芯片上cDSP进行推理吗?
- Slice op question HOT 1
- pplnn run mobilenet v2 model failed. (use cuda) HOT 7
- linux compile error protobuf static assertion failed HOT 3
- malloc_consolidate(): invalid chunk size HOT 2
- pplnn save-input 得到的NDARRAY的 shape不正确 HOT 1
- 如何使用cmake的将ppl.nn和依赖ppl.nn的代码一同编译? HOT 3
- 获取模型推理结果(GetOutputs)耗时长 HOT 2
- Install Error HOT 1
- The compilation passed, but an error was reported in test phase HOT 2
- Floating point exception (core dumped) ? HOT 4
- 使用x86 engine运行resnet50 fp16 onnx模型 core dump
- (Ask) why InferInheritedType handle int8 to fp16 out? HOT 3
- Got wrong output shape when run a Gemm op(transB=0) use cuda HOT 4
- Crash with ONNX Split operator
- 关于全局engine,其他线程引用导致的性能下降问题 HOT 4
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from ppl.nn.