解决TensorRT报错:Assertion Error in trtSmToCask: 0 (Unsupported SM.)

本文介绍了在使用TensorRT时遇到的错误信息,错误源于TensorRT版本7.2.1与CUDA 11.5版本不兼容。解决办法包括重新安装CUDA或TensorRT。建议选择重装CUDA,因为不同TensorRT版本间可能存在代码不一致性,且重装TensorRT可能需要大量代码修改。同时提供了TensorRT、CUDA和cudnn的版本匹配指南。
部署运行你感兴趣的模型镜像

错误信息:

E20211126 11:55:13.654 140412916315904 tensorrt.cpp:10] …/rtSafe/cuda/caskUtils.cpp (98) - Assertion Error in trtSmToCask: 0 (Unsupported SM.)

原因:TensorRT版本与cuda或者cudnn版本不匹配,如笔者的TensorRT版本为7.2.1,而cuda为最新的11.5版本,其中cuda版本过高无法适配:
在这里插入图片描述

解决:重装cuda或者TensorRT
建议:重装cuda,因为不同版本的TensorRT在使用时存在代码不一致的情况,如TensorRT7.2.1和TensorRT8.2.1要求重写虚函数时添加noexcept,如果代码量不大可以考虑重装TensorRT
附上TensorRT、cuda、cudnn各版本的匹配要求:TensorRT release note

您可能感兴趣的与本文相关的镜像

PyTorch 2.5

PyTorch 2.5

PyTorch
Cuda

PyTorch 是一个开源的 Python 机器学习库,基于 Torch 库,底层由 C++ 实现,应用于人工智能领域,如计算机视觉和自然语言处理

[11/27/2025-18:31:50] [I] === Model Options === [11/27/2025-18:31:50] [I] Format: ONNX [11/27/2025-18:31:50] [I] Model: D:\Audio2Face\Audio2Face-3D-SDK\_data\audio2emotion-models\audio2emotion-v2.2\network.onnx [11/27/2025-18:31:50] [I] Output: [11/27/2025-18:31:50] [I] === Build Options === [11/27/2025-18:31:50] [I] Memory Pools: workspace: default, dlaSRAM: default, dlaLocalDRAM: default, dlaGlobalDRAM: default, tacticSharedMem: default [11/27/2025-18:31:50] [I] avgTiming: 8 [11/27/2025-18:31:50] [I] Precision: FP32 [11/27/2025-18:31:50] [I] LayerPrecisions: [11/27/2025-18:31:50] [I] Layer Device Types: [11/27/2025-18:31:50] [I] Calibration: [11/27/2025-18:31:50] [I] Refit: Disabled [11/27/2025-18:31:50] [I] Strip weights: Disabled [11/27/2025-18:31:50] [I] Version Compatible: Disabled [11/27/2025-18:31:50] [I] ONNX Plugin InstanceNorm: Disabled [11/27/2025-18:31:50] [I] ONNX kENABLE_UINT8_AND_ASYMMETRIC_QUANTIZATION_DLA flag: Disabled [11/27/2025-18:31:50] [I] TensorRT runtime: full [11/27/2025-18:31:50] [I] Lean DLL Path: [11/27/2025-18:31:50] [I] Tempfile Controls: { in_memory: allow, temporary: allow } [11/27/2025-18:31:50] [I] Exclude Lean Runtime: Disabled [11/27/2025-18:31:50] [I] Sparsity: Disabled [11/27/2025-18:31:50] [I] Safe mode: Disabled [11/27/2025-18:31:50] [I] Build DLA standalone loadable: Disabled [11/27/2025-18:31:50] [I] Allow GPU fallback for DLA: Disabled [11/27/2025-18:31:50] [I] DirectIO mode: Disabled [11/27/2025-18:31:50] [I] Restricted mode: Disabled [11/27/2025-18:31:50] [I] Skip inference: Disabled [11/27/2025-18:31:50] [I] Save engine: C:\Users\Administrator\AppData\Local\Temp\tmp6s8itunu\network.trt [11/27/2025-18:31:50] [I] Load engine: [11/27/2025-18:31:50] [I] Profiling verbosity: 0 [11/27/2025-18:31:50] [I] Tactic sources: Using default tactic sources [11/27/2025-18:31:50] [I] timingCacheMode: local [11/27/2025-18:31:50] [I] timingCacheFile: [11/27/2025-18:31:50] [I] Enable Compilation Cache: Enabled [11/27/2025-18:31:50] [I] Enable Monitor Memory: Disabled [11/27/2025-18:31:50] [I] errorOnTimingCacheMiss: Disabled [11/27/2025-18:31:50] [I] Preview Features: Use default preview flags. [11/27/2025-18:31:50] [I] MaxAuxStreams: -1 [11/27/2025-18:31:50] [I] BuilderOptimizationLevel: -1 [11/27/2025-18:31:50] [I] MaxTactics: -1 [11/27/2025-18:31:50] [I] Calibration Profile Index: 0 [11/27/2025-18:31:50] [I] Weight Streaming: Disabled [11/27/2025-18:31:50] [I] Runtime Platform: Same As Build [11/27/2025-18:31:50] [I] Debug Tensors: [11/27/2025-18:31:50] [I] Distributive Independence: Disabled [11/27/2025-18:31:50] [I] Mark Unfused Tensors As Debug Tensors: Disabled [11/27/2025-18:31:50] [I] Input(s)s format: fp32:CHW [11/27/2025-18:31:50] [I] Output(s)s format: fp32:CHW [11/27/2025-18:31:50] [I] Input build shape (profile 0): input_values=1x5000+8x30000+128x60000 [11/27/2025-18:31:50] [I] Input calibration shapes: model [11/27/2025-18:31:50] [I] === System Options === [11/27/2025-18:31:50] [I] Device: 0 [11/27/2025-18:31:50] [I] DLACore: [11/27/2025-18:31:50] [I] Plugins: [11/27/2025-18:31:50] [I] setPluginsToSerialize: [11/27/2025-18:31:50] [I] dynamicPlugins: [11/27/2025-18:31:50] [I] ignoreParsedPluginLibs: 0 [11/27/2025-18:31:50] [I] [11/27/2025-18:31:50] [I] === Inference Options === [11/27/2025-18:31:50] [I] Batch: Explicit [11/27/2025-18:31:50] [I] Input inference shape : input_values=8x30000 [11/27/2025-18:31:50] [I] Iterations: 10 [11/27/2025-18:31:50] [I] Duration: 3s (+ 200ms warm up) [11/27/2025-18:31:50] [I] Sleep time: 0ms [11/27/2025-18:31:50] [I] Idle time: 0ms [11/27/2025-18:31:50] [I] Inference Streams: 1 [11/27/2025-18:31:50] [I] ExposeDMA: Disabled [11/27/2025-18:31:50] [I] Data transfers: Enabled [11/27/2025-18:31:50] [I] Spin-wait: Disabled [11/27/2025-18:31:50] [I] Multithreading: Disabled [11/27/2025-18:31:50] [I] CUDA Graph: Disabled [11/27/2025-18:31:50] [I] Separate profiling: Disabled [11/27/2025-18:31:50] [I] Time Deserialize: Disabled [11/27/2025-18:31:50] [I] Time Refit: Disabled [11/27/2025-18:31:50] [I] NVTX verbosity: 0 [11/27/2025-18:31:50] [I] Persistent Cache Ratio: 0 [11/27/2025-18:31:50] [I] Optimization Profile Index: 0 [11/27/2025-18:31:50] [I] Weight Streaming Budget: 100.000000% [11/27/2025-18:31:50] [I] Inputs: [11/27/2025-18:31:50] [I] Debug Tensor Save Destinations: [11/27/2025-18:31:50] [I] Dump All Debug Tensor in Formats: [11/27/2025-18:31:50] [I] === Reporting Options === [11/27/2025-18:31:50] [I] Verbose: Disabled [11/27/2025-18:31:50] [I] Averages: 10 inferences [11/27/2025-18:31:50] [I] Percentiles: 90,95,99 [11/27/2025-18:31:50] [I] Dump refittable layers:Disabled [11/27/2025-18:31:50] [I] Dump output: Disabled [11/27/2025-18:31:50] [I] Profile: Disabled [11/27/2025-18:31:50] [I] Export timing to JSON file: [11/27/2025-18:31:50] [I] Export output to JSON file: [11/27/2025-18:31:50] [I] Export profile to JSON file: [11/27/2025-18:31:50] [I] [11/27/2025-18:31:50] [I] === Device Information === [11/27/2025-18:31:50] [I] Available Devices: [11/27/2025-18:31:50] [I] Device 0: "NVIDIA GeForce RTX 4070 Laptop GPU" UUID: GPU-df57adcc-000e-4631-7e44-f044e50478c6 [11/27/2025-18:31:50] [I] Selected Device: NVIDIA GeForce RTX 4070 Laptop GPU [11/27/2025-18:31:50] [I] Selected Device ID: 0 [11/27/2025-18:31:50] [I] Selected Device UUID: GPU-df57adcc-000e-4631-7e44-f044e50478c6 [11/27/2025-18:31:50] [I] Compute Capability: 8.9 [11/27/2025-18:31:50] [I] SMs: 36 [11/27/2025-18:31:50] [I] Device Global Memory: 8187 MiB [11/27/2025-18:31:50] [I] Shared Memory per SM: 100 KiB [11/27/2025-18:31:50] [I] Memory Bus Width: 128 bits (ECC disabled) [11/27/2025-18:31:50] [I] Application Compute Clock Rate: 1.695 GHz [11/27/2025-18:31:50] [I] Application Memory Clock Rate: 8.001 GHz [11/27/2025-18:31:50] [I] [11/27/2025-18:31:50] [I] Note: The application clock rates do not reflect the actual clock rates that the GPU is currently running at. [11/27/2025-18:31:50] [I] [11/27/2025-18:31:50] [I] TensorRT version: 10.13.3 [11/27/2025-18:31:50] [I] Loading standard plugins [11/27/2025-18:31:50] [I] [TRT] [MemUsageChange] Init CUDA: CPU +1, GPU +0, now: CPU 18935, GPU 1133 (MiB) [11/27/2025-18:31:52] [I] [TRT] [MemUsageChange] Init builder kernel library: CPU +2609, GPU +8, now: CPU 21877, GPU 1141 (MiB) [11/27/2025-18:31:52] [I] Start parsing network model. [11/27/2025-18:31:53] [I] [TRT] ---------------------------------------------------------------- [11/27/2025-18:31:53] [I] [TRT] Input filename: D:\Audio2Face\Audio2Face-3D-SDK\_data\audio2emotion-models\audio2emotion-v2.2\network.onnx [11/27/2025-18:31:53] [I] [TRT] ONNX IR version: 0.0.7 [11/27/2025-18:31:53] [I] [TRT] Opset version: 12 [11/27/2025-18:31:53] [I] [TRT] Producer name: pytorch [11/27/2025-18:31:53] [I] [TRT] Producer version: 2.3.1 [11/27/2025-18:31:53] [I] [TRT] Domain: [11/27/2025-18:31:53] [I] [TRT] Model version: 0 [11/27/2025-18:31:53] [I] [TRT] Doc string: [11/27/2025-18:31:53] [I] [TRT] ---------------------------------------------------------------- [11/27/2025-18:31:54] [I] Finished parsing network model. Parse time: 1.58212 [11/27/2025-18:31:54] [I] Set shape of input tensor input_values for optimization profile 0 to: MIN=1x5000 OPT=8x30000 MAX=128x60000 [11/27/2025-18:31:57] [I] [TRT] Local timing cache in use. Profiling results in this builder pass will not be stored. [11/27/2025-18:31:57] [I] [TRT] Compiler backend is used during engine build. [11/27/2025-18:31:59] [W] [TRT] Tactic Device request: 8999MB Available: 8187MB. Device memory is insufficient to use tactic. [11/27/2025-18:31:59] [W] [TRT] UNSUPPORTED_STATE: Skipping tactic 1 due to insufficient memory on requested size of 9436397568 detected for tactic 0x0000000000000001. [11/27/2025-18:31:59] [W] [TRT] Tactic Device request: 8999MB Available: 8187MB. Device memory is insufficient to use tactic. [11/27/2025-18:31:59] [W] [TRT] UNSUPPORTED_STATE: Skipping tactic 1 due to insufficient memory on requested size of 9436397568 detected for tactic 0x0000000000000001. [11/27/2025-18:32:00] [W] [TRT] Tactic Device request: 8999MB Available: 8187MB. Device memory is insufficient to use tactic. [11/27/2025-18:32:00] [W] [TRT] UNSUPPORTED_STATE: Skipping tactic 1 due to insufficient memory on requested size of 9436397568 detected for tactic 0x0000000000000001. [11/27/2025-18:32:00] [W] [TRT] Tactic Device request: 8999MB Available: 8187MB. Device memory is insufficient to use tactic. [11/27/2025-18:32:00] [W] [TRT] UNSUPPORTED_STATE: Skipping tactic 1 due to insufficient memory on requested size of 9436397568 detected for tactic 0x0000000000000001. [11/27/2025-18:32:00] [W] [TRT] Tactic Device request: 12004MB Available: 8187MB. Device memory is insufficient to use tactic. [11/27/2025-18:32:00] [W] [TRT] UNSUPPORTED_STATE: Skipping tactic 0 due to insufficient memory on requested size of 12588006912 detected for tactic 0x0000000000000000. [11/27/2025-18:32:00] [E] Error[10]: IBuilder::buildSerializedNetworkToStream: Error Code 10: Internal Error (Could not find any implementation for node {ForeignNode[ONNXTRT_squeezeTensor + /model/wav2vec2/feature_extractor/conv_layers.0/Transpose...ONNXTRT_unsqueezeTensor_21]}. In nvinfer1::builder::cgraph::LeafCNode::computeCosts at optimizer/common/tactic/optimizer.cpp:4115) [11/27/2025-18:32:00] [I] Created engine with size: 0 MiB [11/27/2025-18:32:00] [I] Engine built in 6.63394 sec. [11/27/2025-18:32:00] [E] Assertion failure: false && "Attempting to access an empty engine!" Generating benchmark data... Generating 5 minutes audio 1/10 Generating 5 minutes audio 2/10 Generating 5 minutes audio 3/10 Generating 5 minutes audio 4/10 Generating 5 minutes audio 5/10 Generating 5 minutes audio 6/10 Generating 5 minutes audio 7/10 Generating 5 minutes audio 8/10 Generating 5 minutes audio 9/10 Generating 5 minutes audio 10/10 &&&& RUNNING TensorRT.trtexec [TensorRT v101303] [b9] # trtexec --onnx=D:\Audio2Face\Audio2Face-3D-SDK\_data\audio2emotion-models\audio2emotion-v2.2\network.onnx --saveEngine=C:\Users\Administrator\AppData\Local\Temp\tmp7bwtotg_\network.trt --device=0 --minShapes=input_values:1x5000 --maxShapes=input_values:128x60000 --optShapes=input_values:8x30000 [11/27/2025-18:32:03] [I] TF32 is enabled by default. Add --noTF32 flag to further improve accuracy with some performance cost. [11/27/2025-18:32:03] [I] === Model Options === [11/27/2025-18:32:03] [I] Format: ONNX [11/27/2025-18:32:03] [I] Model: D:\Audio2Face\Audio2Face-3D-SDK\_data\audio2emotion-models\audio2emotion-v2.2\network.onnx [11/27/2025-18:32:03] [I] Output: [11/27/2025-18:32:03] [I] === Build Options === [11/27/2025-18:32:03] [I] Memory Pools: workspace: default, dlaSRAM: default, dlaLocalDRAM: default, dlaGlobalDRAM: default, tacticSharedMem: default [11/27/2025-18:32:03] [I] avgTiming: 8 [11/27/2025-18:32:03] [I] Precision: FP32 [11/27/2025-18:32:03] [I] LayerPrecisions: [11/27/2025-18:32:03] [I] Layer Device Types: [11/27/2025-18:32:03] [I] Calibration: [11/27/2025-18:32:03] [I] Refit: Disabled [11/27/2025-18:32:03] [I] Strip weights: Disabled [11/27/2025-18:32:03] [I] Version Compatible: Disabled [11/27/2025-18:32:03] [I] ONNX Plugin InstanceNorm: Disabled [11/27/2025-18:32:03] [I] ONNX kENABLE_UINT8_AND_ASYMMETRIC_QUANTIZATION_DLA flag: Disabled [11/27/2025-18:32:03] [I] TensorRT runtime: full [11/27/2025-18:32:03] [I] Lean DLL Path: [11/27/2025-18:32:03] [I] Tempfile Controls: { in_memory: allow, temporary: allow } [11/27/2025-18:32:03] [I] Exclude Lean Runtime: Disabled [11/27/2025-18:32:03] [I] Sparsity: Disabled [11/27/2025-18:32:03] [I] Safe mode: Disabled [11/27/2025-18:32:03] [I] Build DLA standalone loadable: Disabled [11/27/2025-18:32:03] [I] Allow GPU fallback for DLA: Disabled [11/27/2025-18:32:03] [I] DirectIO mode: Disabled [11/27/2025-18:32:03] [I] Restricted mode: Disabled [11/27/2025-18:32:03] [I] Skip inference: Disabled [11/27/2025-18:32:03] [I] Save engine: C:\Users\Administrator\AppData\Local\Temp\tmp7bwtotg_\network.trt [11/27/2025-18:32:03] [I] Load engine: [11/27/2025-18:32:03] [I] Profiling verbosity: 0 [11/27/2025-18:32:03] [I] Tactic sources: Using default tactic sources [11/27/2025-18:32:03] [I] timingCacheMode: local [11/27/2025-18:32:03] [I] timingCacheFile: [11/27/2025-18:32:03] [I] Enable Compilation Cache: Enabled [11/27/2025-18:32:03] [I] Enable Monitor Memory: Disabled [11/27/2025-18:32:03] [I] errorOnTimingCacheMiss: Disabled [11/27/2025-18:32:03] [I] Preview Features: Use default preview flags. [11/27/2025-18:32:03] [I] MaxAuxStreams: -1 [11/27/2025-18:32:03] [I] BuilderOptimizationLevel: -1 [11/27/2025-18:32:03] [I] MaxTactics: -1 [11/27/2025-18:32:03] [I] Calibration Profile Index: 0 [11/27/2025-18:32:03] [I] Weight Streaming: Disabled [11/27/2025-18:32:03] [I] Runtime Platform: Same As Build [11/27/2025-18:32:03] [I] Debug Tensors: [11/27/2025-18:32:03] [I] Distributive Independence: Disabled [11/27/2025-18:32:03] [I] Mark Unfused Tensors As Debug Tensors: Disabled [11/27/2025-18:32:03] [I] Input(s)s format: fp32:CHW [11/27/2025-18:32:03] [I] Output(s)s format: fp32:CHW [11/27/2025-18:32:03] [I] Input build shape (profile 0): input_values=1x5000+8x30000+128x60000 [11/27/2025-18:32:03] [I] Input calibration shapes: model [11/27/2025-18:32:03] [I] === System Options === [11/27/2025-18:32:03] [I] Device: 0 [11/27/2025-18:32:03] [I] DLACore: [11/27/2025-18:32:03] [I] Plugins: [11/27/2025-18:32:03] [I] setPluginsToSerialize: [11/27/2025-18:32:03] [I] dynamicPlugins: [11/27/2025-18:32:03] [I] ignoreParsedPluginLibs: 0 [11/27/2025-18:32:03] [I] [11/27/2025-18:32:03] [I] === Inference Options === [11/27/2025-18:32:03] [I] Batch: Explicit [11/27/2025-18:32:03] [I] Input inference shape : input_values=8x30000 [11/27/2025-18:32:03] [I] Iterations: 10 [11/27/2025-18:32:03] [I] Duration: 3s (+ 200ms warm up) [11/27/2025-18:32:03] [I] Sleep time: 0ms [11/27/2025-18:32:03] [I] Idle time: 0ms [11/27/2025-18:32:03] [I] Inference Streams: 1 [11/27/2025-18:32:03] [I] ExposeDMA: Disabled [11/27/2025-18:32:03] [I] Data transfers: Enabled [11/27/2025-18:32:03] [I] Spin-wait: Disabled [11/27/2025-18:32:03] [I] Multithreading: Disabled [11/27/2025-18:32:03] [I] CUDA Graph: Disabled [11/27/2025-18:32:03] [I] Separate profiling: Disabled [11/27/2025-18:32:03] [I] Time Deserialize: Disabled [11/27/2025-18:32:03] [I] Time Refit: Disabled [11/27/2025-18:32:03] [I] NVTX verbosity: 0 [11/27/2025-18:32:03] [I] Persistent Cache Ratio: 0 [11/27/2025-18:32:03] [I] Optimization Profile Index: 0 [11/27/2025-18:32:03] [I] Weight Streaming Budget: 100.000000% [11/27/2025-18:32:03] [I] Inputs: [11/27/2025-18:32:03] [I] Debug Tensor Save Destinations: [11/27/2025-18:32:03] [I] Dump All Debug Tensor in Formats: [11/27/2025-18:32:03] [I] === Reporting Options === [11/27/2025-18:32:03] [I] Verbose: Disabled [11/27/2025-18:32:03] [I] Averages: 10 inferences [11/27/2025-18:32:03] [I] Percentiles: 90,95,99 [11/27/2025-18:32:03] [I] Dump refittable layers:Disabled [11/27/2025-18:32:03] [I] Dump output: Disabled [11/27/2025-18:32:03] [I] Profile: Disabled [11/27/2025-18:32:03] [I] Export timing to JSON file: [11/27/2025-18:32:03] [I] Export output to JSON file: [11/27/2025-18:32:03] [I] Export profile to JSON file: [11/27/2025-18:32:03] [I] [11/27/2025-18:32:03] [I] === Device Information === [11/27/2025-18:32:03] [I] Available Devices: [11/27/2025-18:32:03] [I] Device 0: "NVIDIA GeForce RTX 4070 Laptop GPU" UUID: GPU-df57adcc-000e-4631-7e44-f044e50478c6 [11/27/2025-18:32:03] [I] Selected Device: NVIDIA GeForce RTX 4070 Laptop GPU [11/27/2025-18:32:03] [I] Selected Device ID: 0 [11/27/2025-18:32:03] [I] Selected Device UUID: GPU-df57adcc-000e-4631-7e44-f044e50478c6 [11/27/2025-18:32:03] [I] Compute Capability: 8.9 [11/27/2025-18:32:03] [I] SMs: 36 [11/27/2025-18:32:03] [I] Device Global Memory: 8187 MiB [11/27/2025-18:32:03] [I] Shared Memory per SM: 100 KiB [11/27/2025-18:32:03] [I] Memory Bus Width: 128 bits (ECC disabled) [11/27/2025-18:32:03] [I] Application Compute Clock Rate: 1.695 GHz [11/27/2025-18:32:03] [I] Application Memory Clock Rate: 8.001 GHz [11/27/2025-18:32:03] [I] [11/27/2025-18:32:03] [I] Note: The application clock rates do not reflect the actual clock rates that the GPU is currently running at. [11/27/2025-18:32:03] [I] [11/27/2025-18:32:03] [I] TensorRT version: 10.13.3 [11/27/2025-18:32:03] [I] Loading standard plugins [11/27/2025-18:32:03] [I] [TRT] [MemUsageChange] Init CUDA: CPU +5, GPU +0, now: CPU 19468, GPU 1133 (MiB) [11/27/2025-18:32:05] [I] [TRT] [MemUsageChange] Init builder kernel library: CPU +2273, GPU +8, now: CPU 22044, GPU 1141 (MiB) [11/27/2025-18:32:05] [I] Start parsing network model. [11/27/2025-18:32:07] [I] [TRT] ---------------------------------------------------------------- [11/27/2025-18:32:07] [I] [TRT] Input filename: D:\Audio2Face\Audio2Face-3D-SDK\_data\audio2emotion-models\audio2emotion-v2.2\network.onnx [11/27/2025-18:32:07] [I] [TRT] ONNX IR version: 0.0.7 [11/27/2025-18:32:07] [I] [TRT] Opset version: 12 [11/27/2025-18:32:07] [I] [TRT] Producer name: pytorch [11/27/2025-18:32:07] [I] [TRT] Producer version: 2.3.1 [11/27/2025-18:32:07] [I] [TRT] Domain: [11/27/2025-18:32:07] [I] [TRT] Model version: 0 [11/27/2025-18:32:07] [I] [TRT] Doc string: [11/27/2025-18:32:07] [I] [TRT] ---------------------------------------------------------------- [11/27/2025-18:32:07] [I] Finished parsing network model. Parse time: 1.56135 [11/27/2025-18:32:07] [I] Set shape of input tensor input_values for optimization profile 0 to: MIN=1x5000 OPT=8x30000 MAX=128x60000 [11/27/2025-18:32:10] [I] [TRT] Local timing cache in use. Profiling results in this builder pass will not be stored. [11/27/2025-18:32:10] [I] [TRT] Compiler backend is used during engine build. [11/27/2025-18:32:13] [W] [TRT] Tactic Device request: 8999MB Available: 8187MB. Device memory is insufficient to use tactic. [11/27/2025-18:32:13] [W] [TRT] UNSUPPORTED_STATE: Skipping tactic 1 due to insufficient memory on requested size of 9436397568 detected for tactic 0x0000000000000001. [11/27/2025-18:32:13] [W] [TRT] Tactic Device request: 8999MB Available: 8187MB. Device memory is insufficient to use tactic. [11/27/2025-18:32:13] [W] [TRT] UNSUPPORTED_STATE: Skipping tactic 1 due to insufficient memory on requested size of 9436397568 detected for tactic 0x0000000000000001. [11/27/2025-18:32:13] [W] [TRT] Tactic Device request: 8999MB Available: 8187MB. Device memory is insufficient to use tactic. [11/27/2025-18:32:13] [W] [TRT] UNSUPPORTED_STATE: Skipping tactic 1 due to insufficient memory on requested size of 9436397568 detected for tactic 0x0000000000000001. [11/27/2025-18:32:13] [W] [TRT] Tactic Device request: 8999MB Available: 8187MB. Device memory is insufficient to use tactic. [11/27/2025-18:32:13] [W] [TRT] UNSUPPORTED_STATE: Skipping tactic 1 due to insufficient memory on requested size of 9436397568 detected for tactic 0x0000000000000001. [11/27/2025-18:32:14] [W] [TRT] Tactic Device request: 12004MB Available: 8187MB. Device memory is insufficient to use tactic. [11/27/2025-18:32:14] [W] [TRT] UNSUPPORTED_STATE: Skipping tactic 0 due to insufficient memory on requested size of 12588006912 detected for tactic 0x0000000000000000. [11/27/2025-18:32:14] [E] Error[10]: IBuilder::buildSerializedNetworkToStream: Error Code 10: Internal Error (Could not find any implementation for node {ForeignNode[ONNXTRT_squeezeTensor + /model/wav2vec2/feature_extractor/conv_layers.0/Transpose...ONNXTRT_unsqueezeTensor_21]}. In nvinfer1::builder::cgraph::LeafCNode::computeCosts at optimizer/common/tactic/optimizer.cpp:4115) [11/27/2025-18:32:14] [I] Created engine with size: 0 MiB [11/27/2025-18:32:14] [I] Engine built in 6.72844 sec. [11/27/2025-18:32:14] [E] Assertion failure: false && "Attempting to access an empty engine!" 如何解决
最新发布
11-28
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值