@@ -105,7 +105,7 @@ void AddLayer(ConversionCtx* ctx, const torch::jit::Node* n) {
105
105
// Node input has not been converted yet or is a prim op
106
106
TORCHTRT_THROW_ERROR (
107
107
" Unable to retrieve all node inputs for node: "
108
- << util::node_info (n) << " (ctx.AddLayer)\n Specifically failed to retrieve value for input: " << *input_node );
108
+ << util::node_info (n) << " (ctx.AddLayer)\n Specifically failed to retrieve value for input: % " << input-> debugName () );
109
109
}
110
110
}
111
111
@@ -534,18 +534,21 @@ bool VerifyConverterSupportForBlock(const torch::jit::Block* b, bool suppress_er
534
534
if (unsupported_ops.size () != 0 ) {
535
535
std::stringstream unsupported_msg;
536
536
unsupported_msg
537
- << " Method requested cannot be compiled by Torch-TensorRT.TorchScript.\n Unsupported operators listed below:"
537
+ << " Method requested cannot be compiled end to end by Torch-TensorRT.TorchScript.\n Unsupported operators listed below:"
538
538
<< std::endl;
539
539
for (auto s : unsupported_ops) {
540
540
unsupported_msg << " - " << s.second << std::endl;
541
541
}
542
- unsupported_msg << " You can either implement converters for these ops in your application or request implementation"
543
- << std::endl;
544
- unsupported_msg << " https://www.github.com/nvidia/Torch-TensorRT/issues" << std::endl;
545
- unsupported_msg << std::endl << " In Module:" << std::endl;
546
542
547
543
if (!suppress_errors) {
544
+ unsupported_msg << " You can either implement converters for these ops in your application or request implementation"
545
+ << std::endl;
546
+ unsupported_msg << " https://www.github.com/nvidia/Torch-TensorRT/issues" << std::endl;
547
+ unsupported_msg << std::endl << " In Module:" << std::endl;
548
+
548
549
LOG_ERROR (unsupported_msg.str ());
550
+ } else {
551
+ LOG_INFO (unsupported_msg.str ());
549
552
}
550
553
551
554
std::unordered_map<std::string, std::unordered_set<std::string>> unsupported_node_locations;
@@ -572,7 +575,11 @@ bool VerifyConverterSupportForBlock(const torch::jit::Block* b, bool suppress_er
572
575
traceback << str;
573
576
}
574
577
auto tb_str = traceback.str ();
575
- LOG_ERROR (tb_str);
578
+ if (!suppress_errors) {
579
+ LOG_ERROR (tb_str);
580
+ } else {
581
+ LOG_DEBUG (tb_str);
582
+ }
576
583
}
577
584
578
585
return false ;
0 commit comments