@@ -237,7 +237,8 @@ void ReportQuantizationError(
237
237
const QuantizationErrorStats& stat) {
238
238
if (stat.sum_sq == 0 ) {
239
239
LOG (INFO) << " output " << op->debug_def ().output (0 ) << " of operator "
240
- << op << " with type " << op->debug_def ().type ()
240
+ << op << " with type " << op->debug_def ().type () << " and engine "
241
+ << op->debug_def ().engine ()
241
242
<< " has l2 relative error nan (stat.sum_err_sq "
242
243
<< stat.sum_err_sq << " stat.sum_sq 0)"
243
244
<< " and max abs error " << stat.max_abs_err << " (reference is "
@@ -247,8 +248,8 @@ void ReportQuantizationError(
247
248
<< " cnt " << stat.measure_cnt ;
248
249
} else {
249
250
LOG (INFO) << " output " << op->debug_def ().output (0 ) << " of operator "
250
- << op << " with type " << op->debug_def ().type ()
251
- << " has l2 relative error "
251
+ << op << " with type " << op->debug_def ().type () << " and engine "
252
+ << op-> debug_def (). engine () << " has l2 relative error "
252
253
<< std::sqrt (stat.sum_err_sq ) / std::sqrt (stat.sum_sq )
253
254
<< " and max abs error " << stat.max_abs_err << " (reference is "
254
255
<< stat.max_err_ref << " and actual is " << stat.max_err_actual
@@ -305,18 +306,19 @@ static unique_ptr<QuantizationFactory> GetQuantizationFactoryOf_(
305
306
" weight_quantization_kind" ,
306
307
FLAGS_caffe2_dnnlowp_weight_quantization_kind);
307
308
308
- VLOG (2 ) << " Quantization method for op with output " << op_def.output (0 )
309
- << " activation_precision " << activation_precision
310
- << " weight_precision " << weight_precision
311
- << " requantization_multiplier_precision "
312
- << requantization_multiplier_precision
313
- << " eltwise_quantization_precision "
314
- << eltwise_quantization_precision << " preserve_activation_sparsity "
315
- << preserve_activation_sparsity << " preserve_weight_sparsity "
316
- << preserve_weight_sparsity << " force_scale_power_of_two "
317
- << force_scale_power_of_two << " activation_quantization_kind "
318
- << activation_quantization_kind << " weight_quantization_kind "
319
- << weight_quantization_kind;
309
+ LOG (WARNING) << " Quantization method for op with output " << op_def.output (0 )
310
+ << " engine " << op_def.engine () << " activation_precision "
311
+ << activation_precision << " weight_precision "
312
+ << weight_precision << " requantization_multiplier_precision "
313
+ << requantization_multiplier_precision
314
+ << " eltwise_quantization_precision "
315
+ << eltwise_quantization_precision
316
+ << " preserve_activation_sparsity "
317
+ << preserve_activation_sparsity << " preserve_weight_sparsity "
318
+ << preserve_weight_sparsity << " force_scale_power_of_two "
319
+ << force_scale_power_of_two << " activation_quantization_kind "
320
+ << activation_quantization_kind << " weight_quantization_kind "
321
+ << weight_quantization_kind;
320
322
321
323
return unique_ptr<QuantizationFactory>(new QuantizationFactory (
322
324
activation_precision,
0 commit comments