From 6431fb42df1ff3a10618be79ba3dffd0dc9e6381 Mon Sep 17 00:00:00 2001 From: reminisce Date: Fri, 23 Mar 2018 11:17:10 -0700 Subject: [PATCH] Address cr --- tests/python/quantization/test_quantization.py | 3 +++ 1 file changed, 3 insertions(+) diff --git a/tests/python/quantization/test_quantization.py b/tests/python/quantization/test_quantization.py index 73d22e10bc74..7b08f46e836a 100644 --- a/tests/python/quantization/test_quantization.py +++ b/tests/python/quantization/test_quantization.py @@ -121,6 +121,7 @@ def check_requantize(shape, min_calib_range=None, max_calib_range=None): @with_seed() def test_quantized_conv(): if mx.current_context().device_type != 'gpu': + print('skipped testing quantized_conv on cpu since it is not implemented yet') return def check_quantized_conv(data_shape, kernel, num_filter, pad, stride, no_bias): @@ -191,6 +192,7 @@ def check_quantized_conv(data_shape, kernel, num_filter, pad, stride, no_bias): @with_seed() def test_quantized_pooling(): if mx.current_context().device_type != 'gpu': + print('skipped testing quantized_pooling on cpu since it is not implemented yet') return def check_quantized_pooling(data_shape, kernel, pool_type, pad, stride, global_pool): @@ -236,6 +238,7 @@ def check_quantized_pooling(data_shape, kernel, pool_type, pad, stride, global_p @with_seed() def test_quantized_fc(): if mx.current_context().device_type != 'gpu': + print('skipped testing quantized_fc on cpu since it is not implemented yet') return def check_quantized_fc(data_shape, num_hidden, no_bias, flatten=True):