From 22058a28378639cd3fbd68e293f2d62f5a767a4b Mon Sep 17 00:00:00 2001 From: Ziyi Mu Date: Thu, 6 Feb 2020 19:35:56 -0800 Subject: [PATCH] fix custom op makefile (#17516) --- Makefile | 19 +++++++++++-------- example/extensions/lib_custom_op/test_relu.py | 1 + 2 files changed, 12 insertions(+), 8 deletions(-) diff --git a/Makefile b/Makefile index 9e318920302d..2bb10c4a6582 100644 --- a/Makefile +++ b/Makefile @@ -667,18 +667,21 @@ pylint: python3 -m pylint --rcfile=$(ROOTDIR)/ci/other/pylintrc --ignore-patterns=".*\.so$$,.*\.dll$$,.*\.dylib$$" python/mxnet # MXNet extension dynamically loading libraries -EXT_LIBS = custom_op_lib subgraph_lib +EXT_LIBS = build/libcustomop_lib.so build/libsubgraph_lib.so ifeq ($(USE_CUDA), 1) - EXT_LIBS += custom_op_gpu_lib + EXT_LIBS += build/libcustomop_gpu_lib.so endif extension_libs: $(EXT_LIBS) -custom_op_lib: - $(CXX) -shared -fPIC -std=c++11 example/extensions/lib_custom_op/gemm_lib.cc -o build/libcustomop_lib.so -I include/mxnet -custom_op_gpu_lib: - $(NVCC) -shared -std=c++11 -Xcompiler -fPIC example/extensions/lib_custom_op/relu_lib.cu -o build/libcustomop_gpu_lib.so -I include/mxnet -subgraph_lib: - $(CXX) -shared -fPIC -std=c++11 example/extensions/lib_subgraph/subgraph_lib.cc -o build/libsubgraph_lib.so -I include/mxnet +build/libcustomop_lib.so: + @mkdir -p $(@D) + $(CXX) -shared -fPIC -std=c++11 example/extensions/lib_custom_op/gemm_lib.cc -o $@ -I include/mxnet +build/libcustomop_gpu_lib.so: + @mkdir -p $(@D) + $(NVCC) -shared -std=c++11 -Xcompiler -fPIC example/extensions/lib_custom_op/relu_lib.cu -o $@ -I include/mxnet +build/libsubgraph_lib.so: + @mkdir -p $(@D) + $(CXX) -shared -fPIC -std=c++11 example/extensions/lib_subgraph/subgraph_lib.cc -o $@ -I include/mxnet # Cython build cython: diff --git a/example/extensions/lib_custom_op/test_relu.py b/example/extensions/lib_custom_op/test_relu.py index ce2b2fe99cf0..03d02f32d633 100644 --- a/example/extensions/lib_custom_op/test_relu.py +++ b/example/extensions/lib_custom_op/test_relu.py @@ -65,6 +65,7 @@ print("--------start testing larger ndarray---------") a = mx.nd.uniform(shape=(100,100,100), ctx=mx.cpu()) b = mx.nd.uniform(shape=(100,100,100), ctx=mx.gpu()) +mx.nd.waitall() t1 = time.time() r1 = mx.nd.my_relu(a) mx.nd.waitall()