Skip to content

Commit

Permalink
C is not C++
Browse files Browse the repository at this point in the history
  • Loading branch information
uecker committed Jun 28, 2024
1 parent 2af481f commit b919255
Show file tree
Hide file tree
Showing 6 changed files with 14 additions and 12 deletions.
4 changes: 3 additions & 1 deletion Makefile
Original file line number Diff line number Diff line change
Expand Up @@ -96,8 +96,10 @@ LDFLAGS= -lm -pthread
COMMON= -Iinclude/ -I3rdparty/stb/include
CFLAGS=-Wall -Wno-unused-parameter -Wno-unknown-pragmas -fPIC -rdynamic

ifneq ($(USE_CPP), 1)
ifeq ($(USE_CPP), 1)
# C is not C++
CFLAGS+=-fpermissive -Wno-write-strings
else
CFLAGS+=-Wextra
endif

Expand Down
8 changes: 4 additions & 4 deletions src/gaussian_yolo_layer.c
Original file line number Diff line number Diff line change
Expand Up @@ -67,14 +67,14 @@ layer make_gaussian_yolo_layer(int batch, int w, int h, int n, int total, int *m


free(l.output);
if (cudaSuccess == cudaHostAlloc((void*)&l.output, batch*l.outputs * sizeof(float), cudaHostRegisterMapped)) l.output_pinned = 1;
if (cudaSuccess == cudaHostAlloc((void**)&l.output, batch*l.outputs * sizeof(float), cudaHostRegisterMapped)) l.output_pinned = 1;
else {
cudaGetLastError(); // reset CUDA-error
l.output = (float*)calloc(batch * l.outputs, sizeof(float));
}

free(l.delta);
if (cudaSuccess == cudaHostAlloc((void*)&l.delta, batch*l.outputs * sizeof(float), cudaHostRegisterMapped)) l.delta_pinned = 1;
if (cudaSuccess == cudaHostAlloc((void**)&l.delta, batch*l.outputs * sizeof(float), cudaHostRegisterMapped)) l.delta_pinned = 1;
else {
cudaGetLastError(); // reset CUDA-error
l.delta = (float*)calloc(batch * l.outputs, sizeof(float));
Expand Down Expand Up @@ -106,7 +106,7 @@ void resize_gaussian_yolo_layer(layer *l, int w, int h)

if (l->output_pinned) {
CHECK_CUDA(cudaFreeHost(l->output));
if (cudaSuccess != cudaHostAlloc((void*)&l->output, l->batch*l->outputs * sizeof(float), cudaHostRegisterMapped)) {
if (cudaSuccess != cudaHostAlloc((void**)&l->output, l->batch*l->outputs * sizeof(float), cudaHostRegisterMapped)) {
cudaGetLastError(); // reset CUDA-error
l->output = (float*)calloc(l->batch * l->outputs, sizeof(float));
l->output_pinned = 0;
Expand All @@ -115,7 +115,7 @@ void resize_gaussian_yolo_layer(layer *l, int w, int h)

if (l->delta_pinned) {
CHECK_CUDA(cudaFreeHost(l->delta));
if (cudaSuccess != cudaHostAlloc((void*)&l->delta, l->batch*l->outputs * sizeof(float), cudaHostRegisterMapped)) {
if (cudaSuccess != cudaHostAlloc((void**)&l->delta, l->batch*l->outputs * sizeof(float), cudaHostRegisterMapped)) {
cudaGetLastError(); // reset CUDA-error
l->delta = (float*)calloc(l->batch * l->outputs, sizeof(float));
l->delta_pinned = 0;
Expand Down
2 changes: 1 addition & 1 deletion src/gemm.c
Original file line number Diff line number Diff line change
Expand Up @@ -558,7 +558,7 @@ static inline float _mm256_extract_float32(__m256 a, const int index) {

static inline float _dn_castu32_f32(uint32_t a) {
float b;
_Static_assert(sizeof(a) == sizeof(b), "");
static_assert(sizeof(a) == sizeof(b), "");
memcpy(&b, &a, sizeof(b));
return b;
}
Expand Down
2 changes: 1 addition & 1 deletion src/network.c
Original file line number Diff line number Diff line change
Expand Up @@ -657,7 +657,7 @@ int resize_network(network *net, int w, int h)
printf(" try to allocate additional workspace_size = %1.2f MB \n", (float)workspace_size / 1000000);
net->workspace = cuda_make_array(0, workspace_size/sizeof(float) + 1);
net->input_state_gpu = cuda_make_array(0, size);
if (cudaSuccess == cudaHostAlloc((void*)&net->input_pinned_cpu, size * sizeof(float), cudaHostRegisterMapped))
if (cudaSuccess == cudaHostAlloc((void**)&net->input_pinned_cpu, size * sizeof(float), cudaHostRegisterMapped))
net->input_pinned_cpu_flag = 1;
else {
cudaGetLastError(); // reset CUDA-error
Expand Down
2 changes: 1 addition & 1 deletion src/parser.c
Original file line number Diff line number Diff line change
Expand Up @@ -1758,7 +1758,7 @@ network parse_network_cfg_custom(char *filename, int batch, int time_steps)
{
int size = get_network_input_size(net) * net.batch;
net.input_state_gpu = cuda_make_array(0, size);
if (cudaSuccess == cudaHostAlloc((void*)&net.input_pinned_cpu, size * sizeof(float), cudaHostRegisterMapped)) net.input_pinned_cpu_flag = 1;
if (cudaSuccess == cudaHostAlloc((void**)&net.input_pinned_cpu, size * sizeof(float), cudaHostRegisterMapped)) net.input_pinned_cpu_flag = 1;
else {
cudaGetLastError(); // reset CUDA-error
net.input_pinned_cpu = (float*)xcalloc(size, sizeof(float));
Expand Down
8 changes: 4 additions & 4 deletions src/yolo_layer.c
Original file line number Diff line number Diff line change
Expand Up @@ -64,14 +64,14 @@ layer make_yolo_layer(int batch, int w, int h, int n, int total, int *mask, int
l.delta_gpu = cuda_make_array(l.delta, batch*l.outputs);

free(l.output);
if (cudaSuccess == cudaHostAlloc((void*)&l.output, batch*l.outputs*sizeof(float), cudaHostRegisterMapped)) l.output_pinned = 1;
if (cudaSuccess == cudaHostAlloc((void**)&l.output, batch*l.outputs*sizeof(float), cudaHostRegisterMapped)) l.output_pinned = 1;
else {
cudaGetLastError(); // reset CUDA-error
l.output = (float*)xcalloc(batch * l.outputs, sizeof(float));
}

free(l.delta);
if (cudaSuccess == cudaHostAlloc((void*)&l.delta, batch*l.outputs*sizeof(float), cudaHostRegisterMapped)) l.delta_pinned = 1;
if (cudaSuccess == cudaHostAlloc((void**)&l.delta, batch*l.outputs*sizeof(float), cudaHostRegisterMapped)) l.delta_pinned = 1;
else {
cudaGetLastError(); // reset CUDA-error
l.delta = (float*)xcalloc(batch * l.outputs, sizeof(float));
Expand Down Expand Up @@ -102,7 +102,7 @@ void resize_yolo_layer(layer *l, int w, int h)
#ifdef GPU
if (l->output_pinned) {
CHECK_CUDA(cudaFreeHost(l->output));
if (cudaSuccess != cudaHostAlloc((void*)&l->output, l->batch*l->outputs * sizeof(float), cudaHostRegisterMapped)) {
if (cudaSuccess != cudaHostAlloc((void**)&l->output, l->batch*l->outputs * sizeof(float), cudaHostRegisterMapped)) {
cudaGetLastError(); // reset CUDA-error
l->output = (float*)xcalloc(l->batch * l->outputs, sizeof(float));
l->output_pinned = 0;
Expand All @@ -111,7 +111,7 @@ void resize_yolo_layer(layer *l, int w, int h)

if (l->delta_pinned) {
CHECK_CUDA(cudaFreeHost(l->delta));
if (cudaSuccess != cudaHostAlloc((void*)&l->delta, l->batch*l->outputs * sizeof(float), cudaHostRegisterMapped)) {
if (cudaSuccess != cudaHostAlloc((void**)&l->delta, l->batch*l->outputs * sizeof(float), cudaHostRegisterMapped)) {
cudaGetLastError(); // reset CUDA-error
l->delta = (float*)xcalloc(l->batch * l->outputs, sizeof(float));
l->delta_pinned = 0;
Expand Down

0 comments on commit b919255

Please sign in to comment.