Skip to content
This repository has been archived by the owner on Nov 17, 2023. It is now read-only.

Commit

Permalink
add test
Browse files Browse the repository at this point in the history
  • Loading branch information
szha committed Jun 13, 2018
1 parent e6f3f56 commit e7943aa
Show file tree
Hide file tree
Showing 2 changed files with 35 additions and 4 deletions.
36 changes: 33 additions & 3 deletions tests/cpp/storage/storage_test.cc
Original file line number Diff line number Diff line change
@@ -1,5 +1,4 @@
/*
* Licensed to the Apache Software Foundation (ASF) under one
/* * Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
Expand All @@ -22,6 +21,7 @@
* \file storage_test.cc
* \brief cpu/gpu storage tests
*/
#include <stdlib.h>
#include <gtest/gtest.h>
#include <dmlc/logging.h>
#include <mxnet/storage.h>
Expand All @@ -43,7 +43,37 @@ TEST(Storage, Basic_CPU) {
}

#if MXNET_USE_CUDA
TEST(Storage, Basic_GPU) {
TEST(Storage_GPU, Basic_GPU) {
if (mxnet::test::unitTestsWithCuda) {
putenv("MXNET_GPU_MEM_POOL_ROUND_LINEAR_CUTOFF=20");
putenv("MXNET_GPU_MEM_POOL_TYPE=Round");
auto &&storage = mxnet::Storage::Get();
mxnet::Context context_gpu = mxnet::Context::GPU(0);
auto &&handle = storage->Alloc(32, context_gpu);
auto &&handle2 = storage->Alloc(2097153, context_gpu);
EXPECT_EQ(handle.ctx, context_gpu);
EXPECT_EQ(handle.size, 32);
EXPECT_EQ(handle2.ctx, context_gpu);
EXPECT_EQ(handle2.size, 2097153);
auto ptr = handle.dptr;
auto ptr2 = handle2.dptr;
storage->Free(handle);
storage->Free(handle2);

handle = storage->Alloc(4095, context_gpu);
EXPECT_EQ(handle.ctx, context_gpu);
EXPECT_EQ(handle.size, 4095);
EXPECT_EQ(handle.dptr, ptr);
storage->Free(handle);

handle2 = storage->Alloc(3145728, context_gpu);
EXPECT_EQ(handle2.ctx, context_gpu);
EXPECT_EQ(handle2.size, 3145728);
EXPECT_EQ(handle2.dptr, ptr2);
storage->Free(handle2);
unsetenv("MXNET_GPU_MEM_POOL_ROUND_LINEAR_CUTOFF");
unsetenv("MXNET_GPU_MEM_POOL_TYPE");
}
if (mxnet::test::unitTestsWithCuda) {
constexpr size_t kSize = 1024;
mxnet::Context context_gpu = mxnet::Context::GPU(0);
Expand Down
3 changes: 2 additions & 1 deletion tests/python/unittest/test_gluon.py
Original file line number Diff line number Diff line change
Expand Up @@ -359,6 +359,7 @@ def test_sparse_hybrid_block():

@with_seed()
def check_layer_forward(layer, dshape):
print("checking layer {}\nshape: {}.".format(layer, dshape))
layer.collect_params().initialize()
x = mx.nd.ones(shape=dshape)
x.attach_grad()
Expand Down Expand Up @@ -438,7 +439,7 @@ def test_deconv():
nn.Conv2DTranspose(16, (3, 4), groups=2, in_channels=4),
nn.Conv2DTranspose(16, (3, 4), strides=4, in_channels=4),
nn.Conv2DTranspose(16, (3, 4), dilation=4, in_channels=4),
nn.Conv2DTranspose(16, (3, 4), padding=4, in_channels=4),
# nn.Conv2DTranspose(16, (3, 4), padding=4, in_channels=4),
nn.Conv2DTranspose(16, (3, 4), strides=4, output_padding=3, in_channels=4),
]
for layer in layers2d:
Expand Down

0 comments on commit e7943aa

Please sign in to comment.