mirror of
https://github.com/saymrwulf/pytorch.git
synced 2026-05-15 21:00:47 +00:00
Summary: Turns out CuDNN's tensor transform only supports floats. Previous implementation pretended it would work with ints by casting to floats and indeed passed tests for some reason. But rgirdhar found a case where it returned nonsensical results. So rewire int-transposes to use non-cudnn version. Had to refactor a bit for that. Also added a test for the case. Reviewed By: asaadaldien Differential Revision: D6043284 fbshipit-source-id: cc3b14f9fbbdeff421b01da453a1d3c7c5ffd4ac
354 lines
11 KiB
Python
354 lines
11 KiB
Python
# Copyright (c) 2016-present, Facebook, Inc.
|
|
#
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
# you may not use this file except in compliance with the License.
|
|
# You may obtain a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
# See the License for the specific language governing permissions and
|
|
# limitations under the License.
|
|
##############################################################################
|
|
|
|
from __future__ import absolute_import
|
|
from __future__ import division
|
|
from __future__ import print_function
|
|
from __future__ import unicode_literals
|
|
|
|
from caffe2.python import core, workspace
|
|
from hypothesis import given
|
|
from caffe2.proto import caffe2_pb2
|
|
import caffe2.python.hypothesis_test_util as hu
|
|
import hypothesis.strategies as st
|
|
import numpy as np
|
|
import random
|
|
import unittest
|
|
|
|
class TestUtilityOps(hu.HypothesisTestCase):
|
|
|
|
@given(X=hu.tensor(), args=st.booleans(), **hu.gcs)
|
|
def test_slice(self, X, args, gc, dc):
|
|
X = X.astype(dtype=np.float32)
|
|
dim = random.randint(0, X.ndim - 1)
|
|
slice_start = random.randint(0, X.shape[dim] - 1)
|
|
slice_end = random.randint(slice_start, X.shape[dim] - 1)
|
|
starts = np.array([0] * X.ndim).astype(np.int32)
|
|
ends = np.array([-1] * X.ndim).astype(np.int32)
|
|
starts[dim] = slice_start
|
|
ends[dim] = slice_end
|
|
|
|
if args:
|
|
op = core.CreateOperator(
|
|
"Slice", ["X"], ["Y"], starts=starts, ends=ends, device_option=gc
|
|
)
|
|
|
|
def slice_ref(X):
|
|
slc = [slice(None)] * X.ndim
|
|
slc[dim] = slice(slice_start, slice_end)
|
|
return [X[slc]]
|
|
inputs = [X]
|
|
else:
|
|
op = core.CreateOperator(
|
|
"Slice", ["X", "starts", "ends"], ["Y"], device_option=gc
|
|
)
|
|
|
|
def slice_ref(x, starts, ends):
|
|
slc = [slice(None)] * x.ndim
|
|
slc[dim] = slice(slice_start, slice_end)
|
|
return [x[slc]]
|
|
inputs = [X, starts, ends]
|
|
|
|
self.assertReferenceChecks(gc, op, inputs, slice_ref)
|
|
self.assertDeviceChecks(dc, op, inputs, [0])
|
|
self.assertGradientChecks(
|
|
device_option=gc,
|
|
op=op,
|
|
inputs=inputs,
|
|
outputs_to_check=0,
|
|
outputs_with_grads=[0],
|
|
)
|
|
|
|
@given(dtype=st.sampled_from([np.float32, np.int32]),
|
|
ndims=st.integers(min_value=1, max_value=5),
|
|
seed=st.integers(min_value=0, max_value=65536),
|
|
null_axes=st.booleans(),
|
|
engine=st.sampled_from(['CUDNN', None]),
|
|
**hu.gcs)
|
|
def test_transpose(self, dtype, ndims, seed, null_axes, engine, gc, dc):
|
|
dims = (np.random.rand(ndims) * 16 + 1).astype(np.int32)
|
|
X = (np.random.rand(*dims) * 16).astype(dtype)
|
|
|
|
if null_axes:
|
|
axes = None
|
|
op = core.CreateOperator(
|
|
"Transpose",
|
|
["input"], ["output"],
|
|
engine=engine)
|
|
else:
|
|
np.random.seed(int(seed))
|
|
axes = [int(v) for v in list(np.random.permutation(X.ndim))]
|
|
op = core.CreateOperator(
|
|
"Transpose",
|
|
["input"], ["output"],
|
|
axes=axes,
|
|
engine=engine)
|
|
|
|
def transpose_ref(x, axes):
|
|
return (np.transpose(x, axes),)
|
|
|
|
self.assertReferenceChecks(gc, op, [X, axes],
|
|
transpose_ref)
|
|
|
|
@unittest.skipIf(not workspace.has_gpu_support, "No gpu support")
|
|
def test_gpu_transpose_minusones(self):
|
|
'''
|
|
Repro a problem with earlier version of CuDNN Transpose Op that
|
|
casted ints to floats.
|
|
'''
|
|
X = -np.ones((2, 10)).astype(np.int32)
|
|
with core.DeviceScope(core.DeviceOption(caffe2_pb2.CUDA, 0)):
|
|
workspace.FeedBlob("X", X)
|
|
print("X:\n{}\n".format(workspace.FetchBlob("X")))
|
|
op = core.CreateOperator(
|
|
"Transpose",
|
|
["X"],
|
|
["Y"],
|
|
engine='CUDNN'
|
|
)
|
|
workspace.RunOperatorOnce(op)
|
|
Y = workspace.FetchBlob("Y")
|
|
print("Y:\n{}\n".format(Y))
|
|
|
|
for j in list(Y.flatten()):
|
|
self.assertEqual(-1, j)
|
|
|
|
|
|
|
|
@given(m=st.integers(5, 10), n=st.integers(5, 10),
|
|
o=st.integers(5, 10), nans=st.booleans(), **hu.gcs)
|
|
def test_nan_check(self, m, n, o, nans, gc, dc):
|
|
other = np.array([1, 2, 3]).astype(np.float32)
|
|
X = np.random.rand(m, n, o).astype(np.float32)
|
|
if nans:
|
|
x_nan = np.random.randint(0, m)
|
|
y_nan = np.random.randint(0, n)
|
|
z_nan = np.random.randint(0, o)
|
|
X[x_nan, y_nan, z_nan] = float('NaN')
|
|
|
|
# print('nans: {}'.format(nans))
|
|
# print(X)
|
|
|
|
def nan_reference(X, Y):
|
|
if not np.isnan(X).any():
|
|
return [X]
|
|
else:
|
|
return [np.array([])]
|
|
|
|
op = core.CreateOperator(
|
|
"NanCheck",
|
|
["X", "other"],
|
|
["Y"]
|
|
)
|
|
|
|
try:
|
|
self.assertReferenceChecks(
|
|
device_option=gc,
|
|
op=op,
|
|
inputs=[X, other],
|
|
reference=nan_reference,
|
|
)
|
|
if nans:
|
|
self.assertTrue(False, "Did not fail when presented with NaN!")
|
|
except RuntimeError:
|
|
self.assertTrue(nans, "No NaNs but failed")
|
|
|
|
try:
|
|
self.assertGradientChecks(
|
|
device_option=gc,
|
|
op=op,
|
|
inputs=[X],
|
|
outputs_to_check=0,
|
|
outputs_with_grads=[0],
|
|
)
|
|
if nans:
|
|
self.assertTrue(False, "Did not fail when gradient had NaN!")
|
|
except RuntimeError:
|
|
pass
|
|
|
|
@given(n=st.integers(4, 5), m=st.integers(6, 7),
|
|
d=st.integers(2, 3), **hu.gcs)
|
|
def test_elementwise_max(self, n, m, d, gc, dc):
|
|
X = np.random.rand(n, m, d).astype(np.float32)
|
|
Y = np.random.rand(n, m, d).astype(np.float32)
|
|
Z = np.random.rand(n, m, d).astype(np.float32)
|
|
|
|
def max_op(X, Y, Z):
|
|
return [np.maximum(np.maximum(X, Y), Z)]
|
|
|
|
op = core.CreateOperator(
|
|
"Max",
|
|
["X", "Y", "Z"],
|
|
["mx"]
|
|
)
|
|
|
|
self.assertReferenceChecks(
|
|
device_option=gc,
|
|
op=op,
|
|
inputs=[X, Y, Z],
|
|
reference=max_op,
|
|
)
|
|
|
|
@given(n=st.integers(4, 5), m=st.integers(6, 7),
|
|
d=st.integers(2, 3), **hu.gcs)
|
|
def test_elementwise_max_grad(self, n, m, d, gc, dc):
|
|
go = np.random.rand(n, m, d).astype(np.float32)
|
|
X = np.random.rand(n, m, d).astype(np.float32)
|
|
Y = np.random.rand(n, m, d).astype(np.float32)
|
|
Z = np.random.rand(n, m, d).astype(np.float32)
|
|
mx = np.maximum(np.maximum(X, Y), Z)
|
|
|
|
def max_grad_op(mx, go, X, Y, Z):
|
|
def mx_grad(a):
|
|
return go * (mx == a)
|
|
|
|
return [mx_grad(a) for a in [X, Y, Z]]
|
|
|
|
op = core.CreateOperator(
|
|
"MaxGradient",
|
|
["mx", "go", "X", "Y", "Z"],
|
|
["gX", "gY", "gZ"]
|
|
)
|
|
|
|
self.assertReferenceChecks(
|
|
device_option=gc,
|
|
op=op,
|
|
inputs=[mx, go, X, Y, Z],
|
|
reference=max_grad_op,
|
|
)
|
|
|
|
@given(
|
|
inputs=hu.lengths_tensor().flatmap(
|
|
lambda pair: st.tuples(
|
|
st.just(pair[0]),
|
|
st.just(pair[1]),
|
|
hu.dims(max_value=len(pair[1])),
|
|
)
|
|
).flatmap(
|
|
lambda tup: st.tuples(
|
|
st.just(tup[0]),
|
|
st.just(tup[1]),
|
|
hu.arrays(
|
|
tup[2], dtype=np.int32,
|
|
elements=st.integers(
|
|
min_value=0, max_value=len(tup[1]) - 1)),
|
|
)
|
|
),
|
|
**hu.gcs_cpu_only)
|
|
def test_lengths_gather(self, inputs, gc, dc):
|
|
items = inputs[0]
|
|
lengths = inputs[1]
|
|
indices = inputs[2]
|
|
|
|
def lengths_gather_op(items, lengths, indices):
|
|
ends = np.cumsum(lengths)
|
|
return [np.concatenate(
|
|
list(items[ends[i] - lengths[i]:ends[i]] for i in indices))]
|
|
|
|
op = core.CreateOperator(
|
|
"LengthsGather",
|
|
["items", "lengths", "indices"],
|
|
["output"]
|
|
)
|
|
|
|
self.assertReferenceChecks(
|
|
device_option=gc,
|
|
op=op,
|
|
inputs=[items, lengths, indices],
|
|
reference=lengths_gather_op,
|
|
)
|
|
|
|
@given(**hu.gcs)
|
|
def test_size_op(self, gc, dc):
|
|
X = np.array([[1, 2], [3, 4]]).astype(np.float32)
|
|
|
|
def size_op(tensor):
|
|
return [np.prod(tensor.shape)]
|
|
|
|
op = core.CreateOperator(
|
|
"Size",
|
|
["X"],
|
|
["output"]
|
|
)
|
|
|
|
self.assertReferenceChecks(
|
|
device_option=gc,
|
|
op=op,
|
|
inputs=[X],
|
|
reference=size_op,
|
|
)
|
|
|
|
def test_alias_op(self):
|
|
""" Don't use hypothesis because there are only 2 cases to check"""
|
|
for size in [0, 5]:
|
|
X = np.arange(size).astype(np.float32)
|
|
workspace.FeedBlob('X', X)
|
|
|
|
op = core.CreateOperator(
|
|
"Alias",
|
|
["X"],
|
|
["Y"]
|
|
)
|
|
workspace.RunOperatorOnce(op)
|
|
Y = workspace.FetchBlob('Y')
|
|
np.testing.assert_array_equal(X, Y)
|
|
|
|
@given(**hu.gcs)
|
|
def test_range(self, gc, dc):
|
|
names = [
|
|
('stop_',),
|
|
('start_', 'stop_'),
|
|
('start_', 'stop_', 'step_'),
|
|
]
|
|
# Most random values aren't great here, so use a fixed set instead of
|
|
# hypothesis.
|
|
for inputs in (
|
|
(10,),
|
|
(np.float32(10.0),),
|
|
(0,),
|
|
(0, 0),
|
|
(10., 5.0, -1.),
|
|
(2, 10000),
|
|
(2, 10000, 20000),
|
|
(2, 10000, -1),
|
|
):
|
|
inputs = [np.array(v) for v in inputs]
|
|
op = core.CreateOperator(
|
|
"Range",
|
|
names[len(inputs) - 1],
|
|
["Y"]
|
|
)
|
|
|
|
self.assertReferenceChecks(
|
|
device_option=gc,
|
|
op=op,
|
|
inputs=inputs,
|
|
reference=lambda *x: [np.arange(*x)],
|
|
)
|
|
self.assertDeviceChecks(dc, op, inputs, [0])
|
|
|
|
with self.assertRaisesRegexp(RuntimeError, 'Step size cannot be 0'):
|
|
inputs = (np.array(0), np.array(10), np.array(0))
|
|
op = core.CreateOperator(
|
|
"Range",
|
|
names[len(inputs) - 1],
|
|
["Y"]
|
|
)
|
|
self.assertReferenceChecks(
|
|
device_option=gc,
|
|
op=op,
|
|
inputs=inputs,
|
|
reference=lambda *x: [np.arange(*x)],
|
|
)
|