mirror of
https://github.com/saymrwulf/pytorch.git
synced 2026-05-15 21:00:47 +00:00
Summary: Was https://github.com/caffe2/caffe2/pull/1151 Closes https://github.com/caffe2/caffe2/pull/1179 Differential Revision: D5794650 Pulled By: akyrola fbshipit-source-id: 63e7973a88193a3b74ac4ba677df737889cbf0b6
147 lines
3.6 KiB
Python
147 lines
3.6 KiB
Python
from __future__ import absolute_import
|
|
from __future__ import division
|
|
from __future__ import print_function
|
|
from __future__ import unicode_literals
|
|
|
|
from caffe2.python import core
|
|
from hypothesis import assume, given
|
|
import caffe2.python.hypothesis_test_util as hu
|
|
import hypothesis.strategies as st
|
|
import numpy as np
|
|
from caffe2.proto import caffe2_pb2
|
|
|
|
|
|
class TestReductionOps(hu.HypothesisTestCase):
|
|
|
|
@given(n=st.integers(5, 8), **hu.gcs)
|
|
def test_elementwise_sum(self, n, gc, dc):
|
|
X = np.random.rand(n).astype(np.float32)
|
|
|
|
def sum_op(X):
|
|
return [np.sum(X)]
|
|
|
|
op = core.CreateOperator(
|
|
"SumElements",
|
|
["X"],
|
|
["y"]
|
|
)
|
|
|
|
self.assertReferenceChecks(
|
|
device_option=gc,
|
|
op=op,
|
|
inputs=[X],
|
|
reference=sum_op,
|
|
)
|
|
|
|
self.assertGradientChecks(
|
|
device_option=gc,
|
|
op=op,
|
|
inputs=[X],
|
|
outputs_to_check=0,
|
|
outputs_with_grads=[0],
|
|
)
|
|
|
|
@given(n=st.integers(1, 65536),
|
|
dtype=st.sampled_from([np.float32, np.float16]),
|
|
**hu.gcs)
|
|
def test_elementwise_sqrsum(self, n, dtype, gc, dc):
|
|
if dtype == np.float16:
|
|
# fp16 is only supported with CUDA
|
|
assume(gc.device_type == caffe2_pb2.CUDA)
|
|
dc = [d for d in dc if d.device_type == caffe2_pb2.CUDA]
|
|
|
|
X = np.random.rand(n).astype(dtype)
|
|
|
|
def sumsqr_op(X):
|
|
return [np.sum(X * X)]
|
|
|
|
op = core.CreateOperator(
|
|
"SumSqrElements",
|
|
["X"],
|
|
["y"]
|
|
)
|
|
|
|
threshold = 0.01 if dtype == np.float16 else 0.005
|
|
|
|
self.assertReferenceChecks(
|
|
device_option=gc,
|
|
op=op,
|
|
inputs=[X],
|
|
reference=sumsqr_op,
|
|
threshold=threshold,
|
|
)
|
|
|
|
@given(n=st.integers(5, 8), **hu.gcs)
|
|
def test_elementwise_avg(self, n, gc, dc):
|
|
X = np.random.rand(n).astype(np.float32)
|
|
|
|
def avg_op(X):
|
|
return [np.mean(X)]
|
|
|
|
op = core.CreateOperator(
|
|
"SumElements",
|
|
["X"],
|
|
["y"],
|
|
average=1
|
|
)
|
|
|
|
self.assertReferenceChecks(
|
|
device_option=gc,
|
|
op=op,
|
|
inputs=[X],
|
|
reference=avg_op,
|
|
)
|
|
|
|
self.assertGradientChecks(
|
|
device_option=gc,
|
|
op=op,
|
|
inputs=[X],
|
|
outputs_to_check=0,
|
|
outputs_with_grads=[0],
|
|
)
|
|
|
|
@given(batch_size=st.integers(1, 3),
|
|
m=st.integers(1, 3),
|
|
n=st.integers(1, 4),
|
|
**hu.gcs)
|
|
def test_rowwise_max(self, batch_size, m, n, gc, dc):
|
|
X = np.random.rand(batch_size, m, n).astype(np.float32)
|
|
|
|
def rowwise_max(X):
|
|
return [np.max(X, axis=2)]
|
|
|
|
op = core.CreateOperator(
|
|
"RowwiseMax",
|
|
["x"],
|
|
["y"]
|
|
)
|
|
|
|
self.assertReferenceChecks(
|
|
device_option=gc,
|
|
op=op,
|
|
inputs=[X],
|
|
reference=rowwise_max,
|
|
)
|
|
|
|
@given(batch_size=st.integers(1, 3),
|
|
m=st.integers(1, 3),
|
|
n=st.integers(1, 4),
|
|
**hu.gcs)
|
|
def test_columnwise_max(self, batch_size, m, n, gc, dc):
|
|
X = np.random.rand(batch_size, m, n).astype(np.float32)
|
|
|
|
def columnwise_max(X):
|
|
return [np.max(X, axis=1)]
|
|
|
|
op = core.CreateOperator(
|
|
"ColwiseMax",
|
|
["x"],
|
|
["y"]
|
|
)
|
|
|
|
self.assertReferenceChecks(
|
|
device_option=gc,
|
|
op=op,
|
|
inputs=[X],
|
|
reference=columnwise_max,
|
|
)
|