mirror of
https://github.com/saymrwulf/pytorch.git
synced 2026-05-15 21:00:47 +00:00
Fixes #75177 Couldn't find any utility method to get directory name in pytorch repo, hence creating a function for that. Let me know if a new function is not needed. I also referred [this](https://github.com/pytorch/pytorch/blob/master/c10/test/util/tempfile_test.cpp#L15) for directory check. Also I am using TORCH_CHECK to show the error. This is highly verbose with the entire stack visible. Is there any alternative for the same so that it is easier to read? This could happen a frequently, so small and concise error would be more helpful here. Pull Request resolved: https://github.com/pytorch/pytorch/pull/75681 Approved by: https://github.com/albanD
263 lines
8.2 KiB
C++
263 lines
8.2 KiB
C++
#include <gtest/gtest.h>
|
|
|
|
#include <test/cpp/jit/test_utils.h>
|
|
#include <sstream>
|
|
|
|
#include <torch/csrc/jit/mobile/module.h>
|
|
#include <torch/csrc/jit/serialization/export.h>
|
|
#include <torch/csrc/jit/serialization/export_bytecode.h>
|
|
#include <torch/csrc/jit/serialization/import.h>
|
|
#include <torch/csrc/jit/serialization/import_source.h>
|
|
#include <torch/torch.h>
|
|
|
|
#include "caffe2/serialize/istream_adapter.h"
|
|
|
|
namespace torch {
|
|
namespace jit {
|
|
|
|
namespace {
|
|
|
|
Module roundtripThroughMobile(const Module& m) {
|
|
ExtraFilesMap files;
|
|
std::vector<IValue> constants;
|
|
jitModuleToPythonCodeAndConstants(m, &files, &constants);
|
|
CompilationOptions options;
|
|
mobile::Module mobilem = jitModuleToMobile(m, options);
|
|
return jitModuleFromSourceAndConstants(
|
|
mobilem._ivalue(), files, constants, 8);
|
|
}
|
|
|
|
template <class Functor>
|
|
inline void expectThrowsEq(Functor&& functor, const char* expectedMessage) {
|
|
try {
|
|
std::forward<Functor>(functor)();
|
|
} catch (const Error& e) {
|
|
EXPECT_STREQ(e.what_without_backtrace(), expectedMessage);
|
|
return;
|
|
}
|
|
ADD_FAILURE() << "Expected to throw exception with message \""
|
|
<< expectedMessage << "\" but didn't throw";
|
|
}
|
|
|
|
} // namespace
|
|
|
|
TEST(SerializationTest, ExtraFilesHookPreference) {
|
|
// Tests that an extra file written explicitly has precedence over
|
|
// extra files written by a hook
|
|
// TODO: test for the warning, too
|
|
const auto script = R"JIT(
|
|
def forward(self):
|
|
x = torch.rand(5, 5)
|
|
x = x.mm(x)
|
|
return x
|
|
)JIT";
|
|
|
|
auto module =
|
|
std::make_shared<Module>("Module", std::make_shared<CompilationUnit>());
|
|
module->define(script);
|
|
std::ostringstream oss;
|
|
std::unordered_map<std::string, std::string> extra_files;
|
|
extra_files["metadata.json"] = "abc";
|
|
SetExportModuleExtraFilesHook([](const Module&) -> ExtraFilesMap {
|
|
return {{"metadata.json", "def"}};
|
|
});
|
|
module->save(oss, extra_files);
|
|
SetExportModuleExtraFilesHook(nullptr);
|
|
|
|
std::istringstream iss(oss.str());
|
|
caffe2::serialize::IStreamAdapter adapter{&iss};
|
|
std::unordered_map<std::string, std::string> loaded_extra_files;
|
|
loaded_extra_files["metadata.json"] = "";
|
|
auto loaded_module = torch::jit::load(iss, torch::kCPU, loaded_extra_files);
|
|
ASSERT_EQ(loaded_extra_files["metadata.json"], "abc");
|
|
}
|
|
|
|
TEST(SerializationTest, ExtraFileHooksNoSecret) {
|
|
// no secrets
|
|
std::stringstream ss;
|
|
{
|
|
Module m("__torch__.m");
|
|
ExtraFilesMap extra;
|
|
extra["metadata.json"] = "abc";
|
|
m.save(ss, extra);
|
|
}
|
|
ss.seekg(0);
|
|
{
|
|
ExtraFilesMap extra;
|
|
extra["metadata.json"] = "";
|
|
extra["secret.json"] = "";
|
|
jit::load(ss, c10::nullopt, extra);
|
|
ASSERT_EQ(extra["metadata.json"], "abc");
|
|
ASSERT_EQ(extra["secret.json"], "");
|
|
}
|
|
}
|
|
|
|
TEST(SerializationTest, ExtraFileHooksWithSecret) {
|
|
std::stringstream ss;
|
|
{
|
|
SetExportModuleExtraFilesHook([](const Module&) -> ExtraFilesMap {
|
|
return {{"secret.json", "topsecret"}};
|
|
});
|
|
Module m("__torch__.m");
|
|
ExtraFilesMap extra;
|
|
extra["metadata.json"] = "abc";
|
|
m.save(ss, extra);
|
|
SetExportModuleExtraFilesHook(nullptr);
|
|
}
|
|
ss.seekg(0);
|
|
{
|
|
ExtraFilesMap extra;
|
|
extra["metadata.json"] = "";
|
|
extra["secret.json"] = "";
|
|
jit::load(ss, c10::nullopt, extra);
|
|
ASSERT_EQ(extra["metadata.json"], "abc");
|
|
ASSERT_EQ(extra["secret.json"], "topsecret");
|
|
}
|
|
}
|
|
|
|
TEST(SerializationTest, TypeTags) {
|
|
auto list = c10::List<c10::List<int64_t>>();
|
|
list.push_back(c10::List<int64_t>({1, 2, 3}));
|
|
list.push_back(c10::List<int64_t>({4, 5, 6}));
|
|
auto dict = c10::Dict<std::string, at::Tensor>();
|
|
dict.insert("Hello", torch::ones({2, 2}));
|
|
auto dict_list = c10::List<c10::Dict<std::string, at::Tensor>>();
|
|
for (size_t i = 0; i < 5; i++) {
|
|
auto another_dict = c10::Dict<std::string, at::Tensor>();
|
|
another_dict.insert("Hello" + std::to_string(i), torch::ones({2, 2}));
|
|
dict_list.push_back(another_dict);
|
|
}
|
|
auto tuple = std::tuple<int, std::string>(2, "hi");
|
|
struct TestItem {
|
|
IValue value;
|
|
TypePtr expected_type;
|
|
};
|
|
std::vector<TestItem> items = {
|
|
{list, ListType::create(ListType::create(IntType::get()))},
|
|
{2, IntType::get()},
|
|
{dict, DictType::create(StringType::get(), TensorType::get())},
|
|
{dict_list,
|
|
ListType::create(
|
|
DictType::create(StringType::get(), TensorType::get()))},
|
|
{tuple, TupleType::create({IntType::get(), StringType::get()})}};
|
|
// NOLINTNEXTLINE(performance-for-range-copy)
|
|
for (auto item : items) {
|
|
auto bytes = torch::pickle_save(item.value);
|
|
auto loaded = torch::pickle_load(bytes);
|
|
ASSERT_TRUE(loaded.type()->isSubtypeOf(*item.expected_type));
|
|
ASSERT_TRUE(item.expected_type->isSubtypeOf(*loaded.type()));
|
|
}
|
|
}
|
|
|
|
TEST(SerializationTest, TestJitStream_CUDA) {
|
|
torch::jit::Module model;
|
|
std::vector<torch::jit::IValue> inputs;
|
|
// Deserialize the ScriptModule from a file using torch::jit::load().
|
|
// Load the scripted model. This should have been generated by tests_setup.py
|
|
// Refer: TorchSaveJitStream_CUDA in test/cpp/jit/tests_setup.py
|
|
model = torch::jit::load("saved_stream_model.pt");
|
|
|
|
auto output = model.forward(inputs);
|
|
const auto& list_of_elements = output.toTupleRef().elements();
|
|
auto is_stream_s = list_of_elements[0].toBool();
|
|
|
|
// a,b: These are the two input tensors
|
|
// c: This is output tensor generated by the operation torch.cat(a,b)
|
|
auto a = list_of_elements[1].toTensor();
|
|
auto b = list_of_elements[2].toTensor();
|
|
auto c = list_of_elements[3].toTensor();
|
|
// op: this is used to verify if the cat operation produced the same results
|
|
// as that on the GPU with torch.cat
|
|
auto op = at::cat({a, b}, 0);
|
|
|
|
// Check if the stream is set
|
|
ASSERT_TRUE(is_stream_s);
|
|
// Check if the sizes of the outputs (op and c) is same on the GPU and CPU
|
|
ASSERT_EQ(op.sizes(), c.sizes());
|
|
// Check if both the output tensors are equal
|
|
ASSERT_TRUE(op.equal(c));
|
|
}
|
|
|
|
TEST(TestSourceRoundTrip, UpsampleNearest2d) {
|
|
Module m("m");
|
|
m.define(R"(
|
|
def forward(self, input: Tensor, scale:float):
|
|
return torch.upsample_nearest2d(input, [1, 1], float(scale), float(scale))
|
|
)");
|
|
|
|
std::vector<IValue> inputs;
|
|
inputs.emplace_back(torch::rand({1, 3, 128, 128}));
|
|
inputs.emplace_back(at::Scalar(2.0));
|
|
auto ref = m.forward(inputs);
|
|
|
|
Module m2 = roundtripThroughMobile(m);
|
|
auto res = m2.forward(inputs);
|
|
|
|
auto resd = res.toTensor();
|
|
auto refd = ref.toTensor();
|
|
ASSERT_TRUE(resd.equal(refd));
|
|
}
|
|
|
|
TEST(TestSourceRoundTrip, CheckAttrAccess) {
|
|
Module m("m");
|
|
m.register_attribute("mobile_optimized", BoolType::get(), true);
|
|
Module m2 = roundtripThroughMobile(m);
|
|
bool mobile_optimized = m2.attr("mobile_optimized", false).toBool();
|
|
AT_ASSERT(mobile_optimized);
|
|
}
|
|
|
|
TEST(TestSourceRoundTrip,
|
|
MethodInvocation) { // NOLINT (use =delete in gtest)
|
|
const std::vector<std::string> test_programs{
|
|
// test invoking a method with default parameter
|
|
R"(
|
|
def test_func(self, x, b : int = 4):
|
|
return self.foo + x + b
|
|
)",
|
|
// inner method call with default parameter (gets inlined)
|
|
R"(
|
|
def add_with_default_arg(self, x, b : int = 4):
|
|
return self.foo + x + b
|
|
def test_func(self, x):
|
|
return self.add_with_default_arg(x) # invoke method w/ default arg
|
|
)",
|
|
// simple method call
|
|
R"(
|
|
def test_func(self, x):
|
|
b = 4
|
|
return self.foo + x + b
|
|
)",
|
|
};
|
|
for (const auto& test_program : test_programs) {
|
|
Module m("m");
|
|
m.register_parameter("foo", torch::ones({}), false);
|
|
m.define(test_program);
|
|
|
|
const int fortyTwo = 42; // (keep linter happy)
|
|
auto minput = fortyTwo * torch::ones({});
|
|
auto ref = m.run_method("test_func", minput);
|
|
|
|
Module m2 = roundtripThroughMobile(m);
|
|
const auto& test_func = m2.get_method("test_func");
|
|
IValue res;
|
|
for (int i = 0; i < 3; ++i) {
|
|
res = test_func({minput});
|
|
}
|
|
|
|
auto resd = res.toTensor().item<float>();
|
|
auto refd = ref.toTensor().item<float>();
|
|
AT_ASSERT(resd == refd);
|
|
}
|
|
}
|
|
|
|
TEST(SerializationTest, ParentDirNotExist) {
|
|
expectThrowsEq(
|
|
[]() {
|
|
auto t = torch::nn::Linear(5, 5);
|
|
torch::save(t, "./doesnotexist/file.pt");
|
|
},
|
|
"Parent directory ./doesnotexist does not exist.");
|
|
}
|
|
|
|
} // namespace jit
|
|
} // namespace torch
|