|
17 | 17 | #include <executorch/runtime/kernel/operator_registry.h>
|
18 | 18 | #include <executorch/runtime/platform/runtime.h>
|
19 | 19 | #include <executorch/test/utils/DeathTest.h>
|
| 20 | +#include <cstdint> |
| 21 | +#include <cstdio> |
20 | 22 |
|
21 | 23 | using exec_aten::SizesType;
|
22 | 24 | using torch::executor::Error;
|
@@ -272,5 +274,180 @@ TEST_F(RegisterPrimOpsTest, LocalScalarDenseReturnsCorrectValue) {
|
272 | 274 | EXPECT_EQ(stack[1]->toInt(), expected);
|
273 | 275 | }
|
274 | 276 |
|
| 277 | +TEST_F(RegisterPrimOpsTest, TestETView) { |
| 278 | + EXPECT_TRUE(hasOpsFn("executorch_prim::et_view.default")); |
| 279 | + |
| 280 | + testing::TensorFactory<ScalarType::Int> tf; |
| 281 | + |
| 282 | + // *************************************************************************** |
| 283 | + // Make self for tests |
| 284 | + // *************************************************************************** |
| 285 | + auto self = tf.make({3, 2}, {1, 2, 3, 4, 5, 6}); |
| 286 | + auto self_evalue = EValue(self); |
| 287 | + |
| 288 | + // *************************************************************************** |
| 289 | + // Make size for tests |
| 290 | + // *************************************************************************** |
| 291 | + int64_t size[3] = {1, 3, -1}; |
| 292 | + EValue size_as_evals[3] = {EValue(size[0]), EValue(size[1]), EValue(size[2])}; |
| 293 | + EValue* size_wrapped_vals[3] = { |
| 294 | + &size_as_evals[0], &size_as_evals[1], &size_as_evals[2]}; |
| 295 | + int64_t size_unwrapped_vals[3] = {0, 0, 0}; |
| 296 | + EValue size_int_list_evalue = EValue( |
| 297 | + BoxedEvalueList<int64_t>(size_wrapped_vals, size_unwrapped_vals, 3)); |
| 298 | + |
| 299 | + int64_t bad_size1[3] = {-1, 3, -1}; // two inferred dimensions |
| 300 | + EValue bad_size_as_evals1[3] = { |
| 301 | + EValue(bad_size1[0]), EValue(bad_size1[1]), EValue(bad_size1[2])}; |
| 302 | + EValue* bad_size_wrapped_vals1[3] = { |
| 303 | + &bad_size_as_evals1[0], &bad_size_as_evals1[1], &bad_size_as_evals1[2]}; |
| 304 | + int64_t bad_size_unwrapped_vals1[3] = {0, 0, 0}; |
| 305 | + EValue bad_size_int_list_evalue1 = EValue(BoxedEvalueList<int64_t>( |
| 306 | + bad_size_wrapped_vals1, bad_size_unwrapped_vals1, 3)); |
| 307 | + |
| 308 | + int64_t bad_size2[3] = {-2, -3, 1}; // negative size not supported |
| 309 | + EValue bad_size_as_evals2[3] = { |
| 310 | + EValue(bad_size2[0]), EValue(bad_size2[1]), EValue(bad_size2[2])}; |
| 311 | + EValue* bad_size_wrapped_vals2[3] = { |
| 312 | + &bad_size_as_evals2[0], &bad_size_as_evals2[1], &bad_size_as_evals2[2]}; |
| 313 | + int64_t bad_size_unwrapped_vals2[3] = {0, 0, 0}; |
| 314 | + EValue bad_size_int_list_evalue2 = EValue(BoxedEvalueList<int64_t>( |
| 315 | + bad_size_wrapped_vals2, bad_size_unwrapped_vals2, 3)); |
| 316 | + |
| 317 | + // *************************************************************************** |
| 318 | + // Make outs for tests |
| 319 | + // *************************************************************************** |
| 320 | + constexpr int N_GOOD_OUTS = 2; |
| 321 | + Tensor good_outs[N_GOOD_OUTS] = { |
| 322 | + tf.ones({1, 3, 2}), // correct size with nullptr |
| 323 | + tf.ones({1, 3, 2}), // correct size with self data_ptr |
| 324 | + }; |
| 325 | + internal::reset_data_ptr(good_outs[0]); |
| 326 | + ET_CHECK( |
| 327 | + internal::set_tensor_data( |
| 328 | + good_outs[1], self.mutable_data_ptr(), good_outs[1].nbytes()) == |
| 329 | + Error::Ok); |
| 330 | + EValue good_out_evalues[N_GOOD_OUTS] = { |
| 331 | + EValue(good_outs[0]), EValue(good_outs[1])}; |
| 332 | + |
| 333 | + // bad outs expect death |
| 334 | + constexpr int N_BAD_OUTS = 3; |
| 335 | + Tensor bad_outs[N_BAD_OUTS] = { |
| 336 | + tf.ones({1, 3, 2, 1}), // wrong rank |
| 337 | + tf.ones({1, 3, 3}), // wrong size |
| 338 | + tf.ones({1, 3, 2}) // occupied data_ptr |
| 339 | + }; |
| 340 | + EValue bad_out_evalues[N_BAD_OUTS] = { |
| 341 | + EValue(bad_outs[0]), EValue(bad_outs[1]), EValue(bad_outs[2])}; |
| 342 | + |
| 343 | + // *************************************************************************** |
| 344 | + // Run tests |
| 345 | + // *************************************************************************** |
| 346 | + |
| 347 | + constexpr int N_BAD_STACKS = N_BAD_OUTS + 2; |
| 348 | + EValue* bad_stacks[N_BAD_STACKS][3] = { |
| 349 | + // Bad out stacks |
| 350 | + {&self_evalue, &size_int_list_evalue, &bad_out_evalues[0]}, |
| 351 | + {&self_evalue, &size_int_list_evalue, &bad_out_evalues[1]}, |
| 352 | + {&self_evalue, &size_int_list_evalue, &bad_out_evalues[2]}, |
| 353 | + // Bad size stacks |
| 354 | + {&self_evalue, &bad_size_int_list_evalue1, &good_out_evalues[0]}, |
| 355 | + {&self_evalue, &bad_size_int_list_evalue2, &good_out_evalues[0]}}; |
| 356 | + |
| 357 | + // Bad stacks expect death |
| 358 | + for (int i = 0; i < N_BAD_STACKS; i++) { |
| 359 | + ET_EXPECT_DEATH( |
| 360 | + getOpsFn("executorch_prim::et_view.default")(context, bad_stacks[i]), |
| 361 | + ""); |
| 362 | + } |
| 363 | + |
| 364 | + constexpr int N_GOOD_STACKS = N_GOOD_OUTS; |
| 365 | + EValue* good_out_stacks[N_GOOD_STACKS][3] = { |
| 366 | + {&self_evalue, &size_int_list_evalue, &good_out_evalues[0]}, |
| 367 | + {&self_evalue, &size_int_list_evalue, &good_out_evalues[1]}}; |
| 368 | + |
| 369 | + // Good outs expect no death and correct output |
| 370 | + for (int i = 0; i < N_GOOD_STACKS; i++) { |
| 371 | + getOpsFn("executorch_prim::et_view.default")(context, good_out_stacks[i]); |
| 372 | + EXPECT_TENSOR_EQ(good_outs[i], tf.make({1, 3, 2}, {1, 2, 3, 4, 5, 6})); |
| 373 | + EXPECT_EQ(good_outs[i].const_data_ptr(), self.const_data_ptr()); |
| 374 | + } |
| 375 | +} |
| 376 | + |
| 377 | +TEST_F(RegisterPrimOpsTest, TestETViewDynamic) { |
| 378 | + testing::TensorFactory<ScalarType::Int> tf; |
| 379 | + |
| 380 | + auto self = tf.make({3, 1}, {1, 2, 3}); |
| 381 | + auto self_evalue = EValue(self); |
| 382 | + |
| 383 | + int64_t size[3] = {1, 3, -1}; // inferred size should be {1, 3, 1} |
| 384 | + // Construct the size as an EValue int_list |
| 385 | + EValue size_as_evals[3] = {EValue(size[0]), EValue(size[1]), EValue(size[2])}; |
| 386 | + EValue* size_wrapped_vals[3] = { |
| 387 | + &size_as_evals[0], &size_as_evals[1], &size_as_evals[2]}; |
| 388 | + int64_t size_unwrapped_vals[3] = {0, 0, 0}; |
| 389 | + EValue size_int_list_evalue = EValue( |
| 390 | + BoxedEvalueList<int64_t>(size_wrapped_vals, size_unwrapped_vals, 3)); |
| 391 | + |
| 392 | +#ifdef USE_ATEN_LIB |
| 393 | + // ATen mode tensors don't need dynamism specification. |
| 394 | + auto out = tf.make({3, 2, 1}, {0, 0, 0, 0, 0, 0}); |
| 395 | +#else |
| 396 | + auto out = tf.make( |
| 397 | + {3, 2, 1}, {0, 0, 0, 0, 0, 0}, {}, TensorShapeDynamism::DYNAMIC_BOUND); |
| 398 | +#endif |
| 399 | + |
| 400 | + internal::reset_data_ptr(out); |
| 401 | + EValue out_evalue = EValue(out); |
| 402 | + |
| 403 | + EValue* stack[3] = {&self_evalue, &size_int_list_evalue, &out_evalue}; |
| 404 | + |
| 405 | + getOpsFn("executorch_prim::et_view.default")(context, stack); |
| 406 | + |
| 407 | + EXPECT_TENSOR_EQ(out, tf.make({1, 3, 1}, {1, 2, 3})); |
| 408 | + EXPECT_EQ(out.const_data_ptr(), self.const_data_ptr()); |
| 409 | +} |
| 410 | + |
| 411 | +TEST_F(RegisterPrimOpsTest, TestETViewEmpty) { |
| 412 | + testing::TensorFactory<ScalarType::Int> tf; |
| 413 | + |
| 414 | + auto self = tf.make({3, 1, 0}, {}); |
| 415 | + auto self_evalue = EValue(self); |
| 416 | + EXPECT_EQ(self.const_data_ptr(), nullptr); // empty tensor has null data |
| 417 | + |
| 418 | + // Construct the sizes |
| 419 | + int64_t size[3] = {3, 1, -1}; |
| 420 | + EValue size_as_evals[3] = {EValue(size[0]), EValue(size[1]), EValue(size[2])}; |
| 421 | + EValue* size_wrapped_vals[3] = { |
| 422 | + &size_as_evals[0], &size_as_evals[1], &size_as_evals[2]}; |
| 423 | + int64_t size_unwrapped_vals[3] = {0, 0, 0}; |
| 424 | + EValue size_int_list_evalue = EValue( |
| 425 | + BoxedEvalueList<int64_t>(size_wrapped_vals, size_unwrapped_vals, 3)); |
| 426 | + |
| 427 | + int64_t bad_size[3] = {0, 1, -1}; // bad size: cannot infer with 0 |
| 428 | + EValue bad_size_as_evals[3] = { |
| 429 | + EValue(bad_size[0]), EValue(bad_size[1]), EValue(bad_size[2])}; |
| 430 | + EValue* bad_size_wrapped_vals[3] = { |
| 431 | + &bad_size_as_evals[0], &bad_size_as_evals[1], &bad_size_as_evals[2]}; |
| 432 | + int64_t bad_size_unwrapped_vals[3] = {0, 0, 0}; |
| 433 | + EValue bad_size_int_list_evalue = EValue(BoxedEvalueList<int64_t>( |
| 434 | + bad_size_wrapped_vals, bad_size_unwrapped_vals, 3)); |
| 435 | + |
| 436 | + auto out = tf.make({3, 1, 0}, {}, {}); |
| 437 | + EValue out_evalue = EValue(out); |
| 438 | + EXPECT_EQ(out.const_data_ptr(), nullptr); |
| 439 | + |
| 440 | + // good size test |
| 441 | + EValue* stack[3] = {&self_evalue, &size_int_list_evalue, &out_evalue}; |
| 442 | + getOpsFn("executorch_prim::et_view.default")(context, stack); |
| 443 | + EXPECT_TENSOR_EQ(out, tf.make({3, 1, 0}, {})); |
| 444 | + EXPECT_EQ(out.const_data_ptr(), self.const_data_ptr()); |
| 445 | + |
| 446 | + // bad size test |
| 447 | + EValue* bad_stack[3] = {&self_evalue, &bad_size_int_list_evalue, &out_evalue}; |
| 448 | + ET_EXPECT_DEATH( |
| 449 | + getOpsFn("executorch_prim::et_view.default")(context, bad_stack), ""); |
| 450 | +} |
| 451 | + |
275 | 452 | } // namespace executor
|
276 | 453 | } // namespace torch
|
0 commit comments