| // This file is part of Eigen, a lightweight C++ template library |
| // for linear algebra. |
| // |
| // Copyright (C) 2006-2008 Benoit Jacob <jacob.benoit.1@gmail.com> |
| // |
| // This Source Code Form is subject to the terms of the Mozilla |
| // Public License v. 2.0. If a copy of the MPL was not distributed |
| // with this file, You can obtain one at http://mozilla.org/MPL/2.0/. |
| |
| #include "product.h" |
| #include <Eigen/LU> |
| |
| template <typename T> |
| void test_aliasing() { |
| int rows = internal::random<int>(1, 12); |
| int cols = internal::random<int>(1, 12); |
| typedef Matrix<T, Dynamic, Dynamic> MatrixType; |
| typedef Matrix<T, Dynamic, 1> VectorType; |
| VectorType x(cols); |
| x.setRandom(); |
| VectorType z(x); |
| VectorType y(rows); |
| y.setZero(); |
| MatrixType A(rows, cols); |
| A.setRandom(); |
| // CwiseBinaryOp |
| VERIFY_IS_APPROX(x = y + A * x, A * z); // OK because "y + A*x" is marked as "assume-aliasing" |
| x = z; |
| // CwiseUnaryOp |
| VERIFY_IS_APPROX(x = T(1.) * (A * x), |
| A * z); // OK because 1*(A*x) is replaced by (1*A*x) which is a Product<> expression |
| x = z; |
| // VERIFY_IS_APPROX(x = y-A*x, -A*z); // Not OK in 3.3 because x is resized before A*x gets evaluated |
| x = z; |
| } |
| |
| template <int> |
| void product_large_regressions() { |
| { |
| // test a specific issue in DiagonalProduct |
| int N = 1000000; |
| VectorXf v = VectorXf::Ones(N); |
| MatrixXf m = MatrixXf::Ones(N, 3); |
| m = (v + v).asDiagonal() * m; |
| VERIFY_IS_APPROX(m, MatrixXf::Constant(N, 3, 2)); |
| } |
| |
| { |
| // test deferred resizing in Matrix::operator= |
| MatrixXf a = MatrixXf::Random(10, 4), b = MatrixXf::Random(4, 10), c = a; |
| VERIFY_IS_APPROX((a = a * b), (c * b).eval()); |
| } |
| |
| { |
| // check the functions to setup blocking sizes compile and do not segfault |
| // FIXME check they do what they are supposed to do !! |
| std::ptrdiff_t l1 = internal::random<int>(10000, 20000); |
| std::ptrdiff_t l2 = internal::random<int>(100000, 200000); |
| std::ptrdiff_t l3 = internal::random<int>(1000000, 2000000); |
| setCpuCacheSizes(l1, l2, l3); |
| VERIFY(l1 == l1CacheSize()); |
| VERIFY(l2 == l2CacheSize()); |
| std::ptrdiff_t k1 = internal::random<int>(10, 100) * 16; |
| std::ptrdiff_t m1 = internal::random<int>(10, 100) * 16; |
| std::ptrdiff_t n1 = internal::random<int>(10, 100) * 16; |
| // only makes sure it compiles fine |
| internal::computeProductBlockingSizes<float, float, std::ptrdiff_t>(k1, m1, n1, 1); |
| } |
| |
| { |
| // test regression in row-vector by matrix (bad Map type) |
| MatrixXf mat1(10, 32); |
| mat1.setRandom(); |
| MatrixXf mat2(32, 32); |
| mat2.setRandom(); |
| MatrixXf r1 = mat1.row(2) * mat2.transpose(); |
| VERIFY_IS_APPROX(r1, (mat1.row(2) * mat2.transpose()).eval()); |
| |
| MatrixXf r2 = mat1.row(2) * mat2; |
| VERIFY_IS_APPROX(r2, (mat1.row(2) * mat2).eval()); |
| } |
| |
| { |
| Eigen::MatrixXd A(10, 10), B, C; |
| A.setRandom(); |
| C = A; |
| for (int k = 0; k < 79; ++k) C = C * A; |
| B.noalias() = |
| (((A * A) * (A * A)) * ((A * A) * (A * A)) * ((A * A) * (A * A)) * ((A * A) * (A * A)) * ((A * A) * (A * A)) * |
| ((A * A) * (A * A)) * ((A * A) * (A * A)) * ((A * A) * (A * A)) * ((A * A) * (A * A)) * ((A * A) * (A * A))) * |
| (((A * A) * (A * A)) * ((A * A) * (A * A)) * ((A * A) * (A * A)) * ((A * A) * (A * A)) * ((A * A) * (A * A)) * |
| ((A * A) * (A * A)) * ((A * A) * (A * A)) * ((A * A) * (A * A)) * ((A * A) * (A * A)) * ((A * A) * (A * A))); |
| VERIFY_IS_APPROX(B, C); |
| } |
| } |
| |
| // Regression test: row-major GEMV with stride*sizeof > 32000 disables the |
| // 8-row main loop (n8=0). The cleanup must use `for` loops (not `if`) to |
| // process all remaining rows. Without the fix, only 7 out of `rows` results |
| // are computed. This manifests as loss of orthogonality in QR of tall-skinny |
| // matrices, since the Householder application uses row-major GEMV internally. |
| template <int> |
| void bug_gemv_rowmajor_large_stride() { |
| // Direct GEMV test: row-major A with stride (= cols) triggering n8=0. |
| // The threshold is stride * sizeof(Scalar) > 32000. |
| // For double: cols > 4000. For float: cols > 8000. |
| { |
| const int rows = 100; |
| const int cols = 5000; // cols * sizeof(double) = 40000 > 32000 |
| Matrix<double, Dynamic, Dynamic, RowMajor> A(rows, cols); |
| A.setRandom(); |
| VectorXd x = VectorXd::Random(cols); |
| VectorXd y = A * x; |
| VectorXd y_ref = VectorXd::Zero(rows); |
| for (int i = 0; i < rows; ++i) |
| for (int j = 0; j < cols; ++j) y_ref(i) += A(i, j) * x(j); |
| VERIFY_IS_APPROX(y, y_ref); |
| } |
| |
| // QR orthogonality test: this is the high-level symptom. |
| // HouseholderQR of a col-major (m x n) matrix with m > 4000 |
| // uses row-major GEMV internally during Householder application. |
| { |
| const int m = 5000; |
| const int n = 50; |
| MatrixXd A = MatrixXd::Random(m, n); |
| MatrixXd Q = A.householderQr().householderQ() * MatrixXd::Identity(m, n); |
| MatrixXd QtQ = Q.adjoint() * Q; |
| VERIFY_IS_APPROX(QtQ, MatrixXd::Identity(n, n)); |
| } |
| } |
| |
| // Regression test for row-major GEMV run_small_cols bug. |
| // When cols is small (e.g., 2), and loop variables (like n8) are 0 due |
| // to row or stride limits, the remainder loops previously used `if` checks |
| // like `if (i < n4)`. This incorrectly skips rows if multiple remainder |
| // blocks are needed (e.g., 9 rows). |
| template <int> |
| void bug_gemv_run_small_cols() { |
| const int rows = 9; // > 8, covers 8-row loop step but tests remainder cleanup |
| const int cols = 2; // triggers run_small_cols (cols < PacketSize) |
| const int stride = 5000; // 5000 * sizeof(double) > 32000, forces n8 = 0 |
| |
| Matrix<double, Dynamic, Dynamic, RowMajor> A_full(rows, stride); |
| A_full.setRandom(); |
| auto A = A_full.leftCols(cols); |
| |
| VectorXd x = VectorXd::Random(cols); |
| VectorXd y = A * x; |
| VectorXd y_ref = A.eval() * x; // No stride. |
| |
| VERIFY_IS_APPROX(y, y_ref); |
| } |
| |
| // Systematic test of row-major GEMV run_small_cols and main run() remainder paths. |
| // Varies cols from 1-7 (covers float PacketSize=8 and double PacketSize=4 boundaries) |
| // and rows across values that exercise all n8/n4/n2/n1 remainder combinations. |
| template <int> |
| void gemv_small_cols_systematic() { |
| const int test_cols[] = {1, 2, 3, 4, 5, 6, 7}; |
| const int test_rows[] = {1, 2, 3, 4, 5, 7, 8, 9, 11, 13, 15, 16, 17, 25}; |
| |
| // Large stride forces n8=0, exercising all remainder-only paths. |
| { |
| const int stride = 5000; // 5000 * sizeof(double) = 40000 > 32000 |
| for (int ci = 0; ci < 7; ++ci) { |
| for (int ri = 0; ri < 14; ++ri) { |
| int rows = test_rows[ri], cols = test_cols[ci]; |
| Matrix<double, Dynamic, Dynamic, RowMajor> A_full(rows, stride); |
| A_full.setRandom(); |
| auto A = A_full.leftCols(cols); |
| VectorXd x = VectorXd::Random(cols); |
| VectorXd y = A * x; |
| VectorXd y_ref = VectorXd::Zero(rows); |
| for (int i = 0; i < rows; ++i) |
| for (int j = 0; j < cols; ++j) y_ref(i) += A(i, j) * x(j); |
| VERIFY_IS_APPROX(y, y_ref); |
| } |
| } |
| } |
| |
| // Normal stride (n8 active) to cover the 8-row main loop + remainders. |
| for (int ci = 0; ci < 7; ++ci) { |
| for (int ri = 0; ri < 14; ++ri) { |
| int rows = test_rows[ri], cols = test_cols[ci]; |
| Matrix<double, Dynamic, Dynamic, RowMajor> A(rows, cols); |
| A.setRandom(); |
| VectorXd x = VectorXd::Random(cols); |
| VectorXd y = A * x; |
| VectorXd y_ref = VectorXd::Zero(rows); |
| for (int i = 0; i < rows; ++i) |
| for (int j = 0; j < cols; ++j) y_ref(i) += A(i, j) * x(j); |
| VERIFY_IS_APPROX(y, y_ref); |
| } |
| } |
| |
| // Float with large stride: 9000 * sizeof(float) = 36000 > 32000 |
| { |
| const int stride = 9000; |
| for (int ci = 0; ci < 7; ++ci) { |
| for (int ri = 0; ri < 14; ++ri) { |
| int rows = test_rows[ri], cols = test_cols[ci]; |
| Matrix<float, Dynamic, Dynamic, RowMajor> A_full(rows, stride); |
| A_full.setRandom(); |
| auto A = A_full.leftCols(cols); |
| VectorXf x = VectorXf::Random(cols); |
| VectorXf y = A * x; |
| VectorXf y_ref = VectorXf::Zero(rows); |
| for (int i = 0; i < rows; ++i) |
| for (int j = 0; j < cols; ++j) y_ref(i) += A(i, j) * x(j); |
| VERIFY_IS_APPROX(y, y_ref); |
| } |
| } |
| } |
| } |
| |
| // Test the main row-major GEMV n8=0 path (not run_small_cols) with varied row counts. |
| // The n8 threshold is stride*sizeof(Scalar) > 32000. |
| template <int> |
| void gemv_rowmajor_large_stride_varied_rows() { |
| const int test_rows[] = {1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 15, 16, 17, 25, 100}; |
| // Double: cols=5000 (5000*8 > 32000), enough cols to stay on main run() path. |
| { |
| const int cols = 5000; |
| for (int ri = 0; ri < 16; ++ri) { |
| int rows = test_rows[ri]; |
| Matrix<double, Dynamic, Dynamic, RowMajor> A(rows, cols); |
| A.setRandom(); |
| VectorXd x = VectorXd::Random(cols); |
| VectorXd y = A * x; |
| VectorXd y_ref = VectorXd::Zero(rows); |
| for (int i = 0; i < rows; ++i) |
| for (int j = 0; j < cols; ++j) y_ref(i) += A(i, j) * x(j); |
| VERIFY_IS_APPROX(y, y_ref); |
| } |
| } |
| // Float: cols=9000 (9000*4 > 32000). |
| { |
| const int cols = 9000; |
| for (int ri = 0; ri < 16; ++ri) { |
| int rows = test_rows[ri]; |
| Matrix<float, Dynamic, Dynamic, RowMajor> A(rows, cols); |
| A.setRandom(); |
| VectorXf x = VectorXf::Random(cols); |
| VectorXf y = A * x; |
| VectorXf y_ref = VectorXf::Zero(rows); |
| for (int i = 0; i < rows; ++i) |
| for (int j = 0; j < cols; ++j) y_ref(i) += A(i, j) * x(j); |
| VERIFY_IS_APPROX(y, y_ref); |
| } |
| } |
| } |
| |
| // Test extreme aspect ratios that exercise GEMV, outer-product, and thin-GEMM dispatch. |
| template <int> |
| void product_extreme_aspect_ratios() { |
| const int sizes[] = {1, 2, 3, 4, 8, 16, 48, 64, 128}; |
| for (int si = 0; si < 9; ++si) { |
| int s = sizes[si]; |
| for (int ki = 0; ki < 9; ++ki) { |
| int k = sizes[ki]; |
| // Thin result: s x k * k x 2 (2-column GEMM) |
| { |
| MatrixXd A = MatrixXd::Random(s, k); |
| MatrixXd B = MatrixXd::Random(k, 2); |
| MatrixXd C = A * B; |
| MatrixXd Cref = MatrixXd::Zero(s, 2); |
| for (int i = 0; i < s; ++i) |
| for (int j = 0; j < 2; ++j) |
| for (int kk = 0; kk < k; ++kk) Cref(i, j) += A(i, kk) * B(kk, j); |
| VERIFY_IS_APPROX(C, Cref); |
| } |
| // Wide result: 2 x k * k x s (2-row GEMM) |
| { |
| MatrixXd A = MatrixXd::Random(2, k); |
| MatrixXd B = MatrixXd::Random(k, s); |
| MatrixXd C = A * B; |
| MatrixXd Cref = MatrixXd::Zero(2, s); |
| for (int i = 0; i < 2; ++i) |
| for (int j = 0; j < s; ++j) |
| for (int kk = 0; kk < k; ++kk) Cref(i, j) += A(i, kk) * B(kk, j); |
| VERIFY_IS_APPROX(C, Cref); |
| } |
| // GEMV: s x k * k x 1 |
| { |
| MatrixXd A = MatrixXd::Random(s, k); |
| VectorXd x = VectorXd::Random(k); |
| VectorXd y = A * x; |
| VectorXd yref = VectorXd::Zero(s); |
| for (int i = 0; i < s; ++i) |
| for (int kk = 0; kk < k; ++kk) yref(i) += A(i, kk) * x(kk); |
| VERIFY_IS_APPROX(y, yref); |
| } |
| // Vec-mat: 1 x k * k x s |
| { |
| RowVectorXd v = RowVectorXd::Random(k); |
| MatrixXd B = MatrixXd::Random(k, s); |
| RowVectorXd r = v * B; |
| RowVectorXd rref = RowVectorXd::Zero(s); |
| for (int j = 0; j < s; ++j) |
| for (int kk = 0; kk < k; ++kk) rref(j) += v(kk) * B(kk, j); |
| VERIFY_IS_APPROX(r, rref); |
| } |
| } |
| } |
| } |
| |
| template <int> |
| void bug_1622() { |
| typedef Matrix<double, 2, -1, 0, 2, -1> Mat2X; |
| Mat2X x(2, 2); |
| x.setRandom(); |
| MatrixXd y(2, 2); |
| y.setRandom(); |
| const Mat2X K1 = x * y.inverse(); |
| const Matrix2d K2 = x * y.inverse(); |
| VERIFY_IS_APPROX(K1, K2); |
| } |
| |
| EIGEN_DECLARE_TEST(product_large) { |
| for (int i = 0; i < g_repeat; i++) { |
| CALL_SUBTEST_1(product( |
| MatrixXf(internal::random<int>(1, EIGEN_TEST_MAX_SIZE), internal::random<int>(1, EIGEN_TEST_MAX_SIZE)))); |
| CALL_SUBTEST_2(product( |
| MatrixXd(internal::random<int>(1, EIGEN_TEST_MAX_SIZE), internal::random<int>(1, EIGEN_TEST_MAX_SIZE)))); |
| CALL_SUBTEST_2(product(MatrixXd(internal::random<int>(1, 10), internal::random<int>(1, 10)))); |
| |
| CALL_SUBTEST_3(product( |
| MatrixXi(internal::random<int>(1, EIGEN_TEST_MAX_SIZE), internal::random<int>(1, EIGEN_TEST_MAX_SIZE)))); |
| CALL_SUBTEST_4(product(MatrixXcf(internal::random<int>(1, EIGEN_TEST_MAX_SIZE / 2), |
| internal::random<int>(1, EIGEN_TEST_MAX_SIZE / 2)))); |
| CALL_SUBTEST_5(product(Matrix<float, Dynamic, Dynamic, RowMajor>(internal::random<int>(1, EIGEN_TEST_MAX_SIZE), |
| internal::random<int>(1, EIGEN_TEST_MAX_SIZE)))); |
| |
| CALL_SUBTEST_1(test_aliasing<float>()); |
| |
| CALL_SUBTEST_6(bug_1622<1>()); |
| |
| CALL_SUBTEST_7(product(MatrixXcd(internal::random<int>(1, EIGEN_TEST_MAX_SIZE / 2), |
| internal::random<int>(1, EIGEN_TEST_MAX_SIZE / 2)))); |
| CALL_SUBTEST_8(product(Matrix<double, Dynamic, Dynamic, RowMajor>(internal::random<int>(1, EIGEN_TEST_MAX_SIZE), |
| internal::random<int>(1, EIGEN_TEST_MAX_SIZE)))); |
| CALL_SUBTEST_9(product(Matrix<std::complex<float>, Dynamic, Dynamic, RowMajor>( |
| internal::random<int>(1, EIGEN_TEST_MAX_SIZE), internal::random<int>(1, EIGEN_TEST_MAX_SIZE)))); |
| CALL_SUBTEST_10(product(Matrix<std::complex<double>, Dynamic, Dynamic, RowMajor>( |
| internal::random<int>(1, EIGEN_TEST_MAX_SIZE), internal::random<int>(1, EIGEN_TEST_MAX_SIZE)))); |
| CALL_SUBTEST_11(product(Matrix<bfloat16, Dynamic, Dynamic, RowMajor>( |
| internal::random<int>(1, EIGEN_TEST_MAX_SIZE), internal::random<int>(1, EIGEN_TEST_MAX_SIZE)))); |
| CALL_SUBTEST_12(product(Matrix<Eigen::half, Dynamic, Dynamic>(internal::random<int>(1, EIGEN_TEST_MAX_SIZE), |
| internal::random<int>(1, EIGEN_TEST_MAX_SIZE)))); |
| } |
| |
| CALL_SUBTEST_6(product_large_regressions<0>()); |
| CALL_SUBTEST_6(bug_gemv_rowmajor_large_stride<0>()); |
| CALL_SUBTEST_6(bug_gemv_run_small_cols<0>()); |
| CALL_SUBTEST_6(gemv_small_cols_systematic<0>()); |
| CALL_SUBTEST_6(gemv_rowmajor_large_stride_varied_rows<0>()); |
| CALL_SUBTEST_6(product_extreme_aspect_ratios<0>()); |
| |
| // Regression test for bug 714: |
| #if defined EIGEN_HAS_OPENMP |
| omp_set_dynamic(1); |
| for (int i = 0; i < g_repeat; i++) { |
| CALL_SUBTEST_6(product(Matrix<float, Dynamic, Dynamic>(internal::random<int>(1, EIGEN_TEST_MAX_SIZE), |
| internal::random<int>(1, EIGEN_TEST_MAX_SIZE)))); |
| } |
| #endif |
| } |