diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorExecutor.h b/unsupported/Eigen/CXX11/src/Tensor/TensorExecutor.h index 6bbf235cc..f71625ae5 100644 --- a/unsupported/Eigen/CXX11/src/Tensor/TensorExecutor.h +++ b/unsupported/Eigen/CXX11/src/Tensor/TensorExecutor.h @@ -125,7 +125,7 @@ class TensorExecutor int blocksz = std::ceil(static_cast(size)/device.numThreads()) + PacketSize - 1; const Index blocksize = numext::maxi(PacketSize, (blocksz - (blocksz % PacketSize))); - const Index numblocks = size / blocksize; + const unsigned int numblocks = static_cast(size / blocksize); Barrier barrier(numblocks); for (int i = 0; i < numblocks; ++i) { diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorReduction.h b/unsupported/Eigen/CXX11/src/Tensor/TensorReduction.h index 489451215..0ce2517d6 100644 --- a/unsupported/Eigen/CXX11/src/Tensor/TensorReduction.h +++ b/unsupported/Eigen/CXX11/src/Tensor/TensorReduction.h @@ -253,7 +253,7 @@ struct FullReducer { return; } else { const Index blocksize = std::floor(static_cast(num_coeffs) / num_threads); - const Index numblocks = blocksize > 0 ? num_coeffs / blocksize : 0; + const unsigned int numblocks = blocksize > 0 ? static_cast(num_coeffs / blocksize) : 0; eigen_assert(num_coeffs >= numblocks * blocksize); Barrier barrier(numblocks); @@ -300,7 +300,7 @@ struct FullReducer { return; } const Index blocksize = std::floor(static_cast(num_coeffs) / num_threads); - const Index numblocks = blocksize > 0 ? num_coeffs / blocksize : 0; + const unsigned int numblocks = blocksize > 0 ? static_cast(num_coeffs / blocksize) : 0; eigen_assert(num_coeffs >= numblocks * blocksize); Barrier barrier(numblocks);