Spaces:
Sleeping
Sleeping
/*************************************************************************************************** | |
* Copyright (c) 2023 - 2024 NVIDIA CORPORATION & AFFILIATES. All rights reserved. | |
* SPDX-License-Identifier: BSD-3-Clause | |
* | |
* Redistribution and use in source and binary forms, with or without | |
* modification, are permitted provided that the following conditions are met: | |
* | |
* 1. Redistributions of source code must retain the above copyright notice, this | |
* list of conditions and the following disclaimer. | |
* | |
* 2. Redistributions in binary form must reproduce the above copyright notice, | |
* this list of conditions and the following disclaimer in the documentation | |
* and/or other materials provided with the distribution. | |
* | |
* 3. Neither the name of the copyright holder nor the names of its | |
* contributors may be used to endorse or promote products derived from | |
* this software without specific prior written permission. | |
* | |
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" | |
* AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE | |
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE | |
* DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE | |
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL | |
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR | |
* SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER | |
* CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, | |
* OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE | |
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. | |
* | |
**************************************************************************************************/ | |
// Config | |
namespace cute | |
{ | |
/// Copy via cp.async with caching at all levels | |
template <class TS, class TD = TS> | |
struct SM80_CP_ASYNC_CACHEALWAYS | |
{ | |
using SRegisters = TS[1]; | |
using DRegisters = TD[1]; | |
static_assert(sizeof(TS) == sizeof(TD), "cp.async requires sizeof(src_value_type) == sizeof(dst_value_type)"); | |
static_assert(sizeof(TS) == 4 || sizeof(TS) == 8 || sizeof(TS) == 16, "cp.async sizeof(TS) is not supported"); | |
CUTE_HOST_DEVICE static void | |
copy(TS const& gmem_src, | |
TD & smem_dst) | |
{ | |
TS const* gmem_ptr = &gmem_src; | |
uint32_t smem_int_ptr = cast_smem_ptr_to_uint(&smem_dst); | |
asm volatile("cp.async.ca.shared.global.L2::128B [%0], [%1], %2;\n" | |
:: "r"(smem_int_ptr), | |
"l"(gmem_ptr), | |
"n"(sizeof(TS))); | |
CUTE_INVALID_CONTROL_PATH("Support for cp.async instructions has not been enabled"); | |
} | |
}; | |
/// Copy via cp.async with caching at global level | |
template <class TS, class TD = TS> | |
struct SM80_CP_ASYNC_CACHEGLOBAL | |
{ | |
using SRegisters = TS[1]; | |
using DRegisters = TD[1]; | |
static_assert(sizeof(TS) == sizeof(TD), "cp.async requires sizeof(src_value_type) == sizeof(dst_value_type)"); | |
static_assert(sizeof(TS) == 4 || sizeof(TS) == 8 || sizeof(TS) == 16, "cp.async sizeof(TS) is not supported"); | |
CUTE_HOST_DEVICE static void | |
copy(TS const& gmem_src, | |
TD & smem_dst) | |
{ | |
TS const* gmem_ptr = &gmem_src; | |
uint32_t smem_int_ptr = cast_smem_ptr_to_uint(&smem_dst); | |
asm volatile("cp.async.cg.shared.global.L2::128B [%0], [%1], %2;\n" | |
:: "r"(smem_int_ptr), | |
"l"(gmem_ptr), | |
"n"(sizeof(TS))); | |
CUTE_INVALID_CONTROL_PATH("Support for cp.async instructions has not been enabled"); | |
} | |
}; | |
/// Copy via cp.async with caching at all levels | |
template <class TS, class TD = TS> | |
struct SM80_CP_ASYNC_CACHEALWAYS_ZFILL | |
{ | |
using SRegisters = TS[1]; | |
using DRegisters = TD[1]; | |
static_assert(sizeof(TS) == sizeof(TD), "cp.async requires sizeof(src_value_type) == sizeof(dst_value_type)"); | |
static_assert(sizeof(TS) == 4 || sizeof(TS) == 8 || sizeof(TS) == 16, "cp.async sizeof(TS) is not supported"); | |
CUTE_HOST_DEVICE static void | |
copy(TS const& gmem_src, | |
TD & smem_dst, | |
bool pred) | |
{ | |
TS const* gmem_ptr = &gmem_src; | |
uint32_t smem_int_ptr = cast_smem_ptr_to_uint(&smem_dst); | |
int src_size = pred ? sizeof(TS) : 0; | |
asm volatile("cp.async.ca.shared.global.L2::128B [%0], [%1], %2, %3;\n" | |
:: "r"(smem_int_ptr), | |
"l"(gmem_ptr), | |
"n"(sizeof(TS)), | |
"r"(src_size)); | |
CUTE_INVALID_CONTROL_PATH("Support for cp.async instructions has not been enabled"); | |
} | |
}; | |
/// Copy via cp.async with caching at global level | |
template <class TS, class TD = TS> | |
struct SM80_CP_ASYNC_CACHEGLOBAL_ZFILL | |
{ | |
using SRegisters = TS[1]; | |
using DRegisters = TD[1]; | |
static_assert(sizeof(TS) == sizeof(TD), "cp.async requires sizeof(src_value_type) == sizeof(dst_value_type)"); | |
static_assert(sizeof(TS) == 4 || sizeof(TS) == 8 || sizeof(TS) == 16, "cp.async sizeof(TS) is not supported"); | |
CUTE_HOST_DEVICE static void | |
copy(TS const& gmem_src, | |
TD & smem_dst, | |
bool pred) | |
{ | |
TS const* gmem_ptr = &gmem_src; | |
uint32_t smem_int_ptr = cast_smem_ptr_to_uint(&smem_dst); | |
int src_size = pred ? sizeof(TS) : 0; | |
asm volatile("cp.async.cg.shared.global.L2::128B [%0], [%1], %2, %3;\n" | |
:: "r"(smem_int_ptr), | |
"l"(gmem_ptr), | |
"n"(sizeof(TS)), | |
"r"(src_size)); | |
CUTE_INVALID_CONTROL_PATH("Support for cp.async instructions has not been enabled"); | |
} | |
}; | |
//////////////////////////////////////////////////////////////////////////////////////////////////// | |
/// Establishes an ordering w.r.t previously issued cp.async instructions. Does not block. | |
CUTE_HOST_DEVICE | |
void | |
cp_async_fence() | |
{ | |
asm volatile("cp.async.commit_group;\n" ::); | |
} | |
//////////////////////////////////////////////////////////////////////////////////////////////////// | |
/// Blocks until all but N previous cp.async.commit_group operations have committed. | |
template <int N> | |
CUTE_HOST_DEVICE | |
void | |
cp_async_wait() | |
{ | |
if constexpr (N == 0) { | |
asm volatile("cp.async.wait_all;\n" ::); | |
} else { | |
asm volatile("cp.async.wait_group %0;\n" :: "n"(N)); | |
} | |
} | |
template <int N> | |
CUTE_HOST_DEVICE | |
void | |
cp_async_wait(Int<N>) | |
{ | |
return cp_async_wait<N>(); | |
} | |
///////////////////////////////////////////////////////////////////////////////////////////////// | |
} // end namespace cute | |