/*************************************************************************************************** * Copyright (c) 2023 - 2024 NVIDIA CORPORATION & AFFILIATES. All rights reserved. * SPDX-License-Identifier: BSD-3-Clause * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions are met: * * 1. Redistributions of source code must retain the above copyright notice, this * list of conditions and the following disclaimer. * * 2. Redistributions in binary form must reproduce the above copyright notice, * this list of conditions and the following disclaimer in the documentation * and/or other materials provided with the distribution. * * 3. Neither the name of the copyright holder nor the names of its * contributors may be used to endorse or promote products derived from * this software without specific prior written permission. * * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE * DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR * SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER * CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, * OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. * **************************************************************************************************/ #pragma once #include #include // Config #if (defined(__CUDA_ARCH__) && (__CUDA_ARCH__ >= 800)) # define CUTE_ARCH_CP_ASYNC_SM80_ENABLED #endif namespace cute { /// Copy via cp.async with caching at all levels template struct SM80_CP_ASYNC_CACHEALWAYS { using SRegisters = TS[1]; using DRegisters = TD[1]; static_assert(sizeof(TS) == sizeof(TD), "cp.async requires sizeof(src_value_type) == sizeof(dst_value_type)"); static_assert(sizeof(TS) == 4 || sizeof(TS) == 8 || sizeof(TS) == 16, "cp.async sizeof(TS) is not supported"); CUTE_HOST_DEVICE static void copy(TS const& gmem_src, TD & smem_dst) { #if defined(CUTE_ARCH_CP_ASYNC_SM80_ENABLED) TS const* gmem_ptr = &gmem_src; uint32_t smem_int_ptr = cast_smem_ptr_to_uint(&smem_dst); asm volatile("cp.async.ca.shared.global.L2::128B [%0], [%1], %2;\n" :: "r"(smem_int_ptr), "l"(gmem_ptr), "n"(sizeof(TS))); #else CUTE_INVALID_CONTROL_PATH("Support for cp.async instructions has not been enabled"); #endif } }; /// Copy via cp.async with caching at global level template struct SM80_CP_ASYNC_CACHEGLOBAL { using SRegisters = TS[1]; using DRegisters = TD[1]; static_assert(sizeof(TS) == sizeof(TD), "cp.async requires sizeof(src_value_type) == sizeof(dst_value_type)"); static_assert(sizeof(TS) == 4 || sizeof(TS) == 8 || sizeof(TS) == 16, "cp.async sizeof(TS) is not supported"); CUTE_HOST_DEVICE static void copy(TS const& gmem_src, TD & smem_dst) { #if defined(CUTE_ARCH_CP_ASYNC_SM80_ENABLED) TS const* gmem_ptr = &gmem_src; uint32_t smem_int_ptr = cast_smem_ptr_to_uint(&smem_dst); asm volatile("cp.async.cg.shared.global.L2::128B [%0], [%1], %2;\n" :: "r"(smem_int_ptr), "l"(gmem_ptr), "n"(sizeof(TS))); #else CUTE_INVALID_CONTROL_PATH("Support for cp.async instructions has not been enabled"); #endif } }; /// Copy via cp.async with caching at all levels template struct SM80_CP_ASYNC_CACHEALWAYS_ZFILL { using SRegisters = TS[1]; using DRegisters = TD[1]; static_assert(sizeof(TS) == sizeof(TD), "cp.async requires sizeof(src_value_type) == sizeof(dst_value_type)"); static_assert(sizeof(TS) == 4 || sizeof(TS) == 8 || sizeof(TS) == 16, "cp.async sizeof(TS) is not supported"); CUTE_HOST_DEVICE static void copy(TS const& gmem_src, TD & smem_dst, bool pred) { #if defined(CUTE_ARCH_CP_ASYNC_SM80_ENABLED) TS const* gmem_ptr = &gmem_src; uint32_t smem_int_ptr = cast_smem_ptr_to_uint(&smem_dst); int src_size = pred ? sizeof(TS) : 0; asm volatile("cp.async.ca.shared.global.L2::128B [%0], [%1], %2, %3;\n" :: "r"(smem_int_ptr), "l"(gmem_ptr), "n"(sizeof(TS)), "r"(src_size)); #else CUTE_INVALID_CONTROL_PATH("Support for cp.async instructions has not been enabled"); #endif } }; /// Copy via cp.async with caching at global level template struct SM80_CP_ASYNC_CACHEGLOBAL_ZFILL { using SRegisters = TS[1]; using DRegisters = TD[1]; static_assert(sizeof(TS) == sizeof(TD), "cp.async requires sizeof(src_value_type) == sizeof(dst_value_type)"); static_assert(sizeof(TS) == 4 || sizeof(TS) == 8 || sizeof(TS) == 16, "cp.async sizeof(TS) is not supported"); CUTE_HOST_DEVICE static void copy(TS const& gmem_src, TD & smem_dst, bool pred) { #if defined(CUTE_ARCH_CP_ASYNC_SM80_ENABLED) TS const* gmem_ptr = &gmem_src; uint32_t smem_int_ptr = cast_smem_ptr_to_uint(&smem_dst); int src_size = pred ? sizeof(TS) : 0; asm volatile("cp.async.cg.shared.global.L2::128B [%0], [%1], %2, %3;\n" :: "r"(smem_int_ptr), "l"(gmem_ptr), "n"(sizeof(TS)), "r"(src_size)); #else CUTE_INVALID_CONTROL_PATH("Support for cp.async instructions has not been enabled"); #endif } }; //////////////////////////////////////////////////////////////////////////////////////////////////// /// Establishes an ordering w.r.t previously issued cp.async instructions. Does not block. CUTE_HOST_DEVICE void cp_async_fence() { #if defined(CUTE_ARCH_CP_ASYNC_SM80_ENABLED) asm volatile("cp.async.commit_group;\n" ::); #endif } //////////////////////////////////////////////////////////////////////////////////////////////////// /// Blocks until all but N previous cp.async.commit_group operations have committed. template CUTE_HOST_DEVICE void cp_async_wait() { #if defined(CUTE_ARCH_CP_ASYNC_SM80_ENABLED) if constexpr (N == 0) { asm volatile("cp.async.wait_all;\n" ::); } else { asm volatile("cp.async.wait_group %0;\n" :: "n"(N)); } #endif } template CUTE_HOST_DEVICE void cp_async_wait(Int) { return cp_async_wait(); } ///////////////////////////////////////////////////////////////////////////////////////////////// } // end namespace cute