diff --git a/tests/distributed/test_hopper_ll_precision.py b/tests/distributed/test_hopper_ll_precision.py deleted file mode 100644 index 7de0aaca42..0000000000 --- a/tests/distributed/test_hopper_ll_precision.py +++ /dev/null @@ -1,124 +0,0 @@ -import unittest - -import paddle -import paddle.distributed as dist -import paddle.distributed.communication.deep_ep as deep_ep -from paddle.distributed import fleet - - -class TestFusedMoE(unittest.TestCase): - def setUp(self) -> None: - pass - - def test_fused_moe(self): - num_ranks = dist.get_world_size() - if num_ranks <= 1: - return - rank_id = dist.get_rank() - paddle.seed(rank_id + 100) - - strategy = fleet.DistributedStrategy() - strategy.hybrid_configs = {"dp_degree": 1, "mp_degree": num_ranks, "pp_degree": 1} - fleet.init(is_collective=True, strategy=strategy) - - num_tokens, hidden, num_topk, num_experts = 64, 7168, 4, 64 - num_rdma_bytes = deep_ep.Buffer.get_low_latency_rdma_size_hint(num_tokens, hidden, num_ranks, num_experts) - - ep_group = fleet.get_hybrid_communicate_group().get_model_parallel_group() - buffer = deep_ep.Buffer( - ep_group, - num_nvl_bytes=0, - num_rdma_bytes=num_rdma_bytes, - low_latency_mode=True, - num_qps_per_rank=num_experts // num_ranks, - ) - - x = paddle.randn(shape=[num_tokens, hidden], dtype="bfloat16") - scores = paddle.randn([num_tokens, num_experts], dtype="float32").abs() + 1 - topk_info = paddle.topk(scores, num_topk, axis=-1, largest=True, sorted=False) - topk_weight = topk_info[0] - topk_idx = topk_info[1] - - gather_x = [] - dist.all_gather(gather_x, x, ep_group) - gather_x = paddle.stack(gather_x, axis=0) - - gather_topk_idx = [] - dist.all_gather(gather_topk_idx, topk_idx, ep_group) - gather_topk_idx = paddle.concat(gather_topk_idx, axis=0) - - handle = None - - num_tests = 10 - - for _ in range(num_tests): - - dispatch_use_fp8 = False - packed_recv_x, packed_recv_count, handle, event, hook = buffer.low_latency_dispatch( - x, - topk_idx, - None, # expertwise_scale, used in w4a8. - num_tokens, - num_experts, - use_fp8=dispatch_use_fp8, - async_finish=False, - return_recv_hook=True, - ) - - if hook is not None: - hook() - if dispatch_use_fp8: - fp8, scale = packed_recv_x[0], packed_recv_x[1] - fp32 = fp8.cast("float32").reshape([0, 0, hidden // 128, 128]) - scale = scale.transpose([0, 2, 1]).reshape([0, 0, hidden // 128, 1]) - fp32 = fp32 * scale - fp32 = fp32.reshape([0, 0, -1]) - - combined_hidden_states, _, _ = buffer.low_latency_combine( - packed_recv_x, - topk_idx, - topk_weight, - handle, - zero_copy=False, - async_finish=False, - return_recv_hook=False, - ) - - num_local_experts = num_experts // num_ranks - start_ep_id = rank_id * num_local_experts - end_ep_id = start_ep_id + num_local_experts - - num_tokens_send_by_rdma = 0 - for token_id in range(topk_idx.shape[0]): - for dst_expert_id in topk_idx[token_id].numpy().tolist(): - if dst_expert_id not in range(start_ep_id, end_ep_id): - num_tokens_send_by_rdma += 1 - print("num_tokens_send_by_rdma:", num_tokens_send_by_rdma) - - (recv_src_info, recv_layout_range, _, _) = handle - - for ep_id in range(start_ep_id, end_ep_id): - local_ep_id = ep_id - start_ep_id - token_num_this_ep = packed_recv_count[local_ep_id].item() - token_nums_per_rank = [] - begin_idx_per_rank = [] - for rank_id in range(num_ranks): - tmp = recv_layout_range[local_ep_id, rank_id].item() - begin_idx_per_rank.append(tmp >> 32) - token_nums_per_rank.append(tmp & ((1 << 32) - 1)) - assert token_num_this_ep == sum(token_nums_per_rank) - - for rank_id in range(num_ranks): - begin_idx = begin_idx_per_rank[rank_id] - end_idx = begin_idx + token_nums_per_rank[rank_id] - for token_id in range(begin_idx, end_idx): - token = packed_recv_x[local_ep_id, token_id, :] - # 这个token来自rank_id,并且是他的第多少个token呢? - src_token_id = recv_src_info[local_ep_id, token_id].item() - src_token = gather_x[rank_id, src_token_id, :] - # print(token - src_token) - assert (src_token - token).abs().max().item() == 0 - - -if __name__ == "__main__": - unittest.main() diff --git a/tests/distributed/test_hopper_ll_precision_entry.py b/tests/distributed/test_hopper_ll_precision_entry.py deleted file mode 100644 index 96955108ef..0000000000 --- a/tests/distributed/test_hopper_ll_precision_entry.py +++ /dev/null @@ -1,65 +0,0 @@ -# Copyright (c) 2025 PaddlePaddle Authors. All Rights Reserved. -# -# Licensed under the Apache License, Version 2.0 (the "License"); -# you may not use this file except in compliance with the License. -# You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. - -import os -import subprocess -import sys - - -def test_launch(): - """ - test_fused_moe - """ - current_dir = os.path.dirname(os.path.abspath(__file__)) - py_script = os.path.join(current_dir, "./test_hopper_ll_precision.py") - - # 为了方便在PDC的环境下直接python运行这个脚本 - os.environ.pop("PADDLE_ELASTIC_JOB_ID", None) - os.environ.pop("PADDLE_TRAINER_ENDPOINTS", None) - os.environ.pop("DISTRIBUTED_TRAINER_ENDPOINTS", None) - os.environ.pop("FLAGS_START_PORT", None) - os.environ.pop("PADDLE_ELASTIC_TIMEOUT", None) - - os.environ["CUDA_VISIBLE_DEVICES"] = "0,1" - - FD_API_PORT = int(os.getenv("FD_API_PORT", 8188)) - command = [ - sys.executable, - "-m", - "paddle.distributed.launch", - "--gpus", - "0,1", - "--master", - f"127.0.0.1:{FD_API_PORT}", - "--nnodes", - "1", - "--rank", - "0", - py_script, - ] - - process = subprocess.Popen(command, stdout=subprocess.PIPE, stderr=subprocess.PIPE, text=True) - - try: - stdout, stderr = process.communicate(timeout=400) - return_code = process.returncode - except subprocess.TimeoutExpired: - process.kill() - stdout, stderr = process.communicate() - return_code = -1 - print(f"std_out: {stdout}") - assert return_code in (0, 250, 255), f"Process exited with code {return_code}, stdout: {stdout}, stderr: {stderr}" - - -test_launch()