from megatron.core.tensor_parallel.cross_entropy import vocab_parallel_cross_entropy import torch from tests.test_utilities import Utils import numpy as np def test_vocab_parallel_cross_entropy(): Utils.initialize_model_parallel(4,2) vocab_parallel_logits = torch.range(0,7).repeat(16,4).cuda() target = torch.arange(0,32,2).cuda() output = vocab_parallel_cross_entropy(vocab_parallel_logits, target) expected_output = torch.tensor([10.2309, 8.2309, 6.2309, 4.2309, 10.2309, 8.2309, 6.2309, 4.2309, 10.2309, 8.2309, 6.2309, 4.2309, 10.2309, 8.2309, 6.2309, 4.2309]).cuda() assert(torch.equal(torch.round(expected_output), torch.round(output))) Utils.destroy_model_parallel()