Coverage for src/flag_gems/fused/__init__.py: 100%

28 statements  

« prev     ^ index     » next       coverage.py v7.6.9, created at 2026-03-27 02:51 +0800

1from flag_gems.fused.apply_repetition_penalties import apply_repetition_penalties 

2from flag_gems.fused.bincount import bincount 

3from flag_gems.fused.concat_and_cache_mla import concat_and_cache_mla 

4from flag_gems.fused.cross_entropy_loss import cross_entropy_loss 

5from flag_gems.fused.cutlass_scaled_mm import cutlass_scaled_mm 

6from flag_gems.fused.FLA import ( 

7 chunk_gated_delta_rule_fwd, 

8 fused_recurrent_gated_delta_rule_fwd, 

9) 

10from flag_gems.fused.flash_mla import flash_mla 

11from flag_gems.fused.fused_add_rms_norm import fused_add_rms_norm 

12from flag_gems.fused.fused_moe import ( 

13 dispatch_fused_moe_kernel, 

14 fused_experts_impl, 

15 inplace_fused_experts, 

16 invoke_fused_moe_triton_kernel, 

17 outplace_fused_experts, 

18) 

19from flag_gems.fused.geglu import dgeglu, geglu 

20from flag_gems.fused.gelu_and_mul import gelu_and_mul 

21from flag_gems.fused.grouped_topk import grouped_topk 

22from flag_gems.fused.instance_norm import instance_norm 

23from flag_gems.fused.moe_align_block_size import ( 

24 moe_align_block_size, 

25 moe_align_block_size_triton, 

26) 

27from flag_gems.fused.moe_sum import moe_sum 

28from flag_gems.fused.outer import outer 

29from flag_gems.fused.reglu import dreglu, reglu 

30from flag_gems.fused.reshape_and_cache import reshape_and_cache 

31from flag_gems.fused.reshape_and_cache_flash import reshape_and_cache_flash 

32from flag_gems.fused.rotary_embedding import apply_rotary_pos_emb 

33from flag_gems.fused.rwkv_ka_fusion import rwkv_ka_fusion 

34from flag_gems.fused.rwkv_mm_sparsity import rwkv_mm_sparsity 

35from flag_gems.fused.silu_and_mul import silu_and_mul, silu_and_mul_out 

36from flag_gems.fused.skip_layernorm import skip_layer_norm 

37from flag_gems.fused.swiglu import dswiglu, swiglu 

38from flag_gems.fused.topk_softmax import topk_softmax 

39from flag_gems.fused.weight_norm import weight_norm 

40 

41__all__ = [ 

42 "apply_repetition_penalties", 

43 "apply_rotary_pos_emb", 

44 "bincount", 

45 "chunk_gated_delta_rule_fwd", 

46 "concat_and_cache_mla", 

47 "cutlass_scaled_mm", 

48 "cross_entropy_loss", 

49 "dispatch_fused_moe_kernel", 

50 "dgeglu", 

51 "dreglu", 

52 "dswiglu", 

53 "flash_mla", 

54 "fused_add_rms_norm", 

55 "fused_experts_impl", 

56 "fused_recurrent_gated_delta_rule_fwd", 

57 "geglu", 

58 "gelu_and_mul", 

59 "grouped_topk", 

60 "inplace_fused_experts", 

61 "instance_norm", 

62 "invoke_fused_moe_triton_kernel", 

63 "moe_sum", 

64 "moe_align_block_size", 

65 "moe_align_block_size_triton", 

66 "outer", 

67 "outplace_fused_experts", 

68 "reglu", 

69 "reshape_and_cache", 

70 "reshape_and_cache_flash", 

71 "rwkv_ka_fusion", 

72 "rwkv_mm_sparsity", 

73 "silu_and_mul", 

74 "silu_and_mul_out", 

75 "skip_layer_norm", 

76 "swiglu", 

77 "topk_softmax", 

78 "weight_norm", 

79]