0 stringclasses 12 values | 1 float64 0 316k |
|---|---|
megatron.core.transformer.attention.forward.qkv | 265.433105 |
megatron.core.transformer.attention.forward.adjust_key_value | 0.10928 |
megatron.core.transformer.attention.forward.rotary_pos_emb | 0.089152 |
megatron.core.transformer.attention.forward.core_attention | 8,042.949707 |
megatron.core.transformer.attention.forward.linear_proj | 3.619552 |
megatron.core.transformer.transformer_layer._forward_attention.self_attention | 8,313.681641 |
megatron.core.transformer.transformer_layer._forward_attention.self_attn_bda | 484.529053 |
megatron.core.transformer.mlp.forward.linear_fc1 | 3.995008 |
megatron.core.transformer.mlp.forward.activation | 503.244904 |
megatron.core.transformer.mlp.forward.linear_fc2 | 7.014528 |
megatron.core.transformer.transformer_layer._forward_mlp.mlp | 515.026611 |
megatron.core.transformer.transformer_layer._forward_mlp.mlp_bda | 0.496736 |
megatron.core.transformer.attention.forward.qkv | 0.583264 |
megatron.core.transformer.attention.forward.adjust_key_value | 0.083072 |
megatron.core.transformer.attention.forward.rotary_pos_emb | 0.092416 |
megatron.core.transformer.attention.forward.core_attention | 1,546.873657 |
megatron.core.transformer.attention.forward.linear_proj | 0.019744 |
megatron.core.transformer.transformer_layer._forward_attention.self_attention | 1,547.96167 |
megatron.core.transformer.transformer_layer._forward_attention.self_attn_bda | 0.01088 |
megatron.core.transformer.mlp.forward.linear_fc1 | 0.051488 |
megatron.core.transformer.mlp.forward.activation | 0.009344 |
megatron.core.transformer.mlp.forward.linear_fc2 | 0.048928 |
megatron.core.transformer.transformer_layer._forward_mlp.mlp | 0.12176 |
megatron.core.transformer.transformer_layer._forward_mlp.mlp_bda | 0.010912 |
megatron.core.transformer.attention.forward.qkv | 228.266083 |
megatron.core.transformer.attention.forward.adjust_key_value | 0.003008 |
megatron.core.transformer.attention.forward.rotary_pos_emb | 0.002976 |
megatron.core.transformer.attention.forward.core_attention | 6,259.61084 |
megatron.core.transformer.attention.forward.linear_proj | 4.93552 |
megatron.core.transformer.transformer_layer._forward_attention.self_attention | 6,494.400879 |
megatron.core.transformer.transformer_layer._forward_attention.self_attn_bda | 1,700.212524 |
megatron.core.transformer.mlp.forward.linear_fc1 | 9.579488 |
megatron.core.transformer.mlp.forward.activation | 612.642334 |
megatron.core.transformer.mlp.forward.linear_fc2 | 6.07808 |
megatron.core.transformer.transformer_layer._forward_mlp.mlp | 628.885315 |
megatron.core.transformer.transformer_layer._forward_mlp.mlp_bda | 0.45216 |
megatron.core.transformer.attention.forward.qkv | 2.621376 |
megatron.core.transformer.attention.forward.adjust_key_value | 0.002976 |
megatron.core.transformer.attention.forward.rotary_pos_emb | 0.003008 |
megatron.core.transformer.attention.forward.core_attention | 2,150.422363 |
megatron.core.transformer.attention.forward.linear_proj | 1.48992 |
megatron.core.transformer.transformer_layer._forward_attention.self_attention | 2,154.558105 |
megatron.core.transformer.transformer_layer._forward_attention.self_attn_bda | 0.450368 |
megatron.core.transformer.mlp.forward.linear_fc1 | 5.927296 |
megatron.core.transformer.mlp.forward.activation | 0.660736 |
megatron.core.transformer.mlp.forward.linear_fc2 | 6.012608 |
megatron.core.transformer.transformer_layer._forward_mlp.mlp | 12.612736 |
megatron.core.transformer.transformer_layer._forward_mlp.mlp_bda | 0.4496 |
megatron.core.transformer.attention.forward.qkv | 253.147552 |
megatron.core.transformer.attention.forward.adjust_key_value | 0.131328 |
megatron.core.transformer.attention.forward.rotary_pos_emb | 0.109024 |
megatron.core.transformer.attention.forward.core_attention | 8,898.134766 |
megatron.core.transformer.attention.forward.linear_proj | 4.493184 |
megatron.core.transformer.transformer_layer._forward_attention.self_attention | 9,161.310547 |
megatron.core.transformer.transformer_layer._forward_attention.self_attn_bda | 1,760.517944 |
megatron.core.transformer.mlp.forward.linear_fc1 | 3.384544 |
megatron.core.transformer.mlp.forward.activation | 585.555664 |
megatron.core.transformer.mlp.forward.linear_fc2 | 1.131648 |
megatron.core.transformer.transformer_layer._forward_mlp.mlp | 592.144958 |
megatron.core.transformer.transformer_layer._forward_mlp.mlp_bda | 1.110656 |
megatron.core.transformer.attention.forward.qkv | 0.865664 |
megatron.core.transformer.attention.forward.adjust_key_value | 0.085984 |
megatron.core.transformer.attention.forward.rotary_pos_emb | 0.09856 |
megatron.core.transformer.attention.forward.core_attention | 2,296.233643 |
megatron.core.transformer.attention.forward.linear_proj | 0.029664 |
megatron.core.transformer.transformer_layer._forward_attention.self_attention | 2,298.060547 |
megatron.core.transformer.transformer_layer._forward_attention.self_attn_bda | 0.01664 |
megatron.core.transformer.mlp.forward.linear_fc1 | 0.103552 |
megatron.core.transformer.mlp.forward.activation | 0.017632 |
megatron.core.transformer.mlp.forward.linear_fc2 | 0.095488 |
megatron.core.transformer.transformer_layer._forward_mlp.mlp | 0.228448 |
megatron.core.transformer.transformer_layer._forward_mlp.mlp_bda | 0.016832 |
megatron.core.transformer.attention.forward.qkv | 264.525848 |
megatron.core.transformer.attention.forward.adjust_key_value | 0.112352 |
megatron.core.transformer.attention.forward.rotary_pos_emb | 0.086208 |
megatron.core.transformer.attention.forward.core_attention | 7,904.231445 |
megatron.core.transformer.attention.forward.linear_proj | 4.880224 |
megatron.core.transformer.transformer_layer._forward_attention.self_attention | 8,175.945801 |
megatron.core.transformer.transformer_layer._forward_attention.self_attn_bda | 1,810.02124 |
megatron.core.transformer.mlp.forward.linear_fc1 | 1.24016 |
megatron.core.transformer.mlp.forward.activation | 779.075134 |
megatron.core.transformer.mlp.forward.linear_fc2 | 1.036992 |
megatron.core.transformer.transformer_layer._forward_mlp.mlp | 781.970093 |
megatron.core.transformer.transformer_layer._forward_mlp.mlp_bda | 0.467968 |
megatron.core.transformer.attention.forward.qkv | 0.587424 |
megatron.core.transformer.attention.forward.adjust_key_value | 0.07744 |
megatron.core.transformer.attention.forward.rotary_pos_emb | 0.0848 |
megatron.core.transformer.attention.forward.core_attention | 2,649.5 |
megatron.core.transformer.attention.forward.linear_proj | 0.091488 |
megatron.core.transformer.transformer_layer._forward_attention.self_attention | 2,650.628906 |
megatron.core.transformer.transformer_layer._forward_attention.self_attn_bda | 0.03808 |
megatron.core.transformer.mlp.forward.linear_fc1 | 0.369184 |
megatron.core.transformer.mlp.forward.activation | 0.04704 |
megatron.core.transformer.mlp.forward.linear_fc2 | 0.34192 |
megatron.core.transformer.transformer_layer._forward_mlp.mlp | 0.769728 |
megatron.core.transformer.transformer_layer._forward_mlp.mlp_bda | 0.037984 |
megatron.core.transformer.attention.forward.qkv | 250.872482 |
megatron.core.transformer.attention.forward.adjust_key_value | 0.10944 |
megatron.core.transformer.attention.forward.rotary_pos_emb | 0.083904 |
megatron.core.transformer.attention.forward.core_attention | 8,820.628906 |
End of preview. Expand
in Data Studio
No dataset card yet
- Downloads last month
- 35