Searched defs:head_dim (Results 1 – 5 of 5) sorted by relevance
| /third_party/mindspore/mindspore-src/source/tests/st/fusion/ |
| D | test_flash_attention_fusion.py | 29 def attention_compuation(query, key, value, attention_mask, nums_heads_per_partition, head_dim): argument 47 def construct(self, query, key, value, attention_mask, nums_heads_per_partition, head_dim): argument
|
| /third_party/mindspore/mindspore-src/source/tests/st/ops/ascend/test_aclnn_ops/ |
| D | test_flash_attention.py | 31 … def __init__(self, num_heads, head_dim, dropout_rate=0.0, prev_tockens=65536, next_tockens=65536): argument 81 def __init__(self, head_dim, dropout_rate=0., softmax_compute_type=mstype.float16): argument
|
| /third_party/mindspore/mindspore-src/source/mindspore/ccsrc/plugin/device/ascend/kernel/internal/ |
| D | flash_attention_score.cc | 38 int64_t head_dim = q_shape[kDim2] / head_num; in CreateOpParam() local
|
| /third_party/mindspore/mindspore-src/source/tests/ut/python/parallel/ |
| D | test_reshape_and_cache.py | 27 def generate_inputs(bs=1, seq_len=1, num_head=40, head_dim=128, block_size=16, max_seq=2048): argument
|
| /third_party/mindspore/mindspore-src/source/mindspore/lite/test/st/python/akg_custom_ops/ |
| D | test_reshape_and_cache.py | 33 def __init__(self, num_blocks, block_size, kv_head, head_dim): argument
|