Spaces:
Runtime error
Runtime error
File size: 1,522 Bytes
e202b16 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 |
xFormers optimized operators
============================================================
Memory-efficient attention
---------------------------
.. automodule:: xformers.ops
:members: memory_efficient_attention, AttentionOpBase
:show-inheritance:
:imported-members:
Available implementations
~~~~~~~~~~~~~~~~~~~~~~~~~~~
.. automodule:: xformers.ops.fmha.cutlass
:members: FwOp, BwOp
:member-order: bysource
.. automodule:: xformers.ops.fmha.flash
:members: FwOp, BwOp
:member-order: bysource
.. automodule:: xformers.ops.fmha.small_k
:members: FwOp, BwOp
:member-order: bysource
.. automodule:: xformers.ops.fmha.ck
:members: FwOp, BwOp
:member-order: bysource
.. automodule:: xformers.ops.fmha.ck_decoder
:members: FwOp
:member-order: bysource
.. automodule:: xformers.ops.fmha.ck_splitk
:members: FwOp
:member-order: bysource
Attention biases
~~~~~~~~~~~~~~~~~~~~
.. automodule:: xformers.ops.fmha.attn_bias
:members:
:show-inheritance:
:member-order: bysource
Partial Attention
~~~~~~~~~~~~~~~~~~~~
.. automodule:: xformers.ops.fmha
:members: memory_efficient_attention_partial, merge_attentions
:member-order: bysource
Non-autograd implementations
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
.. automodule:: xformers.ops.fmha
:members: memory_efficient_attention_forward, memory_efficient_attention_forward_requires_grad, memory_efficient_attention_backward
:show-inheritance:
:imported-members:
:member-order: bysource
|