An open API service for providing issue and pull request metadata for open source projects.

GitHub / pytorch/pytorch issues and pull requests

Labelled with: module: pt2-dispatcher

#165151 - Custom operator registration doesn't work when libtorch.so changes

Issue - State: open - Opened by Flamefire about 1 month ago - 4 comments
Labels: module: custom-operators, module: pt2-dispatcher

#163588 - Flex Attention Silently Incorrect with bfloat16 and Masks

Issue - State: closed - Opened by zaptrem about 2 months ago - 7 comments
Labels: triaged, oncall: pt2, module: higher order operators, module: pt2-dispatcher, module: flex attention

#162228 - Backpropagation to flex_attention `score_mod` biases fails based on presence of graph breaks

Issue - State: closed - Opened by stefan-baumann 3 months ago - 11 comments
Labels: triaged, module: correctness (silent), oncall: pt2, module: pt2-dispatcher, pt2: ubn

#161473 - flex attention is slower than sdpa

Issue - State: closed - Opened by LinB203 3 months ago - 4 comments
Labels: triaged, oncall: pt2, module: higher order operators, module: pt2-dispatcher, module: flex attention

#161077 - DISABLED test_override_meta (__main__.TestCustomOp)

Issue - State: closed - Opened by pytorch-bot[bot] 3 months ago - 2 comments
Labels: triaged, module: flaky-tests, module: custom-operators, skipped, oncall: pt2, module: pt2-dispatcher

#160955 - NotImplementedError: 64-bit indexing is not yet implemented for triton templates

Issue - State: closed - Opened by sirluk 3 months ago - 1 comment
Labels: triaged, oncall: pt2, module: higher order operators, module: pt2-dispatcher, module: flex attention

#160909 - compile with PrivateUse1 see tensors on "meta"

Issue - State: closed - Opened by zzzwen 3 months ago - 23 comments
Labels: triaged, module: dispatch, oncall: pt2, module: pt2-dispatcher, module: PrivateUse1

#160901 - [dynamo] traced autograd.Function silently incorrect due to bad requires_grad propagation

Issue - State: closed - Opened by xmfan 3 months ago
Labels: high priority, triaged, oncall: pt2, module: dynamo, module: higher order operators, module: pt2-dispatcher

#160840 - `torch.nn.functional.interpolate` generates inconsistent computation results after `torch.compile`

Issue - State: closed - Opened by dlibk 3 months ago - 5 comments
Labels: high priority, triaged, oncall: pt2, module: pt2-dispatcher

#160409 - test_flex_attention fails for float16 & bfloat16 dtypes on GPU

Issue - State: closed - Opened by Flamefire 3 months ago - 1 comment
Labels: triaged, oncall: pt2, module: pt2-dispatcher, module: flex attention

#160385 - TTGIR error for FlexAttention on B200

Issue - State: closed - Opened by drisspg 3 months ago - 2 comments
Labels: module: cuda, triaged, oncall: pt2, upstream triton, module: higher order operators, module: pt2-dispatcher, module: flex attention, Blackwell

#160333 - Tensor subclasses don't work with torch.library.triton_op

Issue - State: open - Opened by zou3519 3 months ago
Labels: tensor subclass, module: pt2-dispatcher

#160018 - Triton kernel generated for torch.compile(create_block_mask) using Flex Attention throws CUDA Illegal memory access

Issue - State: open - Opened by yubofredwang 3 months ago - 2 comments
Labels: module: crash, triaged, oncall: pt2, module: higher order operators, module: pt2-dispatcher, module: flex attention

#159462 - Inductor Fails on Conv1D After Permute with Stride Mismatch Error

Issue - State: closed - Opened by LiSsHhUuAaIi 4 months ago - 3 comments
Labels: module: cpu, triaged, oncall: pt2, module: fakeTensor, module: pt2-dispatcher

#159247 - [Flex Attention] Accuracy issue with kv length not multiple of kv block size

Issue - State: closed - Opened by Valentine233 4 months ago - 3 comments
Labels: triaged, oncall: pt2, module: inductor, module: higher order operators, module: pt2-dispatcher, module: flex attention

#158414 - received a FakeScriptObject input when dispatching DispatchKey.AutocastCUDA. but no python implementation is found

Issue - State: closed - Opened by asglover 4 months ago - 6 comments
Labels: triaged, module: custom-operators, oncall: pt2, module: pt2-dispatcher

#158081 - `torch.compile` errors with inductor raised `GuardOnDataDependentSymNode` exception

Issue - State: open - Opened by StrongerXi 4 months ago - 3 comments
Labels: triaged, oncall: pt2, module: dynamic shapes, module: inductor, module: pt2-dispatcher

#157833 - torch.compile with flex_attention: 'ShapeAsConstantBuffer' object has no attribute 'dtype'

Issue - State: closed - Opened by s22chan 4 months ago - 3 comments
Labels: triaged, oncall: pt2, module: inductor, module: higher order operators, module: pt2-dispatcher, module: flex attention

#157675 - Cannot create a mask for each sequence in a batch with Flex Attention

Issue - State: closed - Opened by luispl77 5 months ago - 3 comments
Labels: oncall: pt2, module: higher order operators, module: pt2-dispatcher, module: flex attention

#157543 - Vmap error raised by mask_mod of FlexAttention

Issue - State: closed - Opened by NiuMa-1234 5 months ago - 2 comments
Labels: triaged, module: vmap, oncall: pt2, module: functorch, module: higher order operators, module: pt2-dispatcher, module: flex attention

#157276 - several `transformers` tests fail with `torch 2.8 RC` but pass with `torch 2.7.1` on `T4` (but both pass on `A10`)

Issue - State: open - Opened by ydshieh 5 months ago - 3 comments
Labels: high priority, triage review, module: cuda, oncall: pt2, module: inductor, module: higher order operators, module: pt2-dispatcher, module: flex attention

#157104 - Document AOTAutograd autocast divergence from eager behavior and provide a workaround

Issue - State: closed - Opened by zou3519 5 months ago - 1 comment
Labels: triaged, oncall: pt2, module: pt2-dispatcher

#156688 - [CPU][flex attention] Llama 3 failed on CPU with PyTorch 2025-06-22 nightly wheel

Issue - State: closed - Opened by LifengWang 5 months ago
Labels: needs reproduction, triaged, module: regression, oncall: pt2, module: higher order operators, oncall: cpu inductor, module: pt2-dispatcher, module: flex attention

#156322 - Provide a way to allow dynamo to trace into an operator defined with `torch.library.custom_op`

Issue - State: closed - Opened by fxmarty-amd 5 months ago - 3 comments
Labels: triaged, module: custom-operators, oncall: pt2, module: pt2-dispatcher

#155721 - Add runtime profiler info for AOTDispatcher prologue

Issue - State: closed - Opened by bdhirsh 5 months ago
Labels: triaged, oncall: pt2, module: pt2-dispatcher

#155386 - OpOverloads should have annotations

Issue - State: open - Opened by zou3519 6 months ago - 1 comment
Labels: triaged, module: custom-operators, module: dispatch, oncall: pt2, module: pt2-dispatcher

#155385 - [custom_op] Custom ops created by @custom_op should get type hints propagated to their OpOverload

Issue - State: open - Opened by zou3519 6 months ago - 1 comment
Labels: triaged, module: custom-operators, oncall: pt2, module: pt2-dispatcher, internal ramp-up task

#154820 - In-place operations are reordered across the forward-backward in autograd function

Issue - State: closed - Opened by zxfseserer 6 months ago - 6 comments
Labels: triaged, oncall: pt2, module: dynamo, module: pt2-dispatcher, ubn

#154554 - flex_attention + NJT output inconsistent with non-NJT results

Issue - State: open - Opened by mauriceweiler 6 months ago - 2 comments
Labels: triaged, oncall: pt2, module: higher order operators, module: pt2-dispatcher, module: flex attention

#154111 - DTensor + torch.compile on CPU: compiled matmul fails with multiple shape inputs

Issue - State: closed - Opened by kberat-synnada 6 months ago - 5 comments
Labels: oncall: distributed, triaged, oncall: pt2, module: aotdispatch, module: dtensor, module: pt2-dispatcher

#153993 - torch.compile custom autograd.Function + dynamic shapes + torch.tensor constructor in backward can cause crash

Issue - State: closed - Opened by bdhirsh 6 months ago - 2 comments
Labels: high priority, oncall: pt2, module: pt2-dispatcher

#153763 - DISABLED test_impl_device_cpu (__main__.TestCustomOp)

Issue - State: closed - Opened by pytorch-bot[bot] 6 months ago - 1 comment
Labels: triaged, module: flaky-tests, module: custom-operators, skipped, oncall: pt2, module: pt2-dispatcher

#152593 - Flex Attention doesn't scale with custom bias

Issue - State: open - Opened by danjenson 7 months ago - 2 comments
Labels: triaged, oncall: pt2, module: pt2-dispatcher, module: flex attention

#152309 - Softmax Decomp Causes Incorrect Gradients when Using `torch.compile` with `F.multi_head_attention_forward`

Issue - State: closed - Opened by defaultd661 7 months ago - 14 comments
Labels: high priority, triaged, module: correctness (silent), oncall: pt2, module: decompositions, module: aotdispatch, module: pt2-dispatcher, module: sdpa, ubn

#152297 - Flex attention: batch-index-dependent block mask causes error with changing batch size

Issue - State: closed - Opened by zhihanyang2022 7 months ago - 1 comment
Labels: triaged, oncall: pt2, module: higher order operators, module: pt2-dispatcher, module: flex attention

#151893 - DISABLED test_builtin_score_mods_different_block_size_float32_score_mod0_BLOCK_SIZE3_cuda_float32 (__main__.TestFlexAttentionCUDA)

Issue - State: closed - Opened by pytorch-bot[bot] 7 months ago - 3 comments
Labels: module: flaky-tests, skipped, oncall: pt2, module: higher order operators, module: pt2-dispatcher, module: flex attention

#151892 - DISABLED test_builtin_score_mods_different_block_size_float32_score_mod3_BLOCK_SIZE2_cuda_float32 (__main__.TestFlexAttentionCUDA)

Issue - State: closed - Opened by pytorch-bot[bot] 7 months ago - 2 comments
Labels: module: flaky-tests, skipped, oncall: pt2, module: higher order operators, module: pt2-dispatcher, module: flex attention

#151891 - DISABLED test_skip_odd_keys_bfloat16_cuda_bfloat16 (__main__.TestFlexAttentionCUDA)

Issue - State: closed - Opened by pytorch-bot[bot] 7 months ago - 1 comment
Labels: module: flaky-tests, skipped, oncall: pt2, module: higher order operators, module: pt2-dispatcher, module: flex attention

#151889 - DISABLED test_builtin_score_mods_dynamic_float16_score_mask_mod7_cuda_float16 (__main__.TestFlexAttentionCUDA)

Issue - State: closed - Opened by pytorch-bot[bot] 7 months ago - 3 comments
Labels: module: flaky-tests, skipped, oncall: pt2, module: higher order operators, module: pt2-dispatcher, module: flex attention

#151365 - flex_attention error in torch.compile

Issue - State: closed - Opened by jjh42 7 months ago - 2 comments
Labels: oncall: pt2, module: pt2-dispatcher, module: flex attention

#151301 - DISABLED test_fake_registration (__main__.TestOpProfiles)

Issue - State: closed - Opened by pytorch-bot[bot] 7 months ago - 14 comments
Labels: triaged, module: flaky-tests, module: custom-operators, skipped, oncall: pt2, module: pt2-dispatcher

#151281 - DISABLED test_duplicate_registration_impl (__main__.TestOpProfiles)

Issue - State: open - Opened by pytorch-bot[bot] 7 months ago - 13 comments
Labels: triaged, module: flaky-tests, module: custom-operators, skipped, oncall: pt2, module: pt2-dispatcher

#151188 - Make auto_functionalize HOPs cacheable

Issue - State: closed - Opened by zou3519 7 months ago
Labels: oncall: pt2, module: higher order operators, module: pt2-dispatcher, dynamo-triage-jan2025

#150056 - FlexAttention inductor tensor has no attribute `get_dtype`

Issue - State: closed - Opened by tsengalb99 8 months ago - 2 comments
Labels: needs reproduction, triaged, oncall: pt2, module: higher order operators, module: pt2-dispatcher, module: flex attention

#149960 - attn_implementation="eager" Buggy on Blackwell

Issue - State: closed - Opened by Oseltamivir 8 months ago - 4 comments
Labels: high priority, triage review, needs reproduction, oncall: pt2, module: higher order operators, module: pt2-dispatcher, module: flex attention

#149801 - `INTERNAL ASSERT FAILED` in `torch.func.vmap` and `torch.scatter_add`

Issue - State: closed - Opened by vwrewsge 8 months ago - 4 comments
Labels: triaged, module: vmap, oncall: pt2, module: functorch, module: pt2-dispatcher

#149789 - flex_attention create_block_mask() + inductor: integer division or modulo by zero

Issue - State: closed - Opened by rmmr 8 months ago - 1 comment
Labels: triaged, oncall: pt2, module: higher order operators, module: pt2-dispatcher, module: flex attention

#149707 - [aot] mark dynamic activations as maybe dynamic

Pull Request - State: closed - Opened by xmfan 8 months ago - 7 comments
Labels: Merged, topic: not user facing, module: inductor, module: dynamo, ciflow/inductor, module: pt2-dispatcher

#149707 - [aot] mark dynamic activations as maybe dynamic

Pull Request - State: open - Opened by xmfan 8 months ago - 5 comments
Labels: topic: not user facing, module: inductor, module: dynamo, ciflow/inductor, module: pt2-dispatcher

#149617 - Torch nightly `torch-2.8.0.dev20250320` breaks torchcodec

Issue - State: closed - Opened by NicolasHug 8 months ago - 1 comment
Labels: module: custom-operators, bug, oncall: pt2, module: pt2-dispatcher

#149468 - torch.library.opcheck doesn't check strides for CPU Tensors

Issue - State: open - Opened by zou3519 8 months ago - 2 comments
Labels: high priority, module: cpp-extensions, triaged, module: custom-operators, module: library, module: pt2-dispatcher

#149105 - [inductor] post grad graph with scatter_upon_const_tensor lowering is not runnable

Issue - State: closed - Opened by xmfan 8 months ago - 2 comments
Labels: triaged, oncall: pt2, module: aotdispatch, module: pt2-dispatcher

#149006 - Missing Additional documentation in autograd.py

Issue - State: closed - Opened by Jason1ien 8 months ago - 3 comments
Labels: module: docs, triaged, module: library, oncall: pt2, module: pt2-dispatcher

#148977 - module.cuda() doesn't work under FakeTensorMode

Issue - State: open - Opened by bdhirsh 8 months ago - 4 comments
Labels: module: nn, triaged, oncall: pt2, module: fakeTensor, module: pt2-dispatcher

#148889 - DISABLED test_make_closure_dynamic_shapes (__main__.DynamicShapesHigherOrderOpTests)

Issue - State: closed - Opened by pytorch-bot[bot] 8 months ago - 3 comments
Labels: triaged, module: flaky-tests, skipped, module: higher order operators, module: pt2-dispatcher

#148835 - [Inductor] Error detected in ReluBackward0

Issue - State: closed - Opened by Cookiee235 9 months ago - 1 comment
Labels: oncall: pt2, module: aotdispatch, module: pt2-dispatcher

#148827 - Inductor may permute inputs to flex attention, leading to assertion error

Issue - State: closed - Opened by Aleko2286 9 months ago - 6 comments
Labels: triaged, oncall: pt2, module: inductor, module: higher order operators, module: pt2-dispatcher, module: flex attention

#148711 - Flex attention significantly slower than SDPA

Issue - State: closed - Opened by nikonikolov 9 months ago - 5 comments
Labels: triaged, oncall: pt2, module: higher order operators, module: pt2-dispatcher, module: flex attention, module: sdpa

#148701 - aot_eager produces wrong output with all_gather_tensor_autograd

Issue - State: open - Opened by eellison 9 months ago - 2 comments
Labels: high priority, oncall: distributed, triaged, actionable, module: correctness (silent), oncall: pt2, module: pt2-dispatcher

#148356 - Inductor layout constraints for custom operators changed from 2.5->2.6, breaking BC

Issue - State: closed - Opened by zou3519 9 months ago
Labels: module: custom-operators, oncall: pt2, module: pt2-dispatcher

#148244 - [inductor] `nn.Upsample-torch.linalg.lu_factor` outputs inconsistent results with eager

Issue - State: closed - Opened by shaoyuyoung 9 months ago - 4 comments
Labels: high priority, triaged, module: correctness (silent), oncall: pt2, module: aotdispatch, module: pt2-dispatcher

#148094 - DISABLED test_split_dynamic (__main__.AutoFunctionalizeTests)

Issue - State: closed - Opened by pytorch-bot[bot] 9 months ago - 6 comments
Labels: triaged, module: flaky-tests, skipped, module: functionalization, oncall: pt2, module: pt2-dispatcher

#148080 - DISABLED test_split (__main__.AutoFunctionalizeTests)

Issue - State: closed - Opened by pytorch-bot[bot] 9 months ago - 7 comments
Labels: triaged, module: flaky-tests, skipped, module: functionalization, oncall: pt2, module: pt2-dispatcher

#148067 - DISABLED test_slice_dynamic (__main__.AutoFunctionalizeTests)

Issue - State: closed - Opened by pytorch-bot[bot] 9 months ago - 5 comments
Labels: high priority, triaged, module: flaky-tests, skipped, module: functionalization, oncall: pt2, module: pt2-dispatcher

#148035 - DISABLED test_slice (__main__.AutoFunctionalizeTests)

Issue - State: closed - Opened by pytorch-bot[bot] 9 months ago - 6 comments
Labels: triaged, module: flaky-tests, skipped, module: functionalization, oncall: pt2, module: pt2-dispatcher

#147822 - `AssertionError: Mixing fake modes NYI` in FakeTensorMode context

Issue - State: closed - Opened by default1360 9 months ago - 2 comments
Labels: triaged, oncall: pt2, module: fakeTensor, module: pt2-dispatcher

#147701 - Compiled `flex_attention` assuming wrong output tensor shape

Issue - State: closed - Opened by mauriceweiler 9 months ago - 5 comments
Labels: triaged, oncall: pt2, module: higher order operators, module: pt2-dispatcher, module: flex attention

#147565 - [dynamo][checkpoint] non-reentrant checkpoint + ambient saved tensor hooks is silently incorrect

Issue - State: closed - Opened by xmfan 9 months ago - 1 comment
Labels: module: activation checkpointing, triaged, oncall: pt2, module: dynamo, module: higher order operators, module: pt2-dispatcher

#147551 - FlexAttention compiled has illegal memory access or device-side assert even though all tensors are contiguous

Issue - State: open - Opened by leijurv 9 months ago - 6 comments
Labels: triaged, oncall: pt2, module: higher order operators, module: pt2-dispatcher, module: flex attention

#147468 - [Triton] [Upstream] FlexDecoding Test Failures opToStageAndCluster[&op].first < numStages && "Op with invalid stage!

Issue - State: closed - Opened by drisspg 9 months ago - 6 comments
Labels: triaged, oncall: pt2, upstream triton, module: higher order operators, module: pt2-dispatcher, module: flex attention

#147463 - Performance regression on modded-nanogpt torch-2.7.0.dev20250208→torch-2.7.0.dev20250209

Issue - State: closed - Opened by YouJiacheng 9 months ago - 16 comments
Labels: triaged, oncall: pt2, module: inductor, module: higher order operators, module: pt2-dispatcher, module: flex attention

#147405 - dynamo should recompile with constant tensors that use ambient device guards

Issue - State: closed - Opened by jamesjwu 9 months ago - 3 comments
Labels: triaged, actionable, oncall: pt2, module: aotdispatch, module: pt2-dispatcher, dynamo-must-fix

#147336 - Investigate FlexAttention performance degradation on low precision inputs

Issue - State: closed - Opened by danielvegamyhre 9 months ago - 4 comments
Labels: triaged, oncall: pt2, upstream triton, module: higher order operators, module: pt2-dispatcher, module: flex attention

#147268 - flex_attention throws `CUDA error: an illegal memory access was encountered`

Issue - State: closed - Opened by mauriceweiler 9 months ago - 1 comment
Labels: triaged, oncall: pt2, module: higher order operators, module: pt2-dispatcher, module: flex attention

#146686 - Unexpected specialization during estimate_runtime

Issue - State: closed - Opened by laithsakka 10 months ago - 3 comments
Labels: triaged, oncall: pt2, module: dynamic shapes, module: aotdispatch, module: pt2-dispatcher

#146624 - [Flex Attention] Cannot determine truth value of Relational

Issue - State: closed - Opened by alexdremov 10 months ago - 2 comments
Labels: triaged, oncall: pt2, module: pt2-dispatcher, module: flex attention

#146212 - [aot_eager] retain_grad is ignored

Issue - State: closed - Opened by anijain2305 10 months ago - 2 comments
Labels: high priority, triage review, triaged, oncall: pt2, module: pt2-dispatcher

#145481 - [custom ops] [2.7 nightly] custom ops with typing.List breaks when importing annotations from future

Issue - State: closed - Opened by dabeschte 10 months ago - 3 comments
Labels: high priority, triaged, module: custom-operators, oncall: pt2, module: pt2-dispatcher

#145460 - Flex Attention not support score_mod with gradients

Issue - State: closed - Opened by LoserCheems 10 months ago - 2 comments
Labels: triaged, oncall: pt2, module: higher order operators, module: pt2-dispatcher, module: flex attention

#145081 - partitioner hangs for some long chains of ops with many users

Issue - State: closed - Opened by bdhirsh 10 months ago
Labels: high priority, triaged, oncall: pt2, module: aotdispatch, module: pt2-dispatcher

#144912 - DISABLED test_flex_attention (__main__.TestCompiledAutograd)

Issue - State: closed - Opened by jeffdaily 10 months ago - 1 comment
Labels: module: rocm, triaged, skipped, oncall: pt2, module: higher order operators, module: compiled autograd, module: pt2-dispatcher, module: flex attention

#144376 - torch.compile post_accumulate_grad_hook ordering is wrong for tiebreakers

Issue - State: open - Opened by xmfan 11 months ago
Labels: triaged, oncall: pt2, module: aotdispatch, module: pt2-dispatcher

#144142 - [torch.compile] Errors on autograd.Function forward returns non-Tensor

Issue - State: closed - Opened by yanboliang 11 months ago
Labels: triaged, oncall: pt2, module: higher order operators, module: pt2-dispatcher, dynamo-autograd-function

#144095 - activation memory budget partitioner can fail with unbacked symints

Issue - State: closed - Opened by bdhirsh 11 months ago - 2 comments
Labels: high priority, triaged, oncall: pt2, module: dynamic shapes, module: aotdispatch, module: pt2-dispatcher

#144072 - Compile error for custom op with optional mutable tensor list argument

Issue - State: closed - Opened by jerrychenhf 11 months ago - 3 comments
Labels: triaged, module: custom-operators, module: functionalization, oncall: pt2, module: aotdispatch, module: pt2-dispatcher

#143832 - flex_attention: OutOfResources

Issue - State: closed - Opened by rmmr 11 months ago - 1 comment
Labels: triaged, oncall: pt2, module: higher order operators, module: pt2-dispatcher, module: flex attention

#143788 - Flex attention with nested tensors, bug in `create_nested_block_mask`

Issue - State: closed - Opened by VivekPanyam 11 months ago - 2 comments
Labels: triaged, module: nestedtensor, oncall: pt2, module: higher order operators, module: pt2-dispatcher, module: flex attention

#143192 - onnx export with dynamo using torch.cond for dynamic models throws error

Issue - State: closed - Opened by ionymikler 11 months ago - 3 comments
Labels: module: onnx, triaged, oncall: pt2, module: higher order operators, export-triage-review, oncall: export, module: pt2-dispatcher

#143180 - Dynamo graph break on FlexAttentionBackwardHOP

Issue - State: closed - Opened by xmfan 11 months ago - 2 comments
Labels: triaged, oncall: pt2, module: higher order operators, module: compiled autograd, module: pt2-dispatcher, module: flex attention

#143163 - Cannot compile a block that contains Flex attention without graph breaks

Pull Request - State: closed - Opened by nicolas-dufour 11 months ago - 2 comments
Labels: triaged, oncall: pt2, module: higher order operators, module: pt2-dispatcher, module: flex attention

#143053 - test_custom_scan_op regression

Issue - State: closed - Opened by eellison 11 months ago - 5 comments
Labels: triaged, bug, oncall: pt2, module: dynamic shapes, module: higher order operators, module: pt2-dispatcher

#142817 - Significant Backward Pass Scaling Bottleneck in Flex Attention: 12x Slower than Forward Pass at 32K Sequence Length

Issue - State: closed - Opened by cora-codes 11 months ago - 6 comments
Labels: module: performance, triaged, oncall: pt2, module: higher order operators, module: pt2-dispatcher, module: flex attention

#142367 - [Break XPU][Inductor UT] The cpu test case `TestFlexAttention::test_GQA_float32_score_mod6` failed on XPU server.

Issue - State: closed - Opened by etaf 12 months ago - 2 comments
Labels: module: cpu, triaged, oncall: pt2, module: higher order operators, module: pt2-dispatcher, module: flex attention

#142313 - [FlexAttention] Using FlexAttention with compiled_autograd complains "Cannot reenter make_fx"

Issue - State: closed - Opened by YouJiacheng 12 months ago - 1 comment
Labels: triaged, module: ddp, oncall: pt2, module: higher order operators, module: compiled autograd, module: pt2-dispatcher, module: flex attention

#141486 - FlexAttention with compiled block mask is slow when varying sequence lengths

Issue - State: closed - Opened by samvanstroud 12 months ago - 5 comments
Labels: triaged, module: nestedtensor, oncall: pt2, module: higher order operators, module: pt2-dispatcher, module: flex attention

#141435 - [Flex Attention] Change the semantics of BlockMask/Adjust

Issue - State: closed - Opened by drisspg 12 months ago
Labels: triaged, oncall: pt2, module: higher order operators, module: pt2-dispatcher, module: flex attention

#141187 - [User Empathy Day] Watermark-Anything AOTAutograd crash

Issue - State: open - Opened by williamwen42 12 months ago
Labels: oncall: pt2, module: pt2-dispatcher, empathy-day

#141183 - FlexAttention: `CUDA error: an illegal memory access was encountered`

Issue - State: closed - Opened by ViktorooReps 12 months ago - 9 comments
Labels: bug, oncall: pt2, module: higher order operators, module: pt2-dispatcher, module: flex attention

#141126 - [FlexAttention] Wrong results for simple block-sparse mask

Issue - State: closed - Opened by alexdremov 12 months ago - 6 comments
Labels: module: mta, oncall: pt2, module: higher order operators, module: pt2-dispatcher, module: flex attention

#140940 - Flex attention blocksize

Issue - State: closed - Opened by liuxin99 about 1 year ago - 2 comments
Labels: triaged, oncall: pt2, module: higher order operators, module: pt2-dispatcher, module: flex attention