Dataset Viewer
Auto-converted to Parquet Duplicate
instance_id
string
repo_id
string
repo_url
string
base_commit
string
language
string
setup_commands
list
test_command
string
test_timeout
int64
refactor_type
string
description
string
files
list
task_type
string
categories
list
bobbyyyan__scorch-refactor_dedupe_ops
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
2071e9b
python
[]
/testbed/run_tests.sh
3,000
There is a lot of duplicated code in src/scorch/ops.py. Please fix that (e.g. make things more modular and maintainable). Note: Do not run the test suite to verify your implementation. You may run build commands (e.g., `pip install -e .`) to check that the code compiles, but do not execute any pytest commands. The ful...
[]
refactor
[ "API/Linear Algebra/Matmul variants", "API/Element-wise/Binary arithmetic" ]
bobbyyyan__scorch-refactor_cost_estimator
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
2071e9b
python
[]
/testbed/run_tests.sh
3,000
extract
Extract the cost-model logic from `src/scorch/compiler/scheduler.py` into a separate `CostEstimator` class or module. The three private static methods `Scheduler._compute_comp_cost`, `Scheduler._compute_workspace_cost`, and `Scheduler._compute_transposition_cost` each independently recompute the same derived quantities...
[]
refactor
[ "Scheduler/IR analyses & scalar opts/Dataflow analyses" ]
bobbyyyan__scorch-refactor_cin_collector
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
b061b53
python
[]
/testbed/run_tests.sh
3,000
consolidate
Consolidate the multiple nearly identical CINVisitor-based classes defined inline within methods of cin.py. The classes TensorAccessGetter, ResultTensorAccessCollector, RHSAccessCollector, WorkspaceGetter, IndexVarCollector, and LoopOrderGetter all follow the same pattern of visiting CIN nodes and collecting specific e...
[]
refactor
[ "IR/CIN nodes" ]
bobbyyyan__scorch-refactor_format_converter
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
2071e9b
python
[]
/testbed/run_tests.sh
3,000
extract
Extract the format conversion methods to_dense and to_sparse from stensor.py into a separate FormatConverter class or module. Both methods contain extensive duplicate code for: generating index variables, creating TensorVar definitions, building assignment expressions using exec/eval, lowering CIN to LLIR, generating C...
[]
refactor
[ "Format/Semantic extensions" ]
bobbyyyan__scorch-refactor_lattice_loop_generator
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
2071e9b
python
[]
/testbed/run_tests.sh
3,000
extract
Extract the gen_single_lattice_loop inner function from the get_lattice_loops method in iter_lattice.py into a separate LatticeLoopGenerator class. This function handles multiple cases based on iterator count and result tensor access patterns. The class should encapsulate loop generation state and provide cleaner metho...
[]
refactor
[ "Scheduler/Loop transformations/Reorder & restructure" ]
bobbyyyan__scorch-feature_block_sparse_format
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
2105686711292cd1d1eb2438035b555fa644bf2a
python
[]
/testbed/run_tests.sh
3,000
Add block sparse format support to scorch's compiler, enabling the full CIN->LLIR->C++ pipeline to generate code for block-structured sparse tensors. In scorch's format system, CSR is `[DENSE, COMPRESSED]` and COO is `[COORDINATE, COORDINATE]` - the level types describe per-dimension storage. Block sparse formats like ...
[]
feature
[ "Format/Block & ELL family" ]
bobbyyyan__scorch-feature_elementwise_mul
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
2105686711292cd1d1eb2438035b555fa644bf2a
python
[]
/testbed/run_tests.sh
3,000
Implement element-wise multiplication (`__mul__`) on `STensor` through the full CIN compilation pipeline, following the same pattern as the existing `__add__` implementation in `stensor.py`. The key difference from addition is **intersection semantics**: when multiplying two sparse tensors, the result is non-zero only ...
[]
feature
[ "API/Element-wise/Binary arithmetic" ]
bobbyyyan__scorch-feature_transpose
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
2105686711292cd1d1eb2438035b555fa644bf2a
python
[]
/testbed/run_tests.sh
3,000
Add a `transpose()` method and `.T` property to `STensor` that physically transposes a 2D sparse tensor, reorganizing its storage layout. This is not a lazy/view transpose - it must produce a new `STensor` with properly restructured index arrays. For CSR format (level types `[DENSE, COMPRESSED]`): transposing produces ...
[]
feature
[ "API/Shape & Layout/Transpose & permute" ]
bobbyyyan__scorch-feature_sum_reduction
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
2105686711292cd1d1eb2438035b555fa644bf2a
python
[]
/testbed/run_tests.sh
3,000
Implement `sum(axis=None)` on `STensor` and as a standalone function in `ops.py` that reduces a sparse tensor along specified dimensions. `axis=None` reduces all dimensions to a scalar. `axis=0` sums along rows (producing a row vector). `axis=1` sums along columns (producing a column vector). For a 2D CSR matrix with `...
[]
feature
[ "API/Reductions & Scans/Aggregate" ]
bobbyyyan__scorch-feature_sddmm
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
2105686711292cd1d1eb2438035b555fa644bf2a
python
[]
/testbed/run_tests.sh
3,000
Implement `sddmm(S, A, B)` as a first-class operation in `ops.py`. SDDMM computes the **sampled dense-dense matrix multiply**: `C = S * (A @ B)` where S is a sparse matrix (the "sample" mask), A and B are dense matrices, `*` is element-wise multiply, and @ is matrix multiply. This is a critical primitive in sparse ML -...
[]
feature
[ "API/Linear Algebra/Matmul variants" ]
bobbyyyan__scorch-feature_autograd
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
2105686711292cd1d1eb2438035b555fa644bf2a
python
[]
/testbed/run_tests.sh
3,000
Add automatic differentiation support for scorch's core operations by implementing custom `torch.autograd.Function` subclasses. Currently, `STensor` extends `torch.nn.Module` and has a `requires_grad` flag, but no gradient computation is implemented. Implement: (1) `SparseAddFunction` - for `C = A + B`, the backward pa...
[]
feature
[ "API/ML Primitives/Autograd" ]
bobbyyyan__scorch-feature_unary_ops
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
2105686711292cd1d1eb2438035b555fa644bf2a
python
[]
/testbed/run_tests.sh
3,000
Add compiler-level support for unary operations on sparse tensors. Currently, CIN only supports `BinaryOp` (add, mul, sub, div). Extend the IR and compilation pipeline to support unary operations: (1) Add a `UnaryOp` CIN node with an `Operation` enum including `ABS`, `NEG`, `RELU`, `SQRT`, `EXP`, `LOG`, `TANH`, `SIGMOI...
[]
feature
[ "API/Element-wise/Unary math" ]
bobbyyyan__scorch-feature_conv1d
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
2105686711292cd1d1eb2438035b555fa644bf2a
python
[]
/testbed/run_tests.sh
3,000
Implement `conv1d(input, kernel, padding=0)` in `ops.py` for sparse 1D signals. The mathematical definition is `y[i] = sum_k x[i + k] * w[k]` where `x` is a sparse 1D input signal and `w` is a (typically dense) convolution kernel. This requires **computed index** support in the CIN pipeline. Scorch already has partial ...
[]
feature
[ "API/Convolution & Pooling" ]
bobbyyyan__scorch-feature_ell_format
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
2105686711292cd1d1eb2438035b555fa644bf2a
python
[]
/testbed/run_tests.sh
3,000
Add ELL (ELLPACK) sparse format support to scorch's format system. ELL format stores a fixed number of entries per row (`max_nnz_per_row`), using two 2D arrays: `indices[num_rows][max_nnz]` for column indices and `values[num_rows][max_nnz]` for values. Rows with fewer non-zeros are padded with a sentinel value (e.g., -...
[]
feature
[ "Format/Block & ELL family" ]
bobbyyyan__scorch-feature_getitem_2d
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
2105686711292cd1d1eb2438035b555fa644bf2a
python
[]
/testbed/run_tests.sh
3,000
Add `__getitem__` support to `STensor` for extracting sub-tensors. Implement three indexing modes: (1) **Integer indexing**: `A[i]` extracts row i from a 2D tensor, returning a 1D STensor. For CSR, this is O(1) - just extract the slice `values[crow[i]:crow[i+1]]` and `col_indices[crow[i]:crow[i+1]]`. For COO, filter en...
[]
feature
[ "API/Indexing & Mutation/Read indexing" ]
bobbyyyan__scorch-feature_batched_matmul
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
2105686711292cd1d1eb2438035b555fa644bf2a
python
[]
/testbed/run_tests.sh
3,000
Add support for batched sparse matrix multiplication. A batched sparse tensor is a 3D STensor with shape `(batch_size, M, N)` where the first dimension is dense (the batch dimension) and the remaining dimensions can be sparse. Implement: (1) Extend `STensor` to support a batch dimension - conceptually this is a list of...
[]
feature
[ "API/Linear Algebra/Matmul variants" ]
bobbyyyan__scorch-feature_sparse_softmax
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
2105686711292cd1d1eb2438035b555fa644bf2a
python
[]
/testbed/run_tests.sh
3,000
Implement `sparse_softmax(A, dim=-1)` that computes softmax over rows (or columns) of a sparse matrix, operating only on the non-zero entries. This is critical for sparse attention in transformers and GNNs. The operation is: for each row i, compute `softmax(A[i,:])` over the non-zero entries only, setting zero entries ...
[]
feature
[ "API/ML Primitives/Activations & losses" ]
bobbyyyan__scorch-feature_dia_format
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
2105686711292cd1d1eb2438035b555fa644bf2a
python
[]
/testbed/run_tests.sh
3,000
Add DIA (diagonal) sparse format support. DIA format is designed for banded/diagonal matrices - it stores the matrix as a set of diagonals, each represented by a dense array. Storage consists of: `offsets` (1D array of diagonal offsets, where 0 is the main diagonal, positive is above, negative is below) and `data` (2D ...
[]
feature
[ "Format/Block & ELL family" ]
bobbyyyan__scorch-feature_elementwise_sub
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Implement element-wise subtraction (`__sub__`) on `STensor` through the full CIN compilation pipeline, following the same pattern as the existing `__add__` implementation. The key difference from addition is asymmetry: subtraction uses union semantics (like addition, the result is non-zero where *either* operand is non...
[]
feature
[ "API/Element-wise/Binary arithmetic" ]
bobbyyyan__scorch-feature_fill_value
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Extend the format system to support configurable non-zero fill values (currently hardcoded to 0.0). Changes required: (1) `format.py` - add a configurable `_fill_value` attribute to the `Format` class with a default of 0.0. (2) `stensor.py` - propagate the fill value through format conversions (`to_dense`, `to_sparse`,...
[]
feature
[ "Format/Semantic extensions" ]
bobbyyyan__scorch-feature_outer_product
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Implement a sparse outer product operation `outer(a, b)` in `ops.py` that takes two 1D sparse vectors and produces a 2D sparse matrix. The CIN expression is `C[i,j] = A[i] * B[j]` (equivalent to einsum `"i,j->ij"`). This is a dimension-increasing operation - verify that the iteration lattice correctly handles it with n...
[]
feature
[ "API/Linear Algebra/Tensor products" ]
bobbyyyan__scorch-feature_dtype_int_float64
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Extend the compilation pipeline to support multiple data types beyond float32, specifically int32, int64, and float64. The dtype maps in `utils.py` already define these types but they aren't exercised through the pipeline. Changes required: (1) `stensor.py` - preserve dtype through all operations and format conversions...
[]
feature
[ "API/Type System/Value dtypes" ]
bobbyyyan__scorch-feature_kron
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Implement a sparse Kronecker product operation `kron(A, B)` in `ops.py`. The Kronecker product produces output `C[i*P+p, j*Q+q] = A[i,j] * B[p,q]` where A is MxN, B is PxQ, and C is (M*P)x(N*Q). This requires new CIN IR nodes for computed indices with multiplication - currently only `IndexVarAdd` exists for additive in...
[]
feature
[ "API/Linear Algebra/Tensor products" ]
bobbyyyan__scorch-feature_trace_diagonal
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Implement `trace(A)` and `diagonal(A, offset=0)` operations for sparse tensors. The CIN expression for diagonal extraction is `d[i] = A[i,i]` - the same index variable appears in both positions. This requires 'locate' capability in the iteration lattice: when iterating over the first dimension, the second dimension mus...
[]
feature
[ "API/Linear Algebra/Tensor products" ]
bobbyyyan__scorch-feature_norm
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Implement `norm(A, dim, ord=2)` for computing per-row or per-column norms of sparse tensors. This requires: (1) UnaryOp lowering - `abs` and `sqrt` are defined in the CIN IR but not lowered through the compilation pipeline. Implement their lowering in `cin_lowerer.py` and `codegen.py` to emit the corresponding C++ math...
[]
feature
[ "API/Reductions & Scans/Aggregate" ]
bobbyyyan__scorch-feature_kernel_fusion
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Implement kernel fusion for chained sparse operations. When multiple operations are composed (e.g., `C = (A @ B) + D`), currently each operation compiles and executes a separate C++ kernel with intermediate materialization. Implement a fusion system that compiles chained operations into a single C++ kernel. Design: (1)...
[]
feature
[ "Scheduler/Loop transformations/Fusion" ]
bobbyyyan__scorch-feature_openmp_parallel
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Extend the code generation pipeline to emit OpenMP parallel directives for auto-generated sparse kernels. Changes required: (1) Add a `ParallelFor` node to the LLIR (Low-Level IR) to represent parallelizable loops. (2) In `cin_lowerer.py`, analyze loop nests for parallelizability - the outermost loop over a dense dimen...
[]
feature
[ "Codegen/Parallelism" ]
bobbyyyan__scorch-feature_cat_stack
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Implement sparse tensor concatenation `cat(tensors, dim)` and stacking `stack(tensors, dim)` in `ops.py`. These are structural operations that operate on storage arrays directly (not through the CIN pipeline). For CSR concatenation along dim=0: concatenate `crow_indices` arrays (with appropriate offset adjustments), co...
[]
feature
[ "API/Shape & Layout/Concat & pad" ]
bobbyyyan__scorch-feature_tensor_factories
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Implement sparse tensor creation utility factory class methods on `STensor`: (1) `STensor.eye(n, format)` - create an nxn identity matrix in the specified sparse format. (2) `STensor.diag(values, format)` - create a diagonal matrix from a 1D tensor of values. (3) `STensor.rand_sparse(shape, density, format)` - create a...
[]
feature
[ "API/Constructors & I/O/Factories" ]
bobbyyyan__scorch-feature_scipy_interop
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Implement SciPy sparse matrix interoperability for `STensor`. Add two methods: (1) `STensor.from_scipy(scipy_matrix)` - construct an STensor from a SciPy sparse matrix, auto-detecting the format (CSR, CSC, or COO). Handle format mapping: scipy's `indptr`/`indices`/`data` arrays map to scorch's `crow_indices`/`col_indic...
[]
feature
[ "API/Constructors & I/O/External I/O" ]
bobbyyyan__scorch-feature_coalesce_coo
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Implement a `coalesce()` method for COO-format sparse tensors that sorts coordinates lexicographically and sums duplicate entries. Add an `is_coalesced` property that returns whether the tensor's coordinates are already sorted and free of duplicates. Support N-dimensional COO tensors (not just 2D). Add an optional `rem...
[]
feature
[ "API/Indexing & Mutation/Canonicalization" ]
bobbyyyan__scorch-feature_setitem
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Implement `__setitem__` on `STensor` for mutating sparse tensor entries in-place. Support four indexing modes: (1) Scalar assignment `A[i, j] = value` - for COO, search for existing entry and update or append; for CSR, search within the row's column range and update or insert with `crow_indices` adjustment; for dense, ...
[]
feature
[ "API/Indexing & Mutation/Write & mutation" ]
bobbyyyan__scorch-feature_bsr_block_matmul
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Implement a hybrid block-sparse computation path that delegates dense block arithmetic to PyTorch dense operators. Add 2D block-structured constructors/converters on `STensor`: `from_bsr(crow_indices, col_indices, values, block_size, shape)` and `to_bsr(block_size)` where values has shape `(nnz_blocks, block_h, block_w...
[]
feature
[ "API/Linear Algebra/Matmul variants" ]
bobbyyyan__scorch-feature_csc_format
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Add native CSC (Compressed Sparse Column) support across `STensor` and ops. Implement `STensor.from_csc(ccol_indices, row_indices, values, shape)` and `STensor.to_csc()`, and extend `STensor.from_torch` to ingest `torch.sparse_csc` tensors. Update format parsing/validation so CSC is representable explicitly (for exampl...
[]
feature
[ "Format/Block & ELL family" ]
bobbyyyan__scorch-feature_pytorch_broadcasting
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Implement full PyTorch-style broadcasting for sparse elementwise binary operations. Today `STensor.__add__` explicitly says broadcasting is TODO, and `__mul__`/`__sub__` paths do not provide broadcast semantics. Add shared broadcast shape inference and index mapping utilities (including implicit leading dimensions, sin...
[]
feature
[ "API/Constructors & I/O/Broadcasting" ]
bobbyyyan__scorch-feature_conv2d
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Implement sparse `conv2d` in `ops.py` for 2D inputs and kernels, analogous to the existing `conv1d` task but with full 2D indexing. Support arguments `stride`, `padding`, and `dilation` (start with `groups=1`), and allow sparse input with dense kernel as the primary path. Lower through CIN/LLIR when feasible, with clea...
[]
feature
[ "API/Convolution & Pooling" ]
bobbyyyan__scorch-feature_singleton_level
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Add full `LevelType.SINGLETON` support throughout the sparse compiler pipeline. `LevelType.SINGLETON` already exists in `format.py`, but parser/iterator/lowering paths do not currently implement execution semantics. Define singleton semantics as one coordinate per parent position, then implement: (1) format parsing in ...
[]
feature
[ "Format/Compressed-style levels" ]
bobbyyyan__scorch-feature_addmm_linear
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Implement fused sparse linear algebra front-end ops `addmm` and `linear`. Add `ops.addmm(input, mat1, mat2, beta=1.0, alpha=1.0)` with semantics matching `torch.addmm`: `beta * input + alpha * (mat1 @ mat2)`, supporting sparse `mat1` with dense or sparse `mat2` where valid. Add `ops.linear(input, weight, bias=None)` as...
[]
feature
[ "API/Linear Algebra/Matmul variants" ]
bobbyyyan__scorch-feature_incremental_insert
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Implement incremental sparse tensor construction APIs centered on the currently unimplemented `STensor.insert`. Add `insert(indices, values, accumulate=True)` that can append or update entries in COO and CSR tensors without forcing full dense materialization, plus a convenience constructor `STensor.from_entries(indices...
[]
feature
[ "API/Indexing & Mutation/Write & mutation" ]
bobbyyyan__scorch-feature_lifecycle_device
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Add full tensor lifecycle/device support by implementing `STensor.validate()`, `STensor.clone()`, and `STensor.to(device)` (with `cpu()`/`cuda()` convenience behavior). `validate()` should check shape/index/value invariants per format (e.g., CSR `crow_indices` length and monotonicity, coordinate bounds, mode-order cons...
[]
feature
[ "API/Constructors & I/O/Introspection" ]
bobbyyyan__scorch-feature_einsum_ellipsis
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Generalize `ops.einsum` parsing and shape handling to support PyTorch-style ellipsis and implicit output inference. The current implementation assumes single-character explicit index lists without ellipsis. Extend it to parse expressions like `...ij,...jk->...ik`, `bij,bjk->bik`, and implicit-output forms (no `->`) whi...
[]
feature
[ "API/Linear Algebra/Einsum" ]
bobbyyyan__scorch-feature_triangular_solve
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Implement sparse triangular solve support: `ops.triangular_solve(A, B, upper=False, unit_diagonal=False, left=True)` for sparse `A` (CSR primary path, COO via conversion) and dense/sparse right-hand side `B`. Use forward substitution for lower-triangular and backward substitution for upper-triangular systems, including...
[]
feature
[ "API/Linear Algebra/Solvers" ]
bobbyyyan__scorch-feature_semiring_matmul
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Add semiring-based sparse matrix multiplication as a first-class feature. Extend `ops.matmul` (and `einsum` where appropriate) with a `semiring` argument supporting at least: `plus_times` (default arithmetic), `min_plus`, `max_plus`, and `logical_or_and`. This requires extending operation/reduction plumbing in CIN/LLIR...
[]
feature
[ "API/Linear Algebra/Matmul variants" ]
bobbyyyan__scorch-feature_int64_indices
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Implement end-to-end 64-bit sparse index support for large tensors. Today `TensorIndex` coerces indices to `torch.int` (int32), which limits representable coordinates and can overflow for very large shapes. Add index-dtype awareness so mode indices can remain int32 or int64, propagate that through storage, lowering, an...
[]
feature
[ "API/Type System/Index dtypes" ]
bobbyyyan__scorch-feature_multi_output_kernels
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Implement first-class multi-output CIN kernels and use them to add sparse `max`/`min` reductions with optional index returns. `cin_lowerer.py` currently assumes a single result tensor (`TODO: need to handle multiple result tensors` / `TODO: deal with multiple outputs`), which blocks APIs like `torch.max(..., dim=...)` ...
[]
feature
[ "API/Reductions & Scans/Argmax-style", "IR/CIN nodes" ]
bobbyyyan__scorch-feature_zero_copy_views
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Implement zero-copy tensor window views using the existing `Window` and `TensorStorageView` scaffolding. Add `STensor.narrow(dim, start, length)` and `STensor.slice_view(offset, shape, step)` that return lightweight views without copying storage whenever possible. Dense views should alias the base value buffer; CSR row...
[]
feature
[ "API/Shape & Layout/Views" ]
bobbyyyan__scorch-feature_torch_sparse_roundtrip
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Add sparse-native PyTorch round-trip APIs without forcing dense conversion. Keep existing `to_torch()` behavior for dense export, and add `STensor.to_torch_sparse(layout='coo'|'csr')` to emit PyTorch sparse tensors directly from stored indices/values. Extend `STensor.from_torch` to robustly ingest both 2D and batched (...
[]
feature
[ "API/Constructors & I/O/External I/O" ]
bobbyyyan__scorch-feature_comparisons_where
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Implement comparison and masking operations for sparse tensors. Add elementwise comparisons (`==`, `!=`, `<`, `<=`, `>`, `>=`) for STensor-STensor and STensor-scalar inputs, returning boolean STensors, plus `ops.where(mask, x, y)` and `STensor.masked_fill(mask, value)`. Extend CIN/LLIR lowering to emit comparison expre...
[]
feature
[ "API/Element-wise/Comparison & predicate" ]
bobbyyyan__scorch-feature_dropout
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Implement sparse dropout for training workflows. Add `ops.dropout(input, p=0.5, training=True, inplace=False, generator=None)` and `STensor.dropout(...)`. For sparse inputs, sample Bernoulli masks over explicitly stored values only, scale retained values by `1/(1-p)` in training mode, and preserve index structure (with...
[]
feature
[ "API/ML Primitives/Regularization" ]
bobbyyyan__scorch-feature_kernel_cache
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Introduce a robust content-addressed kernel compilation cache for generated C++ kernels. Current compilation paths repeatedly call `load_inline(name='kernel', ...)` and cache mostly by CIN string, which can cause redundant compilations and potential collisions across dtype/format/mode-order variants. Add a `KernelCache...
[]
feature
[ "Runtime/Caching & dispatch" ]
bobbyyyan__scorch-feature_affine_gather_scatter
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Generalize computed indexing in CIN/LLIR and expose it through sparse gather/scatter APIs. In `cin.py`, `IndexVarExpr` currently only has `IndexVarAdd(lhs: IndexVar, rhs: IndexVar)`, which blocks common affine forms like `i + 1`, `i - k`, and `i * stride + offset`. Extend the IR with literal constants and affine index ...
[]
feature
[ "API/Indexing & Mutation/Computed indexing" ]
bobbyyyan__scorch-feature_spgemm_symbolic
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Implement symbolic+numeric SpGEMM planning with reusable sparsity structure for repeated sparse matmul. Today `ops.matmul` computes structure and values together each call. Add a symbolic planning path for CSR/COO multiplication that computes only output index structure and metadata once (for example row pointers and c...
[]
feature
[ "API/Linear Algebra/Matmul variants" ]
bobbyyyan__scorch-feature_torch_function_dispatch
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Add PyTorch operator dispatch integration for `STensor` so `torch.*` calls route to scorch implementations automatically. Implement `STensor.__torch_function__` (or an equivalent dispatch layer) for core ops already supported by scorch, including `torch.matmul`, `torch.einsum`, `torch.add`, `torch.sub`, and `torch.mul`...
[]
feature
[ "API/Constructors & I/O/Torch dispatch" ]
bobbyyyan__scorch-feature_serialization
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Implement robust serialization and checkpoint round-tripping for sparse tensors. Add `STensor.to_dict()` / `STensor.from_dict()` plus convenience `save(path)` / `load(path)` helpers that preserve shape, dtype, index dtype, format, mode_order, values, and all mode indices without dense conversion. Ensure compatibility w...
[]
feature
[ "API/Constructors & I/O/Serialization" ]
bobbyyyan__scorch-feature_schedule_autotuner
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Add runtime schedule autotuning for generated sparse kernels. Scorch already has scheduling hooks (`Scheduler.auto_schedule`, tiling support, mode-order changes), but kernel choice is mostly heuristic and static. Implement an autotuner that explores candidate schedules (loop order, tiling factors, workspace choices, an...
[]
feature
[ "Runtime/Tuning & user control" ]
bobbyyyan__scorch-feature_reshape_flatten
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Implement sparse shape-transformation APIs without dense materialization: `STensor.reshape(*shape)`, `STensor.flatten(start_dim=0, end_dim=-1)`, and `STensor.unflatten(dim, sizes)`, plus `ops.reshape` convenience wrappers. The transformation must preserve values and remap indices exactly (including negative dimensions ...
[]
feature
[ "API/Shape & Layout/Reshape" ]
bobbyyyan__scorch-feature_pooling_2d
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Add sparse pooling operators for vision workloads: `ops.max_pool2d` and `ops.avg_pool2d` with parameters `kernel_size`, `stride=None`, `padding=0`, `dilation=1`, and `ceil_mode=False`, plus corresponding `STensor` method wrappers. Support NCHW inputs where activations are sparse and missing entries represent fill value...
[]
feature
[ "API/Convolution & Pooling" ]
bobbyyyan__scorch-feature_fp16_bf16_dtype
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Extend dtype support to include `torch.float16` and `torch.bfloat16` end-to-end in CIN->LLIR->C++ execution, including cached kernels in `csrc`. Add dtype mappings in `utils.py`/`llir.py`, ensure generated C++ uses correct scalar and torch dtype constants, and implement mixed-precision accumulation controls (e.g., `acc...
[]
feature
[ "API/Type System/Value dtypes" ]
bobbyyyan__scorch-feature_user_scheduling_api
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Expose a user-controlled scheduling API for CIN execution instead of relying solely on `Scheduler.auto_schedule`. Add a schedule object or kwargs surface (for `ops.einsum`, `ops.matmul`, and `lower_and_exec_cin`) that can explicitly set loop order, tile sizes per index variable, workspace insertion policy (dense/coo/no...
[]
feature
[ "Runtime/Tuning & user control" ]
bobbyyyan__scorch-feature_matrix_market_io
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Add scientific sparse I/O interoperability for external datasets: Matrix Market (`.mtx`) and SciPy sparse NPZ (`.npz`) round-trips. Implement `STensor.from_matrix_market(path)`, `STensor.to_matrix_market(path)`, `STensor.from_scipy_npz(path)`, and `STensor.to_scipy_npz(path)` with optional dtype/index dtype controls an...
[]
feature
[ "API/Constructors & I/O/External I/O" ]
bobbyyyan__scorch-feature_n_m_sparsity
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Implement semi-structured N:M sparsity support with an initial optimized path for 2:4 sparsity. Add constructors/converters on `STensor` (e.g., `from_semi_structured(values, metadata, pattern=(2,4), dim=-1)` and `to_semi_structured(pattern=(2,4), dim=-1)`), plus validation that each 4-element group has exactly 2 stored...
[]
feature
[ "Format/Block & ELL family" ]
bobbyyyan__scorch-feature_topk_kthvalue
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Implement sparse `topk`/`kthvalue` along a specified dimension: `ops.topk(input, k, dim=-1, largest=True, sorted=True)` and `ops.kthvalue(input, k, dim=-1, keepdim=False)` plus `STensor` method wrappers. Semantics should align with PyTorch, including how implicit fill values (zeros) compete with explicit non-zero entri...
[]
feature
[ "API/Reductions & Scans/Argmax-style" ]
bobbyyyan__scorch-feature_simd_vectorization
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Add SIMD-aware vectorization to generated C++ kernels for dense innermost loops. Extend LLIR/codegen so when an inner loop is contiguous and arithmetic-only, emitted code uses vectorization-friendly constructs (`#pragma omp simd` and/or architecture-gated intrinsics) with safe scalar fallbacks. Ensure correctness for r...
[]
feature
[ "Codegen/Vectorization" ]
bobbyyyan__scorch-feature_complex_dtype
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Extend scorch with end-to-end complex dtype support (`torch.complex64` and `torch.complex128`) across `STensor`, CIN->LLIR lowering, and generated C++ kernels. The current dtype plumbing in `llir.py`/codegen and cached native kernels is real-valued; add complex scalar mappings, kernel argument marshalling, and correct ...
[]
feature
[ "API/Type System/Value dtypes" ]
bobbyyyan__scorch-feature_cuda_backend
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Add a true CUDA codegen backend for CIN-generated kernels so sparse workloads can execute on GPU without host fallback. Today execution is centered around CPU C++ `load_inline`; extend lowering/codegen to emit CUDA-compatible kernels (or CUDA-specialized C++ with `cuda_sources`) and runtime dispatch that chooses CPU or...
[]
feature
[ "Codegen/Backend targets" ]
bobbyyyan__scorch-feature_bitmap_level
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Add bitmap sparse level support to the format system (`LevelType.BITMAP`) and compiler pipeline. A bitmap level stores a dense occupancy bitset plus compacted values, which is efficient for near-dense regions and predictable iteration. Implement bitmap parsing/serialization in `format.py`, storage representation in `Te...
[]
feature
[ "Format/Compressed-style levels" ]
bobbyyyan__scorch-feature_hyb_format
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Implement HYB (ELL+COO) sparse format support for matrices with skewed row densities. HYB stores up to `ell_width` entries per row in an ELL component and spills overflow entries into a COO tail. Add format/storage support plus `STensor.from_hyb(ell_indices, ell_values, coo_indices, coo_values, shape)` and `STensor.to_...
[]
feature
[ "Format/Block & ELL family" ]
bobbyyyan__scorch-feature_csf_format
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Add CSF (Compressed Sparse Fiber) format support for higher-order sparse tensors (3D+), enabling efficient tensor contractions without flattening to COO. Implement hierarchical compressed storage where each sparse mode contributes position/coordinate arrays (generalizing CSR to multiple sparse levels). Extend `TensorFo...
[]
feature
[ "Format/Hierarchical & multi-d" ]
bobbyyyan__scorch-feature_sparse_attention
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Implement fused sparse scaled-dot-product attention for block/coordinate masks: `ops.scaled_dot_product_attention_sparse(Q, K, V, attn_mask_sparse=None, dropout_p=0.0, is_causal=False, training=False)`. The key requirement is to avoid dense `QK^T` materialization: use sparse mask structure to compute only sampled score...
[]
feature
[ "API/ML Primitives/Attention & embedding" ]
bobbyyyan__scorch-feature_int8_quantized
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Add int8 quantized sparse inference as a first-class path. Implement value-only quantization APIs on `STensor` (`quantize_per_tensor`, `quantize_per_channel`, `dequantize`, and `from_torch_quantized`) that preserve sparse index structures while quantizing stored values. Add `ops.matmul_quantized(A, B, bias=None, out_dt...
[]
feature
[ "API/ML Primitives/Quantization" ]
bobbyyyan__scorch-feature_iterative_solvers
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Implement iterative sparse linear solvers for scientific workloads. Add `ops.cg(A, b, x0=None, tol=1e-6, maxiter=None, M=None)` for symmetric positive definite systems and `ops.bicgstab(A, b, x0=None, tol=1e-6, maxiter=None, M=None)` for general non-symmetric systems, with optional Jacobi preconditioning. Reuse existin...
[]
feature
[ "API/Linear Algebra/Solvers" ]
bobbyyyan__scorch-feature_layer_rms_norm
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Add sparse normalization operators for transformer-style models: `ops.layer_norm_sparse` and `ops.rms_norm_sparse`, plus `STensor.layer_norm(...)` and `STensor.rms_norm(...)`. Support affine parameters (`weight`, `bias`) and epsilon controls with semantics matching PyTorch over full normalized dimensions, where implici...
[]
feature
[ "API/ML Primitives/Normalization" ]
bobbyyyan__scorch-feature_einsum_repeated_idx
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Generalize `ops.einsum` to fully support repeated-index semantics within a single operand (diagonal extraction/trace-style behavior) without dense fallback. The current parser/scheduling logic in `ops.py` assumes effectively unique per-operand indices and does not robustly handle expressions like `ii->i`, `bijj->bi`, o...
[]
feature
[ "API/Linear Algebra/Einsum" ]
bobbyyyan__scorch-feature_prune_eliminate_zeros
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Implement explicit-zero management and structured pruning APIs for sparse tensors. Add `STensor.eliminate_zeros(inplace=False, atol=0.0)` to remove stored zeros/near-zeros and rebuild indices correctly for COO/CSR, plus `STensor.prune(threshold=None, topk=None, dim=None, keep_structure=False)` for magnitude-based pruni...
[]
feature
[ "API/Indexing & Mutation/Canonicalization" ]
bobbyyyan__scorch-feature_transpose_matmul
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Add transpose-aware matmul APIs that avoid physical tensor transposition. Extend `ops.matmul` with flags `transpose_a=False` and `transpose_b=False` (and matching `STensor.matmul` kwargs) so callers can request `A^T @ B`, `A @ B^T`, or `A^T @ B^T` directly. Implement this via index remapping in CIN/einsum lowering rath...
[]
feature
[ "API/Linear Algebra/Matmul variants" ]
bobbyyyan__scorch-feature_log_softmax_nll
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Add sparse log-probability training primitives: `ops.sparse_log_softmax(input, dim=-1)` and `ops.sparse_nll_loss(log_probs, target, reduction='mean', ignore_index=-100)`, with `STensor` method wrappers. Build on sparse softmax infrastructure but compute and return log probabilities directly for numerical stability, and...
[]
feature
[ "API/ML Primitives/Activations & losses" ]
bobbyyyan__scorch-feature_block_diag
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Implement block-diagonal sparse packing utilities for variable-size mini-batch workloads. Add `STensor.from_block_diag(tensors)` to pack a list of 2D sparse tensors into a single block-diagonal sparse matrix and `STensor.to_block_diag(block_sizes)` to unpack. Add `ops.block_diag_matmul(A_blockdiag, X, block_sizes)` to ...
[]
feature
[ "API/Shape & Layout/Concat & pad" ]
bobbyyyan__scorch-feature_elementwise_div
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Implement element-wise division (`__truediv__` and `__rtruediv__`) on `STensor` through the full CIN compilation pipeline. `Operation.DIV` already exists in `src/scorch/compiler/cin.py` (line 897: `DIV = "/"`), `AssignOp.DIV_ASSIGN` exists in `src/scorch/compiler/llir.py` (line 81), and `IndexExpr.__sub__` at line 143 ...
[]
feature
[ "API/Element-wise/Binary arithmetic" ]
bobbyyyan__scorch-feature_elementwise_pow
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Implement element-wise power (`__pow__`) on `STensor` that raises each stored value to a given exponent. Unlike add/sub/mul/div, power is not a binary operation between two equally-shaped sparse tensors in the typical case -- the primary use case is `A ** n` where `n` is a scalar (integer or float). This requires a new...
[]
feature
[ "API/Element-wise/Binary arithmetic" ]
bobbyyyan__scorch-feature_matrix_power
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Implement `ops.matrix_power(A, n)` that computes the n-th matrix power of a square sparse matrix `A` by repeated matrix multiplication, reusing the existing `ops.matmul` infrastructure. This is a higher-level operation that does not require new CIN primitives but does require careful handling of sparse format propagati...
[]
feature
[ "API/Linear Algebra/Tensor products" ]
bobbyyyan__scorch-feature_cholesky
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Implement sparse Cholesky factorization for symmetric positive definite (SPD) sparse matrices: `ops.cholesky(A, upper=False)` that returns a sparse lower-triangular `L` such that `A = L @ L^T` (or upper-triangular `U` with `A = U^T @ U` when `upper=True`). This is a two-phase algorithm: symbolic factorization (determin...
[]
feature
[ "API/Linear Algebra/Decompositions" ]
bobbyyyan__scorch-feature_eigenvalue_solvers
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Implement sparse eigenvalue computation for finding dominant eigenvalues and eigenvectors without dense materialization. Add two methods: (1) `ops.power_iteration(A, num_iters=100, tol=1e-6)` for finding the largest-magnitude eigenvalue and its eigenvector, and (2) `ops.lanczos(A, k=6, tol=1e-8, maxiter=None)` for comp...
[]
feature
[ "API/Linear Algebra/Decompositions" ]
bobbyyyan__scorch-feature_cumsum_cumprod
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Implement cumulative reduction operations `ops.cumsum(A, dim)` and `ops.cumprod(A, dim)` for sparse tensors along a specified dimension. These are prefix-scan operations where `cumsum(A, dim=1)[i,j] = sum(A[i, 0:j+1])` and `cumprod(A, dim=1)[i,j] = prod(A[i, 0:j+1])`. The key semantic design decision for sparse tensors...
[]
feature
[ "API/Reductions & Scans/Scans & segment" ]
bobbyyyan__scorch-feature_graph_adjacency
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Implement graph adjacency matrix utility functions essential for graph neural network workloads, directly supporting the existing GCN example in `examples/gcn/scorch_gcn.py` which currently requires manual adjacency construction. Add the following functions to `src/scorch/ops.py`: (1) `ops.degree(A, dim=1)` -- compute ...
[]
feature
[ "API/Constructors & I/O/Factories" ]
bobbyyyan__scorch-feature_hash_level
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Add hash-map based sparse level support (`LevelType.HASH`) to the format system and compiler pipeline. Unlike COMPRESSED (CSR-like sorted arrays with O(log n) or O(nnz) lookup) and COORDINATE (COO unsorted coordinate lists), a HASH level provides O(1) amortized random access to individual entries via a hash table mappi...
[]
feature
[ "Format/Compressed-style levels" ]
bobbyyyan__scorch-feature_sparse_dim_tiling
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Implement true sparse-dimension tiling in the CIN scheduler and lowering pipeline. Today `Scheduler.auto_schedule` explicitly removes sparse index vars from tiling and `IndexVar.size_llir_var` assumes a dense access. Extend `Scheduler.add_tile` and `auto_schedule` so COMPRESSED/COORDINATE dimensions can be strip-mined ...
[]
feature
[ "Scheduler/Loop transformations/Tiling" ]
bobbyyyan__scorch-feature_tile_remainder_predication
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Add remainder-safe predication for tiled loops across dense and sparse domains. Current tiling assumes fixed tile-size loop bounds, which can overrun when dimension sizes (or row-local sparse fiber lengths) are smaller than or not divisible by tile size. Introduce per-tile end bounds: dense loops should use `tile_end =...
[]
feature
[ "Scheduler/Loop transformations/Tiling" ]
bobbyyyan__scorch-feature_segmented_sparse_tiling
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Implement segmented sparse tiling for CSR/COO reductions to improve cache locality in sparse matmul kernels. Add a scheduling transformation that tiles sparse reduction dimensions by nonzero-count segments (position-space tiles) inside each parent fiber (for CSR: per-row `p` segments; for COO: per-leading-coordinate bu...
[]
feature
[ "Scheduler/Loop transformations/Tiling" ]
bobbyyyan__scorch-feature_dual_axis_tiling
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Add dual-axis tiling for mixed sparse-dense kernels (for example SpMM `C[i,n] += A[i,k_sparse] * B[k_sparse,n_dense]`). Support applying two independent tile transforms in one schedule: sparse reduction tiling in position space and dense output-column tiling in coordinate space. Extend scheduler validation so multiple ...
[]
feature
[ "Scheduler/Loop transformations/Tiling" ]
bobbyyyan__scorch-feature_nnz_balanced_partition
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Implement nnz-balanced sparse tile partitioning for parallel execution. Row-wise parallelization is often imbalanced on skewed sparse matrices; add an inspector step that partitions work into tiles/blocks with roughly equal nonzero counts instead of equal row counts. For CSR, build block boundaries from cumulative `cro...
[]
feature
[ "Scheduler/Loop transformations/Tiling" ]
bobbyyyan__scorch-feature_workspace_touched_tracking
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
33532a3
python
[]
/testbed/run_tests.sh
3,000
Add sparse-tile workspace optimization with touched-entry tracking to avoid full-tile clears each iteration. For tiled sparse reductions that use dense workspaces, replace unconditional tile-wide initialization/flush/clear with a touched-index list (and optional small bitmap) that records only entries updated in the cu...
[]
feature
[ "Scheduler/Sparse-specific passes/Workspace transforms" ]
bobbyyyan__scorch-feature_repr_str
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
92fb190
python
[]
/testbed/run_tests.sh
3,000
Implement informative `__repr__` and `__str__` methods on `STensor` that replace the current placeholder returning `"Tensor"`. The output must display the tensor's shape, per-mode format annotations (e.g. `[d, s]` for a dense-then-sparse 2D tensor), number of stored non-zeros (`nnz`), density as a percentage, dtype, an...
[]
feature
[ "API/Constructors & I/O/Introspection" ]
bobbyyyan__scorch-feature_metadata_introspection
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
92fb190
python
[]
/testbed/run_tests.sh
3,000
Add a sparse metadata and introspection API to `STensor`. Implement the following: (1) `nnz` property returning the count of explicitly stored non-zero entries. (2) `density` property returning nnz divided by the total number of elements. (3) `sparsity` property returning 1 minus density. (4) `nonzero()` method returni...
[]
feature
[ "API/Constructors & I/O/Introspection" ]
bobbyyyan__scorch-feature_mode_n_product
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
92fb190
python
[]
/testbed/run_tests.sh
3,000
Implement mode-n tensor-matrix product operations in `ops.py`. Add `ops.mode_n_product(X, M, n)` that multiplies an N-dimensional sparse tensor X by a dense matrix M along mode n. The implementation should dynamically generate einsum subscript strings for arbitrary dimensionality rather than hard-coding cases for speci...
[]
feature
[ "API/Linear Algebra/Tensor products" ]
bobbyyyan__scorch-feature_squeeze_unsqueeze
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
92fb190
python
[]
/testbed/run_tests.sh
3,000
Implement `STensor.squeeze(dim=None)` and `STensor.unsqueeze(dim)` for sparse tensors. `squeeze(dim)` removes a dimension of size 1 at the specified position, updating shape, format (removing the corresponding level type), mode_indices, and mode_order. When `dim=None`, squeeze all dimensions of size 1. `unsqueeze(dim)`...
[]
feature
[ "API/Shape & Layout/Reshape" ]
bobbyyyan__scorch-feature_unfold_refold
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
92fb190
python
[]
/testbed/run_tests.sh
3,000
Implement sparse tensor unfolding (matricization) and refolding. Add `STensor.unfold(mode)` that converts an N-dimensional sparse tensor into a 2D matrix by unfolding along the specified mode: the given mode becomes the row dimension and the remaining modes are combined (in order) into the column dimension. For COO for...
[]
feature
[ "API/Shape & Layout/Reshape" ]
bobbyyyan__scorch-feature_lu_decomposition
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
92fb190
python
[]
/testbed/run_tests.sh
3,000
Implement sparse LU decomposition. Add `ops.lu(A, pivoting=True)` that decomposes a 2D sparse matrix A into lower triangular L, upper triangular U, and permutation matrix P such that P @ A = L @ U. Use a left-looking sparse algorithm that processes columns left to right, computing each column of L and U by solving a sp...
[]
feature
[ "API/Linear Algebra/Decompositions" ]
bobbyyyan__scorch-feature_index_select
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
92fb190
python
[]
/testbed/run_tests.sh
3,000
Implement sparse `index_select` for N-dimensional tensors. Add `ops.index_select(input, dim, index)` that selects slices from the sparse tensor `input` along dimension `dim` according to the entries in `index` (a 1D integer tensor). This is distinct from `__getitem__` (which uses int/slice indexing) and from `gather`/`...
[]
feature
[ "API/Indexing & Mutation/Read indexing" ]
bobbyyyan__scorch-feature_expand_repeat
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
92fb190
python
[]
/testbed/run_tests.sh
3,000
Implement sparse `expand` and `repeat` operations on `STensor`. Add `STensor.expand(*sizes)` for broadcasting-style logical expansion of singleton dimensions to larger sizes without physically duplicating stored values (the expanded dimension's entries are logically shared). Size -1 means keep the current size. Also su...
[]
feature
[ "API/Shape & Layout/Reshape" ]
bobbyyyan__scorch-feature_symmetric_matrix
bobbyyyan__scorch
https://github.com/bobbyyyan/scorch.git
92fb190
python
[]
/testbed/run_tests.sh
3,000
Add sparse symmetric matrix support. Extend `TensorFormat` to include an optional `symmetric` flag indicating that only the lower (or upper) triangle is stored. Add `STensor.from_symmetric(indices, values, shape)` class method that constructs a symmetric sparse matrix storing only one triangle. Add `STensor.to_symmetri...
[]
feature
[ "Format/Semantic extensions" ]
End of preview. Expand in Data Studio

TensorBench

Feature-addition benchmark for LLMs and coding agents, evaluated against the Scorch codebase. Each task asks a model to add a feature (or otherwise extend functionality) to Scorch. Success is defined as the full pytest suite (original + any new tests the model adds) passing after the patch is applied inside a Docker container.

This is the dataset artifact for the TensorBench paper (NeurIPS 2026 Evaluations & Datasets track, double-blind submission).

At a glance

Tasks 199 (194 feature-addition, 5 refactor)
Target codebase bobbyyyan/scorch — a sparse+dense PyTorch compiler
Base commits 5 distinct SHAs across the bench branch
Language Python (with C++ extension)
Test runner pytest -v inside Docker

Files

File Purpose
tensorbench.json Task list. JSON array of 199 task records.
Dockerfile Eval image: python:3.11-slim + Scorch's C++ build deps + the upstream bench clone.
run_tests.sh Container CMD: rebuilds the C++ extension and runs pytest.
croissant.json Croissant metadata (core + RAI fields).

Task schema

Each record in tensorbench.json has:

Field Type Description
instance_id string e.g. bobbyyyan__scorch-feature_kernel_fusion. The feature_* / refactor_* suffix follows the original taxonomy; semantics for both is feature-addition.
repo_id string bobbyyyan__scorch for every task.
repo_url string Upstream Scorch URL.
base_commit string The SHA the task is anchored at. The harness git reset --hards to this before applying patches.
language string python.
setup_commands list[string] Container-side setup hooks.
test_command string /testbed/run_tests.sh.
test_timeout int Seconds. Default 3000.
refactor_type string Mostly empty (legacy field from the original taxonomy).
description string Natural-language task prompt the model receives.
files list[string] Files relevant to the task.
task_type string feature or refactor.
categories list[string] Topical tags (kernel, codegen, format, etc.).

Grading

Run with pytest -v; the grading strategy parses verbose lines (preserving parametrized brackets like [False]) and uses a custom preservation rule:

success  ⇔  after.failed == 0

This handles the common case where an agent adds new tests alongside new code: if every test (original + agent-added) passes, the task is a success. Default exact-match preservation would flag the new tests as "changed" and produce false failures.

How to evaluate a model

The full evaluation harness (codebench) and the project-local benchmark glue (tensorbench) are in the supplementary code submission.

# 1. Install the harness and benchmark glue
git clone <tensorbench-repo>     ~/tensorbench
git clone <codebench-repo>       ~/codebench
pip install -e ~/codebench
pip install -r ~/tensorbench/requirements.txt

# 2. Build the eval image
cd ~/tensorbench
docker build -t scorch-eval -f dockerfiles/scorch/Dockerfile dockerfiles/scorch/

# 3. Generate predictions, then grade them
python run.py predict scorch sonnet-4.5
python run.py eval    scorch sonnet-4.5

Predictions land in predictions/; eval outputs in evaluation_results/<run_id>/.

Anonymity

This is a double-blind submission. Personally identifying information has been scrubbed from the supplementary code. Reviewers should not attempt to identify the authors via the target codebase URL or other public references.

License

MIT for the benchmark scaffolding (this dataset, Dockerfile, grading strategy). The target Scorch codebase is governed by its own upstream license.

Citation

(Anonymized for double-blind review. Citation will be added at camera-ready.)

Downloads last month
16