Package-level declarations
Types
Global hook for the active MemoryTracker.
Provides byte-level read access to a BufferHandle, regardless of its ownership mode.
Ownership / residency mode of a tensor's backing memory.
Factory and conversion utilities for creating BufferHandle instances from common Kotlin types and for slicing existing handles.
Resolves a BufferHandle into a BufferAccessor that can read the underlying bytes. Platform-specific implementations handle file-backed and device-resident buffers; heap-backed handles are resolved generically.
BufferAccessor over a plain ByteArray.
Bridge between KvCacheStore and the SDPA execution path.
Default resolver that handles heap-backed handles directly and delegates file-backed handles to a fileBackedResolver.
Default KV cache implementation using dense FP32 storage.
Configures TurboQuant KV-cache compression for an attention layer.
Resolves KvCache annotations to KvCacheStore instances.
Disables TurboQuant compression for a specific layer.
Configuration for asymmetric K/V encoding policies.
Memory report for a KV cache instance.
Dedicated KV-cache storage abstraction for inference.
Logical numeric type — what the tensor values mean semantically.
Resolves Placement intent into concrete buffer allocation decisions.
Tracks memory allocation events and reports aggregate statistics across all live TensorStorage instances.
Shared contract for all packed/quantized block tensor storage formats.
Declares placement intent for a tensor parameter or property.
High-level placement descriptor: where a tensor lives and how the runtime should manage it.
Diagnostic snapshot of a single tensor's memory characteristics.
A storage specification that captures both logical type AND physical encoding + placement intent. This enables factory routing that goes beyond dtype-only decisions.
Physical storage encoding — how tensor data is laid out in memory.
Runtime descriptor for a tensor's backing memory.
Factory methods for constructing TensorStorage from existing SKaiNET types and from raw data. These bridge the old TensorData world to the new storage model.
KV cache store with TurboQuant compression.
Marks a tensor as an immutable weight that should be file-backed (memory-mapped) when possible.