This document defines vectra’s public contract: which operations are supported, how types and coercion work, what streams and what materializes, and where guarantees stop. It is the reference for what vectra promises.
See ?tbl, ?filter, ?left_join,
and ?explain for function-level documentation.
What vectra is
vectra is an R-native columnar query engine for datasets that don’t
fit in memory. It provides dplyr-style verbs backed by a pure C11
pull-based execution engine and a custom on-disk format
(.vtr). All operations are lazy until
collect() materializes results as an R data.frame.
vectra is not a dplyr backend plugin. It defines its own S3 generics. The verbs share names and semantics with dplyr, but do not depend on it.
Execution model
Pull-based pipeline
Every verb (filter, select,
mutate, …) builds a plan node. Nodes form a tree. No data
moves until collect() calls the root node’s
next_batch() function, which pulls data through the tree
one row group at a time.
A row group (internally: VecBatch) is a set of columnar
arrays, typically thousands to millions of rows. Each column is a typed
array with a validity bitmap for NA support.
Selection vectors (zero-copy filtering)
filter() does not copy rows. It attaches a
selection vector to the batch: an integer array
indexing which physical rows pass the predicate. Downstream nodes read
only the selected rows. This avoids memory allocation and copying for
selective filters.
Data sources
| Function | Format | Streaming |
|---|---|---|
tbl(path) |
.vtr (vectra native) |
yes, row-group-at-a-time |
tbl_csv(path) |
CSV | yes, batch-at-a-time |
tbl_sqlite(path, table) |
SQLite | yes, batch-at-a-time |
tbl_tiff(path) |
GeoTIFF | yes, row-strip-at-a-time |
All sources produce the same vectra_node object. The
query engine does not know or care which source is upstream.
Output sinks
| Function | Format | Streaming |
|---|---|---|
collect() |
R data.frame | materializes full result in R memory |
write_vtr(df, path) |
.vtr |
writes from data.frame |
write_csv(x, path) |
CSV | streams batch-by-batch |
write_sqlite(x, path, table) |
SQLite | streams batch-by-batch |
write_tiff(x, path) |
GeoTIFF | streams batch-by-batch |
Supported verbs
Transformation verbs
| Verb | Streams | Notes |
|---|---|---|
filter(...) |
yes | Zero-copy via selection vector |
select(...) |
yes | Full tidyselect: starts_with(),
where(), -col, etc. |
mutate(...) |
yes | Arithmetic, comparison, boolean, is.na(),
nchar(), substr(), grepl(), math
(abs, sqrt, log,
exp, floor, ceiling,
round, log2, log10,
sign, trunc), if_else(),
between(), %in%, type casting
(as.numeric), tolower(),
toupper(), trimws(), paste0(),
gsub(), sub(), startsWith(),
endsWith(), pmin(), pmax(),
year(), month(), day(),
hour(), minute(), second(),
as.Date()
|
transmute(...) |
yes | Like mutate() but drops unmentioned
columns |
rename(...) |
yes | Full tidyselect rename support |
relocate(...) |
yes | Reorder columns with .before /
.after
|
Aggregation verbs
| Verb | Streams | Notes |
|---|---|---|
group_by(...) |
metadata only | Attaches grouping info; no data moves |
summarise(...) |
materializes | Hash-based or sort-based aggregation |
ungroup() |
metadata only | Removes grouping |
count(...) |
materializes | Sugar for
group_by() |> summarise(n = n())
|
tally(...) |
materializes | Like count() on existing groups |
Supported aggregation functions: n(),
sum(), mean(), min(),
max(), sd(), var(),
first(), last(), any(),
all(), median(), n_distinct().
All accept na.rm = TRUE.
Ordering verbs
| Verb | Streams | Notes |
|---|---|---|
arrange(...) |
materializes | External merge sort with 1 GB spill budget |
slice_head(n) |
yes | Limit node, stops after n rows |
slice_tail(n) |
materializes | Must see all rows to take last n |
slice_min(order_by, n) |
partial | Heap-based top-N; with_ties = TRUE
(default) includes ties |
slice_max(order_by, n) |
partial | Heap-based top-N; with_ties = TRUE
(default) includes ties |
head(n) |
yes | Alias for
slice_head() |> collect()
|
slice(...) |
materializes | Select or exclude rows by position (positive or negative indices) |
distinct(...) |
materializes | Uses hash-based grouping |
Join verbs
| Verb | Streams | Notes |
|---|---|---|
inner_join(x, y) |
build materializes right | Hash join; left streams |
left_join(x, y) |
build materializes right | Hash join; left streams |
right_join(x, y) |
build materializes left | Implemented as swapped left join |
full_join(x, y) |
build materializes right | Hash join + finalize pass |
semi_join(x, y) |
build materializes right | Hash join; returns left rows only |
anti_join(x, y) |
build materializes right | Hash join; returns non-matching left rows |
cross_join(x, y) |
materializes | Cartesian product; no key columns required |
All joins support: by = "col",
by = c("a" = "b"), by = NULL (natural join),
and suffix = c(".x", ".y").
Window functions
Available inside mutate():
| Function | Description |
|---|---|
row_number() |
Sequential row number (respects groups) |
rank(col) |
Min rank with gaps for ties (like
dplyr::min_rank()) |
dense_rank(col) |
Consecutive rank without gaps |
lag(col, n, default) |
Previous value |
lead(col, n, default) |
Next value |
cumsum(col) |
Cumulative sum |
cummean(col) |
Cumulative mean |
cummin(col) |
Cumulative minimum |
cummax(col) |
Cumulative maximum |
ntile(n) |
Divide rows into n roughly equal buckets |
percent_rank(col) |
Relative rank scaled to [0, 1] |
cume_dist(col) |
Cumulative distribution (proportion of values <= current) |
Window functions respect group_by() partitions. They
materialize all data within each partition.
Other verbs
| Verb | Streams | Notes |
|---|---|---|
pull(var) |
materializes | Collects one column as a vector |
bind_rows(...) |
yes | Streaming concat if schemas are compatible |
bind_cols(...) |
materializes | Collects all inputs, then cbind()
|
across(...) |
n/a | Column expansion helper for
mutate()/summarise()
|
explain() |
n/a | Prints the plan tree |
glimpse() |
materializes (preview only) | Shows column types and first few values |
tidyselect support
select(), rename(),
relocate(), distinct(), and
across(.cols) support the full tidyselect vocabulary:
starts_with(),ends_with(),contains(),matches()everything(),last_col()all_of(),any_of()where()predicates (e.g.where(is.numeric))-columnnegation
where() works because vectra builds a 0-row typed proxy
data.frame from the schema, giving tidyselect enough type information to
evaluate predicates.
Supported types
Base types
| C type | R input | R output (default) | R output (bit64 mode) |
|---|---|---|---|
int64 |
integer | double | integer64 |
double |
double | double | double |
bool |
logical | logical | logical |
string |
character | character | character |
R’s 32-bit integer is widened to 64-bit
int64 on write. On read, int64 is returned as
double by default (R has no native 64-bit integer). Set
options(vectra.int64 = "bit64") to get
bit64::integer64 output instead.
Annotated types
The .vtr format version 2 stores per-column annotations
that preserve R type metadata through the write/read cycle:
| R class | Annotation | Storage | Roundtrip |
|---|---|---|---|
| Date | "Date" |
double (days since epoch) | exact |
| POSIXct | "POSIXct\|tz" |
double (seconds since epoch) | exact (tz preserved) |
| factor | "factor\|lev1\|lev2\|..." |
string | exact (levels + order preserved) |
Annotations are metadata. The underlying C engine operates on the
base types only. Type restoration happens at collect()
time.
Date and POSIXct columns support component extraction via
year(), month(), day(),
hour(), minute(), second() in
mutate() and filter() expressions. Use
as.Date("2020-01-01") as a literal in filter comparisons.
Date arithmetic (adding/subtracting days) works via standard
+ and -.
Coercion rules
Arithmetic and comparison expressions
The coercion hierarchy for numeric operations is:
bool < int64 < double
When an expression combines two different numeric types, the narrower type is promoted to the wider type before evaluation. String columns cannot participate in arithmetic or comparison with numeric columns; this is an error.
Join key coercion
Join keys follow the same hierarchy. If the left key is
int64 and the right key is double, the
int64 side is coerced to double for hashing
and comparison. The coercion happens internally; the output column
retains the original left-side type.
Joining a string key against a numeric key is an
error.
bind_rows coercion
When bind_rows() combines tables with different column
types, it computes the common type using the same
bool < int64 < double hierarchy. Per-batch coercion
happens at the C level during streaming — no R fallback needed.
If column names differ across inputs, the R fallback path is used:
all inputs are collected, aligned by column name, and combined with
rbind().
NA semantics
Storage
NAs are tracked by a per-column validity bitmap. Every column of every type supports NA values. The bitmap is bit-packed (1 bit per row, 1 = valid).
Propagation
Arithmetic:
NA + x = NA,NA * x = NAComparison:
NA > x = NA,x == NA = NABoolean:
NA & FALSE = FALSE,NA & TRUE = NA,NA | TRUE = TRUE,NA | FALSE = NAAggregation: NAs are included by default; use
na.rm = TRUEto excludeJoins: NA keys never match (same as SQL NULL semantics)
Window functions:
cumsum()and friends propagate NA forward
Ordering guarantees
filter(),select(),mutate(),rename(),relocate(): preserve input orderarrange(): produces a total order (stable sort)group_by() |> summarise(): output order is not guaranteed (hash-based path) or sorted by key (sort-based path); do not depend on eitherdistinct(): output order is not guaranteedJoins: probe-side order is preserved within each batch; build-side order is not guaranteed
bind_rows(): child order is preserved (first child’s rows, then second’s, etc.)
Streaming vs materializing
Streaming nodes (constant memory per batch)
Scan (
.vtr, CSV, SQLite, TIFF)Filter
Project (select / mutate / rename / relocate / transmute)
Limit (slice_head, head)
Concat (bind_rows)
Materializing nodes
These nodes buffer data in memory:
| Node | What it buffers | Bounded by |
|---|---|---|
| Sort (arrange) | All input rows | 1 GB memory budget, then spills to disk |
| GroupAgg (summarise) | Hash table of groups + accumulators | Number of distinct groups |
| TopN (slice_min/max) | Heap of n rows | Requested n |
| Window | All rows per partition | Partition size |
| Join (build phase) | Right-side table in hash table | Right-side row count |
External sort (spill-to-disk)
arrange() accumulates incoming batches into column
builders in memory. After each batch, the sort node estimates total
memory usage across all builders. When the estimate exceeds the memory
budget (default 1 GB, defined as DEFAULT_MEM_BUDGET), the
node flushes the accumulated data as a sorted run:
The builders are finished into arrays.
The arrays are sorted in-place using a parallel merge sort (OpenMP task spawning above 32,768 rows, sequential below).
The sorted data is written to a temporary
.vtrfile in the system temp directory, split into row groups of 65,536 rows each.The builders are reset and accumulation continues.
This spill cycle repeats as many times as needed. A 10 GB dataset with a 1 GB budget produces approximately 10 spill files.
Once all input is consumed, the sort node enters the merge phase. It
opens all spill files as Vtr1File readers and runs a k-way
merge using a min-heap. Each heap entry holds a reference to one merge
run (one spill file) and tracks the current row group and cursor
position within that run. The merge emits batches of up to 65,536 rows,
reading row groups from spill files on demand. Peak memory during the
merge phase is proportional to k (number of runs) times the row group
size, not the total dataset size.
If all data fits within the 1 GB budget, no spill occurs. The node sorts in memory and emits the result directly. This is the common case for datasets under ~100 million rows (depending on column count and types).
The sort-based group_by() |> summarise() path (used
internally when the engine detects it is advantageous) also benefits
from this spill mechanism. Temporary spill files are deleted when the
sort node is freed.
Join memory model
Joins use a build-right, probe-left hash join:
- Build phase: The entire right-side table is materialized into a hash table in memory. Key columns are hashed using FNV-1a. For single-key joins, the hash is computed directly on the key bytes. For multi-key joins, per-key hashes are combined via XOR and FNV prime multiplication. The hash table uses open addressing. Key hashing is OpenMP-parallelized for batches above 32,768 rows.
- Probe phase: Left-side batches stream through one at a time, probing the hash table. For each left row, the engine computes the key hash, looks up the hash table, and verifies key equality (hash collisions are resolved by comparison). Probe-side hashing is also parallelized.
- Finalize phase (full_join only): Unmatched right-side rows are emitted in chunks of 65,536 rows. A matched-flag array tracks which right-side rows participated in the join.
The memory cost of a join is proportional to the right-side table
size (data arrays plus hash table overhead). The left side streams and
does not accumulate. For this reason, place the smaller table on the
right side of the join. The right_join verb handles this
automatically by swapping sides internally and remapping columns in the
output.
The .vtr file format
The .vtr format is vectra’s native binary columnar
format. It is designed for fast sequential reads with row-group-level
granularity.
Layout
Header:
magic bytes ("VTR1")
version (uint16: 1–4)
n_cols, n_rowgroups
per-column: name + type byte [+ annotation string in v2+]
row group index (byte offsets)
Row groups (repeated):
per-column:
validity bitmap (bit-packed)
[v4] encoding tag (1 byte) + compression tag (1 byte)
[v4] uncompressed_size (uint32)
typed data array (int64/double/bool/string)
[v3+] per-column statistics (min/max)
Version history
Version 1: Base format with typed columns and validity bitmaps.
Version 2: Adds per-column annotation strings for Date, POSIXct, and factor roundtripping.
Version 3: Adds per-column per-rowgroup statistics (min/max) enabling zone-map predicate pushdown.
Version 4 (current): Adds a two-layer encoding and compression stack. Writing always produces v4. All versions (v1–v4) are readable.
Encoding and compression (v4)
v4 applies two transformations to each column chunk (one column in one row group), in order: an encoding step and a compression step.
Encoding transforms data logically to expose redundancy. The encoder picks the best encoding per column per row group automatically:
| Encoding | Applies to | Condition | Mechanism |
|---|---|---|---|
| PLAIN | all types | default | Raw bytes, no transformation |
| DICTIONARY | string | unique ratio < 50% | Builds a string dictionary; stores indices as RLE (run-length encoded) |
| DELTA | int64 | monotonically increasing | Stores first value + deltas (all >= 0) |
DICTIONARY encoding is the most impactful for typical categorical string columns. The encoder counts distinct values in a single pass using an open-addressing hash table (70% load factor, dynamic resizing). If fewer than half the values are unique, it emits a dictionary (offset array + packed strings) followed by RLE-encoded indices. The RLE step collapses runs of repeated indices into (value, count) pairs, which is effective when rows with the same category are clustered (e.g. after a sort). If more than half the values are unique, the encoder aborts dictionary encoding and falls back to PLAIN with zero overhead.
DELTA encoding stores an initial value followed by the difference between consecutive values. It targets auto-increment IDs, timestamps, and other monotonic integer sequences where the deltas are small and compress well.
Compression squeezes bytes physically after encoding. vectra uses a built-in LZ77 compressor (LZ-VTR), approximately 120 lines of C with no external dependencies. The compressor uses a 3-byte minimum match, 256-byte maximum offset, and a hash table for match finding. It skips column chunks smaller than 64 bytes (not worth the overhead). If compression does not reduce size, the chunk is stored uncompressed. There is no configuration knob; the format always attempts compression on eligible chunks.
The two-layer design is intentional. Encoding and compression solve different problems. DICTIONARY and DELTA reduce the entropy of the data (fewer distinct byte patterns, smaller integer ranges). LZ-VTR then exploits the reduced entropy at the byte level. Applying both layers yields better ratios than either layer alone, particularly for RLE-encoded dictionary indices where long runs of identical small integers compress to near zero.
Query optimizer
explain() runs the optimizer before printing so you see
the actual execution plan. Two optimization passes run
automatically:
Predicate pushdown
When a FilterNode sits above a ScanNode
reading a .vtr file (v3+), the filter predicate is attached
to the scan. The scan then applies up to three pruning strategies on its
first next_batch() call, in priority order:
Hash index pushdown (highest priority). If a
.vtrisidecar index exists for the predicate column(s), the scan probes the index to build a row-group bitmap. Row groups not in the bitmap are skipped entirely. This handles==and%in%predicates. For composite indexes, AND-combined equality predicates on the indexed columns are matched and probed as a single composite key. See the Hash indexes section below for details.Binary search on sorted columns. If the
.vtrfile records that a column is sorted and the predicate is a simple comparison (==,<,<=,>,>=) against a literal, the scan binary-searches the row group stats to find the first and last row groups that could contain matching rows. The scan range is narrowed to[first_rg, last_rg). For AND-combined predicates on the same sorted column (e.g.x >= 10 & x < 100), both bounds are applied. For OR-combined predicates, the union of both ranges is used.Zone-map pruning (applied per row group during iteration). Each row group in a v3+ file stores per-column min/max statistics. Before reading a row group’s data from disk, the scan evaluates the pushed-down predicate against the row group’s stats. If the predicate is provably false for the entire row group (e.g.
filter(x > 100)on a row group where max(x) = 50), that row group is skipped entirely without touching the underlying bytes. Zone-map pruning handles comparison operators on numeric and string columns, AND/OR combinations, and%in%predicates (checking whether any set value falls within the row group’s min/max range). String zone maps use a packed 8-byte prefix representation for efficient comparison.
These three strategies compose. The scan first applies hash index and
binary search to narrow the set of candidate row groups, then checks
zone-map stats on each candidate before reading data. In
explain() output, predicate pushdown appears as
predicate pushdown and v3 stats annotations on
the ScanNode.
Column pruning
The optimizer walks the plan tree top-down and determines which
columns each node actually needs from its child. The required column set
at each node is the union of: columns referenced in the node’s own
expressions (filter predicates, mutate expressions, aggregation
functions), columns passed through to the parent, and join key columns.
At scan nodes, unneeded columns are excluded from disk reads by setting
a column mask. For a 100-column .vtr file where only 3
columns are needed, this means 97 columns are never deserialized, never
decompressed, and never decoded. This is visible in
explain() as 3/100 cols (pruned). Column
pruning applies to all .vtr scans regardless of format
version.
Hidden mutate insertion
When summarise() contains nested expressions like
mean(x + y), the optimizer auto-inserts a
ProjectNode (visible as hidden mutate in
explain()) to compute the intermediate expression
x + y before aggregation. This is necessary because the
C-level aggregation nodes (GroupAgg) operate on single columns, not on
expression trees. The hidden mutate computes the intermediate column and
gives it a generated name, which the aggregation node then references.
The same mechanism applies to sum(log(x)),
n_distinct(a + b), and similar patterns. The inserted node
is visible in explain() output but invisible to the user’s
pipeline.
explain() contract
explain() prints the optimized plan tree without
executing it. The output shows:
Node types in execution order (leaf to root)
Per-node annotations: streaming/materializing, column pruning, predicate pushdown, v3 stats, hidden mutate
Grouping columns if present
Output schema (column names and types)
tbl("data.vtr") |>
filter(x > 0) |>
select(id, x) |>
explain()
#> vectra execution plan
#>
#> ProjectNode [streaming]
#> FilterNode [streaming]
#> ScanNode [streaming, 2/5 cols (pruned), predicate pushdown, v3 stats]
#>
#> Output columns (2):
#> id <int64>
#> x <double>The plan tree is a description of what will happen, not a guarantee of how it will happen internally. Node ordering and naming may change between versions.
Hash indexes (.vtri)
vectra supports persistent on-disk hash indexes stored as
.vtri sidecar files alongside .vtr data files.
These indexes enable O(1) row group pruning for equality predicates,
turning full-table scans into targeted reads.
Creating indexes
# Single-column index
create_index(tbl("data.vtr"), "species")
# Case-insensitive index
create_index(tbl("data.vtr"), "species", ci = TRUE)
# Composite (multi-column) index
create_index(tbl("data.vtr"), c("country", "year"))
# Check if an index exists
has_index(tbl("data.vtr"), "species")create_index() reads the .vtr file, hashes
every value in the indexed column(s) per row group, and writes a
.vtri file. The file name encodes the indexed columns:
data.species.vtri for a single-column index,
data.country_year.vtri for a composite.
Index format
The .vtri format is a chained hash table mapping key
hashes to row group indices:
- v1 (single-column): stores one column index, a case-insensitive flag, hash entries, row group entries, a slot heads array, and a next-pointer chain. Hashing uses FNV-1a on the raw column bytes. For case-insensitive indexes, the hash folds ASCII uppercase to lowercase before hashing.
- v2 (composite): stores multiple column indices. The composite hash combines per-column FNV-1a hashes via XOR and multiplication with the FNV prime, producing a single 64-bit hash per row group entry.
How the scan uses indexes
When a ScanNode opens a .vtr file, it
checks for .vtri sidecar files matching the columns
referenced in the pushed-down predicate. For single-column indexes, it
matches == predicates and %in% predicates. For
composite indexes, it matches AND-combined equality predicates where
every indexed column has an == literal clause.
On match, the scan probes the index to produce a row-group bitmap.
For %in% predicates, the scan probes once per set element
and ORs the bitmaps together. Row groups with a 0 bit in the bitmap are
never read from disk. This is the first pruning step, applied before
binary search and zone-map checks.
Performance characteristics
Index probing is O(k) where k is the number of query keys (1 for
==, n for %in%). Each probe is a hash
computation plus a chain walk in the slot array. The index file is
memory-mapped at open time, so repeated probes pay no I/O cost. For
tables with many row groups and selective equality predicates, index
pushdown can reduce I/O by orders of magnitude compared to zone-map
pruning alone.
Materialized blocks
materialize() consumes a vectra node and stores the
result as a persistent in-memory columnar block. Unlike nodes, which are
consumed on collect() and cannot be reused, blocks persist
and support repeated lookups.
blk <- tbl("backbone.vtr") |>
select(taxonID, canonicalName, genus) |>
materialize()Exact lookups
block_lookup() performs hash-based lookups on a string
column. Hash indices are built lazily on first use and cached for
subsequent calls. The return value is a data.frame with a
query_idx column (1-based position in the input keys
vector) plus all columns from the block.
hits <- block_lookup(blk, "canonicalName",
c("Quercus robur", "Pinus sylvestris"))
# Case-insensitive
hits <- block_lookup(blk, "canonicalName",
c("quercus robur"), ci = TRUE)Fuzzy lookups
block_fuzzy_lookup() computes string distances between
query keys and a block column. Three distance methods are available:
Damerau-Levenshtein ("dl", default), Levenshtein
("levenshtein"), and Jaro-Winkler ("jw").
Results are filtered by a maximum normalized distance threshold (default
0.2).
fuzzy <- block_fuzzy_lookup(blk, "canonicalName",
c("Quercus robar", "Pinus silvestris"),
method = "dl", max_dist = 0.2)An optional blocking column reduces the search space by requiring exact matches on a second column before computing distances. This is useful for taxonomic lookups where genus is known:
fuzzy <- block_fuzzy_lookup(blk, "canonicalName",
c("Quercus robar"),
block_col = "genus",
block_keys = c("Quercus"),
n_threads = 4L)Fuzzy lookups are OpenMP-parallelized. The n_threads
parameter controls the thread count (default 4).
Use case
Materialized blocks are designed for repeated lookups against a
reference table. The typical pattern is: load a backbone or codelist
once with materialize(), then probe it many times with
different query vectors. The block stays in memory across calls, and its
internal hash index (for exact lookups) is built once and reused.
OpenMP parallelization
The C engine uses OpenMP for CPU-bound operations where the per-element cost is high enough to justify thread management overhead. Parallelism is conditional: operations fall back to single-threaded execution when the batch size is below a threshold or when the platform does not support OpenMP.
Which operations parallelize
| Operation | Threshold | Schedule | Notes |
|---|---|---|---|
filter() (selection vector build) |
32,768 rows | parallel prefix sum | Two-phase: count matches per thread, then write at offsets |
grepl() with regex |
1,000 rows | dynamic, 64 |
Per-thread regex compilation for thread safety |
levenshtein(), dl_dist(),
jaro_winkler()
|
1,000 rows | dynamic, 64 |
Fuzzy string distance in mutate expressions |
| Sort (merge sort) | 32,768 rows | task |
Recursive task spawning for parallel merge sort |
| Sort (key extraction) | 32,768 rows | static |
Parallel extraction of sort keys into index arrays |
| Join (build-side hashing) | 32,768 rows | static |
Parallel hash computation for build-side keys |
| Join (probe-side hashing) | 32,768 rows | static |
Parallel hash computation for probe-side keys |
| Window (data copy) | 32,768 rows | static |
Parallel copy of partition data |
| Window (group dispatch) | 64 groups | dynamic |
Parallel window computation across groups |
| Collect (column append) | 8 columns | static |
Parallel column-by-column append to R vectors |
| Zone-map stat computation | 32,768 rows |
static with reduction
|
Parallel min/max scan during write |
| Block fuzzy lookup | always (if > 0 keys) |
dynamic, 1 per key /
dynamic, 16 per row |
Parallelized by query key and by block row |
| Literal fill (broadcast) | 32,768 rows | static |
Parallel fill for constant columns |
The general-purpose threshold is 32,768 rows
(VEC_OMP_THRESHOLD). String operations use a lower
threshold of 1,000 rows because their per-element cost (regex
compilation, edit distance matrices) is much higher than arithmetic.
Thread safety
Regex operations (grepl with regex patterns) compile the
POSIX regex once per thread. Each thread owns its own
regex_t instance, allocated in thread-local scope inside
the #pragma omp parallel block. This avoids both contention
and the overhead of recompiling the regex per row.
The filter node uses a parallel prefix sum to build the selection vector without locking. Each thread counts matches in its chunk, a sequential scan computes prefix offsets, then each thread writes selected indices at its computed offset.
Current limitations
slice_tail materializes: There is no reverse-scan optimization.
distinct with .keep_all: Falls back to R when
.keep_all = TRUEwith a column subset.Predicate pushdown is .vtr only: CSV, SQLite, and TIFF scans do not benefit from predicate pushdown, column pruning, or hash index acceleration.
No SIMD: Arithmetic and comparison operations use scalar loops. The compiler may auto-vectorize some patterns, but there are no explicit SIMD intrinsics.
OpenMP availability varies: On macOS, R ships without OpenMP by default. Users must install
libomp(e.g. via Homebrew) and configure the compiler flags. On Windows (rtools) and Linux, OpenMP is typically available out of the box. When OpenMP is not available, all operations run single-threaded with no functional difference.
Fallback behavior
vectra has these fallback paths to base R:
bind_rows with mismatched column names: If column names differ across inputs, all tables are collected and combined via
rbind()in R.distinct with .keep_all and column subset: Falls back to
duplicated()in R (emits a message).slice_tail: Must see all rows to take the last n; returns a data.frame.
slice_min/slice_max with
with_ties = TRUE(the default): Collects all data to identify ties at the boundary; returns a data.frame.reframe: Always collects and evaluates in R; returns a data.frame.
All other operations execute entirely in C. There is no silent fallback to dplyr or any other package.
Grouping preservation
All verbs preserve group_by() metadata:
filter(), select(), mutate(),
rename(), relocate(), arrange(),
and transmute() pass grouping through.
rename() additionally updates group column names to match
the rename. summarise() drops grouping according to its
.groups argument.
Package conflicts
vectra defines its own S3 generics for dplyr-like verbs
(filter, select, mutate, etc.)
and utility functions (glimpse, collect). If
dplyr is also loaded, whichever package was attached last will mask the
other’s generics. vectra’s methods will still dispatch correctly on
vectra_node objects regardless of masking order.