--- title: Linear Re-representation with regress() output: rmarkdown::html_vignette vignette: > %\VignetteIndexEntry{Linear Re-representation with regress()} %\VignetteEngine{knitr::rmarkdown} %\VignetteEncoding{UTF-8} params: family: red css: albers.css resource_files: - albers.css - albers.js includes: in_header: |- --- ```{r setup, include=FALSE} knitr::opts_chunk$set(collapse = TRUE, comment = "#>", fig.width=6, fig.height=4) # Assuming necessary multivarious functions are loaded # e.g., via devtools::load_all() or library(multivarious) library(multivarious) # Implicit dependencies like glmnet or pls might be needed depending on method used. ``` When you already have a known set of basis functions—such as Fourier components, wavelets, splines, or principal components from a reference dataset—the goal isn't to discover a latent space, but rather to express new data in terms of that existing basis. The `regress()` function facilitates this by wrapping several multi-output linear models within the `bi_projector` API. This integration means you automatically inherit standard `bi_projector` methods for tasks like: * `project()`: Map new data from the original space into the basis coefficients. * `inverse_projection()`: Map basis coefficients back to the original data space. * `reconstruct()` / `reconstruct_new()`: Reconstruct data using all or a subset of basis components. * `coef()`: Retrieve the basis coefficients. * `truncate()`: Keep only a subset of basis components. * Plus caching, variable tracking helpers, etc. This vignette demonstrates a typical workflow using an orthonormal Fourier basis, but the `regress()` function works equally well with arbitrary, potentially non-orthogonal, basis dictionaries. --- # 1. Build a design matrix of basis functions First, let's define our basis. We'll use sines and cosines. ```{r build_basis} set.seed(42) n <- 128 # Number of observations (e.g., signals) p <- 32 # Original variables per observation (e.g., time points) k <- 20 # Number of basis functions (<= p often, <= n for lm) ## Create toy data: smooth signals + noise t <- seq(0, 1, length.out = p) Y <- replicate(n, 3*sin(2*pi*3*t) + 2*cos(2*pi*5*t) ) + matrix(rnorm(n*p, sd = 0.3), p, n) # Note: Y is p x n here ## Orthonormal Fourier basis (columns = basis functions) # We create k/2 sine and k/2 cosine terms, plus an intercept freqs <- 1:(k %/% 2) # Integer division for number of frequencies B <- cbind(rep(1, p), # Intercept column do.call(cbind, lapply(freqs, function(f) sin(2*pi*f*t))), do.call(cbind, lapply(freqs, function(f) cos(2*pi*f*t)))) colnames(B) <- c("Intercept", paste0("sin", freqs), paste0("cos", freqs)) # Make columns orthonormal (length 1, orthogonal to each other) B <- scale(B, center = FALSE, scale = sqrt(colSums(B^2))) cat(paste("Dimensions: Y is", nrow(Y), "x", ncol(Y), ", Basis B is", nrow(B), "x", ncol(B), "\n")) # We want coefficients C (k x n) such that Y ≈ B %*% C. ``` --- # 2. Fit multi-output regression Now, we use `regress()` to find the coefficients \(C\) that best represent each signal in \(Y\) using the basis \(B\). ```{r fit_regression} library(multivarious) # Fit using standard linear models (lm) # Y is p x n (32 x 128): 32 time points x 128 signals # B is p x k (32 x 21): 32 time points x 21 basis functions # regress will fit 128 separate regressions, each with 32 observations and 21 predictors fit <- regress(X = B, # Predictors = basis functions (p x k) Y = Y, # Response = signals (p x n) method = "lm", intercept = FALSE) # Basis B already includes an intercept column # The result is a bi_projector object print(fit) ## Conceptual mapping to bi_projector slots: # fit$v : Coefficients (n x k) - Basis coefficients for each signal. # fit$s : Design Matrix (p x k) - The basis matrix B. # Stored for reconstruction. ``` The `bi_projector` structure provides a consistent way to access the core components: the basis (`fit$s`) and the coefficients (`fit$v`). --- # 3. Go back and forth between spaces With the fitted `bi_projector`, we can move freely between the original data space and the basis coefficient space. This section demonstrates the key operations. ## 3.1 Inspecting the fitted coefficients The coefficient matrix `fit$v` has dimensions $n \times k$ (signals $\times$ basis functions). Each row contains the weights that express one signal as a linear combination of basis functions. ```{r inspect_coefficients} coef_matrix_first3 <- fit$v[1:3, ] cat("Coefficient matrix shape (first 3 signals):", nrow(coef_matrix_first3), "x", ncol(coef_matrix_first3), "\n\n") cat("Coefficients for signal 1:\n") print(coef_matrix_first3[1, ]) ``` Notice that `sin3` and `cos5` have the largest coefficients---this matches our generating function $3\sin(2\pi \cdot 3t) + 2\cos(2\pi \cdot 5t)$. ## 3.2 Reconstructing the fitted data Reconstruction recovers the original data from the basis representation. The `bi_projector` stores both the design matrix $B$ (in `fit$s`) and the coefficients (in `fit$v`). Reconstruction is simply their product: $$\hat{Y} = B \cdot C^T = \texttt{fit\$s} \times \texttt{t(fit\$v)}$$ ```{r reconstruct_fitted} Y_hat <- fit$s %*% t(fit$v) max_reconstruction_error <- max(abs(Y_hat - Y)) cat("Reconstruction shape:", nrow(Y_hat), "x", ncol(Y_hat), "\n") cat("Maximum reconstruction error:", format(max_reconstruction_error, digits=3), "\n") ``` The error is non-zero because our signals contain random noise that cannot be captured by the smooth Fourier basis. This is expected and acceptable. ## 3.3 Projecting new data onto the basis A key use case is expressing *new* observations in terms of the same basis. For an orthonormal basis $B$, projection is straightforward: $$\text{coefficients} = B^T \cdot Y_{\text{new}}$$ Let's create a new signal with the same underlying pattern but different noise: ```{r project_new_signal} Y_new_signal <- 3*sin(2*pi*3*t) + 2*cos(2*pi*5*t) + rnorm(p, sd=0.3) Y_new_matrix <- matrix(Y_new_signal, nrow = p, ncol = 1) coef_new <- t(fit$s) %*% Y_new_matrix cat("Basis coefficients for new signal:\n") print(coef_new[, 1]) ``` Again, the `sin3` and `cos5` components dominate, as expected. ## 3.4 Reconstructing new data Finally, we can reconstruct the new signal from its basis coefficients to see how well the basis captures the underlying structure: ```{r reconstruct_new_signal} Y_new_recon <- fit$s %*% coef_new reconstruction_error <- sqrt(mean((Y_new_matrix - Y_new_recon)^2)) cat("Reconstruction RMSE for new signal:", format(reconstruction_error, digits=3), "\n") ``` The reconstruction error reflects the noise that the basis cannot represent. Because $B$ is orthonormal, projection and reconstruction are exact inverses for the signal components that lie within the basis span. --- # 4. Regularisation & PLS If the basis is ill-conditioned or you need feature selection/shrinkage, simply change the `method` argument. `regress()` wraps common regularized models. ```{r regularized_models, eval=FALSE} # Ridge regression (requires glmnet) fit_ridge <- regress(X = B, Y = Y, method = "mridge", lambda = 0.01, intercept = FALSE) # Elastic Net (requires glmnet) fit_enet <- regress(X = B, Y = Y, method = "enet", alpha = 0.5, lambda = 0.02, intercept = FALSE) # Partial Least Squares (requires pls package) - useful if k > p or multicollinearity fit_pls <- regress(X = B, Y = Y, method = "pls", ncomp = 15, intercept = FALSE) # All these return bi_projector objects, so downstream code using # project(), reconstruct(), coef() etc. remains the same. ``` --- # 5. Partial / custom mappings The `bi_projector` interface allows for flexible manipulation: ```{r partial_mappings} # Truncate: Keep only the first 5 basis functions (Intercept + 2 sine + 2 cosine) fit5 <- truncate(fit, ncomp = 5) cat("Dimensions after truncating to 5 components:", "Basis (s):", paste(dim(fit5$s), collapse="x"), ", Coefs (v):", paste(dim(fit5$v), collapse="x"), "\n") # Reconstruction using only first 5 basis functions (manual) # Equivalent to: scores(fit5) %*% t(coef(fit5)) for the selected components Y_hat5 <- fit5$s %*% t(fit5$v) # Partial inverse projection: Map only a subset of coefficients back # e.g., reconstruct using only components 2 through 6 (skip intercept) # Note: partial_inverse_projection is not a standard bi_projector method, # this might require manual slicing of the basis matrix B (fit$s) or coefs (fit$v). # Manual reconstruction example for components 2:6 coef_subset <- fit$v[2:6, , drop=FALSE] # k_sub x n basis_subset <- fit$s[, 2:6, drop=FALSE] # p x k_sub Y_lowHat <- basis_subset %*% coef_subset # p x n reconstruction # Variable usage helpers (Conceptual - actual functions might differ) # `variables_used(fit)` could show which basis functions have non-zero coefficients (esp. for 'enet'). # `vars_for_component(fit, k)` isn't directly applicable here as components are the basis functions themselves. ``` --- # 6. Under-the-hood: Matrix View The core idea is to represent the \(p \times n\) data matrix \(Y\) as a product of the \(p \times k\) basis matrix (stored in `s`) and the \(k \times n\) coefficient matrix (stored in `v`): \[ \underbrace{Y}_{p\times n} \approx \underbrace{s}_{p\times k} \underbrace{v}_{k\times n} \] `regress()` estimates \(v\) (coefficients) based on the chosen method: | `method` | Solver Used (Conceptual) | Regularisation | Key Reference | |----------|--------------------------|----------------------|----------------------| | "lm" | QR decomposition (`lm.fit`)| None | Classical OLS | | "mridge" | `glmnet` (alpha=0) | Ridge (\(\lambda ||\beta||_2^2\)) | Hoerl & Kennard 1970 | | "enet" | `glmnet` | Elastic Net (\(\alpha\)-mix) | Zou & Hastie 2005 | | "pls" | `pls::plsr` | Latent PLS factors | Wold 1984 | The resulting object stores: * `v`: The estimated coefficient matrix (\(k \times n\)). * `s`: The basis/design matrix \(B\) (\(p \times k\)), possibly centered or scaled by the underlying solver. * `sdev`, `center`: Potentially stores scaling/centering info related to \(s\). All other `bi_projector` methods (`project`, `reconstruct`, `inverse_projection`) are derived from these core matrices. --- # 7. Internal Checks (Developer Focus) This section contains internal consistency checks, primarily for package development and CI testing. The code chunk below only runs if the environment variable `_MULTIVARIOUS_DEV_COVERAGE` is set. ```{r internal_checks, eval=nzchar(Sys.getenv("_MULTIVARIOUS_DEV_COVERAGE"))} # This chunk only runs if _MULTIVARIOUS_DEV_COVERAGE is non-empty message("Running internal consistency checks for regress()...") tryCatch({ stopifnot( # Check reconstruction fidelity for lm max(abs(reconstruct(fit) - Y)) < 1e-10, # Check dimensions of inverse projection matrix (n x p) # inverse_projection maps coefficients (k x n) back to data (p x n) # The matrix itself maps k -> p implicitly. Let's check coef matrix dims. nrow(fit$v) == ncol(B), # k rows ncol(fit$v) == ncol(Y) # n columns # Add checks for other methods if evaluated ) message("Regress internal checks passed.") }, error = function(e) { warning("Regress internal checks failed: ", e$message) }) ``` --- # 8. Take-aways * `regress()` provides a convenient way to fit multiple linear models simultaneously when expressing data \(Y\) in a known basis \(B\). * It returns a `bi_projector`, giving immediate access to projection, reconstruction, truncation, and coefficient extraction. * Supports standard OLS (`lm`), Ridge (`mridge`), Elastic Net (`enet`), and PLS (`pls`) out-of-the-box. * Works with any basis dictionary (orthonormal or not). * Can be integrated into larger analysis pipelines using composition (`%>>%`) or cross-validation helpers. Happy re-representing!