Skip to content

Results Class

The SCRIBE package provides a unified ScribeResults class that works consistently across all model types. This class encapsulates all functionality for handling model inference outputs, parameter access, and downstream analysis.

Base Structure

The ScribeResults class provides core functionality for:

  • Accessing variational model parameters and posterior distributions
  • Indexing results by gene (single gene, ranges, boolean indexing)
  • Selecting specific mixture components in mixture models
  • Generating posterior samples and predictive samples
  • Computing log likelihoods and model comparisons
  • Handling metadata from AnnData objects

Basic Usage

After running inference with scribe.fit(), you'll get a results object:

import scribe

results = scribe.fit(adata, n_steps=10_000)

The results object contains several key attributes:

  • params — dictionary of learned variational parameters
  • loss_history — array of ELBO values during training
  • n_cells, n_genes — dataset dimensions
  • model_type — string indicating the type of model
  • model_config — configuration object with model architecture and priors
  • n_components — number of components in mixture models (None for non-mixture models)
  • obs, var, uns — optional metadata if using AnnData

Common Operations

Accessing Parameters and Posterior Distributions

The ScribeResults class provides several methods to access the learned model parameters, either as raw variational parameters, probability distributions, or point estimates:

# Get raw parameters for variational posterior
params = results.params

# Get posterior distributions for parameters
# (returns scipy.stats distributions by default)
distributions = results.get_distributions()

# Get posterior distributions as numpyro distributions
distributions_numpyro = results.get_distributions(backend="numpyro")

# Get maximum a posteriori (MAP) estimates
map_estimates = results.get_map()

Descriptive parameter names

By default, SCRIBE's internal parameter keys use compact math-style names (r, p, mu, phi, gate, p_capture, etc.). For more readable output, pass descriptive_names=True to any parameter-access method. This renames the keys to self-documenting equivalents:

Internal key Descriptive name
r dispersion
p prob
mu expression
phi odds
gate zero_inflation
p_capture capture_prob
phi_capture capture_odds
eta_capture capture_efficiency
mu_eta capture_scaling

Suffixed keys are handled automatically (e.g., r_0 becomes dispersion_0).

# Default internal names
map_estimates = results.get_map()
# >>> dict_keys(['r', 'p', 'p_capture', ...])

# Human-readable names
map_estimates = results.get_map(descriptive_names=True)
# >>> dict_keys(['dispersion', 'prob', 'capture_prob', ...])

The descriptive_names option is supported by get_map(), get_distributions(), get_posterior_samples(), and sample_posterior_parameters().

Tip

Use descriptive_names=True in notebooks and exploratory analysis for clarity. Stick with the default internal names when passing parameters to other SCRIBE functions that expect them.

Subsetting Genes

The ScribeResults object supports indexing operations to extract results for specific genes of interest. You can use integer indexing, slicing, or boolean masks to subset the results:

# Get results for first gene
gene_results = results[0]

# Get results for a set of genes
subset_results = results[0:10]  # First 10 genes

# Boolean indexing
highly_variable = results.var["highly_variable"]
if highly_variable is not None:
    hv_results = results[highly_variable]

Working with Mixture Components

For mixture models, you can access specific components:

# Get results for the first component
component_results = results.get_component(0)

# The component results are a non-mixture ScribeResults object
print(component_results.model_type)  # e.g., "nbdm" instead of "nbdm_mix"

Posterior Sampling

The ScribeResults class provides several methods for generating different types of samples:

  1. Posterior Parameter Samples: Draw samples directly from the fitted parameter distributions using get_posterior_samples(). These samples represent uncertainty in the model parameters as sampled from the variational posterior distribution.

  2. Predictive Samples: Generate new data from the model using get_predictive_samples(). This simulates new count data using the MAP parameter estimates.

  3. Posterior Predictive Check (PPC) Samples: Combine both operations with get_ppc_samples() to generate data for model validation.

# Draw 1000 samples from the posterior distributions of parameters
posterior_samples = results.get_posterior_samples(n_samples=1000)

# Generate new count data using MAP estimates
predictive_samples = results.get_predictive_samples()

# Generate posterior predictive samples for model checking
ppc_samples = results.get_ppc_samples(n_samples=1000)

Note

Generating posterior predictive samples requires simulating entire datasets, which can be computationally intensive. For large datasets, we recommend:

  • Reducing the number of samples
  • Subsetting to fewer genes
  • Using GPU acceleration if available
  • Running sampling in batches

Log Likelihood Computation

Computing the log-likelihood of your data under the fitted model can be valuable for several purposes:

  • Model comparison: Compare different model fits or architectures by their log-likelihood scores
  • Quality control: Identify cells or genes that are poorly explained by the model
  • Outlier detection: Find data points with unusually low likelihood values
  • Model validation: Assess how well the model captures the underlying data distribution
log_liks = results.compute_log_likelihood(
    counts,
    return_by="cell",  # or 'gene'
    batch_size=512,
)

Cell Type Assignment (Mixture Models)

For mixture models, SCRIBE provides methods to compute probabilistic cell type assignments. These assignments quantify how likely each cell belongs to each component (cell type) in the mixture, while also characterizing the uncertainty in these assignments.

The computation involves three key steps:

  1. For each cell, compute the likelihood that it belongs to each component using the full posterior distribution of model parameters
  2. Convert these likelihoods into proper probability distributions over components
  3. (Optional) Fit a Dirichlet distribution to characterize the uncertainty in these assignments

The resulting probabilities can be used to:

  • Make soft assignments of cells to types
  • Identify cells with ambiguous type assignments
  • Quantify uncertainty in cell type classifications
  • Study cells that may be transitioning between states

Two methods are provided:

  • compute_cell_type_assignments() — uses the full posterior distribution to compute assignments and uncertainty
  • compute_cell_type_assignments_map() — uses point estimates for faster but less detailed results
# Compute cell type assignment probabilities
assignments = results.compute_cell_type_assignments(
    counts,
    fit_distribution=True,
)

# Get Dirichlet concentration parameters
concentrations = assignments["concentration"]

# Get mean assignment probabilities
mean_probs = assignments["mean_probabilities"]

# Get assignment probabilities for each posterior sample
sample_probs = assignments["sample_probabilities"]

# Compute using MAP estimates only (faster, less uncertainty info)
map_assignments = results.compute_cell_type_assignments_map(counts)

Entropy Analysis for Mixture Models

For mixture models, SCRIBE provides methods to compute the entropy of component assignments, which serves as a measure of assignment uncertainty. Higher entropy values indicate more uncertainty in the assignments (the cell or gene could belong to multiple components), while lower values indicate more confident assignments (the cell or gene clearly belongs to one component).

The entropy calculation can be performed:

  • Per cell: Measuring how confidently each cell is assigned to a component
  • Per gene: Measuring how component-specific each gene's expression pattern is
  • With optional normalization: Making entropy values comparable across datasets of different sizes
entropies = results.compute_component_entropy(
    counts,
    return_by="cell",  # or 'gene'
    normalize=False,
)

Model-Specific Parameters

The ScribeResults class works with all model types supported by SCRIBE. Each model type has specific parameters available in the params dictionary based on the distributions used.

nbdm_results = scribe.fit(adata, model="nbdm")

# Dispersion parameters (LogNormal distribution)
r_loc = nbdm_results.params["r_loc"]
r_scale = nbdm_results.params["r_scale"]

# Or (Gamma distribution)
r_concentration = nbdm_results.params["r_concentration"]
r_rate = nbdm_results.params["r_rate"]

# Success probability parameters
p_concentration1 = nbdm_results.params["p_concentration1"]  # Alpha
p_concentration0 = nbdm_results.params["p_concentration0"]  # Beta
zinb_results = scribe.fit(adata, model="zinb")

# Additional dropout parameters
gate_concentration1 = zinb_results.params["gate_concentration1"]
gate_concentration0 = zinb_results.params["gate_concentration0"]
nbvcp_results = scribe.fit(adata, model="nbvcp")

# Additional capture probability parameters
p_capture_concentration1 = nbvcp_results.params["p_capture_concentration1"]
p_capture_concentration0 = nbvcp_results.params["p_capture_concentration0"]
zinbvcp_results = scribe.fit(adata, model="zinbvcp")

# Additional dropout and capture probability parameters
gate_concentration1 = zinbvcp_results.params["gate_concentration1"]
gate_concentration0 = zinbvcp_results.params["gate_concentration0"]
p_capture_concentration1 = zinbvcp_results.params["p_capture_concentration1"]
p_capture_concentration0 = zinbvcp_results.params["p_capture_concentration0"]
mix_results = scribe.fit(adata, model="nbdm", n_components=3)

# Mixing weights concentration parameters
mixing_concentration = mix_results.params["mixing_concentration"]

# Component-specific parameters have additional dimensions
# Shape: (n_components, n_genes)
r_concentration = mix_results.params["r_concentration"]

Model Comparison

To compare models, you can use the model comparison utilities:

from scribe import compare_models

# Fit multiple models
nbdm_results = scribe.fit(adata)
zinb_results = scribe.fit(adata, zero_inflation=True)

# Compare models
mc = compare_models(
    [nbdm_results, zinb_results],
    counts=adata.X,
    model_names=["NBDM", "ZINB"],
)
print(mc.summary())

# Per-gene comparison
gene_df = mc.gene_level_comparison("NBDM", "ZINB")

Best Practices

  1. Memory Management:

    • Use batch_size for large datasets
    • Generate posterior samples for specific gene subsets
    • Use compute_log_likelihood with batching for large-scale analyses
  2. Working with Parameters:

    • Access raw parameters through .params
    • Use .get_distributions() for parameter interpretation and sampling
    • Use .get_map() for point estimates
  3. Model Selection:

    • Start with the simplest model (NBDM)
    • Add complexity (zero-inflation, capture probability) as justified by data
    • Consider mixture models for heterogeneous populations
    • Use model comparison tools to select the best model
  4. Diagnostics:

    • Check loss_history for convergence
    • Use posterior predictive checks to evaluate model fit
    • For mixture models, examine entropy of component assignments