Skip to content

Security-FIT/Latium

Repository files navigation

Latium Framework

Running ROME

ROME (and related commands) is driven via the Hydra-based CLI in src/cli.py.

Single intervention:

python -m src.cli command=rome model=gpt2-medium

Batch evaluation:

python -m src.cli command=batch-rome model=gpt2-medium

Compute second-moment statistics (required before running ROME on a new model):

python -m src.cli command=second-moment model=gpt2-medium

The default config is at src/config/config.yaml. Override any value on the command line using Hydra syntax (e.g. model=gpt2-large).

Alternatively, use the console fallback (no Hydra overhead):

python -m src.cli --console rome --config src/config/config.yaml

Running Causal Trace

python -m src.cli command=causal-trace model=gpt2-medium

To inspect the computed noise multiplier without running a full trace:

python -m src.cli command=compute-multiplier model=gpt2-medium

Remote Covariance Pipeline

covariance_a100_remote.sh computes second-moment statistics on a remote GPU node (e.g. A100) and pulls the resulting artifacts back locally.

# Run with default models (deepseek-7b-base, granite4-micro, llama2-7b, mistral-7b-v0.1, mistral-7b-v0.3):
./covariance_a100_remote.sh user@gpu-host

# Override models:
MODEL_KEYS="gpt2-xl gpt-j-6b" ./covariance_a100_remote.sh user@gpu-host /path/to/Latium optim latium

# Arguments: <user@host> [remote_repo_path] [remote_branch] [conda_env]

The script syncs model configs and src/rome/common.py to the remote, runs covariance computation per model, and downloads the .pt artifacts into second_moment_stats/.


Layer Selection Heuristic

src/causal_trace/layer_heuristic.py recommends the best MLP layer for ROME edits using multiple signals (causal trace, weight norms, spectral gap, architectural prior).

# CSV-only (no GPU needed):
python -m src.causal_trace.layer_heuristic \
    --csvs analysis_out/causal_trace_deepseek*.csv \
    --num-layers 30

# Full analysis (GPU + model):
python -m src.causal_trace.layer_heuristic \
    --model deepseek-ai/deepseek-llm-7b-base \
    --layer-template 'model.layers.{}.mlp.down_proj' \
    --num-layers 30 \
    --csvs analysis_out/causal_trace_deepseek*.csv

Running the Structural Benchmark

structural_benchmark.py applies ROME edits across a dataset and evaluates all structural detectors (MSD, blind MSD, spectral, IPR) on the modified weights. Results are written as JSON to analysis_out/.

python structural_benchmark.py \
    --model gpt2-large \
    --n-tests 30 \
    --start-idx 0 \
    --output-dir ./analysis_out \
    --spectral-top-k 50 \
    --trim-first-layers 2 \
    --trim-last-layers 2 \
    --spectral-neighbor-layers 1

Key arguments:

Argument Default Description
--model gpt2-large Model name (must match a config in src/config/model/)
--n-tests 30 Number of ROME edits to benchmark
--start-idx 0 Starting index in the facts dataset
--output-dir ./analysis_out Directory for JSON result files
--spectral-top-k 50 Top-K singular values used by the spectral detector
--trim-first-layers 2 Layers to exclude from the head of the model
--trim-last-layers 2 Layers to exclude from the tail of the model
--n-prompts auto Number of ROME prefix prompts (scales with model size if omitted)

Detection Documentation

Detailed documentation for the detection methods is in the docs/ directory:

  • docs/structural-docs.md - structural detector metrics (L2 discrepancy, relative discrepancy, directional coherence, MSD, IPR, etc.)
  • docs/spectral-docs.md - spectral detector signals and the mathematics behind singular-value z-scores and ratio scores

Models roadmap


Supported Models Causal Trace Weight intervention Success Rate (n=50) (hard test) Notes
gpt2-medium ✔️ ✔️ works
gpt2-large ✔️ ✔️ works
gpt2-xl ✔️ ✔️ works
gpt-j-6b ✔️ ✔️ works
qwen3-0.6b ✔️ ✔️
qwen3-1.7b ✔️ ✔️
qwen3-4b ✔️ ✔️
qwen3-8b ✔️ ✔️ 1.00
granite4-micro ✔️ ✔️ 0.70 Weird architecture
mistral-7b-v0.1 ✔️ ✔️ 0.92
mistral-7b-v0.3 ✔️ ✔️ 0.94
llama2-7b ✔️ ✔️ 0.56 Weird architecture
falcon-7b ✔️ ✔️ 0.96
opt-6.7b ✔️ ✔️ 0.96
deepseek-7b-base ✔️ ✔️ 0.90
llama3 planned
gpt-neo planned
qwen2.5 planned
baichuan planned
chatglm planned
t5 planned

Pipeline Script

pipeline.sh is a simple all-in-one script for running benchmarks locally or on a remote cluster.

# ROME-only, N=20 (default)
bash pipeline.sh

# Run on remote cluster with env setup
bash pipeline.sh --remote ubuntu@132.145.129.234 --setup-env

# Structural benchmark, N=30
bash pipeline.sh --structural --n 30

# Compute covariance first, then benchmark
bash pipeline.sh --compute-cov --n 10

# Specific models only
bash pipeline.sh --models gpt2-xl mistral-7b-v0.1 --n 5
Flag Default Description
--compute-cov off Compute covariance matrices (otherwise uses existing)
--n <int> 20 Number of test edits per model
--structural off Run structural benchmark (default: ROME-only)
--setup-env off Set up conda env + deps on cluster from scratch
--remote <host> (local) SSH target for remote execution
--models <m1 ..> all 13 Override model list
--output-dir <path> ./pipeline_out Output directory

Error codes:


Error code Name of the error Description
1 Help Help invoked. Typically caused by incorrect script usage.
2 Resource already exists Trying to create a resource that already exists.
-1 Unknown An unknow error. Create GitHub issue with the reproduction steps

About

ROME and its surroundings

Topics

Resources

License

Stars

Watchers

Forks

Contributors