PepSeqPred

Residue-level epitope prediction with reproducible evidence.

PepSeqPred predicts epitope masks for protein sequences and supports full developer workflows for preprocessing, training, evaluation, and HPC execution.

Python 3.12+ ESM2 embeddings DDP-ready training Seeded evaluation snapshots

Install

pip install pepseqpred

Quickstart APIs

Pretrained API

from pepseqpred import load_pretrained_predictor

predictor = load_pretrained_predictor(
    model_id="default",
    device="auto"
)
result = predictor.predict_sequence(
    "ACDEFGHIKLMNPQRSTVWY",
    header="example_protein"
)
print(result.binary_mask)

Artifact-path API

from pepseqpred import load_predictor

predictor = load_predictor(
    model_artifact="path/to/ensemble_manifest.json",
    device="auto"
)
result = predictor.predict_sequence(
    "ACDEFGHIKLMNPQRSTVWY"
)
print(result.binary_mask)

Why PepSeqPred

PepSeqPred is built for reproducible residue-level prediction under strong class imbalance. Training and validation are documented as protocol-first; numeric scorecards are based on seeded evaluation snapshots.

Training

  • Ensemble-kfold and seeded runs with deterministic split/train seeds.
  • ID-family-aware splitting to reduce leakage risk across related proteins.
  • DistributedDataParallel support for multi-GPU HPC workflows.
  • Run artifacts include checkpoints, manifests, and run-level CSV/JSON outputs.

Validation

  • Checkpoint selection records threshold, PR-AUC, F1, MCC, AUC, and AUC10.
  • Threshold policy maximizes recall subject to minimum precision constraints.
  • Validation metrics are captured per run with explicit seed provenance.
  • Website currently shows protocol-only validation details to avoid mixed-generation scorecards.

Evaluation

  • Seeded external evaluation compares flagship models across sets 1-10.
  • Class prevalence is very low, so PR metrics are emphasized over accuracy.
  • Paired set statistics include bootstrap confidence intervals and sign tests.
  • Frozen benchmark snapshot is published with source-analysis files.

Flagship2 vs Flagship1 Results

Headline metrics below are means across seeded evaluation sets and paired deltas (flagship2 - flagship1). PR metrics are primary because the positive class is rare.

Positive-class prevalence in this snapshot: . Accuracy can look high under imbalance; PR-AUC and PR lift are more informative for epitope recovery quality.
Metric Flagship1 mean Flagship2 mean Delta (F2 - F1) 95% bootstrap CI Sign-test p
Best-fold PR-AUC (primary) to
Best-fold PR lift to
Best-fold ROC AUC to
Overall PR-AUC to

Reproducibility

PepSeqPred is fully open-source. All pipeline code, scripts, and documentation are available on GitHub.