evaluator
evaluator
¶
Evaluator entry point for synthetic data evaluation.
Orchestrates metric computation and report assembly by delegating
to MultimodalReport and collecting timing information.
Classes:
| Name | Description |
|---|---|
Evaluator |
Orchestrates evaluation of synthetic data against reference data. |
Evaluator(config, generate_results, pii_replacer_time=None, column_statistics=None, train_df=None, test_df=None, workdir=None)
¶
Orchestrates evaluation of synthetic data against reference data.
Computes quality and privacy metrics by delegating to
MultimodalReport, which assembles individual evaluation
components (distribution stability, correlation, PCA, text
similarity, privacy scores, etc.).
Parameters:
| Name | Type | Description | Default |
|---|---|---|---|
config
|
SafeSynthesizerParameters
|
Pipeline configuration controlling which metrics are enabled. |
required |
generate_results
|
GenerateJobResults | DataFrame
|
Synthetic output -- either a |
required |
pii_replacer_time
|
float | None
|
Wall-clock seconds spent on PII replacement, if any. |
None
|
column_statistics
|
dict[str, ColumnStatistics] | None
|
Per-column PII entity counts and transform metadata. |
None
|
train_df
|
DataFrame | None
|
Reference (training) dataframe. |
None
|
test_df
|
DataFrame | None
|
Holdout (test) dataframe used by text-similarity and privacy metrics. |
None
|
workdir
|
Workdir | None
|
Working directory for persisting artifacts. |
None
|
Methods:
| Name | Description |
|---|---|
evaluate |
Run all configured evaluation components and store results. |
Source code in src/nemo_safe_synthesizer/evaluation/evaluator.py
evaluate()
¶
Run all configured evaluation components and store results.
Populates self.report with the completed MultimodalReport
and self.evaluation_time with the elapsed wall-clock seconds.