Processors
Processors are transformations that modify your dataset before or after columns are generated. They run at different stages and can reshape, filter, or augment the data.
When to Use Processors
Processors handle transformations that don't fit the "column" model: restructuring the schema for a specific output format, dropping intermediate columns in bulk, or applying batch-wide operations.
Overview
Each processor:
- Receives the complete batch DataFrame
- Applies its transformation
- Passes the result to the next processor (or to output)
Processors can run at three stages, determined by which callback methods they implement:
| Stage | When it runs | Callback method | Use cases |
|---|---|---|---|
| Pre-batch | After seed columns, before dependent columns | process_before_batch() |
Transform seed data before other columns are generated |
| Post-batch | After each batch completes | process_after_batch() |
Drop columns, transform schema per batch |
| After generation | Once, on final dataset after all batches | process_after_generation() |
Deduplicate, aggregate statistics, final cleanup |
Full Schema Available During Generation
Each batch carries the full dataset schema during generation. Post-batch schema changes such as column dropping only alter past batches, so all columns remain accessible to generators while building follow-up batches.
A processor can implement any combination of these callbacks. The built-in processors use process_after_batch() by default.
Processor Types
🗑️ Drop Columns Processor
Removes specified columns from the output dataset. Dropped columns are saved separately in the dropped-columns directory for reference.
Dropping Columns is More Easily Achieved via drop = True
The Drop Columns Processor is different from others in the sense that it does not need to be explicitly added: setting drop = True when configuring a column will accomplish the same.
Configuration:
import data_designer.config as dd
processor = dd.DropColumnsProcessorConfig(
name="remove_intermediate",
column_names=["temp_calculation", "raw_input", "debug_info"],
)
Behavior:
- Columns specified in
column_namesare removed from the output - Original values are preserved in a separate parquet file
- Missing columns produce a warning but don't fail the build
- Column configs are automatically marked with
drop=Truewhen this processor is added
Use Cases:
- Removing intermediate columns used only for LLM context
- Cleaning up debug or validation columns before final output
- Separating sensitive data from the main dataset
🔄 Schema Transform Processor
Creates an additional dataset with a transformed schema using Jinja2 templates. The output is written to a separate directory alongside the main dataset.
Configuration:
import data_designer.config as dd
processor = dd.SchemaTransformProcessorConfig(
name="chat_format",
template={
"messages": [
{"role": "user", "content": "{{ question }}"},
{"role": "assistant", "content": "{{ answer }}"},
],
"metadata": "{{ category | upper }}",
},
)
Behavior:
- Each key in
templatebecomes a column in the transformed dataset - Values are Jinja2 templates with access to all columns in the batch
- Complex structures (lists, nested dicts) are supported
- Output is saved to the
processors-outputs/{name}/directory - The original dataset passes through unchanged
Template Capabilities:
- Variable substitution:
{{ column_name }} - Filters:
{{ text | upper }},{{ text | lower }},{{ text | trim }} - Nested structures: Arbitrarily deep JSON structures
- Lists:
["{{ col1 }}", "{{ col2 }}"]
Use Cases:
- Converting flat columns to chat message format
- Restructuring data for specific model training formats
- Creating derived views without modifying the source dataset
Using Processors
Add processors to your configuration using the builder's add_processor method:
import data_designer.config as dd
builder = dd.DataDesignerConfigBuilder()
# ... add columns ...
# Drop intermediate columns
builder.add_processor(
dd.DropColumnsProcessorConfig(
name="cleanup",
column_names=["scratch_work", "raw_context"],
)
)
# Transform to chat format
builder.add_processor(
dd.SchemaTransformProcessorConfig(
name="chat_format",
template={
"messages": [
{"role": "user", "content": "{{ question }}"},
{"role": "assistant", "content": "{{ answer }}"},
],
},
)
)
Execution Order
Processors execute in the order they're added. Plan accordingly when one processor's output affects another.
Processor Plugins
You can extend Data Designer with custom processors via the plugin system. A processor plugin is a Python package that provides:
- A config class inheriting from
ProcessorConfigwith aprocessor_type: Literal["your-type"]discriminator - An implementation class inheriting from
Processorthat overrides the desired callback methods - A
Plugininstance connecting the two
Once installed, plugin processors are automatically discovered and can be used with add_processor() like built-in processors.
from my_processor_plugin.config import MyProcessorConfig
builder.add_processor(
MyProcessorConfig(
name="my_processor",
# ... plugin-specific parameters ...
)
)
Entry point configuration in pyproject.toml:
[project.entry-points."data_designer.plugins"]
my-processor = "my_plugin.plugin:my_processor_plugin"
See the plugins overview for the full guide on creating plugins.
Configuration Parameters
Common Parameters
| Parameter | Type | Description |
|---|---|---|
name |
str | Identifier for the processor, used in output directory names |
DropColumnsProcessorConfig
| Parameter | Type | Description |
|---|---|---|
column_names |
list[str] | Columns to remove from output |
SchemaTransformProcessorConfig
| Parameter | Type | Description |
|---|---|---|
template |
dict[str, Any] | Jinja2 template defining the output schema. Must be JSON-serializable. |