Skip to content

Simple, debuggable, step-by-step pipelines for Python

License

Notifications You must be signed in to change notification settings

marek-kan/linepipe

Repository files navigation

linepipe

linepipe sits between ad-hoc Python scripts and full orchestration frameworks. It is best suited for batch pipelines, feature engineering, and reproducible data workflows where simplicity and clarity matter more than scale.

It focuses on:

  • clear data dependencies
  • reproducible execution (linear pipelines)
  • fast iteration and debugging thanks to (local / in-memory) caching
  • simple mental models

Example of a pipeline:

Each node consumes named inputs and produces named outputs, which are stored in an internal registry and made available to downstream nodes.

pipeline = Pipeline(
    nodes=[
        Node(
            func=nodes.create_db_engine,
            inputs=["config.db"],
            outputs=["conn"],
        ),
        Node(
            func=nodes.load_data,
            inputs=["conn"],
            outputs=["df"],
        ),
        Node(
            func=nodes.preprocess_data,
            inputs=["df", "config.preprocessing"],
            outputs=["df_processed"],
        ),
        Node(
            func=nodes.feature_engineering,
            inputs=["df_processed", "config.feature_engineering"],
            outputs=["df_features"],
        ),
        Node(
            func=nodes.train,
            inputs=["df_features", "config.model_params"],
            outputs=["model"],
        ),
    ],
    config=config
)

pipeline.run()

A minimal end-to-end example is available as a Jupyter notebook:

  • examples/basic_usage.ipynb

It walks through:

  • defining nodes
  • building a pipeline
  • running it and inspecting outputs

Installation

pip install linepipe

Optionally:


Why linepipe?

Many pipeline tools are powerful but come with:

  • large configuration surfaces
  • implicit data catalogs
  • steep learning curves

linepipe intentionally keeps things simple:

  • Explicit > implicit
  • Minimal surface area
  • Easy to reason about
  • Easy to debug

Non-goals

  • Distributed execution
  • Dynamic DAG scheduling
  • Dataset catalogs
  • Orchestration / scheduling

linepipe is a building block, not a platform.


Core concepts

Node

A Node wraps a Python callable and declares:

  • which inputs it consumes
  • which outputs it produces
from linepipe.node import Node

def add(a, b):
    return a + b

node = Node(
    func=add,
    inputs=["x", "y"],
    outputs=["sum"],
)

Nodes are executed sequentially inside a Pipeline.


Pipeline

A Pipeline executes a list of Nodes in order, resolving inputs from:

  1. configuration
  2. cached / in-memory outputs
  3. runtime objects
from linepipe.pipeline import Pipeline

pipeline = Pipeline(
    nodes=[node],
    config={},
    x=2,
    y=3,
)

pipeline.run()

Intermediate results can be cached automatically.

Object Registry

During execution, linepipe maintains an internal object registry that:

  • stores intermediate node outputs
    • optionally persists them to disk
  • holds runtime-only objects (e.g. DB connections)
  • injects runtime constants into the pipeline

The registry is created per pipeline run and closed automatically after run is finished. Users can reopen in with pipeline.get_obj_registry().


Inputs and outputs

Resolution rules:

  1. Inputs starting with config. or equal to config are resolved from the pipeline configuration.
  2. All other inputs are resolved from the pipeline's internal object registry (which may contain persisted or in-memory objects)
  3. Runtime constants are injected into the registry at pipeline creation time

Name collisions

Each name in a pipeline must uniquely refer to a single object.

If a name exists both as:

  • a runtime constant, and
  • a persisted cached object

the pipeline will fail fast with an error.

This prevents subtle bugs caused by stale cache values overriding runtime state.

Explicit data flow

Each node declares string-named inputs and outputs:

Node(
    func=transform,
    inputs=["raw_data"],
    outputs=["features"],
)

Config injection

Configuration values can be injected using config.<attr> or you can pass whole config object with "config":

class Config:
    window = 5

Node(
    func=rolling_average,
    inputs=["data", "config.window"],
    outputs=["features"],
)

Caching

Persistent cache

You can turn on a disk-backed cache (shelve) to store intermediate results with use_persistent_cache=True.

pipeline = Pipeline(
    nodes=nodes,
    config=config,
    cache_storage_path="./.cache/pipeline.db",
    use_persistent_cache=True,
)
  • Cached outputs can be reused across runs, as long as output names do not conflict with runtime constants.
  • When persistent caching is enabled, cached objects from previous runs are automatically restored at startup.
  • The cache is closed automatically at the end of execution (i.e. pipeline.run())

You can reopen it for inspection:

ds = pipeline.get_obj_registry()
# placement of registred objects:
print(registry.placement)
registry.close()

Side-effect nodes (writers)

Nodes may have no outputs, useful for:

  • database writes
  • file exports
  • logging
def write_to_db(df):
    ...

Node(
    func=write_to_db,
    inputs=["features"],
    outputs=[],
)

Named partial functions

linepipe provides a helper to adapt generic functions into distinct pipeline nodes.

from linepipe.node import create_named_partial_function
from my_package.pipelines.outputs import nodes

write_val_metrics = Node(
    func=create_named_partial_function(
        func=nodes.write_pipeline_output,
        func_name="write_val_metrics",
        table_name="metrics",
        schema="ml"
    ),
    inputs=["val_metrics"],
    outputs=[],
)

write_predictions = Node(
    func=create_named_partial_function(
        func=nodes.write_pipeline_output,
        func_name="write_predictions",
        table_name="predictions",
        schema="ml"
    ),
    inputs=["predictions"],
    outputs=[],
)

This:

  • pre-binds keyword arguments
  • gives the node a stable, readable name
  • improves logging, history, and introspection

Pipeline composition

Pipelines can be combined using +:

full_pipeline = feature_pipeline + writer_pipeline
full_pipeline.run()

Composition:

  • preserves execution order
  • prevents output name collisions
  • merges runtime constants and configuration
  • Gives you pipeline reusability, for example, same feature-generation pipeline in train and predict

Pipeline visualization

linepipe.viz provides simple, no-dependency ascii graph drawing function draw_ascii_pipeline. It draws nodes in order they are defined along with their inputs and outputs - good for quick check.

from linepipe import viz

print(viz.draw_ascii_pipeline(pipeline))

Output can look like this:

Pipeline (5 nodes)

[load_data]
  outputs:
    - raw_data
      |
      v
[clean_data]
  inputs:
    - raw_data
  outputs:
    - clean_data
      |
      v
[rolling_goals]
  inputs:
    - clean_data
    - config.window
    - config.min_periods
  outputs:
    - rolling_features
      |
      v
[merge]
  inputs:
    - clean_data
    - rolling_features
  outputs:
    - features
      |
      v
[write_features]
  inputs:
    - features
  (no outputs)

If you need something more visual for documentation purposes you can install optional dependency - plotly (pip install linepipe[plot]) and use plot_pipeline function - however, this function is experimental. Graph of pipeline above would look like:

viz.plot_pipeline_graph(full_pipeline)

Pipeline plot


History tracking

Enable execution history for debugging:

pipeline = Pipeline(
    nodes=nodes,
    config=config,
    track_history=True,  # in-memory persistent after `run()` is finished
    use_persistent_cache=False,  # (if True) disk persistent after `run()` is finished
)

After execution:

pipeline.history

Each entry contains:

  • node name
  • resolved inputs
  • outputs

Inputs and outputs are deep-copied at execution time to preserve snapshots. For large objects, this can significantly increase memory usage.


[Optional] Memory profiling

If you wish to profile your nodes you can install optional memory_profiler dependency with pip install linepipe[memory]. This gives you ability to log memory usage during run of a node. For example:

pipeline = Pipeline(
    nodes= [
        Node(
            func=nodes.expensive_function,
            inputs=["df"],
            outputs=["df_processed"],
            profile=True
        )
    ]
)
pipeline.run()

This will produce logs like:

[expensive_function] ΔMem: 0.94 MiB | Peak: 0.94 MiB

Structure is [node_name] ΔMem: {float} MiB | Peak: {float} MiB

  • ΔMem: delta between first and last recorded value
  • Peak: delta between maximum recorded value and first recorded value

License

MIT License.

About

Simple, debuggable, step-by-step pipelines for Python

Resources

License

Stars

Watchers

Forks

Packages

No packages published

Languages