Pathway provides a Python interface and experience created with data developers in mind. You can easily build pipelines by manipulating Pathway tables and rely on the vast resources and libraries of the Python ecosystem. Also, Pathway can seamlessly be integrated into your CI/CD chain as it is inherently compatible with popular tools such as mypy or pytest. Your Pathway pipelines can be automatically tested, built, and deployed, like any other Python workflow.
Pathway also relies on a powerful Rust engine to ensure high performance for your pipelines, no matter if you are dealing with batch or streaming data. Pathway engine makes the utmost of Rust speed and memory safety to provide efficient parallel and distributed processing without being limited by Python's GIL.
Pathway takes the best of both worlds and efficiently associates the convenience of Python with the power of Rust.
Pathway's engine incrementally processes data updates. This means that the minimum work needed by any algorithm or transformation is performed to refresh its results when fresh data arrives.
As a Pathway user, you write code in Python, with Pathway imported as a Python module. Pathway provides a cookiecutter template to help start Pathway projects.
The same code developed with the Pathway module can be used for batch and streaming use cases, depending only on data connector settings and method of launching. For many operations, Pathway returns exactly the same results when run in batch and streaming modes. Exceptions are out-of-order data for which streaming mode may optionally ignore and functions that explicitly depend on processing time.
During rapid prototyping, Pathway code can be written and tested interactively, without waiting for compilation or deployment, with the computation graph being built in the background. This is helpful for instance during data science work in Jupyter notebooks.
The developed code prototype can then run on streaming data sources.
The computation graph is handed down to the runtime engine when executing the line
In terms of the interplay between interactive usability and launching compute graphs, Pathway takes direct inspiration from TensorFlow and PyTorch: just like TensorFlow, we explicitly represent computations as graphs that are executed with a
run() command, however, similarly to PyTorch, we aim to offer the best in class interactive work environment for hands-on, data-driven algorithm design.
(Footnote: Our advisor and Business Angel, Lukasz Kaiser, was a co-author of TensorFlow, and our CTO, Jan Chorowski, was a contributor to Theano).
Pathway calls into the same Rust runtime in any deployment - on either streaming data sources or in interactive mode.
Pathway computes consistent results: each produced output is computed based on data contained in prefixes of the input data streams. For most of the operations, exactly the same outputs are produced:
- If the input data was sent to the engine all at once in batch mode.
- If the inputs were sent in several smaller batches, or in streaming mode.
This equivalence of stream and batch processing facilitates easy development of data processing logic: one can reason in terms of steady-state and consistent results and not worry about all intermediate and wrong states through which an eventually consistent system may pass.
Further details about consistency, coupled with a practical example, can be found here.
Pathway is meant to be deployed in a containerized manner.
Single-machine deployments can easily be achieved using Docker. The deployment can run concurrently on multiple cores using multiple processes or threads.
We provide a
pathway spawn command to aid in launching multi-process and multi-threaded jobs.
The choice between threads and multiple processes depends on the nature of the computation. While communication between threads is faster, Python-heavy workloads may require multiprocess parallelism to bypass the GIL.
For using Pathway on large workloads beyond a single machine, see Distributed deployment.
Pathway tests on offline data snippets can be run locally in any CI/CD pipeline with Python. Tests can cover the handling of temporal (late, out of order) aspects of data by comparing results on multiple revisions. Pathway supports several session-replay mechanisms, such as the demo API. These allow recreating streaming scenarios predictably within standard CI/CD pipelines (Jenkins, GitHub Actions, etc.)
Get realtime alerts with changes in Google Docs using Pathway and LLMs.
Perform signal processing on data streams in realtime, even with out-of-order data.
Monitor your server logs, process your logs, and do realtime anomaly detection.