Getting Started#

All questions are welcome in the Slack channel.

Slack Status

Fugue is an abstraction layer that lets users write code in native Python or pandas and then port it over to Spark and Dask. This section will cover the motivation of Fugue, the benefits of using an abstraction layer, and how to get started. This section is not a complete reference but will be sufficient to get started with writing full workflows in Fugue.


We’ll get started by introducing Fugue and the simplest way to use it with the transform() function. The transform() function can take in a Python or pandas function and scale it out in Spark or Dask without having to modify it. This provides a very simple interface to parallelize Python and pandas code on distributed computing engines.

Type Flexibility#

After seeing an example of the transform() function, we look into the further flexibility Fugue provides by accepting functions with different input and output types. This allows users to define their logic in whatever expression makes the most sense and bring native Python functions to Spark or Dask.


Now that we have seen how functions can be written for Fugue to bring them to Spark or Dask, we look at how the transform() function can be applied with partitioning. In pandas semantics, this would be the equivalent of a groupby-apply(). The difference is partitioning is a core concept in distributed computing as it controls both logical and physical grouping of data.

Decoupling Logic and Execution#

After seeing how the transform function enables the use of Python and pandas code on Spark, we’ll see how we can apply this same principle to entire computing workflows using FugueWorkflow. We’ll show how Fugue allows users to decouple logic from execution, and introduce some of the benefits this provides. We’ll go one step further in showing how we use native Python to make our code truly independent of any framework.

Fugue Interface#

This section will cover some concepts like the Directed Acyclic Graph (DAG) and the need for explicit schema in a distributed computing environment. We’ll show how to pass parameters to Transformers, as well as load and save data. With these, users will be able to start some basic work on data through Fugue.

Joining Data#

Here we’ll show the different ways to join DataFrames in Fugue along with union, intersect, and except. SQL and pandas also have some inconsistencies that users should be aware of when joining. Fugue maintains consistency with SQL (and Spark).


We already covered the transformer, the most commonly used Fugue extension. Extensions are Fugue operations on DataFrames that are used inside the DAG. Here we will cover the creator, processor, cotransformer, and outputter.

Distributed Computing#

The heart of Fugue is distributed computing. In this section, we’ll show the keywords and concepts that allow Fugue to fully utilize the power of distributed computing. This includes partitions, persisting, and broadcasting.


We’ll show a bit of FugueSQL, the SQL interface for using Fugue. This is targeted at heavy SQL users and SQL-lovers who want to use SQL on top of Spark and Dask, or even pandas. FugueSQL is used on DataFrames in memory as opposed to data in databases.

With that, you should be ready to implement data workflows using Fugue.

For full end-to-end examples, check out the Stock Sentiment and COVID-19 examples.

For any questions, feel free to join the Slack channel.