paint-brush
A Quick Introduction to Machine Learning with Dagsterby@stephanbv
3,969 reads
3,969 reads

A Quick Introduction to Machine Learning with Dagster

by stephabauvaMarch 17th, 2021
Read on Terminal Reader
Read this story w/o Javascript
tldt arrow

Too Long; Didn't Read

A Quick Introduction to Machine Learning with Dagster is a quick introduction to a small ML project. It is a data orchestrator for machine learning, analytics, and ETL. It lets you define pipelines in terms of the data flow between reusable, logical components, then test locally and run anywhere. Dagster can schedule and orchestrate Pandas, Spark, SQL, or anything else that Python can invoke. It makes testing easier and deploying faster. It provides a graphical representation of the execution in real time.

Company Mentioned

Mention Thumbnail

Coin Mentioned

Mention Thumbnail
featured image - A Quick Introduction to Machine Learning with Dagster
stephabauva HackerNoon profile picture

This article is a rapid introduction to Dagster using a small ML project. It is beginner-friendly but might also suit more advanced programmers if they don't know Dagster.

Directed Acyclic Graph-ster

Data processing systems typically span multiple runtime, storage, tooling, and organizational boundaries. But all the stages in a data processing system share a fundamental property; they are directed acyclic graphs (DAGs) of functional computations that consume and produce data assets.

So, what is Dagster?

Dagster is a data orchestrator for machine learning, analytics, and ETL.

It lets you define pipelines in terms of the data flow between reusable, logical components, then test locally and run anywhere. With a unified view of pipelines and the assets they produce, Dagster can schedule and orchestrate Pandas, Spark, SQL, or anything else that Python can invoke. It makes testing easier and deploying faster.

Why should you care?

One of the core promises of Dagster is that the developer can write solids that express pure business logic, separated from concerns that are specific to the execution environment. 

Code that's structured in this way is easier to develop and test locally, swapping heavy production dependencies out for lighter test systems or even fixtures and mocks. 

This leads to dramatically shorter feedback loops for developers, which in turn leads to order-of-magnitude improvements in productivity. 

Code written in this way is also easier to reuse in different contexts and more portable when it becomes necessary to change out a third-party system.

The fundamentals

To keep the focus of this article on jump-starting a project, you can read a very brief overview of 4 fundamental elements of Dagster here:

https://docs.dagster.io/tutorial/overview

Then, let’s have a look at a  simple ML program.

The application

To keep it simple, let’s create a text classifier with our old pal sci-kit.

The code as well as further explanation to execute the code can be found at https://github.com/stephanBV/ML_with_DAGs

A solid is a unit of computation that yields a stream of events (similar to a function), while a composite is a collection of solids.

Our script creates a single pipeline which:

  • processes the data
  • searches for optimal parameters between a logistic regression and a random forest
  • trains and tests the models

The data
The raw data is a collection comments from a social media platform, each comment has been labeled ‘positive’ or ‘negative’.

Dagster in action
Once we launch the Dagster’s UI, we get the following DAG:

We can see each solid in light blue and how they are connected to each other, as well as the composites in light purple.

NB: Each solid also has an ‘INFO” section which describes its inputs and outputs.

Furthermore, each composite has an “Expand” button, which let you see the set of solids it contains:

Here, you can see that the composite “process_model” contains the grid search, training, and test solids.

At the top, you can also see three inputs: result, train, and test, which are the outputs of the solids “prepare_grid_random_forest”, “split_data” with the train data, and “split_data” with the test data. 

Parameters can be in a YAML file in your repository or can be added to the Playground section, as follow:

Playground also checks your parameters against your code in real time and will let you know instantly of any errors found. 

If everything is okay, you will see at the bottom right of the window that the name of the solids with the parameters above are in a light-blue colour and the other solids are in grey colour.

As long as nothing is red, you are good to go and you can launch an execution.

Once the pipeline has finished executing, the Runs section provides the list of all executions, as follow. 

You can access the detail of all execution by clicking on its id (e.g. b29a71b4). It provides a graphical representation of the execution duration -

-as well as a table which details all the steps that happened during the process.

 An important part on that table is the step_logs which is particularly useful, especially if you are running on DEBUG mode.

For example, the following is the output of the step_logs from the solid that tests the logistic regression.

Conclusion

To keep this article simple, only one pipeline was represented, although you can have as many pipelines as necessary.

One could be specific to mine the data, another to process the data, another to test the code (with Pytest for e.g.), another could train, test, etc..

Dagster is an amazing tool to experiment, monitor, and maintain your ML pipeline, It kind of makes me think of an Open Source version of Azure ML Studio and Dataiku, but you get much more control over your project. 

We only scratched the surface of Dagster. For more information, visit https://dagster.io/