Quickstart
In this quickstart, you’ll build a complete ML pipeline that:
- Loads the Titanic dataset
- Splits into train/test sets
- Trains a Random Forest classifier
- Evaluates accuracy and generates explanations
Time to complete: ~5 minutes
Prerequisites
Section titled “Prerequisites”Before starting, make sure you have:
- MLOps Desktop installed
- Python 3.9+ with packages:
pip install scikit-learn pandas shap
Create a Pipeline
Section titled “Create a Pipeline”-
Open MLOps Desktop
Launch the app. You’ll see an empty canvas with a toolbar at the top and a node palette on the left.
-
Add a DataLoader node
From the Components panel on the left, drag DataLoader onto the canvas.
Click the node to select it. In the node, click Browse and select a CSV file. For this tutorial, use any classification dataset (or see “Sample Dataset” below).
-
Add a DataSplit node
Drag Data Split from the Components panel onto the canvas.
Connect the DataLoader’s right handle to the DataSplit’s left handle.
Configure DataSplit:
- Test Split: 20% (default)
- Random State: 42
- Stratify: Enable, set column to
Survived(or your target)
-
Add a Trainer node
Drag Trainer onto the canvas and connect from DataSplit.
Configure Trainer:
- Mode: Train (default)
- Model Type: Random Forest Classifier
- Target Column:
Survived
-
Add an Evaluator node
Drag Evaluator onto the canvas and connect from Trainer.
No configuration needed—it auto-detects the model type.
-
Run the pipeline
Click the Run button in the toolbar.
Watch the Logs tab as each node executes:
[DataLoader] Loaded titanic.csv: 891 rows, 12 columns[DataSplit] Split: 712 train, 179 test (stratified by Survived)[Trainer] Training RandomForestClassifier...[Trainer] Training complete[Evaluator] Accuracy: 0.821, F1: 0.756 -
View results
Click the Metrics tab to see:
- Bar chart with Accuracy, Precision, Recall, F1
- Confusion matrix heatmap
Click Explain to generate:
- Feature importance chart
- SHAP beeswarm plot
- Partial dependence plots
Sample Dataset
Section titled “Sample Dataset”If you don’t have a CSV file, create the Titanic dataset:
import pandas as pdfrom sklearn.datasets import fetch_openml
titanic = fetch_openml('titanic', version=1, as_frame=True)df = titanic.framedf.to_csv("titanic.csv", index=False)print(f"Saved titanic.csv with {len(df)} rows")Or use the Iris dataset for a simpler example:
from sklearn.datasets import load_irisimport pandas as pd
iris = load_iris(as_frame=True)df = iris.framedf.to_csv("iris.csv", index=False)print("Saved iris.csv")For Iris, set Target Column to target in the Trainer.
Understanding the Results
Section titled “Understanding the Results”Metrics Tab
Section titled “Metrics Tab”| Metric | What It Means |
|---|---|
| Accuracy | Overall correctness (correct / total) |
| Precision | When we predict positive, how often are we right? |
| Recall | Of all actual positives, how many did we find? |
| F1 Score | Balance of precision and recall |
Confusion Matrix
Section titled “Confusion Matrix” Predicted Died SurvivedActual Died 98 12 Survived 21 48- Diagonal = correct predictions
- Off-diagonal = errors
Explain Section
Section titled “Explain Section”Click Explain to see why your model makes predictions:
- Feature Importance — Which features matter most
- SHAP Beeswarm — How each feature pushes predictions up/down
- Partial Dependence — How changing a feature affects predictions
Output Panel Tabs
Section titled “Output Panel Tabs”Explore the other tabs at the bottom:
| Tab | Purpose |
|---|---|
| Logs | Execution output and errors |
| Data Profile | Dataset statistics |
| Metrics | Model performance charts |
| Runs | History of all pipeline runs |
| Models | Registered models with versioning |
| Trials | Hyperparameter tuning results |
| Serving | HTTP model server |
Save Your Pipeline
Section titled “Save Your Pipeline”Click Save in the toolbar:
- Enter a name (e.g., “titanic-classifier”)
- Click Save
Your pipeline is stored locally and appears in the Load dropdown.
Next Steps
Section titled “Next Steps”Use Optuna to find optimal model settings
Explain PredictionsUnderstand models with SHAP and feature importance
Troubleshooting:
- “No Python found” — See Python Setup
- “Module not found: sklearn” — Run
pip install scikit-learn - “Target column not found” — Check column name matches exactly (case-sensitive)
- Pipeline stuck — Check the Logs tab for error messages