FeaturesUse CasesBlogAPI ReferenceWhy CorePlexMLPricing
Start Free

Train production ML models in minutes

Multi-engine AutoML with H2O and FLAML. Automatically selects algorithms, tunes hyperparameters, and builds stacked ensembles — with parallel engine orchestration for Enterprise plans.

platform.coreplexml.io
AutoML experiment leaderboard with model comparison and metrics

Two engines. One experiment. Best model wins.

Run H2O and FLAML in parallel on the same dataset. Each engine explores different optimization strategies independently. CorePlexML picks the best model across all engines automatically.

H2O: 15+ algorithmsFLAML: Cost-aware sklearnEnterprise: Parallel execution
H2O
XGBoost, GBM, DL, GLM, DRF, Stacked Ensembles
FLAML
Random Forest, Extra Trees, Logistic Regression
Coming soon
AutoGluon Tabular, MLJAR-supervised, TPOT

From raw data to production model in three steps

No manual feature engineering, hyperparameter tuning, or algorithm selection. AutoML handles the entire pipeline.

1. Upload Your Data

Drag and drop CSV, Excel, JSON, or XML files. The platform automatically detects column types, identifies the target variable, and profiles your data for quality issues.

2. Configure & Train

Select your target column, problem type, and engine (H2O or FLAML). AutoML tests multiple algorithms with Bayesian hyperparameter optimization and stacked ensembles. Enterprise plans support parallel multi-engine execution.

3. Evaluate & Deploy

Review the model leaderboard with metrics, SHAP explanations, and feature importance. Deploy the best model to production with one click via MLOps.

Key Capabilities

Everything you need to get the most out of this module.

Automated Algorithm Selection

50+ algorithms tested automatically. XGBoost, GBM, deep learning, GLM, and more — the engine picks the best for your data.

Hyperparameter Tuning

Bayesian optimization finds optimal hyperparameters faster than grid or random search.

Stacked Ensembles

Combine multiple models into powerful ensembles that outperform any single algorithm.

GPU Acceleration

Leverage GPU compute for faster training on large datasets. Automatic fallback to CPU when needed.

15+ algorithms with stacked ensembles

H2O.ai evaluates XGBoost, GBM, Deep Learning, Random Forest, GLM, and Stacked Ensembles with Bayesian hyperparameter tuning.

XGBoost

Gradient-boosted decision trees optimized for speed and performance. Handles missing values natively and supports GPU acceleration.

Gradient Boosting (GBM)

Sequential ensemble method that builds trees correcting previous errors. Excellent for tabular data with complex feature interactions.

Deep Learning

Multi-layer neural networks with configurable architectures. Automatic regularization, dropout, and early stopping for production stability.

Random Forest (DRF)

Parallel ensemble of decision trees with bagging. Robust against overfitting and provides reliable feature importance rankings.

Generalized Linear (GLM)

Interpretable linear models with regularization (L1/L2). Ideal when model explainability is a regulatory requirement.

Stacked Ensembles

Meta-learner that combines predictions from all trained models. Typically achieves the best performance by leveraging model diversity.

Cost-aware optimization with scikit-learn

Microsoft FLAML finds the best model within a time budget using cost-aware search. Lightweight, fast, and ideal for quick iterations.

Random Forest (sklearn)

Scikit-learn Random Forest with automatic hyperparameter search via FLAML's cost-aware optimization. Fast convergence for tabular data.

Extra Trees

Extremely randomized trees with faster training. FLAML optimizes split thresholds and tree depth automatically for best performance.

Logistic Regression (L1)

Sparse regularized classifier via FLAML. Ideal for high-dimensional datasets where feature selection and interpretability matter.

Engine capabilities scale with your plan

Choose your engine per experiment. Enterprise plans unlock parallel multi-engine execution for maximum model coverage.

PlanEnginesPer ExperimentParallel
FreeH2O1
ProH2O, FLAML1 (choose one)
TeamH2O, FLAML1 (choose one)
EnterpriseH2O, FLAMLUp to 3✓ Up to 3
Coming soon: AutoGluon Tabular, MLJAR-supervised, and TPOT engines are in planning. The multi-engine architecture is designed to support additional engines as they become available.
Engines
H2O + FLAML (multi-engine)
Problem Types
Classification & Regression
Execution Mode
Single or Parallel
Explainability
SHAP & Feature Importance
GPU Support
CUDA with CPU fallback
Cross-Validation
Automatic k-fold
Plan Capabilities
Engine access by plan
Coming Soon
AutoGluon, MLJAR, TPOT

Train models programmatically

Use the Python SDK to automate your training pipelines. Full experiment management, from data upload to SHAP explanations.

train_model.py
from coreplexml import CorePlexMLClient

client = CorePlexMLClient(
    base_url="https://api.coreplexml.io",
    api_key="sk_your_api_key"
)

# Upload training data
dataset = client.datasets.upload(
    project_id="proj_abc",
    file_path="customers.csv",
    name="Customer Churn Data"
)

# Start AutoML training with engine selection
experiment = client.experiments.create(
    project_id="proj_abc",
    dataset_version_id=dataset["dataset_version_id"],
    target_column="churn",
    problem_type="classification",
    engine="h2o",  # or "flaml"
    config={"max_models": 20, "balance_classes": True}
)

# Wait for training to complete
result = client.experiments.wait(experiment["id"], timeout=3600)
print(f"Best model: {result['best_model_id']}")
print(f"AUC: {result['metrics']['auc']:.4f}")

# Get feature importance and SHAP values
explain = client.experiments.explain(experiment["id"])
for feat in explain["feature_importance"][:5]:
    print(f"  {feat['feature']}: {feat['importance']:.3f}")

AutoML API

RESTful endpoints for experiment management, model training, and predictions.

POST
/api/experiments

Create AutoML experiment with engine selection (h2o, flaml) and execution mode (single, parallel)

GET
/api/experiments/automl-engines

List available AutoML engines and their capabilities

GET
/api/experiments/capabilities

Get effective AutoML capabilities for the current user and plan

GET
/api/experiments/{id}/engine-runs

List engine runs for a multi-engine experiment

GET
/api/experiments/{id}/status

Check training progress and current status

GET
/api/experiments/{id}/explain

Get feature importance, SHAP values, and explainability data

GET
/api/models/engines

Get distribution of models by engine (h2o, flaml)

POST
/api/models/{id}/predict

Make predictions using any trained model (H2O or FLAML)

From experiment to model analytics

platform.coreplexml.io/experiments#new
Multi-engine AutoML training configuration with H2O and FLAML engine selection, execution mode, and plan-based capabilities

Multi-engine training: H2O + FLAML with plan capabilities

platform.coreplexml.io/experiments/.../logs
Dual engine experiment logs showing H2O and FLAML running in parallel with real-time output side by side

Parallel engine logs: H2O + FLAML running simultaneously

platform.coreplexml.io/experiments/...
Parallel multi-engine experiment with H2O and FLAML running simultaneously, showing engine runs and best model results

Parallel execution: H2O + FLAML with engine runs and results

platform.coreplexml.io/experiments#new
AutoML 4-step experiment wizard with dataset selection and training configuration

4-step AutoML experiment wizard

platform.coreplexml.io/models/...
Model performance dashboard with ROC curve, confusion matrix, and precision-recall

ROC curves, confusion matrix, and precision-recall

platform.coreplexml.io/models/.../features
SHAP values and variable importance charts for model explainability

SHAP values and variable importance

platform.coreplexml.io/models/.../gains
Cumulative gains, lift chart, and K-S statistic for model evaluation

Cumulative gains, lift, and K-S statistics

Ready to get started?

Start building with CorePlexML today. Free tier available — no credit card required.